Skip to content

[BUG] @ruvector/ruvllm SIMD Inference produces garbled output - LLM not functional #103

@stuinfla

Description

@stuinfla

Summary

The @ruvector/ruvllm package advertises built-in SIMD inference capabilities, but the LLM generation produces garbled, unusable output. Users expecting an integrated LLM solution are finding they still need external providers (Ollama/APIs).

Environment

  • Package: @ruvector/ruvllm v0.2.4
  • Platform: darwin-arm64 (Apple Silicon)
  • Node.js: v22.13.1
  • SIMD Support: Yes (NEON detected)

Steps to Reproduce

# Check info shows SIMD is loaded
npx @ruvector/ruvllm info
# Output: Native Module: Loaded, SIMD Support: Yes, Capabilities: NEON

# Attempt to generate text
npx @ruvector/ruvllm generate "Hello, how are you?"

Expected Behavior

Coherent text generation using the built-in SIMD inference, or a clear error message indicating an external LLM provider is required.

Actual Behavior

Garbled, unusable output:

=xmanybyon=``=on)&G&}=#)=wasmany!N=)do<5anduse`55towass>binorxOG6!manywithbytoNhowto5O<=6howOandtoashowNasNO=i<useouthadtop,byxwasG5kinhowmwasinvmandtoto<out5`withO@andfromin`~sx,totoandO<<@beinGk!use,`had<hadp!6]@i{from@use,`as]{`M~haveexG6p>wask\

Additionally, npx @ruvector/ruvllm stats throws an error:

Error: Cannot read properties of undefined (reading 'toFixed')

Impact

This creates confusion for users who expect ruvllm to include a working LLM. The package description states:

"Self-learning LLM orchestration with SONA adaptive learning, HNSW memory, FastGRNN routing, and SIMD inference"

Users interpret "SIMD inference" as meaning the package can perform LLM inference without external dependencies. Currently:

  • They discover this only after attempting generation
  • There's no clear documentation stating Ollama or an API is required
  • The garbled output suggests a broken model rather than a missing dependency

Requested Fix

  1. If SIMD inference is meant to work standalone:

    • Fix the model loading/inference pipeline
    • Ensure coherent output from built-in model
  2. If external LLM is required:

    • Update package description to clarify this
    • Add clear error message: "No LLM configured. Please install Ollama or set ANTHROPIC_API_KEY"
    • Document the requirement in README
  3. Fix the stats command - currently throws undefined error

Additional Context

The HNSW memory, FastGRNN routing, and embedding features work correctly. Only the LLM generation/inference is broken.

Related: ruvector-postgres ONNX embeddings work perfectly (all-MiniLM-L6-v2, BGE models, etc.). The issue is specifically with text generation in ruvllm.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions