-
Notifications
You must be signed in to change notification settings - Fork 342
Description
Summary
The @ruvector/ruvllm package advertises built-in SIMD inference capabilities, but the LLM generation produces garbled, unusable output. Users expecting an integrated LLM solution are finding they still need external providers (Ollama/APIs).
Environment
- Package: @ruvector/ruvllm v0.2.4
- Platform: darwin-arm64 (Apple Silicon)
- Node.js: v22.13.1
- SIMD Support: Yes (NEON detected)
Steps to Reproduce
# Check info shows SIMD is loaded
npx @ruvector/ruvllm info
# Output: Native Module: Loaded, SIMD Support: Yes, Capabilities: NEON
# Attempt to generate text
npx @ruvector/ruvllm generate "Hello, how are you?"Expected Behavior
Coherent text generation using the built-in SIMD inference, or a clear error message indicating an external LLM provider is required.
Actual Behavior
Garbled, unusable output:
=xmanybyon=``=on)&G&}=#)=wasmany!N=)do<5anduse`55towass>binorxOG6!manywithbytoNhowto5O<=6howOandtoashowNasNO=i<useouthadtop,byxwasG5kinhowmwasinvmandtoto<out5`withO@andfromin`~sx,totoandO<<@beinGk!use,`had<hadp!6]@i{from@use,`as]{`M~haveexG6p>wask\
Additionally, npx @ruvector/ruvllm stats throws an error:
Error: Cannot read properties of undefined (reading 'toFixed')
Impact
This creates confusion for users who expect ruvllm to include a working LLM. The package description states:
"Self-learning LLM orchestration with SONA adaptive learning, HNSW memory, FastGRNN routing, and SIMD inference"
Users interpret "SIMD inference" as meaning the package can perform LLM inference without external dependencies. Currently:
- They discover this only after attempting generation
- There's no clear documentation stating Ollama or an API is required
- The garbled output suggests a broken model rather than a missing dependency
Requested Fix
-
If SIMD inference is meant to work standalone:
- Fix the model loading/inference pipeline
- Ensure coherent output from built-in model
-
If external LLM is required:
- Update package description to clarify this
- Add clear error message: "No LLM configured. Please install Ollama or set ANTHROPIC_API_KEY"
- Document the requirement in README
-
Fix the stats command - currently throws undefined error
Additional Context
The HNSW memory, FastGRNN routing, and embedding features work correctly. Only the LLM generation/inference is broken.
Related: ruvector-postgres ONNX embeddings work perfectly (all-MiniLM-L6-v2, BGE models, etc.). The issue is specifically with text generation in ruvllm.