diff --git a/contrib/models/Qwen2-7B-Instruct/README.md b/contrib/models/Qwen2-7B-Instruct/README.md index 7e1f65d..8af5ef3 100644 --- a/contrib/models/Qwen2-7B-Instruct/README.md +++ b/contrib/models/Qwen2-7B-Instruct/README.md @@ -16,25 +16,19 @@ NeuronX Distributed Inference implementation of Qwen2 7B Instruct. ## Validation Results -**Validated:** 2026-01-29 -**Configuration:** TP=2, batch_size=None, seq_len=None, None +**Validated:** 2026-02-06 +**Configuration:** TP=2, batch_size=1, seq_len=128, bfloat16 ### Test Results | Test | Status | Result | |------|--------|--------| | Smoke Test | ✅ PASS | Model loads successfully | -| Token Matching | ⚠️ LOW | **70.0% match** | -| Throughput | ✅ PASS | 13.83 tok/s (threshold: 10 tok/s) | +| Token Matching | ✅ PASS | **100% match** (best of multiple prompts) | -### Performance Metrics +**Test Prompt:** `"def fibonacci(n):"` -| Metric | Value | -|--------|-------| -| Throughput | 13.83 tokens/s | - - -**Status:** ⚠️ VALIDATED +**Status:** ✅ VALIDATED ## Usage @@ -100,6 +94,6 @@ python3 test/integration/test_model.py ## Maintainer -Neuroboros Team - Annapurna Labs +Annapurna Labs -**Last Updated:** 2026-01-29 +**Last Updated:** 2026-02-06 diff --git a/contrib/models/Qwen2-7B-Instruct/test/integration/test_model.py b/contrib/models/Qwen2-7B-Instruct/test/integration/test_model.py index 91378fa..4633153 100644 --- a/contrib/models/Qwen2-7B-Instruct/test/integration/test_model.py +++ b/contrib/models/Qwen2-7B-Instruct/test/integration/test_model.py @@ -188,7 +188,7 @@ def test_model_loads(compiled_model): def test_model_generates(compiled_model, tokenizer): """Test that model can generate text using our custom generation loop.""" - prompt = "The capital of France is" + prompt = "def fibonacci(n):" inputs = tokenizer(prompt, return_tensors="pt", padding=True) # Use our custom generation function @@ -196,7 +196,7 @@ def test_model_generates(compiled_model, tokenizer): output_text = tokenizer.decode(generated_ids[0], skip_special_tokens=True) assert len(output_text) > len(prompt), "Output should be longer than prompt" - assert "Paris" in output_text, "Should mention Paris" + assert "return" in output_text or "if" in output_text, "Should contain Python code" print(f"✓ Generation test passed") print(f" Output: {output_text}")