Skip to content

Commit

Permalink
Fix expected output for ggml test (#35686)
Browse files Browse the repository at this point in the history
fix expected output
  • Loading branch information
MekkCyber authored Jan 14, 2025
1 parent 0506365 commit df2a812
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions tests/quantization/ggml/test_ggml.py
Original file line number Diff line number Diff line change
Expand Up @@ -630,9 +630,9 @@ def test_falcon7b_q2_k(self):
)

text = tokenizer(self.example_text, return_tensors="pt")["input_ids"].to(torch_device)
out = model.generate(text, max_new_tokens=10)
out = model.generate(text, max_new_tokens=16)

EXPECTED_TEXT = "Hello All,\nI am new to this forum."
EXPECTED_TEXT = 'Hello,\nI am trying to use the "get_post_meta"'
self.assertEqual(tokenizer.decode(out[0], skip_special_tokens=True), EXPECTED_TEXT)

def test_falcon7b_weights_conversion_fp16(self):
Expand Down

0 comments on commit df2a812

Please sign in to comment.