diff --git a/tests/quantization/ggml/test_ggml.py b/tests/quantization/ggml/test_ggml.py index 1a9c126f3d0445..0b70b4c3d853bd 100644 --- a/tests/quantization/ggml/test_ggml.py +++ b/tests/quantization/ggml/test_ggml.py @@ -630,9 +630,9 @@ def test_falcon7b_q2_k(self): ) text = tokenizer(self.example_text, return_tensors="pt")["input_ids"].to(torch_device) - out = model.generate(text, max_new_tokens=10) + out = model.generate(text, max_new_tokens=16) - EXPECTED_TEXT = "Hello All,\nI am new to this forum." + EXPECTED_TEXT = 'Hello,\nI am trying to use the "get_post_meta"' self.assertEqual(tokenizer.decode(out[0], skip_special_tokens=True), EXPECTED_TEXT) def test_falcon7b_weights_conversion_fp16(self):