Skip to content

Commit c2b684d

Browse files
committed
style
1 parent 9764cba commit c2b684d

File tree

2 files changed

+4
-3
lines changed

2 files changed

+4
-3
lines changed

tests/models/cohere2/test_modeling_cohere2.py

+4-2
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,6 @@
2626
require_flash_attn,
2727
require_read_token,
2828
require_torch,
29-
require_torch_gpu,
3029
require_torch_large_gpu,
3130
slow,
3231
torch_device,
@@ -285,7 +284,10 @@ def test_export_static_cache(self):
285284
if version.parse(torch.__version__) < version.parse("2.5.0"):
286285
self.skipTest(reason="This test requires torch >= 2.5 to run.")
287286

288-
from transformers.integrations.executorch import TorchExportableModuleWithStaticCache, convert_and_export_with_cache
287+
from transformers.integrations.executorch import (
288+
TorchExportableModuleWithStaticCache,
289+
convert_and_export_with_cache,
290+
)
289291

290292
model_id = "CohereForAI/c4ai-command-r7b-12-2024"
291293
EXPECTED_TEXT_COMPLETION = [

tests/models/glm/test_modeling_glm.py

-1
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,6 @@
2323
is_flaky,
2424
require_flash_attn,
2525
require_torch,
26-
require_torch_accelerator,
2726
require_torch_large_gpu,
2827
require_torch_sdpa,
2928
slow,

0 commit comments

Comments
 (0)