Skip to content

3.3.1 版本 sft 微调internvl3 出现问题。 #4044

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
Kyrie666 opened this issue Apr 30, 2025 · 2 comments
Closed

3.3.1 版本 sft 微调internvl3 出现问题。 #4044

Kyrie666 opened this issue Apr 30, 2025 · 2 comments

Comments

@Kyrie666
Copy link

rank1: File "/usr/local/lib/python3.11/site-packages/transformers/modeling_utils.py", line 4455, in from_pretrained
rank1: ) = cls._load_pretrained_model(
rank1: ^^^^^^^^^^^^^^^^^^^^^^^^^^^
rank1: File "/usr/local/lib/python3.11/site-packages/transformers/modeling_utils.py", line 4841, in _load_pretrained_model
rank1: caching_allocator_warmup(model_to_load, expanded_device_map)
rank1: File "/usr/local/lib/python3.11/site-packages/transformers/modeling_utils.py", line 5829, in caching_allocator_warmup
rank1: re.compile("|".join([re.escape(plan) for plan in model._tp_plan]))
rank1: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

这个问题曾经遇到过,当时的解决方案是将transformers 固定为4.50.2版本,请问你们内部测试这个模型的时候使用的事哪个版本

@Jintao-Huang
Copy link
Collaborator

@Kyrie666
Copy link
Author

huggingface/transformers#37663
fixed .thanks

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants