Skip to content

Conversation

@hmellor
Copy link
Member

@hmellor hmellor commented Nov 28, 2025

RoPE config went through a secondary refactor in huggingface/transformers#42255 which:

  • Deletes standardize_rope_params and rope_config_validation
  • Adds a mixin which is inherited by PreTrainedConfig which does RoPE config patching and validation automatically

Therefore, for Transformers v5 compatibility, this PR:

  • Removes any imports of these deleted methods
  • Removes manual RoPE config patching for Tranformers v5 in patch_rope_parameters as it's now handled inside PreTrainedConfig

@chatgpt-codex-connector
Copy link

Codex usage limits have been reached for code reviews. Please check with the admins of this repo to increase the limits by adding credits.

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request correctly addresses the breaking changes related to RoPE configuration in Transformers v5. The removal of manual patching for transformers>=5.0.0 and the deletion of calls to the now-removed rope_config_validation function align with the library's new automatic handling of RoPE parameters. The changes are clean, well-targeted, and ensure compatibility with upcoming Transformers releases.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant