Add max_eval_batches argument to TrainingArguments #41524
+31
−0
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Add max_eval_batches argument to TrainingArguments
Description
Adds a
max_eval_batches
parameter toTrainingArguments
that allows users to limit the number of batches used during evaluation.Fixes #31561
Motivation
When working with large evaluation datasets, running evaluation on the entire dataset can be very slow. During development, hyperparameter tuning, or quick iteration, it's often sufficient to evaluate on a subset of the data.
This is similar to PyTorch Lightning's
limit_val_batches
parameter.Changes
max_eval_batches
parameter toTrainingArguments
Trainer.evaluation_loop
Usage Example