What's Changed
- ♨️ [GRPO] Fix potential hang in
get_high_entropy_mask
by @akakakakakaa in #4041 - Aux loss is already included in the loss returned by Transformers by @pramodith in #4078
- Fix get_peft_model() so that prepare_model_for_kbit_training does not reapply to an instance of PeftModel, thus freezing all the layers by @Hoesu in #4081
- 🐯 fix: use_liger_kernel with IterableDataset by @jue-jue-zi in #4087
- [SFTrainer]: Fix DFT Loss by @pramodith in #4112
- ⚡ Fix Flash Attention x Padding-Free loss by @qgallouedec in #4170
New Contributors
Full Changelog: v0.23.0...v0.23.1