-
Notifications
You must be signed in to change notification settings - Fork 2.2k
Add DataLoader to MathQA benchmarking script #2954
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Add DataLoader to MathQA benchmarking script #2954
Conversation
|
Thanks for the PR. Note that this may not be quite as easy to do anymore as we added a peft/method_comparison/MetaMathQA/utils.py Line 296 in c5a905d
Honestly, I'm not sure if adding the standard torch You could give it a try and see if you can simplify/improve the training code this way, but I think the expected value of working on this wouldn't be too big. |
|
So, in the open tasks mentioned, which of the pending tasks are considered to have a higher priority? cc: @BenjaminBossan |
|
Hmm, good question. Perhaps it would be a good idea to remove the Among the open tasks, I think that 1. ensuring that we get similar results with Another way to contribute would be by adding experiment settings, see the discussions at the end of #2310. This requires access to hardware that can run the experiments at a decent speed. |
|
Can the vLLM addition and clean-up of prints also be considered testing issues. Would you like me to open a seperate PR specifying which are test issues and code issues in the readme? cc: @BenjaminBossan |
Regarding vLLM, over time we have sped up generation times quite well, so it's less of an issue now. As mentioned, for testing, it's still a bit on the slow side. I think the main value of adding vLLM would be more of a learning experience, I'm not sure if it's worth it to add it as a dependency. Regarding the prints, it's not super high value and I'd have to check how to best stream them, it's more of a personal judgment.
You could do a PR with that change, yes. Please also remove the |
|
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. |
|
not stale |
Refs #2310
As described in the open tasks for the benchmarking of LoRA methods, I have added a dataloader to data.py script. Before integrating it in run.py, I had a few queries:
Is the
default_data_collatorcollator ideal for benchmarking or something more advanced likeDataCollatorWithPaddingpreferred?Are there any tests needed for checking improvement in throughput or memory needed?
Could you please review?
cc: @BenjaminBossan