-
Notifications
You must be signed in to change notification settings - Fork 100
Add PEFT benchmarking script in thunder/benchmarks
#2254
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
for more information, see https://pre-commit.ci
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Great add! Left some small comments
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you @riccardofelluga @KaelanDt
At some point we might look at splitting out the various requirements here, but not now.
What does this PR do?
This PR picks up and replaces #1978 with some tweaks. It introduces PEFT benchmarks for single gpu and multi-gpu with fsdp2 as deistributed strategy.
PR review
Can be tested by running:
Multi-GPU:
Side notes
I've extended the logging from the original PR and removed the prints. Now this uses the log levels to print more or less information depending on
--verbose
flag.@crcrpar in the fsdp2 part I carried it over from the
benchmark_litgpt.py
file, but I've commented out the litgpt specific layers. I've left it there for now but I am planning to remove them after the review, let me know if you think I should find the equivalent layers in transformers and adapt that code.cc. @wprazuch