-
Notifications
You must be signed in to change notification settings - Fork 18
✨ Use shared CachedRequestData as vllm:main #273
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: Prashant Gupta <[email protected]>
👋 Hi! Thank you for contributing to vLLM support on Spyre.
Or this can be done with
Now you are good to go 🚀 |
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
@prashantgupta24 , I think the other breaking change that you mentioned is the sampling metadata one, right? I've opened a hacky PR to temporarily fix this: #278 |
Yep, thanks! |
Signed-off-by: Prashant Gupta <[email protected]>
closing in favor of #283 |
# Description This branch has a fix for: - Caching the token_ids (now the new tokens are cached in `execute_model` instead of `update_states`. This is because of vllm-project/vllm#20291. ) - Changes from the `CachedRequestData` (#273) ## Related Issues Fix for #271 --------- Signed-off-by: Prashant Gupta <[email protected]> Signed-off-by: Max de Bayser <[email protected]> Co-authored-by: Max de Bayser <[email protected]>
Description
This is more complicated than I thought originally :)
Alright, all tests are passing locally. There seems to be another breaking change in
vllm:main
that will have to be addressed to make themain
tests pass. Thedefault
tests fail because this code is not backward compatible yet 😅Related Issues
fix #271