-
Notifications
You must be signed in to change notification settings - Fork 56
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to add a new model for OmniQuant? #22
Comments
If you want to quantize a new model with the same architecture with supported model, you can just set the
|
Thanks for you reply, my question is how to add a new model which architecture is not supported yet. |
You have to add new files int_{your model}_layer.py in |
Have you tried to add bloom models? I met some problems in issue #29 . |
@Louym No, I haven't tried to add bloom models. I don't know the details of your implementation (and I'm not a contributor of this repo), and attached error has various reasons. Anyway, I hope you can solve this problem, and I hope my experience can be of help.
|
Thanks for your brilliant work, after explord the project for several days, I found that OmniQuant is portable for edge devices, like Jetson or phones. And wondering how can I add more models into OmniQuant, do you have any tutorials about this? Or maybe we can start from CodeLlama, since it has the similiar architecture with Llama-2, and Llama-2 is already supported.
Also apologies in advance if this seems to be something obvious because I'm new in LLM field.
The text was updated successfully, but these errors were encountered: