use llama.cpp quantization with QAT finetune #11170
Unanswered
anyinlover
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I'm finetuning a 1B model with QAT using torchtune library, but I wonder is it match to llama.cpp quantization algorithm, does anyone have some experience?
Beta Was this translation helpful? Give feedback.
All reactions