Chat template for Pixtral 12B model #11706
Replies: 1 comment 2 replies
-
I've been doing some experimenting with Pixtral 12B on VLLM. This is the chat template I use if it helps. It's an expanded version of the default Mistral template. It allows for images to be provided in all response types, allows non-alternating roles, and inline system messages. Works pretty well if you want to use it.
If you use it make sure you also pass the content format arg as it might default to OpenAI format and break. e.g. use Also if you're using a chat template with pixtral you need to use a HF version of the model, not the mistral version. I've got some quants published. W4A16 brings model size down to ~9GB https://huggingface.co/nintwentydo/pixtral-12b-2409-W4A16-G128 And these are some others I've tried I'd recommend avoiding bitsandbytes though - prompt processing is already slow on pixtral compared to other models and bnb has been by far the slowest for me. |
Beta Was this translation helpful? Give feedback.
-
I need help on hosting Pixtral 12b model using Vllm with the chat template enabled. If any anyone has chat template for Pixtral 12B model, drop your template and thoughts.
Beta Was this translation helpful? Give feedback.
All reactions