Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Llama-3.2-11B-Vision-Instruct-GGUF Support Request #125

Open
boricuapab opened this issue Nov 17, 2024 · 2 comments
Open

Llama-3.2-11B-Vision-Instruct-GGUF Support Request #125

boricuapab opened this issue Nov 17, 2024 · 2 comments

Comments

@boricuapab
Copy link

boricuapab commented Nov 17, 2024

Is this gguf llama 3.2 vision model supported?

https://huggingface.co/leafspark/Llama-3.2-11B-Vision-Instruct-GGUF/tree/main

I tried running the 20 gig 11b model from meta, but it's taking extremely long to run and never finishes on a 32 gig ram/16gig vram setup.

@heshengtao
Copy link
Owner

I found that the model you gave me doesn't seem to be able to be successfully called by llama cpp python, or it may be my version problem, I upgraded the VLM-GGUF loader node. Now you can choose the chat format, but if none of these options successfully load the clip model, I seriously doubt that llama cpp python does not support this model yet. https://github.com/abetlen/llama-cpp-python/tree/7ecdd944624cbd49e4af0a5ce1aa402607d58dcc?tab=readme-ov-file#multi-modal-models all the models written here can be called by party, because party relies on llama cpp python to call GGUF.

@power88
Copy link

power88 commented Nov 24, 2024

Consider use Llama-3.2-vision-11b via Ollama? Now Ollama support Llama-3.2-vision-11b.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants