You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This is one of the best-performing open-weight models, and should be possible to host locally when quantized. Possible use cases include being an LLM judge.
Your contribution
N/A
The text was updated successfully, but these errors were encountered:
Feature request
Possible quantized model to use: https://huggingface.co/neuralmagic/Meta-Llama-3.1-405B-Instruct-quantized.w8a8
May be able to fit on 8x A100 (40 or 80GB) on GCP.
Motivation / references
This is one of the best-performing open-weight models, and should be possible to host locally when quantized. Possible use cases include being an LLM judge.
Your contribution
N/A
The text was updated successfully, but these errors were encountered: