Skip to content

Issues: ModelCloud/GPTQModel

Beta
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Issues list

[BUG] RuntimeError: b_q_weight type is not kInt bug Something isn't working
#1625 opened Jun 2, 2025 by wumaotegan
Torch Linear instead of Triton Linear
#1617 opened May 20, 2025 by MekkCyber
How to Quantize LLaVA based models?
#1611 opened May 12, 2025 by Himanshunitrr
multi GPU setting
#1610 opened May 12, 2025 by chesterout
mistral3
#1561 opened Apr 27, 2025 by ewof
[BUG]TypeError: internvl_chat isn't supported yet. bug Something isn't working
#1556 opened Apr 23, 2025 by Maglanyulan
[Benchmark] Reproduce GPTQv2 results bug Something isn't working
#1545 opened Apr 16, 2025 by eldarkurtic
GPTQModel with PEFT
#1534 opened Apr 11, 2025 by BUGBOY101
[BUG] vllm support for QQQ format checkpoints bug Something isn't working
#1501 opened Apr 4, 2025 by jmkuebler
[BUG] ValueError: Quantization: Failed due to NaN loss bug Something isn't working
#1497 opened Apr 2, 2025 by it-dainb
[BUG]cant install with torch-rocm bug Something isn't working
#1473 opened Mar 20, 2025 by yggdrasil75
[FEATURE] ADD VPTQ
#1463 opened Mar 14, 2025 by Qubitium
[BUG] RuntimeError: Numpy is not available bug Something isn't working
#1403 opened Mar 9, 2025 by davidray222
[KERNEL] AllSpark + Exllama vLLM
#1359 opened Mar 1, 2025 by Qubitium
ProTip! Find all open issues with in progress development work with linked:pr.