-
Notifications
You must be signed in to change notification settings - Fork 67
Issues: NVIDIA/TensorRT-Model-Optimizer
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Support for W4A16 and W4A8 Quantization in TensorRT Model Optimizer
feature request
New feature or request
#189
opened Apr 30, 2025 by
david-PHR
Cannot serve modelopt quantized nvfp4 model on TensorRT LLM
bug
Something isn't working
#187
opened Apr 27, 2025 by
enisaras
[BUG] modelopt restore quantized models using 'AutoModelForCausalLM.from_pretrained' doesn't work for mixtral-8x7b
bug
Something isn't working
#186
opened Apr 27, 2025 by
wanzhenchn
[BUG] modelopt restore using 'from_pretrained' doesn't work in diffusers
bug
Something isn't working
#185
opened Apr 25, 2025 by
ishan-modi
Support more Quantization methods for "onnx_ptq"?
feature request
New feature or request
#184
opened Apr 24, 2025 by
s101010tw
[BUG] Issue processing NF4 double quantization
bug
Something isn't working
#183
opened Apr 22, 2025 by
ishan-modi
Qwen2_MoE AWQ(w4a16/w4a8) quantization failed with Nan AssertionError
#182
opened Apr 22, 2025 by
wanzhenchn
Torch Quantization: Allow restoring quantized model and re-running calibration on new data (PTQ)
feature request
New feature or request
#179
opened Apr 16, 2025 by
david-PHR
Explicit INT8 Quantization Fails to Fuse Concat-Conv Block Compared to Implicit Mode
#174
opened Apr 9, 2025 by
patrickgrommelt
Getting Real quantization not supported for this format error when using mtq.compress(model)
#171
opened Apr 5, 2025 by
RivenSama
int4 quantization output onnx does not load
bug
Something isn't working
#156
opened Mar 13, 2025 by
thejaswi01
PyTorch Quantization Failed to Quantize Scaled Dot Product
#149
opened Mar 7, 2025 by
YixuanSeanZhou
Restore functionality: lm_head option to disable quantization
#138
opened Feb 20, 2025 by
michaelfeil
Previous Next
ProTip!
Exclude everything labeled
bug
with -label:bug.