Skip to content

Pull requests: intel/auto-round

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

[WIP] WIP address feedback on MTP params
#1528 opened Mar 11, 2026 by Copilot AI Draft
1 of 8 tasks
support MTP params: copy, fp8 dequant, and WOQ RTN quantization
#1526 opened Mar 10, 2026 by xin3he Loading…
2 of 9 tasks
fix dynamic int8 w8a8 export issue with tuning
#1525 opened Mar 10, 2026 by thuang6 Loading…
6 tasks
Support GLM-Image model quantizaiton
#1512 opened Mar 8, 2026 by lvliang-intel Loading…
2 of 9 tasks
Support block-wise fp8 quant
#1487 opened Mar 3, 2026 by mengniwang95 Loading…
1 of 6 tasks
Enhance llmc CI on GPU and XPU
#1483 opened Mar 2, 2026 by chensuyue Loading…
1 of 9 tasks
0.12.0
Enable CUDA CI
#1473 opened Feb 27, 2026 by XuehaoSun Draft
3 of 6 tasks
0.12.0
Support Qwen3 and Qwen2.5 Omni model quantization
#1404 opened Feb 4, 2026 by lvliang-intel Loading…
2 of 9 tasks
refactor init of compressor engineering ready only add when the PR is ready to merge
#1339 opened Jan 26, 2026 by n1ck-guo Loading…
1 of 9 tasks
Add asym for XPU backend.
#1316 opened Jan 22, 2026 by luoyu-intel Draft
Fix ignore_layers not working for FP8 models
#1286 opened Jan 15, 2026 by Copilot AI Loading…
11 tasks done
fix disable_opt_rtn spelling error
#1250 opened Jan 9, 2026 by WeiweiZhang1 Loading…
add per-task lm_eval args for exprimental usage
#1017 opened Nov 11, 2025 by WeiweiZhang1 Loading…
[WIP] [STEP 2] split compressor into few quantizers
#841 opened Sep 23, 2025 by n1ck-guo Loading…
ProTip! Adding no:label will show everything without a label.