-
Notifications
You must be signed in to change notification settings - Fork 313
Pull requests: vllm-project/llm-compressor
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Fix deprecated torch_dtype usage in transformers loading
#2109
opened Dec 11, 2025 by
jangel97
Loading…
add kv quant example
autoround
For any PR / issue related to autoround support
#2100
opened Dec 5, 2025 by
mengniwang95
Loading…
Linearize gpt_oss model and add separate example to qunatize it to w4a8
#2091
opened Dec 3, 2025 by
isharif168
Loading…
feat: add importance-aware mixed-precision quantization
#2083
opened Dec 2, 2025 by
wangwenmingaa
Loading…
[Performance] Batched calibration
ready
When a PR is ready for review
#2054
opened Nov 20, 2025 by
kylesayrs
Loading…
[Misc] Remove When a PR is ready for review
is_moe_model
ready
#2053
opened Nov 20, 2025 by
kylesayrs
Loading…
Modernize transformers module with type hints and generic types
#2034
opened Nov 14, 2025 by
sugatmahanti
Loading…
[Sequential Onloading] Support onloading and offloading frozen dataclasses
#2016
opened Nov 10, 2025 by
kylesayrs
Loading…
Implement When a PR is ready for review
propagate_error argument
ready
#2008
opened Nov 10, 2025 by
kylesayrs
Loading…
[AWQ] Allow users to disable quantization during AWQ
#1973
opened Oct 28, 2025 by
brian-dellabetta
•
Draft
[AWQ] Generalize AWQ quantization
ready
When a PR is ready for review
#1961
opened Oct 22, 2025 by
kylesayrs
Loading…
4 tasks done
[Oneshot] Add validation for empty dataset and enhance oneshot function parameters
#1957
opened Oct 21, 2025 by
ArkaSanka
Loading…
[Attention] Support FP4 attention quantization
nvfp4
For any PR / issue related to NVFP4 support
#1924
opened Oct 14, 2025 by
kylesayrs
Loading…
Previous Next
ProTip!
no:milestone will show everything without a milestone.