-
Notifications
You must be signed in to change notification settings - Fork 200
Pull requests: NVIDIA/TensorRT-Model-Optimizer
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
[5680954,5620660@2][ONNX][Autocast] Update value info in converted graph
#611
opened Nov 26, 2025 by
gcunhase
Loading…
Add checkpoint save/load to ForwardHook + add IterativeChannelContributionHook
#610
opened Nov 26, 2025 by
danielkorzekwa
Loading…
Support attention quantization for diffusers >= 0.35.0
#608
opened Nov 25, 2025 by
shengliangxu
•
Draft
Add pruning checkpoints for the compress algorithm
#607
opened Nov 25, 2025 by
danielkorzekwa
Loading…
Fix extra args and --component-dtype default value
#605
opened Nov 24, 2025 by
shengliangxu
Loading…
Convert compressed-tensor int4 format to GPTQ int4 format
#590
opened Nov 20, 2025 by
Edwardf0t1
Loading…
Product Rename: TensorRT Model Optimizer to Model Optimizer
#583
opened Nov 20, 2025 by
kevalmorabia97
Loading…
1 of 2 tasks
[OMNIML-2852] [2/n] Add Core Sparse Attention Infrastructure
#527
opened Nov 7, 2025 by
kaix-nv
Loading…
[Draft] [5526696] Add kv cache quantization support for onnx quantization
#486
opened Oct 31, 2025 by
zhanghaoc
Loading…
Preserve original rope scaling type in export due to transformers library AutoConfig issue
#452
opened Oct 17, 2025 by
Edwardf0t1
Loading…
[1/2] Registry interface for custom quantization functional backend
#449
opened Oct 17, 2025 by
realAsma
Loading…
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.