Skip to content

Pull requests: vllm-project/compressed-tensors

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

[Bugfix] Forward quantize better wrapping
#521 opened Nov 18, 2025 by kylesayrs Loading…
support wInt4aFp8 for moe
#518 opened Nov 12, 2025 by Wangzheee Loading…
[WIP] fix qparams decompression bug Something isn't working
#514 opened Nov 10, 2025 by shanjiaz Loading…
[MXFP4] Add calibration support
#509 opened Nov 4, 2025 by dsikka Loading…
[Attention] Support FP4 attention quantization
#491 opened Oct 14, 2025 by kylesayrs Loading…
[KV Cache] support kv cache int8 per channel quant
#398 opened Jul 19, 2025 by Eviannn Loading…
Optimize sparse 2:4 compression performance
#358 opened Jun 16, 2025 by rahul-tuli Draft
8 tasks done
relax setuptools_scm version requirement
#343 opened Jun 6, 2025 by envolution Loading…
ProTip! Add no:assignee to see everything that’s not assigned.