Skip to content

Commit e1464c3

Browse files
authored
[Quantization] Enable compressed-tensors AWQ for Turing GPU (#29732)
Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
1 parent a491b09 commit e1464c3

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

vllm/model_executor/layers/quantization/compressed_tensors/schemes/compressed_tensors_wNa16.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -79,8 +79,8 @@ def __init__(
7979

8080
@classmethod
8181
def get_min_capability(cls) -> int:
82-
# ampere and up
83-
return 80
82+
# Turing and up
83+
return 75
8484

8585
def create_weights(
8686
self,

0 commit comments

Comments
 (0)