Skip to content

Commit 34d8036

Browse files
committed
fix ruff
Signed-off-by: Pavani Majety <pmajety@nvidia.com>
1 parent 7798af4 commit 34d8036

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

vllm/model_executor/layers/quantization/modelopt.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -236,7 +236,7 @@ def get_quant_method(
236236
if "vision_tower" in prefix or "vision_model" in prefix:
237237
return UnquantizedLinearMethod()
238238
return ModelOptFp8LinearMethod(self)
239-
elif isinstance(layer, Attention) or isinstance(layer, MLAAttention):
239+
elif isinstance(layer, (Attention, MLAAttention)):
240240
return ModelOptFp8KVCacheMethod(self)
241241
elif isinstance(layer, FusedMoE):
242242
return ModelOptFp8MoEMethod(self, layer)
@@ -921,7 +921,7 @@ def get_quant_method(
921921
if "vision_tower" in prefix or "vision_model" in prefix:
922922
return UnquantizedLinearMethod()
923923
return ModelOptNvFp4LinearMethod(self)
924-
elif isinstance(layer, Attention) or isinstance(layer, MLAAttention):
924+
elif isinstance(layer, (Attention, MLAAttention)):
925925
return ModelOptFp8KVCacheMethod(self)
926926
elif isinstance(layer, FusedMoE):
927927
if skip_layer:

0 commit comments

Comments
 (0)