Skip to content

Commit cec275e

Browse files
GuanHIsotr0py
andauthored
[Bugfix] resolve Qwen3-VL GPTQModel quantized model loading failure (#28663)
Signed-off-by: GuanH <guansdrailib@gmail.com> Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn> Co-authored-by: Isotr0py <mozf@mail2.sysu.edu.cn>
1 parent e2741f6 commit cec275e

File tree

2 files changed

+6
-3
lines changed

2 files changed

+6
-3
lines changed

vllm/model_executor/models/qwen3_vl.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1138,7 +1138,9 @@ def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""):
11381138
self.config = config
11391139

11401140
self.quant_config = quant_config
1141-
self.model = Qwen3LLMModel(vllm_config=vllm_config, prefix=prefix)
1141+
self.model = Qwen3LLMModel(
1142+
vllm_config=vllm_config, prefix=maybe_prefix(prefix, "model")
1143+
)
11421144

11431145
if get_pp_group().is_last_rank:
11441146
if config.tie_word_embeddings:

vllm/model_executor/models/utils.py

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -117,9 +117,10 @@ class AutoWeightsLoader:
117117
environment variable `VLLM_LOGGING_LEVEL=DEBUG`.
118118
"""
119119

120-
# Models trained using early version ColossalAI
121-
# may include these tensors in checkpoint. Skip them.
120+
# Models trained using early version ColossalAI or quantized by
121+
# GPTQModel may include these tensors in checkpoint. Skip them.
122122
ROTARY_EMBEDS_UNUSED_WEIGHTS = [
123+
"rotary_pos_emb.inv_freq",
123124
"rotary_emb.inv_freq",
124125
"rotary_emb.cos_cached",
125126
"rotary_emb.sin_cached",

0 commit comments

Comments
 (0)