Skip to content

Commit 0fc8fa7

Browse files
authored
fix: gptq marlin weight loading failure (vllm-project#23066)
1 parent 21e3943 commit 0fc8fa7

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

vllm/model_executor/layers/quantization/gptq_marlin.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -56,7 +56,7 @@ def get_moe_quant_method(
5656
# Dynamic per module/layer rules may override base config
5757
override_config(cloned_config, prefix=prefix)
5858

59-
return moe_method_cls(cloned_config)
59+
return moe_method_cls(cloned_config, layer.moe_config)
6060
return None
6161

6262

0 commit comments

Comments
 (0)