File tree Expand file tree Collapse file tree 1 file changed +6
-1
lines changed
tensorrt_llm/_torch/modules/fused_moe Expand file tree Collapse file tree 1 file changed +6
-1
lines changed Original file line number Diff line number Diff line change 44
55from ...distributed .ops import reducescatter
66from ...model_config import ModelConfig
7- from ...utils import Fp4QuantizedTensor
7+ from ...utils import Fp4QuantizedTensor , get_sm_version
88from .interface import MoE , MoEWeightLoadingMode
99from .quantization import (DeepSeekFP8BlockScalesFusedMoEMethod ,
1010 NVFP4TRTLLMGenFusedMoEMethod )
@@ -68,6 +68,11 @@ def __init__(
6868 weight_loading_mode = weight_loading_mode ,
6969 )
7070
71+ sm_version = get_sm_version ()
72+ if sm_version >= 120 :
73+ raise NotImplementedError (
74+ "TRTLLMGenFusedMoE does not support SM120 and above." )
75+
7176 assert not self .smart_router , "Smart router is not supported in TRTLLMGenFusedMoE."
7277
7378 self .num_slots = self .num_experts
You can’t perform that action at this time.
0 commit comments