Skip to content

Commit 88140a8

Browse files
committed
Address code rabbit comment
Signed-off-by: Jin Li <[email protected]>
1 parent 4e653f6 commit 88140a8

File tree

2 files changed

+10
-1
lines changed

2 files changed

+10
-1
lines changed

tensorrt_llm/_torch/custom_ops/trtllm_gen_custom_ops.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -540,7 +540,7 @@ def fp8_block_scale_moe_runner(
540540

541541
@fp8_block_scale_moe_runner.register_fake
542542
def _(
543-
rrouting_logits: torch.Tensor,
543+
routing_logits: torch.Tensor,
544544
routing_bias: torch.Tensor,
545545
hidden_states: torch.Tensor,
546546
hidden_states_scale: torch.Tensor,

tensorrt_llm/llmapi/llm_args.py

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1797,6 +1797,15 @@ class TorchCompileConfig(BaseModel):
17971797
description=
17981798
"The maximum number of CUDA streams to use for torch.compile.")
17991799

1800+
@field_validator('max_num_streams')
1801+
@classmethod
1802+
def validate_torch_compile_max_num_streams(cls, v):
1803+
"""Validate torch_compile_config.max_num_streams >= 1."""
1804+
if v < 1:
1805+
raise ValueError(
1806+
"torch_compile_config.max_num_streams must be >= 1")
1807+
return v
1808+
18001809

18011810
class TorchLlmArgs(BaseLlmArgs):
18021811
# Just a dummy BuildConfig to allow code reuse with the TrtLlmArgs

0 commit comments

Comments
 (0)