Skip to content

Commit f63ece4

Browse files
xinhe-nvlancelly
authored andcommitted
test: [CI] Add failed cases into waives.txt (NVIDIA#6333)
Signed-off-by: Xin He (SW-GPU) <[email protected]> Signed-off-by: Lanyu Liao <[email protected]>
1 parent 980bc9e commit f63ece4

File tree

5 files changed

+18
-4
lines changed

5 files changed

+18
-4
lines changed

tests/integration/defs/accuracy/test_cli_flow.py

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -211,6 +211,7 @@ class TestLlama3_3NemotronSuper49Bv1(CliFlowAccuracyTestHarness):
211211
def test_auto_dtype_tp2(self):
212212
self.run(tasks=[MMLU(self.MODEL_NAME)], tp_size=2, dtype='auto')
213213

214+
@skip_pre_hopper
214215
@pytest.mark.skip(
215216
reason="nemotron-nas scripts have to accommodate fp8 flags")
216217
@pytest.mark.skip_less_device(2)
@@ -811,14 +812,14 @@ class TestLlama3_1_8BInstruct(CliFlowAccuracyTestHarness):
811812
def test_auto_dtype(self):
812813
self.run(dtype='auto')
813814

814-
@skip_pre_ada
815+
@skip_pre_hopper
815816
def test_fp8_prequantized(self, mocker):
816817
mocker.patch.object(
817818
self.__class__, "MODEL_PATH",
818819
f"{llm_models_root()}/llama-3.1-model/Llama-3.1-8B-Instruct-FP8")
819820
self.run(quant_algo=QuantAlgo.FP8, kv_cache_quant_algo=QuantAlgo.FP8)
820821

821-
@skip_pre_ada
822+
@skip_pre_hopper
822823
@skip_post_blackwell
823824
def test_medusa_fp8_prequantized(self, mocker):
824825
# nvidia/Llama-3.1-8B-Medusa-FP8
@@ -958,6 +959,7 @@ class TestLlama3_3_70BInstruct(CliFlowAccuracyTestHarness):
958959
def test_auto_dtype_tp8(self):
959960
self.run(tasks=[MMLU(self.MODEL_NAME)], tp_size=8, dtype='auto')
960961

962+
@skip_pre_hopper
961963
@pytest.mark.skip_less_device(4)
962964
@pytest.mark.skip_device_not_contain(["H100", "H200", "B200"])
963965
def test_fp8_prequantized_tp4(self, mocker):

tests/integration/defs/accuracy/test_llm_api_pytorch.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -307,6 +307,7 @@ def test_auto_dtype(self):
307307
task = CnnDailymail(self.MODEL_NAME)
308308
task.evaluate(llm)
309309

310+
@skip_pre_hopper
310311
def test_fp8_prequantized(self):
311312
model_path = f"{llm_models_root()}/llama-3.2-models/Llama-3.2-1B-FP8"
312313
with LLM(model_path) as llm:
@@ -1478,6 +1479,7 @@ def test_auto_dtype_tp2(self):
14781479
task.evaluate(llm,
14791480
extra_evaluator_kwargs=dict(apply_chat_template=True))
14801481

1482+
@skip_pre_hopper
14811483
@pytest.mark.skip_less_device(2)
14821484
@pytest.mark.skip_device_not_contain(["H100", "B200"])
14831485
def test_fp8_prequantized_tp2(self):
@@ -1507,6 +1509,7 @@ def test_auto_dtype(self):
15071509
task.evaluate(llm,
15081510
extra_evaluator_kwargs=dict(apply_chat_template=True))
15091511

1512+
@skip_pre_hopper
15101513
@pytest.mark.skip_device_not_contain(["H100", "B200"])
15111514
def test_fp8_prequantized(self):
15121515
model_path = f"{llm_models_root()}/Llama-3.1-Nemotron-Nano-8B-v1-FP8"
@@ -1547,6 +1550,7 @@ def test_auto_dtype(self, cuda_graph, tp_size, pp_size, ep_size):
15471550
# task.evaluate(llm,
15481551
# extra_evaluator_kwargs=dict(apply_chat_template=True))
15491552

1553+
@skip_pre_hopper
15501554
@pytest.mark.skip_less_device(8)
15511555
@pytest.mark.skip_device_not_contain(["H100", "B200"])
15521556
@parametrize_with_ids("cuda_graph", [False, True])

tests/integration/defs/test_e2e.py

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1938,8 +1938,12 @@ def test_ptp_quickstart_advanced_mixed_precision(llm_root, llm_venv):
19381938
("llava-v1.6-mistral-7b", "llava-v1.6-mistral-7b-hf"),
19391939
("qwen2-vl-7b-instruct", "Qwen2-VL-7B-Instruct"),
19401940
("qwen2.5-vl-7b-instruct", "Qwen2.5-VL-7B-Instruct"),
1941-
("mistral-small-3.1-24b-instruct", "Mistral-Small-3.1-24B-Instruct-2503"),
1942-
("gemma-3-27b-it", "gemma/gemma-3-27b-it"),
1941+
pytest.param("mistral-small-3.1-24b-instruct",
1942+
"Mistral-Small-3.1-24B-Instruct-2503",
1943+
marks=pytest.mark.skip_less_device_memory(80000)),
1944+
pytest.param("gemma-3-27b-it",
1945+
"gemma/gemma-3-27b-it",
1946+
marks=pytest.mark.skip_less_device_memory(80000)),
19431947
])
19441948
def test_ptp_quickstart_multimodal(llm_root, llm_venv, model_name, model_path,
19451949
modality, use_cuda_graph):

tests/integration/test_lists/qa/llm_sanity_test.txt

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -109,6 +109,8 @@ test_e2e.py::test_ptp_quickstart_multimodal[qwen2.5-vl-7b-instruct-Qwen2.5-VL-7B
109109
test_e2e.py::test_ptp_quickstart_multimodal[qwen2.5-vl-7b-instruct-Qwen2.5-VL-7B-Instruct-image-True]
110110
test_e2e.py::test_ptp_quickstart_multimodal[qwen2.5-vl-7b-instruct-Qwen2.5-VL-7B-Instruct-video-False]
111111
test_e2e.py::test_ptp_quickstart_multimodal[qwen2.5-vl-7b-instruct-Qwen2.5-VL-7B-Instruct-video-True]
112+
test_e2e.py::test_ptp_quickstart_multimodal[gemma-3-27b-it-gemma/gemma-3-27b-it-image-False]
113+
test_e2e.py::test_ptp_quickstart_multimodal[gemma-3-27b-it-gemma/gemma-3-27b-it-image-True]
112114
test_e2e.py::test_ptp_scaffolding[DeepSeek-R1-Distill-Qwen-7B-DeepSeek-R1/DeepSeek-R1-Distill-Qwen-7B]
113115
test_e2e.py::test_ptp_star_attention_example[Llama3.1-8B-BF16-llama-3.1-model/Meta-Llama-3.1-8B]
114116
test_e2e.py::test_qwen_e2e_cpprunner_large_new_tokens[DeepSeek-R1-Distill-Qwen-1.5B-DeepSeek-R1-Distill-Qwen-1.5B]

tests/integration/test_lists/waives.txt

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -421,6 +421,7 @@ triton_server/test_triton_llm.py::test_llava_onevision[test_video-False-1---Fals
421421
triton_server/test_triton.py::test_cpp_unit_tests[cpp-unit-tests] SKIP (https://nvbugs/5401088)
422422
accuracy/test_llm_api_pytorch.py::TestGemma3_27BInstruct::test_auto_dtype SKIP (https://nvbugs/5401114)
423423
test_e2e.py::test_ptp_quickstart_multimodal[gemma-3-27b-it-gemma/gemma-3-27b-it-image-True] SKIP (https://nvbugs/5401114)
424+
test_e2e.py::test_ptp_quickstart_multimodal[gemma-3-27b-it-gemma/gemma-3-27b-it-image-False] SKIP (https://nvbgus/5401114)
424425
examples/test_recurrentgemma.py::test_llm_recurrentgemma_1gpu[use_cpp_session-recurrentgemma-2b-use_paged_cache-int4_awq-float16-enable_attn_plugin-enable_gemm_plugin] SKIP (https://nvbugs/5401233)
425426
examples/test_recurrentgemma.py::test_llm_recurrentgemma_2gpu[recurrentgemma-2b] SKIP (https://nvbugs/5401233)
426427
examples/test_multimodal.py::test_llm_multimodal_general[VILA1.5-3b-pp:1-tp:1-float16-bs:1-cpp_e2e:False-nb:1] SKIP (https://nvbugs/5401156)
@@ -440,3 +441,4 @@ unittest/trt/attention/test_gpt_attention.py -k "partition0" SKIP (https://nvbug
440441
unittest/trt/attention/test_gpt_attention.py -k "partition1" SKIP (https://nvbugs/5412456)
441442
unittest/trt/attention/test_gpt_attention.py -k "partition2" SKIP (https://nvbugs/5412456)
442443
unittest/trt/attention/test_gpt_attention.py -k "partition3" SKIP (https://nvbugs/5412456)
444+
test_e2e.py::test_ptp_quickstart_multimodal[qwen2-vl-7b-instruct-Qwen2-VL-7B-Instruct-image-False] SKIP (https://nvbugs/5414909)

0 commit comments

Comments
 (0)