@@ -1130,23 +1130,21 @@ def test_nvfp4_tp2(self):
11301130# This class has extensively parameterized test methods, which yield totally 200 test cases.
11311131# This is because this model requires high test coverage over the feature combinations.
11321132# Normally we should not parameterize test methods so extensively -- just test on the typical/important feature combinations.
1133+ @skip_pre_hopper
11331134class TestDeepSeekV3Lite (LlmapiAccuracyTestHarness ):
11341135 MODEL_NAME = "deepseek-ai/DeepSeek-V3-Lite"
11351136 MODEL_PATH = f"{ llm_models_root ()} /DeepSeek-V3-Lite/bf16"
11361137
11371138 @pytest .mark .skip_less_device_memory (60000 )
11381139 # Chunked Prefill for MLA can only be enabled on SM100
1139- @parametrize_with_ids (
1140- "enable_chunked_prefill" ,
1141- [False , pytest .param (True , marks = skip_pre_hopper )])
1140+ @parametrize_with_ids ("enable_chunked_prefill" , [False , True ])
11421141 @parametrize_with_ids ("torch_compile" , [False , True ])
11431142 @parametrize_with_ids ("attention_dp,cuda_graph,overlap_scheduler" ,
11441143 [(False , False , False ), (True , False , False ),
11451144 (False , True , False ), (False , False , True ),
11461145 (False , True , True ), (True , True , True )])
11471146 # Only Hopper and Blackwell MLA kernel supports MTP
1148- @parametrize_with_ids ("mtp_nextn" ,
1149- [0 , pytest .param (2 , marks = skip_pre_hopper )])
1147+ @parametrize_with_ids ("mtp_nextn" , [0 , 2 ])
11501148 def test_bfloat16 (self , mtp_nextn , attention_dp , cuda_graph ,
11511149 overlap_scheduler , torch_compile , enable_chunked_prefill ):
11521150 kv_cache_config = KvCacheConfig (free_gpu_memory_fraction = 0.75 )
0 commit comments