@@ -106,7 +106,7 @@ size_t dispatchNVFP4xNVFP4GemmClusterShapeSm10x(T* D, void const* A, void const*
106106 break ;
107107 default :
108108 throw std::runtime_error (
109- " [TensorRT- LLM Error][FP4][dispatch_gemm_cluster_shape] Config is invalid for FP4 GEMM." );
109+ " [TensorRT LLM Error][FP4][dispatch_gemm_cluster_shape] Config is invalid for FP4 GEMM." );
110110 break ;
111111 }
112112}
@@ -187,7 +187,7 @@ size_t dispatchNVFP4xNVFP4GemmClusterShapeSm120(T* D, void const* A, void const*
187187 break ;
188188 default :
189189 throw std::runtime_error (
190- " [TensorRT- LLM Error][FP4][dispatch_gemm_cluster_shape] Config is invalid for FP4 GEMM." );
190+ " [TensorRT LLM Error][FP4][dispatch_gemm_cluster_shape] Config is invalid for FP4 GEMM." );
191191 break ;
192192 }
193193}
@@ -215,16 +215,16 @@ size_t dispatchNVFP4xNVFP4GemmCTAShapeSm120(T* D, void const* A, void const* B,
215215 occupancy);
216216 break ;
217217 case tkc::CutlassTileConfigSM120::Undefined:
218- throw std::runtime_error (" [TensorRT- LLM Error][FP4][sm120][dispatch_gemm_cta_shape] Gemm config undefined." );
218+ throw std::runtime_error (" [TensorRT LLM Error][FP4][sm120][dispatch_gemm_cta_shape] Gemm config undefined." );
219219 break ;
220220 case tkc::CutlassTileConfigSM120::ChooseWithHeuristic:
221221 throw std::runtime_error (
222- " [TensorRT- LLM Error][FP4][sm120][dispatch_gemm_cta_shape] Gemm config should have already been set by "
222+ " [TensorRT LLM Error][FP4][sm120][dispatch_gemm_cta_shape] Gemm config should have already been set by "
223223 " heuristic." );
224224 break ;
225225 default :
226226 throw std::runtime_error (
227- " [TensorRT- LLM Error][FP4][sm120][dispatch_gemm_cta_shape] Config is invalid for FP4 GEMM." );
227+ " [TensorRT LLM Error][FP4][sm120][dispatch_gemm_cta_shape] Config is invalid for FP4 GEMM." );
228228 break ;
229229 }
230230}
@@ -267,7 +267,7 @@ size_t dispatchMXFP8xMXFP4GemmClusterShapeSm100(T* D, void const* A, void const*
267267 break ;
268268 default :
269269 throw std::runtime_error (
270- " [TensorRT- LLM Error][FP4][dispatch_gemm_cluster_shape] Config is invalid for FP4 GEMM." );
270+ " [TensorRT LLM Error][FP4][dispatch_gemm_cluster_shape] Config is invalid for FP4 GEMM." );
271271 break ;
272272 }
273273}
@@ -303,15 +303,15 @@ size_t dispatchMXFP8xMXFP4GemmCTAShapeSm100(T* D, void const* A, void const* B,
303303 occupancy);
304304 break ;
305305 case tkc::CutlassTileConfigSM100::Undefined:
306- throw std::runtime_error (" [TensorRT- LLM Error][FP4][dispatch_gemm_cta_shape] Gemm config undefined." );
306+ throw std::runtime_error (" [TensorRT LLM Error][FP4][dispatch_gemm_cta_shape] Gemm config undefined." );
307307 break ;
308308 case tkc::CutlassTileConfigSM100::ChooseWithHeuristic:
309309 throw std::runtime_error (
310- " [TensorRT- LLM Error][FP4][dispatch_gemm_cta_shape] Gemm config should have already been set by "
310+ " [TensorRT LLM Error][FP4][dispatch_gemm_cta_shape] Gemm config should have already been set by "
311311 " heuristic." );
312312 break ;
313313 default :
314- throw std::runtime_error (" [TensorRT- LLM Error][FP4][dispatch_gemm_cta_shape] Config is invalid for FP4 GEMM." );
314+ throw std::runtime_error (" [TensorRT LLM Error][FP4][dispatch_gemm_cta_shape] Config is invalid for FP4 GEMM." );
315315 break ;
316316 }
317317}
@@ -348,7 +348,7 @@ size_t CutlassFp4GemmRunner<T, fp4GemmType>::dispatchToArch(T* D, void const* A,
348348 else
349349 {
350350 throw std::runtime_error (
351- " [TensorRT- LLM Error][CutlassFp4GemmRunner][GEMM Dispatch] Arch unsupported for CUTLASS FP4 GEMM" );
351+ " [TensorRT LLM Error][CutlassFp4GemmRunner][GEMM Dispatch] Arch unsupported for CUTLASS FP4 GEMM" );
352352 }
353353 }
354354 else if constexpr (fp4GemmType == FP4GemmType::W4A4_NVFP4_NVFP4)
@@ -376,13 +376,13 @@ size_t CutlassFp4GemmRunner<T, fp4GemmType>::dispatchToArch(T* D, void const* A,
376376 else
377377 {
378378 throw std::runtime_error (
379- " [TensorRT- LLM Error][CutlassFp4GemmRunner][GEMM Dispatch] Arch unsupported for CUTLASS FP4 GEMM" );
379+ " [TensorRT LLM Error][CutlassFp4GemmRunner][GEMM Dispatch] Arch unsupported for CUTLASS FP4 GEMM" );
380380 }
381381 }
382382 else
383383 {
384384 throw std::runtime_error (
385- " [TensorRT- LLM Error][CutlassFp4GemmRunner][GEMM Dispatch] FP4 Gemm type unsupported for CUTLASS FP4 GEMM" );
385+ " [TensorRT LLM Error][CutlassFp4GemmRunner][GEMM Dispatch] FP4 Gemm type unsupported for CUTLASS FP4 GEMM" );
386386 }
387387}
388388
0 commit comments