@@ -2724,7 +2724,7 @@ def launchTestJobs(pipeline, testFilter)
27242724 " RTXPro6000D-4_GPUs-PyTorch-Post-Merge-2" : [" rtx-pro-6000d-x4" , " l0_rtx_pro_6000" , 2 , 2 , 4 ],
27252725 ]
27262726
2727- parallelJobs = x86TestConfigs. collectEntries{key , values -> [key, [createKubernetesPodConfig(LLM_DOCKER_IMAGE , values[0 ], " amd64" , values[4 ] ?: 1 , key. contains(" - Perf- " )), {
2727+ parallelJobs = x86TestConfigs. collectEntries{key , values -> [key, [createKubernetesPodConfig(LLM_DOCKER_IMAGE , values[0 ], " amd64" , values[4 ] ?: 1 , key. contains(" Perf" )), {
27282728 def config = VANILLA_CONFIG
27292729 if (key. contains(" single-device" )) {
27302730 config = SINGLE_DEVICE_CONFIG
@@ -2735,7 +2735,7 @@ def launchTestJobs(pipeline, testFilter)
27352735 if (key. contains(" Pybind" )) {
27362736 config = PYBIND_CONFIG
27372737 }
2738- runLLMTestlistOnPlatform(pipeline, values[0 ], values[1 ], config, key. contains(" - Perf- " ), key, values[2 ], values[3 ])
2738+ runLLMTestlistOnPlatform(pipeline, values[0 ], values[1 ], config, key. contains(" Perf" ), key, values[2 ], values[3 ])
27392739 }]]}
27402740 fullSet = parallelJobs. keySet()
27412741
@@ -2753,10 +2753,10 @@ def launchTestJobs(pipeline, testFilter)
27532753 " DGX_B200-4_GPUs-PyTorch-Post-Merge-1" : [" b200-trtllm" , " l0_dgx_b200" , 1 , 1 , 4 , 1 , true ],
27542754 " DGX_B300-4_GPUs-PyTorch-Post-Merge-1" : [" b300-x4" , " l0_dgx_b300" , 1 , 1 , 4 ],
27552755 // Perf sanity post merge test
2756- " DGX_B200-4_GPUs-PyTorch-PerfSanity -Post-Merge-1" : [" b200-x4" , " perf_sanity_l0_dgx_b200" , 1 , 2 , 4 ],
2757- " DGX_B200-4_GPUs-PyTorch-PerfSanity -Post-Merge-2" : [" b200-x4" , " perf_sanity_l0_dgx_b200" , 2 , 2 , 4 ],
2758- // "DGX_B200-8_GPUs-PyTorch-PerfSanity -Post-Merge-1": ["b200-x8", "perf_sanity_l0_dgx_b200", 1, 1, 8],
2759- // "DGX_B300-4_GPUs-PyTorch-PerfSanity -Post-Merge-1": ["b300-x4", "perf_sanity_l0_dgx_b300", 1, 1, 4],
2756+ " DGX_B200-4_GPUs-PyTorch-Perf-Sanity -Post-Merge-1" : [" b200-x4" , " perf_sanity_l0_dgx_b200" , 1 , 2 , 4 ],
2757+ " DGX_B200-4_GPUs-PyTorch-Perf-Sanity -Post-Merge-2" : [" b200-x4" , " perf_sanity_l0_dgx_b200" , 2 , 2 , 4 ],
2758+ // "DGX_B200-8_GPUs-PyTorch-Perf-Sanity -Post-Merge-1": ["b200-x8", "perf_sanity_l0_dgx_b200", 1, 1, 8],
2759+ // "DGX_B300-4_GPUs-PyTorch-Perf-Sanity -Post-Merge-1": ["b300-x4", "perf_sanity_l0_dgx_b300", 1, 1, 4],
27602760 ]
27612761 fullSet + = x86SlurmTestConfigs. keySet()
27622762
@@ -2768,7 +2768,7 @@ def launchTestJobs(pipeline, testFilter)
27682768 if (key. contains(" llvm" )) {
27692769 config = LLVM_CONFIG
27702770 }
2771- runLLMTestlistOnSlurm(pipeline, values[0 ], values[1 ], config, key. contains(" - Perf- " ), key, values[2 ], values[3 ], values[4 ] ?: 1 , values[5 ] ?: 1 , values[6 ] ?: false )
2771+ runLLMTestlistOnSlurm(pipeline, values[0 ], values[1 ], config, key. contains(" Perf" ), key, values[2 ], values[3 ], values[4 ] ?: 1 , values[5 ] ?: 1 , values[6 ] ?: false )
27722772 }]]}
27732773
27742774 parallelJobs + = parallelSlurmJobs
@@ -2797,7 +2797,7 @@ def launchTestJobs(pipeline, testFilter)
27972797 " GB200-8_GPUs-2_Nodes-PyTorch-Post-Merge-1" : [" gb200-oci-trtllm" , " l0_gb200_multi_nodes" , 1 , 3 , 8 , 2 ],
27982798 " GB200-8_GPUs-2_Nodes-PyTorch-Post-Merge-2" : [" gb200-oci-trtllm" , " l0_gb200_multi_nodes" , 2 , 3 , 8 , 2 ],
27992799 " GB200-8_GPUs-2_Nodes-PyTorch-Post-Merge-3" : [" gb200-oci-trtllm" , " l0_gb200_multi_nodes" , 3 , 3 , 8 , 2 ],
2800- " GB200-8_GPUs-2_Nodes-PyTorch-PerfSanity -Post-Merge-1" : [" gb200-oci-trtllm" , " perf_sanity_l0_gb200_multi_nodes" , 1 , 1 , 8 , 2 ],
2800+ " GB200-8_GPUs-2_Nodes-PyTorch-Perf-Sanity -Post-Merge-1" : [" gb200-oci-trtllm" , " perf_sanity_l0_gb200_multi_nodes" , 1 , 1 , 8 , 2 ],
28012801 ]
28022802 fullSet + = multiNodesSBSAConfigs. keySet()
28032803
@@ -2815,7 +2815,7 @@ def launchTestJobs(pipeline, testFilter)
28152815 if (key. contains(" llvm" )) {
28162816 config = LLVM_CONFIG
28172817 }
2818- runLLMTestlistOnSlurm(pipeline, values[0 ], values[1 ], config, key. contains(" - Perf- " ), key, values[2 ], values[3 ], values[4 ] ?: 1 , values[5 ] ?: 1 , values[6 ] ?: false )
2818+ runLLMTestlistOnSlurm(pipeline, values[0 ], values[1 ], config, key. contains(" Perf" ), key, values[2 ], values[3 ], values[4 ] ?: 1 , values[5 ] ?: 1 , values[6 ] ?: false )
28192819 }]]}
28202820 parallelJobs + = parallelSlurmJobs
28212821
@@ -2828,7 +2828,7 @@ def launchTestJobs(pipeline, testFilter)
28282828 if (key. contains(" llvm" )) {
28292829 config = LLVM_CONFIG
28302830 }
2831- runLLMTestlistOnSlurm(pipeline, values[0 ], values[1 ], config, key. contains(" - Perf- " ), key, values[2 ], values[3 ], values[4 ] ?: 1 , values[5 ] ?: 2 , values[6 ] ?: false )
2831+ runLLMTestlistOnSlurm(pipeline, values[0 ], values[1 ], config, key. contains(" Perf" ), key, values[2 ], values[3 ], values[4 ] ?: 1 , values[5 ] ?: 2 , values[6 ] ?: false )
28322832 }]]}
28332833
28342834 parallelJobs + = parallelMultiNodesSBSAJobs
0 commit comments