@@ -52,7 +52,7 @@ Some these benchmarks are rather slow or take a long time to run on the referenc
5252| llama2_70b_lora | [ llama2_70b_lora] ( https://github.com/mlcommons/training/tree/master/llama2_70b_lora ) | pytorch | SCROLLS GovReport | 70B
5353| llama3.1_405b | [ large_language_model_pretraining] ( https://github.com/mlcommons/training/tree/master/large_language_model_pretraining ) | NeMo | C4 | 405B
5454| dlrm_dcnv2 | [ recommendation_v2] ( https://github.com/mlcommons/training/tree/master/recommendation_v2/torchrec_dlrm ) | torchrec | Criteo 3.5TB multi-hot | 167M
55- | rgat | [ graph_neural_network] ( https://github.com/mlcommons/training/tree/master/graph_neural_network ) | GLT | IGBFull | 25M
55+ | rgat | [ graph_neural_network] ( https://github.com/mlcommons/training/tree/master/graph_neural_network ) | GLT | IGBH-Full | 25M
5656
5757* Framework here is given for the reference implementation. Submitters are free to use their own frameworks to run the benchmark.
5858
@@ -71,7 +71,7 @@ Some these benchmarks are rather slow or take a long time to run on the referenc
7171| llama3.1_405b | [ large_language_model_pretraining] ( https://github.com/mlcommons/training/tree/master/large_language_model_pretraining ) | NeMo | C4 | 405B
7272| llama2_70b_lora | [ llama2_70b_lora] ( https://github.com/mlcommons/training/tree/master/llama2_70b_lora ) | pytorch | SCROLLS GovReport | 70B
7373| dlrm_dcnv2 | [ recommendation_v2] ( https://github.com/mlcommons/training/tree/master/recommendation_v2/torchrec_dlrm ) | torchrec | Criteo 3.5TB multi-hot | 167M
74- | rgat | [ graph_neural_network] ( https://github.com/mlcommons/training/tree/master/graph_neural_network ) | GLT | IGBFull | 25M
74+ | rgat | [ graph_neural_network] ( https://github.com/mlcommons/training/tree/master/graph_neural_network ) | GLT | IGBH-Full | 25M
7575
7676# MLPerf Training v4.1 (Submission Deadline Oct 11, 2024)
7777* Framework here is given for the reference implementation. Submitters are free to use their own frameworks to run the benchmark.
@@ -84,7 +84,7 @@ Some these benchmarks are rather slow or take a long time to run on the referenc
8484| GPT3 | [ language/llm] ( https://github.com/mlcommons/training/tree/master/large_language_model ) | paxml,megatron-lm | C4 | 175B
8585| LLama2 70B-LoRA | [ language/LLM fine-tuning] ( https://github.com/mlcommons/training/tree/master/llama2_70b_lora ) | pytorch | SCROLLS GovReport | 70B
8686| DLRMv2 | [ recommendation] ( https://github.com/mlcommons/training/tree/master/recommendation_v2/torchrec_dlrm ) | torchrec | Criteo 3.5TB multi-hot | 167M
87- | RGAT | [ GNN] ( https://github.com/mlcommons/training/tree/master/graph_neural_network ) | pytorch | IGBFull | 25M
87+ | RGAT | [ GNN] ( https://github.com/mlcommons/training/tree/master/graph_neural_network ) | pytorch | IGBH-Full | 25M
8888
8989* Note model parameter count is not the same as active parameter that are being trained in the benchmark.
9090
0 commit comments