This is a repository of reference implementations for the MLPerf training benchmarks. These implementations are valid as starting points for benchmark implementations but are not fully optimized and are not intended to be used for "real" performance measurements of software frameworks or hardware.
Please see the MLPerf Training Benchmark paper for a detailed description of the motivation and guiding principles behind the benchmark suite. If you use any part of this benchmark (e.g., reference implementations, submissions, etc.) in academic work, please cite the following:
@misc{mattson2019mlperf,
title={MLPerf Training Benchmark},
author={Peter Mattson and Christine Cheng and Cody Coleman and Greg Diamos and Paulius Micikevicius and David Patterson and Hanlin Tang and Gu-Yeon Wei and Peter Bailis and Victor Bittorf and David Brooks and Dehao Chen and Debojyoti Dutta and Udit Gupta and Kim Hazelwood and Andrew Hock and Xinyuan Huang and Atsushi Ike and Bill Jia and Daniel Kang and David Kanter and Naveen Kumar and Jeffery Liao and Guokai Ma and Deepak Narayanan and Tayo Oguntebi and Gennady Pekhimenko and Lillian Pentecost and Vijay Janapa Reddi and Taylor Robie and Tom St. John and Tsuguchika Tabaru and Carole-Jean Wu and Lingjie Xu and Masafumi Yamazaki and Cliff Young and Matei Zaharia},
year={2019},
eprint={1910.01500},
archivePrefix={arXiv},
primaryClass={cs.LG}
}
These reference implementations are still very much "alpha" or "beta" quality. They could be improved in many ways. Please file issues or pull requests to help us improve quality.
We provide reference implementations for benchmarks in the MLPerf suite, as well as several benchmarks under development.
Each reference implementation provides the following:
- Code that implements the model in at least one framework.
- A Dockerfile which can be used to run the benchmark in a container.
- A script which downloads the appropriate dataset.
- A script which runs and times training the model.
- Documentation on the dataset, model, and machine setup.
Follow instructions on the Readme of each benchmark. Generally, a benchmark can be run with the following steps:
- Setup docker & dependencies. There is a shared script (install_cuda_docker.sh) to do this. Some benchmarks will have additional setup, mentioned in their READMEs.
- Download the dataset using
./download_dataset.sh
. This should be run outside of docker, on your host machine. This should be run from the directory it is in (it may make assumptions about CWD). - Optionally, run
verify_dataset.sh
to ensure the was successfully downloaded. - Build and run the docker image, the command to do this is included with each Benchmark.
Each benchmark will run until the target quality is reached and then stop, printing timing results.
Some these benchmarks are rather slow or take a long time to run on the reference hardware. We expect to see significant performance improvements with more hardware and optimized implementations.
*Framework here is given for the reference implementation. Submitters are free to use their own frameworks to run the benchmark.
model | reference implementation | framework | dataset |
---|---|---|---|
resnet50v1.5 | vision/classification_and_detection | tensorflow2 | Imagenet |
RetinaNet | vision/object detection | pytorch | OpenImages |
3DUnet | vision/image segmentation | pytorch | KiTS19 |
Stable Diffusionv2 | image generation | pytorch | LAION-400M-filtered |
BERT-large | language/nlp | tensorflow | Wikipedia 2020/01/01 |
GPT3 | language/llm | paxml,megatron-lm | C4 |
LLama2 70B-LoRA | language/LLM fine-tuning | pytorch | SCROLLS govtReport |
DLRMv2 | recommendation | torchrec | Criteo 4TB multi-hot |
RGAT | GNN | pytorch | IGBFull |