Skip to main content Skip to secondary navigation
Journal Article

DAWNBench: An End-to-End Deep Learning Benchmark and Competition

Abstract

Despite considerable research on systems, algorithms and hardware to speed up deep learning workloads, there is no standard means of evaluating end-to-end deep learning performance. Existing benchmarks measure proxy metrics, such as time to process one minibatch of data, that do not indicate whether the system as a whole will produce a high-quality result. In this work, we introduce DAWNBench, a benchmark and competition focused on end-to-end training time to achieve a state-of-the-art accuracy level, as well as inference time with that accuracy. Using time to accuracy as a target metric, we explore how different optimizations, including choice of optimizer, stochastic depth, and multi-GPU training, affect end-to-end training performance. Our results demonstrate that optimizations can interact in non-trivial ways when used in conjunction, producing lower speed-ups and less accurate models. We believe DAWNBench will provide a useful, reproducible means of evaluating the many trade-offs in deep learning systems.

Project page

A benchmark suite for end-to-end deep learning training and inference.
Author(s)
Cody Coleman
Deepak Narayanan
Daniel Kang
Tian Zhao
Jian Zhang
Luigi Nardi
Peter Bailis
Kunle Olukotun
Chris RĂ©
Matei Zaharia
Journal Name
31st Conference on Neural Information Processing Systems (NIPS 2017)
Publication Date
2017