TY - GEN
T1 - Fathom
T2 - 2016 IEEE International Symposium on Workload Characterization, IISWC 2016
AU - Adolf, Robert
AU - Rama, Saketh
AU - Reagen, Brandon
AU - Wei, Gu Yeon
AU - Brooks, David
N1 - Publisher Copyright:
© 2016 IEEE.
PY - 2016/10/3
Y1 - 2016/10/3
N2 - Deep learning has been popularized by its recent successes on challenging artificial intelligence problems. One of the reasons for its dominance is also an ongoing challenge: the need for immense amounts of computational power. Hardware architects have responded by proposing a wide array of promising ideas, but to date, the majority of the work has focused on specific algorithms in somewhat narrow application domains. While their specificity does not diminish these approaches, there is a clear need for more flexible solutions. We believe the first step is to examine the characteristics of cutting edge models from across the deep learning community. Consequently, we have assembled Fathom: a collection of eight archetypal deep learning workloads for study. Each of these models comes from a seminal work in the deep learning community, ranging from the familiar deep convolutional neural network of Krizhevsky et al., to the more exotic memory networks from Facebook's AI research group. Fathom has been released online, and this paper focuses on understanding the fundamental performance characteristics of each model. We use a set of application-level modeling tools built around the TensorFlow deep learning framework in order to analyze the behavior of the Fathom workloads. We present a breakdown of where time is spent, the similarities between the performance profiles of our models, an analysis of behavior in inference and training, and the effects of parallelism on scaling.
AB - Deep learning has been popularized by its recent successes on challenging artificial intelligence problems. One of the reasons for its dominance is also an ongoing challenge: the need for immense amounts of computational power. Hardware architects have responded by proposing a wide array of promising ideas, but to date, the majority of the work has focused on specific algorithms in somewhat narrow application domains. While their specificity does not diminish these approaches, there is a clear need for more flexible solutions. We believe the first step is to examine the characteristics of cutting edge models from across the deep learning community. Consequently, we have assembled Fathom: a collection of eight archetypal deep learning workloads for study. Each of these models comes from a seminal work in the deep learning community, ranging from the familiar deep convolutional neural network of Krizhevsky et al., to the more exotic memory networks from Facebook's AI research group. Fathom has been released online, and this paper focuses on understanding the fundamental performance characteristics of each model. We use a set of application-level modeling tools built around the TensorFlow deep learning framework in order to analyze the behavior of the Fathom workloads. We present a breakdown of where time is spent, the similarities between the performance profiles of our models, an analysis of behavior in inference and training, and the effects of parallelism on scaling.
UR - http://www.scopus.com/inward/record.url?scp=84994709898&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84994709898&partnerID=8YFLogxK
U2 - 10.1109/IISWC.2016.7581275
DO - 10.1109/IISWC.2016.7581275
M3 - Conference contribution
AN - SCOPUS:84994709898
T3 - Proceedings of the 2016 IEEE International Symposium on Workload Characterization, IISWC 2016
SP - 148
EP - 157
BT - Proceedings of the 2016 IEEE International Symposium on Workload Characterization, IISWC 2016
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 25 September 2016 through 27 September 2016
ER -