Abstract
At Facebook, machine learning provides a wide range of capabilities that drive many aspects of user experience including ranking posts, content understanding, object detection and tracking for augmented and virtual reality, speech and text translations. While machine learning models are currently trained on customized datacenter infrastructure, Facebook is working to bring machine learning inference to the edge. By doing so, user experience is improved with reduced latency (inference time) and becomes less dependent on network connectivity. Furthermore, this also enables many more applications of deep learning with important features only made available at the edge. This paper takes a datadriven approach to present the opportunities and design challenges faced by Facebook in order to enable machine learning inference locally on smartphones and other edge platforms.
Original language | English (US) |
---|---|
Title of host publication | Proceedings - 25th IEEE International Symposium on High Performance Computer Architecture, HPCA 2019 |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
Pages | 331-344 |
Number of pages | 14 |
ISBN (Electronic) | 9781728114446 |
DOIs | |
State | Published - Mar 26 2019 |
Event | 25th IEEE International Symposium on High Performance Computer Architecture, HPCA 2019 - Washington, United States Duration: Feb 16 2019 → Feb 20 2019 |
Publication series
Name | Proceedings - 25th IEEE International Symposium on High Performance Computer Architecture, HPCA 2019 |
---|
Conference
Conference | 25th IEEE International Symposium on High Performance Computer Architecture, HPCA 2019 |
---|---|
Country/Territory | United States |
City | Washington |
Period | 2/16/19 → 2/20/19 |
Keywords
- Edge Inference
- Machine learning
ASJC Scopus subject areas
- Software
- Hardware and Architecture
- Computer Networks and Communications