@inproceedings{a007b24455b84706b92d67918a0368d4,
title = "Weightless: Lossy weight encoding for deep neural network compression",
abstract = "The large memory requirements of deep neural networks limit their deployment and adoption on many devices. Model compression methods effectively reduce the memory requirements of these models, usually through applying transformations such as weight pruning or quantization. In this paper, we present a novel scheme for lossy weight encoding co-designed with weight simplification techniques. The encoding is based on the Bloomier filter, a probabilistic data structure that can save space at the cost of introducing random errors. Leveraging the ability of neural networks to tolerate these imperfections and by re-training around the errors, the proposed technique, named Weightless, can compress weights by up to 496 × without loss of model accuracy. This results in up to a 1.51 × improvement over the state-of-the-art.",
author = "Brandon Reagen and Udit Gupta and Robert Adolf and Mitzenmacher, {Michael M.} and Rush, {Alexander M.} and Wei, {Gu Yeon} and David Brooks",
note = "Publisher Copyright: {\textcopyright} 35th International Conference on Machine Learning, ICML 2018.All Rights Reserved.; 35th International Conference on Machine Learning, ICML 2018 ; Conference date: 10-07-2018 Through 15-07-2018",
year = "2018",
language = "English (US)",
series = "35th International Conference on Machine Learning, ICML 2018",
publisher = "International Machine Learning Society (IMLS)",
pages = "6886--6899",
editor = "Andreas Krause and Jennifer Dy",
booktitle = "35th International Conference on Machine Learning, ICML 2018",
}