TY - GEN
T1 - Efficient Learning of Sparse Representations with an Energy-Based Model
AU - Ranzato, Marc'Aurelio
AU - Poultney, Christopher
AU - Chopra, Sumit
AU - LeCun, Yann
N1 - Publisher Copyright:
© NIPS 2006.All rights reserved
PY - 2006
Y1 - 2006
N2 - We describe a novel unsupervised method for learning sparse, overcomplete features. The model uses a linear encoder, and a linear decoder preceded by a sparsifying non-linearity that turns a code vector into a quasi-binary sparse code vector. Given an input, the optimal code minimizes the distance between the output of the decoder and the input patch while being as similar as possible to the encoder output. Learning proceeds in a two-phase EM-like fashion: (1) compute the minimum-energy code vector, (2) adjust the parameters of the encoder and decoder so as to decrease the energy. The model produces “stroke detectors” when trained on handwritten numerals, and Gabor-like filters when trained on natural image patches. Inference and learning are very fast, requiring no preprocessing, and no expensive sampling. Using the proposed unsupervised method to initialize the first layer of a convolutional network, we achieved an error rate slightly lower than the best reported result on the MNIST dataset. Finally, an extension of the method is described to learn topographical filter maps.
AB - We describe a novel unsupervised method for learning sparse, overcomplete features. The model uses a linear encoder, and a linear decoder preceded by a sparsifying non-linearity that turns a code vector into a quasi-binary sparse code vector. Given an input, the optimal code minimizes the distance between the output of the decoder and the input patch while being as similar as possible to the encoder output. Learning proceeds in a two-phase EM-like fashion: (1) compute the minimum-energy code vector, (2) adjust the parameters of the encoder and decoder so as to decrease the energy. The model produces “stroke detectors” when trained on handwritten numerals, and Gabor-like filters when trained on natural image patches. Inference and learning are very fast, requiring no preprocessing, and no expensive sampling. Using the proposed unsupervised method to initialize the first layer of a convolutional network, we achieved an error rate slightly lower than the best reported result on the MNIST dataset. Finally, an extension of the method is described to learn topographical filter maps.
UR - http://www.scopus.com/inward/record.url?scp=85112276587&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85112276587&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85112276587
T3 - NIPS 2006: Proceedings of the 19th International Conference on Neural Information Processing Systems
SP - 1137
EP - 1144
BT - NIPS 2006
A2 - Scholkopf, Bernhard
A2 - Platt, John C.
A2 - Hofmann, Thomas
PB - MIT Press Journals
T2 - 19th International Conference on Neural Information Processing Systems, NIPS 2006
Y2 - 4 December 2006 through 7 December 2006
ER -