Abstract
We introduce DropConnect, a generalization of Dropout (Hinton et al., 2012), for regularizing large fully-connected layers within neural networks. When training with Dropout, a randomly selected subset of activations are set to zero within each layer. DropConnect instead sets a randomly selected subset of weights within the network to zero. Each unit thus receives input from a random subset of units in the previous layer. We derive a bound on the generalization performance of both Dropout and DropConnect. We then evaluate DropConnect on a range of datasets, comparing to Dropout, and show state-of-the-art results on several image recognition benchmarks by aggregating multiple DropConnect-trained models.
Original language | English (US) |
---|---|
Title of host publication | 30th International Conference on Machine Learning, ICML 2013 |
Publisher | International Machine Learning Society (IMLS) |
Pages | 2095-2103 |
Number of pages | 9 |
Edition | PART 3 |
State | Published - 2013 |
Event | 30th International Conference on Machine Learning, ICML 2013 - Atlanta, GA, United States Duration: Jun 16 2013 → Jun 21 2013 |
Other
Other | 30th International Conference on Machine Learning, ICML 2013 |
---|---|
Country/Territory | United States |
City | Atlanta, GA |
Period | 6/16/13 → 6/21/13 |
ASJC Scopus subject areas
- Human-Computer Interaction
- Sociology and Political Science