TY - JOUR
T1 - Regularizing neural networks via minimizing hyperspherical energy
AU - Lin, Rongmei
AU - Liu, Weiyang
AU - Liu, Zhen
AU - Feng, Chen
AU - Yu, Zhiding
AU - Rehg, James M.
AU - Xiong, Li
AU - Song, Le
N1 - Funding Information:
Acknowledgements. The research is partially supported by NSF BigData program under IIS-1838200, NSF CPS program under CMMI-1932187, and USDOT via C2SMART under 69A3551747124.
Publisher Copyright:
© 2020 IEEE.
PY - 2020
Y1 - 2020
N2 - Inspired by the Thomson problem in physics where the distribution of multiple propelling electrons on a unit sphere can be modeled via minimizing some potential energy, hyperspherical energy minimization has demonstrated its potential in regularizing neural networks and improving their generalization power. In this paper, we first study the important role that hyperspherical energy plays in neural network training by analyzing its training dynamics. Then we show that naively minimizing hyperspherical energy suffers from some difficulties due to highly non-linear and non-convex optimization as the space dimensionality becomes higher, therefore limiting the potential to further improve the generalization. To address these problems, we propose the compressive minimum hyperspherical energy (CoMHE) as a more effective regularization for neural networks. Specifically, CoMHE utilizes projection mappings to reduce the dimensionality of neurons and minimizes their hyperspherical energy. According to different designs for the projection mapping, we propose several distinct yet well-performing variants and provide some theoretical guarantees to justify their effectiveness. Our experiments show that CoMHE consistently outperforms existing regularization methods, and can be easily applied to different neural networks.
AB - Inspired by the Thomson problem in physics where the distribution of multiple propelling electrons on a unit sphere can be modeled via minimizing some potential energy, hyperspherical energy minimization has demonstrated its potential in regularizing neural networks and improving their generalization power. In this paper, we first study the important role that hyperspherical energy plays in neural network training by analyzing its training dynamics. Then we show that naively minimizing hyperspherical energy suffers from some difficulties due to highly non-linear and non-convex optimization as the space dimensionality becomes higher, therefore limiting the potential to further improve the generalization. To address these problems, we propose the compressive minimum hyperspherical energy (CoMHE) as a more effective regularization for neural networks. Specifically, CoMHE utilizes projection mappings to reduce the dimensionality of neurons and minimizes their hyperspherical energy. According to different designs for the projection mapping, we propose several distinct yet well-performing variants and provide some theoretical guarantees to justify their effectiveness. Our experiments show that CoMHE consistently outperforms existing regularization methods, and can be easily applied to different neural networks.
UR - http://www.scopus.com/inward/record.url?scp=85094836952&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85094836952&partnerID=8YFLogxK
U2 - 10.1109/CVPR42600.2020.00695
DO - 10.1109/CVPR42600.2020.00695
M3 - Conference article
AN - SCOPUS:85094836952
SN - 1063-6919
SP - 6916
EP - 6925
JO - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
JF - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
M1 - 9156828
T2 - 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020
Y2 - 14 June 2020 through 19 June 2020
ER -