IMPLICIT REGULARIZATION FOR GROUP SPARSITY

Jiangyuan Li, Thanh V. Nguyen, Chinmay Hegde, Raymond K.W. Wong

    Research output: Contribution to conferencePaperpeer-review

    Abstract

    We study the implicit regularization of gradient descent towards structured sparsity via a novel neural reparameterization, which we call a “diagonally grouped linear neural network”. We show the following intriguing property of our reparameterization: gradient descent over the squared regression loss, without any explicit regularization, biases towards solutions with a group sparsity structure. In contrast to many existing works in understanding implicit regularization, we prove that our training trajectory cannot be simulated by mirror descent. We analyze the gradient dynamics of the corresponding regression problem in the general noise setting and obtain minimax-optimal error rates. Compared to existing bounds for implicit sparse regularization using diagonal linear networks, our analysis with the new reparameterization shows improved sample complexity. In the degenerate case of size-one groups, our approach gives rise to a new algorithm for sparse linear regression. Finally, we demonstrate the efficacy of our approach with several numerical experiments.

    Original languageEnglish (US)
    StatePublished - 2023
    Event11th International Conference on Learning Representations, ICLR 2023 - Kigali, Rwanda
    Duration: May 1 2023May 5 2023

    Conference

    Conference11th International Conference on Learning Representations, ICLR 2023
    Country/TerritoryRwanda
    CityKigali
    Period5/1/235/5/23

    ASJC Scopus subject areas

    • Language and Linguistics
    • Computer Science Applications
    • Education
    • Linguistics and Language

    Fingerprint

    Dive into the research topics of 'IMPLICIT REGULARIZATION FOR GROUP SPARSITY'. Together they form a unique fingerprint.

    Cite this