Towards provable learning of polynomial neural networks using low-rank matrix estimation

Mohammadreza Soltani, Chinmay Hegde

    Research output: Contribution to conferencePaperpeer-review

    Abstract

    We study the problem of (provably) learning the weights of a two-layer neural network with quadratic activations. In particular, we focus on the under-parametrized regime where the number of neurons in the hidden layer is (much) smaller than the dimension of the input. Our approach uses a lifting trick, which enables us to borrow algorithmic ideas from low-rank matrix estimation. In this context, we propose two novel, nonconvex training algorithms which do not need any extra tuning parameters other than the number of hidden neurons. We support our algorithms with rigorous theoretical analysis, and show that the proposed algorithms enjoy linear convergence, fast running time per iteration, and near-optimal sample complexity. Finally, we complement our theoretical results with several numerical experiments.

    Original languageEnglish (US)
    Pages1417-1426
    Number of pages10
    StatePublished - 2018
    Event21st International Conference on Artificial Intelligence and Statistics, AISTATS 2018 - Playa Blanca, Lanzarote, Canary Islands, Spain
    Duration: Apr 9 2018Apr 11 2018

    Conference

    Conference21st International Conference on Artificial Intelligence and Statistics, AISTATS 2018
    Country/TerritorySpain
    CityPlaya Blanca, Lanzarote, Canary Islands
    Period4/9/184/11/18

    ASJC Scopus subject areas

    • Statistics and Probability
    • Artificial Intelligence

    Fingerprint

    Dive into the research topics of 'Towards provable learning of polynomial neural networks using low-rank matrix estimation'. Together they form a unique fingerprint.

    Cite this