The statistical cost of robust kernel hyperparameter tuning

Raphael A. Meyer, Christopher Musco

    Research output: Contribution to journalConference articlepeer-review

    Abstract

    This paper studies the statistical complexity of kernel hyperparameter tuning in the setting of active regression under adversarial noise. We consider the problem of finding the best interpolant from a class of kernels with unknown hyperparameters, assuming only that the noise is square-integrable. We provide finite-sample guarantees for the problem, characterizing how increasing the complexity of the kernel class increases the complexity of learning kernel hyperparameters. For common kernel classes (e.g. squared-exponential kernels with unknown lengthscale), our results show that hyperparameter optimization increases sample complexity by just a logarithmic factor, in comparison to the setting where optimal parameters are known in advance. Our result is based on a subsampling guarantee for linear regression under multiple design matrices which may be of independent interest.

    Original languageEnglish (US)
    JournalAdvances in Neural Information Processing Systems
    Volume2020-December
    StatePublished - 2020
    Event34th Conference on Neural Information Processing Systems, NeurIPS 2020 - Virtual, Online
    Duration: Dec 6 2020Dec 12 2020

    ASJC Scopus subject areas

    • Computer Networks and Communications
    • Information Systems
    • Signal Processing

    Fingerprint

    Dive into the research topics of 'The statistical cost of robust kernel hyperparameter tuning'. Together they form a unique fingerprint.

    Cite this