UCL Discovery
UCL home » Library Services » Electronic resources » UCL Discovery

Nonlinear Meta-learning Can Guarantee Faster Rates

Meunier, Dimitri; Li, Zhu; Gretton, Arthur; Kpotufe, Samory; (2025) Nonlinear Meta-learning Can Guarantee Faster Rates. SIAM Journal on Mathematics of Data Science , 7 (4) pp. 1594-1615. 10.1137/24m1662977. Green open access

[thumbnail of 24m1662977.pdf]
Preview
PDF
24m1662977.pdf - Published Version

Download (657kB) | Preview

Abstract

Many recent theoretical works on meta-learning aim to achieve guarantees in leveraging similar representational structures from related tasks towards simplifying a target task. The main aim of theoretical guarantees on the subject is to establish the extent to which convergence rates—in learning a common representation—may scale with the number N of tasks (as well as the number of samples per task). First steps in this setting demonstrate this property when both the shared representation amongst tasks and task-specific regression functions are linear. This linear setting readily reveals the benefits of aggregating tasks, e.g., via averaging arguments. In practice, however, the representation is often highly nonlinear, introducing nontrivial biases in each task that cannot easily be averaged out as in the linear case. In the present work, we derive theoretical guarantees for meta-learning with nonlinear representations. In particular, assuming the shared nonlinearity maps to an infinite dimensional reproducing kernel Hilbert space, we show that additional biases can be mitigated with careful regularization that leverages the smoothness of task-specific regression functions, yielding improved rates that scale with the number of tasks as desired.

Type: Article
Title: Nonlinear Meta-learning Can Guarantee Faster Rates
Open access status: An open access version is available from UCL Discovery
DOI: 10.1137/24m1662977
Publisher version: https://doi.org/10.1137/24m1662977
Language: English
Additional information: This version is the version of record. For information on re-use, please refer to the publisher's terms and conditions.
Keywords: kernel methods, subspace approximation, nonparametric statistics
UCL classification: UCL
UCL > Provost and Vice Provost Offices > School of Life and Medical Sciences
UCL > Provost and Vice Provost Offices > School of Life and Medical Sciences > Faculty of Life Sciences
UCL > Provost and Vice Provost Offices > School of Life and Medical Sciences > Faculty of Life Sciences > Gatsby Computational Neurosci Unit
URI: https://discovery.ucl.ac.uk/id/eprint/10215595
Downloads since deposit
50Downloads
Download activity - last month
Download activity - last 12 months
Downloads by country - last 12 months

Archive Staff Only

View Item View Item