UCL Discovery
UCL home » Library Services » Electronic resources » UCL Discovery

Make Haste Slowly: A Theory of Emergent Structured Mixed Selectivity in Feature Learning ReLU Networks

Jarvis, Devon; Klein, Richard; Rosman, Benjamin; Saxe, Andrew M; (2025) Make Haste Slowly: A Theory of Emergent Structured Mixed Selectivity in Feature Learning ReLU Networks. In: Proceedings of the Thirteenth International Conference on Learning Representations (ICLR 2025). (pp. pp. 1-35). OpenReview.net: Singapore, Singapore. Green open access

[thumbnail of 12212_Make_Haste_Slowly_A_Theo.pdf]
Preview
Text
12212_Make_Haste_Slowly_A_Theo.pdf - Published Version

Download (4MB) | Preview

Abstract

In spite of finite dimension ReLU neural networks being a consistent factor behind recent deep learning successes, a theory of feature learning in these models remains elusive. Currently, insightful theories still rely on assumptions including the linearity of the network computations, unstructured input data and architectural constraints such as infinite width or a single hidden layer. To begin to address this gap we establish an equivalence between ReLU networks and Gated Deep Linear Networks, and use their greater tractability to derive dynamics of learning. We then consider multiple variants of a core task reminiscent of multi-task learning or contextual control which requires both feature learning and nonlinearity. We make explicit that, for these tasks, the ReLU networks possess an inductive bias towards latent representations which are not strictly modular or disentangled but are still highly structured and reusable between contexts. This effect is amplified with the addition of more contexts and hidden layers. Thus, we take a step towards a theory of feature learning in finite ReLU networks and shed light on how structured mixed-selective latent representations can emerge due to a bias for node-reuse and learning speed.

Type: Proceedings paper
Title: Make Haste Slowly: A Theory of Emergent Structured Mixed Selectivity in Feature Learning ReLU Networks
Event: Thirteenth International Conference on Learning Representations (ICLR 2025)
Open access status: An open access version is available from UCL Discovery
Publisher version: https://openreview.net/forum?id=27SSnLl85x
Language: English
Additional information: This is an Open Access article published under a Creative Commons Attribution 4.0 International (CC BY 4.0) Licence, https://creativecommons.org/licenses/by/4.0/
Keywords: Gated Deep Linear Networks, Feature Learning Dynamics, Structured Mixed Selectivity, ReLU Networks
UCL classification: UCL
UCL > Provost and Vice Provost Offices > School of Life and Medical Sciences
UCL > Provost and Vice Provost Offices > School of Life and Medical Sciences > Faculty of Life Sciences
UCL > Provost and Vice Provost Offices > School of Life and Medical Sciences > Faculty of Life Sciences > Gatsby Computational Neurosci Unit
URI: https://discovery.ucl.ac.uk/id/eprint/10207673
Downloads since deposit
0Downloads
Download activity - last month
Download activity - last 12 months
Downloads by country - last 12 months

Archive Staff Only

View Item View Item