UCL Discovery
UCL home » Library Services » Electronic resources » UCL Discovery

Artificial Neuronal Ensembles with Learned Context Dependent Gating

Tilley, Matthew; Miller, Michelle; Freedman, David J; (2023) Artificial Neuronal Ensembles with Learned Context Dependent Gating. In: Proceedings of the Eleventh International Conference on Learning Representations, ICLR 2023. (pp. pp. 1-13). International Conference on Learning Representations Green open access

[thumbnail of 5379_artificial_neuronal_ensembles_.pdf]
Preview
Text
5379_artificial_neuronal_ensembles_.pdf - Published Version

Download (708kB) | Preview

Abstract

Biological neural networks are capable of recruiting different sets of neurons to encode different memories. However, when training artificial neural networks on a set of tasks, typically, no mechanism is employed for selectively producing anything analogous to these neuronal ensembles. Further, artificial neural networks suffer from catastrophic forgetting, where the network's performance rapidly deteriorates as tasks are learned sequentially. By contrast, sequential learning is possible for a range of biological organisms. We introduce Learned Context Dependent Gating (LXDG), a method to flexibly allocate and recall `artificial neuronal ensembles', using a particular network structure and a new set of regularization terms. Activities in the hidden layers of the network are modulated by gates, which are dynamically produced during training. The gates are outputs of networks themselves, trained with a sigmoid output activation. The regularization terms we have introduced correspond to properties exhibited by biological neuronal ensembles. The first term penalizes low gate sparsity, ensuring that only a specified fraction of the network is used. The second term ensures that previously learned gates are recalled when the network is presented with input from previously learned tasks. Finally, there is a regularization term responsible for ensuring that new tasks are encoded in gates that are as orthogonal as possible from previously used ones. We demonstrate the ability of this method to alleviate catastrophic forgetting on continual learning benchmarks. When the new regularization terms are included in the model along with Elastic Weight Consolidation (EWC) it achieves better performance on the benchmark `permuted MNIST' than with EWC alone. The benchmark `rotated MNIST' demonstrates how similar tasks recruit similar neurons to the artificial neuronal ensemble.

Type: Proceedings paper
Title: Artificial Neuronal Ensembles with Learned Context Dependent Gating
Event: The Eleventh International Conference on Learning Representations, ICLR 2023
Location: Kigali, Rwanda
Dates: 1st-5th May 2023
Open access status: An open access version is available from UCL Discovery
DOI: https://openreview.net/forum?id=dBk3hsg-n6
Publisher version: https://openreview.net/forum?id=dBk3hsg-n6
Language: English
Additional information: This version is the version of record. For information on re-use, please refer to the publisher's terms and conditions.
Keywords: Continual Learning, Catastrophic Forgetting
UCL classification: UCL
UCL > Provost and Vice Provost Offices > School of Life and Medical Sciences
UCL > Provost and Vice Provost Offices > School of Life and Medical Sciences > Faculty of Life Sciences
UCL > Provost and Vice Provost Offices > School of Life and Medical Sciences > Faculty of Life Sciences > Gatsby Computational Neurosci Unit
URI: https://discovery.ucl.ac.uk/id/eprint/10206343
Downloads since deposit
0Downloads
Download activity - last month
Download activity - last 12 months
Downloads by country - last 12 months

Archive Staff Only

View Item View Item