TY  - JOUR
TI  - A neural-level model of spatial memory and imagery
N1  - This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
ID  - discovery10056563
UR  - https://doi.org/10.7554/eLife.33752
Y1  - 2018/09/04/
JF  - eLife
VL  - 7
SN  - 2050-084X
A1  - Bicanski, A
A1  - Burgess, N
KW  - computational model
KW  -  episodic memory
KW  -  neuroscience
KW  -  none
KW  -  scene construction
KW  -  spatial cognition
KW  -  spatially selective cells
KW  -  trace cells
N2  - We present a model of how neural representations of egocentric spatial experiences in parietal cortex interface with viewpoint-independent representations in medial temporal areas, via retrosplenial cortex, to enable many key aspects of spatial cognition. This account shows how previously reported neural responses (place, head-direction and grid cells, allocentric boundary- and object-vector cells, gain-field neurons) can map onto higher cognitive function in a modular way, and predicts new cell types (egocentric and head-direction-modulated boundary- and object-vector cells). The model predicts how these neural populations should interact across multiple brain regions to support spatial memory, scene construction, novelty-detection, 'trace cells', and mental navigation. Simulated behavior and firing rate maps are compared to experimental data, for example showing how object-vector cells allow items to be remembered within a contextual representation based on environmental boundaries, and how grid cells could update the viewpoint in imagery during planning and short-cutting by driving sequential place cell activity.
AV  - public
IS  - e33752
ER  -