» Articles » PMID: 38288081

Learning Dynamics of Deep Linear Networks with Multiple Pathways

Overview
Date 2024 Jan 30
PMID 38288081
Authors
Affiliations
Soon will be listed here.
Abstract

Not only have deep networks become standard in machine learning, they are increasingly of interest in neuroscience as models of cortical computation that capture relationships between structural and functional properties. In addition they are a useful target of theoretical research into the properties of network computation. Deep networks typically have a serial or approximately serial organization across layers, and this is often mirrored in models that purport to represent computation in mammalian brains. There are, however, multiple examples of parallel pathways in mammalian brains. In some cases, such as the mouse, the entire visual system appears arranged in a largely parallel, rather than serial fashion. While these pathways may be formed by differing cost functions that drive different computations, here we present a new mathematical analysis of learning dynamics in networks that have parallel computational pathways driven by the same cost function. We use the approximation of deep linear networks with large hidden layer sizes to show that, as the depth of the parallel pathways increases, different features of the training set (defined by the singular values of the input-output correlation) will typically concentrate in one of the pathways. This result is derived analytically and demonstrated with numerical simulation with both linear and non-linear networks. Thus, rather than sharing stimulus and task features across multiple pathways, parallel network architectures learn to produce sharply diversified representations with specialized and specific pathways, a mechanism which may hold important consequences for codes in both biological and artificial systems.

Citing Articles

Flexible task abstractions emerge in linear networks with fast and bounded units.

Sandbrink K, Bauer J, Proca A, Saxe A, Summerfield C, Hummos A ArXiv. 2025; .

PMID: 39876939 PMC: 11774440.


Transition to chaos separates learning regimes and relates to measure of consciousness in recurrent neural networks.

Mastrovito D, Liu Y, Kusmierz L, Shea-Brown E, Koch C, Mihalas S bioRxiv. 2024; .

PMID: 38798582 PMC: 11118502. DOI: 10.1101/2024.05.15.594236.

References
1.
Yamins D, DiCarlo J . Using goal-driven deep learning models to understand sensory cortex. Nat Neurosci. 2016; 19(3):356-65. DOI: 10.1038/nn.4244. View

2.
Cadena S, Denfield G, Walker E, Gatys L, Tolias A, Bethge M . Deep convolutional models improve predictions of macaque V1 responses to natural images. PLoS Comput Biol. 2019; 15(4):e1006897. PMC: 6499433. DOI: 10.1371/journal.pcbi.1006897. View

3.
Felleman D, Van Essen D . Distributed hierarchical processing in the primate cerebral cortex. Cereb Cortex. 1991; 1(1):1-47. DOI: 10.1093/cercor/1.1.1-a. View

4.
Goodale M, Milner A . Separate visual pathways for perception and action. Trends Neurosci. 1992; 15(1):20-5. DOI: 10.1016/0166-2236(92)90344-8. View

5.
Wang Q, Gao E, Burkhalter A . Gateways of ventral and dorsal streams in mouse visual cortex. J Neurosci. 2011; 31(5):1905-18. PMC: 3040111. DOI: 10.1523/JNEUROSCI.3488-10.2011. View