» Articles » PMID: 36161906

Data-driven Emergence of Convolutional Structure in Neural Networks

Overview
Specialty Science
Date 2022 Sep 26
PMID 36161906
Authors
Affiliations
Soon will be listed here.
Abstract

Exploiting data invariances is crucial for efficient learning in both artificial and biological neural circuits. Understanding how neural networks can discover appropriate representations capable of harnessing the underlying symmetries of their inputs is thus crucial in machine learning and neuroscience. Convolutional neural networks, for example, were designed to exploit translation symmetry, and their capabilities triggered the first wave of deep learning successes. However, learning convolutions directly from translation-invariant data with a fully connected network has so far proven elusive. Here we show how initially fully connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs, resulting in localized, space-tiling receptive fields. These receptive fields match the filters of a convolutional network trained on the same task. By carefully designing data models for the visual scene, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs, which has long been recognized as the hallmark of natural images. We provide an analytical and numerical characterization of the pattern formation mechanism responsible for this phenomenon in a simple model and find an unexpected link between receptive field formation and tensor decomposition of higher-order input correlations. These results provide a perspective on the development of low-level feature detectors in various sensory modalities and pave the way for studying the impact of higher-order statistics on learning in neural networks.

Citing Articles

Prediction of Member Forces of Steel Tubes on the Basis of a Sensor System with the Use of AI.

Li H, Chung H Sensors (Basel). 2025; 25(3).

PMID: 39943558 PMC: 11820130. DOI: 10.3390/s25030919.


Local kernel renormalization as a mechanism for feature learning in overparametrized convolutional neural networks.

Aiudi R, Pacelli R, Baglioni P, Vezzani A, Burioni R, Rotondo P Nat Commun. 2025; 16(1):568.

PMID: 39794337 PMC: 11724055. DOI: 10.1038/s41467-024-55229-3.


Parallel development of object recognition in newborn chicks and deep neural networks.

Pandey L, Lee D, Wood S, Wood J PLoS Comput Biol. 2024; 20(12):e1012600.

PMID: 39621774 PMC: 11651591. DOI: 10.1371/journal.pcbi.1012600.


Comparison of the Capacity of Several Machine Learning Tools to Assist Immunofluorescence-Based Detection of Anti-Neutrophil Cytoplasmic Antibodies.

Bertin D, Bongrand P, Bardin N Int J Mol Sci. 2024; 25(6).

PMID: 38542244 PMC: 10969855. DOI: 10.3390/ijms25063270.


Ensemble Learning, Deep Learning-Based and Molecular Descriptor-Based Quantitative Structure-Activity Relationships.

Matsuzaka Y, Uesawa Y Molecules. 2023; 28(5).

PMID: 36903654 PMC: 10005768. DOI: 10.3390/molecules28052410.


References
1.
Vidne M, Ahmadian Y, Shlens J, Pillow J, Kulkarni J, Litke A . Modeling the impact of common noise inputs on the network activity of retinal ganglion cells. J Comput Neurosci. 2011; 33(1):97-121. PMC: 3560841. DOI: 10.1007/s10827-011-0376-2. View

2.
van Bergen R, Kriegeskorte N . Going in circles is the way forward: the role of recurrence in visual inference. Curr Opin Neurobiol. 2020; 65:176-193. DOI: 10.1016/j.conb.2020.11.009. View

3.
Hubel D, Wiesel T . Receptive fields, binocular interaction and functional architecture in the cat's visual cortex. J Physiol. 1962; 160:106-54. PMC: 1359523. DOI: 10.1113/jphysiol.1962.sp006837. View

4.
DiCarlo J, Johnson K, Hsiao S . Structure of receptive fields in area 3b of primary somatosensory cortex in the alert monkey. J Neurosci. 1998; 18(7):2626-45. PMC: 6793113. View

5.
Knudsen E, Konishi M . Center-surround organization of auditory receptive fields in the owl. Science. 1978; 202(4369):778-80. DOI: 10.1126/science.715444. View