» Articles » PMID: 34170027

Do Humans and Deep Convolutional Neural Networks Use Visual Information Similarly for the Categorization of Natural Scenes?

Overview
Journal Cogn Sci
Specialty Psychology
Date 2021 Jun 25
PMID 34170027
Citations 4
Authors
Affiliations
Soon will be listed here.
Abstract

The investigation of visual categorization has recently been aided by the introduction of deep convolutional neural networks (CNNs), which achieve unprecedented accuracy in picture classification after extensive training. Even if the architecture of CNNs is inspired by the organization of the visual brain, the similarity between CNN and human visual processing remains unclear. Here, we investigated this issue by engaging humans and CNNs in a two-class visual categorization task. To this end, pictures containing animals or vehicles were modified to contain only low/high spatial frequency (HSF) information, or were scrambled in the phase of the spatial frequency spectrum. For all types of degradation, accuracy increased as degradation was reduced for both humans and CNNs; however, the thresholds for accurate categorization varied between humans and CNNs. More remarkable differences were observed for HSF information compared to the other two types of degradation, both in terms of overall accuracy and image-level agreement between humans and CNNs. The difficulty with which the CNNs were shown to categorize high-passed natural scenes was reduced by picture whitening, a procedure which is inspired by how visual systems process natural images. The results are discussed concerning the adaptation to regularities in the visual environment (scene statistics); if the visual characteristics of the environment are not learned by CNNs, their visual categorization may depend only on a subset of the visual information on which humans rely, for example, on low spatial frequency information.

Citing Articles

Novel applications of Convolutional Neural Networks in the age of Transformers.

Ersavas T, Smith M, Mattick J Sci Rep. 2024; 14(1):10000.

PMID: 38693215 PMC: 11063149. DOI: 10.1038/s41598-024-60709-z.


Automatic identification of stone-handling behaviour in Japanese macaques using LabGym artificial intelligence.

Ardoin T, Sueur C Primates. 2024; 65(3):159-172.

PMID: 38520479 DOI: 10.1007/s10329-024-01123-x.


Evaluation of Structural Retinal Layer Alterations in Retinitis Pigmentosa.

Yavuzer K, Citirik M, Yavuzer B Rom J Ophthalmol. 2024; 67(4):326-336.

PMID: 38239428 PMC: 10793365. DOI: 10.22336/rjo.2023.53.


Luminance and timing control during visual presentation of natural scenes.

De Cesarei A, Marzocchi M, Codispoti M HardwareX. 2022; 12:e00376.

PMID: 36437839 PMC: 9682347. DOI: 10.1016/j.ohx.2022.e00376.

References
1.
Cadieu C, Hong H, Yamins D, Pinto N, Ardila D, Solomon E . Deep neural networks rival the representation of primate IT cortex for core visual object recognition. PLoS Comput Biol. 2014; 10(12):e1003963. PMC: 4270441. DOI: 10.1371/journal.pcbi.1003963. View

2.
Yamins D, DiCarlo J . Using goal-driven deep learning models to understand sensory cortex. Nat Neurosci. 2016; 19(3):356-65. DOI: 10.1038/nn.4244. View

3.
Guclu U, van Gerven M . Deep Neural Networks Reveal a Gradient in the Complexity of Neural Representations across the Ventral Stream. J Neurosci. 2015; 35(27):10005-14. PMC: 6605414. DOI: 10.1523/JNEUROSCI.5023-14.2015. View

4.
De Cesarei A, Codispoti M, Schupp H, Stegagno L . Selectively attending to natural scenes after alcohol consumption: an ERP analysis. Biol Psychol. 2005; 72(1):35-45. DOI: 10.1016/j.biopsycho.2005.06.009. View

5.
Torralba A, Oliva A . Statistics of natural image categories. Network. 2003; 14(3):391-412. View