» Articles » PMID: 32347528

The Linguistic Analysis of Scene Semantics: LASS

Overview
Publisher Springer
Specialty Social Sciences
Date 2020 Apr 30
PMID 32347528
Citations 5
Authors
Affiliations
Soon will be listed here.
Abstract

In this paper, we define a new method for analyzing object-scene contextual relationships using computational linguistics: Linguistic Analysis of Scene Semantics, or LASS. LASS uses linguistic semantic similarity relationships between scene object and context labels embedded in a vector-space language model: Facebook Research's fastText. Importantly, the use of fastText permits semantic similarity score calculation between any set of strings and thus elements of any set of image data for which labels are available. Scene semantic similarity scores are then embedded in object segmentation mask locations in the image, creating a semantic similarity map. LASS can also be fully automated by generating context and object labels, as well as object segmentation masks, using deep learning. We compare semantic similarity maps between human- and neural network-generated annotations on a corpus of images taken from the LabelMe database. Semantic similarity maps produced by the fully automated LASS have a number of desirable properties, while maintaining a high degree of spatial and semantic similarity to them. Finally, we use LASS to evaluate the distribution of semantically consistent scene elements in space. Both show relatively uniform distributions of semantic relatedness to scene context, suggesting that contextually appropriate objects are likely to be found in all image regions. Taken together, these results suggest that LASS is accurate, automatic, flexible, and useful in a number of research contexts such as scene grammar and novelty detection.

Citing Articles

Quantifying task-related gaze.

Walter K, Freeman M, Bex P Atten Percept Psychophys. 2024; 86(4):1318-1329.

PMID: 38594445 PMC: 11093728. DOI: 10.3758/s13414-024-02883-w.


Visual search patterns during exploration of naturalistic scenes are driven by saliency cues in individuals with cerebral visual impairment.

Walter K, Manley C, Bex P, Merabet L Sci Rep. 2024; 14(1):3074.

PMID: 38321069 PMC: 10847433. DOI: 10.1038/s41598-024-53642-8.


Low-level factors increase gaze-guidance under cognitive load: A comparison of image-salience and semantic-salience models.

Walter K, Bex P PLoS One. 2022; 17(11):e0277691.

PMID: 36441789 PMC: 9704686. DOI: 10.1371/journal.pone.0277691.


Semantic object-scene inconsistencies affect eye movements, but not in the way predicted by contextualized meaning maps.

Pedziwiatr M, Kummerer M, Wallis T, Bethge M, Teufel C J Vis. 2022; 22(2):9.

PMID: 35171232 PMC: 8857618. DOI: 10.1167/jov.22.2.9.


Cognitive load influences oculomotor behavior in natural scenes.

Walter K, Bex P Sci Rep. 2021; 11(1):12405.

PMID: 34117336 PMC: 8196072. DOI: 10.1038/s41598-021-91845-5.

References
1.
Mohr J, Seyfarth J, Lueschow A, Weber J, Wichmann F, Obermayer K . BOiS-Berlin Object in Scene Database: Controlled Photographic Images for Visual Search Experiments with Quantified Contextual Priors. Front Psychol. 2016; 7:749. PMC: 4876128. DOI: 10.3389/fpsyg.2016.00749. View

2.
Olshausen B, Field D . How close are we to understanding v1?. Neural Comput. 2005; 17(8):1665-99. DOI: 10.1162/0899766054026639. View

3.
Rust N, Movshon J . In praise of artifice. Nat Neurosci. 2005; 8(12):1647-50. DOI: 10.1038/nn1606. View

4.
Anderson S, Chiu E, Huette S, Spivey M . On the temporal dynamics of language-mediated vision and vision-mediated language. Acta Psychol (Amst). 2010; 137(2):181-9. DOI: 10.1016/j.actpsy.2010.09.008. View

5.
Hayhoe M, Shrivastava A, Mruczek R, Pelz J . Visual memory and motor planning in a natural task. J Vis. 2003; 3(1):49-63. DOI: 10.1167/3.1.6. View