TY - GEN
T1 - Visually Analyzing Universal Quantifiers in Photograph Captions
AU - Sato, Yuri
AU - Mineshima, Koji
N1 - Funding Information:
supported by JSPS KAKENHI Grant Number
Funding Information:
This work was supported by JSPS KAKENHI Grant Number JP20K12782 to the first author.
Publisher Copyright:
© 2022, Springer Nature Switzerland AG.
PY - 2022
Y1 - 2022
N2 - Universal quantifiers have been the subject of much work in symbolic and diagrammatic logic. However, little attention is paid to the question of how they can be visually grounded, that is, depicted in real images such as photographs. To investigate this question, we focus on universal quantifiers such as “all” and “every” used in an image captioning dataset and present a qualitative analysis of these expressions. The analysis revealed that although the use of universal quantifiers in image captions is rare, there are interesting patterns in their usage in terms of the semantics of visual representations. We distinguish two ways in which universal quantifiers are used in image captions. One is object-based quantification, which involves quantification over multiple discrete objects in a definite domain. The other is region-based quantification, where some property is ascribed to a salient continuous region in an image. We compare these two ways of visually representing universal quantification with two major representation systems studied in diagrammatic logic.
AB - Universal quantifiers have been the subject of much work in symbolic and diagrammatic logic. However, little attention is paid to the question of how they can be visually grounded, that is, depicted in real images such as photographs. To investigate this question, we focus on universal quantifiers such as “all” and “every” used in an image captioning dataset and present a qualitative analysis of these expressions. The analysis revealed that although the use of universal quantifiers in image captions is rare, there are interesting patterns in their usage in terms of the semantics of visual representations. We distinguish two ways in which universal quantifiers are used in image captions. One is object-based quantification, which involves quantification over multiple discrete objects in a definite domain. The other is region-based quantification, where some property is ascribed to a salient continuous region in an image. We compare these two ways of visually representing universal quantification with two major representation systems studied in diagrammatic logic.
KW - Grounding
KW - Image caption
KW - Object
KW - Photograph
KW - Region
KW - Universal quantifier
KW - Visual representation
UR - http://www.scopus.com/inward/record.url?scp=85139071720&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85139071720&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-15146-0_34
DO - 10.1007/978-3-031-15146-0_34
M3 - Conference contribution
AN - SCOPUS:85139071720
SN - 9783031151453
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 373
EP - 377
BT - Diagrammatic Representation and Inference - 13th International Conference, Diagrams 2022, Proceedings
A2 - Giardino, Valeria
A2 - Linker, Sven
A2 - Burns, Richard
A2 - Bellucci, Francesco
A2 - Boucheix, Jean-Michel
A2 - Viana, Petrucio
PB - Springer Science and Business Media Deutschland GmbH
T2 - 13th International Conference on Theory and Application of Diagrams, Diagrams 2022, co-located with the IEEE Symposium on Visual Languages and Human-Centric Computing, VL/HCC 2022
Y2 - 13 September 2022 through 17 September 2022
ER -