Question Answering for Visual Navigation in Human-Centered Environments
2021 (English)In: Advances in Soft Computing: 20th Mexican International Conference on Artificial Intelligence, MICAI 2021, Mexico City, Mexico, October 25–30, 2021, Proceedings, Part II / [ed] Ildar Batyrshin, Alexander Gelbukh, Grigori Sidorov, Springer Nature, 2021, p. 31-45Conference paper, Published paper (Refereed)
Abstract [en]
In this paper, we propose an HISNav VQA dataset - a challenging dataset for a Visual Question Answering task that is aimed at the needs of Visual Navigation in human-centered environments. The dataset consists of images of various room scenes that were captured using the Habitat virtual environment and of questions important for navigation tasks using only visual information. We also propose a baseline for a HISNav VQA dataset, a Vector Semiotic Architecture, and demonstrate its performance. The Vector Semiotic Architecture is a combination of a Sign-Based World Model and Vector Symbolic Architectures. The Sign-Based World Model allows representing various aspects of an agent’s knowledge, and Vector Symbolic Architectures serve on a low computational level. The Vector Semiotic Architecture addresses the symbol grounding problem that plays an important role in the Visual Question Answering Task.
Place, publisher, year, edition, pages
Springer Nature, 2021. p. 31-45
Series
Lecture Notes in Artificial Intelligence, ISSN 0302-9743, E-ISSN 1611-3349
Keywords [en]
Visual question answering, Semiotic approach, Vector symbolic architecture, Habitat, Visual Navigation
National Category
Computer Sciences
Research subject
Dependable Communication and Computation Systems
Identifiers
URN: urn:nbn:se:ltu:diva-90559DOI: 10.1007/978-3-030-89820-5_3ISI: 000769230400003Scopus ID: 2-s2.0-85118354614OAI: oai:DiVA.org:ltu-90559DiVA, id: diva2:1657347
Conference
20th Mexican International Conference on Artificial Intelligence (MICAI 2021), Mexico City, Mexico, [ONLINE], October 25–30, 2021
Note
Funder: Russian Foundation for Basic Research, RFBR (19-37-90164)
2022-05-102022-05-102023-10-14Bibliographically approved