Visual Navigation subject to Embodied Mismatch

In the embodied visual navigation task, the agent navigates to a target location based on the visual observation it collects during the interaction with the environment. And various approaches have been proposed to learn robust navigation strategies for this task. However, existing approaches assume...

Full description

Saved in:
Bibliographic Details
Published inIEEE transactions on cognitive and developmental systems Vol. 15; no. 4; p. 1
Main Authors Liu, Xinzhu, Guo, Di, Liu, Huaping, Zhang, Xinyu, Sun, Fuchun
Format Journal Article
LanguageEnglish
Published Piscataway IEEE 01.12.2023
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:In the embodied visual navigation task, the agent navigates to a target location based on the visual observation it collects during the interaction with the environment. And various approaches have been proposed to learn robust navigation strategies for this task. However, existing approaches assume that the action spaces in the training and testing phases are the same, which is usually not the case in reality. And thus it is difficult to directly apply these approaches on practical scenarios. In this paper, we consider the situation where the action spaces in the training and testing phases are different, and a novel task of visual navigation subject to embodied mismatch is proposed. To solve the proposed task, we establish a two-stage robust adversary learning framework which can learn a robust policy to adapt the learned model to a new action space. In the first stage, an adversary training mechanism is used to learn a robust feature representation of the state. In the second stage, an adaptation training is used to transfer the learned strategy to a new action space with fewer training samples. Experiments of three types of embodied visual navigation tasks are conducted in 3D indoor scenes demonstrating the effectiveness of the proposed approach.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:2379-8920
2379-8939
DOI:10.1109/TCDS.2023.3238840