This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Temporal-Spatial Object Relations Modeling for Vision-and-Language Navigation

Bowen Huang ID , Yanwei Zheng1 ID , , Chuanlin Lan ID , Xinpeng Zhao ID , Yifei Zou ID , , Dongxiao yu ID 1Corresponding author. Bowen Huang, Yanwei Zheng, Xinpeng Zhao, Yifei Zou, and Dongxiao Yu are with the School of Computer Science and Technology, Institute of Intelligent Computing, Shandong University, Qingdao 266237, China (e-mail: huangbw@mail.sdu.edu.cn; zhengyw@sdu.edu.cn; zhaoxp1001@gmail.com; yfzou@sdu.edu.cn, dxyu@sdu.edu.cn). Chuanlin Lan is a Ph.D. student in the Department of Electrical Engineering at City University of Hong Kong, Hong Kong 999077, China(e-mail: cllan2-c@my.cityu.edu.hk).
Abstract

Vision-and-Language Navigation (VLN) is a challenging task where an agent is required to navigate to a natural language described location via vision observations. The navigation abilities of the agent can be enhanced by the relations between objects, which are usually learned using internal objects or external datasets. The relationships between internal objects are modeled employing graph convolutional network (GCN) in traditional studies. However, GCN tends to be shallow, limiting its modeling ability. To address this issue, we utilize a cross attention mechanism to learn the connections between objects over a trajectory, which takes temporal continuity into account, termed as Temporal Object Relations (TOR). The external datasets have a gap with the navigation environment, leading to inaccurate modeling of relations. To avoid this problem, we construct object connections based on observations from all viewpoints in the navigational environment, which ensures complete spatial coverage and eliminates the gap, called Spatial Object Relations (SOR). Additionally, we observe that agents may repeatedly visit the same location during navigation, significantly hindering their performance. For resolving this matter, we introduce the Turning Back Penalty (TBP) loss function, which penalizes the agent’s repetitive visiting behavior, substantially reducing the navigational distance. Experimental results on the REVERIE, SOON, and R2R datasets demonstrate the effectiveness of the proposed method.

Index Terms:
vision-and-language navigation, temporal object relations, spatial object relations, turning back penalty

I Introduction

In recent years, advancements in vision-and-language navigation (VLN) have demonstrated significant potential to revolutionize intelligent transportation systems, by enabling more intuitive and effective ways for autonomous vehicles to interpret and navigate complex urban environments [1, 2]. The goal of VLN [3, 4, 5, 6] is to guide an agent to a designated location, which is described by a natural language instruction. Although related vision-and-language problems have been extensively studied [7, 8, 9, 10], the VLN task remains highly challenging due to the dynamic nature of real-world environments and the complexity of the instructions involved.

Significant achievements have been made in the field of VLN [4, 11]. Most previous methods [12, 13, 14, 15] employ RNN models (e.g., GRUs and LSTMs) or transformer-based architectures to receive and store visual information, aligning it with the instruction to predict actions at each step. In recent developments, a number of techniques [16, 17, 18] have incorporated topological maps and semantic graphs to retain historical data, showing promising outcomes. Concurrently, a subset of research [19, 20, 21] has started to underscore the importance of modeling relationships between objects in navigational environments.

Refer to caption
(a) Graph-structured features
Refer to caption
(b) External knowledge
Refer to caption
(c) Temporal-spatial object relations modeling
Figure 1: Three methods of learning the connections between objects

A method for learning the relations between objects is constructing graph-structured feature representations [19, 22]. As depicted in Fig. 1(a), a graph-based navigation state is maintained utilizing GCN at each location during the agent’s navigation process (such as moving from position 1 to position 3). However, due to the issue of over-smoothing, GCN networks are typically kept shallow, which can impede their ability to accurately learn relationships. Another approach of modeling the object relations is introducing external knowledge. The external knowledge primarily originates from three sources: it is either derived from publicly available image-text datasets [23, 24, 17]; acquired through the use of the pretrained ConceptNet system [25, 26]; or obtained by employing large language models [27, 28, 29]. Nevertheless, as shown in Fig. 1(b), this external knowledge does not directly come from within the navigational environment, resulting in a significant gap between the object relations it forms and those in the real environment.

To address the above problems, we present two fundamental modules: the Temporal Object Relations (TOR) module and the Spatial Object Relations (SOR) module. As illustrated in Fig. 1(c), the TOR module captures the variations in positional relationships among objects from the moving trajectory of the agent. At each position along its trajectory, a cross attention mechanism is used to calculate a relationship matrix based on objects found and nouns in the instruction. This matrix is consistently updated as the agent moves, thereby modeling the object relations from a continuous temporal perspective. On the other hand, the SOR module is designed to model the spatial relations among objects from all locations within a navigational environment. Each viewpoint is considered equally in the environment, creating a relational database based on the objects discovered at these locations. This database includes the information of all viewpoints and ensures complete spatial coverage.

While these modules enhance the agent’s ability to understand and navigate complex environments, they also introduce a new challenge. The detailed tracking and continuous updating of object relations can lead the agent to explore new locations that are not part of the correct path or to conduct multiple explorations at the same location. Exploration across new locations is helpful, granting agents critical environmental insights and informing their navigational decisions [16, 30]. However, repetitive revisits to the same viewpoint do not enhance navigational success but rather impair efficiency. To counteract this, we introduce the Turning Back Penalty (TBP) loss function. Specifically, during the training process of the agent, it penalizes the agent each time it passes a previously visited location. This effectively mitigates the issue of revisits, thereby improving navigation efficiency. Our primary contributions can be summarized as follows:

  • \bullet

    We propose the TOR and SOR modules, which learn the interdependent relations among different objects from the dimensions of time and space, respectively.

  • \bullet

    We introduce the TBP loss function, which effectively alleviates the problem of excessive path length caused by repeated visits to the same location by the agent.

  • \bullet

    Extensive experiments have been conducted on the REVERIE [6], SOON [31], and R2R [4] datasets to demonstrate the superiority of our method over existing approaches in visual-and-language navigation.

The rest of the paper is organized as follows. Section II reviews relevant research about VLN. Section III introduces the details of our method. In Section IV, we present the training methodology and parameter settings of our model, and evaluate it on three datasets. We conclude the paper in Section V.

II Related work

II-A Vision-and-Language Navigation

VLN [32, 33, 34, 35, 36] has received significant research interests in recent years with the continual improvement. Early methods [12, 37, 32] usually utilize recurrent neural networks (RNNs) to encode historical observations and actions, which are represented as a state vector. In order to capture environment layouts, Wang et al. [34] employs a structured scene memory to accurately memorize the percepts during navigation. Tan et al. [14] proposes a two-stage training approach to enhance the generalization ability of agent. Ma et al. [38] use a progress monitor as a learnable heuristic for search.

More recently, transformer-based architectures have been shown successful in VLN tasks [39], notably by leveraging pre-trained architectures. PRESS [40] proposes a stochastic sampling scheme to reduce the considerable gap between the expert actions in training and sampled actions in test. VLN-BERT [11] employs recurrent units in transformer architecture to predict actions. To learn general navigation oriented textual representations, both AirBERT [3] and HM3D-AutoVLN [36] introduce expansive VLN dataset to enhance the interaction between various modalities. DUET [16] adeptly merges local observations with the overarching topological map through the use of graph transformers. This streamlines action planning and bolsters cross-modal comprehension. GridMM [18] builds a top-down egocentric and dynamically growing grid memory map to structure the visited environment. Different from them, our work propose an object-relations model designed to enhance the agent’s understanding of its environment.

II-B Object Relations Modeling

Recently, some studies have begun to focus on utilizing the relationships between objects to guide agent navigation [19, 41]. ORG [19] improves visual representation learning by integrating object relationships, including category closeness and spatial correlations. SEvol [22] propose a novel structured state-evolution model to learn the object-level relationship. CKR [26] propose a knowledge-enabled entity relationship reasoning module to learn the internal-external correlations among room- and object-entities. KERM [17] constructs an external knowledge base to assist in establishing relationships between the various entities described in the instructions. OAAM [42] utilizes two learnable attention modules to highlight language relating to objects and actions within a given instruction. VLMaps [43] translates natural language commands into a sequence of open-vocabulary navigation goals using large language models, resulting in objects that are spatially defined. Our method differs from others in modeling object relationships. It both considers temporal continuity and ensures complete spatial coverage. This integration results in a more accurate representation of the connections between objects.

II-C Training Regimes

Previous studies [16] mostly train the agent with the supervision from a pseudo interactive demonstrator similar to the DAgger algorithm [44]. Anderson et al. [4] introduces two distinct training regimes, teacher-forcing and student-forcing, and utilizes cross entropy loss at each step to maximize the likelihood of the ground-truth target action. Ma et al. [45] introduces a self-monitoring loss function that enhances the agent’s performance by estimating its navigation progress. Tan et al. [14] introduce an environment dropout method and enable the agent to navigate in environments with incomplete information and improving its generalization. Wang et al. [32] introduces a loss function that integrates reinforcement learning and self-supervised learning to optimize the agent’s matching capability across different modalities. These methods accumulate penalties for the agent as the number of exploratory steps increases. However, due to the poor balancing of penalty intensity, this may lead to either excessive revisitation of the same location by the agent or insufficient exploration of the environment. In contrast, our TBP loss function avoids such pitfalls by preventing the agent from retracing its steps while ensuring ample exploration of the environment.

Refer to caption
Figure 2: The overall network architecture. (a) The baseline utilizes a dual-scale encoder to encode local panoramic features, global historical features, and instruction features for action prediction of the agent. (b) At each time step tt, our method employ two modules to learn temporal and spatial object relations. Then the object relation features are combined with the image features for action prediction. Finally, we designs a novel TBP loss function to supervise the training of the agent in order to reduce its tendency to backtrack.

III Method

In the VLN task, the agent is initially located at a starting node in a previously unseen environment. The environment is represented by a weighted undirected graph 𝒢={𝒱,}\mathcal{G}=\left\{\mathcal{V},\mathcal{E}\right\}, where 𝒱\mathcal{V} denotes navigable nodes and \mathcal{E} denotes edges. The agent needs to explore this environment to reach the target location, guided by a natural language instruction. The instruction embedding consisting of LL words is 𝒲={𝒘i}i=1L\mathcal{W}=\left\{\boldsymbol{w}_{i}\right\}_{i=1}^{L}. At each time step tt, the agent observes a panoramic view of the current node. The panorama is divided into nn different perspective images t={𝒓i}i=1n\mathcal{R}_{t}=\left\{\boldsymbol{r}_{i}\right\}_{i=1}^{n}, where 𝒓i\boldsymbol{r}_{i} denotes the image feature of the i-th perspective and the direction encoding of that perspective. In addition, mm object features 𝒪t={𝒐i}i=1m\mathcal{O}_{t}=\left\{\boldsymbol{o}_{i}\right\}_{i=1}^{m} are extracted from the panorama. This is done using annotated object bounding boxes or automatic object detectors [46], enhancing the agent’s fine-grained visual perception.

III-A Overview of Our Approach

As shown in Fig. 2(a), we adopt the architecture of DUET [16] as the baseline, It consists of three inputs: panoramic visual features of the current location, a topological map, and an instruction. At time step tt, the topological map is represented as 𝒢t={𝒱t,t}\mathcal{G}_{t}=\left\{\mathcal{V}_{t},\mathcal{E}_{t}\right\}. Here, 𝒢t\mathcal{G}_{t} is a subset of the overall map 𝒢\mathcal{G} and encapsulates the state of the environment after tt steps of navigation. 𝒱t\mathcal{V}_{t} contains three kinds of nodes: visited nodes (circular nodes), the current node (pentagram nodes), and navigable nodes (triangular nodes).

Our method, as illustrated in Fig. 2(b), employs two modules to calculate temporal object features t\mathcal{M}_{t} and spatial object features 𝒩t\mathcal{N}_{t} at each step tt, respectively. These features are then combined with 𝒪t\mathcal{O}_{t} and t\mathcal{R}_{t}, in order to generate the panoramic feature 𝒬t\mathcal{Q}_{t}, which is fed into a dual-scale encoder to predict the agent’s action. To further enhance the agent’s performance, we introduce the TBP loss function. It can help to prevent repeated explorations and reduce the length of the agent’s path.

III-B Object Relations

In the VLN task, there are connections between the various objects that the agent perceives. In our method, these connections are learned from two dimensions of time and space, significantly enhancing the accuracy of the navigation.

III-B1 Object nouns features

To extract object-related noun features from the instruction, we begin by choosing word embeddings describing objects from 𝒲\mathcal{W}. Specifically, our process begins by obtaining labels for all objects from the MatterPort3D simulator [47]. These labels are then compiled into a noun database, denoted as DD. For a given natural language instruction 𝒲\mathcal{W}, we iterate through each word. If a word is found within DD, it is selected as an object-related token.

Upon acquiring these object-related embeddings, we enhance them with positional embeddings as described in [48]. These positional embeddings correspond to the respective word’s location within the sentence. Additionally, a type embedding specific to text, as outlined in [49], is also incorporated. Next, we input all noun tokens into a text encoder that consists of a multi-layer transformer. This process generates contextual noun representations, which we refer to as 𝒲^={𝒘^1,𝒘^2,,𝒘^L^}{\mathcal{\hat{W}}}=\{\boldsymbol{\hat{w}}_{1},\boldsymbol{\hat{w}}_{2},\dots,\boldsymbol{\hat{w}}_{\hat{L}}\}.

Refer to caption
(a) As the agent interacts with the environment, it continuously updates its understanding of these relationships.
Refer to caption
(b) The spatial object relationship matrix is constructed by considering the objects visible at each node in the training environment.
Figure 3: Learning methods for two kinds of relationships

III-B2 Temporal object relations

In the study of object relations, to circumvent the issue of inadequate learning capabilities resulting from the shallow nature of GCN, we have designed a temporal object relations (TOR) module. This module employs a cross attention mechanism to focus on objects observed during navigation and noun embeddings in the instruction. Through this approach, we learn a relationship matrix accurately as the agent progresses along its exploration trajectory. This ensures temporal continuity in the agent’s learning process.

Fig. 3(a) depicts how the agent, when arriving at a new location, establishes connections between the objects it perceives and the relevant nouns from the navigation instruction. Specifically, when the agent reaches a location, it obtains a panoramic view of that position. The agent employs a cross attention mechanism to learn the associations between all objects discovered at this location and the nouns mentioned in the instruction. This process is consistently applied at each location along the agent’s navigational path. It sets the stage for the agent to learn and progressively refine the inter-object relationships across temporal dimensions.

In our approach, we treat the object features, denoted as OtO_{t}, as the query, and the noun features, W^\hat{W}, as the key. We employ a cross attention mechanism to compute the relationship matrix TT, which can be expressed in the following way:

𝐓=FC(𝒪t)𝒲^\mathbf{T}={\rm FC}\left(\mathcal{O}_{t}\right)\mathcal{\hat{W}} (1)

where FC\rm FC is a fully connected layer.

We leverage 𝐓\mathbf{T} to derive the temporal object features t\mathcal{M}_{t}. It is formalized as follow:

t=𝐓𝒲^\mathcal{M}_{t}=\mathbf{T}\mathcal{\hat{W}} (2)

III-B3 Spatial object relations

Due to the gap between external knowledge and the navigation environment, the agent is unable to accurately learn the relationships between objects based on external knowledge. To bridge this gap, we introduce the spatial object relations (SOR) module. This module considers all viewpoints in the environment and ensures complete spatial coverage, which help the agent learn a more precise object relations.

During the establishment of spatial object relations, the agent proceeds to update an initially zero-valued relationship matrix 𝐄\mathbf{E} based on the objects identified at each respective location. As illustrated in Fig. 3(b), for objects xx and yy observed concurrently at the same location, a shorter distance between them correlates with a stronger association. Consequently, we update the relationship matrix based on the distances between objects as seen by the agent at each location within the environment. The update rule for 𝐄\mathbf{E} is formalized as follows:

𝐄(x,y)+=k1k2𝒗x𝒗y2+k3𝒅x𝒅y2\mathbf{E}\left(x,y\right)+=\frac{k_{1}}{k_{2}\|\boldsymbol{v}_{x}-\boldsymbol{v}_{y}\|_{2}+k_{3}\|\boldsymbol{d}_{x}-\boldsymbol{d}_{y}\|_{2}} (3)

where 𝒗x\boldsymbol{v}_{x} and 𝒗y\boldsymbol{v}_{y} denote the perspectives from the current position in relation to objects xx and yy, respectively, and 𝒅x\boldsymbol{d}_{x} and 𝒅y\boldsymbol{d}_{y} represent the respective depths of objects xx and yy from the current position. Here, k1k_{1}, k2k_{2}, and k3k_{3} are predefined constants, and 2\|\cdot\|_{2} denotes the L2 norm.

In the course of the agent’s training, the pertinent matrix 𝐄\mathbf{E}^{\prime} is retrieved from 𝐄\mathbf{E}, guided by the objects detected at the agent’s current location and the nouns encapsulated within the instruction. Assuming the agent currently discovers c1c_{1} objects, and the current instruction contains c2c_{2} nouns, the calculation of 𝐄\mathbf{E}^{\prime} is as follows:

𝐄=[𝐄p(1)q(1)𝐄p(1)q(2)𝐄p(1)q(3)𝐄p(1)q(c2)𝐄p(2)q(1)𝐄p(2)q(2)𝐄p(2)q(3)𝐄p(2)q(c2)𝐄p(c1)q(1)𝐄p(c1)q(2)𝐄p(c1)q(3)𝐄p(c1)q(c2)]\mathbf{E}^{\prime}=\begin{bmatrix}\mathbf{E}_{p(1)q(1)}&\mathbf{E}_{p(1)q(2)}&\mathbf{E}_{p(1)q(3)}&\dots&\mathbf{E}_{p(1)q(c_{2})}\\ \mathbf{E}_{p(2)q(1)}&\mathbf{E}_{p(2)q(2)}&\mathbf{E}_{p(2)q(3)}&\dots&\mathbf{E}_{p(2)q(c_{2})}\\ \dots&\dots&\dots&\dots&\dots\\ \mathbf{E}_{p(c_{1})q(1)}&\mathbf{E}_{p(c_{1})q(2)}&\mathbf{E}_{p(c_{1})q(3)}&\dots&\mathbf{E}_{p(c_{1})q(c_{2})}\\ \end{bmatrix}

(4)

where p(i)p(i) is the index of the i-th object and q(j)q(j) is the index of the j-th noun. Subsequently, matrix multiplication is employed to derive the environmental object feature 𝒩t\mathcal{N}_{t}, as per the following equation:

𝒩t=𝐄𝒲^\mathcal{N}_{t}=\mathbf{E}^{\prime}\mathcal{\hat{W}} (5)

Upon obtaining both the temporal object features t\mathcal{M}_{t} and the spatial object features 𝒩t\mathcal{N}_{t}, the final object relationship feature 𝒬t\mathcal{Q}_{t} is computed utilizing the equation:

𝒬t=α1𝒪t+α2t+α3𝒩t\mathcal{Q}_{t}=\alpha_{1}\mathcal{O}_{t}+\alpha_{2}\mathcal{M}_{t}+\alpha_{3}\mathcal{N}_{t} (6)

where α1\alpha_{1}, α2\alpha_{2}, and α3\alpha_{3} are three learnable parameters, summing to 1. Subsequently, a concatenation of 𝒬t\mathcal{Q}_{t} and the image feature t\mathcal{R}_{t} is performed to yield the panoramic feature t=[t,𝒬t]\mathcal{F}_{t}=\left[\mathcal{R}_{t},\mathcal{Q}_{t}\right]

III-C Turning back penalty

After introducing the TOR and SOR modules, we observed the agent tended to revisit the same location multiple times. This usually culminates in an elongated navigation path. For instance, an agent’s journey initiates at point aa and concludes at point ee, successfully navigating through the route abcdbea\rightarrow b\rightarrow c\rightarrow d\rightarrow b\rightarrow e. It can be observed that the direct path composed of abea\rightarrow b\rightarrow e would be optimal. The exploration of vertices cc and dd represents additional exploration by the agent, while revisiting vertex bb indicates redundant exploration. Such additional exploration is beneficial as it enables the agent to acquire new knowledge, thereby enhancing its navigational skills. Conversely, redundant exploration does not improve navigation efficiency and leads to unnecessarily prolonged paths, which is disadvantageous.

To address the issue of the agent frequently revisiting the same location, we have developed a new loss function, named Turning Back Penalty (TBP). This function introduces a punitive measure to discourage the agent from redundant navigation, fostering a more streamlined and direct trajectory. Concretely, let us consider a scenario wherein the agent is positioned at location aa, and it has a set of rr navigable positions, denoted as {b1,b2,,br}\left\{b_{1},b_{2},\dots,b_{r}\right\}.

LTBP=i=1repidij=1repjL_{TBP}=\sum_{i=1}^{r}\frac{e^{p_{i}}d_{i}}{\sum_{j=1}^{r}e^{p_{j}}} (7)

where pip_{i} symbolizes the probability of transitioning from location aa to location bib_{i}, and did_{i} represents the cumulative length of the paths that have been traversed repetitively by the agent in the course of navigating from aa to bib_{i}.

III-D Training and Inference

III-D1 Pretrainging

Building upon the foundational work presented in [50, 51, 39], the advantageous impacts of pretraining transformer-based models within the domain of Vision-and-Language Navigation (VLN) have been well-established. In this work, we employ four auxiliary tasks to pretrain our proposed model.

Masked Language Modeling (MLM)

In accordance with [48], we randomly mask out certain words in the instructions with a probability of 15%. Subsequently, we use the contextual information within the instructions to predict the masked words.

Masked Region Classification (MRC)

In the MRC [52] task, the model need to identify the semantic labels of images where certain views are masked. We employ a random masking of view images at a rate of 15%. In alignment with the methodology in [16], we use an image classification model [53] pre-trained on ImageNet to generate target labels for these view images.

Single-step Action Prediction (SAP)

In the SAP [54] task, the agent is required to infer its next action based on its previous actions. The SAP loss in behavior cloning given a demonstration path PP^{*} is as follows:

LSAP=t=1Tlogp(at𝒲,𝒫<t)L_{SAP}=\sum_{t=1}^{T}-\log p\left(a_{t}^{*}\mid\mathcal{W},\mathcal{P}_{<t}^{*}\right) (8)

where 𝒫<t\mathcal{P}_{<t}^{*} represents a partial demonstration path, ata_{t}^{*} is the expert action of 𝒫<t\mathcal{P}_{<t}^{*}.

Object Grounding (OG)

For tasks with object annotations, the object grounding [55] loss is employed.

LOG=logp(o𝒲,𝒫T)L_{OG}=-\log p\left(o^{*}\mid\mathcal{W},\mathcal{P}_{T}\right) (9)

where oo^{*} refers to the object category at the agent’s final destination 𝒫T\mathcal{P}_{T} .

III-D2 Fine-tuning and Inference

In the phase of fine-tuning, we employ a triad of loss functions: Single-step Action Prediction (SAP), Object Grounding (OG), and Turning Back Penalty (TBP), to meticulously guide and refine the learning process of the agent. Distinct from the approach of utilizing a demonstration path in the pre-training phase, during fine-tuning, the model is supervised by a pseudo-interactive demonstrator. This pseudo-interactive demonstrator operates by employing the shortest path algorithm at each decision point, selecting a node as the ensuing location. This selection is made such that it adheres to the criterion of minimizing the overall path length from the agent’s current location to the target destination. The cumulative loss function that governs the fine-tuning process is formulated as follows:

L=λ1LSAP+λ2LOG+λ3LTBPL=\lambda_{1}L_{SAP}+\lambda_{2}L_{OG}+\lambda_{3}L_{TBP} (10)

In this expression, λ1\lambda_{1}, λ2\lambda_{2}, and λ3\lambda_{3} serve as balance factors, ensuring a harmonious integration of the individual loss components.

For inference, the agent predicts one action at each step. If the action is not a stop action, it executes the action and moves to the appropriate position. Otherwise, the agent stops at the current position. When the number of actions executed by the agent exceeds the maximum number of actions specified, it is forced to stop and the node with the maximum stop probability is taken as the final position. When the agent stops, it will take the object with the maximum prediction score as the target.

IV Experiments

IV-A Datasets

In this paper, we mainly focus on using relationships between objects to improve the ability of agent to perform visual-and-language navigation, so we choose two datasets with object annotations, REVERIE [6] and SOON [31], to evaluate our model. Additionally, we also present the results of our model on the R2R [4] dataset, which lacks object annotations.

REVERIE dataset mainly consists of instructions that describe target locations and objects of interest, averaging 21 words per instruction. It provides the agent with bounding boxes for each object in various panoramas. The agent must correctly identify and choose the right object bounding box at the end of its navigation. Paths demonstrated by experts in this dataset vary in length, ranging from 4 to 7 steps.

SOON dataset provides detailed instructions that accurately identify target rooms and objects, averaging 47 words in length. Unlike other datasets, SOON does not include predefined bounding boxes for objects. This requires the agent to predict objects’ central locations within the panoramas. To facilitate this, we utilize an automatic object detection approach as described in [46]. This method helps us generate potential bounding boxes for the objects. The lengths of expert demonstrations in SOON are varied, ranging from 2 to 21 steps, with an average of approximately 9.5 steps.

R2R dataset encompasses a total of 21,567 words, with an average instruction length of 29 words. Due to the absence of object annotations in this dataset, we substitute object features with features from panoramic images. In experiments, we solely train the agent using temporal object relations.

IV-B Evaluation Metrics

To assess the performance of our method in comparison to previous works, we adopt the conventional evaluation metrics for visual-and-language navigation task, as delineated in [4, 6]. These metrics encompass: (1) Trajectory Length (TL)—the agent’s average path length in meters; (2) Navigation Error (NE)—average distance in meters between agent’s final location and the target; (3) Success Rate (SR)—the percentage of instructions that are successfully executed, with an NE smaller than 3 meters; (4) Oracle SR (OSR)—SR given the oracle stop policy; (5) SPL—SR weighted by Path Length; (6) Remote Grounding Success (RGS)—the percentage of instructions that are executed successfully.; (7) RGSPL—RGS penalized by Path Length. Except for TL and NE, all metrics are higher the better.

IV-C Implementation Details

IV-C1 Model Architectures

For the extraction of object features within the REVERIE dataset, which conveniently offers bounding boxes, we employ the ViT-B/16 model [53], pretrained on the ImageNet dataset. In contrast, when dealing with the SOON dataset, object bounding boxes are extracted using the BUTD object detector [46] due to its absence of provided bounding boxes. For the R2R dataset, we do not employ the spatial object relations module because of the unavailability of object annotations.

Further delving into our architecture, we incorporate a dual-scale graph transformer [16]. The specific configuration of this transformer includes setting the number of layers for the language encoder, panorama encoder, coarse-scale cross-modal encoder, and fine-scale cross-modal encoder to 9, 2, 4, and 4, respectively. The parameters for this segment of our model are initialized using the pretrained LXMERT model [49].

In the process of computing the relationship matrix \mathcal{E}, we assigned specific values to the parameters k1k_{1}, k2k_{2}, and k3k_{3}, setting them at 2, 2, and 5e45e^{-4}, respectively. Furthermore, in order to compute the object relationship feature 𝒬\mathcal{Q}, the weight parameters α1\alpha_{1}, α2\alpha_{2}, and α3\alpha_{3} were meticulously initialized to 0.8, 0.1, and 0.1, respectively. Due to lacking spatial object relations module, we set α1\alpha_{1}, α2\alpha_{2}, and α3\alpha_{3} as 0.8, 0.2, and 0 when using R2R dataset. Analogously, for the precise calculation of the loss function LL, we established the values of the weight parameters λ1\lambda_{1}, λ2\lambda_{2}, and λ3\lambda_{3} at 1, 1, and 0.2, respectively.

TABLE I: Comparison with the state-of-the-art methods on the REVERIE dataset. The baseline indicates the replicated results of DUET.
Methods Val seen Val Unseen Test Unseen
Navigation Grounding Navigation Grounding Navigation Grounding
TL OSR SR SPL RGS RGSPL TL OSR SR SPL RGS RGSPL TL OSR SR SPL RGS RGSPL
Seq2Seq [4] 12.88 35.70 29.59 24.01 18.97 14.96 11.07 8.07 4.20 2.84 2.16 1.63 10.89 6.88 3.99 3.09 2.00 1.58
RCM [37] 10.70 29.44 23.33 21.82 13.23 15.36 11.98 14.23 9.39 6.97 4.89 3.89 10.60 11.68 7.84 6.67 3.67 3.14
VLNBERT [11] 13.44 53.90 51.79 47.96 38.23 35.61 16.78 35.02 30.67 24.90 18.77 15.27 15.68 32.91 29.61 23.99 16.50 13.51
AirBERT [3] 15.16 49.98 47.01 42.34 32.75 30.01 18.71 34.51 27.89 21.88 18.23 14.18 17.91 34.20 30.28 23.61 16.83 13.28
HOP [35] 13.80 54.88 53.76 47.19 38.65 33.85 16.46 36.24 31.78 26.11 18.85 15.73 16.38 33.06 30.17 24.34 17.69 14.34
HAMT [54] 12.79 47.65 43.29 40.19 27.20 15.18 14.08 36.84 32.95 30.20 18.92 17.28 13.62 33.41 30.40 26.67 14.88 13.08
CKR [26] 12.16 61.91 57.27 53.57 39.07 - 26.26 31.44 19.14 11.84 11.45 - 22.46 30.40 22.00 14.25 11.60 -
DUET [16] 13.86 73.68 71.75 63.94 57.41 51.14 22.11 51.07 46.98 33.73 32.15 23.03 21.30 56.91 52.51 36.06 31.88 22.06
KERM [17] 12.84 79.20 76.88 70.45 61.00 56.07 21.85 55.21 50.44 35.38 34.51 24.45 17.32 57.58 52.43 39.21 32.39 23.64
BEVBert [56] - - - - - - - 56.40 51.78 36.37 34.71 24.44 - 57.26 52.81 36.41 32.06 22.09
GridMM [18] - - - - - - 23.20 57.48 51.37 36.47 34.57 24.56 19.97 59.55 53.13 36.60 34.87 23.45
Baseline 13.84 73.79 71.68 63.90 57.34 51.11 22.12 51.09 46.98 33.75 32.15 23.05 18.19 55.31 50.67 36.27 31.87 22.65
Ours 14.00 83.06 80.46 73.12 64.02 58.29 22.00 55.55 50.30 36.84 35.27 25.98 20.25 61.08 55.31 40.37 35.16 24.99
TABLE II: Comparison with the state-of-the-art methods on the SOON dataset. The baseline indicates the replicated results of DUET.
Methods Val Unseen Test Unseen
TL OSR SR SPL RGSPL TL OSR SR SPL RGSPL
GBE [57] 28.96 28.54 19.52 13.34 1.16 27.88 21.45 12.90 9.23 0.45
DUET [16] 36.20 50.91 36.28 22.58 3.75 41.83 43.00 33.44 21.42 4.17
KERM [17] 35.83 51.62 38.05 23.16 4.04 - - - - -
GridMM [18] 38.92 53.39 37.46 24.81 3.91 46.20 48.02 36.27 21.25 4.15
Baseline 36.18 50.88 36.17 22.54 3.75 40.73 41.74 32.05 20.79 4.55
Ours 38.75 55.46 40.12 26.00 5.04 40.05 47.73 37.09 23.60 6.35

IV-C2 Training Details

For pretraining, we set the batch size as 32 using 1 NVIDIA RTX3090 GPU. For the REVERIE dataset, we combine the original dataset with augmented data synthesized by DUET [16] to pretrain our model with 100k iterations. Then we fine-tune the pretrained model with the batch size of 16 for 20k iterations on 1 NVIDIA RTX3090 GPU. For the SOON dataset, we only use the original data with automatically cleaned object bounding boxes, sharing the same settings in DUET [16]. We pretrain the model with 40k iterations. Then we fine-tune the pretrained model with the batch size of 4 for 40k iterations on 1 NVIDIA RTX3090 GPU. For the R2R dataset, additional augmented R2R data in [39] is used in pretraining. We pretrain the model for 200k iterations with batch size of 64 and then fine-tune it for 20k iterations with batch size of 8.

IV-D Performance Comparison

The results of our method on the REVERIE and SOON datasets are depicted in Table I and Table II, respectively. Across most metrics on these two datasets, our approach achieves superior performance. Specifically, in the test split of REVERIE, as detailed in Table I, our method achieves notable improvements over the baseline: 4.64% on SR, 4.10% on SPL, 3.29% on RGS, and 2.34% on RGSPL. This substantial enhancement underscores the robustness and efficacy of our technique. Moreover, even when compared to the current state-of-the-art method, GridMM[18], our method still shows advancements of 2.18%, 3.77%, 0.29%, and 1.54% on SR, SPL, RGS, and RGSPL, respectively, highlighting the superior capability of our approach. As indicated in Table II, on the more intricate SOON dataset, our method also manifests exceptional performance, surpassing the current state-of-the-art. This underscores our method’s proficiency in grasping inter-object relations, thereby enhancing the agent’s navigational prowess.

In our study, we also conducted experiments on the R2R dataset, which lacks object annotations. As shown in Table III, the absence of explicit object information impeded the agent’s ability to accurately learn the relationships between objects in the navigation environment. Our method does not exhibit significant improvements over other methods in both the val unseen and test splits across various metrics. Interestingly, we observes a remarkable phenomenon: on the val seen split, our method significantly outperforms both the baseline and other approaches. This can be attributed to the agent’s repeated exposure to the panoramas in the val seen split during training. Despite the lack of explicit object annotations, the agent often manages to infer the presence and relationships of objects based on these panoramic features. These findings further illustrate the vital role of object relationships in enabling an agent to accurately complete navigation tasks. This aspect of our research highlights the significance of understanding and integrating object interactions within the navigational context for improved agent performance.

IV-E Ablation Study

IV-E1 Ablation of object relations.

To explore the effects of our proposed modules, TOR and SOR, on the agent’s navigation skills, we integrated them separately into the baseline method. We then conducted experiments with these integrations on the val unseen split of the REVERIE and SOON datasets. As illustrated in Table IV, both TOR and SOR notably enhance the navigation performance of the agent. However, we observed that SOR contributes to a more modest improvement in navigation performance compared to TOR. This is attributed to the limited scale of the current datasets used for training the agent. Solely relying on the objects observed at each position within the environment is inadequate to accurately capture the relationships among different objects. When both modules are employed in tandem, the navigational prowess of the agent is further amplified.

TABLE III: Comparison with the state-of-the-art methods on the R2R dataset. The baseline indicates the replicated results of DUET.
Methods Val seen Val Unseen Test Unseen
TL NE SR SPL TL NE SR SPL TL NE SR SPL
Seq2Seq [4] 11.33 6.01 39 - 8.39 7.81 22 - 8.13 7.85 20 18
RCM [37] - 3.33 70 67 - 5.28 55 50 - 5.15 55 51
PREVALENT [39] 10.32 3.67 69 65 10.19 4.71 58 53 10.51 5.30 54 51
EntityGraph [33] 10.13 3.47 67 65 9.99 4.73 57 53 10.29 4.75 55 52
VLNBERT [11] 11.13 2.90 72 68 12.01 3.93 63 57 12.35 4.09 63 57
AirBERT [3] 11.09 2.68 75 70 11.78 4.01 62 56 12.41 4.13 62 67
HOP [35] 11.26 2.72 75 70 12.27 3.80 64 57 12.68 3.83 64 59
HAMT [54] - - - - 11.46 2.29 66 61 12.27 3.93 65 60
DUET [16] - - - - 13.94 3.31 72 60 14.73 3.65 69 59
KERM [17] 12.16 2.19 79.73 73.79 13.54 3.22 71.95 60.91 14.60 3.61 69.73 59.25
BEVBert [56] - - - - - 2.81 75 64 - 3.13 73 62
GridMM [18] - - - - 13.27 2.83 75 64 14.43 3.35 73 62
Baseline 13.41 2.35 79 72 13.92 3.25 71 60 15.2 3.42 70 60
Ours 12.67 2.05 82 76 13.57 3.05 71 61 15.4 3.27 72 62
TABLE IV: Ablation of the object relations on the val unseen split of REVERIE and SOON. In this table, TOR and EOR respectively represent the trajectory-object relationships module and the environment-object relationships module. All experiments are conducted without utilizing TBP loss function.
Dataset TOR EOR TL OSR SR SPL RGS RGSPL
REVERIE ×\times ×\times 22.12 51.09 46.98 33.75 32.15 23.05
×\times 22.31 52.85 48.08 33.77 32.55 23.07
×\times 23.07 55.44 49.22 33.32 33.46 22.90
23.52 54.47 49.64 34.56 34.05 23.41
SOON ×\times ×\times 36.18 50.88 36.17 22.54 6.02 3.75
×\times 37.07 51.92 38.05 25.31 5.46 3.51
×\times 36.46 54.28 39.09 25.13 7.08 4.61
39.91 56.49 40.71 24.98 7.37 4.56
Refer to caption
Figure 4: The distribution of trajectory lengths predicted on the val seen and val unseen splits of REVERIE dataset.
TABLE V: Ablation of TBP loss on the val unseen split of REVERIE and SOON. Both the TOR and EOR modules are employed in these experiments.
Dataset TBP TL OSR SR SPL RGS RGSPL
REVERIE ×\times 23.52 54.47 49.64 34.56 34.05 23.41
22.00 55.55 50.30 36.79 35.27 25.98
SOON ×\times 39.91 56.49 40.71 24.98 7.37 4.56
38.75 55.46 40.12 26.00 7.67 5.04

IV-E2 Ablation of TBP loss.

In Table IV, we observes an intriguing phenomenon. The inclusion of object relationships does indeed significantly enhance the agent’s success rates in navigation (OSR, SR, and RGS). However, this enhancement also leads to an increase in the trajectory length (TL) of the agent. As a result, there is no pronounced improvement in metrics such as SPL and RGSPL. This suggests that the object relationship module leads the agent to engage in excessive redundant exploration, resulting in elongated navigation paths. Upon integrating the TBP loss function, we observes a significant reduction in the agent’s revisits to the same location. This is illustrated in Table V. This change leads to a more efficient task completion and shows clear improvements in metrics such as TL, SPL.

Additionally, Fig. 4 depicts the distribution of navigational path lengths corresponding to successful navigation instances (specifically when SR is 1) for both the val seen and val unseen splits of the REVERIE dataset, comparing scenarios with and without the integration of the TBP loss function. The figure reveals that with the TBP loss, there’s a notable increase in the proportion of agent paths falling between 0 and 10, while paths longer than 20 significantly diminish. This solidly validates that our TBP loss function effectively curtails the navigation path length of the agent.

IV-E3 Punish turning back during inference.

TABLE VI: The results of punishing turning back during inference. ξ=i\xi=i means dividing the probability of the agent reaching a certain position by ii.
No. ξ\xi TL OSR SR SPL RGS RGSPL
1 0.5 28.10 59.24 38.00 22.17 27.78 16.66
2 1 22.00 55.55 50.30 36.79 35.27 25.98
3 2 22.53 51.32 47.71 32.91 32.06 22.22
4 4 22.01 49.53 46.49 32.35 31.47 21.91
5 6 21.90 48.79 45.81 32.03 31.10 21.75
6 8 21.84 48.59 45.61 31.93 30.99 21.73

It is also a intuitive way punish turning back during inference. To assess whether penalizing the agent’s repetitive visiting behavior during inference improves its navigational abilities, we have designed several experiments. The experimental outcomes are presented in Table VI.

Our findings indicate that penalizing the agent’s repetitive visiting behavior during inference does not enhance its navigational performance. As evidenced in the last five rows of Table VI, the navigation success rate of the agent decreases with increasing penalty intensity. This decline in performance is attributed to the fact that such penalties during inference prevent the agent from correcting its navigational errors. Additionally, the first two rows of Table VI reveal that when repetitive visits by the agent are encouraged, there is a more significant drop in navigational ability. This is due to the agent engaging in more unproductive exploration, substantially increasing the length of the navigational path.

IV-F Qualitative Results

Refer to caption
(a) The figures show the attention maps of the temporal object relations and spatial object relations. The attention is indicated from low to high with the color blue to red.
Refer to caption
(b) In this figure, orange lines emphasize instances where the agent backtracks on its route.
Figure 5: Visualization of attention maps and navigation examples

IV-F1 Visualization of object relations.

To demonstrate that our method can accurately learn the various relationships between objects, we visualized attention heatmaps of different object interactions after training. Fig. 5(a) showcases the object relationships learned by both the TOR and SOR modules. From this, we observes that both modules are capable of accurately modeling object relationships. The spatial object relations provide a detailed and comprehensive depiction of inherent connections between objects, such as chairs and rooms, or decor and thermostats being closely associated. The spatial object relations supplement the temporal object relations. Certain information not captured by the TOR module is correctly learned in SOR. For instance, the environmental object relationships suggest a connection between decor and bathroom, which aligns with real-world scenarios but is not precisely represented in the trajectory-object relationships.

IV-F2 Visualization of the navigation trajectories.

To elucidate the effectiveness of our proposed approach, we have rendered a comparative visualization of the navigation trajectories generated by both DUET and our method. As illustrated in Fig. 5(b), both techniques can accurately reach navigation targets. However, DUET tends to involve the agent in excessive exploratory actions. This often results in repeated visits to the same places, hindering navigational efficiency. Conversely, our method substantially diminishes the agent’s inclination to backtrack, facilitating the selection of more direct routes that expedite the completion of the navigation tasks. Furthermore, our analysis reveals that at the onset of navigation, our method opted for a proximal route, in contrast to DUET which embarked on a relatively longer path with a greater number of actions. This indicates that our method can effectively understand the relationships between objects encountered by the agent and the specified targets. Consequently, it identifies more efficient paths. This aids in completing navigational tasks more effectively.

V Conclusion

In this study, we introduced Temporal-Spatial Object Relations Modules and a Turning Back Penalty (TBP) loss function that together enhance agent navigation. By learning the connections between various objects, the agent can more effectively complete navigation tasks. The application of the TBP loss function successfully prevents repetitive visits to the same location by the agent, thereby enhancing navigational efficiency. It is noteworthy that the object relationships we model might not be entirely accurate because of the limited datasets. Moving forward, our future endeavors will focus on devising more sophisticated object relationship modeling techniques, expanding dataset scales, and honing the precision and efficiency of navigational tasks.

References

  • [1] M. Nawaz, J. K.-T. Tang, K. Bibi, S. Xiao, H.-P. Ho, and W. Yuan, “Robust cognitive capability in autonomous driving using sensor fusion techniques: A survey,” IEEE Transactions on Intelligent Transportation Systems, vol. 25, no. 5, pp. 3228–3243, 2024.
  • [2] C. Sun, R. Zhang, Y. Lu, Y. Cui, Z. Deng, D. Cao, and A. Khajepour, “Toward ensuring safety for autonomous driving perception: Standardization progress, research advances, and perspectives,” IEEE Transactions on Intelligent Transportation Systems, vol. 25, no. 5, pp. 3286–3304, 2024.
  • [3] P.-L. Guhur, M. Tapaswi, S. Chen, I. Laptev, and C. Schmid, “Airbert: In-domain pretraining for vision-and-language navigation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2021, pp. 1634–1643.
  • [4] P. Anderson, Q. Wu, D. Teney, J. Bruce, M. Johnson, N. Sünderhauf, I. Reid, S. Gould, and A. van den Hengel, “Vision-and-language navigation: Interpreting visually-grounded navigation instructions in real environments,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018, pp. 3674–3683.
  • [5] W. Zhu, Y. Qi, P. Narayana, K. Sone, S. Basu, X. Wang, Q. Wu, M. Eckstein, and W. Y. Wang, “Diagnosing vision-and-language navigation: What really matters,” in Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2022, pp. 5981–5993.
  • [6] Y. Qi, Q. Wu, P. Anderson, X. Wang, W. Y. Wang, C. Shen, and A. van den Hengel, “Reverie: Remote embodied visual referring expression in real indoor environments,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 9979–9988.
  • [7] W. Kang, J. Mun, S. Lee, and B. Roh, “Noise-aware learning from web-crawled image-text data for image captioning,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 2942–2952.
  • [8] Y. Zhang, C.-H. Ho, and N. Vasconcelos, “Toward unsupervised realistic visual question answering,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2023, pp. 15 613–15 624.
  • [9] J. Qi, Y. Niu, J. Huang, and H. Zhang, “Two causal principles for improving visual dialog,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 10 857–10 866.
  • [10] Y. Qiao, C. Deng, and Q. Wu, “Referring expression comprehension: A survey of methods and datasets,” IEEE Transactions on Multimedia, vol. 23, pp. 4426–4440, 2021.
  • [11] Y. Hong, Q. Wu, Y. Qi, C. Rodriguez-Opazo, and S. Gould, “Vln bert: A recurrent vision-and-language bert for navigation,” in 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 1643–1653.
  • [12] D. Fried, R. Hu, V. Cirik, A. Rohrbach, J. Andreas, L.-P. Morency, T. Berg-Kirkpatrick, K. Saenko, D. Klein, and T. Darrell, “Speaker-follower models for vision-and-language navigation,” Advances in Neural Information Processing Systems, vol. 31, 2018.
  • [13] A. Moudgil, A. Majumdar, H. Agrawal, S. Lee, and D. Batra, “Soat: A scene- and object-aware transformer for vision-and-language navigation,” in Advances in Neural Information Processing Systems, M. Ranzato, A. Beygelzimer, Y. Dauphin, P. Liang, and J. W. Vaughan, Eds., vol. 34.   Curran Associates, Inc., 2021, pp. 7357–7367.
  • [14] H. Tan, L. Yu, and M. Bansal, “Learning to navigate unseen environments: Back translation with environmental dropout,” in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 2019, pp. 2610–2621.
  • [15] B. Lin, Y. Zhu, Z. Chen, X. Liang, J. Liu, and X. Liang, “Adapt: Vision-language navigation with modality-aligned action prompts,” in 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 15 375–15 385.
  • [16] S. Chen, P.-L. Guhur, M. Tapaswi, C. Schmid, and I. Laptev, “Think global, act local: Dual-scale graph transformer for vision-and-language navigation,” in 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 16 516–16 526.
  • [17] X. Li, Z. Wang, J. Yang, Y. Wang, and S. Jiang, “Kerm: Knowledge enhanced reasoning for vision-and-language navigation,” in 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023, pp. 2583–2592.
  • [18] Z. Wang, X. Li, J. Yang, Y. Liu, and S. Jiang, “Gridmm: Grid memory map for vision-and-language navigation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2023, pp. 15 625–15 636.
  • [19] H. Du, X. Yu, and L. Zheng, “Learning object relation graph and tentative policy for visual navigation,” in Computer Vision – ECCV 2020, A. Vedaldi, H. Bischof, T. Brox, and J.-M. Frahm, Eds.   Cham: Springer International Publishing, 2020, pp. 19–34.
  • [20] R. Dang, Z. Shi, L. Wang, Z. He, C. Liu, and Q. Chen, “Unbiased directed object attention graph for object navigation,” in Proceedings of the 30th ACM International Conference on Multimedia, 2022, pp. 3617–3627.
  • [21] R. Dang, L. Wang, Z. He, S. Su, J. Tang, C. Liu, and Q. Chen, “Search for or navigate to? dual adaptive thinking for object navigation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2023, pp. 8250–8259.
  • [22] J. Chen, C. Gao, E. Meng, Q. Zhang, and S. Liu, “Reinforced structured state-evolution for vision-language navigation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2022, pp. 15 450–15 459.
  • [23] C.-W. Kuo and Z. Kira, “Beyond a pre-trained object detector: Cross-modal textual and visual context for image captioning,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 17 969–17 979.
  • [24] R. Krishna, Y. Zhu, O. Groth, J. Johnson, K. Hata, J. Kravitz, S. Chen, Y. Kalantidis, L.-J. Li, D. A. Shamma et al., “Visual genome: Connecting language and vision using crowdsourced dense image annotations,” International journal of computer vision, vol. 123, pp. 32–73, 2017.
  • [25] R. Speer, J. Chin, and C. Havasi, “Conceptnet 5.5: An open multilingual graph of general knowledge,” in Proceedings of the AAAI conference on artificial intelligence, vol. 31, 2017.
  • [26] C. Gao, J. Chen, S. Liu, L. Wang, Q. Zhang, and Q. Wu, “Room-and-object aware knowledge reasoning for remote embodied referring expression,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 3064–3073.
  • [27] R. Schumann, W. Zhu, W. Feng, T.-J. Fu, S. Riezler, and W. Y. Wang, “VELMA: Verbalization Embodiment of LLM Agents for Vision and Language Navigation in Street View,” arXiv e-prints, p. arXiv:2307.06082, Jul. 2023.
  • [28] G. Zhou, Y. Hong, and Q. Wu, “NavGPT: Explicit Reasoning in Vision-and-Language Navigation with Large Language Models,” arXiv e-prints, p. arXiv:2305.16986, May 2023.
  • [29] D. Shah, B. Osiński, b. ichter, and S. Levine, “Lm-nav: Robotic navigation with large pre-trained models of language, vision, and action,” in Proceedings of The 6th Conference on Robot Learning, ser. Proceedings of Machine Learning Research, K. Liu, D. Kulic, and J. Ichnowski, Eds., vol. 205.   PMLR, 14–18 Dec 2023, pp. 492–504.
  • [30] H. Wang, W. Wang, W. Liang, S. C. Hoi, J. Shen, and L. V. Gool, “Active perception for visual-language navigation,” International Journal of Computer Vision, vol. 131, no. 3, pp. 607–625, 2023.
  • [31] F. Zhu, X. Liang, Y. Zhu, Q. Yu, X. Chang, and X. Liang, “Soon: Scenario oriented object navigation with graph-based exploration,” in 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 12 684–12 694.
  • [32] X. Wang, Q. Huang, A. Celikyilmaz, J. Gao, D. Shen, Y.-F. Wang, W. Y. Wang, and L. Zhang, “Reinforced cross-modal matching and self-supervised imitation learning for vision-language navigation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 6629–6638.
  • [33] Y. Hong, C. Rodriguez, Y. Qi, Q. Wu, and S. Gould, “Language and visual entity relationship graph for agent navigation,” Advances in Neural Information Processing Systems, vol. 33, pp. 7685–7696, 2020.
  • [34] H. Wang, W. Wang, W. Liang, C. Xiong, and J. Shen, “Structured scene memory for vision-language navigation,” in Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, 2021, pp. 8455–8464.
  • [35] Y. Qiao, Y. Qi, Y. Hong, Z. Yu, P. Wang, and Q. Wu, “Hop: History-and-order aware pretraining for vision-and-language navigation,” in 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 15 397–15 406.
  • [36] S. Chen, P.-L. Guhur, M. Tapaswi, C. Schmid, and I. Laptev, “Learning from unlabeled 3d environments for vision-and-language navigation,” in European Conference on Computer Vision.   Springer, 2022, pp. 638–655.
  • [37] F. Zhu, Y. Zhu, X. Chang, and X. Liang, “Vision-language navigation with self-supervised auxiliary reasoning tasks,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 10 009–10 019.
  • [38] C.-Y. Ma, Z. Wu, G. AlRegib, C. Xiong, and Z. Kira, “The regretful agent: Heuristic-aided navigation through progress estimation,” in Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, 2019, pp. 6732–6740.
  • [39] W. Hao, C. Li, X. Li, L. Carin, and J. Gao, “Towards learning a generic agent for vision-and-language navigation via pre-training,” 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13 134–13 143, 2020.
  • [40] X. Li, C. Li, Q. Xia, Y. Bisk, A. Celikyilmaz, J. Gao, N. A. Smith, and Y. Choi, “Robust navigation with language pretraining and stochastic sampling,” in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 2019, pp. 1494–1499.
  • [41] S. Zhang, X. Song, Y. Bai, W. Li, Y. Chu, and S. Jiang, “Hierarchical object-to-zone graph for object navigation,” in 2021 IEEE/CVF International Conference on Computer Vision (ICCV), 2021, pp. 15 110–15 120.
  • [42] Y. Qi, Z. Pan, S. Zhang, A. van den Hengel, and Q. Wu, “Object-and-action aware model for visual language navigation,” in European Conference on Computer Vision.   Springer, 2020, pp. 303–317.
  • [43] C. Huang, O. Mees, A. Zeng, and W. Burgard, “Visual language maps for robot navigation,” in 2023 IEEE International Conference on Robotics and Automation (ICRA), 2023, pp. 10 608–10 615.
  • [44] S. Ross, G. Gordon, and D. Bagnell, “A reduction of imitation learning and structured prediction to no-regret online learning,” in Proceedings of the fourteenth international conference on artificial intelligence and statistics.   JMLR Workshop and Conference Proceedings, 2011, pp. 627–635.
  • [45] C. Ma, J. Lu, Z. Wu, G. AlRegib, Z. Kira, R. Socher, and C. Xiong, “Self-monitoring navigation agent via auxiliary progress estimation,” in 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019.   OpenReview.net, 2019.
  • [46] P. Anderson, X. He, C. Buehler, D. Teney, M. Johnson, S. Gould, and L. Zhang, “Bottom-up and top-down attention for image captioning and visual question answering,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018, pp. 6077–6086.
  • [47] A. Chang, A. Dai, T. Funkhouser, M. Halber, M. Niebner, M. Savva, S. Song, A. Zeng, and Y. Zhang, “Matterport3d: Learning from rgb-d data in indoor environments,” in 2017 International Conference on 3D Vision (3DV).   IEEE, 2017, pp. 667–676.
  • [48] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of deep bidirectional transformers for language understanding,” in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers).   Minneapolis, Minnesota: Association for Computational Linguistics, Jun. 2019, pp. 4171–4186.
  • [49] H. H. Tan and M. Bansal, “Lxmert: Learning cross-modality encoder representations from transformers,” in Conference on Empirical Methods in Natural Language Processing, 2019.
  • [50] J. Krantz, E. Wijmans, A. Majumdar, D. Batra, and S. Lee, “Beyond the nav-graph: Vision-and-language navigation in continuous environments,” in Computer Vision – ECCV 2020, A. Vedaldi, H. Bischof, T. Brox, and J.-M. Frahm, Eds.   Cham: Springer International Publishing, 2020, pp. 104–120.
  • [51] A. Pashevich, C. Schmid, and C. Sun, “Episodic transformer for vision-and-language navigation,” 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 15 922–15 932, 2021.
  • [52] J. Lu, D. Batra, D. Parikh, and S. Lee, “Vilbert: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks,” Advances in neural information processing systems, vol. 32, 2019.
  • [53] A. Kolesnikov, A. Dosovitskiy, D. Weissenborn, G. Heigold, J. Uszkoreit, L. Beyer, M. Minderer, M. Dehghani, N. Houlsby, S. Gelly, T. Unterthiner, and X. Zhai, “An image is worth 16x16 words: Transformers for image recognition at scale,” in Proceedings of the International Conference on Learning Representations (ICLR), 2021.
  • [54] S. Chen, P.-L. Guhur, C. Schmid, and I. Laptev, “History aware multimodal transformer for vision-and-language navigation,” in NeurIPS, 2021.
  • [55] X. Lin, G. Li, and Y. Yu, “Scene-intuitive agent for remote embodied visual grounding,” in 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 7032–7041.
  • [56] D. An, Y. Qi, Y. Li, Y. Huang, L. Wang, T. Tan, and J. Shao, “Bevbert: Multimodal map pre-training for language-guided navigation,” Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023.
  • [57] M. Li, Z. Wang, T. Tuytelaars, and M.-F. Moens, “Layout-aware dreamer for embodied visual referring expression grounding,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, 2023, pp. 1386–1395.