Next Article in Journal
Retention Time Trajectory Matching for Peak Identification in Chromatographic Analysis
Next Article in Special Issue
Accurate Visual Simultaneous Localization and Mapping (SLAM) against Around View Monitor (AVM) Distortion Error Using Weighted Generalized Iterative Closest Point (GICP)
Previous Article in Journal
Augmented Reality-Assisted Deep Reinforcement Learning-Based Model towards Industrial Training and Maintenance for NanoDrop Spectrophotometer
Previous Article in Special Issue
Deep Learning-Aided Inertial/Visual/LiDAR Integration for GNSS-Challenging Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Outdoor Vision-and-Language Navigation Needs Object-Level Alignment

1
Department of Electronics and Electrical Engineering, Faculty of Science and Technology, Keio University, 3-14-1, Hiyoshi, Kohoku-ku, Yokohama 223-8522, Japan
2
National Institute of Advanced Industrial Science and Technology (AIST), Tsukuba 305-8560, Japan
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(13), 6028; https://doi.org/10.3390/s23136028
Submission received: 19 May 2023 / Revised: 14 June 2023 / Accepted: 27 June 2023 / Published: 29 June 2023
(This article belongs to the Special Issue Artificial Intelligence (AI) and Machine-Learning-Based Localization)

Abstract

:
In the field of embodied AI, vision-and-language navigation (VLN) is a crucial and challenging multi-modal task. Specifically, outdoor VLN involves an agent navigating within a graph-based environment, while simultaneously interpreting information from real-world urban environments and natural language instructions. Existing outdoor VLN models predict actions using a combination of panorama and instruction features. However, these methods may cause the agent to struggle to understand complicated outdoor environments and ignore the details in the environments to fail to navigate. Human navigation often involves the use of specific objects as reference landmarks when navigating to unfamiliar places, providing a more rational and efficient approach to navigation. Inspired by this natural human behavior, we propose an object-level alignment module (OAlM), which guides the agent to focus more on object tokens mentioned in the instructions and recognize these landmarks during navigation. By treating these landmarks as sub-goals, our method effectively decomposes a long-range path into a series of shorter paths, ultimately improving the agent’s overall performance. In addition to enabling better object recognition and alignment, our proposed OAlM also fosters a more robust and adaptable agent capable of navigating complex environments. This adaptability is particularly crucial for real-world applications where environmental conditions can be unpredictable and varied. Experimental results show our OAlM is a more object-focused model, and our approach outperforms all metrics on a challenging outdoor VLN Touchdown dataset, exceeding the baseline by 3.19% on task completion (TC). These results highlight the potential of leveraging object-level information in the form of sub-goals to improve navigation performance in embodied AI systems, paving the way for more advanced and efficient outdoor navigation.

1. Introduction

Vision-and-language navigation (VLN) is a challenging task wherein agents must navigate a real-world visual environment by following instructions in natural language. The agent must comprehend the instructions and ground them in the observable environment with only visual perception. Moreover, it needs to reason about its position in relation to objects and understand how these relationships change as it moves through the environment. Ultimately, the agent should generate a series of corrective actions to reach the destination. Enabling agents to understand the relationship between instructional words and perceived features in the environment is one of the critical challenges of VLN [1].
To help the agent understand the relationship between instructions and environmental features, approaches such as the OAAM (object-and-action aware model) [2] and ROAA (room-and-object aware attention) [3] extract the semantic connections between different modalities (actions, scenes, observed objects, and objects mentioned in instructions) and then softly align them with an attention mechanism. However, their effectiveness is limited in outdoor environments due to inherent differences in environmental conditions. Compared to indoor VLN [4], outdoor VLN faces a more complex real-world urban environment, and it is considered a more challenging task for an agent to be able to handle more spatial information and significant visual feature differences such as textures, lighting, and scene layouts [5,6]. Furthermore, outdoor environments also involve dynamic elements such as pedestrians and vehicles, which indoor models may struggle to handle effectively.
As illustrated in Figure 1 [6,7,8,9,10], which largely relies on simplistic encoder–decoder models, our approach places significant emphasis on landmarks. These existing models concatenate instruction features and panoramic features and use a decoder to predict a sequence of actions from these concatenated features. While these models can produce agents capable of basic navigation, their capacity to understand specific semantics is lacking. However, such an approach does not align with human intuition. When navigating to a place in an unfamiliar environment, humans tend to use visible objects such as buildings and other objects seen along the way as reference landmarks [11]. Therefore, the objects are essential clues in outdoor VLN. We observed that existing models hardly focus on object tokens by visualizing the paths generated by them and the attention heatmap of the instructions. In many cases, the agent ignores landmarks given by the instructions, making turns or stopping in incorrect places.
Inspired by humans’ use of landmarks to navigate unfamiliar places and the significant results of object-aware indoor VLN models [2,3,12,13], several recent studies of indoor VLN models in particular [14,15,16] have demonstrated the utility of using object detectors. Therefore, we believe that better use of landmarks in outdoor VLN scenarios is a more rational way to navigate, and may improve navigation performance.
We, therefore, proposed an Object-level Alignment Module (OAlM), which is a straightforward yet effective module that guides the agent to focus more on the landmarks indicated in the navigation instructions. The OAlM consists of four components: the landmark extractor, landmark detector, match checker, and index controller. The landmark extractor identifies and extracts landmarks from the instructions, which serve as critical reference points that the agent should aim to pass during navigation. Simultaneously, a landmark detector scans the environment and identifies landmarks, such as text on signboards or objects explicitly mentioned in the instructions. The index controller indicates which reference landmark the agent should aim for next. It effectively sequences the order of landmarks to guide the navigation process. Finally, the match checker is responsible for assessing whether the agent has successfully reached the current reference landmark at each step. Based on this verification, the index controller adjusts the points accordingly.
We have extensively experimented on the Touchdown dataset [6] with three baseline outdoor VLN models, RCONCAT [6], GA [7], and ORAR [10]. We added the OAlM module into the baselines and compared the results with/without our OAlM to show that landmark alignment is necessary for better outdoor VLN. Our ablation studies and qualitative results further verify that the improvement indeed comes from agents being able to make better use of landmarks. In summary, the main contributions of this paper are as follows:
  • We developed an object-level alignment module (OAlM) and used OCR and object-detection technology to identify the prominent landmarks. OAlM pays more attention to landmarks and helps the agent understand the environment to improve navigation ability.
  • We extensively evaluated our approach and found that it outperforms all metrics on the Touchdown dataset against multiple existing methods, exceeding the baseline by 3.19% on task completion (TC).

2. Related Works

2.1. Vision-and-Language Navigation

An agent needs to reach the target destination by following a set of natural language instructions. One of the foundational benchmarks in this area, Room-to-Room (R2R) [4] for indoor scenarios was proposed, along with a multimodal Seq2Seq baseline model. Then several indoor VLN datasets were created by extending R2R with other languages, such as XL-R2R [17] and RxR [18]. However, outdoor environments are usually more complex and diverse compared to indoor environments. Recognizing this, the Touchdown [6] dataset was proposed, making the first outdoor VLN benchmark, which is based on Google Street View (https://developers.google.com/maps/documentation/streetview/intro, accessed on 10 September 2022). Subsequently, additional outdoor VLN datasets, such as StreetLearn [5,19], StreetNav [20], Map2seq [21], and Talk2Nav [22] for outdoor VLN have also been proposed.

2.2. Outdoor Vision-and-Language Navigation Models

Numerous methods have been proposed to address the outdoor VLN task. RCONCAT [6] is the baseline model proposed in the original Touchdown paper, and it encodes the trajectory and the instructions in an LSTM-based model. ARC [8] proposed ARC+l2s, which cascades the action prediction into a binary stopping decision and subsequent direction classification. VLN-Transformer [9] uses a pretrained BERT [23] applied to an external multimodal dataset to enrich the navigation data. GA [7] uses gated attention to compute a fused representation of instructions and images to predict actions. ORAR [10] adds junction-type embedding and a heading delta to train a general model, reducing the performance gap between seen and unseen environments. However, these outdoor VLN models use LSTM or other encoders to encode the instructions and a decoder LSTM to predict the actions. This architecture limits attention to particular semantics, such as objects, which are essential clues in VLN.

2.3. Object-Aware Vision-and-Language Navigation

To handle fine-grained information, such as objects, in order to facilitate simultaneous understanding of instructions and visuals, vision-and-language pretrained models such as ViLBERT [24] have been widely adopted in VLN because they can provide good joint features for better understanding of instructions and environmental features. ORIST [25] was trained with specific features such as objects and rooms. OAAM [2] extracted object and action tokens from instructions and then matched them to visual perceptions and orientation to predict actions. ROAA [3] was designed to perceive the room and object-type information from instruction and visual perceptions, while the knowledge-enabled entity relationship reasoning (KERR) module was designed to learn the internal–external correlations between room and object for the agent to take appropriate actions. SOAT [12] uses a scene classification network and an object detector to produce features that match these two distinct visual cues. M-Track [26] was proposed as a milestone builder and milestone checker to guide agents complete a task step by step and was tested on the ALFRED [27] dataset. The scenario involves an agent completing household tasks in an interactive visual environment. While these approaches all perform well in indoor scenarios, they are unsuitable for outdoor VLN, which has a different and more complex environmental graph. The vast urban areas being navigated lead to a much larger space for the agent to explore and also contain a wider range of encountered objects in the visual environment. Therefore, outdoor VLN requires special treatment in recognizing these diverse objects and using them to understand the environment better.

3. Proposed Methods: OAlM

In this section, we introduce the model that we use to analyze navigation performance in outdoor VLN. First, we give a definition for the VLN problem. Afterwards, we provide a more formal description of the model architecture.

3.1. VLN Problem Definition

The VLN task requires an agent in an environment to follow natural language instructions X = x 1 , x 2 , , x l , along a trajectory and stop at a target location. The length of the instructions is denoted by l, and a word token is denoted by x i . The environment is an undirected graph consisting of nodes v V and labeled edges ( v , u ) E , where u is one of the outgoing neighbors of node v. Each node is associated with a panoramic RGB image and each edge connects nodes to neighboring panoramas at the heading angle α ( v , u ) . At time t, the agent’s state is defined by s t S , s t = ( v t , α ( v t 1 , v t ) ) , where v t is the agent-observed panoramic view, and α ( v t 1 , v t ) is the heading angle between previous state’s view v t 1 and the current state’s view v t . Following the instructions X, the agent executes an action a t { FORWARD , LEFT , RIGHT , STOP } and is updated to the next state s t + 1 . The agent must produce a sequence of state-action pairs s 1 , a 1 , s 2 , a 2 , , s n , a n , where a n = STOP , to reach the goal location.

3.2. Overview

As shown in Figure 2, the model takes navigation instructions as the input and outputs a sequence of agent actions. At each decoding timestep, a new visual representation of the current agent state within the environment is computed, and the alignment between the objects in the environment and the given instructions is computed. The action predictor then decodes a fused text and visual representation and eventually predicts the next action.
The intuition behind the architecture of this model is to first concatenate visual and instruction features at each time step because high-level features (e.g., objects) can outperform lower-level features extracted by the CNN [14]. We then use these high-level features to check whether the agent is on the correct route. According to the results of this check and the fused features, the Action-Predictor decides what action to take. We also propose an object-level alignment module (OAlM), which allows the model to pay more attention to object information from the environment and the instructions.

3.3. OAlM

The alignment module is made up of four parts: a landmark extractor, a landmark detector, a match checker, and an index controller. The landmark extractor and landmark detector detect landmarks from the instructions and the panoramas. We set the ordered landmarks mentioned in the instructions as references. At each timestep, the match checker determines whether the agent has reached the reference by comparing it to the observed landmark in the environment. The index controller sets the first landmark as the current reference landmark at the start of navigation, and updates to the next landmark when the agent reaches the current reference landmark.
  • Landmark Detector: At each timestep t, the landmark detector identifies the text on signboards and objects at the current node v t as LM v t . To obtain the high quality LM v t , we set a threshold.
    LM v t = Landmark-Detector ( v t )
    LM v t = { lm v t ( 1 ) , lm v t ( 2 ) , , lm v t ( n ) } is a set of observed landmarks, where n is the maximum number of landmarks at a panoramic view location v t .
  • Landmark Extractor: To extract this sequence from the natural language instructions, we employed a standard large language model, which in our prototype was GPT-3 [28]. We used a prompt with three examples of correct landmark extractions, and the model was then used to extract a list of landmarks from the instructions. We evaluated the performance on 20 test pairs, the accuracy of GPT-3 outputs was 99%. Therefore, the model’s output was reliable and robust to small changes in the input prompts. For instance, if the prompts give the instructions “Continue down the long block to the next light with Starbucks on the near left corner. Turn right, then stop. Dunkin’ Donuts should be ahead on your right”, the landmark extractor will output “Starbucks, Dunkin’ Donuts” as the ordered landmarks. In this case, the agent would check to see if it has arrived at Starbucks first and then check Dunkin’ Donuts. We extract the list of ordered landmarks LM X = { lm X ( 1 ) , lm X ( 2 ) , , lm X ( d ) } from instructions X, where d is the maximum number of ordered landmarks extracted.
  • Match Checker and Index Controller: The match checker determines if an agent has reached a known landmark, while the index controller keeps track of the reference landmark that the agent should arrive at next. At the beginning of navigation, the index controller points to the first of the ordered landmarks lm X ( 1 ) obtained using the landmark extractor.
    cos ( A , B ) = A · B A B
    m t = 1 n i = 1 n cos ( lm v t ( i ) , lm X ( t ) )
  • With Equation (2), the match checker calculates the cosine similarity between all of the landmarks observed in the environment LM v t and lm X ( t ) and uses this as a matching score m t to check if the agent has reached the current reference landmark lm X ( t ) . If the matching score m t is over the set threshold, the index controller will indicate to the agent the next reference landmark lm X ( t + 1 ) , t < d 1 until the agent reaches the last landmark.

3.4. Inference of OAlM

We implemented our OAlM for use with the following existing VLN models: GA [7], RCONCAT [6], and ORAR [10]. These models were chosen because of their similar architectures and allow seamless integration of our OAlM. To explain in detail, these models embed and encode the instructions X by LSTM [29] or Bi-LSTM [30] to obtain instruction features X t . Following each timestep t, these models utilize a pretrained ResNet-50 [31] or CNNs to extract panorama features v t . Subsequently, the instruction features X t , and the panorama features v t are concatenated to predict the action a t at each timestep t, as indicated by Equation (4). In this euqation, the Action-Predictor is represented as a simple FFNN layer.
In our implementation, we add the matching score m t , introduced above. The Match Checker calculates the matching score and helps the model to confirm whether the sub-goal has been reached. We concatenate m t with the panorama features v t , and the instruction features X t , as expressed in Equation (5). Once this concatenated vector is obtained, it is subsequently input to the action predictor to determine the action a t .
a t = Action-Predictor ( [ X t v t ] )
a t = Action-Predictor ( [ X t v t m t ] )

4. Experiments

In this section, we present extensive experiments on the outdoor VLN Touchdown dataset [6] to evaluate whether our OAlM can improve the performance of models on the outdoor VLN task.

4.1. Implementation Details

The proposed framework and the baselines are implemented in PyTorch [32]. We used ResNet [31] to extract panorama features. For the landmark detector, we used OCR [33] and object-detection technology to recognize landmarks from the panoramas. More specifically, we used EasyOCR (https://github.com/JaidedAI/EasyOCR, accessed on 12 October 2022) to recognize the text on signboards, and we used Mask R-CNN [34] (which is based on ResNet50 [31] pretrained with the ImageNet [35] and then fine-tuned on MSCOCO [36] dataset) to detect objects from the panoramas. The landmark detector outputs results with the confidence score of each Intersection over Union (IoU), and we separated out the results scoring over 0.7, 0.8, and 0.9 to conduct experiments at each threshold value. Additionally, we extracted the ordered landmarks in the instructions (e.g., “Starbucks, MacDonald’s”) with GPT-3 [28]. Then, the extracted landmarks were lower-cased into byte-pair encodings [37] with a vocabulary of 2000 tokens and embedded with a size of 256. We then calculated the cosine similarity of these embeddings.
We used the training settings in the original papers for each model. For GA and RCONCAT, we selected the model with the best shortest-path distance (SPD) performance on the development set after 100 epochs. For ORAR, we selected the model with the best SPD performance on the development set after 150 epochs. Additionally, since the experimental results for the original ORAR code were found to be unstable, we calculated the average of the results from three experiments for evaluation.

4.2. Dataset

Touchdown [6] is a dataset that uses urban scenarios to create a large navigation environment based on Google Street View (https://developers.google.com/maps/documentation/streetview/overview, accessed on 10 September 2022). There are 29,641 nodes and 61,319 undirected edges to simulate New York City’s environment. Each node includes a 360-degree RGB panoramic image. The dataset contains 9326 trajectories for the navigation task, and each trajectory is paired with human-written navigation instructions according to the panoramas. There were 6525 samples in the training set, 1391 samples in the development set, and 1409 samples in the test set.

4.3. Metrics

The following metrics were used to evaluate VLN performance:
(1)
Task Completion (TC): the accuracy of navigating to the correct location. The correct location is defined as the exact goal panorama or one of its neighboring panoramas.
(2)
Shortest-Path Distance (SPD) [6]: the mean distance between the agent’s final position and the goal position in the environment graph.
(3)
Success Weighted by Edit Distance (SED): the normalized Levenshtein edit distance [38] between the predicted path and the ground-truth path, with points only awarded for successful paths.
(4)
Coverage Weighted by Length Score (CLS) [39]: a measurement of the fidelity of the agent’s path with respect to the ground-truth path.
(5)
Normalized Dynamic Time Warping (nDTW) [40]: the minimized cumulative distance between the predicted path and the ground-truth path.
(6)
Success Weighted Dynamic Time Warping (SDTW): the nDTW value where the summation is only of successful navigation.

4.4. Results

In this section, we report the outdoor VLN performance and the quality of object-level alignment to validate the effectiveness of our proposed OAlM. We compare the results with and without OAlM and discuss the influence of object-level alignment in outdoor VLN.
We compare the performance of our approach with the baseline GA [7], RCONCAT [6], and ORAR [10] models, since these models are not treated as specific object features on VLN.

4.4.1. Quantitative Results

OAlM with OCR:Table 1 summarizes the results of the navigation experiments on the Touchdown dataset with and without OAlM (and using OCR as landmark detector). According to Table 1, the majority of the OAlM model’s results in each metric have outperformed the baseline ORAR. Moreover, OAlM, with an OCR score of 0.7, gave the best results in each metric. Of these, OAlM resulted in a 1% improvement in TC and showed significant improvements in the path alignment metrics (CLS, sDTW), demonstrating the effectiveness of OAlM in instruction following and goal achievement. The results of the RCONCAT and GA models show that our proposed OAlM outperformed the baseline in each metric and in each setting. Our OAlM brings about a 3% improvement in both goal-oriented metrics (TC, SED), which demonstrates that our OAlM stops better than the baselines.
OAlM with object detection:Table 2 summarizes the results of navigation with and without OAlM (using Mask R-CNN as a landmark detector). The top of Table 2 illustrates that most metrics were better than baseline when using the landmark detected by Mask R-CNN on the ORAR model. For RCONCAT, although TC, SED, and CLS were below the baseline in experiments utilizing 0.8 score object detection as environment landmarks. Experiment using object-detection scores of 0.9 shows a significant improvement over baseline. Additionally, for the GA model, the SPD was lower than the baseline in all experiment settings. However, other metrics are outperformed on each experiment setting. Furthermore, when given the landmarks of Mask R-CNN with 0.8 score, OAlM brings a 3% improvement in TC and SED.

4.4.2. Qualitative Results

Visualization of trajectories: We provide visualizations of four qualitative examples to illustrate further how our OAlM learns to stop better and turn at the correct place in Figure 3. As seen in Figure 3, when navigating, the baseline tends to ignore landmarks that represent significant turning points or stopping places in instructions. Specifically, in Figure 3a,b, the baseline agent fails to understand the instructions and its current environment, causing it to stop at the wrong place. In Figure 3c, the baseline agent misses a turn and ignores the following instructions that describe the stopping place. Finally, in Figure 3d, the baseline agent turns at the wrong location and cannot understand the whole sentence, then ignores the following description of the observed environment. In contrast to the baseline, our agent stops and turns in the right places.
On the other hand, we provide some failed navigation examples in Figure 4. When we analyzed the failure cases, we found that in many instances, our model was just one step away from the goal, as seen in Figure 4a. These instances accounted for 33% of the failure cases, indicating that even in the majority of the failed scenarios, our model was remarkably close to a successful navigation path. In contrast, the failure cases for the baseline models were unevenly distributed, which we believe demonstrates an advantage of our approach. However, our model does struggle in certain complex situations. For instance, when there were two parallel paths in the environment, the agent chose the incorrect one, as shown in Figure 4b. Some cases of failure, such as Figure 4c, have difficult statements in the instructions that make the agent confused.
OAlM is an object-focused model: To further prove that our proposed OAlM is more object-focused than previous methods, we ran masking experiments similar to ORAR [10] and DiagnoseVLN [41]. Figure 5 shows the changes in task completion rates when masked object tokens are used. The model with the addition of OAlM shows a greater change in the task completion rate, indicating that OAlM indeed enables the model to learn objects from the instructions. However, In Figure 5a, an intriguing result can be observed at a certain data point, where the impact on our model upon masking 20% of the object tokens appears to be minimal. One possible explanation for this might be the robustness of RCOANCAT in using panorama features and instruction features, making it less dependent on specific object tokens. In this particular case, the object detection’s proficiency in extracting critical visual information might have compensated for the effects of object token masking in the instructions. Moreover, when we mask off the object token, the effect on the navigation effect is not very significant compared to ORAR (Figure 5b).
OAlM can stop in the correct position: As mentioned earlier, previous outdoor VLN methods tended to be off on the prediction of stop locations. In the comparison of the stop location predicted by the baselines and our OAlM, as summarized in Table 3, our OAlM has far fewer stop-location errors from failed predicted paths.

4.5. Discussion

The results presented above indicate that the navigation ability of agents can be improved if objects with a certain level of accuracy are used for object alignment. Although some of the experimental results were below baseline, these results were generally overly dependent on the accuracy of existing OCR and object detection. Before the experiments, we expected that the results of OAlM with Mask R-CNN would be better than OAlM with OCR, since the Touchdown dataset describes landmarks with unspecified nouns, which are similar to the results of Mask R-CNN. However, a comparison of Table 1 and Table 2 shows that OAlM with OCR actually performs better. This is because the Mask R-CNN used in this paper has only 81 classes, among which the number of classes that can be recognized from the panorama is quite limited.
In addition, we have investigated the performance of navigation when using both objects detected by OCR and object detection, and Table 4 lists the experimental results of using both OCR and object-detection results for OAlM. However, using both results did not further improve the navigational performance and irregular. Although not shown in Table 1 and Table 2, the results of the experiment using the lower threshold (lower than 0.7) landmarks were actually lower than baseline. On the other hand, Table 4 shows that when both lower scores of landmarks identified by OCR and object detection are used for OAlM, navigation is outperforming. Comparing Table 1, Table 2, and Table 4, besides the SPD in GA model, the results using only either OCR or object detection yield the best results. This is because not all objects detected by OCR and object detection are correct, and the match score is likely to be noise. Moreover, our OAlM is a simple model that cannot handle such complex combinations. In the future, we plan to develop a model that can use multiple object information in a balanced manner.
Based on the quantitative and qualitative results, our proposed OAlM solves the problem of ignoring objects that existed in previous methods. The OAlM method also improves the navigation capability of the agent by focusing on the objects in the instructions and panoramic images.

5. Conclusions

We introduced a novel object-level alignment module for outdoor VLN and showed that explicit landmark-level alignment benefits VLN in an outdoor environment, which is considerably more complex and diverse than indoor environments. Our results demonstrated the effectiveness of using OAlM with three baseline models. Experiments illustrated that object-level alignment can improve outdoor VLN performance, indicating that objects from natural language instructions and the environment are important clues in outdoor VLN. Our proposed OAlM has the advantage of being more interpretable for humans and more accurate in navigation, both in determining where to turn and in stopping.
Limitation: Our OAlM extracted and detected landmarks from instructions and the corresponding environment. A shortcoming in our work is that our landmark detector did not use OCR and object detection was retrained for outdoor environments. However, from our current experiments, we have found that focusing on objects is useful for outdoor navigation. A further shortcoming is that our model’s use of both OCR and object-detection features will result in reduced performance, while separately they both contribute positively to the navigation performance. This may be due to the lack of a special structure designed to handle these two features simultaneously and effectively.
In the future, we plan to focus on investigating how to improve our model’s ability to process both OCR and object-detection inputs effectively simultaneously. We believe that existing OCR and object-detection techniques should be applied to outdoor environments in order to make more use of objects in the environment.

Author Contributions

Y.S. proposed and implemented the approaches and conducted the experiments. She also wrote the paper together with Y.Q., Y.A. and H.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by JST SPRING, Grant Number JPMJSP2123.

Institutional Review Board Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
VLNVision-and-Lagnauge Navigation
LSTMLong Short-Term Memory
CNNConvolutional Neural Network
GPT-3Generative Pre-trained Transformer 3
FFNNFeed-Forward Neural Network
OCROptical Character Recognition

References

  1. Gu, J.; Stefani, E.; Wu, Q.; Thomason, J.; Wang, X. Vision-and-Language Navigation: A Survey of Tasks, Methods, and Future Directions. In Proceedings of the Association for Computational Linguistics, Dublin, Ireland, 22–27 May 2022. [Google Scholar]
  2. Qi, Y.; Pan, Z.; Zhang, S.; van den Hengel, A.; Wu, Q. Object-and-Action Aware Model for Visual Language Navigation. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  3. Gao, C.; Chen, J.; Liu, S.; Wang, L.; Zhang, Q.; Wu, Q. Room-and-Object Aware Knowledge Reasoning for Remote Embodied Referring Expression. In Proceedings of the Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  4. Anderson, P.; Wu, Q.; Teney, D.; Bruce, J.; Johnson, M.; Sünderhauf, N.; Reid, I.; Gould, S.; van den Hengel, A. Vision-and-Language Navigation: Interpreting Visually-Grounded Navigation Instructions in Real Environments. In Proceedings of the Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  5. Mirowski, P.; Grimes, M.K.; Malinowski, M.; Hermann, K.M.; Anderson, K.; Teplyashin, D.; Simonyan, K.; Kavukcuoglu, K.; Zisserman, A.; Hadsell, R. Learning to Navigate in Cities Without a Map. In Proceedings of the Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
  6. Chen, H.; Suhr, A.; Misra, D.; Snavely, N.; Artzi, Y. TOUCHDOWN: Natural Language Navigation and Spatial Reasoning in Visual Street Environments. In Proceedings of the Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  7. Chaplot, D.S.; Sathyendra, K.M.; Pasumarthi, R.K.; Rajagopal, D.; Salakhutdinov, R. Gated-Attention Architectures for Task-Oriented Language Grounding. In Proceedings of the Association for the Advancement of Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  8. Xiang, J.; Wang, X.; Wang, W.Y. Learning to Stop: A Simple yet Effective Approach to Urban Vision-Language Navigation. In Proceedings of the Empirical Methods in Natural Language Processing, Online, 16–20 November 2020. [Google Scholar]
  9. Zhu, W.; Wang, X.; Fu, T.J.; Yan, A.; Narayana, P.; Sone, K.; Basu, S.; Wang, W.Y. Multimodal Text Style Transfer for Outdoor Vision-and-Language Navigation. In Proceedings of the EAssociation for Computational Linguistics, Online, 19–23 April 2021. [Google Scholar]
  10. Schumann, R.; Riezler, S. Analyzing Generalization of Vision and Language Navigation to Unseen Outdoor Areas. In Proceedings of the Association for Computational Linguistics, Dublin, Ireland, 22–27 May 2022. [Google Scholar]
  11. Chan, E.; Baumann, O.; Bellgrove, M.; Mattingley, J. From Objects to Landmarks: The Function of Visual Location Information in Spatial Navigation. Front. Psychol. 2012, 3, 304. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Moudgil, A.; Majumdar, A.; Agrawal, H.; Lee, S.; Batra, D. SOAT: A Scene- and Object-Aware Transformer for Vision-and-Language Navigation. In Proceedings of the Neural Information Processing Systems, Online, 6–14 December 2021. [Google Scholar]
  13. Zhu, F.; Liang, X.; Zhu, Y.; Yu, Q.; Chang, X.; Liang, X. SOON: Scenario Oriented Object Navigation with Graph-Based Exploration. In Proceedings of the Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  14. Hu, R.; Fried, D.; Rohrbach, A.; Klein, D.; Darrell, T.; Saenko, K. Are You Looking? Grounding to Multiple Modalities in Vision-and-Language Navigation. In Proceedings of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019. [Google Scholar] [CrossRef]
  15. Zhang, Y.; Tan, H.; Bansal, M. Diagnosing the Environment Bias in Vision-and-Language Navigation. In Proceedings of the International Joint Conference on Artificial Intelligence, Online, 19–26 August 2020. [Google Scholar]
  16. Majumdar, A.; Shrivastava, A.; Lee, S.; Anderson, P.; Parikh, D.; Batra, D. Improving Vision-and-Language Navigation with Image-Text Pairs from the Web. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  17. Yan, A.; Wang, X.E.; Feng, J.; Li, L.; Wang, W.Y. Cross-Lingual Vision-Language Navigation. arXiv 2019, arXiv:1910.11301. [Google Scholar] [CrossRef]
  18. Ku, A.; Anderson, P.; Patel, R.; Ie, E.; Baldridge, J. Room-Across-Room: Multilingual Vision-and-Language Navigation with Dense Spatiotemporal Grounding. In Proceedings of the Empirical Methods in Natural Language Processing, Online, 16–20 November 2020. [Google Scholar]
  19. Mehta, H.; Artzi, Y.; Baldridge, J.; Ie, E.; Mirowski, P. Retouchdown: Releasing Touchdown on StreetLearn as a Public Resource for Language Grounding Tasks in Street View. In Proceedings of the Empirical Methods in Natural Language Processing-SpLU, Online, 16–20 November 2020. [Google Scholar] [CrossRef]
  20. Hermann, K.; Malinowski, M.; Mirowski, P.; Banki-Horvath, A.; Anderson, K.; Hadsell, R. Learning to Follow Directions in Street View. Assoc. Adv. Artif. Intell. 2020, 34, 11773–11781. [Google Scholar] [CrossRef]
  21. Schumann, R.; Riezler, S. Generating Landmark Navigation Instructions from Maps as a Graph-to-Text Problem. In Proceedings of the Association for Computational Linguistics, Bangkok, Thailand, 1–6 August 2021. [Google Scholar]
  22. Vasudevan, A.B.; Dai, D.; Van Gool, L. Talk2Nav: Long-Range Vision-and-Language Navigation with Dual Attention and Spatial Memory. Int. J. Comput. Vis. 2021, 129, 246–266. [Google Scholar] [CrossRef]
  23. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the NAAssociation for Computational Linguistics, Minneapolis, MN, USA, 2–7 June 2019. [Google Scholar] [CrossRef]
  24. Lu, J.; Batra, D.; Parikh, D.; Lee, S. ViLBERT: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks. In Proceedings of the Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  25. Qi, Y.; Pan, Z.; Hong, Y.; Yang, M.H.; Hengel, A.v.d.; Wu, Q. The Road to Know-Where: An Object-and-Room Informed Sequential BERT for Indoor Vision-Language Navigation. In Proceedings of the International Conference on Computer Vision, Online, 11–17 October 2021. [Google Scholar]
  26. Song, C.H.; Kil, J.; Pan, T.Y.; Sadler, B.M.; Chao, W.L.; Su, Y. One Step at a Time: Long-Horizon Vision-and-Language Navigation With Milestones. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 15482–15491. [Google Scholar]
  27. Shridhar, M.; Thomason, J.; Gordon, D.; Bisk, Y.; Han, W.; Mottaghi, R.; Zettlemoyer, L.; Fox, D. ALFRED: A Benchmark for Interpreting Grounded Instructions for Everyday Tasks. In Proceedings of the Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  28. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Few-Shot Learners. In Proceedings of the Neural Information Processing Systems, Online, 6–12 December 2020; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H., Eds.; Volume 33. [Google Scholar]
  29. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  30. Graves, A.; Fernández, S.; Schmidhuber, J. Bidirectional LSTM Networks for Improved Phoneme Classification and Recognition. In Proceedings of the Artificial Neural Networks: Formal Models and Their Applications—ICANN, Warsaw, Poland, 11–15 September 2005; Duch, W., Kacprzyk, J., Oja, E., Zadrożny, S., Eds.; 2005. [Google Scholar]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; Available online: https://openaccess.thecvf.com/content_cvpr_2016/html/He_Deep_Residual_Learning_CVPR_2016_paper.html (accessed on 15 October 2021).
  32. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Proceedings of the Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  33. Srihari, S.N.; Shekhawat, A.; Lam, S.W. Optical Character Recognition (OCR). In Encyclopedia of Computer Science; Van Nostrand Reinhold Company: New York, NY, USA, 2003; pp. 1326–1333. [Google Scholar]
  34. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  35. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  36. Lin, T.Y.; Maire, M.; Belongie, S.; Bourdev, L.; Girshick, R.; Hays, J.; Perona, P.; Ramanan, D.; Zitnick, C.L.; Dollár, P. Microsoft COCO: Common Objects in Context. arXiv 2015, arXiv:1405.0312. [Google Scholar]
  37. Sennrich, R.; Haddow, B.; Birch, A. Neural Machine Translation of Rare Words with Subword Units. In Proceedings of the Association for Computational Linguistics, Berlin, Germany, 7–12 August 2016. [Google Scholar] [CrossRef]
  38. Levenshtein, V. Leveinshtein Distance. 1965. Available online: https://rybn.org/ANTI/ADMXI/documentation/ALGORITHM_DOCUMENTATION/HARMONY_OF_THE_SPEARS/LEVENSHTEIN_EDIT_DISTANCE/ABOUT/NIST_Levenshtein_Edit_Distance.pdf (accessed on 10 November 2021).
  39. Jain, V.; Magalhaes, G.; Ku, A.; Vaswani, A.; Ie, E.; Baldridge, J. Stay on the Path: Instruction Fidelity in Vision-and-Language Navigation. In Proceedings of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019. [Google Scholar] [CrossRef]
  40. Magalhaes, G.I.; Jain, V.; Ku, A.; Ie, E.; Baldridge, J. General Evaluation for Instruction Conditioned Navigation using Dynamic Time Warping. In Proceedings of the Neural Information Processing Systems Visually Grounded Interaction and Language (ViGIL) Workshop, Vancouver, BC, Canada, 13 December 2019. [Google Scholar]
  41. Zhu, W.; Qi, Y.; Narayana, P.; Sone, K.; Basu, S.; Wang, X.; Wu, Q.; Eckstein, M.; Wang, W.Y. Diagnosing Vision-and-Language Navigation: What Really Matters. In Proceedings of the NAAssociation for Computational Linguistics, Dublin, Ireland, 22–27 May 2022. [Google Scholar]
Figure 1. Comparison of existing outdoor vision-and-language navigation models with simple feature concatenation (Top) and our object-alignment module focused on landmarks for navigation (Bottom).
Figure 1. Comparison of existing outdoor vision-and-language navigation models with simple feature concatenation (Top) and our object-alignment module focused on landmarks for navigation (Bottom).
Sensors 23 06028 g001
Figure 2. Illustration of the proposed Object-level Alignment Module (OAlM), using instruction feature, visual features, and the match score of landmarks to predict action at each time.
Figure 2. Illustration of the proposed Object-level Alignment Module (OAlM), using instruction feature, visual features, and the match score of landmarks to predict action at each time.
Sensors 23 06028 g002
Figure 3. Visualizations of baseline and OAlM models. Two examples from the test set are shown where our approach is successful, but the baseline model stops too late or turns at the wrong place. Left: trajectory generated by baseline. Right: trajectory generated by our OAlM. (a) A case where the baseline model stops too late. (b) A case where the baseline model stops too early. (c) A case where the baseline model misses a turn. (d) A case where the baseline model turns at the wrong place.
Figure 3. Visualizations of baseline and OAlM models. Two examples from the test set are shown where our approach is successful, but the baseline model stops too late or turns at the wrong place. Left: trajectory generated by baseline. Right: trajectory generated by our OAlM. (a) A case where the baseline model stops too late. (b) A case where the baseline model stops too early. (c) A case where the baseline model misses a turn. (d) A case where the baseline model turns at the wrong place.
Sensors 23 06028 g003
Figure 4. Visualization of the failed navigation examples; the top part of each figure is the instruction, the left part represents the trajectory generated by our OAlM, and the right part represents the scene observed by the agent during partial real navigation. The green or orange circled parts in each figure on the left correspond to the specific scenario on the right, which is also marked in green or orange. (a) A case where the OAlM model stops early and is just one step away from the goal. (b) A case of navigation failure due to the complex road conditions. The red arrow in the right figure shows the correct road. The blue arrow shows the model-predicted road. (c) A case where the OAlM model’s inaccuracy causes navigation failures.
Figure 4. Visualization of the failed navigation examples; the top part of each figure is the instruction, the left part represents the trajectory generated by our OAlM, and the right part represents the scene observed by the agent during partial real navigation. The green or orange circled parts in each figure on the left correspond to the specific scenario on the right, which is also marked in green or orange. (a) A case where the OAlM model stops early and is just one step away from the goal. (b) A case of navigation failure due to the complex road conditions. The red arrow in the right figure shows the correct road. The blue arrow shows the model-predicted road. (c) A case where the OAlM model’s inaccuracy causes navigation failures.
Sensors 23 06028 g004
Figure 5. Navigation performance when masking out object tokens from instructions. (a) Change in TC for RCONCAT. (b) Change in TC for ORAR.
Figure 5. Navigation performance when masking out object tokens from instructions. (a) Change in TC for RCONCAT. (b) Change in TC for ORAR.
Sensors 23 06028 g005
Table 1. Navigation results when adding OCR to OAlM. The blue numbers in the table indicate results that were worse than the baseline, while the red numbers indicate results that were better than the baseline. Bold numbers indicate the highest score in the column (per model).
Table 1. Navigation results when adding OCR to OAlM. The blue numbers in the table indicate results that were worse than the baseline, while the red numbers indicate results that were better than the baseline. Bold numbers indicate the highest score in the column (per model).
ExperimentTCSPDSEDCLSnDTWsDTW
ORAR (baseline)24.2317.3023.7056.8737.2022.87
OAlM (+OCR 0.70)25.8616.9425.2758.2138.2324.29
OAlM (+OCR 0.80) 24.5117.3923.9256.5336.3723.06
OAlM (+OCR 0.90) 25.8417.4525.1756.3737.4524.20
RCONCAT (baseline)8.9422.488.5543.2318.207.98
OAlM (+OCR 0.70) 11.6421.1511.2644.1120.9510.74
OAlM (+OCR 0.80) 9.9421.479.6744.2820.349.21
OAlM (+OCR 0.90) 11.0720.9910.4144.1921.109.94
GA (baseline)9.8720.349.4247.7721.518.92
OAlM (+OCR 0.70) 10.3620.589.9148.0122.559.24
OAlM (+OCR 0.80) 11.4320.2011.1149.0323.8410.51
OAlM (+OCR 0.90) 10.9319.6210.6347.8522.6810.09
Table 2. Navigation results adding object-detection results to OAlM. The blue numbers in the table indicate results that were worse than the baseline, while the red numbers indicate results that were better than the baseline. Bold numbers indicate the highest score in the column (per model).
Table 2. Navigation results adding object-detection results to OAlM. The blue numbers in the table indicate results that were worse than the baseline, while the red numbers indicate results that were better than the baseline. Bold numbers indicate the highest score in the column (per model).
ExperimentTCSPDSEDCLSnDTWsDTW
ORAR (baseline)24.2317.3023.7056.8737.2022.87
OAlM (+OD 0.70)24.7016.7624.1057.6636.9923.04
OAlM (+OD 0.80)24.9417.0824.3256.9137.0623.39
OAlM (+OD 0.90)24.7216.5224.1258.3337.6323.20
RCONCAT (baseline)8.9422.488.5543.2318.207.98
OAlM (+OD 0.70)9.6521.849.3143.4020.248.83
OAlM (+OD 0.80)8.6621.458.4342.8720.528.11
OAlM (+OD 0.90)10.1520.569.7745.9921.229.23
GA (baseline)9.8720.349.4247.7721.518.92
OAlM (+OD 0.70) 11.5720.3911.2347.8922.6110.60
OAlM (+OD 0.80)13.0620.3612.5648.4923.3611.86
OAlM (+OD 0.90)10.8620.7910.5748.1722.9810.06
Table 3. Percentage of stop-location errors of failed predicted paths.
Table 3. Percentage of stop-location errors of failed predicted paths.
BaselinesOAlM
ORAR32.764.38
RCONCAT15.347.87
GA23.235.84
Table 4. Navigation results adding both OCR and object-detection results to OAlM. The blue numbers in the table indicate results that were worse than the baseline, while the red numbers indicate results that were better than the baseline.
Table 4. Navigation results adding both OCR and object-detection results to OAlM. The blue numbers in the table indicate results that were worse than the baseline, while the red numbers indicate results that were better than the baseline.
ExperimentTCSPDSEDCLSnDTWsDTW
ORAR (baseline)24.2317.3023.7056.8737.2022.87
+OCR&OD 0.4024.9616.5324.4858.6538.0623.48
+OCR&OD 0.5024.8116.8524.1057.0836.9023.19
+OCR&OD 0.6022.7318.2022.2355.0235.0121.41
+OCR&OD 0.7022.9216.9022.8757.4136.3721.43
+OCR&OD 0.8023.3517.1022.8257.1436.5021.99
+OCR&OD 0.9022.2718.4721.7353.5734.3420.85
RCONCAT (baseline)8.9422.488.5543.2318.207.98
+OCR&OD 0.4010.1521.459.5843.6320.058.90
+OCR&OD 0.509.4421.519.0344.0820.258.49
+OCR&OD 0.608.5221.788.2842.3018.827.91
+OCR&OD 0.709.0121.698.5544.4619.057.73
+OCR&OD 0.809.6521.459.2044.4320.248.56
+OCR&OD 0.908.8721.028.4944.3520.007.93
GA (baseline)9.8720.349.4247.7721.518.92
+OCR&OD 0.4011.5019.8211.1949.1923.7610.57
+OCR&OD 0.5010.0119.709.7348.3322.179.22
+OCR&OD 0.6011.7119.1711.2849.0523.6710.69
+OCR&OD 0.707.6723.237.3442.9516.326.74
+OCR&OD 0.809.7921.479.4446.4420.608.80
+OCR&OD 0.9010.4320.9710.1746.1920.639.70
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, Y.; Qiu, Y.; Aoki, Y.; Kataoka, H. Outdoor Vision-and-Language Navigation Needs Object-Level Alignment. Sensors 2023, 23, 6028. https://doi.org/10.3390/s23136028

AMA Style

Sun Y, Qiu Y, Aoki Y, Kataoka H. Outdoor Vision-and-Language Navigation Needs Object-Level Alignment. Sensors. 2023; 23(13):6028. https://doi.org/10.3390/s23136028

Chicago/Turabian Style

Sun, Yanjun, Yue Qiu, Yoshimitsu Aoki, and Hirokatsu Kataoka. 2023. "Outdoor Vision-and-Language Navigation Needs Object-Level Alignment" Sensors 23, no. 13: 6028. https://doi.org/10.3390/s23136028

APA Style

Sun, Y., Qiu, Y., Aoki, Y., & Kataoka, H. (2023). Outdoor Vision-and-Language Navigation Needs Object-Level Alignment. Sensors, 23(13), 6028. https://doi.org/10.3390/s23136028

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop