• Corpus ID: 246652464

Navigating to Objects in Unseen Environments by Distance Prediction

@article{Zhu2022NavigatingTO,
  title={Navigating to Objects in Unseen Environments by Distance Prediction},
  author={Minzhao Zhu and Binglei Zhao and Tao Kong},
  journal={ArXiv},
  year={2022},
  volume={abs/2202.03735}
}
—Object Goal Navigation (ObjectNav) task is to navigate an agent to an object category in unseen environments without a pre-built map. In this paper, we solve this task by predicting the distance to the target using semantically-related objects as cues. Based on the estimated distance to the target object, our method directly choose optimal mid-term goals that are more likely to have a shorter path to the target. Specifically, based on the learned knowledge, our model takes a bird’s-eye view… 

References

SHOWING 1-10 OF 63 REFERENCES

Object Goal Navigation using Goal-Oriented Semantic Exploration

A modular system called, `Goal-Oriented Semantic Exploration' which builds an episodic semantic map and uses it to explore the environment efficiently based on the goal object category and outperforms a wide range of baselines including end-to-end learning-based methods as well as modular map- based methods.

Learning to Map for Active Semantic Goal Navigation

This work proposes a novel framework that actively learns to generate semantic maps outside the field of view of the agent and leverages the uncertainty over the semantic classes in the unobserved areas to decide on long term goals.

Learning hierarchical relationships for object-goal navigation

Memory-utilized Joint hierarchical Object Learning for Navigation in Indoor Rooms (MJOLNIR), a target-driven visual navigation algorithm, which considers the inherent relationship between "target" objects, along with the more salient "parent" objects occurring in its surrounding, and learns to converge much faster than other algorithms.

ObjectNav Revisited: On Evaluation of Embodied Agents Navigating to Objects

This document summarizes the consensus recommendations of this working group on ObjectNav and makes recommendations on subtle but important details of evaluation criteria, the agent's embodiment parameters, and the characteristics of the environments within which the task is carried out.

Cognitive Mapping and Planning for Visual Navigation

The Cognitive Mapper and Planner is based on a unified joint architecture for mapping and planning, such that the mapping is driven by the needs of the task, and a spatial memory with the ability to plan given an incomplete set of observations about the world.

Exploiting Scene-specific Features for Object Goal Navigation

A new reduced dataset is introduced that speeds up the training of navigation models, a notoriously complex task, and the SMTSC model is proposed, an attention-based model capable of exploiting the correlation between scenes and objects contained in them.

Visual Navigation with Spatial Attention

The attention model is shown to improve the agent’s policy and to achieve state-of-the-art results on commonly-used datasets.

SSCNav: Confidence-Aware Semantic Scene Completion for Visual Semantic Navigation

This paper introduces SSCNav, an algorithm that explicitly models scene priors using a confidence-aware semantic scene completion module to complete the scene and guide the agent's navigation planning and demonstrates that the proposed scenepletion module improves the efficiency of the downstream navigation policies.

Seeing the Un-Scene: Learning Amodal Semantic Maps for Room Navigation

A learning-based approach for room navigation using semantic maps that learns to predict top-down belief maps of regions that lie beyond the agent's field of view while modeling architectural and stylistic regularities in houses.

Occupancy Anticipation for Efficient Exploration and Navigation

This work proposes occupancy anticipation, where the agent uses its egocentric RGB-D observations to infer the occupancy state beyond the visible regions, which facilitates efficient exploration and navigation in 3D environments.
...