Publications

What Do We Learn from a Large-Scale Study of Pre-Trained Visual Representations in Sim and Real Environments?
What Do We Learn from a Large-Scale Study of Pre-Trained Visual Representations in Sim and Real Environments?

We conduct a study on using pre-trained visual representations (PVRs) to train robots for real-world tasks.

Where are we in the search for an Artificial Visual Cortex for Embodied Intelligence?
Where are we in the search for an Artificial Visual Cortex for Embodied Intelligence?

We present the largest and most comprehensive empirical study of visual foundation models for Embodied AI (EAI).

HomeRobot: Open-Vocabulary Mobile Manipulation
HomeRobot: Open-Vocabulary Mobile Manipulation

We propose a combined simulation and real-world benchmark on the problem of Open-Vocabulary Mobile Manipulation (OVMM).

Navigating to Objects Specified by Images
Navigating to Objects Specified by Images

We present a modular system that can perform well on the Instance ImageNav task in both simulation and the real world.

Habitat-matterport 3d semantics dataset
Habitat-matterport 3d semantics dataset

We present Habitat-Matterport 3D Semantics (HM3DSEM), the largest dataset of 3D real-world spaces with densely annotated semantics.

OVRL: Offline Visual Representation Learning for Embodied Navigation
OVRL: Offline Visual Representation Learning for Embodied Navigation

In this work we propose OVRL, a two-stage representation learning strategy for visual navigation tasks in Embodied AI.

Learning to Prevent Monocular SLAM Failure using Reinforcement Learning
Learning to Prevent Monocular SLAM Failure using Reinforcement Learning

In this work, we develop a novel formulation based on Reinforcement Learning that generates fail safe trajectories while using Monocular SLAM for localization.