• Publications
  • Influence
AI2-THOR: An Interactive 3D Environment for Visual AI
TLDR
AI2-THOR consists of near photo-realistic 3D indoor scenes, where AI agents can navigate in the scenes and interact with objects to perform tasks and facilitate building visually intelligent models.
RoboTHOR: An Open Simulation-to-Real Embodied AI Platform
TLDR
RoboTHOR offers a framework of simulated environments paired with physical counterparts to systematically explore and overcome the challenges of simulation-to-real transfer, and a platform where researchers across the globe can remotely test their embodied models in the physical world.
Artificial Agents Learn Flexible Visual Representations by Playing a Hiding Game
TLDR
This work is the first to show that embodied adversarial reinforcement learning agents playing cache, a variant of hide-and-seek, in a high fidelity, interactive, environment, learn representations of their observations encoding information such as occlusion, object permanence, free space, and containment.
Beyond Sentential Semantic Parsing: Tackling the Math SAT with a Cascade of Tree Transducers
We present an approach for answering questions that span multiple sentences and exhibit sophisticated cross-sentence anaphoric phenomena, evaluating on a rich source of such questions – the math
ManipulaTHOR: A Framework for Visual Object Manipulation
TLDR
This work proposes a framework for object manipulation built upon the physics-enabled, visually rich AI2-THOR framework and presents a new challenge to the Embodied AI community known as ArmPointNav, which extends the popular point navigation task to object manipulation and offers new challenges including 3D obstacle avoidance.
Interactive Visualization for Linguistic Structure
TLDR
The library is not tied to any particular linguistic representation, but provides a general-purpose API for the interactive exploration of hierarchical linguistic structure, and offers several important features, including expand/collapse functionality, positional and color cues, and explicit visual support for sequential structure.
A Case Study in Hybrid Multi-threading and Hierarchical Reinforcement Learning Approach for Cooperative Multi-agent Systems
TLDR
Experimental results show that this multi-agent system can reduce the time process and still maintain independence of agents.
Iconary: A Pictionary-Based Game for Testing Multimodal Communication with Drawings and Text
TLDR
This work proposes models to play Iconary, a collaborative game of drawing and guessing based on Pictionary, that poses a novel challenge for the research community, and proposes models that are skillful players and able to employ world knowledge in language models toplay with words unseen during training.