Target-driven visual navigation in indoor scenes using deep reinforcement learning
- Yuke Zhu, Roozbeh Mottaghi, Ali Farhadi
- Computer ScienceIEEE International Conference on Robotics and…
- 16 September 2016
This paper proposes an actor-critic model whose policy is a function of the goal as well as the current state, which allows better generalization and proposes the AI2-THOR framework, which provides an environment with high-quality 3D scenes and a physics engine.
Multimodal Model-Agnostic Meta-Learning via Task-Aware Modulation
- Risto Vuorio, Shao-Hua Sun, Hexiang Hu, Joseph J. Lim
- Computer ScienceNeural Information Processing Systems
- 30 October 2019
This paper proposes a multimodal MAML (MMAML) framework, which is able to modulate its meta-learned prior parameters according to the identified mode, allowing more efficient fast adaptation and demonstrating the effectiveness of the model in modulating the meta-learning prior in response to the characteristics of tasks.
Multi-view to Novel View: Synthesizing Novel Views With Self-learned Confidence
- Shao-Hua Sun, Minyoung Huh, Yuan-Hong Liao, Ning Zhang, Joseph J. Lim
- Computer ScienceEuropean Conference on Computer Vision
- 8 September 2018
This paper proposes an end-to-end trainable framework that learns to exploit multiple viewpoints to synthesize a novel view without any 3D supervision, and introduces a self-learned confidence aggregation mechanism.
Learning and Using the Arrow of Time
- D. Wei, Joseph J. Lim, Andrew Zisserman, W. Freeman
- Computer ScienceIEEE/CVF Conference on Computer Vision and…
- 1 June 2018
A ConvNet suitable for extended temporal footprints and for class activation visualization, and the effect of artificial cues, such as cinematographic conventions, on learning is studied, which achieves state-of-the-art performance on large-scale real-world video datasets.
Accelerating Reinforcement Learning with Learned Skill Priors
- Karl Pertsch, Youngwoon Lee, Joseph J. Lim
- Computer ScienceConference on Robot Learning
- 22 October 2020
This work proposes a deep latent variable model that jointly learns an embedding space of skills and the skill prior from offline agent experience, and extends common maximum-entropy RL approaches to use skill priors to guide downstream learning.
Demo2Vec: Reasoning Object Affordances from Online Videos
- Kuan Fang, Te-Lin Wu, Daniel Yang, S. Savarese, Joseph J. Lim
- Computer Science, PhysicsIEEE/CVF Conference on Computer Vision and…
- 1 June 2018
The Demo2Vec model is designed which learns to extract embedded vectors of demonstration videos and predicts the interaction region and the action label on a target image of the same object.
High-fidelity facial and speech animation for VR HMDs
- Kyle Olszewski, Joseph J. Lim, Shunsuke Saito, Hao Li
- Computer ScienceACM Transactions on Graphics
- 11 November 2016
This work introduces a novel system for HMD users to control a digital avatar in real-time while producing plausible speech animation and emotional expressions and demonstrates the quality of the system on a variety of subjects and evaluates its performance against state-of-the-art real- time facial tracking techniques.
Neural Program Synthesis from Diverse Demonstration Videos
- Shao-Hua Sun, Hyeonwoo Noh, S. Somasundaram, Joseph J. Lim
- Computer ScienceInternational Conference on Machine Learning
- 3 July 2018
This work proposes a neural program synthesizer that is able to explicitly synthesize underlying programs from behaviorally diverse and visually complicated demonstration videos and introduces a summarizer module as part of the model to improve the network’s ability to integrate multiple demonstrations varying in behavior.
Learning to Coordinate Manipulation Skills via Skill Behavior Diversification
- Youngwoon Lee, Jingyun Yang, Joseph J. Lim
- Computer ScienceInternational Conference on Learning…
- 30 April 2020
The proposed framework is able to efficiently learn sub-skills with diverse behaviors and coordinate them to solve challenging collaborative control tasks such as picking up a long bar, placing a block inside a container while pushing the container with two robot arms, and pushing a box with two ant agents.
Program Guided Agent
- Shao-Hua Sun, Te-Lin Wu, Joseph J. Lim
- Computer ScienceInternational Conference on Learning…
- 30 April 2020
Experimental results on a 2D Minecraft environment not only demonstrate that the proposed framework learns to reliably accomplish program instructions and achieves zero-shot generalization to more complex instructions but also verify the efficiency of the proposed modulation mechanism for learning the multitask policy.
...
...