Corpus ID: 212725223

Cooperation without Coordination: Hierarchical Predictive Planning for Decentralized Multiagent Navigation

@article{Wang2020CooperationWC,
  title={Cooperation without Coordination: Hierarchical Predictive Planning for Decentralized Multiagent Navigation},
  author={Rose E. Wang and J. Kew and Dennis Lee and T. Lee and T. Zhang and Brian Ichter and Jie Tan and Aleksandra Faust},
  journal={ArXiv},
  year={2020},
  volume={abs/2003.06906}
}
Decentralized multiagent planning raises many challenges, such as adaption to changing environments inexplicable by the agent's own behavior, coordination from noisy sensor inputs like lidar, cooperation without knowing other agents' intents. To address these challenges, we present hierarchical predictive planning (HPP) for decentralized multiagent navigation tasks. HPP learns prediction models for itself and other teammates, and uses the prediction models to propose and evaluate navigation… Expand
Spatial Intention Maps for Multi-Agent Mobile Manipulation
TLDR
Experiments across a variety of multi-agent environments show that incorporating spatial intention maps improves performance for different mobile manipulation tasks while significantly enhancing cooperative behaviors. Expand

References

SHOWING 1-10 OF 31 REFERENCES
Long-Range Indoor Navigation With PRM-RL
TLDR
This article uses probabilistic roadmaps (PRMs) as the sampling-based planner, and AutoRL as the RL method in the indoor navigation context, and shows that PRM-RL with AutoRL is more successful than several baselines, is robust to noise, and can guide robots over hundreds of meters in the face of noise and obstacles in both simulation and on robots. Expand
R-MADDPG for Partially Observable Environments and Limited Communication
TLDR
A deep recurrent multiagent actor-critic framework (R-MADDPG) for handling multiagent coordination under partial observable set-tings and limited communication and demonstrates that the resulting framework learns time dependencies for sharing missing observations, handling resource limitations, and developing different communication patterns among agents. Expand
Multiagent Planning with Factored MDPs
TLDR
This work presents a principled and efficient planning algorithm for cooperative multiagent dynamic systems that avoids the exponential blowup in the state and action space and is an efficient alternative to more complicated algorithms even in the single agent case. Expand
Ad Hoc Autonomous Agent Teams: Collaboration without Pre-Coordination
TLDR
This paper defines the concept of ad hoc team agents, specifies an evaluation paradigm, and provides examples of possible theoretical and empirical approaches to challenge to encourage progress towards this ambitious, newly realistic, and increasingly important research goal. Expand
Efficient Hierarchical Robot Motion Planning Under Uncertainty and Hybrid Dynamics
TLDR
This work proposes a hierarchical POMDP planner that develops cost-optimized motion plans for hybrid dynamics models and evaluates the approach on a navigation task in the simulated domain and on an assembly task with a robotic manipulator, showing that the approach can solve tasks having high observation noise and nonlinear dynamics effectively with significantly lower computational costs compared to direct planning approaches. Expand
Learning Navigation Behaviors End-to-End With AutoRL
TLDR
Empirical evaluations show that AutoRL policies do not suffer from the catastrophic forgetfulness that plagues many other deep reinforcement learning algorithms, generalize to new environments and moving obstacles, are robust to sensor, actuator, and localization noise, and can serve as robust building blocks for larger navigation tasks. Expand
RL-RRT: Kinodynamic Motion Planning via Learning Reachability Estimators From RL Policies
TLDR
RL-RRT is introduced that uses an obstacle-avoiding policy that maps a robot's sensor observations to actions, which is used as a local planner during planning and as a controller during execution, and also provides a shorter path finish time than a steering function free method. Expand
Counterfactual Multi-Agent Policy Gradients
TLDR
A new multi-agent actor-critic method called counterfactual multi- agent (COMA) policy gradients, which uses a centralised critic to estimate the Q-function and decentralised actors to optimise the agents' policies. Expand
Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments
TLDR
An adaptation of actor-critic methods that considers action policies of other agents and is able to successfully learn policies that require complex multi-agent coordination is presented. Expand
Getting Robots Unfrozen and Unlost in Dense Pedestrian Crowds
TLDR
This work enhances the robot's mobility and unfreeze the robot in the crowd using a reinforcement learning-based local navigation policy developed in the previous work which naturally takes into account the coordination between robots and humans. Expand
...
1
2
3
4
...