26/10/2020

Reinforcement Learning for Zone Based Multiagent Pathfinding under Uncertainty

Jiajing Ling, Tarun Gupta, Akshat Kumar

Keywords: multiagent path finding, multiagent reinforcement learning, decision making under uncertainty, decentralized partially observable MDPs

Abstract: We present a new framework for the problem of multiple agents finding their paths from respective source to destination nodes in a graph (also called MAPF). Most existing approaches assume that all agents move at fixed speed, and that a single node accommodates only a single agent. Motivated by the emerging applications of autonomous vehicles such as drone traffic management, we present zone-based path finding (or ZBPF) where agents move among zones (e.g., geofenced airblocks for drones), and agents' movements require uncertain travel time. Furthermore, each zone can accommodate multiple agents (as per its capacity). We also develop a 3D simulator for ZBPF in the ml-agents platform of the Unity3D game engine, which provides a clean interface from the simulation environment to learning algorithms. We develop a novel formulation of the ZBPF problem using difference-of-convex functions (DC) programming. The resulting approach utilizes samples from the simulator to optimize agent policies. We also present a multiagent credit assignment scheme that helps our learning approach converge faster. Empirical results in a number of 2D and 3D instances show that our approach can effectively minimize congestion in zones, while ensuring agents reach their final destinations.

 0
 0
 0
 0
This is an embedded video. Talk and the respective paper are published at ICAPS 2020 virtual conference. If you are one of the authors of the paper and want to manage your upload, see the question "My papertalk has been externally embedded..." in the FAQ section.

Comments

Post Comment
no comments yet
code of conduct: tbd Characters remaining: 140

Similar Papers