26/10/2020

Guidelines for Action Space Definition in Reinforcement Learning-Based Traffic Signal Control Systems

Maxime Treca, Julian Garbiso, Dominique Barth

Keywords: Learning, Real-time scheduling, Traffic signal control, Reinforcement Learning, Action space, Mobility

Abstract: Traffic signal control is an urban planning tool with important economic, social and environmental implications. Reinforcement learning applied to traffic signal control (RL-TSC) has shown promising results compared to existing methods. If previous works in the RL-TSC literature have focused on optimizing state and reward definitions, the impact of the agent's action space definition remains largely unexplored. Indeed, typical RL-TSC models feature either phase-based controllers — which determine a signal duration in one go — or step-based controllers — which can decide to extend a phase duration interactively — without comparing their respective merits. In this paper, we provide guidelines for optimally defining RL-TSC actions by comparing different action types in a simulated network featuring different traffic demand patterns. Our results show that an agent's performance and convergence speed both increase with its interaction frequency with the environment. However, certain methods with lower observation frequencies — that can be achieved with realistic sensing technologies — have reasonably similar performance compared to higher frequency ones in all scenarios, and even outperform them under specific traffic conditions.

 0
 0
 0
 0
This is an embedded video. Talk and the respective paper are published at ICAPS 2020 virtual conference. If you are one of the authors of the paper and want to manage your upload, see the question "My papertalk has been externally embedded..." in the FAQ section.

Comments

Post Comment
no comments yet
code of conduct: tbd Characters remaining: 140

Similar Papers