论文标题

出席灯:流量信号控制的基于普遍注意力的强化学习模型

AttendLight: Universal Attention-Based Reinforcement Learning Model for Traffic Signal Control

论文作者

Oroojlooy, Afshin, Nazari, Mohammadreza, Hajinezhad, Davood, Silva, Jorge

论文摘要

我们建议出席灯,这是交通信号控制问题的端到端增强学习(RL)算法。此问题的先前方法的缺点是,他们需要为每个新的交叉路口进行培训,并具有不同的结构或交通流量分布。出席灯通过训练单个通用模型来解决与任何数量的道路,车道,相位(可能的信号)和交通流量的交集来解决此问题。为此,我们提出了一个深入的RL模型,该模型包含了两个注意力模型。引入了第一个注意模型来处理不同数量的道路灯笼。第二个注意模型旨在在交集中使用任意数量的阶段实现决策。结果,我们提出的模型适用于任何相交配置,只要在训练集中表示类似的配置即可。实验是使用合成和现实世界标准基准数据集进行的。我们显示的结果涵盖了与三到四条通往道路的交叉点;一个,两条车道的单向/双向道路;不同数量的阶段;和不同的交通流量。我们考虑了两个制度:(i)单环境培训,单部门和(ii)多环境培训,多部署。在这两个制度中,所有情况下的所有情况都超过了古典和其他基于RL的方法。

We propose AttendLight, an end-to-end Reinforcement Learning (RL) algorithm for the problem of traffic signal control. Previous approaches for this problem have the shortcoming that they require training for each new intersection with a different structure or traffic flow distribution. AttendLight solves this issue by training a single, universal model for intersections with any number of roads, lanes, phases (possible signals), and traffic flow. To this end, we propose a deep RL model which incorporates two attention models. The first attention model is introduced to handle different numbers of roads-lanes; and the second attention model is intended for enabling decision-making with any number of phases in an intersection. As a result, our proposed model works for any intersection configuration, as long as a similar configuration is represented in the training set. Experiments were conducted with both synthetic and real-world standard benchmark data-sets. The results we show cover intersections with three or four approaching roads; one-directional/bi-directional roads with one, two, and three lanes; different number of phases; and different traffic flows. We consider two regimes: (i) single-environment training, single-deployment, and (ii) multi-environment training, multi-deployment. AttendLight outperforms both classical and other RL-based approaches on all cases in both regimes.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源