Please wait a minute...
Journal of ZheJiang University (Engineering Science)  2024, Vol. 58 Issue (8): 1671-1680    DOI: 10.3785/j.issn.1008-973X.2024.08.014
    
Traffic signal control method based on asynchronous advantage actor-critic
Baolin YE1,2(),Ruitao SUN1,2,Weimin WU3,Bin CHEN2,Qing YAO4
1. School of Information Science and Engineering, Zhejiang Sci-Tech University, Hangzhou 310018, China
2. Jiaxing Key Laboratory of Smart Transportations, Jiaxing University, Jiaxing 314001, China
3. State Key Laboratory of Industrial Control Technology, Institute ofCyber-Systems and Control, Zhejiang University, Hangzhou 310027, China
4. School of Computer Science and Technology, Zhejiang Sci-Tech University, Hangzhou 310018, China
Download: HTML     PDF(1889KB) HTML
Export: BibTeX | EndNote (RIS)      

Abstract  

A single intersection traffic signal control method based on the asynchronous advantage actor-critic (A3C) algorithm was proposed aiming at high cost of model learning and decision making in the existing traffic signal control methods based on deep reinforcement learning. Vehicle weight gain network was constructed from two different dimensions at the input side of the model, namely intersections and lanes, in order to preprocess the collected vehicle state information. A new reward mechanism was designed and an A3C algorithm that integrated vehicle weight gain networks was proposed. The simulation test results based on the microscopic traffic simulation software simulation of urban mobility (SUMO) show that the proposed method achieves better traffic signal control performance under three different traffic flow conditions of low, medium and high levels compared with traditional traffic signal control methods and benchmark reinforcement learning methods.



Key wordstraffic signal control      deep reinforcement learning      A3C      weight gain network     
Received: 26 August 2023      Published: 23 July 2024
CLC:  TP 393  
Fund:  国家自然科学基金资助项目(61603154);浙江省自然科学基金资助项目 ( LTGS23F030002); 嘉兴市应用性基础研究项目(2023AY11034);浙江省尖兵领雁研发攻关计划资助项目(2023C01174);工业控制技术国家重点实验室开放课题资助项目 (ICT2022B52).
Cite this article:

Baolin YE,Ruitao SUN,Weimin WU,Bin CHEN,Qing YAO. Traffic signal control method based on asynchronous advantage actor-critic. Journal of ZheJiang University (Engineering Science), 2024, 58(8): 1671-1680.

URL:

https://www.zjujournals.com/eng/10.3785/j.issn.1008-973X.2024.08.014     OR     https://www.zjujournals.com/eng/Y2024/V58/I8/1671


基于异步优势演员-评论家的交通信号控制方法

针对现有基于深度强化学习的交通信号控制方法的模型学习和决策成本高的问题,提出基于异步优势演员-评论家(A3C)算法的单交叉口交通信号控制方法. 在模型输入端分别从交叉口和车道2个不同维度构建车辆权重增益网络,对采集的车辆状态信息进行预处理. 设计新的奖励机制,提出融合车辆权重增益网络的A3C算法. 基于微观交通仿真软件SUMO的仿真测试结果表明,相比于传统的交通信号控制方法和基准强化学习方法,所提方法在低、中、高3种不同的交通流量状态下,均能够取得更好的交通信号控制效益.


关键词: 交通信号控制,  深度强化学习,  A3C,  权重增益网络 
Fig.1 Diagram of intersection
Fig.2 Diagram of phase setting of intersection
Fig.3 Diagram of action space
RwRpWchange_ratePchange_rate
4(0.35, +∞)
3(0.25, 0.35]
2(0.15, 0.25]
1[0, 0.15]
?1[?0.15, 0)
?2[?0.25, ?0.15)
?3[?0.35, ?0.25)
?4(?∞, ?0.35)
Tab.1 Reward setting rules
Fig.4 Diagram of information exchange between vehicle weight gain network and network of A2C
Fig.5 Diagram of intersection vehicle weight gain network
Fig.6 Diagram of lane vehicle weight gain network
算法1 XVWG-A3C算法
1)设定全局共享参数向量$ \boldsymbol{\theta } $$ {\boldsymbol{\theta }}_{v} $,全局共享计数器T. 假设线程特定的参数为$ {\boldsymbol{\theta }}' $$ {\boldsymbol{\theta }}_{v}' $. 2)初始化线程步骤计数器t←13)训练A3C网络和权重增益网络XVWG:1 repeat2 重置梯度$ {\mathrm{d}}\boldsymbol{\theta } $←0和d$ {\boldsymbol{\theta }}_{v} $0,重置线程参数$ {\boldsymbol{\theta }}' $=$ \boldsymbol{\theta } $, $ {\boldsymbol{\theta }}_{v}' $= $ {\boldsymbol{\theta }}_{v} $tstart = t;3 初始化神经网络的输入s,根据式(6)、(7)计算经过权重网络之后的状态st4 repeat5 根据策略函数$ \pi ({a_t}|{s_t};{{\boldsymbol{\theta}} '}) $执行动作at,得到奖励rt和经过权重网络之和的状态st+1;6 更新tt+1,TT+1;7 until 得到最终状态st 或者t?tstart==tmax7 若st+1为最终状态,R=0. 否则$ R = V({s_{t+1}},{\boldsymbol{\theta}} _v^{'}) $8 for $ i \in \{ t - 1, \cdots ,{t_{{\mathrm{start}}}}\} $do9 $ R \leftarrow {r_i}+\gamma R $10 计算累计梯度$ {{\boldsymbol{\theta}} ^{'}} \leftarrow {\mathrm{d}}{\boldsymbol{\theta}} + {\nabla _{{{\boldsymbol{\theta}} ^{'}}}}\ln \pi ({a_i}|{s_i};{{\boldsymbol{\theta}} ^{'}})(R - V({s_i};{\boldsymbol{\theta}} _v^{'})) $; $ {\boldsymbol{\theta}} _v^{'} \leftarrow {\mathrm{d}}{{\boldsymbol{\theta}} _v}+\partial {(R - V({s_i};{\boldsymbol{\theta}} _v^{'}))^2}/\partial {\boldsymbol{\theta}} _v^{'} $11 end for12 使用$ {\mathrm{d}}\boldsymbol{\theta } $和d$ {\boldsymbol{\theta }}_{v} $异步更新$ \boldsymbol{\theta } $$ {\boldsymbol{\theta }}_{v} $,使用式(18)~(22)更新车辆权重增益网络的参数Ψ;
13 until T > Tmax
 
Fig.7 Framework of A3C based traffic signal control fused with vehicle weight gain network
实验对比参数数值
融合权重增益网络的A3C
算法、传统A3C算法
Actor网络学习率0.000 02
Critic网络学习率0.000 2
Actor网络神经元数量200
Critic网络神经元数量100
折扣因子0.9
训练步数200
训练时间/s7 200
DQN算法学习率0.000 02
神经元数量200
折扣因子0.9
训练步数200
训练时间/s7 200
Tab.2 Parameter setting of various deep reinforcement learning model in comparative experiment
参数数值
车道长度/m100
平均车辆长度/m5
最小车辆间隔/m2.5
车辆最大速度/(m·s?1)13.89
车辆最大加速度/(m·s?2)2.6
车辆最大减速度/(m·s?2)4.6
黄灯时间/s3
相位保持时绿灯持续时间/s5
相位最小绿灯时间/s15
车辆直行概率0.5
车辆左转概率0.3
车辆右转概率0.2
Tab.3 Parameter setting of traffic simulation environment
Fig.8 Cumulative round reward
Fig.9 Average waiting time of vehicles
Fig.10 Average queue length of vehicles
Fig.11 Average number of vehicle stops
控制方法低流量中流量高流量
W/sL/mPW/sL/mPW/sL/mP
固定配时13.855.830.7715.686.280.9117.318.621.01
自适应5.403.730.557.854.000.5713.344.610.63
DQN6.143.540.47`7.383.810.558.324.050.59
A3C5.883.430.466.903.730.497.453.830.56
LVWG-A3C5.133.280.436.113.520.467.223.760.52
IVWG-A3C4.723.180.404.923.260.436.313.500.47
Tab.4 Test result of different traffic signal control methods under isolated intersection condition
[1]   YE B-L, WU W, RUAN K, et al A survey of model predictive control methods for traffic signal control[J]. IEEE/CAA Journal of Automatica Sinica, 2019, 6 (3): 623- 640
doi: 10.1109/JAS.2019.1911471
[2]   彭渠栩 优化双向“绿波带”关键路口控制参数算法的研究[J]. 应用数学进展, 2023, 12 (2): 781
PENG Quxu Research on the algorithm for optimizing the key intersection control parameters of two-way “Green Wave Belt”[J]. Advances in Applied Mathematics, 2023, 12 (2): 781
[3]   刘建伟, 高峰, 罗雄麟 基于值函数和策略梯度的深度强化学习综述[J]. 计算机学报, 2019, 42 (6): 1406- 1438
LIU Jianwei, GAO Feng, LUO Xionglin Review of deep reinforcement learning based on value functions and policy gradients[J]. Chinese Journal of Computers, 2019, 42 (6): 1406- 1438
doi: 10.11897/SP.J.1016.2019.01406
[4]   刘义, 何均宏 强化学习在城市交通信号灯控制方法中的应用[J]. 科技导报, 2019, 37 (6): 84- 90
LIU Yi, HE Junhong application of reinforcement learning in city traffic signal control methods.[J]. Science and Technology Review, 2019, 37 (6): 84- 90
[5]   FARID A, HUSSAIN F, KHAN K, et al A fast and accurate real-time vehicle detection method using deep learning for unconstrained environments[J]. Applied Sciences, 2023, 13 (5): 3059
doi: 10.3390/app13053059
[6]   MNIH V, KAVUKCUOGLU K, SILVER D, et al Human-level control through deep reinforcement learning[J]. Nature, 2015, 518 (7540): 529- 533
doi: 10.1038/nature14236
[7]   WANG M, WU L, LI J, et al Traffic signal control with reinforcement learning based on region-aware cooperative strategy[J]. IEEE Transactions on Intelligent Transportation Systems, 2021, 23 (7): 6774- 6785
[8]   WANG Z, YANG K, LI L, et al Traffic signal priority control based on shared experience multi-agent deep reinforcement learning[J]. IET Intelligent Transport Systems, 2023, 17 (7): 1363- 1379
[9]   MA D, ZHOU B, SONG X, et al A deep reinforcement learning approach to traffic signal control with temporal traffic pattern mining[J]. IEEE Transactions on Intelligent Transportation Systems, 2021, 23 (8): 11789- 11800
[10]   BOUKTIF S, CHENIKI A, OUNI A Traffic signal control using hybrid action space deep reinforcement learning[J]. Sensors, 2021, 21 (7): 2302
doi: 10.3390/s21072302
[11]   CHU T, WANG J, CODECA L, et al Multi-agent deep reinforcement learning for large-scale traffic signal control[J]. IEEE Transactions on Intelligent Transportation Systems, 2019, 21 (3): 1086- 1095
[12]   刘智敏, 叶宝林, 朱耀东, 等 基于深度强化学习的交通信号控制方法[J]. 浙江大学学报: 工学版, 2022, 56 (6): 1249- 1256
LIU Zhimin, YE Baolin, ZHU Yaodong, et al Traffic signal control methods based on deep reinforcement learning[J]. Journal of Zhejiang University: Engineering Science, 2022, 56 (6): 1249- 1256
[13]   WU T, ZHOU P, LIU K, et al Multi-agent deep reinforcement learning for urban traffic light control in vehicular networks[J]. IEEE Transactions on Vehicular Technology, 2020, 69 (8): 8243- 8256
doi: 10.1109/TVT.2020.2997896
[14]   赵乾, 张灵, 赵刚, 等 双环相位结构约束下的强化学习交通信号控制方法[J]. 交通运输工程与信息学报, 2023, 21 (1): 19- 28
ZHAO Qian, ZHANG Ling, ZHAO Gang, et al Reinforcement learning traffic signal control method under dual-ring phase structure constraints[J]. Journal of Transportation Engineering and Information, 2023, 21 (1): 19- 28
[15]   王安麟, 孙晓龙, 钟馥声 一种基于通行优先度规则的城市交通信号自组织控制方法[J]. 重庆交通大学学报: 自然科学版, 2018, 37 (2): 96
WANG Anlin, SUN Xiaolong, ZHONG Fusheng A self-organized control method for urban traffic signals based on priority rules for passage[J]. Journal of Chongqing Jiaotong University: Natural Science, 2018, 37 (2): 96
[16]   ASIAIN E, CLEMPNER J B, POZNYAK A S Controller exploitation-exploration reinforcement learning architecture for computing near-optimal policies[J]. Soft Computing, 2019, 23 (11): 3591- 3604
doi: 10.1007/s00500-018-3225-7
[17]   TROIA S, SAPIENZA F, VARÉ L, et al On deep reinforcement learning for traffic engineering in SD-WAN[J]. IEEE Journal on Selected Areas in Communications, 2021, 39 (7): 2198- 2212
doi: 10.1109/JSAC.2020.3041385
[18]   TAN K L, SHARMA A, SARKAR S Robust deep reinforcement learning for traffic signal control[J]. Journal of Big Data Analytics in Transportation, 2020, 2: 263- 274
doi: 10.1007/s42421-020-00029-6
[19]   LI M, LI Z, XU C, et al Deep reinforcement learning-based vehicle driving strategy to reduce crash risks in traffic oscillations[J]. Transportation Research Record, 2020, 2674 (10): 42- 54
doi: 10.1177/0361198120937976
[20]   CHU T, WANG J, CODECÀ L, et al Multi-agent deep reinforcement learning for large-scale traffic signal control[J]. IEEE Transactions on Intelligent Transportation Systems, 2020, 21 (3): 1086- 1095
doi: 10.1109/TITS.2019.2901791
[1] Meng ZHANG,Dian-hai WANG,Sheng JIN. Deep reinforcement learning approach to signal control combined with domain experience[J]. Journal of ZheJiang University (Engineering Science), 2023, 57(12): 2524-2532.
[2] Yu-feng JIANG,Dong-sheng CHEN. Assembly strategy for large-diameter peg-in-hole based on deep reinforcement learning[J]. Journal of ZheJiang University (Engineering Science), 2023, 57(11): 2210-2216.
[3] Xia HUA,Xin-qing WANG,Ting RUI,Fa-ming SHAO,Dong WANG. Vision-driven end-to-end maneuvering object tracking of UAV[J]. Journal of ZheJiang University (Engineering Science), 2022, 56(7): 1464-1472.
[4] Zhi-min LIU,Bao-Lin YE,Yao-dong ZHU,Qing YAO,Wei-min WU. Traffic signal control method based on deep reinforcement learning[J]. Journal of ZheJiang University (Engineering Science), 2022, 56(6): 1249-1256.
[5] Qi-lin DENG,Juan LU,Yong-hui CHEN,Jian FENG,Xiao-ping LIAO,Jun-yan MA. Optimization method of CNC milling parameters based on deep reinforcement learning[J]. Journal of ZheJiang University (Engineering Science), 2022, 56(11): 2145-2155.
[6] Yi-fan MA,Fan-yu ZHAO,Xin WANG,Zhong-he JIN. Satellite earth observation task planning method based on improved pointer networks[J]. Journal of ZheJiang University (Engineering Science), 2021, 55(2): 395-401.
[7] Kai LU,Xin TIAN,Guan-rong LIN,Xing-dong DENG. Simultaneous optimization model of signal phase design and timing at intersection[J]. Journal of ZheJiang University (Engineering Science), 2020, 54(5): 921-930.
[8] LI Wen-jing, SUN Feng, LI Xi-yao, MA Dong-fang. Time-of-day breakpoints for traffic signal control using dynamic recurrence order clustering[J]. Journal of ZheJiang University (Engineering Science), 2018, 52(6): 1150-1156.