计算机与现代化 ›› 2023, Vol. 0 ›› Issue (07): 30-35.doi: 10.3969/j.issn.1006-2475.2023.07.006

• 人工智能 • 上一篇    下一篇

基于D3QN的交通灯控制优化

  

  1. (太原科技大学计算机科学与技术学院,山西 太原 030024)
  • 出版日期:2023-07-26 发布日期:2023-07-27
  • 作者简介:张国有(1972—),男,山西忻州人,副教授,博士,研究方向:群机器人,系统建模,E-mail:2003072@tyust.edu.cn; 通信作者:宋世峰(1997—),男,山西代县人,硕士研究生,研究方向:强化学习,智能交通,E-mail: 18734484910@163.com。
  • 基金资助:
    国家自然科学基金面上项目(62072325)

Traffic Light Control Optimization Based On D3QN

  1. (School of Computer Science and Technology, Taiyuan University of Science and Technology, Taiyuan 030024, China)
  • Online:2023-07-26 Published:2023-07-27

摘要: 交通灯在控制路口车辆通行上起着至关重要的作用。现阶段城市路口的交通灯多采用固定配时、固定相位变换的控制策略,难以满足不同的车流情况。设计出能够根据路口车流情况实时调整交通灯变换的控制方案成为智能交通领域的研究热点之一。而城市路口车流具有动态变化性,难以直接对其展开研究。为了设计一种合适的交通灯动态控制方案,本文引入深度强学习技术。将十字路口交通灯控制问题抽象成强化学习模型,采用D3QN算法对该模型进行求解。在此基础上综合考虑处于不同状态的车辆,改进状态输入和奖励函数。最终在交通模拟器SUMO上进行不同车流下的仿真实验。实验结果表明,模型训练趋于稳定后,改进奖励函数和状态输入的D3QN算法的平均队列长度在3种车流量下对比传统的固定控制策略和自适应控制策略均有明显提升,对比DQN和DDQN算法也有一定的优化,控制效果更佳。

关键词: 交通灯控制, 深度强化学习, D3QN算法, 状态输入, 奖励函数, 车流情况

Abstract: Traffic lights play a vital role in controlling traffic at intersections. At present, the traffic lights at urban intersections mostly adopt the control strategy of fixed timing and fixed phase transformation, which is difficult to meet different traffic flow conditions. It has become one of the research hotspots in the field of intelligent transportation to design a control scheme that can adjust the traffic light transformation in real time according to the traffic flow at the intersection. However, the traffic flow at urban intersections is dynamic, so it is difficult to study it directly. In order to design an appropriate traffic light dynamic control scheme, the deep strong learning technology is introduced. The intersection traffic light control problem is abstracted into a reinforcement learning model, which is solved by D3QN algorithm. On this basis, considering the vehicles in different states, the state input and reward function are improved. Finally, the simulation experiments under different traffic flows are carried out on the traffic simulator SUMO. The experimental results show that after the model training becomes stable, the average queue length of the D3QN algorithm with improved reward function and state input is significantly improved compared with the traditional fixed control strategy and adaptive control strategy under three traffic flows, and the control effect is better then DQN algorithms and DDQN algorithms.

Key words: traffic light control, deep reinforcement learning, D3QN algorithm, status input, reward function, traffic flow

中图分类号: