东北大学学报(自然科学版) ›› 2023, Vol. 44 ›› Issue (10): 1369-1376.DOI: 10.12068/j.issn.1005-3026.2023.10.001

• 信息与控制 •    下一篇

基于双决斗深度Q网络的自动换道决策模型

张雪峰, 王照乙   

  1. (东北大学 理学院, 辽宁 沈阳110819)
  • 发布日期:2023-10-27
  • 通讯作者: 张雪峰
  • 作者简介:张雪峰(1966-),男,辽宁辽阳人,东北大学副教授.
  • 基金资助:
    国家重点研发计划项目(2020YFB1710003).

Automatic Lane Change Decision Model Based on Dueling Double Deep Q-network

ZHANG Xue-feng, WANG Zhao-yi   

  1. School of Sciences, Northeastern University, Shenyang 110819, China.
  • Published:2023-10-27
  • Contact: ZHANG Xue-feng
  • About author:-
  • Supported by:
    -

摘要: 汽车自动变道需要在保证不发生碰撞的情况下,以尽可能快的速度行驶,规则性地控制不仅对意外情况不具有鲁棒性,而且不能对间隔车道的情况做出反应.针对这些问题,提出了一种基于双决斗深度Q网络(dueling double deep Q-network, D3QN)强化学习模型的自动换道决策模型,该算法对车联网反馈的环境车信息处理之后,通过策略得到动作,执行动作后根据奖励函数对神经网络进行训练,最后通过训练的网络以及强化学习来实现自动换道策略.利用Python搭建的三车道环境以及车辆仿真软件CarMaker进行仿真实验,得到了很好的控制效果,结果验证了本文算法的可行性和有效性.

关键词: 车道变换;自动驾驶;强化学习;深度学习;深度强化学习

Abstract: Automatic lane change of vehicles requires driving at the fastest possible speed while ensuring no collision situations. However, regular control is not robust enough to handle unexpected situations or respond to lane separation. To solve these problems, an automatic lane change decision model based on dueling double deep Q-network(D3QN) reinforcement learning model is proposed. The algorithm processes the environmental vehicle information fed back by the internet of vehicles, and then obtains actions through strategies. After the actions are executed, the neural network is trained according to given reward function, and finally the automatic lane change strategy is realized through the trained network and reinforcement learning. The three-lane environment built by Python and the vehicle simulation software CarMaker are used to carry out simulation experiments. The results show that the algorithm proposed has a good control effect, making it feasible and effective.

Key words: lane change; driverless vehicles; reinforcement learning; deep learning; deep reinforcement leaning

中图分类号: