[1]

Liang C, Ghazel M, Ci Y, Zheng W. 2024. Analyzing rear-end collision risk relevant to autonomous vehicles by using a humanlike brake model. Journal of Transportation Engineering, Part A: Systems 150(7):04024031

doi: 10.1061/JTEPBS.TEENG-8250
[2]

Liang C, Ghazel M, Xie C, Zheng W, Chen W. 2024. A dynamic synchronous interactive functional validation approach for electric vehicles. IEEE Transactions on Intelligent Vehicles 00:1−14

doi: 10.1109/TIV.2024.3393559
[3]

Liang C, Ghazel M, Zheng W, Chen W. 2025. Dynamic cumulative human-like brake control modeling for autonomous vehicle collision analysis. IEEE Transactions on Vehicular Technology 74:3976−90

doi: 10.1109/TVT.2024.3497583
[4]

Ma Y, Liu Q, Fu J, Liufu K, Li Q. 2023. Collision-avoidance Lane change control method for enhancing safety for connected vehicle platoon in mixed traffic environment. Accident Analysis and Prevention 184:106999

doi: 10.1016/j.aap.2023.106999
[5]

Li G, Qiu Y, Yang Y, Li Z, Li S, et al. 2023. Lane change strategies for autonomous vehicles: a deep reinforcement learning approach based on transformer. IEEE Transactions on Intelligent Vehicles 8:2197−211

doi: 10.1109/TIV.2022.3227921
[6]

He X, Yang H, Hu Z, Lv C. 2023. Robust lane change decision making for autonomous vehicles: an observation adversarial reinforcement learning approach. IEEE Transactions on Intelligent Vehicles 8(1):184−93

doi: 10.1109/TIV.2022.3165178
[7]

Wang R, Yang T, Liang C, Wang M, Ci Y. 2025. Reliable autonomous driving environment perception: uncertainty quantification of semantic segmentation. Journal of Transportation Engineering, Part A: Systems 151(3):04024117

doi: 10.1061/jtepbs.teeng-8660
[8]

Mirchevska B, Pek C, Werling M, Althoff M, Boedecker J. 2018. High-level decision making for safe and reasonable autonomous lane changing using reinforcement learning. 21st International Conference on Intelligent Transportation Systems (ITSC), 4–7 November 2018, Maui, HI, USA. USA: IEEE. pp. 2156–62 DOI: 10.1109/ITSC.2018.8569448

[9]

Wang S, Yin X, Li P, Zhang M, Wang X. 2020. Trajectory tracking control for mobile robots using reinforcement learning and PID. Iranian Journal of Science and Technology, Transactions of Electrical Engineering 44:1059−68

doi: 10.1007/s40998-019-00286-4
[10]

Hoel CJ, Wolff K, Laine L. 2018. Automated speed and lane change decision making using deep reinforcement learning. 21st International Conference on Intelligent Transportation Systems (ITSC), 4–7 November 2018, Maui, HI, USA. USA: IEEE. pp. 2148–55 doi: 10.1109/ITSC.2018.8569568

[11]

Clemmons J, Jin Y. 2023. Reinforcement learning-based guidance of autonomous vehicles. 2023 24th International Symposium on Quality Electronic Design (ISQED), 5−7 April 2023, San Francisco, CA, USA. USA: IEEE. pp. 1−6 doi: 10.1109/ISQED57927.2023.10129362

[12]

Wang J, Zhang Q, Zhao D, Chen Y. 2019. Lane change decision-making through deep reinforcement learning with rule-based constraints. 2019 International Joint Conference on Neural Networks (IJCNN), 14−19 July 2019, Budapest, Hungary, 2019. USA: IEEE. pp. 1−6 doi: 10.1109/IJCNN.2019.8852110

[13]

Zhang X, Zhang J, Lin Y, Xie L. 2021. Research on Decision Model of Autonomous Vehicle Based on Deep Reinforcement Learning. 2021 IEEE 11th International Conference on Electronics Information and Emergency Communication (ICEIEC), 18−20 June 2021, Beijing, China. USA: IEEE. pp. 1−5 doi: 10.1109/ICEIEC51955.2021.9463817

[14]

Chen W, Xie G, Ji W, Fei R, Hei X. 2021. Decision making for overtaking of unmanned vehicle based on deep Q-learning. 2021 IEEE 10th Data Driven Control and Learning Systems Conference (DDCLS), 14−16 May 2021, Suzhou, China. USA: IEEE. pp. 350−53 doi: 10.1109/ddcls52934.2021.9455523

[15]

Sangjin K. 2021. Reinforcement learning based decision making for self driving & shared control between human driver and machine. Doctoral dissertation. Texas A&M University, USA

[16]

Wang, S, Zhang B, Liang Q, Wang X. 2024. Research on decision making of intelligent vehicle based on composite priority experience replay. International Journal of Intelligent Decision Technologies 18:599−612

doi: 10.3233/IDT-230271
[17]

Hessel M, Modayil J, Van Hasselt H, Schaul T, Ostrovski G, et al. 2018. Rainbow: combining improvements in deep reinforcement learning. Proceedings of the AAAI Conference on Artificial Intelligence 32(1):11796

doi: 10.1609/aaai.v32i1.11796
[18]

Zhang H, Xiong K, Bai J. 2018. Improved deep deterministic policy gradient algorithm based on prioritized sampling. Proceedings of 2018 Chinese Intelligent Systems Conference. Lecture Notes in Electrical Engineering. vol 528. Singapore: Springer. pp. 205−15 doi: 10.1007/978-981-13-2288-4_21

[19]

Yuan W, Li Y, Zhuang H, Wang C, Yang M. 2021. Prioritized Experience Replay-Based Deep Q Learning: Multiple-Reward Architecture for Highway Driving Decision Making. IEEE robotics and automation magazine 28(4):21−31

doi: 10.1109/MRA.2021.3115980
[20]

Yavas U, Kumbasar T, Ure NK. 2020. A new approach for tactical decision making in lane changing: sample efficient deep Q Learning with a safety feedback reward. 2020 IEEE Intelligent Vehicles Symposium (IV), 19 October 2020 − 13 November 2020, Las Vegas, NV, USA. USA: IEEE. pp. 1156−61 doi: 10.1109/iv47402.2020.9304594

[21]

Alzubaidi A, Al Sumaiti AS, Byon YJ, Al Hosani K. 2023. Emergency vehicle aware lane change decision model for autonomous vehicles using deep reinforcement learning. IEEE Access 11:27127−37

doi: 10.1109/ACCESS.2023.3253503
[22]

Hoel CJ, Wolff K, Laine L. 2020. Tactical decision-making in autonomous driving by reinforcement learning with uncertainty estimation. 2020 IEEE intelligent vehicles symposium (IV), 19 October 2020 − 13 November 2020, Las Vegas, NV, USA. USA: IEEE. pp. 1563−69 doi: 10.1109/iv47402.2020.9304614

[23]

Lu Z, Farhi N, Christoforou Z, Haddadou N. 2021. Imitation of real lane-change decisions using reinforcement learning. IFAC-PapersOnLine 54(2):203−9

doi: 10.1016/j.ifacol.2021.06.023
[24]

Erdmann J. 2015. SUMO's lane-changing model. Modeling Mobility with Open Data: 2nd SUMO Conference, Berlin, Germany, 2014. Cham, Switzerland: Springer International Publishing. pp. 105−23 doi: 10.1007/978-3-319-15024-6_7