AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
PDF (12.8 MB)
Collect
Submit Manuscript AI Chat Paper
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Research Article | Open Access

Multi-level objective control of AVs at a saturated signalized intersection with multi-agent deep reinforcement learning approach

Wenfeng Lin1Xiaowei Hu1( )Jian Wang2
School of Transportation Science and Engineering, Harbin Institute of Technology, Harbin 15000, China
School of Management, Harbin Institute of Technology, Harbin 15000, China
Show Author Information

Abstract

Reinforcement learning (RL) can free automated vehicles (AVs) from the car-following constraints and provide more possible explorations for mixed behavior. This study uses deep RL as AVs’ longitudinal control and designs a multi-level objectives framework for AVs’ trajectory decision-making based on multi-agent DRL. The saturated signalized intersection is taken as the research object to seek the upper limit of traffic efficiency and realize the specific target control. The simulation results demonstrate the convergence of the proposed framework in complex scenarios. When prioritizing throughputs as the primary objective and emissions as the secondary objective, both indicators exhibit a linear growth pattern with increasing market penetration rate (MPR). Compared with MPR is 0%, the throughputs can be increased by 69.2% when MPR is 100%. Compared with linear adaptive cruise control (LACC) under the same MPR, the emissions can also be reduced by up to 78.8%. Under the control of the fixed throughputs, compared with LACC, the emission benefits grow nearly linearly as MPR increases, it can reach 79.4% at 80% MPR. This study employs experimental results to analyze the behavioral changes of mixed flow and the mechanism of mixed autonomy to improve traffic efficiency. The proposed method is flexible and serves as a valuable tool for exploring and studying the behavior of mixed flow behavior and the patterns of mixed autonomy.

References

[1]
Abousleiman, R., Rawashdeh, O., 2015. Energy consumption model of an electric vehicle. In: 2015 IEEE Transportation Electrification Conference and Expo (ITEC), 1–5.
[2]

Aragon-Gómez, R., Clempner, J. B., 2020. Traffic-signal control reinforcement learning approach for continuous-time Markov games. Eng Appl Artif Intell, 89, 103415.

[3]

Aslani, M., Mesgari, M. S., Wiering, M., 2017. Adaptive traffic signal control with actor-critic methods in a real-world traffic network with different traffic disruption events. Transp Res Part C Emerg Technol, 85, 732–752.

[4]

Bhatnagar, S., Sutton, R. S., Ghavamzadeh, M., Lee, M., 2009. Natural actor-critic algorithms. Automatica, 45, 2471–2482.

[5]

Chen, C., Wang, J., Xu, Q., Wang, J., Li, K., 2021. Mixed platoon control of automated and human-driven vehicles at a signalized intersection: Dynamical analysis and optimal control. Transp Res Part C Emerg Technol, 127, 103138.

[6]

Dai, P., Liu, K., Zhuge, Q., Sha, E. H. M., Lee, V. C. S., Son, S. H., 2016. Quality-of-experience-oriented autonomous intersection control in vehicular networks. IEEE Trans Intell Transport Syst, 17, 1956–1967.

[7]

Ding, H., Li, W., Xu, N., Zhang, J., 2022. An enhanced eco-driving strategy based on reinforcement learning for connected electric vehicles: Cooperative velocity and lane-changing control. J Intell Connect Veh, 5, 316–332.

[8]

Fayazi, S. A., Vahidi, A., 2018. Mixed-integer linear programming for optimal scheduling of autonomous vehicle intersection crossing. IEEE Trans Intell Veh, 3, 287–299.

[9]

Feng, Y., Yu, C., Liu, H. X., 2018. Spatiotemporal intersection control in a connected and automated vehicle environment. Transp Res Part C Emerg Technol, 89, 364–383.

[10]

Guan, Y., Ren, Y., Li, S. E., Sun, Q., Luo, L., Li, K., 2020. Centralized cooperation for connected and automated vehicles at intersections by proximal policy optimization. IEEE Trans Veh Technol, 69, 12597–12608.

[11]

Han, X., Ma, R., Zhang, H. M., 2020. Energy-aware trajectory optimization of CAV platoons through a signalized intersection. Transp Res Part C Emerg Technol, 118, 102652.

[12]
He, Y., Liu, Y., Yang, L., Qu, X., 2023. Deep adaptive control: Deep reinforcement learning-based adaptive vehicle trajectory control algorithms for different risk levels. IEEE Trans Intell Veh. Http://doi.org/10.1109/TIV.2023.3303408.
[13]
Hoel, C. J., Driggs-Campbell, K., Wolff, K., Laine, L., Kochenderfer, M. J., 2019. Combining planning and deep reinforcement learning in tactical decision making for autonomous driving. https://arxiv.org/abs/1905.02680.pdf
[14]
Isele, D., Rahimi, R., Cosgun, A., Subramanian, K., Fujimura, K., 2017. Navigating occluded intersections with autonomous vehicles using deep reinforcement learning. https://arxiv.org/abs/1705.01196.pdf
[15]

Jiang, H., Hu, J., An, S., Wang, M., Park, B. B., 2017. Eco approaching at an isolated signalized intersection under partially connected and automated vehicles environment. Transp Res Part C Emerg Technol, 79, 290–307.

[16]

Jing, S., Hui, F., Zhao, X., Rios-Torres, J., Khattak, A. J., 2019. Cooperative game approach to optimal merging sequence and on-ramp merging control of connected and automated vehicles. IEEE Trans Intell Transport Syst, 20, 4234–4244.

[17]

Kesting, A., Treiber, M., Schönhof, M., Helbing, D., 2008. Adaptive cruise control design for active congestion avoidance. Transp Res Part C Emerg Technol, 16, 668–683.

[18]

Kiran, B. R., Sobh, I., Talpaert, V., Mannion, P., Al Sallab, A. A., Yogamani, S. et al., 2022. Deep reinforcement learning for autonomous driving: A survey. IEEE Trans Intell Transport Syst, 23, 4909–4926.

[19]
Kreidieh, A. R., Wu, C., Bayen, A. M., 2018. Dissipating stop-and-go waves in closed and open networks via deep reinforcement learning. In: 2018 21st International Conference on Intelligent Transportation Systems (ITSC), 1475–1480.
[20]
Liang, E., Liaw, R., Nishihara, R., Moritz, P., Fox, R., Gonzalez, J., et al., 2017. Ray rLLib: A composable and scalable reinforcement learning library. http://arxiv.org/abs/1712.09381
[21]

Li, G., Li, S., Li, S., Qin, Y., Cao, D., Qu, X. et al., 2020. Deep reinforcement learning enabled decision-making for autonomous driving at intersections. Automot Innov, 3, 374–385.

[22]

Liu, C., Lin, C. W., Shiraishi, S., Tomizuka, M., 2018. Distributed conflict resolution for connected autonomous vehicles. IEEE Trans Intell Veh, 3, 18–29.

[23]
Liu, Y., Wu, F., Liu, Z., Wang, K., Wang, F., Qu, X., 2023. Can language models be used for real-world urban-delivery route optimization? Innovation, 4, 100520.
[24]
Lopez, P.A., Behrisch, M., Bieker-Walz, L., Erdmann, J., Flötteröd, Y.P., Hilbrich, R., et al., 2018. Microscopic traffic simulation using SUMO. In: 2018 21st international conference on intelligent transportation systems (ITSC), 2575–2582.
[25]

Makantasis, K., Kontorinaki, M., Nikolos, I., 2020. Deep reinforcement-learning-based driving policy for autonomous road vehicles. IET Intell Transp Syst, 14, 13–24.

[26]

Morales Medina, A. I., van de Wouw, N., Nijmeijer, H., 2018. Cooperative intersection control based on virtual platooning. IEEE Trans Intell Transport Syst, 19, 1727–1740.

[27]

Meng, X., Cassandras, C. G., 2022. Eco-driving of autonomous vehicles for nonstop crossing of signalized intersections. IEEE Trans Automat Sci Eng, 19, 320–331.

[28]

Mushtaq, A., Haq, I. U., Imtiaz, M. U., Khan, A., Shafiq, O., 2021. Traffic flow management of autonomous vehicles using deep reinforcement learning and smart rerouting. IEEE Access, 9, 51005–51019.

[29]

Phan, T. T., Ngoduy, D., Le, L. B., 2020. Space distribution method for autonomous vehicles at a signalized multi-lane intersection. IEEE Trans Intell Transport Syst, 21, 5283–5294.

[30]
Schulman, J., Moritz, P., Levine, S., Jordan, M., Abbeel, P., 2015. High-dimensional continuous control using generalized advantage estimation. https://arxiv.org/abs/1506.02438.pdf
[31]
Schulman, J., Wolski, F., Dhariwal, P., Radford, A., Klimov, O., 2017. Proximal policy optimization algorithms. https://arxiv.org/abs/1707.06347.pdf
[32]
Silver, D., Lever, G., Heess, N., Degris, T., Wierstra, D., Riedmiller, M., 2014. Deterministic policy gradient algorithms. 31st Int Conf Mach Learn ICML 2014, 1, 605–619.
[33]

Stryszowski, M., Longo, S., Velenis, E., Forostovsky, G., 2021. A framework for self-enforced interaction between connected vehicles: Intersection negotiation. IEEE Trans Intell Transport Syst, 22, 6716–6725.

[34]

Treiber, M., Hennecke, A., Helbing, D., 2000. Congested traffic states in empirical observations and microscopic simulations. Phys Rev E Stat Phys Plasmas Fluids Relat Interdiscip Topics, 62, 1805–1824.

[35]

Vahidi, A., Sciarretta, A., 2018. Energy saving potentials of connected and automated vehicles. Transp Res Part C Emerg Technol, 95, 822–843.

[36]

Wang, K., Wang, Y., Du, H., Nam, K., 2021. Game-theory-inspired hierarchical distributed control strategy for cooperative intersection considering priority negotiation. IEEE Trans Veh Technol, 70, 6438–6449.

[37]
Wang, Z., Bapst, V., Heess, N., Mnih, V., Munos, R., Kavukcuoglu, K., et al., 2016. Sample efficient actor-critic with experience replay. http://arxiv.org/abs/1611.01224
[38]
Wu, C., Kreidieh, A., Parvate, K., Vinitsky, E. and Bayen, A. M., 2017. Flow: Architecture and benchmarking for reinforcement learning in traffic control. https://arxiv.org/abs/1710.05465
[39]

Wu, J., Qu, X., 2022. Intersection control with connected and automated vehicles: A review. J Intell Connect Veh, 5, 260–269.

[40]

Xiao, L., Gao, F., 2011. Practical string stability of platoon of adaptive cruise control vehicles. IEEE Trans Intell Transport Syst, 12, 1184–1194.

[41]

Ye, Y., Zhang, X., Sun, J., 2019. Automated vehicle’s behavior decision making using deep reinforcement learning and high-fidelity simulation environment. Transp Res Part C Emerg Technol, 107, 155–170.

[42]

Yu, H., Jiang, R., He, Z., Zheng, Z., Li, L., Liu, R. et al., 2021. Automated vehicle-involved traffic flow studies: A survey of assumptions, models, speculations, and perspectives. Transp Res Part C Emerg Technol, 127, 103101.

[43]

Zhang, Y., Zhou, Y., Lu, H., Fujita, H., 2021. Cooperative multi-agent actor-critic control of traffic network flow based on edge computing. Future Gener Comput Syst, 123, 128–141.

[44]

Zhang, Z., Yang, X. T., 2021. Analysis of highway performance under mixed connected and regular vehicle environment. J Intell Connect Veh, 4, 68–79.

[45]

Zhou, M., Yu, Y., Qu, X., 2020. Development of an efficient driving strategy for connected and automated vehicles at signalized intersections: A reinforcement learning approach. IEEE Trans Intell Transport Syst, 21, 433–443.

Journal of Intelligent and Connected Vehicles
Pages 250-263
Cite this article:
Lin W, Hu X, Wang J. Multi-level objective control of AVs at a saturated signalized intersection with multi-agent deep reinforcement learning approach. Journal of Intelligent and Connected Vehicles, 2023, 6(4): 250-263. https://doi.org/10.26599/JICV.2023.9210021

370

Views

15

Downloads

0

Crossref

1

Scopus

Altmetrics

Received: 15 June 2023
Revised: 18 September 2023
Accepted: 09 October 2023
Published: 30 December 2023
© The author(s) 2023.

This is an open access article under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/).

Return