Discover the SciOpen Platform and Achieve Your Research Goals with Ease.
Search articles, authors, keywords, DOl and etc.
Despite significant progress in autonomous vehicles (AVs), the development of driving policies that ensure both the safety of AVs and traffic flow efficiency has not yet been fully explored. In this paper, we propose an enhanced human-in-the-loop reinforcement learning method, termed the Human as AI mentor-based deep reinforcement learning (HAIM-DRL) framework, which facilitates safe and efficient autonomous driving in mixed traffic platoon. Drawing inspiration from the human learning process, we first introduce an innovative learning paradigm that effectively injects human intelligence into AI, termed Human as AI mentor (HAIM). In this paradigm, the human expert serves as a mentor to the AI agent. While allowing the agent to sufficiently explore uncertain environments, the human expert can take control in dangerous situations and demonstrate correct actions to avoid potential accidents. On the other hand, the agent could be guided to minimize traffic flow disturbance, thereby optimizing traffic flow efficiency. In detail, HAIM-DRL leverages data collected from free exploration and partial human demonstrations as its two training sources. Remarkably, we circumvent the intricate process of manually designing reward functions; instead, we directly derive proxy state-action values from partial human demonstrations to guide the agents’ policy learning. Additionally, we employ a minimal intervention technique to reduce the human mentor’s cognitive load. Comparative results show that HAIM-DRL outperforms traditional methods in driving safety, sampling efficiency, mitigation of traffic flow disturbance, and generalizability to unseen traffic scenarios.
Andreotti, E.Selpi, Boyraz, P., 2023. Potential impact of autonomous vehicles in mixed traffic from simulation using real traffic flow. J. Int. Con. Veh. 6, 1–15.
Ansariyar, A., Tahmasebi, M., 2022. Investigating the effects of gradual deployment of market penetration rates (MPR) of connected vehicles on delay time and fuel consumption. J. Intell. Connect. Veh. 5, 188–198.
Aradi, S., 2022. Survey of deep reinforcement learning for motion planning of autonomous vehicles. IEEE Trans. Intell. Transport. Syst. 23, 740–759.
Bando, M., Hasebe, K., Nakayama, A., Shibata, A., Sugiyama, Y., 1995. Dynamical model of traffic congestion and numerical simulation. Phys. Rev. E. 51, 1035–1042.
Booth, S., Knox, W.B., Shah, J., Niekum, S., Stone, P., Allievi, A., 2023. The perils of trial-and-error reward design: Misdesign through overfitting and invalid task specifications. Proc. AAAI Conf. Artif. Intell. 37, 5920–5929.
Chen, C., Wang, J., Xu, Q., Wang, J., Li, K., 2021a. Mixed platoon control of automated and human-driven vehicles at a signalized intersection: Dynamical analysis and optimal control. Transp. Res. Part C Emerg. Technol. 127, 103138.
Chen, D., Srivastava, A., Ahn, S., Li, T., 2020. Traffic dynamics under speed disturbance in mixed traffic with automated and non-automated vehicles. Transp. Res. Part C Emerg. Technol. 113, 293–313.
Chen, S., Dong, J., Ha, P., Li, Y., Labi, S., 2021b. Graph neural network and reinforcement learning for multi-agent cooperative control of connected autonomous vehicles. Comput. Aided. Civ. Infrastruct. Eng. 36, 838–857.
Chen, S., Zong, S., Chen, T., Huang, Z., Chen, Y., Labi, S., 2023. A taxonomy for autonomous vehicles considering ambient road infrastructure. Sustainability 15, 11258.
Di, X., Shi, R., 2021. A survey on autonomous vehicle control in the era of mixed-autonomy: From physics-based to AI-guided driving policy learning. Transp. Res. Part C Emerg. Technol. 125, 103008.
Ding, H., Li, W., Xu, N., Zhang, J., 2022. An enhanced eco-driving strategy based on reinforcement learning for connected electric vehicles: Cooperative velocity and lane-changing control. J. Intell. Connect. Veh. 5, 316–332.
Dong, J., Chen, S., Li, Y., Du, R., Steinfeld, A., Labi, S., 2021. Space-weighted information fusion using deep reinforcement learning: The context of tactical control of lane-changing autonomous vehicles and connectivity range assessment. Transp. Res. Part C Emerg. Technol. 128, 103192.
Dong, J., Chen, S., Miralinaghi, M., Chen, T., Labi, S., 2022. Development and testing of an image transformer for explainable autonomous driving systems. J. Intell. Connect. Veh. 5, 235–249.
Dong, J., Chen, S., Miralinaghi, M., Chen, T., Li, P., Labi, S., 2023. Why did the AI make that decision? Towards an explainable artificial intelligence (XAI) for autonomous driving systems. Transp. Res. Part C Emerg. Technol. 156, 104358.
Du, R., Chen, S., Dong, J., Chen, T., Fu, X., Labi, S., 2023. Dynamic urban traffic rerouting with fog-cloud reinforcement learning. Computer Aided Civil. Eng. 1–21.
Feng, S., Sun, H., Yan, X., Zhu, H., Zou, Z., Shen, S., et al., 2023. Dense reinforcement learning for safety validation of autonomous vehicles. Nature 615, 620–627.
Ha, P., Chen, S., Dong, J., Labi, S., 2023. Leveraging vehicle connectivity and autonomy for highway bottleneck congestion mitigation using reinforcement learning. Transp. A Transp. Sci. 1–26.
Han, Y., Wang, M., Leclercq, L., 2023. Leveraging reinforcement learning for dynamic traffic control: A survey and challenges for field implementation. Commun. Transport. Res. 3, 100104.
Han, Y., Wang, M., Li, L., Roncoli, C., Gao, J., Liu, P., 2022. A physics-informed reinforcement learning-based strategy for local and coordinated ramp metering. Transp. Res. Part C Emerg. Technol. 137, 103584.
Huang, L., Guo, H., Zhang, R., Wang, H., Wu, J., 2018. Capturing drivers’ lane changing behaviors on operational level by data driven methods. IEEE Access 6, 57497–57506.
Huang, Z., Liu, H., Wu, J., Lv, C., 2023b. Conditional predictive behavior planning with inverse reinforcement learning for human-like autonomous driving. IEEE Trans. Intell. Transp. Syst. 24, 7244–7258.
Jiang, L., Xie, Y., Evans, N.G., Wen, X., Li, T., Chen, D., 2022. Reinforcement Learning based cooperative longitudinal control for reducing traffic oscillations and improving platoon stability. Transp. Res. Part C Emerg. Technol. 141, 103744.
Kesting, A., Treiber, M., Helbing, D., 2007. General lane-changing model MOBIL for car-following models. Transp. Res. Rec. 1999, 86–94.
Kiran, B.R., Sobh, I., Talpaert, V., Mannion, P., Al Sallab, A.A., Yogamani, S., et al., 2022. Deep reinforcement learning for autonomous driving: A survey. IEEE Trans. Intell. Transport. Syst. 23, 4909–4926.
Knox, W.B., Allievi, A., Banzhaf, H., Schmitt, F., Stone, P., 2023. Reward (Mis)design for autonomous driving. Artif. Intell. 316, 103829.
Krishna, R., Lee, D., Li, F.F., Bernstein, M.S., 2022. Socially situated artificial intelligence enables learning from human interaction. Proc. Natl. Acad. Sci. USA 119, 39.
Le Mero, L., Yi, D., Dianati, M., Mouzakitis, A., 2022. A survey on imitation learning techniques for end-to-end autonomous vehicles. IEEE Trans. Intell. Transport. Syst. 23, 14128–14147.
Li, J., Wu, P., Li, R., Pian, Y., Huang, Z., Xu, L., et al., 2022a. ST-CRMF: Compensated residual matrix factorization with spatial-temporal regularization for graph-based time series forecasting. Sensors 22, 5877.
Li, Q., Peng, Z., Feng, L., Zhang, Q., Xue, Z., Zhou, B., 2023. MetaDrive: Composing diverse driving scenarios for generalizable reinforcement learning. IEEE Trans. Pattern Anal. Mach. Intell. 1–14.
Lin, Y., Huang, Z., Wu, P., Xu, L., 2021. RSSI positioning method of vehicles in tunnels based on semi-supervised extreme learning machine. J. Traffic Transp. Eng. 243–255.
Liu, C., Sheng, Z., Chen, S., Shi, H., Ran, B., 2023. Longitudinal control of connected and automated vehicles among signalized intersections in mixed traffic flow with deep reinforcement learning approach. Phys. A Stat. Mech. Appl. 629, 129189.
Ma, H., An, B., Li, L., Zhou, Z., Qu, X., Ran, B., 2023. Anisotropy safety potential field model under intelligent and connected vehicle environment and its application in car-following modeling. J. Int. Con. Veh. 6, 79–90.
Mahdinia, I., Mohammadnazar, A., Arvin, R., Khattak, A.J., 2021. Integration of automated vehicles in mixed traffic: Evaluating changes in performance of following human-driven vehicles. Accid. Anal. Prev. 152, 106006.
Mohammadian, S., Zheng, Z., Haque, M.M., Bhaskar, A., 2023. Continuum modeling of freeway traffic flows: State-of-the-art, challenges and future directions in the era of connected and automated vehicles. Commun. Transport. Res. 3, 100107.
Muhammad, K., Ullah, A., Lloret, J., Del Ser, J., de Albuquerque, V.H.C., 2021. Deep learning for safe autonomous driving: Current challenges and future directions. IEEE Trans. Intell. Transp. Syst. 22, 4316–4336.
Olovsson, T., Svensson, T., Wu, J., 2022. Future connected vehicles: Communications demands, privacy and cyber-security. Commun. Transp. Res. 2, 100056.
Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C., Mishkin, P., et al., 2022. Training language models to follow instructions with human feedback. Adv. Neurl inf. Process. Syst. 35, 27730–27744.
Qu, X., Lin, H., Liu, Y., 2023. Envisioning the future of transportation: Inspiration of ChatGPT and large models. Commun. Transport. Res. 3, 100103.
Sharma, A., Zheng, Z., Kim, J., Bhaskar, A., Haque, M.M., 2021. Assessing traffic disturbance, efficiency, and safety of the mixed traffic flow of connected vehicles and traditional vehicles by considering human factors. Transp. Res. Part C Emerg. Technol. 124, 102934.
Shi, H., Chen, D., Zheng, N., Wang, X., Zhou, Y., Ran, B., 2023. A deep reinforcement learning based distributed control strategy for connected automated vehicles in mixed traffic platoon. Transp. Res. Part C Emerg. Technol. 148, 104019.
Shi, X., Wang, Z., Li, X., Pei, M., 2021. The effect of ride experience on changing opinions toward autonomous vehicle safety. Commun. Transport. Res. 1, 100003.
Silver, D., Hubert, T., Schrittwieser, J., Antonoglou, I., Lai, M., Guez, A., et al., 2018. A general reinforcement learning algorithm that Masters chess, shogi, and Go through self-play. Science 362, 1140–1144.
Stern, R.E., Chen, Y., Churchill, M., Wu, F., Delle Monache, M.L., Piccoli, B., et al., 2019. Quantifying air quality benefits resulting from few autonomous vehicles stabilizing traffic. Transp. Res. Part D Transp. Environ. 67, 351–365.
Treiber, M., Hennecke, A., Helbing, D., 2000. Congested traffic states in empirical observations and microscopic simulations. Phys. Rev. E. 62, 1805–1824.
Wang, W., Zhang, Y., Gao, J., Jiang, Y., Yang, Y., Zheng, Z., et al., 2023. GOPS: A general optimal control problem solver for autonomous driving and industrial control applications. Commun. Transport. Res. 3, 100096.
Wu, J., Huang, Z., Hu, Z., Lv, C., 2023. Toward human-in-the-loop AI: Enhancing deep reinforcement learning via real-time human guidance for autonomous driving. Engineering 21, 75–91.
Wu, J., Qu, X., 2022. Intersection control with connected and automated vehicles: A review. J. Intell. Connect. Veh. 5, 260–269.
Wu, P., Huang, Z., Pian, Y., Xu, L., Li, J., Chen, K., 2020. A combined deep learning method with attention-based LSTM model for short-term traffic speed forecasting. J. Adv. Transp. 2020, 1–15.
Wu, Y., Chen, H., Zhu, F., 2019. DCL-AIM: Decentralized coordination learning of autonomous intersection management for connected and automated vehicles. Transp. Res. Part C Emerg. Technol. 103, 246–260.
Xu, M., Di, Y., Ding, H., Zhu, Z., Chen, X., Yang, H., 2023. AGNP: Network-wide short-term probabilistic traffic speed prediction and imputation. Commun. Transport. Res. 3, 100099.
Yue, L., Abdel-Aty, M., Wang, Z., 2022. Effects of connected and autonomous vehicle merging behavior on mainline human-driven vehicle. J. Intell. Connect. Veh. 5, 36–45.
Zhou, H., Zhou, A., Li, T., Chen, D., Peeta, S., Laval, J., 2022. Congestion-mitigating MPC design for adaptive cruise control based on Newell’s car following model: History outperforms prediction. Transp. Res. Part C Emerg. Technol. 142, 103801.
Zhu, J., Easa, S., Gao, K., 2022. Merging control strategies of connected and autonomous vehicles at freeway on-ramps: A comprehensive review. J. Intell. Connect. Veh. 5, 99–111.
Zhu, M., Wang, Y., Pu, Z., Hu, J., Wang, X., Ke, R., 2020. Safe, efficient, and comfortable velocity control based on reinforcement learning for autonomous driving. Transp. Res. Part C Emerg. Technol. 117, 102662.
Zhu, Z., Zhao, H., 2022. A survey of deep RL and IL for autonomous driving policy learning. IEEE Trans. Intell. Transport. Syst. 23, 14043–14065.
Zhuo, J., Zhu, F., 2023. Evaluation of platooning configurations for connected and automated vehicles at an isolated round about in a mixed traffic environment. J. Int. Con. Veh. 6, 136–148.
This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).