Discover the SciOpen Platform and Achieve Your Research Goals with Ease.
Search articles, authors, keywords, DOl and etc.
Model-based reinforcement learning (RL) is anticipated to exhibit higher sample efficiency than model-free RL by utilizing a virtual environment model. However, obtaining sufficiently accurate representations of environmental dynamics is challenging because of uncertainties in complex systems and environments. An inaccurate environment model may degrade the sample efficiency and performance of model-based RL. Furthermore, while model-based RL can improve sample efficiency, it often still requires substantial training time to learn from scratch, potentially limiting its advantages over model-free approaches. To address these challenges, this paper introduces a knowledge-informed model-based residual reinforcement learning framework aimed at enhancing learning efficiency by infusing established expert knowledge into the learning process and avoiding the issue of beginning from zero. Our approach integrates traffic expert knowledge into a virtual environment model, employing the intelligent driver model (IDM) for basic dynamics and neural networks for residual dynamics, thus ensuring adaptability to complex scenarios. We propose a novel strategy that combines traditional control methods with residual RL, facilitating efficient learning and policy optimization without the need to learn from scratch. The proposed approach is applied to connected automated vehicle (CAV) trajectory control tasks for the dissipation of stop-and-go waves in mixed traffic flows. The experimental results demonstrate that our proposed approach enables the CAV agent to achieve superior performance in trajectory control compared with the baseline agents in terms of sample efficiency, traffic flow smoothness and traffic mobility.
Andreotti, E., Selpi, Boyraz, P., 2023. Potential impact of autonomous vehicles in mixed traffic from simulation using real traffic flow. J. Int. Con. Veh. 6, 1–15.
Chee, K.Y., Jiahao, T.Z., Hsieh, M.A., 2022. KNODE-MPC: a knowledge-based data-driven predictive control framework for aerial robots. IEEE Rob. Autom. Lett. 7, 2819–2826.
Chen, D., Hajidavalloo, M.R., Li, Z., Chen, K., Wang, Y., Jiang, L., et al., 2023a. Deep multi-agent reinforcement learning for highway on-ramp merging in mixed traffic. IEEE Trans. Intell. Transport. Syst. 24, 11623–11638.
Chen, S., Dong, J., Ha, P.Y.J., Li, Y., Labi, S., 2021. Graph neural network and reinforcement learning for multi-agent cooperative control of connected autonomous vehicles. Computer Aided Civil Eng. 36, 838–857.
Chen, S., Zong, S., Chen, T., Huang, Z., Chen, Y., Labi, S., 2023b. A taxonomy for autonomous vehicles considering ambient road infrastructure. Sustainability 15, 11258.
Chua, K., Calandra, R., McAllister, R., Levine, S., 2018. Deep reinforcement learning in a handful of trials using probabilistic dynamics models. Adv. Neural Inf. Process. Syst 31.
Ding, H., Li, W., Xu, N., Zhang, J., 2022. An enhanced eco-driving strategy based on reinforcement learning for connected electric vehicles: cooperative velocity and lane-changing control. J. Intell. Connect. Veh. 5, 316–332.
Dong, J., Chen, S., Miralinaghi, M., Chen, T., Labi, S., 2022. Development and testing of an image transformer for explainable autonomous driving systems. J. Intell. Connect. Veh. 5, 235–249.
Dong, J., Chen, S., Miralinaghi, M., Chen, T., Li, P., Labi, S., 2023. Why did the AI make that decision? Towards an explainable artificial intelligence (XAI) for autonomous driving systems. Transport. Res. C Emerg. Technol. 156, 104358.
Du, R., Chen, S., Dong, J., Chen, T., Fu, X., Labi, S., 2024. Dynamic urban traffic rerouting with fog-cloud reinforcement learning. Computer Aided Civil Eng. 39, 793–813.
Feng, S., Song, Z., Li, Z., Zhang, Y., Li, L., 2021. Robust platoon control in mixed traffic flow based on tube model predictive control. IEEE Trans. Intell. Veh. 6, 711–722.
Garriga, J.L., Soroush, M., 2010. Model predictive control tuning methods: a review. Ind. Eng. Chem. Res. 49, 3505–3515.
Gong, S., Du, L., 2018. Cooperative platoon control for a mixed traffic flow including human drive vehicles and connected and autonomous vehicles. Transp. Res. Part B Methodol. 116, 25–61.
Guo, J., Cheng, L., Wang, S., 2023. CoTV: cooperative control for traffic light signals and connected autonomous vehicles using deep reinforcement learning. IEEE Trans. Intell. Transport. Syst. 24, 10501–10512.
Ha, P.Y.J., Chen, S., Dong, J., Labi, S., 2023. Leveraging vehicle connectivity and autonomy for highway bottleneck congestion mitigation using reinforcement learning. Transp. A Transp. Sci. 1–26.
Han, Y., Wang, M., Leclercq, L., 2023. Leveraging reinforcement learning for dynamic traffic control: a survey and challenges for field implementation. Commun. Transp. Res. 3, 100104.
Han, Y., Wang, M., Li, L., Roncoli, C., Gao, J., Liu, P., 2022. A physics-informed reinforcement learning-based strategy for local and coordinated ramp metering. Transport. Res. C Emerg. Technol. 137, 103584.
Hornik, K., Stinchcombe, M., White, H., 1989. Multilayer feedforward networks are universal approximators. Neural Network. 2, 359–366.
Hou, J., Chen, G., Li, Z., He, W., Gu, S., Knoll, A., et al., 2024. Hybrid residual multiexpert reinforcement learning for spatial scheduling of high-density parking lots. IEEE Trans. Cybern. 54, 2771–2783.
Hou, X., Gan, M., Zhang, J., Zhao, S., Ji, Y., 2023. Vehicle ride comfort optimization in the post-braking phase using residual reinforcement learning. Adv. Eng. Inf. 58, 102198.
Huang, Z., Chen, S., Pian, Y., Sheng, Z., Ahn, S., Noyce, D.A., 2024a. Toward C-V2X enabled connected transportation system: RSU-based cooperative localization framework for autonomous vehicles. IEEE Trans. Intell. Transport. Syst. 1–15.
Huang, Z., Sheng, Z., Ma, C., Chen, S., 2024b. Human as AI mentor: enhanced human-in-the-loop reinforcement learning for safe and efficient autonomous driving. Commun. Transp. Res. 4, 100127.
Janner, M., Fu, J., Zhang, M., Levine, S., 2019. When to trust your model: model-based policy optimization. Adv. Neural Inf. Process. Syst. 32.
Kabzan, J., Hewing, L., Liniger, A., Zeilinger, M.N., 2019. Learning-based model predictive control for autonomous racing. IEEE Rob. Autom. Lett. 4, 3363–3370.
Karniadakis, G.E., Kevrekidis, I.G., Lu, L., Perdikaris, P., Wang, S., Yang, L., 2021. Physics-informed machine learning. Nat. Rev. Phys. 3, 422–440.
Kiran, B.R., Sobh, I., Talpaert, V., Mannion, P., Al Sallab, A.A., Yogamani, S., et al., 2022. Deep reinforcement learning for autonomous driving: a survey. IEEE Trans. Intell. Transport. Syst. 23, 4909–4926.
Lee, H., Kim, N., Cha, S.W., 2020. Model-based reinforcement learning for eco-driving control of electric vehicles. IEEE Access 8, 202886–202896.
Li, J., Yu, C., Shen, Z., Su, Z., Ma, W., 2023. A survey on urban traffic control under mixed traffic environment with connected automated vehicles. Transport. Res. C Emerg. Technol. 154, 104258.
Liao, H., Shen, H., Li, Z., Wang, C., Li, G., Bie, Y., et al., 2024. GPT-4 enhanced multimodal grounding for autonomous driving: leveraging cross-modal attention with large language models. Commun. Transp. Res. 4, 100116.
Lidstrom, K., Sjoberg, K., Holmberg, U., Andersson, J., Bergh, F., Bjade, M., et al., 2012. A modular CACC system integration and design. IEEE Trans. Intell. Transport. Syst.13, 1050–1061.
Liu, C., Sheng, Z., Chen, S., Shi, H., Ran, B., 2023a. Longitudinal control of connected and automated vehicles among signalized intersections in mixed traffic flow with deep reinforcement learning approach. Phys. Stat. Mech. Appl. 629, 129189.
Liu, C., Sheng, Z., Li, P., Chen, S., Luo, X., Ran, B., 2024. A distributed deep reinforcement learning-based longitudinal control strategy for connected automated vehicles combining attention mechanism. Transp. Lett. 1–17.
Liu, Y., Wu, F., Liu, Z., Wang, K., Wang, F., Qu, X., 2023b. Can language models be used for real-world urban-delivery route optimization? Innovation 4, 100520.
Milanes, V., Shladover, S.E., Spring, J., Nowakowski, C., Kawazoe, H., Nakamura, M., 2014. Cooperative adaptive cruise control in real traffic situations. IEEE Trans. Intell. Transport. Syst. 15, 296–305.
Mo, Z., Shi, R., Di, X., 2021. A physics-informed deep learning paradigm for car-following models. Transport. Res. C Emerg. Technol. 130, 103240.
Moerland, T.M., Broekens, J., Plaat, A., Jonker, C.M., 2023. Model-based reinforcement learning: a survey. FNT. Mach. Learn. 16, 1–118.
Newell, G.F., 2002. A simplified car-following theory: a lower order model. Transp. Res. Part B Methodol. 36, 195–205.
O'Connell, M., Shi, G., Shi, X., Azizzadenesheli, K., Anandkumar, A., Yue, Y., et al., 2022. Neural-Fly enables rapid learning for agile flight in strong winds. Sci. Robot. 7, 195–205.
Olovsson, T., Svensson, T., Wu, J., 2022. Future connected vehicles: communications demands, privacy and cyber-security. Commun. Transp. Res. 2, 100056.
Pan, T., Guo, R., Lam, W.H.K., Zhong, R., Wang, W., He, B., 2021. Integrated optimal control strategies for freeway traffic mixed with connected automated vehicles: a model-based reinforcement learning approach. Transport. Res. C Emerg. Technol. 123, 102987.
Peng, B., Keskin, M.F., Kulcsár, B., Wymeersch, H., 2021. Connected autonomous vehicles for improving mixed traffic efficiency in unsignalized intersections with deep reinforcement learning. Commun. Transp. Res. 1, 100017.
Qu, X., Lin, H., Liu, Y., 2023. Envisioning the future of transportation: inspiration of ChatGPT and large models. Commun. Transp. Res. 3, 100103.
Sheng, Z., Huang, Z., Chen, S., 2024a. Kinematics-aware multigraph attention network with residual learning for heterogeneous trajectory prediction. J. Intell. Connect. Veh. 7, 138–150.
Sheng, Z., Liu, L., Xue, S., Zhao, D., Jiang, M., Li, D., 2023. A cooperation-aware lane change method for automated vehicles. IEEE Trans. Intell. Transport. Syst. 24, 3236–3251.
Shi, H., Chen, D., Zheng, N., Wang, X., Zhou, Y., Ran, B., 2023. A deep reinforcement learning based distributed control strategy for connected automated vehicles in mixed traffic platoon. Transport. Res. C Emerg. Technol. 148, 104019.
Shi, R., Mo, Z., Huang, K., Di, X., Du, Q., 2022. A physics-informed deep learning paradigm for traffic state and fundamental diagram estimation. IEEE Trans. Intell. Transport. Syst. 23, 11688–11698.
Staessens, T., Lefebvre, T., Crevecoeur, G., 2022. Adaptive control of a mechatronic system using constrained residual reinforcement learning. IEEE Trans. Ind. Electron. 69, 10447–10456.
Stern, R.E., Cui, S., Delle Monache, M.L., Bhadani, R., Bunting, M., Churchill, M., et al., 2018. Dissipation of stop-and-go waves via control of autonomous vehicles: field experiments. Transport. Res. C Emerg. Technol. 89, 205–221.
Treiber, M., Hennecke, A., Helbing, D., 2000. Congested traffic states in empirical observations and microscopic simulations. Phys. Rev. E 62, 1805–1824.
Vahidi, A., Eskandarian, A., 2003. Research advances in intelligent collision avoidance and adaptive cruise control. IEEE Trans. Intell. Transport. Syst. 4, 143–153.
Wang, S., Li, Z., Wang, B., Li, M., 2024. Collision avoidance motion planning for connected and automated vehicle platoon merging and splitting with a hybrid automaton architecture. IEEE Trans. Intell. Transport. Syst. 25, 1445–1464.
Wang, W., Zhang, Y., Gao, J., Jiang, Y., Yang, Y., Zheng, Z., et al., 2023. GOPS: a general optimal control problem solver for autonomous driving and industrial control applications. Commun. Transp. Res. 3, 100096.
Wu, C., Kreidieh, A.R., Parvate, K., Vinitsky, E., Bayen, A.M., 2022. Flow: a modular learning framework for mixed autonomy traffic. IEEE Trans. Robot. 38, 1270–1286.
Wu, J., Huang, Z., Lv, C., 2023. Uncertainty-aware model-based reinforcement learning: methodology and application in autonomous driving. IEEE Trans. Intell. Veh. 8, 194–203.
Wu, J., Qu, X., 2022. Intersection control with connected and automated vehicles: a review. J. Intell. Connect. Veh. 5, 260–269.
Yang, J., Zhao, D., Lan, J., Xue, S., Zhao, W., Tian, D., et al., 2023. Eco-driving of general mixed platoons with CAVs and HDVs. IEEE Trans. Intell. Veh. 8, 1190–1203.
Yang, Z., Zheng, Z., Kim, J., Rakha, H., 2024. Eco-driving strategies using reinforcement learning for mixed traffic in the vicinity of signalized intersections. Transport. Res. C Emerg. Technol. 165, 104683.
Yu, M., Long, J., 2022. An eco-driving strategy for partially connected automated vehicles at a signalized intersection. IEEE Trans. Intell. Transport. Syst. 23, 15780–15793.
Yue, L., Abdel-Aty, M., Wang, Z., 2022. Effects of connected and autonomous vehicle merging behavior on mainline human-driven vehicle. J. Intell. Connect. Veh. 5, 36–45.
Zhang, R., Hou, J., Chen, G., Li, Z., Chen, J., Knoll, A., 2022. Residual policy learning facilitates efficient model-free autonomous racing. IEEE Rob. Autom. Lett. 7, 11625–11632.
Zheng, Y., Wang, J., Li, K., 2020. Smoothing traffic flow via control of autonomous vehicles. IEEE Internet Things J. 7, 3882–3896.
Zhou, D., Gayah, V., 2023. Improving deep reinforcement learning-based perimeter metering control methods with domain control knowledge. Transport. Res. Rec. 2677, 384–405.
Zhou, Y., Ahn, S., Wang, M., Hoogendoorn, S., 2020. Stabilizing mixed vehicular platoons with connected automated vehicles: an H-infinity approach. Transp. Res. Part B Methodol. 132, 152–170.
Zhu, J., Easa, S., Gao, K., 2022. Merging control strategies of connected and autonomous vehicles at freeway on-ramps: a comprehensive review. J. Intell. Connect. Veh. 5,99–111.
This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).