Future missions to the Moon and beyond are likely to involve low-thrust propulsion technologies due to their propellant efficiency. However, these still present a difficult trajectory design problem, owing to the near continuous thrust, lack of control authority and chaotic dynamics. Lyapunov control laws can generate sub-optimal trajectories for such missions with minimal computational cost and are suitable for feasibility studies and as initial guesses for optimisation methods. In this work a Reinforced Lyapunov Controller is used to design optimal low-thrust transfers from geostationary transfer orbit towards lunar polar orbit. Within the reinforcement learning (RL) framework, a dual-actor network setup is used, one in each of the Earth- and Moon-centred inertial frames respectively. A key contribution of this paper is the demonstration of a forwards propagated trajectory, removing the need to define a patch point a priori. This is enabled by an adaptive patch distance and extensive initial geometry exploration during the RL training. Results for both time- and fuel-optimal transfers are presented, along with a Monte Carlo analysis of the robustness to disturbances for such transfers. Phasing is introduced where necessary to aid rendezvous with the Moon. The results demonstrate the potential for such techniques to provide a basis for the design and guidance of low-thrust lunar transfers.
Tsuda, Y., Yoshikawa, M., Abe, M., Minamino, H., Nakazawa, S. System design of the Hayabusa 2—Asteroid sample return mission to 1999 JU3. Acta Astronautica, 2013, 91: 356–362.
Benkhoff, J., van Casteren, J., Hayakawa, H., Fujimoto, M., Laakso, H., Novara, M., Ferri, P., Middleton, H. R., Ziethe, R. BepiColombo—Comprehensive exploration of Mercury: Mission overview and science goals. Planetary and Space Science, 2010, 58(1–2): 2–20.
Gao, Y., Li, X. Optimization of low-thrust many-revolution transfers and Lyapunov-based guidance. Acta Astronautica, 2010, 66(1–2): 117–129.
Shannon, J. L., Ozimek, M. T., Atchison, J. A., Christine, M. Q-law aided direct trajectory optimization for the high-fidelity, many-revolution low-thrust orbit transfer problem. Advances in the Astronautical Sciences, 2019, 168: 781–800.
Jagannatha, B. B., Bouvier, J. B. H., Ho, K. Preliminary design of low-energy, low-thrust transfers to halo orbits using feedback control. Journal of Guidance, Control, and Dynamics, 2019, 42(2): 260–271.
Epenoy, R., Pérez-Palau, D. Lyapunov-based low-energy low-thrust transfers to the Moon. Acta Astronautica, 2019, 162: 87–97.
Holt, H., Armellin, R., Baresi, N., Hashida, Y., Turconi, A., Scorsoglio, A., Furfaro, R. Optimal Q-laws via reinforcement learning with guaranteed stability. Acta Astronautica, 2021, 187: 511–528.
Zavoli, A., Federici, L. Reinforcement learning for robust trajectory design of interplanetary missions. Journal of Guidance, Control, and Dynamics, 2021, 44(8): 1440–1453.
LaFarge, N. B., Miller, D., Howell, K. C., Linares, R. Autonomous closed-loop guidance using reinforcement learning in a low-thrust, multi-body dynamical environment. Acta Astronautica, 2021, 186: 1–23.
Sutton, R. S., Barto, A. G. Reinforcement Learning: An Introduction. MIT Press, 2018.
Shirobokov, M., Trofimov, S., Ovchinnikov, M. Survey of machine learning techniques in spacecraft control design. Acta Astronautica, 2021, 186: 87–97.
Izzo, D., Blazquez, E., Ferede, R., Origer, S., De Wagter, C., de Croon, G. C. Optimality principles in spacecraft neural guidance and control. Science Robotics, 2024, 9(91): eadi6421.
Malyuta, D., Yu, Y., Elango, P., Açıkmeşe, B. Advances in trajectory optimization for space vehicle control. Annual Reviews in Control, 2021, 52: 282–315.
Scorsoglio, A., Furfaro, R., Linares, R., Massari, M. Relative motion guidance for near-rectilinear lunar orbits with path constraints via actor-critic reinforcement learning. Advances in Space Research, 2023, 71(1): 316–335.
Petropoulos, A. E. Refinements to the Q-law for the low-thrust orbit transfers. Advances in the Astronautical Sciences, 2005, 120: 963–982.
Pontani, M., Pustorino, M. Nonlinear Earth orbit control using low-thrust propulsion. Acta Astronautica, 2021, 179: 296–310.
Grondman, I., Busoniu, L., Lopes, G. A. D., Babuska, R. A survey of actor-critic reinforcement learning: Standard and natural policy gradients. IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews), 2012, 42(6): 1291–1307.
DeMars, K. J., Jah, M. K. Probabilistic initial orbit determination using Gaussian mixture models. Journal of Guidance, Control, and Dynamics, 2013, 36(5): 1324–1335.
Betts, J. T., Erb, S. O. Optimal low thrust trajectories to the Moon. SIAM Journal on Applied Dynamical Systems, 2003, 2(2): 144–170.
Shannon, J. L., Ozimek, M. T., Atchison, J. A., Hartzell, C. M. Rapid design of high-fidelity low-thrust transfers to the Moon. Journal of Spacecraft and Rockets, 2022, 59(5): 1522–1535.
Narayanaswamy, S., Damaren, C. J. Equinoctial Lyapunov control law for low-thrust rendezvous. Journal of Guidance, Control, and Dynamics, 2023, 46(4): 781–795.