TY - JOUR
T1 - Self-Evolutionary Reinforcement Learning for Autonomous and Robust Multiimpulse Orbital Transfers
AU - Ren, He
AU - Gui, Haichao
AU - Zhong, Rui
N1 - Publisher Copyright:
© 1965-2011 IEEE.
PY - 2025
Y1 - 2025
N2 - Autonomous, robust, and fuel-efficient trajectory optimization is critical for future spacecraft missions, enabling real-time orbital transfers under dynamic mission requirements. N-impulse transfers between noncoplanar Earth orbits present significant computational challenges for traditional iterative methods, which rely on terminal conditions and lack adaptability to real-time state changes. To address this, we propose a self-evolutionary reinforcement learning (SERL) algorithm that integrates the evolutionary principles of competition and mutual learning from particle swarm optimization into the reinforcement learning framework. This integration enhances the algorithm's generalization and exploration efficiency. SERL divides the N-impulse transfer process into N-1 intermediate transition orbits, allowing the spacecraft to select intermediate orbits based on its observed states. Compared to traditional methods, SERL offers two key advantages. First, it generates transfer trajectories in real time once the target orbits are defined, enabling onboard deployment and improving spacecraft autonomy. Second, it bases decisions on immediate state observations rather than predefined terminal conditions, allowing it to handle complex tasks, such as dynamically changing target orbits during mission execution. Numerical simulations validate SERL's ability to improve computational efficiency and adaptability for challenging orbital transfer scenarios.
AB - Autonomous, robust, and fuel-efficient trajectory optimization is critical for future spacecraft missions, enabling real-time orbital transfers under dynamic mission requirements. N-impulse transfers between noncoplanar Earth orbits present significant computational challenges for traditional iterative methods, which rely on terminal conditions and lack adaptability to real-time state changes. To address this, we propose a self-evolutionary reinforcement learning (SERL) algorithm that integrates the evolutionary principles of competition and mutual learning from particle swarm optimization into the reinforcement learning framework. This integration enhances the algorithm's generalization and exploration efficiency. SERL divides the N-impulse transfer process into N-1 intermediate transition orbits, allowing the spacecraft to select intermediate orbits based on its observed states. Compared to traditional methods, SERL offers two key advantages. First, it generates transfer trajectories in real time once the target orbits are defined, enabling onboard deployment and improving spacecraft autonomy. Second, it bases decisions on immediate state observations rather than predefined terminal conditions, allowing it to handle complex tasks, such as dynamically changing target orbits during mission execution. Numerical simulations validate SERL's ability to improve computational efficiency and adaptability for challenging orbital transfer scenarios.
KW - Multiimpulse orbit transfer
KW - onboard optimization method
KW - self-evolutionary reinforcement learning (SERL)
UR - https://www.scopus.com/pages/publications/105013181105
U2 - 10.1109/TAES.2025.3596223
DO - 10.1109/TAES.2025.3596223
M3 - 文章
AN - SCOPUS:105013181105
SN - 0018-9251
VL - 61
SP - 16634
EP - 16646
JO - IEEE Transactions on Aerospace and Electronic Systems
JF - IEEE Transactions on Aerospace and Electronic Systems
IS - 6
ER -