摘要
Existing nonlinear guidance methods are difficult to reconcile performance optimality with stability assurance. This study proposes a concept of robust incremental learning for approximate optimal control of nonlinear terminal guidance problems. It transitions incrementally and stably from a traditional analytically formulated guidance law to an approximate optimal guidance policy. Specifically, we propose an incremental policy iteration algorithm, where a base guidance law is utilized to mitigate the initial instability and warm-start the learning process. Then, a robustness enhancement technique is proposed to theoretically guarantee the stability of learning process, where the guidance command is refined leveraging a virtual Lyapunov-based energy function. As a result, a robust and efficient learning method for nonlinear optimal guidance problems is developed. Simulation results for a specific impact-angle-constrained guidance problem verify advantages of the proposed method on efficiency, stability, and optimality.
| 源语言 | 英语 |
|---|---|
| 页(从-至) | 6041-6052 |
| 页数 | 12 |
| 期刊 | IEEE Transactions on Aerospace and Electronic Systems |
| 卷 | 61 |
| 期 | 3 |
| DOI | |
| 出版状态 | 已出版 - 2025 |
指纹
探究 'Robust Incremental Learning of Approximate Dynamic Programming for Nonlinear Optimal Guidance' 的科研主题。它们共同构成独一无二的指纹。引用此
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver