TY - GEN
T1 - Unmanned Aerial Vehicle Autonomous Visual Landing through Visual Attention-Based Deep Reinforcement Learning
AU - Wang, Shaofan
AU - Li, Ke
AU - Chen, Jiaao
AU - Zhang, Tao
N1 - Publisher Copyright:
© 2023 Technical Committee on Control Theory, Chinese Association of Automation.
PY - 2023
Y1 - 2023
N2 - Unmanned aerial vehicle (UAV) autonomous landing is still an open and challenging issue. State-of-art work focused on the use of meticulously designed hand-crafted geometric features and complex sensor-data fusion to identify fiducial marker and guide a UAV. To address these issues, this study proposed a novel end-to-end control method based on deep reinforcement learning (DRL) that only requires low-resolution images of the environment ahead. A convolutional neural network (CNN) function approximator combined with visual attention blocks was adopted for direct frame-to-action prediction. The input frames were acquired from a low-cost monocular camera integrated with the UAV, without any other sensors. The combination of a dueling deep Q-network (DQN) with diverse visual attention modules was implemented and compared with the original dueling DQN. The simulation results showed that the UAV could autonomously land on a marker in a high-fidelity virtual simulation environment with rich scene transformation, regardless of initial relative positions. Moreover, the test of the proposed visual landing algorithm in multiple actual landing scenarios proved that the algorithm can accurately extract and localize the landing marker by using real images.
AB - Unmanned aerial vehicle (UAV) autonomous landing is still an open and challenging issue. State-of-art work focused on the use of meticulously designed hand-crafted geometric features and complex sensor-data fusion to identify fiducial marker and guide a UAV. To address these issues, this study proposed a novel end-to-end control method based on deep reinforcement learning (DRL) that only requires low-resolution images of the environment ahead. A convolutional neural network (CNN) function approximator combined with visual attention blocks was adopted for direct frame-to-action prediction. The input frames were acquired from a low-cost monocular camera integrated with the UAV, without any other sensors. The combination of a dueling deep Q-network (DQN) with diverse visual attention modules was implemented and compared with the original dueling DQN. The simulation results showed that the UAV could autonomously land on a marker in a high-fidelity virtual simulation environment with rich scene transformation, regardless of initial relative positions. Moreover, the test of the proposed visual landing algorithm in multiple actual landing scenarios proved that the algorithm can accurately extract and localize the landing marker by using real images.
KW - attention mechanism
KW - Autonomous visual landing
KW - deep reinforcement learning
KW - dueling deep Q-network
UR - https://www.scopus.com/pages/publications/85175534647
U2 - 10.23919/CCC58697.2023.10240825
DO - 10.23919/CCC58697.2023.10240825
M3 - 会议稿件
AN - SCOPUS:85175534647
T3 - Chinese Control Conference, CCC
SP - 4143
EP - 4148
BT - 2023 42nd Chinese Control Conference, CCC 2023
PB - IEEE Computer Society
T2 - 42nd Chinese Control Conference, CCC 2023
Y2 - 24 July 2023 through 26 July 2023
ER -