TY - GEN
T1 - Forward long-distance 3D reconstruction in rail transit scenarios based on occupancy networks
AU - Li, Hongbo
AU - Wang, Zhangyu
AU - Yang, Songyue
AU - Wang, Hao
AU - Liu, Wentao
AU - Yu, Guizhen
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - In rail transit autonomous driving scenarios, real-time three-dimensional (3D) reconstruction is crucial for understanding scenes and ensuring the safety of the driving environment. Urban rail trains, with their substantial weight and long braking distances, necessitate an extended forward perception range in 3D space. To address this challenge, this paper proposes a method for forward long-distance 3D scene reconstruction tailored for rail transit scenarios based on occupancy networks. Firstly, a 3D feature representation method using three mutually perpendicular spatial planes is proposed to mitigate the high computational complexity of spatial voxel features. Secondly, considering the characteristics of forward binocular vision in rail transit scenarios, we employ self-attention and cross-attention mechanisms to fuse features between different images. Thirdly, due to the limitations in projection distance of the Light Detection and Ranging (LiDAR) point cloud used for supervision, we introduce a method to generate long-distance dense ground truth during the training stage. By pioneering the application of occupancy networks in rail transit scenarios, this approach significantly extends the forward perception range of autonomous driving trains, achieving an impressive 77.79% mean Intersection over Union (mIoU) accuracy.
AB - In rail transit autonomous driving scenarios, real-time three-dimensional (3D) reconstruction is crucial for understanding scenes and ensuring the safety of the driving environment. Urban rail trains, with their substantial weight and long braking distances, necessitate an extended forward perception range in 3D space. To address this challenge, this paper proposes a method for forward long-distance 3D scene reconstruction tailored for rail transit scenarios based on occupancy networks. Firstly, a 3D feature representation method using three mutually perpendicular spatial planes is proposed to mitigate the high computational complexity of spatial voxel features. Secondly, considering the characteristics of forward binocular vision in rail transit scenarios, we employ self-attention and cross-attention mechanisms to fuse features between different images. Thirdly, due to the limitations in projection distance of the Light Detection and Ranging (LiDAR) point cloud used for supervision, we introduce a method to generate long-distance dense ground truth during the training stage. By pioneering the application of occupancy networks in rail transit scenarios, this approach significantly extends the forward perception range of autonomous driving trains, achieving an impressive 77.79% mean Intersection over Union (mIoU) accuracy.
KW - occupancy networks
KW - rail transit autonomous driving
KW - three-dimensional view conversion
UR - https://www.scopus.com/pages/publications/85215530334
U2 - 10.1109/INDIN58382.2024.10774340
DO - 10.1109/INDIN58382.2024.10774340
M3 - 会议稿件
AN - SCOPUS:85215530334
T3 - IEEE International Conference on Industrial Informatics (INDIN)
BT - Proceedings - 2024 IEEE 22nd International Conference on Industrial Informatics, INDIN 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 22nd IEEE International Conference on Industrial Informatics, INDIN 2024
Y2 - 18 August 2024 through 20 August 2024
ER -