Skip to main navigation Skip to search Skip to main content

Geometric-driven structure recovery from a single omnidirectional image based on planar depth map learning

  • Ming Meng
  • , Likai Xiao
  • , Zhong Zhou*
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Scene structure recovery is a crucial process for assisting scene reconstruction and understanding by extracting vital scene structure information and has been widely used in smart city, VR/AR and intelligent robot navigation. Omnidirectional image with a 180° or 360° field of view (FoV) provides greater visual information, making them a significant research topic in computer vision and computational photography. However, indoor omnidirectional scene structure recovery faces challenges like severe occlusion of critical local regions caused by cluttered objects and large nonlinear distortion. To address these limitations, we propose a geometric-driven indoor structure recovery method based on planar depth map learning, aiming to mitigate the interference caused by occlusions in critical local regions. Our approach involves designing an OmniPDMNet, a planar depth map learning network for omnidirectional image, which uses upsampling and a feature-based objective loss function to accurately estimate high-precision planar depth map. Furthermore, we leverage prior knowledge from the omnidirectional depth map and introduce it into the structure recovery network (OmniSRNet) to extract global structural features and enhance the overall quality of structure recovery. We also introduce a distortion-aware module for feature extraction from omnidirectional image, allowing adaptability to omnidirectional geometric distortion and enhancing the performance of both OmniPDMNet and OmniSRNet. Finally, we conduct extensive experiments on omnidirectional dataset focusing on planar depth and structure recovery demonstrate that our proposed method achieves state-of-the-art performance.

Original languageEnglish
Pages (from-to)24407-24433
Number of pages27
JournalNeural Computing and Applications
Volume35
Issue number34
DOIs
StatePublished - Dec 2023

UN SDGs

This output contributes to the following UN Sustainable Development Goals (SDGs)

  1. SDG 11 - Sustainable Cities and Communities
    SDG 11 Sustainable Cities and Communities

Keywords

  • Distortion-aware learning
  • Omnidirectional image
  • Planar depth map learning
  • Structure recovery

Fingerprint

Dive into the research topics of 'Geometric-driven structure recovery from a single omnidirectional image based on planar depth map learning'. Together they form a unique fingerprint.

Cite this