Indoor semantic scene understanding using 2D-3D fusion

Document Type

Conference Proceeding

Publication Title

2021 Digital Image Computing: Techniques and Applications (DICTA)

Publisher

IEEE

School

School of Science / Graduate Research

RAS ID

43592

Comments

Gopinathan, M., Truong, G., & Abu-Khalaf, J. (2021, November-December). Indoor semantic scene understanding using 2D-3D fusion [Paper presentation]. 2021 Digital Image Computing: Techniques and Applications (DICTA), Gold Coast, Australia.

https://doi.org/10.1109/DICTA52665.2021.9647182

Abstract

Seamless Human-Robot Interaction is the ultimate goal of developing service robotic systems. For this, the robotic agents have to understand their surroundings to better complete a given task. Semantic scene understanding allows a robotic agent to extract semantic knowledge about the objects in the environment. In this work, we present a semantic scene understanding pipeline that fuses 2D and 3D detection branches to generate a semantic map of the environment. The 2D mask proposals from state-of-the-art 2D detectors are inverse-projected to the 3D space and combined with 3D detections from point segmentation networks. Unlike previous works that were evaluated on collected datasets, we test our pipeline on an active photo-realistic robotic environment BenchBot. Our novelty includes the rectification of 3D proposals using projected 2D detections and modality fusion based on object size. This work is done as part of the Robotic Vision Scene Understanding Challenge (RVSU). The performance evaluation demonstrates that our pipeline has improved on baseline methods without significant computational bottleneck.

DOI

10.1109/DICTA52665.2021.9647182

Access Rights

subscription content

Share

 
COinS