Identifying Reflected Images From Object Detector in Indoor Environment Utilizing Depth Information

We observed that mirror reflection severely degrades person detection performance in an indoor environment, which is an essential task for service robots. To address this problem, we propose a new real-time method to identify reflected virtual images in an indoor environment utilizing 3D depth infor...

Full description

Saved in:
Bibliographic Details
Published inIEEE robotics and automation letters Vol. 6; no. 2; pp. 635 - 642
Main Authors Park, Daehee, Park, Yong-Hwa
Format Journal Article
LanguageEnglish
Published Piscataway IEEE 01.04.2021
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:We observed that mirror reflection severely degrades person detection performance in an indoor environment, which is an essential task for service robots. To address this problem, we propose a new real-time method to identify reflected virtual images in an indoor environment utilizing 3D depth information. Images reflected by the mirror are similar to real objects, so it is a non-trivial task to differentiate them. Conventional object detectors, which do not deal with this problem, obviously recognize reflected images as real objects. The proposed method compares the geometric relationship between the 3D spatial information of the detected object and its surrounding environment where the object locates. It analyzes the layout of surrounding indoor space utilizing semantic segmentation and plane detection method. With the estimated layout of indoor space, detected object candidates are examined whether they are real or reflected images utilizing 3D depth information. To verify the proposed method, a large indoor dataset was newly acquired and examined in a dedicated Living-lab environment. The performance of the algorithm is verified by comparing conventional detectors with the proposed method in the acquired Living-lab dataset .
ISSN:2377-3766
2377-3766
DOI:10.1109/LRA.2020.3047796