MCGMapper: Light-Weight Incremental Structure from Motion and Visual Localization With Planar Markers and Camera Groups
Structure from Motion (SfM) and visual localization in indoor texture-less scenes and industrial scenarios present prevalent yet challenging research topics. Existing SfM methods designed for natural scenes typically yield low accuracy or map-building failures due to insufficient robust feature extr...
Saved in:
Main Authors | , , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
26.05.2024
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Structure from Motion (SfM) and visual localization in indoor texture-less
scenes and industrial scenarios present prevalent yet challenging research
topics. Existing SfM methods designed for natural scenes typically yield low
accuracy or map-building failures due to insufficient robust feature extraction
in such settings. Visual markers, with their artificially designed features,
can effectively address these issues. Nonetheless, existing marker-assisted SfM
methods encounter problems like slow running speed and difficulties in
convergence; and also, they are governed by the strong assumption of unique
marker size. In this paper, we propose a novel SfM framework that utilizes
planar markers and multiple cameras with known extrinsics to capture the
surrounding environment and reconstruct the marker map. In our algorithm, the
initial poses of markers and cameras are calculated with Perspective-n-Points
(PnP) in the front-end, while bundle adjustment methods customized for markers
and camera groups are designed in the back-end to optimize the 6-DOF pose
directly. Our algorithm facilitates the reconstruction of large scenes with
different marker sizes, and its accuracy and speed of map building are shown to
surpass existing methods. Our approach is suitable for a wide range of
scenarios, including laboratories, basements, warehouses, and other industrial
settings. Furthermore, we incorporate representative scenarios into simulations
and also supply our datasets with pose labels to address the scarcity of
quantitative ground-truth datasets in this research field. The datasets and
source code are available on GitHub. |
---|---|
DOI: | 10.48550/arxiv.2405.16599 |