Object Detection Based on the Fusion of Sparse LiDAR Point Cloud and Dense Stereo Pseudo Point Cloud
3D object detection is critical for unmanned intelligent vehicles to perceive their environment. LiDAR and stereo cameras are commonly used sensors, but each has limitations. LiDAR provides high-precision 3D point cloud data, but sparse distant and small objects pose challenges. Stereo cameras gener...
Saved in:
Published in | 2024 4th International Conference on Neural Networks, Information and Communication (NNICE) pp. 860 - 863 |
---|---|
Main Authors | , , , |
Format | Conference Proceeding |
Language | English |
Published |
IEEE
19.01.2024
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | 3D object detection is critical for unmanned intelligent vehicles to perceive their environment. LiDAR and stereo cameras are commonly used sensors, but each has limitations. LiDAR provides high-precision 3D point cloud data, but sparse distant and small objects pose challenges. Stereo cameras generate dense pseudo point clouds, but lack accurate positioning. This article combines the strengths of dense stereo cameras and precise LiDAR to improve detection performance for distant and small objects. We propose two innovations: (1) a method that combines 2D object detection and threshold filtering to accurately identify small distant objects, and (2) a local Graph based Depth Correction (GDC) that introduces a mask operation to correct and fusion small and distant object areas, filter out invalid pseudo point clouds, and improve accuracy. Experimental results on the PV-RCNN network demonstrate improved detection accuracy for distant small objects, with increased performance levels across Easy, Moderate, and Hard categories for pedestrian small objects (15.88%, 16.86%, and 14.41% respectively). |
---|---|
DOI: | 10.1109/NNICE61279.2024.10498214 |