Robust 3d model reconstruction based on continuous point cloud for autonomous vehicles

Hongwei Gao, Jiahui Yu, Jian Sun, Wei Yang, Yueqiu Jiang, Lei Zhu, Zhaojie Ju

Research output: Contribution to journalArticlepeer-review

Abstract

Continuous point cloud stitching can reconstruct a 3D model and play an essential role in autonomous vehicles. However, most existing methods are based on binocular stereo vision, which increases space and material costs, and these systems also achieve poor matching accuracies and speeds. In this paper, a novel point cloud stitching method based on the monocular vision system is proposed to solve these problems. First, the calibration and parameter acquisition based on monocular vision are presented. Next, the region-growing algorithm in sparse matching and dense matching is redesigned to improve the matching density. Finally, an Iterative Closest Point (ICP)-based splicing method is proposed for monocular zoom stereo vision. The point cloud data are spliced by introducing the rotation matrix and translation factor obtained in the matching process. In the experiments, the proposed method is evaluated on two datasets: self-collected and public datasets. The results show that the proposed method achieves a higher matching accuracy than the binocular-based systems, and it also outperforms other recent approaches. In addition, the 3D model generated using this method has a wider viewing angle, a more precise outline, and more distinct layers than the state-of-the-art algorithms.

Original languageEnglish
Pages (from-to)3169-3186
Number of pages18
JournalSensors and Materials
Volume33
Issue number9
DOIs
Publication statusPublished - 16 Sept 2021

Keywords

  • dense 3d point cloud
  • match optimization
  • monocular zoom stereo vision
  • region growing

Fingerprint

Dive into the research topics of 'Robust 3d model reconstruction based on continuous point cloud for autonomous vehicles'. Together they form a unique fingerprint.

Cite this