3D Cross-source point cloud registration (3DCSR) benchmark


It is worthy of noting that the accuracy of registration algorithms is heavily dependent on the closeness of point cloud data in scale, size and sparsity. These factors are in turn heavily dependent on the hardware used. As different applications start to relay more and more on multi-source data, the data fusion of these different kinds data is emerging as more and more of a problem. Cross-source point cloud registration has wide applications such as building construction, augmented reality, and driverless vehicles. For example, in construction, project information is in the form of the 3D CAD model and in the form of real-time LiDAR scans (potentially captured using different hardware). The fusion of this data is vital to enable contractors to track progress and evaluate construction quality. However, until now there were no data sets for cross-source point cloud registration which would contain plentiful 3D data from recent popular sensors.
This benchmark aims to close the gap and provides a labelled 3D cross-source point cloud pairs with recent three popular sensors, that are RGB camera, depth sensor and Lidar sensor. The benchmark is captured in indoor working environment, which contains the popular objects in the working space, such as chairs, desks, computers, lights, walls, flowers, cupboard and so on.

Cross-source challenges


Since the point clouds are captured from the different types of sensors, and different types of sensors contain different imaging mechanisms, the cross-source challenges in the registration problem are much more complicated than the same-source challenges. These challenges are summarized below:

  • Noise and outliers: The captured noise in the environment and the sensors are different at different acquisition times. Thus, captured point clouds of the same scene will almost always contain a different set of noise and outliers around the same 3D position.
  • Partial overlap: Captured point clouds from different viewpoints and at different times will only have partial overlap.
  • Density difference: Due to different imaging mechanisms (e.g. active Lidar and passive RGB camera) and different resolutions, the captured point clouds will usually be of different densities.
  • Scale variation: Since different imaging mechanisms may have different physical metrics, the captured point clouds may be of different scales.

Data structure


The dataset contains two folders: kinect_lidar and kinect_sfm. The ground truth transformations are labbled by one computer science expert and cross-check by two other experts.

  • Kinect_lidar: This folder contains 19 scene folders that stored data from Kinect and Lidar sensors . Each folder is captured from one scene and croped into many parts. Each part is a pair of point clouds. We provide the Kinect point cloud, Lidar point cloud and the ground truth transformation between these two point clouds.
  • Kinect_sfm: This folder contains 2 scene folders that stored data from Kinect and RGB sensors. Each folder is captured from one scene and croped into many parts. Each part is a pair of point clouds. The RGB images have already constructed into a point cloud by using structure from motion software (VSFM). Finnaly, we provide the Kinect point cloud, SFM point cloud and the ground truth transformation between these two point clouds.

Research Paper


  • Download the paper that describes the cross-source point cloud registration benchmark dataset.


      @inproceedings{huang2021comprehensive,
       title={{A comprehensive survey on point cloud registration}},
       author={Huang, Xiaoshui and Mei, Guofeng and Zhang, Jian and Abbas, Rana},
       journal={arXiv preprint arXiv:2103.02690},
       year = {2021},
     }

Copyright


The datasets provided on this page are published under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 License. This means that you must attribute the work in the manner specified by the authors, you may not use this work for commercial purposes and if you alter, transform, or build upon this work, you may distribute the resulting work only under the same license. If you are interested in commercial usage you can contact us for further options.

Download

We provide several download links, including data set, data loader to read this dataset and several baseline algorithms.

The cross-source point cloud datasets could download from this link.
The data loader is a toolset to read the benchmark dataset by using python.
client
Dataloader
Download
We also provide several baseline algorithms. Get them from the link below.
client
Baseline algorithms
Download

Results

The evaluation results are listed below. If you have new results, please contact jian.zhang@uts.edu.au to add to this table.

Registration Method Registration Recall (%) Translation ErrorRotation ErrorTime(s)
DGR [2] 36.60.044.260.87
FMR [1] 17.80.104.660.28
PointNetLK [3] 0.050.0912.542.25
FGR [4] 1.490.0710.742.23
ICP [5] 24.30.385.710.19
JRMPC [6] 1.00.718.5718.1
RANSAC [7] 3.470.138.300.03
GCTR [8] 0.50.177.4615.8

References

[1] Huang, Xiaoshui, Guofeng Mei, and Jian Zhang. "Feature-metric Registration: A Fast Semi-supervised Approach for Robust Point Cloud Registration without Correspondences." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020.

[2] Choy, Christopher, Wei Dong, and Vladlen Koltun. "Deep global registration." Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2020.

[3] Aoki, Yasuhiro, et al. "Pointnetlk: Robust & efficient point cloud registration using pointnet." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2019.

[4] Zhou, Qian-Yi, Jaesik Park, and Vladlen Koltun. "Fast global registration." European Conference on Computer Vision. Springer, Cham, 2016.

[5] Peng, Furong, et al. "Street view cross-sourced point cloud matching and registration." 2014 IEEE International Conference on Image Processing (ICIP). IEEE, 2014.

[6] Huang, Xiaoshui, et al. "A coarse-to-fine algorithm for registration in 3D street-view cross-source point clouds." 2016 International Conference on Digital Image Computing: Techniques and Applications (DICTA). IEEE, 2016.

[7] Mellado, Nicolas, Matteo Dellepiane, and Roberto Scopigno. "Relative scale estimation and 3D registration of multi-modal geometry using Growing Least Squares." IEEE transactions on visualization and computer graphics 22.9 (2015): 2160-2173.

[8] Huang, Xiaoshui, et al. "Fast registration for cross-source point clouds by using weak regional affinity and pixel-wise refinement." 2019 IEEE International Conference on Multimedia and Expo (ICME). IEEE, 2019.

Contact Us

If you have any questions about the cross-source point cloud registration benchmark, please contact Prof. Jian Zhang ( jian.zhang@uts.edu.au ).