Toronto 3D Save

A Large-scale Mobile LiDAR Dataset for Semantic Segmentation of Urban Roadways

Project README

Toronto-3D: A Large-scale Mobile LiDAR Dataset for Semantic Segmentation of Urban Roadways

[Paper] [Download] [Results] [Codes]

Image

Toronto-3D is a large-scale urban outdoor point cloud dataset acquired by an MLS system in Toronto, Canada for semantic segmentation. This dataset covers approximately 1 km of road and consists of about 78.3 million points. Here is an overview of the dataset and the tiles. The approximate location of the dataset is at (43.726, -79.417).

Image

Point clouds has 10 attributes and classified in 8 labelled object classes. There is a data preparation tip to handle UTM coordinates to avoid problems. There are also some known issues.

Details on the dataset can be found at CVPRW2020. Revisions on the labels will lead to different results from the published paper, and updated results will be updated here.

If you have questions, or any suggestions to help us improve the dataset, please contact Weikai Tan.


Semantic segmentation results (%)

More results to be added

Default: point coordinates only

Method OA mIoU Road Road mrk. Natural Bldg Util. line Pole Car Fence
PointNet++ 84.88 41.81 89.27 0.00 69.0 54.1 43.7 23.3 52.0 3.0
PointNet++ MSG 92.56 59.47 92.90 0.00 86.13 82.15 60.96 62.81 76.41 14.43
PointNet++ * 91.66 58.01 92.71 7.68 84.30 81.83 67.44 63.30 60.92 5.92
DGCNN 94.24 61.79 93.88 0.00 91.25 80.39 62.40 62.32 88.26 15.81
KPFCNN 95.39 69.11 94.62 0.06 96.07 91.51 87.68 81.56 85.66 15.72
MS-PCNN 90.03 65.89 93.84 3.83 93.46 82.59 67.80 71.95 91.12 22.50
TGNet 94.08 61.34 93.54 0.00 90.83 81.57 65.26 62.98 88.73 7.85
MS-TGNet 95.71 70.50 94.41 17.19 95.72 88.83 76.01 73.97 94.24 23.64
RandLA-Net Hu, et al., 2021 92.95 77.71 94.61 42.62 96.89 93.01 86.51 78.07 92.85 37.12
Rim et al., 2021 72.55 66.87 92.74 14.75 88.66 93.52 81.03 67.71 39.65 56.90
MappingConvSeg Yan, et al., 2021 93.17 77.57 95.02 39.27 96.77 93.32 86.37 79.11 89.81 40.89
Use RGB
RandLA-Net Hu, et al., 2021 (RGB) 94.37 81.77 96.69 64.21 96.92 94.24 88.06 77.84 93.37 42.86
Rim et al., 2021 (RGB) 83.60 71.03 92.84 27.43 89.90 95.27 85.59 74.50 44.41 58.30
MappingConvSeg Yan, et al., 2021 94.72 82.89 97.15 67.87 97.55 93.75 86.88 82.12 93.72 44.11
Others
Han et al., 2021 (Intensity + Normal) 93.60 70.80 92.20 53.80 92.80 86.00 72.20 72.50 75.70 21.20

* use same radii and k as TGNet


Codes for training your own network


Point cloud attributes

  • XYZ
  • RGB
  • Intensity
  • GPS time
  • Scan angle rank

Classes

  • Road (label 1)
  • Road marking (label 2)
  • Natural (label 3)
  • Building (label 4)
  • Utility line (label 5)
  • Pole (label 6)
  • Car (label 7)
  • Fence (label 8)
  • unclassified (label 0)

Data preparation tip

The XY coordinates are stored in UTM format. The Y coordinate may exceed decimal digits in float type commonly used in point cloud processing algorithms. Directly read and process the coordinates could result in loss of detail and wrong geometric features.

I set a UTM_OFFSET = [627285, 4841948, 0] to subtract from the raw coordinates. You may use any other numbers to reduce number of digits.

Example of potential issues during grid_subsampling operation used in KPConv and RandLA-Net: both subsampled to grid size 6cm

without offset with offset

Known issues

  1. Point RGB assignments on taller vehicles.

Image

  1. Point RGB artifact assignments on moving vehicles.

Image

  1. Point acquisition on moving vehicles.

Image


Download

Dataset can be downloaded at OneDrive or 百度网盘(提取码:aewp). Check Changelog for changes.

Toronto-3D belongs to Mobile Sensing and Geodata Science Lab, University of Waterloo. Toronto-3D is distributed under the CC BY-NC 4.0 License

Citation

Please consider citing our work:

@inproceedings{tan2020toronto3d,
    title={{Toronto-3D}: A large-scale mobile lidar dataset for semantic segmentation of urban roadways},
    author={Tan, Weikai and Qin, Nannan and Ma, Lingfei and Li, Ying and Du, Jing and Cai, Guorong and Yang, Ke and Li, Jonathan},
    booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
    pages={202--203},
    year={2020}
}

Acknowledgements

Teledyne Optech is acknowledged for providing mobile LiDAR point cloud data collected by Maverick. Thanks Jing Du and Dr. Guorong Cai from Jimei University for point cloud labelling.

Thanks Intel ISL for including our dataset in the Open3D-ML 3D Machine Learning module.


Changelog

  • [2023-02-07] Added code for RandLA-Net

  • [2020-04-23] Uploaded newest version. Fixed some labelling errors. Major revision on cars.

  • [2020-03-22] Uploaded newest version.

Open Source Agenda is not affiliated with "Toronto 3D" Project. README Source: WeikaiTan/Toronto-3D

Open Source Agenda Badge

Open Source Agenda Rating