skip to main content
10.1145/3394171.3413661acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
research-article

Campus3D: A Photogrammetry Point Cloud Benchmark for Hierarchical Understanding of Outdoor Scene

Published: 12 October 2020 Publication History

Abstract

Learning on 3D scene-based point cloud has received extensive attention as its promising application in many fields, and well-annotated and multisource datasets can catalyze the development of those data-driven approaches. To facilitate the research of this area, we present a richly-annotated 3D point cloud dataset for multiple outdoor scene understanding tasks and also an effective learning framework for its hierarchical segmentation task. The dataset was generated via the photogrammetric processing on unmanned aerial vehicle (UAV) images of the National University of Singapore (NUS) campus, and has been point-wisely annotated with both hierarchical and instance-based labels. Based on it, we formulate a hierarchical learning problem for 3D point cloud segmentation and propose a measurement evaluating consistency across various hierarchies. To solve this problem, a two-stage method including multi-task (MT) learning and hierarchical ensemble (HE) with consistency consideration is proposed. Experimental results demonstrate the superiority of the proposed method and potential advantages of our hierarchical annotations. In addition, we benchmark results of semantic and instance segmentation, which is accessible online at https://3d.dataset.site with the dataset and all source codes.

Supplementary Material

ZIP File (mmfp1456aux.zip)
The supplementary material contains a 'mmfp1456_supp.pdf' file. It provides more information on proposed the dataset and method, including descriptions of dataset, experiment settings, and visualized statistics.
MP4 File (3394171.3413661.mp4)
This video presents the work of the paper ?Campus3D: A Photogrammetry Point Cloud Benchmark for Hierarchical Understanding of Outdoor Scene? for ACM Multimedia 2020. The presenters are Xinke Li and Chongshou Li who are the authors of the paper. \r\nKey contribution of the paper is highlighted first in the presentation, followed by the motivation, dataset introduction and the hierarchical annotation. Notably, animations are integrated in slides to better visualize the proposed dataset. Based on the dataset, a two-stage method is utilized to solve the hierarchical learning problem. Some key concepts are introduced including hierarchical consistency and consistency rate. Finally, a benchmark of 3D understanding task is built, containing two efficient sampling methods with fast local query. \r\nOverall, this video is recorded for better understanding on the work of paper, specifically, (1) a better visual illustration of the dataset, (2) a succinct summary of the whole work.

References

[1]
Iro Armeni, Sasha Sax, Amir R Zamir, and Silvio Savarese. 2017. Joint 2d-3d-semantic data for indoor scene understanding. arXiv preprint arXiv:1702.01105 (2017).
[2]
Iro Armeni, Ozan Sener, Amir R Zamir, Helen Jiang, Ioannis Brilakis, Martin Fischer, and Silvio Savarese. 2016. 3d semantic parsing of large-scale indoor spaces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 1534--1543.
[3]
Jens Behley, Martin Garbade, Andres Milioto, Jan Quenzel, Sven Behnke, Cyrill Stachniss, and Jurgen Gall. 2019. SemanticKITTI: A dataset for semantic scene understanding of lidar sequences. In Proceedings of the IEEE International Conference on Computer Vision. 9297--9307.
[4]
Cindy Cappelle, Maan E El Najjar, Francc ois Charpillet, and Denis Pomorski. 2012. Virtual 3D city model for navigation in urban areas. Journal of Intelligent & Robotic Systems, Vol. 66, 3 (2012), 377--399.
[5]
Ludovico Carozza, David Tingdahl, Frédéric Bosché, and Luc Van Gool. 2014. Markerless vision-based augmented reality for urban planning. Computer-Aided Civil and Infrastructure Engineering, Vol. 29, 1 (2014), 2--17.
[6]
Angel Chang, Angela Dai, Thomas Funkhouser, Maciej Halber, Matthias Niessner, Manolis Savva, Shuran Song, Andy Zeng, and Yinda Zhang. 2017. Matterport3d: Learning from rgb-d data in indoor environments. arXiv preprint arXiv:1709.06158 (2017).
[7]
Arnis Cirulis and Kristaps Brigis Brigmanis. 2013. 3D outdoor augmented reality for architecture and urban planning. Procedia Computer Science, Vol. 25 (2013), 71--79.
[8]
CloudCompare. [n.d.]. CloudCompare: 3D point cloud and mesh processing software, Open Source Project. https://www.danielgm.net/cc/
[9]
Angela Dai, Angel X Chang, Manolis Savva, Maciej Halber, Thomas Funkhouser, and Matthias Nießner. 2017. Scannet: Richly-annotated 3d reconstructions of indoor scenes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 5828--5839.
[10]
Michael Firman. 2016. RGBD datasets: Past, present and future. In Proceedings of the IEEE conference on computer vision and pattern recognition workshops. 19--31.
[11]
Jan-Michael Frahm, Pierre Fite-Georgel, David Gallup, Tim Johnson, Rahul Raguram, Changchang Wu, Yi-Hung Jen, Enrique Dunn, Brian Clipp, Svetlana Lazebnik, et almbox. 2010. Building rome on a cloudless day. In European Conference on Computer Vision. Springer, 368--381.
[12]
Andreas Geiger, Philip Lenz, Christoph Stiller, and Raquel Urtasun. 2013. Vision meets robotics: The KITTI dataset. The International Journal of Robotics Research, Vol. 32, 11 (2013), 1231--1237.
[13]
Andreas Geiger, Philip Lenz, and Raquel Urtasun. 2012. Are we ready for autonomous driving? the kitti vision benchmark suite. In 2012 IEEE Conference on Computer Vision and Pattern Recognition. IEEE, 3354--3361.
[14]
Timo Hackel, Nikolay Savinov, Lubor Ladicky, Jan D Wegner, Konrad Schindler, and Marc Pollefeys. 2017. SEMANTIC3D. NET: A NEW LARGE-SCALE POINT CLOUD CLASSIFICATION BENCHMARK. ISPRS Annals of Photogrammetry, Remote Sensing & Spatial Information Sciences, Vol. 4 (2017).
[15]
Qingyong Hu, Bo Yang, Linhai Xie, Stefano Rosa, Yulan Guo, Zhihua Wang, Niki Trigoni, and Andrew Markham. 2019 RandLA-Net: Efficient semantic segmentation of large-scale point clouds. arXiv preprint arXiv:1911.11236 (2019).
[16]
Binh-Son Hua, Quang-Hieu Pham, Duc Thanh Nguyen, Minh-Khoi Tran, Lap-Fai Yu, and Sai-Kit Yeung. 2016. Scenenn: A scene meshes dataset with annotations. In 2016 Fourth International Conference on 3D Vision (3DV). IEEE, 92--101.
[17]
Thomas H Kolbe, Gerhard Gröger, and Lutz Plümer. 2005. CityGML: Interoperable access to 3D city models. In Geo-information for disaster management. Springer, 883--899.
[18]
Loic Landrieu and Martin Simonovsky. 2018. Large-scale point cloud semantic segmentation with superpoint graphs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 4558--4567.
[19]
Minglei Li, Liangliang Nan, Neil Smith, and Peter Wonka. 2016. Reconstructing building mass models from UAV images. Computers & Graphics, Vol. 54 (2016), 84--93.
[20]
Yangyan Li, Rui Bu, Mingchao Sun, Wei Wu, Xinhan Di, and Baoquan Chen. 2018. PointCNN: Convolution On X-Transformed Points. In Advances in Neural Information Processing Systems. 828--838.
[21]
Madeleine Manyoky, Ulrike Wissen Hayek, Kurt Heutschi, Reto Pieren, and Adrienne Grêt-Regamey. 2014. Developing a GIS-based visual-acoustic 3D simulation for wind farm assessment. ISPRS International Journal of Geo-Information, Vol. 3, 1 (2014), 29--48.
[22]
Quang-Hieu Pham, Thanh Nguyen, Binh-Son Hua, Gemma Roig, and Sai-Kit Yeung. 2019. JSIS3D: joint semantic-instance segmentation of 3d point clouds with multi-task pointwise networks and multi-value conditional random fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 8827--8836.
[23]
Charles R Qi, Hao Su, Kaichun Mo, and Leonidas J Guibas. 2017a. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition . 652--660.
[24]
Charles Ruizhongtai Qi, Li Yi, Hao Su, and Leonidas J Guibas. 2017b. Pointnet: Deep hierarchical feature learning on point sets in a metric space. In Advances in neural information processing systems. 5099--5108.
[25]
Xavier Roynard, Jean-Emmanuel Deschaud, and Francc ois Goulette. 2018. Paris-Lille-3D: A large and high-quality ground-truth urban point cloud dataset for automatic segmentation and classification. The International Journal of Robotics Research, Vol. 37, 6 (2018), 545--557.
[26]
Andrés Serna, Beatriz Marcotegui, Francc ois Goulette, and Jean-Emmanuel Deschaud. 2014. Paris-rue-Madame database: a 3D mobile laser scanner dataset for benchmarking urban detection, segmentation and classification methods. In 4th International Conference on Pattern Recognition, Applications and Methods ICPRAM 2014.
[27]
Nathan Silberman and Rob Fergus. 2011. Indoor scene segmentation using a structured light sensor. In 2011 IEEE international conference on computer vision workshops (ICCV workshops). IEEE, 601--608.
[28]
Nathan Silberman, Derek Hoiem, Pushmeet Kohli, and Rob Fergus. 2012. Indoor segmentation and support inference from rgbd images. In European Conference on Computer Vision. Springer, 746--760.
[29]
Shuran Song, Samuel P Lichtenberg, and Jianxiong Xiao. 2015. Sun rgb-d: A rgb-d scene understanding benchmark suite. In Proceedings of the IEEE conference on computer vision and pattern recognition. 567--576.
[30]
Hugues Thomas, Charles R Qi, Jean-Emmanuel Deschaud, Beatriz Marcotegui, Francc ois Goulette, and Leonidas J Guibas. 2019. Kpconv: Flexible and deformable convolution for point clouds. In Proceedings of the IEEE International Conference on Computer Vision. 6411--6420.
[31]
Bruno Vallet, Mathieu Brédif, Andrés Serna, Beatriz Marcotegui, and Nicolas Paparoditis. 2015. TerraMobilita/iQmulus urban point cloud analysis benchmark. Computers & Graphics, Vol. 49 (2015), 126--133.
[32]
Yannick Verdie, Florent Lafarge, and Pierre Alliez. 2015. Lod generation for urban scenes. ACM Transactions on Graphics, Vol. 34, ARTICLE (2015), 30.
[33]
Lei Wang, Yuchun Huang, Yaolin Hou, Shenman Zhang, and Jie Shan. 2019 a. Graph attention convolution for point cloud semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 10296--10305.
[34]
Shenlong Wang, Simon Suo, Wei-Chiu Ma, Andrei Pokrovsky, and Raquel Urtasun. 2018a. Deep parametric continuous convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2589--2597.
[35]
Weiyue Wang, Ronald Yu, Qiangui Huang, and Ulrich Neumann. 2018b. Sgpn: Similarity group proposal network for 3d point cloud instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2569--2578.
[36]
Xinlong Wang, Shu Liu, Xiaoyong Shen, Chunhua Shen, and Jiaya Jia. 2019 b. Associatively segmenting instances and semantics in point clouds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 4096--4105.
[37]
Yue Wang, Yongbin Sun, Ziwei Liu, Sanjay E Sarma, Michael M Bronstein, and Justin M Solomon. 2019 c. Dynamic graph cnn for learning on point clouds. ACM Transactions on Graphics (TOG), Vol. 38, 5 (2019), 146.
[38]
Matthew J Westoby, James Brasington, Niel F Glasser, Michael J Hambrey, and Jennifer M Reynolds. 2012. 2. Structure-from-Motion photogrammetry: A low-cost, effective tool for geoscience applications. Geomorphology, Vol. 179 (2012), 300--314.
[39]
Changchang Wu, Sameer Agarwal, Brian Curless, and Steven M Seitz. 2011. Multicore bundle adjustment. In CVPR 2011. IEEE, 3057--3064.
[40]
Jianxiong Xiao, Andrew Owens, and Antonio Torralba. 2013. Sun3d: A database of big spaces reconstructed using sfm and object labels. In Proceedings of the IEEE International Conference on Computer Vision. 1625--1632.
[41]
Saining Xie, Sainan Liu, Zeyu Chen, and Zhuowen Tu. 2018. Attentional shapecontextnet for point cloud recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 4606--4615.
[42]
Jiancheng Yang, Qiang Zhang, Bingbing Ni, Linguo Li, Jinxian Liu, Mengdie Zhou, and Qi Tian. 2019. Modeling point clouds with self-attention and gumbel subset sampling. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 3323--3332.
[43]
Hengshuang Zhao, Li Jiang, Chi-Wing Fu, and Jiaya Jia. 2019. PointWeb: Enhancing local neighborhood features for point cloud processing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 5565--5573.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
MM '20: Proceedings of the 28th ACM International Conference on Multimedia
October 2020
4889 pages
ISBN:9781450379885
DOI:10.1145/3394171
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 12 October 2020

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. hierarchical learning
  2. instance segmentation
  3. point cloud
  4. scene understanding
  5. semantic segmentation

Qualifiers

  • Research-article

Funding Sources

  • National University of Singapore Institute of Operations Research and Analytics (IORA)
  • National Research Foundation of Singapore

Conference

MM '20
Sponsor:

Acceptance Rates

Overall Acceptance Rate 2,145 of 8,556 submissions, 25%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)83
  • Downloads (Last 6 weeks)4
Reflects downloads up to 11 Jan 2025

Other Metrics

Citations

Cited By

View all

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media