Point Cloud Convolution Network Based on Spatial Location Correspondence
Abstract
:1. Introduction
- (1)
- We reveal the essence of discrete convolution—the summation of products based on correspondence. What matters is the correspondence. We believe that convolution is an operation that is related not to the order, but to the correspondence between the convolution range and the convolution kernel. We argue that the convolution value does not change as long as the correspondence between the convolution range and the elements in a convolution kernel is kept constant.
- (2)
- We found that spatial location correspondence satisfies 3D point clouds, which can solve the problem of disorder in point clouds; furthermore, we analyzed different correspondence styles, suggesting that point clouds should adopt N-to-M correspondences, which can solve the problem of irregularity in point clouds. These are not covered in other existing convolution networks.
- (3)
- We propose a general convolution framework for point clouds according to the spatial location correspondence and give an example of a convolution network based on this framework. We carried out several experiments on point cloud tasks, such as classification and semantic segmentation. All of our results achieved consistency with the current mainstream networks.
2. Related Work
2.1. Projection-Based
2.1.1. Multi-View-Based
2.1.2. Voxel-Based
2.2. Point-Based
2.3. Graph-Based
2.4. Convolution-Based
2.5. Transformer-Based
3. Materials and Methods
3.1. The Mathematical Nature of Convolution
3.2. Spatial Location Correspondence
3.3. Point Convolution Framework
- (1)
- First, determine a suitable point cloud neighborhood system.
- (2)
- Second, determine how the convolution kernel’s coordinate points are generated and the appropriate size of the convolution kernel. In this study, we generated the kernel points from the covariance matrix of a sample.
- (3)
- Third, determine the range of influence of each of the convolution kernel points based on the Euclidean spatial location.
- (4)
- Finally, apply the convolution operation according to the correspondence.
3.4. An Example of a Network
4. Results and Discussion
4.1. Classification Tasks
ModelNet40 Classification
4.2. Semantic Segmentation Tasks
4.2.1. S3DIS: Semantic Segmentation for Indoor Scenes
4.2.2. Semantic3D: LiDAR Semantic Segmentation
4.2.3. SensatUrban: Photogrammetric Point Cloud Datasets at the City Level
4.3. Discussion
4.3.1. The Way in Which the Kernel Points Were Generated
4.3.2. The Number of Kernel Points
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Armeni, I.; Sener, O.; Zamir, A.R.; Jiang, H.; Brilakis, I.; Fischer, M.; Savarese, S. 3D Semantic Parsing of Large-Scale Indoor Spaces. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE: Las Vegas, NV, USA; pp. 1534–1543. [Google Scholar]
- Hackel, T.; Savinov, N.; Ladicky, L.; Wegner, J.D.; Schindler, K.; Pollefeys, M. Semantic3D. Net: A New Large-Scale Point Cloud Classification Benchmark. arXiv 2017, arXiv:1704.03847. [Google Scholar]
- Hu, Q.; Yang, B.; Khalid, S.; Xiao, W.; Trigoni, N.; Markham, A. Towards Semantic Segmentation of Urban-Scale 3D Point Clouds: A Dataset, Benchmarks and Challenges. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; IEEE: Nashville, TN, USA; pp. 4975–4985. [Google Scholar]
- Artificial Intelligence Computing Leadership from NVIDIA. Available online: https://www.nvidia.com/en-us/ (accessed on 18 July 2022).
- TensorFlow. Available online: https://tensorflow.google.cn/?hl=en (accessed on 18 July 2022).
- PyTorch. Available online: https://pytorch.org/ (accessed on 18 July 2022).
- Wang, Y. DGCNN: Learning Point Cloud Representations by Dynamic Graph CNN. Ph.D. Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 2020. [Google Scholar]
- Thomas, H.; Qi, C.R.; Deschaud, J.-E.; Marcotegui, B.; Goulette, F.; Guibas, L.J. KPConv: Flexible and Deformable Convolution for Point Clouds. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
- Maturana, D.; Scherer, S. VoxNet: A 3D Convolutional Neural Network for Real-Time Object Recognition. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–3 October 2015; pp. 922–928. [Google Scholar]
- Li, Y.; Bu, R.; Sun, M.; Wu, W.; Di, X.; Chen, B. PointCNN: Convolution On X-Transformed Points. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 31. [Google Scholar]
- Hua, B.-S.; Tran, M.-K.; Yeung, S.-K. Pointwise Convolutional Neural Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; IEEE: Salt Lake City, UT, USA; pp. 984–993. [Google Scholar]
- Tatarchenko, M.; Park, J.; Koltun, V.; Zhou, Q.-Y. Tangent Convolutions for Dense Prediction in 3D. arXiv 2018, arXiv:1807.02443. [Google Scholar]
- Wang, S.; Suo, S.; Ma, W.-C.; Pokrovsky, A.; Urtasun, R. Deep Parametric Continuous Convolutional Neural Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2589–2597. [Google Scholar]
- Deuge, M.D.; Quadros, A.; Hung, C.; Douillard, B. Unsupervised Feature Learning for Classification of Outdoor 3D Scans. In Proceedings of the Australasian Conference on Robotics and Automation, Sydney, Australia, 2–4 December 2013. [Google Scholar]
- Hu, Q.; Yang, B.; Xie, L.; Rosa, S.; Guo, Y.; Wang, Z.; Trigoni, N.; Markham, A. RandLA-Net: Efficient Semantic Segmentation of Large-Scale Point Clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
- Graham, B.; Engelcke, M.; van der Maaten, L. 3D Semantic Segmentation with Submanifold Sparse Convolutional Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; IEEE: Salt Lake City, UT, USA; pp. 9224–9232. [Google Scholar]
- Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
- Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
- Simonovsky, M.; Komodakis, N. Dynamic Edge-Conditioned Filters in Convolutional Neural Networks on Graphs. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 29–38. [Google Scholar]
- Landrieu, L.; Simonovsky, M. Large-Scale Point Cloud Semantic Segmentation with Superpoint Graphs. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4558–4567. [Google Scholar]
- Guo, M.-H.; Cai, J.; Liu, Z.-N.; Mu, T.-J.; Martin, R.; Hu, S. PCT: Point Cloud Transformer. Undefined 2021, 7, 187–199. [Google Scholar] [CrossRef]
- Zhao, H.; Jiang, L.; Jia, J.; Torr, P.; Koltun, V. Point Transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
- Engel, N.; Belagiannis, V.; Dietmayer, K. Point Transformer. IEEE Access 2021, 9, 134826–134840. [Google Scholar] [CrossRef]
- Groh, F.; Wieschollek, P.; Lensch, H.P.A. Flex-Convolution (Million-Scale Point-Cloud Learning Beyond Grid-Worlds). arXiv 2020, arXiv:1803.07289. [Google Scholar]
- Lei, H.; Akhtar, N.; Mian, A. Octree Guided CNN With Spherical Kernels for 3D Point Clouds. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; IEEE: Long Beach, CA, USA; pp. 9623–9632. [Google Scholar]
- Wu, W.; Qi, Z.; Fuxin, L. PointConv: Deep Convolutional Networks on 3D Point Clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
- Lei, H.; Akhtar, N.; Mian, A. Spherical Kernel for Efficient Graph Convolution on 3D Point Clouds. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 3664–3680. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Su, H.; Maji, S.; Kalogerakis, E.; Learned-Miller, E. Multi-View Convolutional Neural Networks for 3D Shape Recognition. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 945–953. [Google Scholar]
- Lin, Z.-H.; Huang, S.-Y.; Wang, Y.-C.F. Convolution in the Cloud: Learning Deformable Kernels in 3D Graph Convolution Networks for Point Cloud Analysis. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1797–1806. [Google Scholar]
- Thomas, H.; Goulette, F.; Deschaud, J.-E.; Marcotegui, B.; LeGall, Y. Semantic Classification of 3D Point Clouds with Multiscale Spherical Neighborhoods. In Proceedings of the 2018 International Conference on 3D Vision (3DV), Verona, Italy, 5–8 September 2018; pp. 390–398. [Google Scholar]
- Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3D ShapeNets: A Deep Representation for Volumetric Shapes. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; IEEE: Boston, MA, USA; pp. 1912–1920. [Google Scholar]
- Li, J.; Chen, B.M.; Lee, G.H. SO-Net: Self-Organizing Network for Point Cloud Analysis. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 9397–9406. [Google Scholar]
- Xu, Y.; Fan, T.; Xu, M.; Zeng, L.; Qiao, Y. SpiderCNN: Deep Learning on Point Sets with Parameterized Convolutional Filters. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
- Tchapmi, L.; Choy, C.; Armeni, I.; Gwak, J.; Savarese, S. SEGCloud: Semantic Segmentation of 3D Point Clouds. In Proceedings of the 2017 International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017; pp. 537–547. [Google Scholar]
- Semantic3D. Available online: http://semantic3d.net/ (accessed on 16 November 2022).
Networks | Modelnet40 |
---|---|
OA (%) | |
PointNet [17] | 89.2 |
PointNet++ [18] | 90.7 |
SO-Net [32] | 90.9 |
SpiderCNN [33] | 90.5 |
FlexConv [24] | 90.2 |
DGCNN [7] | 92.2 |
SPH3D [27] | 92.1 |
PointConv [26] | 92.5 |
KPConv [8] | 92.9 |
Ours—Covariance | 92.7 |
Ours—Random | 91.5 |
Method | mIOU | Acc | Ceil. | Floor | Wall | Beam | Col. | Win. | Door | Table | Chair | Book. | Sofa | Broad | Clut. |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
PointNet [17] | 41.1 | 49.0 | 88.8 | 97.3 | 69.8 | 0.1 | 3.9 | 46.3 | 10.8 | 58.9 | 52.6 | 58.9 | 40.3 | 5.9 | 26.4 |
SEGCloud [34] | 48.9 | 57.4 | 90.1 | 96.1 | 69.9 | 0.0 | 18.4 | 38.4 | 23.1 | 70.4 | 75.9 | 70.4 | 58.4 | 40.9 | 13.0 |
Tanentconv [12] | 52.6 | 62.2 | 90.5 | 97.7 | 74.0 | 0.0 | 20.7 | 39.0 | 31.3 | 77.5 | 69.4 | 77.5 | 38.5 | 57.3 | 48.8 |
SPGraph [20] | 58.0 | 66.5 | 89.4 | 96.9 | 78.1 | 0.0 | 42.8 | 48.9 | 61.6 | 75.4 | 84.7 | 75.4 | 69.8 | 52.6 | 2.1 |
Paramconv [13] | 58.3 | 67.1 | 92.3 | 96.2 | 75.9 | 0.3 | 6.0 | 69.5 | 63.5 | 65.6 | 66.9 | 65.6 | 47.3 | 68.9 | 59.1 |
SPH3d [27] | 59.5 | - | 93.3 | 97.1 | 81.1 | 0.0 | 33.2 | 45.8 | 43.8 | 79.7 | 86.9 | 71.5 | 33.2 | 54.1 | 53.7 |
KPConv [8] | 65.4 | 70.9 | 92.6 | 97.3 | 81.4 | 0.0 | 16.5 | 54.5 | 69.5 | 80.2 | 90.1 | 80.2 | 74.6 | 66.4 | 63.7 |
Ours | 64.6 | 69.6 | 93.0 | 97.4 | 82.3 | 0.2 | 29.3 | 60.3 | 62.1 | 78.6 | 89.1 | 78.3 | 75.2 | 67.2 | 26.8 |
Method | mIOU | OA | Man. | Natural. | High Veg. | Low Veg. | Building | Hard. | Scan. | Cars |
---|---|---|---|---|---|---|---|---|---|---|
SEGCloud [34] | 59.1 | 88.6 | 82.0 | 77.3 | 79.7 | 22.9 | 91.1 | 18.4 | 37.3 | 64.4 |
SPGraph [20] | 73.0 | 84.0 | 97.4 | 92.6 | 87.9 | 44.0 | 83.2 | 31.0 | 63.5 | 76.2 |
KPConv [8] | 74.6 | 92.9 | 90.9 | 82.2 | 84.2 | 47.9 | 94.9 | 40.0 | 77.3 | 79.7 |
RandLA-NET [15] | 77.4 | 94.8 | 95.6 | 91.4 | 86.6 | 51.5 | 95.7 | 51.5 | 69.8 | 76.8 |
Ours—colored | 74.4 | 92.7 | 90.5 | 90.7 | 82.6 | 46.5 | 90.6 | 50.6 | 70.3 | 73.7 |
Model | mIOU | OA | Man. | Natural. | High Veg. | Low veg. | Building | Hard. | Scan. | Cars |
---|---|---|---|---|---|---|---|---|---|---|
Non-colored | 79.0 | 95.9 | 97.7 | 78..8 | 77.3 | 35.7 | 98.9 | 70.3 | 74.2 | 99.2 |
Method | OA | mIOU | Grou. | Veg. | Buil. | Wall | Brid. | Park. | Rail | Traf. | Stre. | Car | Foo. | Bike | Water |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
PointNet [17] | 80.78 | 23.71 | 67.96 | 89.52 | 80.05 | 0.00 | 0.00 | 3.95 | 0.00 | 31.55 | 0.00 | 35.14 | 0.00 | 0.00 | 0.00 |
PointNet++ [18] | 84.30 | 32.92 | 72.46 | 94.24 | 84.77 | 2.72 | 2.09 | 25.79 | 0.00 | 31.54 | 11.42 | 38.84 | 7.12 | 0.00 | 56.93 |
SPGraph [20] | 85.27 | 37.29 | 69.93 | 94.55 | 88.87 | 32.83 | 12.58 | 15.77 | 15.48 | 30.63 | 22.96 | 56.42 | 0.54 | 0.00 | 44.24 |
SparseConv [16] | 88.66 | 42.66 | 74.10 | 97.90 | 94.20 | 63.30 | 7.50 | 24.20 | 0.00 | 30.10 | 34.00 | 74.40 | 0.00 | 0.00 | 54.80 |
KPConv [8] | 93.20 | 57.58 | 87.10 | 98.91 | 95.33 | 74.40 | 28.69 | 41.38 | 0.00 | 55.99 | 54.43 | 85.67 | 40.39 | 0.00 | 86.30 |
RandLA-Net [15] | 89.78 | 52.69 | 80.11 | 98.07 | 91.58 | 48.88 | 40.75 | 51.62 | 0.00 | 56.67 | 33.23 | 80.14 | 32.63 | 0.00 | 71.31 |
Ours | 91.6 | 56.92 | 86.56 | 98.08 | 92.35 | 68.6 | 35.68 | 49.56 | 5.63 | 57.86 | 40.36 | 86.25 | 38.36 | 0.00 | 80.63 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Xv, J.; Deng, F.; Liu, H. Point Cloud Convolution Network Based on Spatial Location Correspondence. ISPRS Int. J. Geo-Inf. 2022, 11, 591. https://doi.org/10.3390/ijgi11120591
Xv J, Deng F, Liu H. Point Cloud Convolution Network Based on Spatial Location Correspondence. ISPRS International Journal of Geo-Information. 2022; 11(12):591. https://doi.org/10.3390/ijgi11120591
Chicago/Turabian StyleXv, Jiabin, Fei Deng, and Haibing Liu. 2022. "Point Cloud Convolution Network Based on Spatial Location Correspondence" ISPRS International Journal of Geo-Information 11, no. 12: 591. https://doi.org/10.3390/ijgi11120591
APA StyleXv, J., Deng, F., & Liu, H. (2022). Point Cloud Convolution Network Based on Spatial Location Correspondence. ISPRS International Journal of Geo-Information, 11(12), 591. https://doi.org/10.3390/ijgi11120591