Partition-Based Point Cloud Completion Network with Density Refinement
<p>It shows the structure of our proposed point cloud completion model, which is inspired by the PointNet architecture. The model takes global and local feature representations as input and generates a set of points as output. Our approach combines global and local point cloud density with spatial distance for denoising optimization. The figure also illustrates the transformation networks A and B in our model, hereafter referred to as transform_A and transform_B. S* represents a set of coordinates at any position.</p> "> Figure 2
<p>Comparison of collection methods in different sensory fields. (<b>A</b>). Perceptual field metric based on the radius criterion is the most commonly used method among researchers. It focuses on capturing local information within a specific radius. (<b>B</b>). Our proposed approach combines the perceptual field metric with the global sensory field approach. By incorporating two 3D geometric forms, we enhance the perception of the local field, resulting in a more comprehensive understanding of the data. (<b>C</b>). The global sensory field approach, although valuable for capturing global information, is used less frequently due to its high parametric characteristics, which present computational challenges.</p> "> Figure 3
<p>Diagram showing subregion optimization in point cloud refinement. The cube is divided into several subregions, each with cores representing its regional perceptual field. To achieve better refinement results, only the discrete points of each region are predicted in the refinement block. (<b>A</b>) shows the structural shape of point clouds in a single differentiated region, with yellow dashed lines explaining the simplified structural lines of point clouds in this region. (<b>B</b>) shows the segmentation method in the segmentation region, where the center point K is the center of the sphere. The number of segmentation regions can be determined based on the actual situation and experimental results. The figure shows the effect of dividing a small region into four regions.</p> "> Figure 4
<p>Test results score schematic chart. In this figure, we analyze the precision test results of different models and our proposed model in terms of CD and EMD metrics. We also compare the results of integrating our point cloud density refinement module (CD-D, EMD-M). Furthermore, we compare the accuracy of 300 and 3000 iterations.</p> "> Figure 5
<p>After multiple experimental comparisons, the experimental results of this model are good, with good completion ability for manually incomplete point clouds and strong optimization ability for overall point cloud density.</p> ">
Abstract
:1. Introduction
- (1)
- Proposed an encoder-decoder architecture that models pairwise interactions between point cloud elements to infer missing elements for point cloud completion.
- (2)
- Introduced a spatially sliced sensory field that transforms the input point cloud into uniform perceptual fields for better local feature representation in the transformer model.
- (3)
- Developed a geometric density-aware block to improve the exploitation of the 3D geometric structure and preserve fine details in the point cloud.
2. Related Work
2.1. Point Cloud Convolution
2.2. Point Cloud Completion
3. Method
3.1. Regional Experience Field
3.2. Global and Sub-Regional Convolution
3.3. Sub-Regional Optimization
4. Experiments and Evaluations
4.1. Loss Function
4.2. Implementation Details
4.3. Optimization
- (1)
- First, we need to estimate the local density of each point , i.e., the number of points in its -neighborhood. We can use either the k-nearest neighbor algorithm or the mesh partitioning algorithm to achieve this step.
- (2)
- Then, we can determine which points are outliers or noise points based on the density threshold and remove them. This step can be represented by the following equation: . where is the probability that belongs to an outlier, is a classification function, and is the set of points in the -neighborhood of . We define uniformly dispersed core points in the tangent region, with the distance between and defined as , we need to estimate the local density of each core point and find the farthest distance points with the distance limit ( is determined empirically).
- (3)
- The farthest points from the cores of the cat-off region are defined as “potentially discrete points”. The region represented by each is sampled or filtered to reduce the amount of data and preserve the main features. We can use uniform sampling, nearest neighbor interpolation, bilateral filtering, etc. to achieve this step. This step can be represented by the following equation: , where is the offset estimated for the point (or zero if there is no offset), and is a denoising or refinement function that is the other points in the cluster to which belongs.
- (4)
- The final denoised and refined point cloud is obtained as: , where is the final prediction for point . The comparison results of the point cloud detail optimization after our design are shown in Table 4. The CD-D(CD With Density) and EMD-D(EMD With Density) are representations of the results after the optimization of our model combining density and spatial location.
4.4. Results
5. Conclusions and Discussions
- (1)
- Local-Global Fusion: We introduce the concept of perceptual fields, which divide the input point cloud into uniform local regions. By combining global and local information, our method effectively captures the overall shape and fine-grained details, preserving sharp edges and detailed structures.
- (2)
- Transformer-based Model: We utilize a transformer model to process the feature vectors obtained from each perceptual field. This allows us to capture long-range dependencies and effectively infer missing elements in the point cloud.
- (3)
- Geometric Density-aware Block: We design a geometric density-aware block to leverage the inherent 3D geometric structure of the point cloud. This block enhances the preservation of important geometric features and improves the accuracy of the completed point cloud.
- (1)
- Lack of scalability: Our proposed method may face scalability issues when dealing with large-scale point clouds, as it involves the partitioning of the point cloud.
- (2)
- Limited effectiveness for complex objects: While our method achieves state-of-the-art results on several benchmark datasets, it may have limited effectiveness for complex objects with more intricate shapes and details.
- (3)
- Limited generalizability: Our approach may have limited generalizability to point clouds from different domains or with different characteristics, as it was designed specifically for point cloud completion tasks.
Author Contributions
Funding
Institutional Review Board Statement
Data Availability Statement
Conflicts of Interest
References
- Fei, B.; Yang, W.; Chen, W.M.; Li, Z.; Li, Y.; Ma, T.; Hu, X.; Ma, L. Comprehensive review of deep learning-based 3D point cloud completion processing and analysis. IEEE Trans. Intell. Transport. Syst. 2022, 23, 22862–22883. [Google Scholar] [CrossRef]
- Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the 2017 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Hangzhou, China, 7–9 August 2017; pp. 652–660. [Google Scholar]
- Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. In Proceedings of the 2017 Conference on Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, 4–9 July 2017; pp. 652–660. [Google Scholar]
- Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E. Dynamic graph cnn for learning on point clouds. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef] [Green Version]
- Pistilli, F.; Fracastoro, G.; Valsesia, D.; Magli, E. Learning Robust Graph-Convolutional Representations for Point Cloud Denoising. IEEE J. Sel. Top. Signal Process. 2021, 15, 402–414. [Google Scholar] [CrossRef]
- Wu, W.; Qi, Z.; Fuxin, L. PointConv: Deep Convolutional Networks on 3D Point Clouds. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition(CVPR), Seoul, Republic of Korea, 4–6 November 2019; pp. 9613–9622. [Google Scholar]
- Yuan, W.; Khot, T.; Held, D.; Mertz, C.; Hebert, M. PCN: Point Completion Network. In Proceedings of the 2018 International Conference on 3D Vision (3DV), Verona, Italy, 5–8 September 2018; pp. 728–737. [Google Scholar]
- Najibi, M.; Rastegari, M.; Davis, L.S. G-CNN: An Iterative Grid Based Object Detector. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2369–2377. [Google Scholar]
- Wagh, N.; Varatharajah, Y. Eeg-gcnn: Augmenting electroencephalogram-based neurological disease diagnosis using a domain-guided graph convolutional neural network. In Proceedings of the Machine Learning for Health(PMLR), Virtually, 24 August 2020; pp. 367–378. [Google Scholar]
- Feng, F.; Huang, W.; He, X.; Xin, X.; Wang, Q.; Chua, T.S. Should graph convolution trust neighbors? a simple causal inference method. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR), Virtually, 11–15 July 2021; pp. 1208–1218. [Google Scholar]
- Afrasiabi, S.; Mohammadi, M.; Afrasiabi, M.; Parang, B. Modulated Gabor filter based deep convolutional network for electrical motor bearing fault classification and diagnosis. IET Sci. Meas. Technol. 2021, 15, 154–162. [Google Scholar] [CrossRef]
- Dai, X.; Fu, R.; Zhao, E.; Zhang, Z.; Lin, Y.; Wang, F.Y.; Li, L. DeepTrend 2.0: A light-weighted multi-scale traffic prediction model using detrending. Transp. Res. Part C Emerg. Technol. 2019, 103, 142–157. [Google Scholar] [CrossRef]
- Shafqat, W.; Byun, Y.C. Incorporating similarity measures to optimize graph convolutional neural networks for product recommendation. Appl. Sci. 2021, 11, 1366. [Google Scholar] [CrossRef]
- Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef] [Green Version]
- Rana, M.A.; Mahmood, T.; Afzal, M. A Survey of Convolutional Neural Networks: Architectures, Algorithms, and Applications. IEEE Access 2021, 9, 129521–129549. [Google Scholar]
- Chiang, W.L.; Liu, X.; Si, S.; Li, Y.; Bengio, S.; Hsieh, C.J. Cluster-gcn: An efficient algorithm for training deep and large graph convolutional networks. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 257–266. [Google Scholar]
- Zhao, L.; Song, Y.; Zhang, C.; Liu, Y.; Wang, P.; Lin, T.; Deng, M.; Li, H. T-gcn: A temporal graph convolutional network for traffic prediction. IEEE Trans. Intell. Transport. Syst. 2019, 21, 3848–3858. [Google Scholar] [CrossRef] [Green Version]
- Abu-El-Haija, S.; Kapoor, A.; Perozzi, B.; Lee, J. N-gcn: Multi-scale graph convolution for semi-supervised node classification. In Proceedings of the Uncertainty in Artificial Intelligence (PMLR), Online, 22–25 July 2019; pp. 841–851. [Google Scholar]
- Ben-Shabat, Y.; Lindenbaum, M.; Fischer, A. Nesti-net: Normal estimation for unstructured 3d point clouds using convolutional neural networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seoul, Republic of Korea, 4–6 November 2019; pp. 10112–10120. [Google Scholar]
- Hermosilla, P.; Ritschel, T.; Ropinski, T. Total denoising: Unsupervised learning of 3D point cloud cleaning. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seoul, Republic of Korea, 4–6 November 2019; pp. 52–60. [Google Scholar]
- Xie, H.; Yao, H.; Zhou, S.; Mao, J.; Zhang, S.; Sun, W. Grnet: Gridding residual network for dense point cloud completion. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 365–381. [Google Scholar]
- Zhang, Q.; Chen, Z.; Xue, Y.; Zhang, J. Attention-guided graph convolutional network for 3D point cloud classification. Signal Process. Image Commun. 2021, 99, 116317. [Google Scholar] [CrossRef]
- Wang, J.; Li, Y.; Sun, M.; Liu, Y.; Rosman, G. Learning to grasp objects with a robot hand-eye system using simulated depth images and point clouds. Sci. Robot. 2020, 5, 7695. [Google Scholar]
- Cao, S.; Zhao, H.; Liu, P. Semantic Segmentation for Point Clouds via Semantic-Based Local Aggregation and Multi-Scale Global Pyramid. Machines 2023, 11, 11. [Google Scholar] [CrossRef]
- Oh, H.; Ahn, S.; Kim, J.; Lee, S. Blind deep S3D image quality evaluation via local to global feature aggregation. IEEE Trans. Image Process. 2017, 26, 4923–4936. [Google Scholar] [CrossRef] [PubMed]
- Tchapmi, L.P.; Kosaraju, V.; Rezatofighi, H.; Reid, I.; Savarese, S. TopNet: Structural Point Cloud Decoder. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seoul, Republic of Korea, 4–6 November 2019; pp. 383–392. [Google Scholar]
- Huang, W.; Xue, Y.; Hu, L.; Liuli, H. S-EEGNet: Electroencephalogram signal classification based on a separable convolution neural network with bilinear interpolation. IEEE Access 2020, 8, 131636–131646. [Google Scholar] [CrossRef]
- Gao, H.; Yuan, H.; Wang, Z.; Ji, S. Pixel transposed convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 1218–1227. [Google Scholar] [CrossRef] [PubMed]
- Dai, Y.; Lu, H.; Shen, C. Learning affinity-aware upsampling for deep image matting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 6841–6850. [Google Scholar]
- Thomas, H.; Qi, C.R.; Deschaud, J.E.; Marcotegui, B.; Goulette, F.; Guibas, L.J. KPConv: Flexible and Deformable Convolution for Point Clouds. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–3 November 2019; pp. 6410–6419. [Google Scholar]
- Wang, L.; Huang, Y.; Hou, Y.; Zhang, S.; Shan, J. Graph attention convolution for point cloud semantic segmentation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seoul, Republic of Korea, 4–6 November 2019; pp. 10296–10305. [Google Scholar]
- Komarichev, A.; Zhong, Z.; Hua, J. A-cnn: Annularly convolutional neural networks on point clouds. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seoul, Republic of Korea, 4–6 November 2019; pp. 7421–7430. [Google Scholar]
- Ji, S.; Wang, Y.; Zhang, Y.; Liu, L.; Ye, Q. Squeeze-and-Excitation Networks for 3D Deep Shape Analysis of Human Organs. IEEE Trans. Med. Imaging 2020, 39, 1654–1664. [Google Scholar]
- Mousavirad, S.J.; Oliva, D.; Hinojosa, S.; Schaefer, G. Differential Evolution-based Neural Network Training Incorporating a Centroid-based Strategy and Dynamic Opposition-based Learning. In Proceedings of the 2021 IEEE Congress on Evolutionary Computation (CEC), Krakow, Poland, 28 June–1 July 2021; pp. 1233–1240. [Google Scholar]
- Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 277–281. [Google Scholar] [CrossRef] [Green Version]
- Lin, C.J.; Jeng, S.Y.; Chen, M.K. Using 2D CNN with Taguchi parametric optimization for lung cancer recognition from CT images. Appl. Sci. 2020, 10, 2591. [Google Scholar] [CrossRef] [Green Version]
- Deng, D. DBSCAN Clustering Algorithm Based on Density. In Proceedings of the 2020 7th International Forum on Electrical Engineering and Automation (IFEEA), Guangzhou, China, 18–20 December 2020; pp. 949–953. [Google Scholar]
- Wen, Y.; Zhang, K.; Li, Z.; Qiao, Y. A discriminative feature learning approach for deep face recognition. In Proceedings of the 2016 European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 8–16 October 2016; pp. 499–515. [Google Scholar]
- Fan, H.; Zhang, Y.; Hua, Z.; Li, J.; Sun, T.; Ren, M. ShapeNets: Image Representation Based on the Shape. In Proceedings of the 2016 IEEE 14th Intl Conf on Dependable, Autonomic and Secure Computing (DASC), Auckland, New Zealand, 8–12 August 2016; pp. 196–201. [Google Scholar]
- Borgefors, G. Hierarchical chamfer matching: A parametric edge matching algorithm. IEEE Trans. Pattern Anal. Mach. Intell. 1988, 10, 849–865. [Google Scholar] [CrossRef]
- Vahdat, A.; Williams, F.; Gojcic, Z.; Litany, O.; Fidler, S.; Kreis, K. LION: Latent Point Diffusion Models for 3D Shape Generation. In Proceedings of the 2022 Conference on Neural Information Processing Systems (NeurIPS), New Orleans, LA, USA, 28 November–9 December 2022; pp. 10021–10039. [Google Scholar]
- Lu, D.; Lu, X.; Sun, Y.; Wang, J. Deep feature-preserving normal estimation for point cloud filtering. Comput.-Aided Design 2020, 125, 102860. [Google Scholar] [CrossRef]
- Yan, X.; Zheng, C.; Li, Z.; Wang, S.; Cui, S. Pointasnl: Robust point clouds processing using nonlocal neural networks with adaptive sampling. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Washington, DC, USA, 14–19 June 2020; pp. 5589–5598. [Google Scholar]
Method | Avg. | Airplane | Car | Chair | Guitar | Sofa |
---|---|---|---|---|---|---|
PCN | 13.17 | 11.74 | 13.56 | 14.58 | 12.79 | 13.2 |
AtlasNet | 13.96 | 13.01 | 13.85 | 14.03 | 14.26 | 14.56 |
TopNet | 11.184 | 10.4 | 11.5 | 13.08 | 10.93 | 10.55 |
Our | 10.186 | 10.7 | 9.45 | 10.95 | 9.3 | 10.53 |
Method | Avg. | Airplane | Car | Chair | Guitar | Sofa |
---|---|---|---|---|---|---|
PCN | 11.18 | 10.2 | 10.2 | 15.22 | 10.76 | 9.53 |
AtlasNet | 12.097 | 11.16 | 10.3 | 15.1 | 10.43 | 13.4 |
Our | 10.15 | 10.15 | 9.73 | 15.6 | 9.98 | 8.41 |
Method | NOP*80% | NOP*50% | NOP*30% |
---|---|---|---|
PCN | 14.86 | 27.95 | null |
AtlasNet | 15.09 | 29.56 | null |
TopNet | 13.012 | 37.69 | null |
Our | 12.007 | 28.82 | null |
Method | CD | CD-D | CD-D 3000 | EMD | EMD-D | EMD-D 3000 |
---|---|---|---|---|---|---|
PCN | 13.174 | 12.174 | 11.174 | 11.18 | 10.25 | 10.25 |
AtlasNet | 13.96 | 15.66 | 12.66 | 12.097 | 10.21 | 9.21 |
TopNet | 11.184 | 9.98 | 9.98 | NULL | NULL | NULL |
Our | 10.186 | 8.73 | 8.34 | 10.15 | 9.43 | 9.2 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Li, J.; Si, G.; Liang, X.; An, Z.; Tian, P.; Zhou, F. Partition-Based Point Cloud Completion Network with Density Refinement. Entropy 2023, 25, 1018. https://doi.org/10.3390/e25071018
Li J, Si G, Liang X, An Z, Tian P, Zhou F. Partition-Based Point Cloud Completion Network with Density Refinement. Entropy. 2023; 25(7):1018. https://doi.org/10.3390/e25071018
Chicago/Turabian StyleLi, Jianxin, Guannan Si, Xinyu Liang, Zhaoliang An, Pengxin Tian, and Fengyu Zhou. 2023. "Partition-Based Point Cloud Completion Network with Density Refinement" Entropy 25, no. 7: 1018. https://doi.org/10.3390/e25071018