Conversion of Point Cloud Data to 3D Models Using PointNet++ and Transformer
- Autores: Sorokin M.I.1, Zhdanov D.D.1, Zhdanov A.D.1
-
Afiliações:
- ITMO University
- Edição: Nº 3 (2024)
- Páginas: 33-41
- Seção: COMPUTER GRAPHICS AND VISUALIZATION
- URL: https://modernonco.orscience.ru/0132-3474/article/view/675693
- DOI: https://doi.org/10.31857/S0132347424030044
- EDN: https://elibrary.ru/QARZUT
- ID: 675693
Citar
Resumo
This work presents an approach to reconstructing 3D models from point cloud data, based on the use of modern neural network architectures. The basis of the method is PointNet++ and Transformer. PointNet++ plays a central role, providing efficient feature extraction and encoding of complex 3D scene geometries. This is achieved by recursively applying PointNet++ to nested partitions of the input point set in metric space. Convex decomposition, an important step in the approach, allows transforming complex three-dimensional objects into a set of simpler convex shapes. This simplifies data processing and makes the reconstruction process more manageable. The Transformer then trains the model on these features, allowing for the generation of high-quality reconstructions. It is important to note that the Transformer is used exclusively to determine the position of walls and object boundaries. This combination of technologies allows achieving high accuracy in the reconstruction of 3D models. The main idea of the method is to segment the point cloud into small fragments, which are then restored as polygonal meshes. To restore missing points in point cloud data, a method based on the L1-Median algorithm and local point cloud features is used. This approach can adapt to various geometric structures and correct topological connection errors. The proposed method was compared with several modern approaches and showed its potential in various fields, including architecture, engineering, digitization of cultural heritage, and augmented and mixed reality systems. This underscores its' wide applicability and significant potential for further development and application in various fields.
Palavras-chave
Texto integral

Sobre autores
M. Sorokin
ITMO University
Autor responsável pela correspondência
Email: vergotten@gmail.com
Rússia, Saint Petersburg
D. Zhdanov
ITMO University
Email: ddzhdanov@mail.ru
Rússia, Saint Petersburg
A. Zhdanov
ITMO University
Email: andrew.gtx@gmail.com
Rússia, Saint Petersburg
Bibliografia
- Liu N., Lin B., Lv G., Zhu A.X., Zhou L. A Delaunay triangulation algorithm based on dual-spatial data organization // PFG–Journal of Photogrammetry, Remote Sensing and Geoinformation Science. 2019. P. 19–31.
- Ivanovsky S.A., Preobrazhensky A.S., Simonchik S.K. Algorithms for computational geometry. Convex hulls: simple algorithms // Computer tools in education. 2007. P. 3–16.
- Attali D., Lieutier A., Salinas L. Vietoris-Rips complexes also provide topologically correct reconstructions of sampled shapes // Proceedings of the twenty-seventh annual symposium on Computational geometry. ACM. 2011. P. 491–500.
- Guennebaud G., Gross M. Algebraic point set surfaces. ACM Transactions on Graphics (TOG). 2007. V. 26. No. 3. Article 23.
- Qi C.R., Su H., Mo K., Guibas L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation // Proceedings of the IEEE conference on computer vision and pattern recognition. IEEE. 2017. P. 652–660.
- Groueix T., Fisher M., Kim V.G., Russell B.C., Aubry M. Atlasnet: A papier-mâché approach to learning 3d surface generation // Proceedings of the IEEE conference on computer vision and pattern recognition. IEEE. 2018. P. 216–224.
- Yuan W., Khot T., Held D., Mertz C., Hebert M. Pcn: Point completion network // 2018 International Conference on 3D Vision (3DV). IEEE. 2018. https://doi.org/10.1109/3DV.2018.00088
- Yinyu Nie, Ji Hou, Xiaoguang Han, Matthias Nießner. RfD-Net: Point Scene Understanding by Semantic Instance Reconstruction // Computer Vision and Pattern Recognition. 2020.
- Ji Hou, Angela Dai, Matthias Nießner. RevealNet: Seeing Behind Objects in RGB-D Scans // Computer Vision and Pattern Recognition. 2019.
- Xiaoxue Chen, Hao Zhao, Guyue Zhou, Ya-Qin Zhang. PQ-Transformer: Jointly Parsing 3D Objects and Layouts from Point Clouds // Computer Vision and Pattern Recognition. 2021.
- Huan-ang Gao, Beiwen Tian, Pengfei Li, Xiaoxue Chen, Hao Zhao, Guyue Zhou, Yurong Chen, Hongbin Zha. From Semi-supervised to Omni-supervised Room Layout Estimation Using Point Clouds // Computer Vision and Pattern Recognition. 2023.
- Zhiqin Chen, Andrea Tagliasacchi, Hao Zhang. Learning Mesh Representations via Binary Space Partitioning Tree Networks // Computer Vision and Pattern Recognition. 2021.
- Wei Cao, Jiayi Wu, Yufeng Shi, Dong Chen. Restoration of Individual Tree Missing Point Cloud Based on Local Features of Point Cloud. Terrestrial and Mobile Mapping in Complex Indoor and Outdoor Environments. 2022.
- Qi C.R., Yi L., Su H., Guibas L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space // Proceedings of the Computer Vision and Pattern Recognition. 2017.
- Zhaofeng Niu, Yuichiro Fujimoto, Masayuki Kanbara, Taishi Sawabe, Hirokazu Kato. DFusion: Denoised TSDF Fusion of Multiple Depth Maps with Sensor Pose Noises // Computer Vision and Machine Learning for Intelligent Sensing Systems. 2022.
Arquivos suplementares
