Conversion of Point Cloud Data to 3D Models Using PointNet++ and Transformer

Capa

Texto integral

Acesso aberto Acesso aberto
Acesso é fechado Acesso está concedido
Acesso é fechado Somente assinantes

Resumo

This work presents an approach to reconstructing 3D models from point cloud data, based on the use of modern neural network architectures. The basis of the method is PointNet++ and Transformer. PointNet++ plays a central role, providing efficient feature extraction and encoding of complex 3D scene geometries. This is achieved by recursively applying PointNet++ to nested partitions of the input point set in metric space. Convex decomposition, an important step in the approach, allows transforming complex three-dimensional objects into a set of simpler convex shapes. This simplifies data processing and makes the reconstruction process more manageable. The Transformer then trains the model on these features, allowing for the generation of high-quality reconstructions. It is important to note that the Transformer is used exclusively to determine the position of walls and object boundaries. This combination of technologies allows achieving high accuracy in the reconstruction of 3D models. The main idea of the method is to segment the point cloud into small fragments, which are then restored as polygonal meshes. To restore missing points in point cloud data, a method based on the L1-Median algorithm and local point cloud features is used. This approach can adapt to various geometric structures and correct topological connection errors. The proposed method was compared with several modern approaches and showed its potential in various fields, including architecture, engineering, digitization of cultural heritage, and augmented and mixed reality systems. This underscores its' wide applicability and significant potential for further development and application in various fields.

Texto integral

Acesso é fechado

Sobre autores

M. Sorokin

ITMO University

Autor responsável pela correspondência
Email: vergotten@gmail.com
Rússia, Saint Petersburg

D. Zhdanov

ITMO University

Email: ddzhdanov@mail.ru
Rússia, Saint Petersburg

A. Zhdanov

ITMO University

Email: andrew.gtx@gmail.com
Rússia, Saint Petersburg

Bibliografia

  1. Liu N., Lin B., Lv G., Zhu A.X., Zhou L. A Delaunay triangulation algorithm based on dual-spatial data organization // PFG–Journal of Photogrammetry, Remote Sensing and Geoinformation Science. 2019. P. 19–31.
  2. Ivanovsky S.A., Preobrazhensky A.S., Simonchik S.K. Algorithms for computational geometry. Convex hulls: simple algorithms // Computer tools in education. 2007. P. 3–16.
  3. Attali D., Lieutier A., Salinas L. Vietoris-Rips complexes also provide topologically correct reconstructions of sampled shapes // Proceedings of the twenty-seventh annual symposium on Computational geometry. ACM. 2011. P. 491–500.
  4. Guennebaud G., Gross M. Algebraic point set surfaces. ACM Transactions on Graphics (TOG). 2007. V. 26. No. 3. Article 23.
  5. Qi C.R., Su H., Mo K., Guibas L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation // Proceedings of the IEEE conference on computer vision and pattern recognition. IEEE. 2017. P. 652–660.
  6. Groueix T., Fisher M., Kim V.G., Russell B.C., Aubry M. Atlasnet: A papier-mâché approach to learning 3d surface generation // Proceedings of the IEEE conference on computer vision and pattern recognition. IEEE. 2018. P. 216–224.
  7. Yuan W., Khot T., Held D., Mertz C., Hebert M. Pcn: Point completion network // 2018 International Conference on 3D Vision (3DV). IEEE. 2018. https://doi.org/10.1109/3DV.2018.00088
  8. Yinyu Nie, Ji Hou, Xiaoguang Han, Matthias Nießner. RfD-Net: Point Scene Understanding by Semantic Instance Reconstruction // Computer Vision and Pattern Recognition. 2020.
  9. Ji Hou, Angela Dai, Matthias Nießner. RevealNet: Seeing Behind Objects in RGB-D Scans // Computer Vision and Pattern Recognition. 2019.
  10. Xiaoxue Chen, Hao Zhao, Guyue Zhou, Ya-Qin Zhang. PQ-Transformer: Jointly Parsing 3D Objects and Layouts from Point Clouds // Computer Vision and Pattern Recognition. 2021.
  11. Huan-ang Gao, Beiwen Tian, Pengfei Li, Xiaoxue Chen, Hao Zhao, Guyue Zhou, Yurong Chen, Hongbin Zha. From Semi-supervised to Omni-supervised Room Layout Estimation Using Point Clouds // Computer Vision and Pattern Recognition. 2023.
  12. Zhiqin Chen, Andrea Tagliasacchi, Hao Zhang. Learning Mesh Representations via Binary Space Partitioning Tree Networks // Computer Vision and Pattern Recognition. 2021.
  13. Wei Cao, Jiayi Wu, Yufeng Shi, Dong Chen. Restoration of Individual Tree Missing Point Cloud Based on Local Features of Point Cloud. Terrestrial and Mobile Mapping in Complex Indoor and Outdoor Environments. 2022.
  14. Qi C.R., Yi L., Su H., Guibas L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space // Proceedings of the Computer Vision and Pattern Recognition. 2017.
  15. Zhaofeng Niu, Yuichiro Fujimoto, Masayuki Kanbara, Taishi Sawabe, Hirokazu Kato. DFusion: Denoised TSDF Fusion of Multiple Depth Maps with Sensor Pose Noises // Computer Vision and Machine Learning for Intelligent Sensing Systems. 2022.

Arquivos suplementares

Arquivos suplementares
Ação
1. JATS XML
2. Fig. 1. ScanNet and ShapeNetv2 datasets

Baixar (337KB)
3. Fig. 2. Semantic segmentation results at the level of classes and scene instances

Baixar (768KB)
4. Fig. 3. Visualisation of layout extraction results

Baixar (104KB)
5. Рис. 4. Сравнение предсказанных (слева) и истинных (справа) 3D-моделей

Baixar (131KB)
6. Fig. 5. Formation of 3D scene model from predicted data

Baixar (191KB)
7. Fig. 6. Segmented point clouds and reconstructed 3D scene models

Baixar (542KB)

Declaração de direitos autorais © Russian Academy of Sciences, 2024