1
|
Wang H, Tian Y. Sequential Point Clouds: A Survey. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2024; 46:5504-5523. [PMID: 38354073 DOI: 10.1109/tpami.2024.3365970] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/16/2024]
Abstract
Point clouds have garnered increasing research attention and found numerous practical applications. However, many of these applications, such as autonomous driving and robotic manipulation, rely on sequential point clouds, essentially adding a temporal dimension to the data (i.e., four dimensions) because the information of the static point cloud data could provide is still limited. Recent research efforts have been directed towards enhancing the understanding and utilization of sequential point clouds. This paper offers a comprehensive review of deep learning methods applied to sequential point cloud research, encompassing dynamic flow estimation, object detection & tracking, point cloud segmentation, and point cloud forecasting. This paper further summarizes and compares the quantitative results of the reviewed methods over the public benchmark datasets. Ultimately, the paper concludes by addressing the challenges in current sequential point cloud research and pointing towards promising avenues for future research.
Collapse
|
2
|
Li J, Cheng B, Niu N, Gao G, Ying S, Shi J, Zeng T. A fine-grained orthodontics segmentation model for 3D intraoral scan data. Comput Biol Med 2024; 168:107821. [PMID: 38064844 DOI: 10.1016/j.compbiomed.2023.107821] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/27/2023] [Revised: 11/01/2023] [Accepted: 12/04/2023] [Indexed: 01/10/2024]
Abstract
With the widespread application of digital orthodontics in the diagnosis and treatment of oral diseases, more and more researchers focus on the accurate segmentation of teeth from intraoral scan data. The accuracy of the segmentation results will directly affect the follow-up diagnosis of dentists. Although the current research on tooth segmentation has achieved promising results, the 3D intraoral scan datasets they use are almost all indirect scans of plaster models, and only contain limited samples of abnormal teeth, so it is difficult to apply them to clinical scenarios under orthodontic treatment. The current issue is the lack of a unified and standardized dataset for analyzing and validating the effectiveness of tooth segmentation. In this work, we focus on deformed teeth segmentation and provide a fine-grained tooth segmentation dataset (3D-IOSSeg). The dataset consists of 3D intraoral scan data from more than 200 patients, with each sample labeled with a fine-grained mesh unit. Meanwhile, 3D-IOSSeg meticulously classified every tooth in the upper and lower jaws. In addition, we propose a fast graph convolutional network for 3D tooth segmentation named Fast-TGCN. In the model, the relationship between adjacent mesh cells is directly established by the naive adjacency matrix to better extract the local geometric features of the tooth. Extensive experiments show that Fast-TGCN can quickly and accurately segment teeth from the mouth with complex structures and outperforms other methods in various evaluation metrics. Moreover, we present the results of multiple classical tooth segmentation methods on this dataset, providing a comprehensive analysis of the field. All code and data will be available at https://github.com/MIVRC/Fast-TGCN.
Collapse
Affiliation(s)
- Juncheng Li
- School of Communication Information Engineering, Shanghai University, Shanghai, China.
| | - Bodong Cheng
- School of Computer Science and Technology, East China Normal University, Shanghai, China.
| | - Najun Niu
- School of Stomatology, Nanjing Medical University, Nanjing, China.
| | - Guangwei Gao
- Institute of Advanced Technology, Nanjing University of Posts and Telecommunications, Nanjing, China.
| | - Shihui Ying
- Department of Mathematics, School of Science, Shanghai University, Shanghai, China.
| | - Jun Shi
- School of Communication Information Engineering, Shanghai University, Shanghai, China.
| | - Tieyong Zeng
- Department of Mathematics, The Chinese University of Hong Kong, New Territories, Hong Kong.
| |
Collapse
|
3
|
Wu L, Hou Y, Xu J, Zhao Y. Robust Mesh Segmentation Using Feature-Aware Region Fusion. SENSORS (BASEL, SWITZERLAND) 2022; 23:416. [PMID: 36617011 PMCID: PMC9824490 DOI: 10.3390/s23010416] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/30/2022] [Revised: 12/19/2022] [Accepted: 12/28/2022] [Indexed: 06/17/2023]
Abstract
This paper introduces a simple but powerful segmentation algorithm for 3D meshes. Our algorithm consists of two stages: over-segmentation and region fusion. In the first stage, adaptive space partition is applied to perform over-segmentation, which is very efficient. In the second stage, we define a new intra-region difference, inter-region difference, and fusion condition with the help of various shape features and propose an iterative region fusion method. As the region fusion process is feature aware, our algorithm can deal with complex 3D meshes robustly. Massive qualitative and quantitative experiments also validate the advantages of the proposed algorithm.
Collapse
Affiliation(s)
- Lulu Wu
- School of Mathematical Sciences, Ocean University of China, Qingdao 266100, China
| | - Yu Hou
- School of Mathematical Sciences, Ocean University of China, Qingdao 266100, China
| | - Junli Xu
- School of Mathematics and Physics, Qingdao University of Science and Technology, Qingdao 266061, China
| | - Yong Zhao
- School of Mathematical Sciences, Ocean University of China, Qingdao 266100, China
| |
Collapse
|
4
|
Liu Z, Tang H, Zhao S, Shao K, Han S. PVNAS: 3D Neural Architecture Search With Point-Voxel Convolution. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2022; 44:8552-8568. [PMID: 34469291 DOI: 10.1109/tpami.2021.3109025] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
3D neural networks are widely used in real-world applications (e.g., AR/VR headsets, self-driving cars). They are required to be fast and accurate; however, limited hardware resources on edge devices make these requirements rather challenging. Previous work processes 3D data using either voxel-based or point-based neural networks, but both types of 3D models are not hardware-efficient due to the large memory footprint and random memory access. In this paper, we study 3D deep learning from the efficiency perspective. We first systematically analyze the bottlenecks of previous 3D methods. We then combine the best from point-based and voxel-based models together and propose a novel hardware-efficient 3D primitive, Point-Voxel Convolution (PVConv). We further enhance this primitive with the sparse convolution to make it more effective in processing large (outdoor) scenes. Based on our designed 3D primitive, we introduce 3D Neural Architecture Search (3D-NAS) to explore the best 3D network architecture given a resource constraint. We evaluate our proposed method on six representative benchmark datasets, achieving state-of-the-art performance with 1.8-23.7× measured speedup. Furthermore, our method has been deployed to the autonomous racing vehicle of MIT Driverless, achieving larger detection range, higher accuracy and lower latency.
Collapse
|
5
|
Zhang C, Wan H, Shen X, Wu Z. PVT: Point‐voxel transformer for point cloud learning. INT J INTELL SYST 2022. [DOI: 10.1002/int.23073] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Affiliation(s)
- Cheng Zhang
- School of Media and Design Hangzhou Dianzi University Hangzhou China
| | - Haocheng Wan
- School of Media and Design Hangzhou Dianzi University Hangzhou China
| | - Xinyi Shen
- Department of Medical Physics and Bioengineering University College London London UK
| | - Zizhao Wu
- School of Media and Design Hangzhou Dianzi University Hangzhou China
| |
Collapse
|
6
|
Zhao Y, Zhang L, Liu Y, Meng D, Cui Z, Gao C, Gao X, Lian C, Shen D. Two-Stream Graph Convolutional Network for Intra-Oral Scanner Image Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:826-835. [PMID: 34714743 DOI: 10.1109/tmi.2021.3124217] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Precise segmentation of teeth from intra-oral scanner images is an essential task in computer-aided orthodontic surgical planning. The state-of-the-art deep learning-based methods often simply concatenate the raw geometric attributes (i.e., coordinates and normal vectors) of mesh cells to train a single-stream network for automatic intra-oral scanner image segmentation. However, since different raw attributes reveal completely different geometric information, the naive concatenation of different raw attributes at the (low-level) input stage may bring unnecessary confusion in describing and differentiating between mesh cells, thus hampering the learning of high-level geometric representations for the segmentation task. To address this issue, we design a two-stream graph convolutional network (i.e., TSGCN), which can effectively handle inter-view confusion between different raw attributes to more effectively fuse their complementary information and learn discriminative multi-view geometric representations. Specifically, our TSGCN adopts two input-specific graph-learning streams to extract complementary high-level geometric representations from coordinates and normal vectors, respectively. Then, these single-view representations are further fused by a self-attention module to adaptively balance the contributions of different views in learning more discriminative multi-view representations for accurate and fully automatic tooth segmentation. We have evaluated our TSGCN on a real-patient dataset of dental (mesh) models acquired by 3D intraoral scanners. Experimental results show that our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
Collapse
|
7
|
Zhao Y, Zhang L, Yang C, Tan Y, Liu Y, Li P, Huang T, Gao C. 3D Dental model segmentation with graph attentional convolution network. Pattern Recognit Lett 2021. [DOI: 10.1016/j.patrec.2021.09.005] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
8
|
Guo Y, Wang H, Hu Q, Liu H, Liu L, Bennamoun M. Deep Learning for 3D Point Clouds: A Survey. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2021; 43:4338-4364. [PMID: 32750799 DOI: 10.1109/tpami.2020.3005434] [Citation(s) in RCA: 202] [Impact Index Per Article: 50.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Point cloud learning has lately attracted increasing attention due to its wide applications in many areas, such as computer vision, autonomous driving, and robotics. As a dominating technique in AI, deep learning has been successfully used to solve various 2D vision problems. However, deep learning on point clouds is still in its infancy due to the unique challenges faced by the processing of point clouds with deep neural networks. Recently, deep learning on point clouds has become even thriving, with numerous methods being proposed to address different problems in this area. To stimulate future research, this paper presents a comprehensive review of recent progress in deep learning methods for point clouds. It covers three major tasks, including 3D shape classification, 3D object detection and tracking, and 3D point cloud segmentation. It also presents comparative results on several publicly available datasets, together with insightful observations and inspiring future research directions.
Collapse
|
9
|
Chen XT, Li Y, Fan JH, Wang R. RGAM: A novel network architecture for 3D point cloud semantic segmentation in indoor scenes. Inf Sci (N Y) 2021. [DOI: 10.1016/j.ins.2021.04.069] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
10
|
|
11
|
Point Cloud Semantic Segmentation Using a Deep Learning Framework for Cultural Heritage. REMOTE SENSING 2020. [DOI: 10.3390/rs12061005] [Citation(s) in RCA: 27] [Impact Index Per Article: 5.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
In the Digital Cultural Heritage (DCH) domain, the semantic segmentation of 3D Point Clouds with Deep Learning (DL) techniques can help to recognize historical architectural elements, at an adequate level of detail, and thus speed up the process of modeling of historical buildings for developing BIM models from survey data, referred to as HBIM (Historical Building Information Modeling). In this paper, we propose a DL framework for Point Cloud segmentation, which employs an improved DGCNN (Dynamic Graph Convolutional Neural Network) by adding meaningful features such as normal and colour. The approach has been applied to a newly collected DCH Dataset which is publicy available: ArCH (Architectural Cultural Heritage) Dataset. This dataset comprises 11 labeled points clouds, derived from the union of several single scans or from the integration of the latter with photogrammetric surveys. The involved scenes are both indoor and outdoor, with churches, chapels, cloisters, porticoes and loggias covered by a variety of vaults and beared by many different types of columns. They belong to different historical periods and different styles, in order to make the dataset the least possible uniform and homogeneous (in the repetition of the architectural elements) and the results as general as possible. The experiments yield high accuracy, demonstrating the effectiveness and suitability of the proposed approach.
Collapse
|