Next Article in Journal
Ultrasound Imaging Comparison of Crural Fascia Thickness and Muscle Stiffness in Stroke Patients with Spasticity
Previous Article in Journal
A Novel Method for 3D Lung Tumor Reconstruction Using Generative Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale 3D Cephalometric Landmark Detection Based on Direct Regression with 3D CNN Architectures

1
Medical Device Development Center, Daegu-Gyeongbuk Medical Innovation Foundation (K-MEDI Hub), Daegu 41061, Republic of Korea
2
Daegu-Gyeongbuk Research Division, Electronics and Telecommunications Research Institute (ETRI), Daegu 42994, Republic of Korea
3
Department of Oral and Maxillofacial Surgery, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul 06351, Republic of Korea
*
Authors to whom correspondence should be addressed.
Diagnostics 2024, 14(22), 2605; https://doi.org/10.3390/diagnostics14222605
Submission received: 9 September 2024 / Revised: 13 November 2024 / Accepted: 19 November 2024 / Published: 20 November 2024
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Background: Cephalometric analysis is important in diagnosing and planning treatments for patients, traditionally relying on 2D cephalometric radiographs. With advancements in 3D imaging, automated landmark detection using deep learning has gained prominence. However, 3D imaging introduces challenges due to increased network complexity and computational demands. This study proposes a multi-scale 3D CNN-based approach utilizing direct regression to improve the accuracy of maxillofacial landmark detection. Methods: The method employs a coarse-to-fine framework, first identifying landmarks in a global context and then refining their positions using localized 3D patches. A clinical dataset of 150 CT scans from maxillofacial surgery patients, annotated with 30 anatomical landmarks, was used for training and evaluation. Results: The proposed method achieved an average RMSE of 2.238 mm, outperforming conventional 3D CNN architectures. The approach demonstrated consistent detection without failure cases. Conclusions: Our multi-scale-based 3D CNN framework provides a reliable method for automated landmark detection in maxillofacial CT images, showing potential for other clinical applications.

1. Introduction

Cephalometric analysis involves the quantitative measurements of anatomical landmarks in the craniofacial region and plays an important role in patient diagnosis and surgical planning. Traditionally, cephalometric analysis has been conducted with 2D cephalometric radiographs using lateral view analysis. However, with advancements in computer vision technology, 3D imaging techniques such as computed tomography (CT) and cone beam computed tomography (CBCT) are being applied, leading to research on 3D cephalometric analysis.
Cephalometric analysis is conducted by detecting craniofacial anatomical landmarks and performing linear and angular measurements between the landmarks. In the process of landmark analysis, the accurate detection of anatomical landmarks is an essential step for accurate analysis. However, conventional landmark detection involves time-consuming manual measurement by medical professionals, which are prone to measurement variability [1,2]. To address these issues, several studies reported an automated detection of anatomical landmarks using deep learning technologies, which improved the accuracy of cephalometric landmark detection.
Automatic 2D cephalometric measurement focuses on landmark detection in lateral radiographs, with various methods developed to enhance accuracy. Yao et al. applied self-supervised learning with a multi-task U-Net, leveraging cascade cosine similarity maps and pseudo-labels [1]. Chen et al. developed a semi-supervised approach using shape-regulated self-training with U-Net and the spatial configuration network (SCN) [2]. Zhu et al. proposed a universal anatomical landmark detection model consisting of a local network and a global network [3]. Yao et al. proposed a cephalometric landmark detection network comprising a global module and a local module based on a U-Net architecture [4].
While 2D cephalograms remain commonly used in medical practice, which contains clear limitations, such as the superimposition of anatomical structures, 3D cephalometric imaging (e.g., CT, CBCT) provides detailed access to complex craniofacial and maxillomandibular structures, enabling precise analysis. However, the increased dimensionality of 3D imaging introduces more trainable parameters and network complexity in deep learning-based approaches. Consequently, training with 3D images presents challenges, and various approaches have been developed to address these difficulties in 3D cephalometric landmark detection.
A method for automatic 3D landmark detection is the 2.5-dimensional approach, which involves learning slices of 3D images. Ma et al. proposed a patch-based deep learning model that uses 2D convolution on extracted patches to achieve 3D landmark extraction [5]. Dot et al. developed a model incorporating a coarse-to-fine structure into the SCN [6]. This approach detects landmark regions of interest (ROIs) through coarse detection, and then performs fine detection using heatmaps within each ROI to extract 3D landmarks.
A widely used approach for 3D landmark detection is to localize landmark coordinates using regression methods with the 3D convolutional filter [7]. Zheng et al. proposed a method combining 3D CNNs with a bootstrapped detector, leveraging Haar wavelet features for landmark detection [8]. Yun et al. proposed a coarse-to-fine structure integrated with a variational auto-encoder (VAE) to detect 93 3D landmarks [9]. Liu et al. reported a multi-stage CNN model that employs a 3D U-Net for cephalometric landmark detection and bone segmentation, treating these tasks as voxel classification [10]. Nishimoto et al. proposed a multi-phased deep regression model that uses progressively smaller cropped voxels to achieve refined localization [11].
Recent studies have introduced 3D detection methods to simultaneously address 3D localization and classification problems. These studies commonly employ two-stage networks, which include a region proposal network (RPN) and Bounding Box Recognition, to identify regions containing landmarks and estimate the coordinates. Lang et al. reported a method for automatic landmark detection applying a 3D R-CNN [12]. The approach applied a 3D Mask R-CNN with a coarse-to-fine structure considering three stages of low- and high-resolution images. A root/leaf structured loss function was introduced to group anatomical landmarks by dividing them into nine geometric regions, and the method was validated using CBCT images from 49 patients. Chen et al. proposed a landmark detection method using 3D Faster R-CNN and 3D U-Net [13]. By applying a coarse-to-fine structure, the method first identifies approximate locations within CBCT images and then uses a multi-scale U-Net to detect landmark coordinates. Heatmap regression was used to train the U-Net, and the proposed method was validated using 80 sets of clinical CBCT data. Sahlsten et al. implemented a deep learning method using a stacked hourglass network to automate the identification of 3D cephalometric landmarks [14]. This single-stage model, built on a stacked hourglass network architecture, incorporated 3D convolutional layers optimized for computational efficiency. Model performance was evaluated based on the mean Euclidean distance and success detection rate across 46 cephalometric landmarks, using a dataset of 309 CBCT scans.
Previous studies have proposed methods for automatically detecting anatomical landmarks in 3D medical images. These methods are based on training 3D convolutional filters in deep learning and aim to improve accuracy through the application of auto-encoders, detection, segmentation, or stacked hourglass network architectures. Studies applying RCNN-type networks have shown relatively high accuracy by incorporating region proposal networks or heatmap regression. This study employed a stacked hourglass network architecture that achieved an accuracy of less than 2 mm. However, these approaches can still result in failure cases when detecting cephalometric anatomical landmarks, which may limit their applicability in clinical settings. In contrast, direct regression methods can produce trained landmarks without failure cases, but the accuracy required for clinical application has not yet been reported.
In this paper, a method for 3D cephalometric landmark detection by applying a coarse-to-fine structure into direct regression is presented. The proposed method employs multi-scale direct regression with a 3D CNN to extract 3D local patches for each landmark from the global maxillofacial region, followed by refining the landmark coordinates within local patches. The optimal base network was identified through performance comparisons of previous 3D CNN architectures for direct regression, and the proposed method was validated using 150 clinical datasets. The contributions of this study are summarized as follows:
(1)
Performance comparison of 3D CNN architectures for landmark direct regression;
(2)
Development of an automatic cephalometric landmark detection network using multi-scale direct regression;
(3)
Validation using 150 sets of maxillofacial clinical data (CT).

2. Methods

2.1. Method Overview

A multi-scale cephalometric landmark detection method based on direct regression with 3D CNN aims to enhance detection accuracy through a hierarchical learning strategy on multi-scale volumes. This method includes coarse detection to extract the class and coordinates of all landmarks from the global input volume, 3D region of interest (ROI) processing to transform these results into local volumes, and fine localization to precisely determine the landmark coordinate within a local volume, as illustrated in Figure 1.
Stage 1 involves coarse detection to determine the coordinates and classes of landmarks across the entire input volume, resulting in the identification of classes and coordinates for 30 anatomical landmarks. Subsequently, 3D ROI processing is performed to generate local volumes corresponding to each landmark. Stage 2 focuses on fine localization, using the local volumes and classes obtained from Stage 1. Sub-sampled local volumes at a scale of 64 × 64 × 64 with the corresponding class are used to precisely localize landmark coordinates at the local scale. Further details on the networks employed in each stage are discussed in Section 2.5.

2.2. Data Description

CT or CBCT imaging is a widely accepted modality for modern maxillofacial diagnosis and treatment planning. Clinical maxillofacial CT images from 150 patients between 1 January 2010 and 31 July 2022 were collected retrospectively at Samsung Medical Center, with approval from the Institutional Review Board of the Department of Oral and Maxillofacial Surgery, Samsung Medical Center, Seoul, Korea (SMC 2022-08-091-001). Data from patients with severe diseases, those still in growth, and those with craniofacial anomalies were excluded to ensure the dataset was suitable for normal oral and maxillofacial diagnosis and treatment planning.
Collected data were anonymized and processed to create the medical imaging dataset. The dataset comprised only the 3D reconstructed volume data of the skull, excluding facial soft tissues. Volumes were generated by segmenting the maxillofacial skull regions, and all patient metadata were removed for anonymization. For supervised learning, 3D anatomical landmark data were established as ground truths. The ground truth consisted of 30 anatomical landmarks: (11apex, 13, 16, 21apex, 23, 26, A_point, ANS, B, Ba, CoL, CoR, Gn, GoL, GoR, L1, Lt_FZS, Menton, Na, Nasaltip, OrL, OrR, PNS, PoL, PoR, Pog, Rt_FZS, S, Sp, U1). Clinical experts at Samsung Medical Center manually identified 30 anatomical landmarks used in maxillofacial surgical planning, and 3D coordinates for these landmarks were annotated on each anonymized maxillofacial volume.

2.3. Data Preprocessing

Data preprocessing is required to convert maxillofacial volumetric data (or segmented mesh data) into a normalized format suitable for 3D CNN application. In prior studies [7,8,9,10,11,12,13], voxel data have been used as the input format for the training of inference of 3D CNNs. In this study, maxillofacial data were provided in mesh format (e.g., STL) due to the internal policy for personal medical information protection. Consequently, a preprocessing algorithm was developed and applied to prepare the maxillofacial volume data for CNN training, as shown in Figure 2.
The preprocessing algorithm was developed using Open3D [15] and the Visualization Toolkit [16] open-source libraries. To standardize the coordinate system for the maxillofacial volumetric data, the origin was set at the front bottom left point of a bounding cube encompassing the 3D data. To expedite the voxelization process, uniform sampling was applied to select 50,000 points. The sampled volume was voxelized using the Open3D library, and the data were normalized to the scale of 128 × 128 × 128 voxels.

2.4. Three-Dimensional CNN Architectures

Convolutional Neural Networks (CNNs) are deep learning architectures optimized for feature extraction and regression in spatially structured data, such as images. Consisting of multiple layers (e.g., convolutional, pooling, Fully Connected Layers), CNNs automatically extract and learn features from low to high levels, making them effective for object recognition and classification.
Three-dimensional CNNs extend conventional CNNs to handle three-dimensional data. By using 3D Convolution and 3D pooling, these networks can learn from additional spatial or temporal dimensions and are applied in areas like robotics, autonomous driving, and medical imaging (e.g., CT, MRI). Three-dimensional CNNs have also been employed in diagnostics, particularly for predicting protein–ligand interactions [17,18,19] and assessing the binding affinity between RNA and small molecules [20]. Notably, 3D CNNs have shown promising results in medical imaging tasks such as tumor detection, classification, and tissue segmentation using 3D images such as CT, MRI, and PET [21,22,23,24].
In this study, 3D CNNs were applied to solve detection tasks in 3D maxillofacial images via voxel-wise regression. Three-dimensional CNN architectures, including ResNet, DenseNet, Inception, and InceptionResNet, as shown in Figure 3, were trained on 3D maxillofacial images to assess landmark detection accuracy. We trained and evaluated each network using the maxillofacial clinical dataset and compared the landmark detection accuracy to select the optimal network for 3D maxillofacial landmark detection.

2.4.1. ResNet Architecture

ResNet is a model that introduced residual learning to overcome the degradation of learning performance associated with increased network depth [25]. Typically, deepening a neural network to solve complex problems leads to a significant increase in trainable parameters, causing the gradient vanishing problem. Residual learning addresses this by having each layer learn a residual function F(x) = H(x) − x instead of the direct mapping H(x). By approximating the residuals to zero, optimization becomes easier, allowing identity mapping without adding parameters or computational complexity. This approach mitigates the gradient vanishing issue, enabling the training of deeper networks. ResNet is available in configurations with various depths, such as 50, 101, or 152 layers, and its effectiveness has been validated using datasets like ImageNet and CIFAR-10.

2.4.2. DenseNet Architecture

DenseNet (Densely Connected Convolutional Networks) is a CNN architecture designed to maximize feature map reuse by connecting all layers in a feed-forward fashion [26]. In DenseNet, each layer receives input from all preceding layers and passes its output to all subsequent layers, creating a dense connectivity pattern within dense blocks. This approach enhances feature integration and reuse, differing from ResNet by connecting features rather than summing them. A network with L layers will have L (L + 1)/2 direct connections. To handle the increasing number of feature maps, the growth rate k is set as a hyper-parameter, typically k = 12, allowing for narrower layers and fewer parameters. Dense blocks reduce network complexity and alleviate gradient vanishing through feature reuse. By stacking dense blocks, structural expansion is possible, leading to improved performance as network depth increases. The effectiveness of the architecture has been validated using object recognition benchmarks such as CIFAR-10, CIFAR-100, SVHN, and ImageNet.

2.4.3. Inception Architecture

Inception networks aim for high performance and computational efficiency by incorporating dimensional reduction and parallel structures [27]. The key element of the Inception architecture is the Inception module, which simultaneously utilizes multiple filter sizes, such as 1 × 1, 3 × 3, and 5 × 5 convolution layers, combined with pooling layers. This design allows for balanced expansion in both depth and width, facilitating the extraction and reuse of diverse features through parallel convolutional filters. The use of 1 × 1 convolutions within these parallel structures also aids in dimensional reduction, enhancing computational efficiency. Inception-v3 introduced additional optimization techniques based on the Inception module. By stacking 3 × 3 convolution layers, as in VGGNet [28], an effective 7 × 7 convolution layer was created, further improving computational efficiency. Label smoothing was also employed to reduce overfitting and enhance generalization performance. Inception networks have shown high performance on large-scale image classification benchmarks like ImageNet, demonstrating their versatility across various computer vision tasks.

2.4.4. InceptionResNet Architecture

InceptionResNet is a network model that combines the Inception module, which uses multiple filter sizes simultaneously, with identity mapping via shortcut connections [29]. The main feature of the InceptionResNet architecture is the addition of residual connections following the Inception module, which uses parallel combinations of filters. This setup allows the network to perform residual learning on the Inception module’s parallel structures. By adjusting the activation scaling value of the residual connections, the model’s training stability is improved, alleviating gradient vanishing and increasing the learning speed. InceptionResNet has demonstrated high performance on the ImageNet dataset and has shown that applying residual connections can significantly enhance training speed.

2.5. Proposed Architecture

In order to achieve multi-scale 3D cephalometric landmark detection based on direct regression with the 3D CNN, we sought to experimentally identify the suitable base architecture for landmark detection from existing 3D CNN structures. A comparative experiment was conducted to identify the optimal CNN architecture for cephalometric landmark detection, with detailed results provided in Section 3. Among the existing CNN architectures, the 169-layer DenseNet demonstrated the lowest error in landmark detection experiments using the maxillofacial clinical dataset. Consequently, DenseNet169 was selected as the base architecture for developing the multi-scale CNN structure in this study.
For the implementation of multi-scale direct regression, a coarse detection architecture was designed to detect the coordinates and classes of all landmarks across the entire volume, while a fine localization architecture was developed to precisely localize individual landmark coordinates within local regions, as depicted in Figure 4 and Figure 5. This study used 30 landmarks for training and inference. In the coarse detection, all landmarks were regressed simultaneously to generate 30 local volumes and classes. The fine localization then used these local volumes and classes to localize landmark coordinates at a local scale.

2.5.1. Coarse Detection

Coarse detection utilizes a multi-output learning model to extract the coordinates and classes of all 30 landmarks from a scale of 128 × 128 × 128 input volume. The model is designed to learn the coordinates and classes of the 30 landmarks simultaneously. To address the regression problem of landmark coordinates, mean absolute error (MAE) loss and mean squared error (MSE) metrics are employed. For classifying landmark classes, binary cross-entropy loss and binary accuracy metrics are used. The loss weights for the multi-output model are set to 0.9 for regression and 0.1 for classification to ensure balanced learning for both outputs. With this network configuration, the coarse detection structure processes a scale of 128 × 128 × 128 input volume and outputs the 30 × 3 coordinates and 30 × 30 classes for the 30 landmarks.

2.5.2. Three-Dimensional Region of Interest (ROI) Processing

Coarse detection provides the coordinates of 30 anatomical landmarks and their corresponding classes from the maxillofacial image. In this study, local patches were generated for multi-scale learning based on the coarse detection results by creating 3D ROIs using the inferred landmark coordinates. To create these 3D ROIs, each of the 30 landmarks identified in the 128 × 128 × 128 input volume was used as a central coordinate. A local volume with dimensions of 64 along the x, y, and z axes was then defined around each landmark. As a result, 30 sets of 64 × 64 × 64 local volumes were generated from an input volume, each centered on a landmark, to support local learning.

2.5.3. Fine Localization

The objective of fine localization is to precisely localize the coordinates of landmarks within a local volume that includes the landmark, focusing on the local-scale features. To accomplish this, a multi-input learning model was employed, using the local volume and corresponding class from the coarse detection results to infer the final landmark coordinates. The fine localization network uses two inputs—a 64 × 64 × 64 local volume and the corresponding class—and it outputs 3 × 1 fine coordinates. The local volume is processed through dense blocks and then concatenated with the class input before the final dense layer for learning. Since the task is to predict landmark coordinates, a regression approach was applied using mean absolute error (MAE) as the loss function and mean squared error (MSE) as the metric. With this network configuration, the fine localization network utilizes the multi-inputs of local volume and class to localize coordinates within the local region. The resulting coordinates are initially specific to the local region but are converted to the global scale to serve as the final landmark coordinates.

3. Experimental Results

This section presents a comparative experiment with existing 3D CNN architectures and the performance evaluation of the proposed method. We assessed 3D CNN models, including ResNet [25], DenseNet [26], Inceptionv3 [27], and Inception-ResNetV2 [29], for their performance in 3D cephalometric landmark detection. Additionally, we tested the landmark detection capabilities of the proposed method.
The 150 sets of maxillofacial clinical CT images from Samsung Medical Center, mentioned in Section 2.2, were normalized and used for network training and evaluation. Of the 150 datasets, 120 were used for training and 30 were used for validation and testing.
Each CNN model was trained with the following hyper-parameters: Adam optimizer, MSE loss function, batch size of 41,000 epochs, and a learning rate of 0.001. The proposed method involved two training phases: coarse (using Adam optimizer, MAE-BinaryCrossEntropy loss function, batch size of 41,000 epochs, learning rate of 0.001) and fine (using Adam optimizer, MAE loss function, batch size of 15,1000 epochs, learning rate of 0.001).
Experiments were conducted on a Precision 7920 workstation (Dell, Round Rock, TX, USA) with Microsoft Windows 10 Pro for Workstations, an Intel Xeon Gold 5218R CPU, and an NVIDIA RTX A6000 GPU, using Python 3.9.16, Keras 2.10.0, Open3D 0.17.0 [15], and Visualization Toolkit 9.2.6 [16].
Table 1 compares the network parameters, accuracy, and inference time for maxillofacial landmark detection. Accuracy is measured as the mean and variance of the root-mean-square error (RMSE) between network predictions and the ground truth. Parameters indicate the number of trainable parameters, and times represent the average inference time per prediction. Regarding the accuracy of existing CNN architectures, Inception-ResNetV2 had the highest error at 2.978 ± 0.493 mm, while DenseNet169 had the lowest error at 2.464 ± 0.382 mm. The proposed method achieved an error of 2.238 ± 0.364 mm, reducing the error by approximately 0.74 mm compared to Inception-ResNetV2 and by about 0.226 mm compared to DenseNet169. Table 2 shows the RMSE for each of the 30 maxillofacial landmarks across different networks. The proposed method achieved the lowest error of 1.457 ± 0.674 mm for landmark GN and the highest error of 3.038 ± 1.862 mm for landmark GoL.

4. Discussion

In this study, a multi-scale 3D CNN architecture for automatic cephalometric landmark detection is proposed. This approach employs a coarse-to-fine strategy, performing global-scale detection followed by local-scale localization to automatically extract landmarks. The coarse-to-fine architecture enables the use of both global-scale and local-scale features, allowing for localization refinement by subdividing the volume scale. In the proposed method, multi-output learning and multi-input learning structures were applied. This approach enabled feature learning across the entire maxillofacial volume to regress landmark coordinates and classes, while also allowing for local refinement within a subdivided volume scale in the local region. Unlike previous landmark detection studies [5,6,7,8,9,10,11,12,13], to the best of the authors’ knowledge, this is the first research integrates both a coarse-to-fine structure and direct regression.
To validate the proposed method, 150 sets of clinical CT data and 30 anatomical landmarks identified by clinical experts were used. The network’s ability to detect cephalometric landmarks was evaluated by comparing the RMSE between the network-predicted coordinates and those determined by experts. Comparative experiments with existing CNN architectures revealed that the proposed method achieved 2.238 mm of the lowest average accuracy and 0.364 mm of the lowest standard deviation. In comparison to conventional CNN models, our results show an accuracy improvement of approximately 25% over Inception-ResNetV2, 23% over InceptionV3, 20% over ResNet152, and 9% over DenseNet169. In the proposed method, DenseNet169 was applied as the backbone network for feature reuse, suggesting that the application of a multi-stage structure contributed to a 9% improvement in accuracy. An individual error comparison for the 30 cephalometric anatomical landmarks showed that the proposed method achieved the lowest error for 24 landmarks, followed by DenseNet201, which recorded the lowest error for two landmarks. DenseNet169, ResNet152, ResNet101, and ResNet50 each recorded the lowest error for one landmark. Among the 3D CNN structures, deep learning architectures designed for feature reuse (e.g., DenseNet) were found to potentially lower the accuracy of landmark detection.
Objective accuracy comparisons across 3D cephalometric landmark detection studies are challenging due to variations in clinical data (e.g., the number of images and landmarks). Some state-of-the-art studies [12,13] reported ~1 mm accuracy with 80 CBCT scans and 18 landmarks, and ~2 mm accuracy with 49 CBCT scans and 105 landmarks. These studies have the advantage of generating ROIs through the network, enabling precise localization refinement. However, they use region-proposal-based approaches, resulting in failure cases, with reported accuracies excluding these failure cases. In contrast, the proposed method, utilizing multi-scale direct regression for landmark inference, achieved a ~2.24 mm error without failure cases.
In the design of the multi-scale detection architecture, the trainable parameter was achieved at 39,056,605, with an average inference time of 3.835 s. In the process of designing and training deep learning networks, trainable parameters are directly related to computational cost, training time, and network complexity. From this perspective, the DenseNet169 backbone network was chosen in the proposed method for its efficient feature reuse capability, which minimizes the number of trainable parameters and enables deeper learning without adding a significant computational burden. The two-stage learning strategy increases network complexity; although the trainable parameters have increased compared to the base architecture (DenseNet169), it remains lower than that of the shallower ResNet50 architecture. The inference time, representing the time taken for a single prediction, was higher compared to existing 3D CNN architectures. Since the goal of this study is to apply the method in pre-operative procedures such as patient diagnosis or surgical planning, prioritizing higher accuracy over reduced inference time will improve its clinical applicability.
For clinical applicability, cephalometric analysis generally requires an accuracy below 2 mm [1,2,3], indicating the need for further accuracy enhancements in future research. Specifically, achieving a detection accuracy below 2 mm is important for clinical acceptance, but the absence of detection failures is also crucial to maintain clinical viability. In this study, we aimed to achieve landmark detection accuracy without any failure cases to ensure this viability. Furthermore, the capability to detect 30 landmarks without failure cases suggests potential clinical utility, especially if post-processing steps are introduced to allow clinicians to refine landmarks.
In deep learning, model generalization is important for applicability across diverse clinical environments, making the construction of high-quality, massive datasets an important task. Building those clinical datasets for research presents practical challenges, such as the volume of data accumulated at hospitals and patient protection policies. Particularly in medical fields with relatively low demand, building extensive datasets can remain a challenging issue. In this research, due to these limitations, 150 CT datasets were used to implement the proposed method. However, in future studies, methods such as federated learning, which enables data collaboration across multiple hospitals, may provide a viable solution for these limitations.
Some limitations remain. First, although the multi-scale design enhances accuracy, it also increases inference time due to the multi-stage detection process. For applications in broader fields, especially during surgical procedures, further optimization may be needed to reduce computation time. Secondly, this study focused only on the skull region of the maxillofacial volume, excluding surrounding soft tissues to ensure patient privacy. This approach required additional preprocessing steps for anonymization and limited the ability to learn features of soft tissues.
Considering recent advancements in deep learning model structures, there remains potential to further enhance accuracy in future research. In this study, a 3D CNN architecture and a coarse-to-fine strategy were applied to improve landmark detection accuracy. The coarse-to-fine structure has been reported to enhance localization accuracy [12,13], a result also demonstrated in this study for cephalometric landmark detection. However, studies have been published that advance conventional CNNs, and future research could consider applying Vision Transformer (ViT) [30] or stacked hourglass architecture [14] to further improve accuracy.
This study focused on using only the skull from maxillofacial images, requiring segmentation to isolate the skull for deep learning network learning. To ensure patient privacy, soft tissues surrounding the skull were removed during the anonymization process and were not used in this study. Future research aims to develop deep learning networks that can directly utilize clinical data by implementing training servers within healthcare institutions and exporting only the trained models (e.g., deep network weights).

5. Conclusions

In this paper, a multi-scale image-based CNN architecture was proposed for automatic 3D cephalometric landmark detection. This method applies direct regression in multi-scale learning, enabling landmark detection without failure cases. Experimental results with clinical data indicate that the proposed approach enhances cephalometric landmark detection performance compared to existing CNN methods. Although the proposed method requires prior segmentation of the skull volume, future research aims to develop algorithms that can be directly used in clinical settings by operating training servers within healthcare institutions. We expect that the concept of the proposed method has the potential for application in surgical planning and medical analysis involving maxillofacial imaging.

Author Contributions

Conceptualization, funding acquisition, and project administration, C.S.; methodology, Y.J.; supervision, H.H. and J.-W.P.; validation, J.-Y.P. and J.A.; writing—review and editing, J.S. and E.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Korea Medical Device Development Fund grant funded by the Korea government (the Ministry of Science and ICT; the Ministry of Trade, Industry and Energy; the Ministry of Health & Welfare; the Ministry of Food and Drug Safety) (Project Number: 1711174476, RS-2022-00141072). This work was supported by the Commercialization Promotion Agency for R&D Outcomes (COMPA) grant funded by the Korean Government (Ministry of Science and ICT) (RS-2023-00304776).

Institutional Review Board Statement

This study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of Samsung Medical Center (IRB File No. SMC 2022-08-091-001).

Informed Consent Statement

Written informed consent was waived by the Institutional Review Board.

Data Availability Statement

The original contributions presented in the study are included in this article, and further inquiries can be directed to the corresponding authors.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yao, Q.; Quan, Q.; Xiao, L.; Kevin Zhou, S. One-shot medical landmark detection. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2021, Proceedings of the 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Proceedings, Part II 24; Springer: Berlin/Heidelberg, Germany, 2021; pp. 177–188. [Google Scholar]
  2. Chen, R.; Ma, Y.; Liu, L.; Chen, N.; Cui, Z.; Wei, G.; Wang, W. Semi-supervised anatomical landmark detection via shape-regulated self-training. Neurocomputing 2022, 471, 335–345. [Google Scholar] [CrossRef]
  3. Zhu, H.; Yao, Q.; Xiao, L.; Zhou, S.K. You only learn once: Universal anatomical landmark detection. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2021, Proceedings of the 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Proceedings, Part V 24; Springer: Berlin/Heidelberg, Germany, 2021; pp. 85–95. [Google Scholar]
  4. Yao, J.; Zeng, W.; He, T.; Zhou, S.; Zhang, Y.; Guo, J.; Tang, W. Automatic localization of cephalometric landmarks based on convolutional neural network. Am. J. Orthod. Dentofac. Orthop. 2022, 161, e250–e259. [Google Scholar] [CrossRef] [PubMed]
  5. Ma, Q.; Kobayashi, E.; Fan, B.; Nakagawa, K.; Sakuma, I.; Masamune, K.; Suenaga, H. Automatic 3D landmarking model using patch-based deep neural networks for CT image of oral and maxillofacial surgery. Int. J. Med. Robot. Comput. Assist. Surg. 2020, 16, e2093. [Google Scholar] [CrossRef] [PubMed]
  6. Dot, G.; Schouman, T.; Chang, S.; Rafflenbeul, F.; Kerbrat, A.; Rouch, P.; Gajny, L. Automatic three-dimensional cephalometric landmarking via deep learning. J. Dent. Res. 2022, 101, 1380–1387. [Google Scholar] [CrossRef] [PubMed]
  7. Kang, S.H.; Jeon, K.; Kim, H.-J.; Seo, J.K.; Lee, S.-H. Automatic three-dimensional cephalometric annotation system using three-dimensional convolutional neural networks: A developmental trial. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2020, 8, 210–218. [Google Scholar] [CrossRef]
  8. Zheng, Y.; Liu, D.; Georgescu, B.; Nguyen, H.; Comaniciu, D. 3D deep learning for efficient and robust landmark detection in volumetric data. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Proceedings of the 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part I 18; Springer: Berlin/Heidelberg, Germany, 2015; pp. 565–572. [Google Scholar]
  9. Yun, H.S.; Jang, T.J.; Lee, S.M.; Lee, S.-H.; Seo, J.K. Learning-based local-to-global landmark annotation for automatic 3D cephalometry. Phys. Med. Biol. 2020, 65, 085018. [Google Scholar] [CrossRef] [PubMed]
  10. Liu, Q.; Deng, H.; Lian, C.; Chen, X.; Xiao, D.; Ma, L.; Chen, X.; Kuang, T.; Gateno, J.; Yap, P.-T. SkullEngine: A multi-stage CNN framework for collaborative CBCT image segmentation and landmark detection. In Machine Learning in Medical Imaging, Proceedings of the 12th International Workshop, MLMI 2021, Held in Conjunction with MICCAI 2021, Strasbourg, France, 27 September 2021; Proceedings 12; Springer: Berlin/Heidelberg, Germany, 2021; pp. 606–614. [Google Scholar]
  11. Nishimoto, S.; Saito, T.; Ishise, H.; Fujiwara, T.; Kawai, K.; Kakibuchi, M. Three-Dimensional Craniofacial Landmark Detection in Series of CT Slices Using Multi-Phased Regression Networks. Diagnostics 2023, 13, 1930. [Google Scholar] [CrossRef] [PubMed]
  12. Lang, Y.; Wang, L.; Yap, P.-T.; Lian, C.; Deng, H.; Thung, K.-H.; Xiao, D.; Yuan, P.; Shen, S.G.; Gateno, J. Automatic detection of craniomaxillofacial anatomical landmarks on CBCT images using 3D mask R-CNN. In Graph Learning in Medical Imaging, Proceedings of the First International Workshop, GLMI 2019, Held in Conjunction with MICCAI 2019, Shenzhen, China, 17 October 2019; Proceedings 1; Springer: Berlin/Heidelberg, Germany, 2019; pp. 130–137. [Google Scholar]
  13. Chen, X.; Lian, C.; Deng, H.H.; Kuang, T.; Lin, H.-Y.; Xiao, D.; Gateno, J.; Shen, D.; Xia, J.J.; Yap, P.-T. Fast and accurate craniomaxillofacial landmark detection via 3D faster R-CNN. IEEE Trans. Med. Imaging 2021, 40, 3867–3878. [Google Scholar] [CrossRef] [PubMed]
  14. Sahlsten, J.; Järnstedt, J.; Jaskari, J.; Naukkarinen, H.; Mahasantipiya, P.; Charuakkra, A.; Vasankari, K.; Hietanen, A.; Sundqvist, O.; Lehtinen, A. Deep learning for 3D cephalometric landmarking with heterogeneous multi-center CBCT dataset. PLoS ONE 2024, 19, e0305947. [Google Scholar] [CrossRef]
  15. Zhou, Q.-Y.; Park, J.; Koltun, V. Open3D: A modern library for 3D data processing. arXiv 2018, arXiv:1801.09847 2018. [Google Scholar]
  16. Schroeder, W.; Martin, K.; Lorensen, B. The Visualization Toolkit, 4th ed.; Kitware: New York, NY, USA, 2006. [Google Scholar]
  17. Jiménez, J.; Skalic, M.; Martinez-Rosell, G.; De Fabritiis, G. K deep: Protein–ligand absolute binding affinity prediction via 3d-convolutional neural networks. J. Chem. Inf. Model. 2018, 58, 287–296. [Google Scholar] [CrossRef] [PubMed]
  18. Li, Y.; Rezaei, M.A.; Li, C.; Li, X. DeepAtom: A framework for protein-ligand binding affinity prediction. In Proceedings of the 2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), San Diego, CA, USA, 18–21 November 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 303–310. [Google Scholar]
  19. Wang, Y.; Qiu, Z.; Jiao, Q.; Chen, C.; Meng, Z.; Cui, X. Structure-based protein-drug affinity prediction with spatial attention mechanisms. In Proceedings of the 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Houston, TX, USA, 9–12 December 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 92–97. [Google Scholar]
  20. Sun, S.; Gao, L. Contrastive pre-training and 3D convolution neural network for RNA and small molecule binding affinity prediction. Bioinformatics 2024, 40, btae155. [Google Scholar] [CrossRef] [PubMed]
  21. Chen, H.; Dou, Q.; Yu, L.; Qin, J.; Heng, P.-A. VoxResNet: Deep voxelwise residual networks for brain segmentation from 3D MR images. NeuroImage 2018, 170, 446–455. [Google Scholar] [CrossRef] [PubMed]
  22. Milletari, F.; Navab, N.; Ahmadi, S.-A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 565–571. [Google Scholar]
  23. Kompanek, M.; Tamajka, M.; Benesova, W. Volumetric data augmentation as an effective tool in mri classification using 3d convolutional neural network. In Proceedings of the 2019 International Conference on Systems, Signals and Image Processing (IWSSIP), Osijek, Croatia, 5–7 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 115–119. [Google Scholar]
  24. Alakwaa, W.; Nassef, M.; Badr, A. Lung cancer detection and classification with 3D convolutional neural network (3D-CNN). Int. J. Adv. Comput. Sci. Appl. 2017, 8, 409–417. [Google Scholar] [CrossRef]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  26. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  27. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  28. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  29. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  30. Tamhane, A.; Mida, T.E.; Posner, E.; Bouhnik, M. Colonoscopy landmark detection using vision transformers. In MICCAI Workshop on Imaging Systems for GI Endoscopy; Springer: Berlin/Heidelberg, Germany, 2022; pp. 24–34. [Google Scholar]
Figure 1. Diagram of multi-scale cephalometric landmark detection architecture. Stage 1 involves coarse detection to identify coordinates and classes from the entire input volume, followed by generating local volumes through 3D region of interest (ROI) processing. Stage 2 focuses on fine localization using these local volumes and classes.
Figure 1. Diagram of multi-scale cephalometric landmark detection architecture. Stage 1 involves coarse detection to identify coordinates and classes from the entire input volume, followed by generating local volumes through 3D region of interest (ROI) processing. Stage 2 focuses on fine localization using these local volumes and classes.
Diagnostics 14 02605 g001
Figure 2. Steps for data preprocessing to convert volumetric CT data into normalized voxel data.
Figure 2. Steps for data preprocessing to convert volumetric CT data into normalized voxel data.
Diagnostics 14 02605 g002
Figure 3. Convolutional Neural Network (CNN) architectures including ResNet, DenseNet, Inception, and InceptionResNet.
Figure 3. Convolutional Neural Network (CNN) architectures including ResNet, DenseNet, Inception, and InceptionResNet.
Diagnostics 14 02605 g003
Figure 4. DenseNet169-based multi-output learning model for coarse detection.
Figure 4. DenseNet169-based multi-output learning model for coarse detection.
Diagnostics 14 02605 g004
Figure 5. DenseNet169-based multi-input learning model for fine localization.
Figure 5. DenseNet169-based multi-input learning model for fine localization.
Diagnostics 14 02605 g005
Table 1. Comparison of trainable parameters, accuracy, and inference time for landmark detection of the proposed method and existing CNN architectures.
Table 1. Comparison of trainable parameters, accuracy, and inference time for landmark detection of the proposed method and existing CNN architectures.
ParametersAccuracy (mm)Time (s)
Inception-ResNetV267,647,8022.978 ± 0.4930.246
Inceptionv334,299,7702.917 ± 0.5010.225
ResNet5046,377,5012.787 ± 0.5450.118
ResNet10185,475,8692.754 ± 0.4730.142
ResNet152117,680,6692.783 ± 0.4900.302
DenseNet20125,732,7622.505 ± 0.3910.172
DenseNet12111,418,5222.467 ± 0.4350.232
DenseNet16918,850,9702.464 ± 0.3820.334
DenseNet169 + Coarse-to-fine (Proposed)39,056,6052.238 ± 0.3643.835
Table 2. The accuracy for each of the 30 cephalometric landmark coordinates across the proposed method and existing CNN architectures.
Table 2. The accuracy for each of the 30 cephalometric landmark coordinates across the proposed method and existing CNN architectures.
Inception
ResNetV2
Inceptionv3ResNet50ResNet101ResNet152DenseNet201DenseNet121DenseNet169Proposed
11apex2.497 ± 0.9222.620 ± 1.1772.302 ± 1.1362.087 ± 0.9562.379 ± 1.0811.959 ± 0.8892.068 ± 0.8731.925 ± 0.8761.815 ± 0.945
132.412 ± 1.0902.473 ± 1.0852.268 ± 0.9362.330 ± 0.8892.450 ± 1.1062.145 ± 1.0982.197 ± 0.9442.153 ± 0.9402.028 ± 0.913
163.221 ± 1.1803.343 ± 1.2702.865 ± 1.1462.823 ± 1.3883.201 ± 1.4632.882 ± 1.2262.956 ± 1.1212.973 ± 1.2472.875 ± 1.208
21apex2.477 ± 0.8652.506 ± 1.2552.221 ± 0.9312.130 ± 0.9762.475 ± 0.9001.947 ± 0.7761.951 ± 0.8191.838 ± 0.7851.953 ± 0.731
232.748 ± 1.2263.008 ± 1.5242.574 ± 1.0652.572 ± 1.1592.803 ± 1.2652.278 ± 1.1662.345 ± 1.2282.286 ± 1.1292.053 ± 1.074
263.097 ± 1.1473.395 ± 1.4522.996 ± 1.3002.969 ± 1.2713.243 ± 1.6522.886 ± 1.2882.986 ± 1.1922.978 ± 1.3572.900 ± 1.244
A-point2.484 ± 1.0012.595 ± 1.0622.265 ± 0.9532.306 ± 0.9952.319 ± 0.8432.083 ± 0.8602.133 ± 0.9132.179 ± 0.9542.106 ± 0.985
ANS2.778 ± 1.3563.003 ± 1.3672.611 ± 1.2052.869 ± 1.4222.394 ± 1.1882.416 ± 1.3252.414 ± 1.3792.495 ± 1.3222.656 ± 1.676
B2.933 ± 1.4082.824 ± 1.2552.537 ± 1.4072.746 ± 1.3282.550 ± 1.1982.821 ± 1.1482.354 ± 1.2552.515 ± 1.2072.041 ± 1.118
Ba3.271 ± 1.8222.427 ± 1.1323.102 ± 1.6613.124 ± 1.6313.051 ± 1.7422.680 ± 1.5042.418 ± 1.2262.683 ± 1.3862.131 ± 1.135
CoL3.444 ± 1.0733.383 ± 1.5223.661 ± 1.8613.821 ± 1.7673.847 ± 1.2853.274 ± 1.4453.319 ± 1.3002.907 ± 1.3732.832 ± 1.393
CoR3.618 ± 1.5323.351 ± 1.4863.549 ± 1.4353.543 ± 1.2363.326 ± 1.4593.060 ± 1.3062.946 ± 1.1172.862 ± 1.0352.860 ± 1.126
Gn3.189 ± 1.4382.650 ± 1.4402.367 ± 1.2192.373 ± 0.9612.319 ± 0.8002.196 ± 1.1711.884 ± 1.0751.938 ± 0.9271.457 ± 0.674
GoL4.448 ± 2.0794.608 ± 1.9494.104 ± 2.0344.184 ± 2.1974.104 ± 2.0993.567 ± 2.1013.660 ± 1.7793.666 ± 1.8663.038 ± 1.862
GoR4.213 ± 1.6233.740 ± 1.4763.875 ± 1.4593.741 ± 1.7313.417 ± 1.5743.473 ± 1.5793.386 ± 1.6543.469 ± 1.7452.664 ± 1.444
L12.645 ± 1.0512.803 ± 1.0292.932 ± 0.9812.900 ± 0.9792.924 ± 0.9632.730 ± 0.8772.502 ± 0.9372.734 ± 0.9422.291 ± 0.934
Lt. FZS2.281 ± 1.1982.654 ± 1.2822.592 ± 1.1802.419 ± 1.0202.367 ± 1.2542.053 ± 0.8432.040 ± 0.8532.120 ± 0.8451.860 ± 0.891
Menton3.048 ± 1.7432.687 ± 1.2492.358 ± 1.2942.292 ± 0.8842.253 ± 0.9512.091 ± 1.1072.042 ± 1.0791.935 ± 0.9811.533 ± 0.746
Na2.297 ± 1.1932.394 ± 1.1022.228 ± 1.0732.197 ± 0.9122.218 ± 0.9911.732 ± 0.9301.708 ± 0.9181.767 ± 0.9921.699 ± 0.805
Nasal tip3.001 ± 1.5983.009 ± 1.1842.574 ± 1.2732.657 ± 1.1632.756 ± 1.2112.662 ± 1.1252.795 ± 1.4072.710 ± 1.4002.689 ± 1.535
OrL2.967 ± 1.5053.093 ± 1.4842.864 ± 1.4182.559 ± 1.2642.863 ± 1.4242.644 ± 1.0902.639 ± 1.0282.633 ± 1.1292.318 ± 1.211
OrR3.183 ± 1.1843.092 ± 1.2102.601 ± 0.8412.876 ± 1.1662.785 ± 1.3252.572 ± 0.9962.624 ± 1.0622.664 ± 1.0032.312 ± 0.980
PNS2.889 ± 1.5742.698 ± 1.8512.826 ± 1.7872.828 ± 1.7402.855 ± 1.7412.687 ± 1.6562.604 ± 1.6362.804 ± 1.4382.597 ± 1.451
PoL3.570 ± 1.5533.031 ± 1.2293.583 ± 1.5913.295 ± 1.6493.321 ± 1.5543.041 ± 1.3002.830 ± 1.4132.819 ± 1.2162.774 ± 1.251
PoR3.310 ± 1.4192.704 ± 1.1143.402 ± 1.6272.865 ± 1.1992.886 ± 1.3382.459 ± 1.1692.525 ± 1.0242.658 ± 0.9732.253 ± 0.978
Pog3.045 ± 1.3692.681 ± 1.2472.467 ± 1.1692.508 ± 0.9662.603 ± 0.7212.324 ± 0.9371.968 ± 0.9842.018 ± 0.8471.687 ± 0.813
Rt. FZS2.838 ± 1.2532.594 ± 1.0612.140 ± 1.0252.371 ± 1.5162.307 ± 1.1722.140 ± 1.1732.102 ± 1.0851.991 ± 1.0911.747 ± 1.102
S2.232 ± 1.3742.606 ± 1.3002.544 ± 1.2132.319 ± 1.3112.381 ± 1.3051.971 ± 1.1192.039 ± 1.0461.872 ± 1.1091.845 ± 0.940
Sp2.780 ± 1.8082.879 ± 1.2412.662 ± 1.5512.537 ± 1.4262.699 ± 1.5892.149 ± 1.2882.297 ± 1.3702.122 ± 1.3082.033 ± 1.374
U12.430 ± 1.0162.659 ± 1.0622.532 ± 0.9702.375 ± 1.0532.382 ± 1.0472.224 ± 1.0622.272 ± 0.9692.201 ± 0.8972.091 ± 0.872
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, C.; Jeong, Y.; Huh, H.; Park, J.-W.; Paeng, J.-Y.; Ahn, J.; Son, J.; Jung, E. Multi-Scale 3D Cephalometric Landmark Detection Based on Direct Regression with 3D CNN Architectures. Diagnostics 2024, 14, 2605. https://doi.org/10.3390/diagnostics14222605

AMA Style

Song C, Jeong Y, Huh H, Park J-W, Paeng J-Y, Ahn J, Son J, Jung E. Multi-Scale 3D Cephalometric Landmark Detection Based on Direct Regression with 3D CNN Architectures. Diagnostics. 2024; 14(22):2605. https://doi.org/10.3390/diagnostics14222605

Chicago/Turabian Style

Song, Chanho, Yoosoo Jeong, Hyungkyu Huh, Jee-Woong Park, Jun-Young Paeng, Jaemyung Ahn, Jaebum Son, and Euisung Jung. 2024. "Multi-Scale 3D Cephalometric Landmark Detection Based on Direct Regression with 3D CNN Architectures" Diagnostics 14, no. 22: 2605. https://doi.org/10.3390/diagnostics14222605

APA Style

Song, C., Jeong, Y., Huh, H., Park, J. -W., Paeng, J. -Y., Ahn, J., Son, J., & Jung, E. (2024). Multi-Scale 3D Cephalometric Landmark Detection Based on Direct Regression with 3D CNN Architectures. Diagnostics, 14(22), 2605. https://doi.org/10.3390/diagnostics14222605

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop