Abstract
Cephalometric analysis is the primary diagnosis method in orthodontics. In our previous study, the algorithm was developed to estimate cephalometric landmarks from lateral facial photographs of patients with normal occlusion. This study evaluates the estimation accuracy by the algorithm trained on a dataset of 2320 patients with added malocclusion patients and the analysis values. The landmarks were estimated from the input of lateral facial photographs as training data using trained CNN-based algorithms. The success detection rate (SDR) was calculated based on the mean radial error (MRE) of the distance between the estimated and actual coordinates. Furthermore, the estimated landmarks were used to measure angles and distances as a cephalometric analysis. In the skeletal Class II malocclusion, MRE was 0.42 ± 0.15 mm, and in the skeletal Class III malocclusion, MRE was 0.46 ± 0.16 mm. We conducted a cephalometric analysis using the estimated landmarks and examined the differences with actual data. In both groups, no significant differences were observed for any of the data. Our new algorithm for estimating the landmarks from lateral facial photographs of malocclusion patients resulted in an error of less than 0.5 mm; the error in cephalometric analysis was less than 0.5°. Therefore, the algorithm can be clinically valuable.
Similar content being viewed by others
Introduction
Lateral cephalometric analysis, originally proposed by Broadbent in 1931 to assess anteroposterior and vertical skeletal morphology, is a fundamental diagnostic tool in orthodontics and an integral part of orthodontic treatment planning1. Cephalometric analysis relies on identifying radiological landmarks and subsequent measurements of various angles, distances, and ratios to delineate anatomical structures in the craniofacial region2. Remarkably, manual landmark plotting continues to be the standard practice, even when utilizing professional software within the orthodontic field3. Hence, the quality of cephalometric analysis heavily relies on the orthodontist’s knowledge and experience, which can result in variability among different raters4. The recognition of the value of automated landmark detection for cephalometric analysis is growing, driven by it potential to streamline the labor-intensive nature and reduce reliance on experienced orthodontists5,6.
In recent years, the application of artificial intelligence (AI) in the medical field, particularly computer vision and AI-based image analysis, has gained significant traction. Deep learning (DL), a subset of machine learning, and the use of convolutional neural networks (CNN) have been notably advocated for their suitability in computer vision tasks7. One prevalent approach is supervised learning, where an algorithm is trained on data pairs and corresponding labels8. During the training phase, the algorithm learns the statistical relationships between input data and labels to predict new, unseen data2.
Efforts have been made to create specialized datasets for automated cephalometric analysis and incorporate them into DL models for landmark detection9,10. However, despite promising initial results, challenges persist, particularly those related to the accuracy of landmark detection, which still needs to be improved11. Additionally, concerns about X-ray exposure persist, particularly in cephalograms and cone beam computed tomography (CBCT) used in maxillofacial procedures12. These concerns highlight the need to study alternative techniques for orthodontic pre-treatment examinations that reduce or eliminate X-ray exposure.
Conversely, facial landmark detection involves localizing predefined facial landmarks such as the eyes, nose, mouth, and chin in facial images. Previous studies have utilized models such as the active shape model (ASM)13 and constrained local model (CLM)14. However, these models have struggled to maintain robustness against appearance variations. To address these challenges, studies have explored the cascade regression approach15; however, this approach has limitations regarding accuracy improvement16. Deep neural networks (DNN) have recently emerged as a potent alternative17, with CNN-based approaches, including the hourglass structure18 and heat-map-based regression19,20,21, demonstrating outstanding results. Sun et al. introduced the high-resolution network V2 (HRNetV2)22, designed to preserve high resolution throughout the network while reducing parameters and computational costs compared to conventional methods. Other studies have focused on architecture-agnostic learning algorithms, such as ADNet, based on LAB and Awing, which achieved high performance19,23,24. Other methods of improved accuracy are called fine-tuning. It is the process of adjusting the parameters of a model that has been pre-trained with a large amount of data by learning additional data according to the target task to perform. Combining pre-training and fine-tuning has the advantage of providing AI with high-quality learning even when there is insufficient training data, and is used in AI development.
Our previous study has developed a novel approach to estimating cephalometric landmarks for cephalon analysis from lateral facial photographs by using CNN-based algorithm25. Even though landmark estimation was performed from lateral facial photographs, this method showed outstanding accuracy, surpassing the previously reported accuracy of landmark estimation from X-ray cephalograms. A two-stage learning method wherein HRNetV2 is employed for cephalometric landmark localization, and a multilayer perceptron (MLP) is used to refine the positions of the landmarks in our algorithm, thereby enhancing accuracy. However, this study’s results, which achieved high accuracy, were conducted using data from patients with normal occlusion after orthodontic treatment as training data.
Given the potential clinical applications, where orthodontists often need to analyze malocclusions, it becomes crucial to investigate further whether a similarly high level of accuracy can be achieved in malocclusion patients with skeletal Class II malocclusion and skeletal Class III malocclusion. this is a critical step towards the clinical implementation of our novel approach.
Therefore, by adding malocclusion data to the algorithm in the previous study25, this study was conducted to evaluate the accuracy and reliability of detecting estimated cephalometric landmarks from lateral facial photographs using a modified our algorithm that can adapt to malocclusion.
Additionally, it evaluated the accuracy and reliability of cephalometric analysis values derived from the estimated cephalometric landmarks.
Material and methods
The study protocol was approved by the Tokyo Dental College Ethics Committee (No. 1091). All experiments were conducted in accordance with the principles of the Declaration of Helsinki and the relevant guidelines and regulations. Informed written consent, including facial images, was obtained from all patients.
Data preparation
Patients with malocclusion who underwent orthodontic treatment, including surgical orthodontic treatment, at Tokyo Dental College between September 1981 and September 2021 were recruited, and their pre-treatment data were collected for analysis. All participants included in the study exhibited an ANB angle of 4° or more, indicating skeletal Class II malocclusion (CL II group) or an ANB angle of 1° or less, and a Wits appraisal of -1 mm or less, indicative of Skeletal Class III mallocclusion (CL III group) (Table 1).
The exclusion criteria included patients with congenital abnormality such as cleft lip and palate, all syndromes that cause maxillofacial deformity, patients with missing molars or multiple missing teeth of more than 3 teeth, and those with a history of maxillofacial trauma. A total of 200 patients were selected for each group, and the lateral cephalograms were obtained during participants examinations.
Lateral cephalograms were acquired using a CX-150S (ASAHIROENTGEN, Tokyo, Japan), and lateral facial photographs were captured in the JPEG format using an EOS Kiss X90 camera (Canon, Tokyo, Japan). The acquisition of lateral cephalograms adhered to globally recognized standards1. The cartridge was positioned parallel to the mid-sagittal plane of the head, and the head was stabilized using ear rods so that the central axis of the X-ray beam passed through the axis of the left and right ear rods. The distance from the X-ray tube to the mid-sagittal plane of the head was 150 cm, and the distance from the mid-sagittal plane of the head to the cartridge was 15 cm. For the lateral facial photographs, the camera-to-mid-sagittal plane distance was maintained at 150 cm, and ear rods were used to secure the head for capturing images.
The plotting and transfer of landmarks from the cephalogram to lateral facial photographs were performed according to the guidelines reported in a previous study25. A brief description of this process is as follows. A total of 23 landmarks were selected as representative landmarks of the skeleton and teeth (Fig. 1, Supplementary information 1). These are the main landmarks used in major cephalometric analyses, such as Downs, Northwestern, and Ricketts26,27,28, and the landmarks were plotted and tracing soft tissue using Quick Ceph Studio (version 5.0.2, Quick Ceph Systems, San Diego, California). A line traced from the forehead to the upper lip on the cephalogram was matched to a corresponding line on the lateral facial photograph in the software. This alignment ensured that the magnification and position of the cephalogram and the lateral facial photograph matched (Fig. 2.)
Procedure for plotting landmarks and superimposing the cephalometric tracing on the facial profile image. (a) cephalogram. (b) plotting landmark and tracing on cephalogram. Landmarks and tracing are excluded from cephalograms. (c) lateral facial photograph. (d) superimposition of landmarks, tracing and lateral facial photograph
Error analysis for the landmarks plotting on the cephalogram and transferred landmarks on the lateral facial photograph
Two manual procedures were conducted: (1) plotting the landmarks and (2) transferring the landmarks from the cephalogram to the lateral facial photograph. These error tests were based on the methods reported in25. A brief description of this process is as follows. Prior to each manual procedure, an error test was conducted to validate the precision of each task. To assess the repeatability and reproducibility errors in landmark plotting, 23 landmarks from five randomly selected patients were plotted three times for each case, with a two-week interval between repetitions. Plotting and tracing for this study were performed by two orthodontists: one orthodontic specialist with five years of experience and one orthodontic supervisor with 15 years of experience. The accuracy of plotting and tracing was subsequently verified by another orthodontist: an orthodontic supervisor with 36 years of experience and a board member of the Japanese Orthodontic Society. The Shapiro–Wilk test was employed for normality checks to calculate the intraclass correlation coefficients (ICC) (Supplementary information 2). The results indicated that both repeatability and reproducibility errors remained within a 1-mm range, and the ICC was above 0.9 for all landmarks. Consequently, No problems were identified related to errors with the plotted landmarks29. Subsequently, error tests were conducted to superimpose trace lines on the lateral facial photographs at five landmarks (sella, porion, menton, gonion, and basion). The error distances between and within each landmark were measured consistently. The findings revealed that the errors were within a 1-mm range, and the ICC was above 0.9, indicating that the superimposition errors were negligible30. Each image has 1200 × 1200 pixels and the pixel size was 0.35 mm and each cephalograms has 1980 × 1648 pixels and the pixel size was 0.15 mm. The cephalograms is then matched to the pixel size of the lateral facial photograph.
Input of training data and estimation of test data
In this study, a computer equipped with a GeForce GTX 1080, a 3.70 GHz Intel(R) Core (TM) i7-8700 K CPU, and 16 GB of memory was used. For the landmark estimation, we adopted the neural network used in25, which consisted of two stages: HRNetV2 and MLP. The training dataset comprised 160 randomly selected participants from the CL II group (n = 200) and 160 randomly selected participants from the CL III group (n = 200). The fine-tuning process was applied as a learning process in this study. (Fig. 3). The test data comprised the remaining 40 from the CL II group and CL III group that were not included in the training dataset. Five-fold cross-validation was performed on all images in this study, and the results were averaged across five estimations.
Accuracy assessment of the estimated landmark
To validate the accuracy of landmarks estimation from lateral facial photographs using AI, we calculated the mean radial error (MRE) according to the methodology proposed in31. Furthermore, we assessed the displacement in the horizontal (x) and vertical (y) directions. The radial error was defined as R (R = \(\sqrt{\Delta {x}^{2}+\Delta {y}^{2}}\)) , where Δx and Δy the Euclidean distance between the estimated landmarks and the actual landmarks along the respective x and y axes. The MRE can be defined as follows:
Where Ri represents the radial error for the ith landmark, and n denotes the total number of detected landmarks. The success detection rate (SDR) is the proportion of estimated landmarks within a predefined threshold, and it is defined as follows:
Where, nd represents the number of successfully detected landmarks. The thresholds were set to 2.0, 2.5, 3.0, and 4.0-mm, respectively. The average SDR for all images at each threshold was calculated.
Accuracy cephalometric of analysis using estimated landmarks
-
1)
Accuracy: Using the estimated landmarks, we conducted angle and distance landmarks for the cephalometric analysis of skeletal and denture patterns. Student’s t-test was used to compare the cephalometric analysis results between the estimated and actual landmarks in the two groups.
-
2)
Agreement: Brand-Altman analysis was performed to examine the agreement of cephalometric analysis values by AI.
-
3)
Correlation: To investigate the influence of anterior–posterior and vertical positional discrepancies in the 23 landmarks of skeletal malocclusions, we examined the bivariate correlations with the ANB angle, wits appraisal, and FMA for each group. The Kolmogorov–Smirnov test was used to determine whether the data were normally distributed; we used Pearson’s correlation coefficient for correlation analysis. All statistical analysis was conducted using IBM SPSS Statistics software (version 27.0, IBM, Armonk, New York).
Results
Accuracy of landmark estimation
Figure 4 shows an example of the results of landmark estimation on training data, and Tables 2 and 3 list the estimation accuracy of 23 landmarks of the mean MRE and the SDR for each threshold.
In the CL II group, the mean MRE was 0.42 ± 0.15 mm, with values lower than 0.35 mm for Or and Pt, indicating high accuracy. However, landmarks such as Pog, Ar, and R3 show values above 0.5 mm. The SDR showed 100.00% at all thresholds. In the CL III group, the mean MRE was 0.46 ± 0.16 mm. High accuracy (< 0.35 mm) was observed for Po, Pt, and ANS, whereas landmarks such as S, N, A, B, Pog, Is, Ar, and U1-R showed values above 0.5 mm. The SDR was 99.96% at a mean of all landmarks for the 2.0-mm threshold and 100.00% for other thresholds.
Evaluation of cephalometric analysis using estimated measurement points by AI from the lateral facial photograph
-
1)
Accuracy: We conducted a cephalometric analysis using the estimated landmarks and examined the differences with the actual data. In both groups, no significant differences were observed for any of the data. In the Cl II group, the error was approximately less than 0.1° for convexity, ANB angle, U1 to FH, less than 0.1 mm for L1 to Apo (mm), and approximately 0.5°for facial, gonial, and interincisal angles, and FMIA. In the CL III group, the error was less than 0.1° for the cant of occlusal plane, SNB angle, SNP angle, facial axis, and U1 to FH, and approximately 1.0°for interincisal angle and FMIA (Tables 4 and 5).
-
2)
Agreement: In the horizontal (x) direction of the CL II group, random errors were detected at N, Or, Po, and Ba. Fixed errors were detected as S, A, B, Pog, Me, Gn, Go, Ii, Is, Cc, Pt, R1, PNS, ANS, Ar, R3, U1-R, and L1-R biased in the negative direction, and R3 biased in the positive direction. Proportional errors were detected in Mo. For the CL II group in the vertical direction (y), random errors were detected in S, N, Or, A, Pog, Me, Gn, Go, Is, PNS, ANS, Ar, Mo, and L1-R. Fixed errors were detected for Po in the positive direction and for B, Ii, Ba, Cc, Pt, R1, R3, and U1-R in the negative direction. For the CL III group in the horizontal direction (x), random errors were detected for N, Or, Ii, Cc, Pt, R3, and Mo. Fixed errors were positively biased at S and negatively biased at A, B, Pog, Gn, Go, R1, PNS, ANS, Ar, U1-R, and L1-R. Proportional errors were detected for Po, Me, and Ba. Random errors were detected in Po, A, Me, Gn, Go, Pt, R1, and L1-R in the vertical direction (y) of the CL III group. Fixed errors were observed in the positive direction for S and N and the negative direction for Or, B, Pog, Ii, Is, Ba, PNS, R3, and U1-R. Proportional errors were detected for Cc, ANS, Ar, and Mo (Tables 6 and 7 and Supplementary information 3).
-
3)
Correlation: Bivariate correlation coefficients between the average MRE at each landmark and the measured values of the ANB angle, wits appraisal, and FMA were determined. A positive correlation (r = 0.485) was observed between the ANB angle and Is in the CL II group (Table 8).
Limited correlation was observed between Wits appraisal and any of the measurement points. However, a negative correlation (r = − 0.495) was observed between FMA and Cc.
In the CL III group, negative correlations were observed between the ANB angle and A (r = − 0.464), Go (r = − 0.436), and Ii (r = − 0.482) (Table 9).
Regarding Wits appraisal, a negative correlation was observed with A (r = − 0.492) and Ii (r = − 0.479). In FMA, weak positive correlations were observed with Po (r = 0.364) and B (r = 0.364), whereas weak negative correlations were noted with Cc (r = − 0.319) and Ar (r = − 0.379).
Discussion
This study enabled the acquisition of a highly accurate cephalometric analysis value from the lateral facial photographs of malocclusion patients.
The proposed algorithm and the accuracy of landmarks estimation
The algorithm implemented in this study is based on our previous work25, where we predicted landmarks from lateral facial photographs of patients with normal occlusion. This algorithm comprises two stages: HRNetV2 and MLP. Initially, we used HRNetV2 for heatmap regression to estimate the positions of all landmarks in the input image. HRNetV2 applies multiple convolutional layers, enabling a precise understanding of the spatial relationships between each landmark and facial structures. In this step, the model learns the relationship between the local features of the face and landmarks. Subsequently, we introduced coordinate regression using MLP, which significantly improved the accuracy of landmark estimation. Learning complex spatial relationships with only MLP might not be accurate; however, combination of MLP and HR NetV2 can improve the accuracy of landmark estimation. Moreover, MLP can integrate input position information effectively because it is fully connected from the input to the output layer. This enables the model to learn underlying spatial relationships. Therefore, MLP can estimate the structural features between landmarks, i.e., it can estimate the relative positions of landmarks. In this study, the two-stage approach, comprising the coarse estimation through HRNetV2 heatmap regression and the fine estimation through MLP, enabled the accurate detection of all landmarks25.
Even though our study inferred landmarks from lateral facial photographs, the results outperformed those of previous studies that inferred landmarks from cephalograms32,33,34,35. This outcome may be unexpected from the perspective of humans; however, for the AI, it is possible that the HR NetV2 heatmap regression-based methods could capture facial soft tissue features such as eyes, nose, and mouth, in color facial photographs than in radiograph images36.
In our previous study25, we used lateral facial photographs of patients with normal occlusion after orthodontic treatment as the training data and the test data, and the mean MRE for each landmark was 0.61 ± 0.50 mm. In contrast, this study focused on pre-treatment patients with skeletal Class II and III malocclusion. Malocclusion data are considered more variable than that of normal occlusion, which may lead to greater variability in the prediction of landmarks. However, as a result, the mean MRE for each landmark in patients with the CL II group was 0.42 ± 0.15 mm, and that in patients with the CL III group was 0.46 ± 0.16 mm, which is even more accurate than those of normal occlusion. Moreover, there were almost no differences between each landmark.
In this study, several methods were attempted as a preliminary step in the research, using the algorithms from our previous study to learn and estimate data on malocclusion (Table 10). The methods include learning only skeletal Class II malocclusion data to predict skeletal Class II malocclusion, learning only skeletal Class III malocclusion data to predict skeletal Class III malocclusion, learning a combination of normal occlusion data and skeletal Class II malocclusion and skeletal Class III malocclusion data to predict the mixed data, and adding fine-tuning methods to these methods. Fine-tuning is the process of selecting pre-trained data, adding new layers tailored to the target task, and retraining the entire model with these new layers for refinement37,38. After conducting these preliminary steps, finally, we adopted method 6, because it achieved the most accurate estimation of the landmarks. This study was able to predict landmarks more accurately than the previous study for several reasons. One reason is that we employed fine-tuning methods to improve accuracy, and the other is that the constructed AI algorithms were additionally trained with malocclusion data. We obtained more precise results because we attempted several fine-tuning processes and selected the one with the highest accuracy. This can be considered an advancement in machine learning.
Cephalometric measurements
-
1)
Accuracy: Most skeletal cephalometric measurements based on estimated landmarks revealed a discrepancy of approximately 0.5° between the actual and estimated values. Clinically, an error of less than 1° in skeletal measurements is considered acceptable39.
-
2)
Agreement: Using a Bland–Altman analysis, we evaluated the agreement and error trends between the actual and estimated values. Measurements contain errors, which can be either random or systematic errors. A random error occurs in the true value and is caused by uncontrollable causes.
In this analysis, the influence of individual participant differences was removed by considering the difference between the two measurement methods (the actual data and the estimated data) for each participant. Systematic errors, which have a certain biased tendency toward the true value and are caused by controllable factors (e.g., the habit of the measurer or the inadequacy of the measuring instrument), were further classified into fixed and proportional errors. Fixed errors are those that have a constant bias in a specific direction, regardless of the true value. A proportional error is an error that occurs in a specific direction with respect to the true value. A scatter plot was created to investigate these systematic errors. The scatter plot considered the difference between the estimated and actual values on the vertical axis and the average of the estimated and actual values on the horizontal axis. Then, the vertical (Y-direction) and anterior–posterior (X-direction) errors were classified in this study, and the limit of agreement (LOA) of errors was calculated to ensure the reliability of our findings.
-
3)
Correlation: We selected the ANB angle, Wits appraisal, and FMA as the skeletal type measurement items in this study to examine the relationship between skeletal deformity and the error of each landmark. The ANB angle and Wits appraisal are generally used to indicate the anteroposterior position of the skeleton, with the ANB angle being more relevant than the Wits appraisal, particularly in skeletal Class II malocclusion and the Wits appraisal being more relevant than the ANB angle in skeletal Class III malocclusion40,41. FMA was selected for this study because it is associated with vertical position of the skeleton.
In skeletal Class II malocclusion, the greater the anteroposterior skeletal deformity, namely, the larger the ANB angle, the larger is the error in the position estimation of the upper incisor edge. In contrast, in skeletal Class III malocclusion, the more significant the anteroposterior skeletal deformity, namely, the smaller the Wits appraisal, the larger is the error in the position estimation of the lower incisor edge. Notably, errors in locating the incisor edge of the forward-positioned in overjet could cause errors in inclination. This might be because the position of incisor teeth located forward of overjet is prone to errors in estimated position due to the thickness of the lips42,43. Therefore, the results suggest that the greater the anterior–posterior skeletal deviations, the greater is the error in estimating the incisal position of the anterior teeth.
Regarding the skeletal structure, the results indicated that the larger the skeletal Class III malocclusion, the more the position of A shifted. As noted in the results of this study, the direction of the error was mainly horizontal, and the error was in the backward direction. These findings indicated the similar tendency in both ANB angle and Wits appraisal. Moreover, in skeletal vertical positioning, the greater the tendency for a long face in skeletal Class II malocclusion, the smaller is the error in estimating the position of the condylar head.
However, all landmarks had errors of less than 0.5 mm, which is considered a suitable accuracy level for clinical applications44.
Study limitation
A previous study reported that the greater the sample size of the training data, the greater is the accuracy45. Although no studies have yet been reported on detecting landmarks from photographs, Moon et al.46 estimated the landmarks from the cephalogram and reported that 2300 is the optimal sample size for training data. Arik et al.2, Gilmour et al.32, Li et al.33, Kwon et al.34, and Oh et al.35 used a total of 400 data, and Kim et al.9 used total data counts of 2075 (Table 11).
In this study, we additionally trained using the Takahashi et al. algorithm25, thus using 160 skeletal Class II malocclusion and 160 skeletal Class III malocclusion patients in addition to the 2000 patients from the previous study. Thus, 2320 patients were used as the number of training data. Therefore, based on these related reports, we consider this number as the optimum sample size for the number of training data.
In this study, as several researchers worked on the task, assessing both repeatability and reproducibility errors was crucial. A previous study reported that errors between landmarks in the cephalometric analysis are acceptable in the range of 2 mm46. The errors calculated in the present study were, on average less, than 0.5 mm. Furthermore, this value is considerably smaller than the acceptable errors among orthodontists reported previously, as the inter-measurement error between the two orthodontists in this study was within 1 mm. This result implies that the training data already has an error of at least 1 mm. Namely, the accuracy estimated by the algorithm in this study is considered suitable and high.
Furthermore, this research includes manual steps in the preparation of the test data. These are the plotting of landmarks and the transfer of landmarks from cephalograms to lateral photographs. The methodology in this manual are likely affects the accuracy of the prediction, so it was necessary for us to be highly sensitive in this point.
The training data used in this study consisted mainly of participants of Japanese descent, representing a specific racial background. Hence, errors may arise when facial photographs of participants from different races are used as test data for inference, as there will be differences in the facial morphology, soft tissue, and skeletal structures of the lateral facial photographs. Therefore, to implement the approach for various races, the training data of various races must be considered, and the accuracy must be improved. Thus, we plan to add more training data from various races and conduct further research.
Conclusion
This study developed a machine learning algorithm by adding skeletal Class II and III malocclusion data to the previous algorithm using normal occlusion data. As a result, this algorithm can be applied not only to those with normal occlusion but also to Class II Class III patients. Furthermore, the predictive algorithm was highly accurate, suggesting that a clinically accurate cephalometric analysis could be performed without conventional lateral cephalograms. The findings of this study provide a practical approach to automating cephalometric analysis solely using patients’ lateral facial photographs, thereby eliminating the need for X-ray exposure. Given prevalent concerns regarding X-ray radiation exposure, this development could benefit patients, particularly children. The results of this study have considerable clinical significance.
Data availability
The datasets generated and analyzed during the study are not publicly available because sensitive information in them may violate patient privacy and our institution’s ethics policy. However, the datasets are available from the corresponding author on reasonable request. Data usage agreements may be required.
References
Broadbent, B. H. A new X-ray technique and its application to orthodontia. Angle Orthod. 1, 45–66 (1931).
Arik, S. Ö., Ibragimov, B. & Xing, L. Fully automated quantitative cephalometry using convolutional neural networks. J. Med. Imag. 4, 014501 (2017).
Nishimoto, S., Sotsuka, Y., Kawai, K., Ishise, H. & Kakibuchi, M. Personal computer-based cephalometric landmark detection with deep learning, using cephalograms on the internet. J. Craniofac. 30, 91–95 (2019).
Wang, C.-W. et al. Evaluation and comparison of anatomical landmark detection methods for cephalometric X-ray images: A grand challenge. IEEE Trans. Med. Imag. 34, 1890–1900 (2015).
Forsyth, D. B. & Davis, D. N. Assessment of an automated cephalometric analysis system. Eur. J. Orthod. 18, 471–478 (1996).
Huang, F. J. & LeCun, Y. Large-scale learning with SVM and convolutional for generic object categorization. in Proceeding of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, (CVPR’06). 284–91(2006).
Kunz, F., Stellzig-Eisenhauer, A., Zeman, F. & Boldt, J. Artificial intelligence in orthodontics: Evaluation of a fully automated cephalometric analysis using a customized convolutional neural network. J. Orofac. Orthop. 81, 52–68 (2020).
LeCun, Y., Bengio, Y. & Hinton, G. Deep learning. Nature 521, 436–444 (2015).
Kim, H. et al. Web-based fully automated cephalometric analysis by deep learning. Comput. Methods Programs Biomed. 194, 105513 (2020).
Kim, M.-J. et al. Automatic cephalometric landmark identification system based on the multi-stage convolutional neural networks with CBCT combination images. Sensors 21, 505 (2021).
Song, Y., Qiao, X., Iwamoto, Y., Chen, Y.-W. & Chen, Y. An efficient deep learning based coarse-to-fine cephalometric landmark detection method. IEICE Trans. Inf. Syst. 104, 1359–1366 (2021).
Valentin, J. The 2007 recommendations of the international commission on radiological protection. ICRP Publ. 103(37), 2–4 (2007).
Milborrow, S. & Nicolls F. Locating facial features with an extended active shape model. in Proceeding of the European Conference on Computer Vision. 504–513 (Springer, 2008).
Cristinacce, D., Cootes, T. F.et al. Feature detection and tracking with constrained local models. in Proceeding of the 17th British Machine Vision Conference. 1.3 (Citeseer, 2006).
Dollár, P., Welinder, P. & Perona, P. Cascaded pose regression. in Proceedings of the IEEE Conference on Computer Vision Pattern Recognition. 1078–1085 (IEEE, 2010).
Sun, X., Wei, Y., Liang, S., Tang, X. & Sun, J. Cascaded hand pose regression. in Proceedings of the IEEE Conference on Computer Vision Pattern Recognition. 824–832 (2015).
Khabarlak, L. & Koriashkina, K. Fast facial landmark detection and applications: A survey. J. Comput. Sci. Technol. 22, e02 (2022).
Newell, A., Yang, K. & Deng, J. Stacked hourglass networks for human post estimation. in Proceedings of the European Conference Computer Vision. 483–499. (Springer, 2016).
Huang, Y., Yang, H., Li, C., Kim, J. & Wei, F. Adnet: Leveraging error-bias towards normal direction in face alignment. arXiv:2109.05721(2021).
Bulat, A., Sanchez, E. & Tzimiropoulos, G. Subpixel heatmap regression for facial landmark localization. arXiv:2111.02360(2021).
Sun, K. et al. High-resolution representations for labeling pixels and regions. arXiv:1904.04514(2019).
Sun, K., Xiao, B., Liu, D. & Wang, J. Deep high-resolution representation learning for human pose estimation. in Proceedings of the IEEE Conference on Computer Vision Pattern Recognition. 5693–5703 (2019).
Wu, W.et al. Look at boundary: A boundary-aware face alignment algorithm. in Proceedings of the IEEE Conference on Computer Vision Pattern Recognition. 2129–2138 (2018).
Wang, X., Bo, L. & Fuxin, L. Adaptive wing loss for robust face alignment via heatmap regression. in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV). 6971–6981 (2019).
Takahashi, K., Shimamura, Y., Tachiki, C., Nishii, Y. & Hagiwara, M. Cephalometric landmark detection without X-rays combining coordinate regression and heatmap regression. Sci. Rep. 13, 20011 (2023).
Downs, W. B. Variation in facial relationships: Their significance in treatment and prognosis. Am. J. Orthod. 34, 812–840 (1984).
Ricketts, R. M. A foundation for cephalometric communication. Am. J. Orthod. 46, 330–357 (1960).
Graber, T. M. A critical review of clinical cephalometric radiography. Am. J. Orthod. 40, 1–26 (1954).
Shrout, P. E. & Fleiss, J. L. Intraclass Correlations: Uses in assessing rater reliability. Psychol Bull. 86, 420–428 (1979).
Landis, J. R. & Koch, G. G. The measurement of observer agreement for categorical data. Biometrics 33, 159–174 (1977).
Wang, C.-W. et al. A benchmark for comparison of dental radiography analysis algorithms. Med. Image Anal. 31, 63–76 (2016).
Gilmour, L.& Ray, N. Locating cephalometric x-ray landmarks with foveated pyramid attention. in Proceeding of the International Conference on Medicine Imaging Deep Learning (MIDL). 262–276 (PMLR, 2020).
Li, W. et al. Structured landmark detection via topology-adapting deep graph learning. in Proceedings of the European Conference on Computer Vision. 266–283 (Springer, 2020).
Kwon, H. J., Koo, H. I., Park, J. & Cho, N. I. Multistage probabilistic approach for the localization of cephalometric landmarks. IEEE Access 9, 21306–21314 (2021).
Oh, K. et al. Deep anatomical context feature learning for cephalometric landmark detection. IEEE J. Biomed. Heal. Inf. 25, 806–817 (2020).
Huang, Z. Y. et al. A study on computer vision for facial emotion recognition. Sci. Rep. 13, 8425 (2023).
Michiel, A. et al. Fine-Tuning Language Models to Find Agreement among Humans with Diverse Preferences. arXiv:2211.15006 (2022).
Devlin, J., Chang, M. -W., Lee, K., Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT). 4171–4186 (2019).
Shibata, Y., Hiraide, T., Shibasaki, Y. & Fukuhara, T. About the error of the facial profile roentgenographic cephalogram. J. Showa. 1, 170–178 (1982).
Jacobson, A. Update on the wits appraisal angle orthodontist. Angle Orthod. 58, 205–219 (1988).
Thayer, T. A. Effects of functional versus bisected occlusal planes on the Wits appraisal. Am. J. Orthod. Dentofacial. Orthop. 97, 422–426 (1990).
Chang, Q. et al. Automatic analysis of lateral cephalograms based on high-resolution net. Am. J. Orthod. Dentofacial. Orthop. 163, 501–508 (2023).
Jeelani, W., Fida, M. & Shaikh, A. The maxillary incisor display at rest: Analysis of the underlying components. Dental Press J. Orthod. 23, 48–55 (2018).
Maddalone, M., Losi, F., Rota, E. & Baldoni, M. G. Relationship between the position of the incisors and the thickness of the soft tissues in the upper jaw: Cephalometric evaluation. Int. J. Clin. Pediatr. Dent. 12, 391–397 (2019).
Krizhevsky, A., Sutskever, I. & Hinton, G. E. ImageNet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 25, 1097–1105 (2012).
Moon, J. H. et al. How much deep learning is enough for automatic identification to be reliable?. Angle Orthod. 90, 823–830 (2020).
Funding
This works was supported by JSPS KAKENHI Grant Number JP24K13203.
Author information
Authors and Affiliations
Contributions
Y.S. contributed to the study design and study concept, data collection, data labeling experiments, and manuscript writing. K.T. contributed to the study concept, building and learning the deep learning models. C.T. contributed to the study design, data collection, data labeling and critical revision of the manuscript. T.T. contributed to the study design, data collection, and data labeling. S.M. contributed to the study design, and critical revision of the manuscript. M.H. contributed to the study concept and critical revision of the manuscript. Y.N. contributed to the study design, data collection, data labeling, and critical revision of the manuscript. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
About this article
Cite this article
Shimamura, Y., Tachiki, C., Takahashi, K. et al. Accuracy of cephalometric landmark and cephalometric analysis from lateral facial photograph by using CNN-based algorithm. Sci Rep 14, 31089 (2024). https://doi.org/10.1038/s41598-024-82230-z
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41598-024-82230-z






