Svoboda | Graniru | BBC Russia | Golosameriki | Facebook
Next Article in Journal
Exploring Health Literacy and Vascular Access Decision Making: A Scoping Review
Previous Article in Journal
Experiences of Patients with Breast Cancer Participating in a Prehabilitation Program: A Qualitative Study
Previous Article in Special Issue
Reliability of the AI-Assisted Assessment of the Proximity of the Root Apices to Mandibular Canal
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparison of Three Commercially Available, AI-Driven Cephalometric Analysis Tools in Orthodontics

by
Wojciech Kazimierczak
1,2,*,
Grzegorz Gawin
2,
Joanna Janiszewska-Olszowska
3,
Marta Dyszkiewicz-Konwińska
4,
Paweł Nowicki
2,
Natalia Kazimierczak
2,
Zbigniew Serafin
1 and
Kaan Orhan
5,6,7
1
Department of Radiology and Diagnostic Imaging, Collegium Medicum, Nicolaus Copernicus University in Torun, Jagiellońska 13-15, 85-067 Bydgoszcz, Poland
2
Kazimierczak Private Medical Practice, Dworcowa 13/u6a, 85-009 Bydgoszcz, Poland
3
Department of Interdisciplinary Dentistry, Pomeranian Medical University in Szczecin, 70-111 Szczecin, Poland
4
Department of Diagnostics, Poznań University of Medical Sciences, 61-701 Poznań, Poland
5
Department of DentoMaxillofacial Radiology, Faculty of Dentistry, Ankara University, Ankara 06500, Turkey
6
Medical Design Application and Research Center (MEDITAM), Ankara University, Ankara 06500, Turkey
7
Department of Oral Diagnostics, Faculty of Dentistry, Semmelweis University, 1085 Budapest, Hungary
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2024, 13(13), 3733; https://doi.org/10.3390/jcm13133733
Submission received: 11 June 2024 / Revised: 20 June 2024 / Accepted: 24 June 2024 / Published: 26 June 2024
(This article belongs to the Special Issue Emerging Technologies for Dental Imaging)

Abstract

:
Background: Cephalometric analysis (CA) is an indispensable diagnostic tool in orthodontics for treatment planning and outcome assessment. Manual CA is time-consuming and prone to variability. Methods: This study aims to compare the accuracy and repeatability of CA results among three commercial AI-driven programs: CephX, WebCeph, and AudaxCeph. This study involved a retrospective analysis of lateral cephalograms from a single orthodontic center. Automated CA was performed using the AI programs, focusing on common parameters defined by Downs, Ricketts, and Steiner. Repeatability was tested through 50 randomly reanalyzed cases by each software. Statistical analyses included intraclass correlation coefficients (ICC3) for agreement and the Friedman test for concordance. Results: One hundred twenty-four cephalograms were analyzed. High agreement between the AI systems was noted for most parameters (ICC3 > 0.9). Notable differences were found in the measurements of angle convexity and the occlusal plane, where discrepancies suggested different methodologies among the programs. Some analyses presented high variability in the results, indicating errors. Repeatability analysis revealed perfect agreement within each program. Conclusions: AI-driven cephalometric analysis tools demonstrate a high potential for reliable and efficient orthodontic assessments, with substantial agreement in repeated analyses. Despite this, the observed discrepancies and high variability in part of analyses underscore the need for standardization across AI platforms and the critical evaluation of automated results by clinicians, particularly in parameters with significant treatment implications.

1. Introduction

Artificial intelligence (AI), a term coined in 1956 by John McCarthy, describes the ability of machines to imitate logical human behavior [1]. Recent advancements in AI technology have led to the incorporation of this technology into many fields of everyday life, including internet search engines (Google), private online assistants (Siri, Alexa), and housekeeping (iRobot). The development of AI has also made its way into the field of medicine, particularly radiology, where medical imaging in 2023 constituted approximately 85% of the FDA-approved AI programs [2]. With its significant role in imaging for treatment planning and outcome monitoring, orthodontics is one of the fields of dentistry where AI tools are being implemented most rapidly.
Since 1931, when Broadbent and Hofrath simultaneously developed a standardized method to obtain lateral cephalometric radiographs, cephalometric analysis (CA) has remained a fundamental tool used in orthodontics [3]. It allows for the precise assessment of the mandible, maxilla, and cranial base in the sagittal and vertical dimensions [4]. It involves the use of X-ray lateral cephalograms of the head and face to obtain precise linear and angular measurements between predefined landmarks. CA allows for the assessment of projected growth directions in children and adolescents, the diagnosis of malocclusion, precise treatment planning, and posttreatment evaluation. In addition to orthodontics, CA is a valuable tool in orthognathic surgery planning, ensuring precise assessment and intervention [5]. Moreover, cephalometry is used to measure changes in the pharynx and other anatomical structures in patients with obstructive sleep apnea, especially after surgical treatment [6]. Despite its high diagnostic value, CA remains a burdensome task—it is associated with the labor-intensive and time-consuming process of identifying cephalometric landmarks. Currently, the time-consuming manual measurements have been replaced with digital CA software, which facilitates quicker measurements and calculations, as well as the automatic presentation of analysis results. The digitalization of CAs has been shown to reduce the number of errors resulting from manual measurements made with a ruler and protractor [7].
The main drawbacks of manual CA are its high operator dependency and significant variability in landmark identification [8,9,10,11,12,13,14,15,16,17,18,19]. Even analyses conducted by expert orthodontists suffer from significant intrareader variability [20]. Recent advancements in artificial intelligence (AI) technology have led to the introduction of AI in many areas of dental radiology, including CA [21,22]. Compared to manual CA, automated CA is relatively stable and repeatable [23]. Several studies have reported the reliability of AI algorithms in CA, demonstrating both high diagnostic accuracy and reduced analysis time [19,24,25,26,27]. Hopefully, this approach can lead to improved workflow and productivity in dental practices.
The integration of AI in dental diagnostics has paved the way for the development of AI-based commercially available programs such as AudaxCeph (Audax, Ljubljana, Slovenia), WebCeph (Assemble Circle, Seoul, Republic of Korea), and CephX (ORCA Dental AI, Las Vegas, NV, USA). AI algorithms in CA utilize deep learning (DL) and convolutional neural networks (CNNs) to automate the identification of anatomical landmarks on radiographs. These algorithms are trained on large datasets of labeled images to learn the patterns and features associated with specific anatomical landmarks. These programs automate the identification of cephalometric points, evaluate landmarks, calculate angles and distances, and generate automated analysis reports with diagnoses. The primary advantage of such software is the ability to automatically perform CA within seconds [13]. Studies conducted to date have demonstrated a high degree of agreement between manual and AI-automated CA performed by the mentioned software [28,29,30,31]. However, to the best of our knowledge, the agreement among the results from automated CA has not been assessed.
The present study aimed to compare the agreement of the results from randomly selected, three commercially available AI tools in automated CA in one patient cohort and to assess the repeatability of the AI results. Our hypothesis is that AI-driven cephalometric analysis tools demonstrate high accuracy and interchangeability.

2. Materials and Methods

2.1. Patient Population, Sample Size Calculation

The material of this retrospective study initially consisted of 130 lateral cephalograms obtained from patients aged 12 to 20 years from the patient archives of a single, private orthodontic center. The cephalograms were selected from the initial records of new patients admitted between 2018 and 2023. After initial screening, the images were anonymized. All the cephalograms were performed on the same digital panoramic machine, Hyperion X9-Pro (MyRay, Verona, Italy). The primary indication for lateral cephalograms was orthodontic treatment planning. The selected lateral cephalograms were manually uploaded into the databases of the chosen AI programs without any image modifications (cropping, contrast adjustments, filters, etc.).
The sample size was validated according to a paper by Bonnet titled “Sample Size Requirements for Estimating Intraclass Correlations with Desired Precision” [31]. The sample size calculations were conducted using a web-based calculator (https://wnarifin.github.io/ssc/sssnsp.html, accessed on 6 May 2024). The following assumptions were made: ICC was calculated for each of the assessed parameters with a precision of 0.1, a confidence level of 90%, and a number of raters of 3.

2.2. Eligibility Criteria

All patients, aged 12–20 years, with lateral cephalograms acquired during the treatment planning, were consecutively enrolled in this study. Patients aged 12–20 years were selected for this study as this age range represents the common period during which orthodontic treatment is initiated and actively managed. Adolescents and young adults are the primary demographic for orthodontic interventions, making this sample representative of the population typically undergoing cephalometric analysis in clinical practice. The eligibility criteria are listed in Table 1.

2.3. Automatic Cephalometric Analysis

The selected lateral cephalograms were manually uploaded into the following databases of the chosen AI programs: CephX, WebCeph, and AudaxCeph. The selection of the programs included in this study was based on the commercial availability, widespread use in clinical practice, and the ability to perform fully automated cephalometric analyses. This ensures the relevance and applicability of our findings to practitioners.
The software automatically selected types of CAs and generated automatic reports. For the analysis, the common measurements from all three programs according to Downs, Ricketts, and Steiner were utilized. No manual adjustments to cephalometric landmarks were made to assess the fully automatic process of CA. The analyzed parameters are listed in Table 2.

2.4. Repeatability Analysis

Fifty randomly selected subjects were reuploaded as the new patients and reanalyzed by all three evaluated platforms. The intraclass correlation coefficient (ICC3) values for repeated CAs were calculated to assess the agreement among the results.

2.5. Statistical Analysis

The concordance of measurements of quantitative variables was assessed with ICC type 3 (according to the Shrout and Fleiss classification) [32]. The Friedman test was used to compare three or more repeated measures of quantitative variables. Paired Wilcoxon tests with Bonferroni correction served as post hoc procedures. The paired Wilcoxon test was used to compare two repeated measures of quantitative variables. The significance level for all the statistical tests was set to 0.05. All the analyses were conducted with R software, version 4.3.3.

3. Results

3.1. Population, Sample Size Calculation

After the exclusion of six cephalometric radiographs due to poor image quality (2), presence of artifacts (1), or significant double borders of the mandible (3), 124 patients (59 men, 65 women) aged 12–20 years (mean age of 14.4 years) were included in this study.
The minimum sample size calculation (n = 104) showed that our sample was sufficient for the validity of the results.
Figure 1 shows the cephalogram of a sample patient with superimposed cephalometric points.

3.2. The Results from Automated CA

The results of the analyses are presented in Table 3. Most of the analyses revealed similar mean calculation values; however, significant discrepancies were found in some of the analyzed parameters. The largest differences were demonstrated in the measurements of angular values. The greatest discrepancies were observed in the results of the angle convexity analysis, where CephX had a mean value of 176.32°, AudaxCeph had a mean value of 7.18°, and WebCeph had a mean value of 7.99°. Similar discrepancies were evident in the measurements of the occlusal plane angle, with CephX reporting a mean value of 42.8°, AudaxCeph of 6.11°, and WebCeph of 5.86°; the angle of the lower incisor (LI) to the occlusal plane—CephX 69.23°, AudaxCeph 20.31°, and WebCeph 20.62°; and the angle of the LI to the mandibular plane—CephX 87.13°, AudaxCeph 6.98°, and WebCeph 6.87°. The results of the other analyses performed showed some minor discrepancies.
Due to differences in the number of analyses performed by the AI programs, some measurements were performed only by part of the programs. A summary of the results of the analyses performed only by CephX and AudaXCeph can be found in Table 4.

3.3. Concordance Analysis

The results of the concordance analysis showed good to excellent concordance of the results of the analyses for most of the parameters. However, some of the parameters showed poor and fair concordance. The detailed results of the concordance analysis of all three selected platforms are shown in Table 5. Comparisons of the results of the sample analyses showing excellent and poor concordance are shown in Figure 2 and Figure 3, respectively.
The results of the concordance analysis of the two CA platforms are shown in Table 6.

3.4. Repeatability Analysis

The results of the repeated analyses for 50 patients, as performed by each of the three programs, showed perfect agreement; each program returned the same results for all the repeated analyses performed.

4. Discussion

The present study aimed to compare the variability of the CA results of three commercially available AI-automated CA tools—CephX, WebCeph, and AudaxCeph. Our results demonstrated a high level of agreement among the AI-driven automated systems in the CA for most of the parameters evaluated. The repeatability analysis showed perfect agreement within each program, indicating that the automated systems can produce consistent results when reanalyzing the same radiographs, proving the accuracy of the algorithms. This suggests that AI-driven tools can offer a reliable alternative to traditional methods, with the added benefits of speed and consistency. Notably, significant discrepancies were observed in some angular measurements, such as angle convexity and occlusal plane angles, indicating that different methodologies were adopted by the selected platforms.
AI-driven tools offer significant advantages in CA, potentially leading to improved diagnostic accuracy, consistency in landmark identification, and a reduction in the time required for cephalometric analysis. These tools have the potential to enhance clinical decision-making, streamline workflows, and reduce the risk of human error, ultimately leading to better patient outcomes. However, the results of our study indicate significant variability in the results of several analyses among the programs. These discrepancies could be attributed to differences in algorithms, methods of evaluation, and landmark recognition capabilities across the three platforms. The highest mean differences were found in the facial angle of convexity, defined as angle convexity (CephX), angle of convexity (AudaXCeph), and N-A-Pg (WebCeph). Considering the average results of the analysis, along with their SDs, the only possibility for explaining these differences was a completely different measurement method. The inspection of the CA results of individual patients revealed that CephX indicated entirely different normal ranges than the two other programs. CephX indicated a correct value of 180 ± 5°, while the other programs indicated a value of 0 ± 5°. After eliminating CephX from this analysis, AudaXCeph and WebCeph showed very similar results, with 7.18 ± 4.53 and 7.99 ± 4.51, respectively.
The angle of convexity is usually defined as the angle formed by the intersection of two lines drawn from the most anterior point on the maxilla (Point A) and the most anterior point on the mandible (Point B) to the point on the forehead (nasion). This measurement is used to assess the relationship between the mandible and maxilla and the overall facial profile. A larger angle of convexity usually indicates a more convex facial profile, which can be associated with a protruding upper jaw or a receding lower jaw. A smaller angle indicates a straighter or more concave facial profile. The angle of convexity is a substantial parameter of CA, although its definition varies among authors. Some use the soft tissue glabellar point [33,34], the frontal point (Fr) [35], the NS point [36], or the N′ point at the depression of the nose as cranial reference points [37,38]. Godt et al. have proven that variance in the definition of landmarks used in facial convexity measurement methods can lead to significant discrepancies in the obtained angle values [39]. The differences in facial convexity automated measurement values and the normal ranges between CephX and the two other employed programs clearly show that the adapted methodology was different (Figure 3A). This example clearly shows that CephX indicated entirely different mean values and CI for angle convexity compared to AudaXCeph and WebCeph, suggesting a different measurement approach.
Similar discrepancies were found in the results of the occlusal plane calculations. These calculations were defined as the FH to the occipital plane by both CephX and AudaXCeph and as the cant of the occlusal plane by WebCeph. Again, the mean values shown by WebCeph and AudXCeph were similar, but the mean results of the CephX calculations were entirely different, yielding the results of 5.86 ± 4.05, 6.11 ± 4.01, and 42.8 ± 72.23, respectively. Notably, the results obtained by CephX showed significant variability (SD = 72.23, min = 0.13, max 179.93) compared to the significantly lower variability of the results from the rest of the analyzed platforms (Figure 3B).
The occlusal plane in the CA is an imaginary surface drawn through the incisal edges and occlusal surfaces of the teeth. It represents the mean curvature of the surface drawn rather than the actual plane. The measurement of the occlusal plane typically involves determining its angle relative to other anatomical planes or structures in the head and neck. As shown in a review by Mazurkiewicz et al., the occlusal plane might be evaluated with many different methods and devices; thus, discrepancies might be obtained [40]. The cant of the occlusal plane is defined as the vertical alignment of the teeth when there is a difference between the left and right sides. This involves either an upward or downward rotation of one side over the other in the transverse plane. The different methodologies adopted by CephX explain the discrepancies among the selected programs. However, the variability of the results from the CephX analyses (SD = 72.23, min = 0.13, max = 179.93) indicates significant discrepancies in the obtained results and raises considerable doubts about the reliability of these measurements. The results of the CephX occlusal plane measurements showed values as high as 179.93 in some patients, whereas the indicated average normal value was 9.3 ± 3.8. The same patients, assessed by AudaxCeph and WebCeph, had values of 4.68589 and 4.1, respectively. The differences shown in Figure 3B raise concerns regarding the reliability of CephX in occlusal plane measurements.
Another parameter that showed significant discrepancies was the lower incisor (LI) to the occlusal plane (Figure 3C). The mean results for CephX, AudaxCeph, and WebCeph were 69.23°, 20.31°, and 20.62°, respectively. Similarly, for the LI to the mandibular plane, the mean values were 87.13°, 6.98°, and 6.87°, respectively (Figure 3D). These significant differences in the mean values indicate that the three AI programs may have used different methods for locating the landmarks and performing the measurements. It is worth noting that these parameters are crucial in the diagnosis and treatment planning for many orthodontic cases, including those involving anterior open bite, deep overbite, and skeletal Class III malocclusion [41]. Hence, the discrepancies in these measurements among the different AI programs may lead to different diagnoses and treatment plans.
Despite the significant discrepancies mentioned above in the results, the vast majority of the results showed perfect agreement (ICC type 3 > 0.9). These discrepancies are in some cases related to the significantly different results of one of the three programs, likely due to different measurement methodologies. In addition to the drastically different values indicated by CephX, as discussed above, other parameters should also be mentioned. The results of the mandibular arc analysis showed variability in the mean results of all three tested platforms (Figure 4A). Moreover, the results of CephX analyses showed the highest variability, which was not comparable to that of other platforms, ranging from 15.94 to 69.51. Similarly, AudaxCeph showed high variability in MAND. 1 to APo (lower incisor—A point—pogonion) measurements, with a maximum value measured at 38.42 mm. In contrast, the maximum values indicated by CephX and WebCeph were 7.79 and 7.0, respectively (Figure 4B). The other parameters with fair agreement were the linear measurements of the distance between the lower incisor N and B or A point (/I to NA,/I to NB), with the highest variability in the AudaxCeph measurements (Figure 4C,D). However, all the indicated parameters (mandibular arc, MAND. 1 to Apo,/I to NA,/I to NB) presented 95% confidence intervals (CIs), and the standard deviations (SDs) calculated were comparable to those of other platforms, indicating likely incidental errors in landmark identification. Notably, fair concordance was found only for the distance measurements of/I to NA and/I to NB parameters, whereas the angular measurements showed perfect agreement.
To date, we have conducted two studies on CephX reliability. In the 2023 study [42], we assessed the concordance of repeated automatic CA from computed tomography (CT) images. Our results displayed good-to-excellent concordance in a large majority of measurements, except for two angular measurements: lower incisor–nasion–B-point (LI-N-B) and prosthion–nasion–A-point (Pr-N-A). The repeated analyses showed concordance results of 0 and 0.302, respectively, indicating significant discrepancies in the repeated measurements. However, it is important to note that the concordance in the repeated CA results of the current study was perfect, accounting for 1.0. These facts might demonstrate both the improvement of the automated CA module and the challenges of performing CA based on CT images. Our 2024 study [43] investigated the reliability of automated facial asymmetry assessment. Our results showed very low agreement between automatic and manual facial asymmetry measurements (ICC < 0.3). Furthermore, we found a total lack of agreement between the AI and manual asymmetry rate analyses (ICC type 3  =  0). As of the date of manuscript preparation (April 2024), CephX’s automated asymmetry assessment module is no longer available.
A recent study by Yassir et al. [44] evaluated the accuracy and reliability of WebCeph in CA. The authors reported problems with landmark identification and soft tissue delineation, and inconsistency of measurements are inherent features of the program’s automated analyses. Similar to our results, most of the discrepancies were found in the angular measurements. The 2024 study by Silva et al. [45] compared the accuracy of WebCeph and CefBot for landmark identification to that of 10 experienced readers. The authors concluded that CefBot exhibited excellent reliability and was ready for use in clinical practice, while WebCeph produced significant errors in landmark identification. A study on AudaXCeph’s tracing reliability by Ristau et al. showed that the program was not significantly different from that of experienced orthodontists [31]. However, some discrepancies in lower incisor apex measurements were found. These problems with lower incisor identification were also found in our study, yielding the high variability shown in the measurements presented in Figure 4C,D. Ribwar and Azeez compared the results of CephX and manual tracings [46]. The authors have shown that, except for several parameters, the results of the automated analysis showed high agreement with the manual method. The article concluded that CephX is adequate for clinical use.
Along with a growing body of evidence on the use of AI-automated CA in experimental studies, recent meta-analyses have been published providing a comprehensive overview of its accuracy and reliability [47,48,49,50,51,52,53]. However, most of the tested AI models were experimental and not available for common users. Furthermore, the results depend on the predefined thresholds. As expected, the accuracy sharply decreases when the threshold is lower than 2 mm [47,49,52]. A systematic review and meta-analysis by Schwendicke et al. [48] assessed the accuracy of deep learning (DL) for cephalometric landmark detection on 2D and 3D radiographs. The meta-analysis, which included 19 studies published between 2017 and 2020, revealed that DL exhibited relatively high accuracy in detecting cephalometric landmarks. However, the body of evidence suffers from a high risk of bias, highlighting the need for further studies to demonstrate the robustness and generalizability of DL for landmark detection. Rauniyar et al. [52] conducted a systematic review to determine the accuracy of identifying cephalometric landmarks using AI and compared the results with those of a manual tracing group. The review concluded that AI showed extremely positive and promising results compared to manual tracing, indicating its potential in CA. A meta-regression conducted by Serafin et al. [53] in a meta-analysis on AI-automated cephalometric landmarks indicated a significant relationship between the mean error in landmarking and the year of publication (p-value = 0.012). The authors concluded that the accuracy of the AI algorithms in this task has risen significantly in studies published between 2021 and 2023. These results give hope for the further development of algorithms, their refinement, and the possibility of their application in daily clinical practice.
The AI-driven CA tools used in this study are trained on large datasets of labeled images using deep learning and convolutional neural networks. These algorithms learn to identify anatomical landmarks by recognizing patterns and features from the training data. However, the specific details of these datasets and training processes are proprietary. The program’s developers refrain from providing information on this topic and treat it as a trade secret. The three AI-driven CA tools used in this study—CephX, WebCeph, and AudaxCeph—were selected based on their commercial availability, widespread use in clinical practice, and the ability to perform fully automated cephalometric analyses. These programs are among the most used AI tools in orthodontics, making this study relevant to practitioners. While other AI-driven tools exist, our access was limited to these three programs. Future studies should aim to include a broader range of AI tools to provide a more comprehensive evaluation.
A reduction in analysis time without compromising accuracy can potentially enhance productivity in orthodontic practices. Moreover, the consistency of AI-driven systems reduces the risk of human error, thus providing more standardized outcomes. However, the variations observed in certain measurements highlight the need for standardization among different AI platforms. This variation could lead to different orthodontic diagnoses and treatment plans, raising concerns about the interchangeability of these systems. It also indicates the necessity of thoroughly familiarizing oneself with the methodology of the platform used before its application. Additionally, our study demonstrated the presence of evident calculation results stemming from erroneous cephalometric landmark identification. However, practitioners must be aware of the limitations of these tools, especially regarding the discrepancies found in some angular measurements. It is imperative that practitioners critically evaluate the results from AI-driven systems and, if necessary, confirm the findings manually, especially when the measurements have significant implications for treatment decisions [54].
Despite some discrepancies, this study found high levels of agreement and repeatability in the results obtained from the AI programs for most cephalometric parameters. This suggests the potential of AI to deliver consistent and reliable analyses. The perfect repeatability of results across all evaluated programs underscores the consistency of automated analysis. AI reduces the risk of human error associated with manual CA. This standardization, coupled with high accuracy in landmark detection, can significantly improve the reliability of the assessments. In our view, our study offers valuable data that can guide the further development and refinement of AI algorithms, potentially expanding their use in clinical settings. However, this implies the need for further studies—evaluating different software with broader and more heterogeneous study groups—to provide a comprehensive evaluation of existing and future technologies. Our findings underscore the need for methodological standardization and algorithm improvement, as both factors influenced our results, indicating that current AI tools still require human supervision. Moreover, there is a vast yet insufficiently explored area of cone beam computed tomography (CBCT)-based CA. While some pilot studies have already assessed the accuracy of AI algorithms, further evaluation is still needed. Future studies should consider including a human comparison group (preferably multi-reader evaluation) to evaluate the performance of AI algorithms against experienced orthodontists. This would provide a more comprehensive understanding of the advantages and limitations of AI in cephalometry.
The findings of the present study should be interpreted considering several limitations. First, this study was retrospective in nature and relied on archived cephalograms, which might have affected the quality of some images. Moreover, there are geographic and ethnic limitations which may influence the generalizability of the results, especially in more diverse populations. Second, this study included a relatively small sample size, which may limit the generalizability of the findings. The results of the analyses were not compared to any “golden standard” based on the expert readers’ consensus; however, this was not the aim of this study. While this study demonstrates the high repeatability of AI-driven cephalometric analyses, the absence of a gold standard limits the ability to determine the most accurate tool. Future research should include manual evaluations by expert orthodontists as a benchmark to compare the AI tools, providing a comprehensive assessment of their accuracy. A further limitation of this study is the human involvement in the image selection, which implies that the process was not entirely AI-driven. Knowledge about image quality was required to exclude suboptimal radiographs, which could impact the analysis. Last, this study evaluated only three AI programs, and many other commercially available AI-driven automated CA tools were not included in the analysis.
Future studies should focus on evaluating additional AI-driven cephalometric analysis tools to provide a more comprehensive comparison. It would be beneficial to conduct studies involving larger and more diverse patient populations to enhance the generalizability of the findings. Moreover, integrating a human comparison group, including multi-reader evaluations by experienced orthodontists, could offer deeper insights into the performance of AI algorithms.

5. Conclusions

In conclusion, AI-driven automated CA tools can provide a quick and consistent alternative to traditional manual methods. However, significant discrepancies exist in the measurements of some parameters among different AI programs, which may potentially lead to varying diagnoses. Moreover, some parameters assessed by the selected AI platforms exhibited significant variability, indicating severe inaccuracies in landmark identification. Therefore, clinicians should be aware of these discrepancies, carefully interpreting the results of automated CA in conjunction with clinical findings and assessing the accuracy of landmark identification.

Author Contributions

Conceptualization, W.K., methodology, W.K.; software, W.K.; validation, W.K. and P.N.; formal analysis, G.G. and W.K.; investigation, W.K., N.K. and P.N.; resources, W.K.; data curation, W.K.; writing—original draft preparation, W.K.; writing—review and editing, W.K., J.J.-O., Z.S., M.D.-K. and K.O.; visualization, W.K.; supervision, W.K., Z.S. and K.O.; project administration, W.K.; funding acquisition, W.K. and Z.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study was conducted in accordance with the Declaration of Helsinki and approved by the Ethics Committee of Collegium Medicum, Nicolaus Copernicus University in Torun, Poland (protocol No. KB 227/2023, 10.04.20223), for studies involving humans.

Informed Consent Statement

Patient consent was waived due to the retrospective nature of this study and patient data anonymization.

Data Availability Statement

The data are available on request.

Acknowledgments

Łukasz Deryło, statistician, was involved in the data analysis.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. McCarthy, J.; Minsky, M.L.; Rochester, N.; Shannon, C.E. A Proposal for the Dartmouth Summer Research Project on Artificial Intelligence. AI Mag. 2006, 27, 12. [Google Scholar]
  2. McNabb, N.K.; Christensen, E.W.; Rula, E.Y.; Coombs, L.; Dreyer, K.; Wald, C.; Treml, C. Projected Growth in FDA-Approved Artificial Intelligence Products Given Venture Capital Funding. J. Am. Coll. Radiol. 2023, 21, 617–623. [Google Scholar] [CrossRef] [PubMed]
  3. Hans, M.G.; Martin Palomo, J.; Valiathan, M. History of Imaging in Orthodontics from Broadbent to Cone-Beam Computed Tomography. Am. J. Orthod. Dentofac. Orthop. 2015, 148, 914–921. [Google Scholar] [CrossRef]
  4. Leonardi, R.; Giordano, D.; Maiorana, F.; Spampinato, C. Automatic Cephalometric Analysis: A Systematic Review. Angle Orthod. 2008, 78, 145–151. [Google Scholar] [CrossRef]
  5. Devereux, L.; Moles, D.; Cunningham, S.J.; McKnight, M. How Important Are Lateral Cephalometric Radiographs in Orthodontic Treatment Planning? Am. J. Orthod. Dentofac. Orthop. 2011, 139, e175–e181. [Google Scholar] [CrossRef]
  6. Faria, A.C.; Xavier, S.P.; Silva, S.N.; Trawitzki, L.V.V.; De Mello-Filho, F.V. Cephalometric Analysis of Modifications of the Pharynx Due to Maxillo-Mandibular Advancement Surgery in Patients with Obstructive Sleep Apnea. Int. J. Oral Maxillofac. Surg. 2013, 42, 579–584. [Google Scholar] [CrossRef]
  7. Liu, J.K.; Chen, Y.T.; Cheng, K.S. Accuracy of Computerized Automatic Identification of Cephalometric Landmarks. Am. J. Orthod. Dentofac. Orthop. 2000, 118, 535–540. [Google Scholar] [CrossRef]
  8. Dobratulin, K.; Gaidel, A.; Kapishnikov, A.; Ivleva, A.; Aupova, I.; Zelter, P. The Efficiency of Deep Learning Algorithms for Detecting Anatomical Reference Points on Radiological Images of the Head Profile. In Proceedings of the ITNT 2020—6th IEEE International Conference on Information Technology and Nanotechnology, Samara, Russia, 26–29 May 2020. [Google Scholar]
  9. Park, J.H.; Hwang, H.W.; Moon, J.H.; Yu, Y.; Kim, H.; Her, S.B.; Srinivasan, G.; Aljanabi, M.N.A.; Donatelli, R.E.; Lee, S.J. Automated Identification of Cephalometric Landmarks: Part 1—Comparisons between the Latest Deep-Learning Methods YOLOV3 and SSD. Angle Orthod. 2019, 89, 903–909. [Google Scholar] [CrossRef] [PubMed]
  10. Kim, H.; Shim, E.; Park, J.; Kim, Y.J.; Lee, U.; Kim, Y. Web-Based Fully Automated Cephalometric Analysis by Deep Learning. Comput. Methods Programs Biomed. 2020, 194, 105513. [Google Scholar] [CrossRef]
  11. Grau, V.; Alcañiz, M.; Juan, M.C.; Monserrat, C.; Knoll, C. Automatic Localization of Cephalometric Landmarks. J. Biomed. Inform. 2001, 34, 146–156. [Google Scholar] [CrossRef]
  12. Yao, J.; Zeng, W.; He, T.; Zhou, S.; Zhang, Y.; Guo, J.; Tang, W. Automatic Localization of Cephalometric Landmarks Based on Convolutional Neural Network. Am. J. Orthod. Dentofac. Orthop. 2022, 161, E250–E259. [Google Scholar] [CrossRef]
  13. Nishimoto, S.; Sotsuka, Y.; Kawai, K.; Ishise, H.; Kakibuchi, M. Personal Computer-Based Cephalometric Landmark Detection with Deep Learning, Using Cephalograms on the Internet. J. Craniofacial Surg. 2019, 30, 91–95. [Google Scholar] [CrossRef]
  14. Kunz, F.; Stellzig-Eisenhauer, A.; Zeman, F.; Boldt, J. Artificial Intelligence in Orthodontics: Evaluation of a Fully Automated Cephalometric Analysis Using a Customized Convolutional Neural Network. J. Orofac. Orthop. 2020, 81, 52–68. [Google Scholar] [CrossRef] [PubMed]
  15. Yu, H.J.; Cho, S.R.; Kim, M.J.; Kim, W.H.; Kim, J.W.; Choi, J. Automated Skeletal Classification with Lateral Cephalometry Based on Artificial Intelligence. J. Dent. Res. 2020, 99, 249–256. [Google Scholar] [CrossRef]
  16. Lee, J.H.; Yu, H.J.; Kim, M.J.; Kim, J.W.; Choi, J. Automated Cephalometric Landmark Detection with Confidence Regions Using Bayesian Convolutional Neural Networks. BMC Oral Health 2020, 20, 270. [Google Scholar] [CrossRef]
  17. Hwang, H.-W.; Moon, J.-H.; Kim, M.-G.; Donatelli, R.E.; Lee, S.-J. Evaluation of Automated Cephalometric Analysis Based on the Latest Deep Learning Method. Angle Orthod. 2021, 91, 329–335. [Google Scholar] [CrossRef] [PubMed]
  18. Hwang, H.W.; Park, J.H.; Moon, J.H.; Yu, Y.; Kim, H.; Her, S.B.; Srinivasan, G.; Aljanabi, M.N.A.; Donatelli, R.E.; Lee, S.J. Automated Identification of Cephalometric Landmarks: Part 2—Might It Be Better than Human? Angle Orthod. 2020, 90, 69–76. [Google Scholar] [CrossRef] [PubMed]
  19. Nishimoto, S.; Kawai, K.; Fujiwara, T.; Ishise, H.; Kakibuchi, M. Locating Cephalometric Landmarks with Multi-Phase Deep Learning. J. Dent. Health Oral Res. 2023, 4, 1–13. [Google Scholar] [CrossRef]
  20. Stabrun, A.E.; Danielsen, K. Precision in Cephalometric Landmark Indentification. Eur. J. Orthod. 1982, 4, 185–196. [Google Scholar] [CrossRef]
  21. Kazimierczak, N.; Kazimierczak, W.; Serafin, Z.; Nowicki, P.; Nożewski, J.; Janiszewska-Olszowska, J. AI in Orthodontics: Revolutionizing Diagnostics and Treatment Planning—A Comprehensive Review. J. Clin. Med. 2024, 13, 344. [Google Scholar] [CrossRef]
  22. Abesi, F.; Jamali, A.S.; Zamani, M. Accuracy of Artificial Intelligence in the Detection and Segmentation of Oral and Maxillofacial Structures Using Cone-Beam Computed Tomography Images: A Systematic Review and Meta-Analysis. Pol. J. Radiol. 2023, 88, e256. [Google Scholar] [CrossRef] [PubMed]
  23. Kiełczykowski, M.; Kamiński, K.; Perkowski, K.; Zadurska, M.; Czochrowska, E. Application of Artificial Intelligence (AI) in a Cephalometric Analysis: A Narrative Review. Diagnostics 2023, 13, 2640. [Google Scholar] [CrossRef] [PubMed]
  24. Kang, S.H.; Jeon, K.; Kang, S.H.; Lee, S.H. 3D Cephalometric Landmark Detection by Multiple Stage Deep Reinforcement Learning. Sci. Rep. 2021, 11, 17509. [Google Scholar] [CrossRef] [PubMed]
  25. Chung, E.J.; Yang, B.E.; Park, I.Y.; Yi, S.; On, S.W.; Kim, Y.H.; Kang, S.H.; Byun, S.H. Effectiveness of Cone-Beam Computed Tomography-Generated Cephalograms Using Artificial Intelligence Cephalometric Analysis. Sci. Rep. 2022, 12, 20585. [Google Scholar] [CrossRef] [PubMed]
  26. Bulatova, G.; Kusnoto, B.; Grace, V.; Tsay, T.P.; Avenetti, D.M.; Sanchez, F.J.C. Assessment of Automatic Cephalometric Landmark Identification Using Artificial Intelligence. Orthod. Craniofac. Res. 2021, 24, 37–42. [Google Scholar] [CrossRef] [PubMed]
  27. Tsolakis, I.A.; Tsolakis, A.I.; Elshebiny, T.; Matthaios, S.; Palomo, J.M. Comparing a Fully Automated Cephalometric Tracing Method to a Manual Tracing Method for Orthodontic Diagnosis. J. Clin. Med. 2022, 11, 6854. [Google Scholar] [CrossRef] [PubMed]
  28. Meriç, P.; Naoumova, J. Web-Based Fully Automated Cephalometric Analysis: Comparisons between App-Aided, Computerized, and Manual Tracings. Turk. J. Orthod. 2020, 33, 142–149. [Google Scholar] [CrossRef]
  29. Alqahtani, H. Evaluation of an Online Website-Based Platform for Cephalometric Analysis. J. Stomatol. Oral Maxillofac. Surg. 2020, 121, 53–57. [Google Scholar] [CrossRef]
  30. Prince, S.T.T.; Srinivasan, D.; Duraisamy, S.; Kannan, R.; Rajaram, K. Reproducibility of Linear and Angular Cephalometric Measurements Obtained by an Artificial-Intelligence Assisted Software (WebCeph) in Comparison with Digital Software (AutoCEPH) and Manual Tracing Method. Dent. Press. J. Orthod. 2023, 28, e2321214. [Google Scholar] [CrossRef]
  31. Ristau, B.; Coreil, M.; Chapple, A.; Armbruster, P.; Ballard, R. Comparison of AudaxCeph®’s Fully Automated Cephalometric Tracing Technology to a Semi-Automated Approach by Human Examiners. Int. Orthod. 2022, 20, 100691. [Google Scholar] [CrossRef]
  32. Shrout, P.E.; Fleiss, J.L. Intraclass Correlations: Uses in Assessing Rater Reliability. Psychol. Bull. 1979, 86, 420–428. [Google Scholar] [CrossRef] [PubMed]
  33. Mauchamp, O.; Sassouni, V. Growth and Prediction of the Skeletal and Soft-Tissue Profiles. Am. J. Orthod. 1973, 64, 83–94. [Google Scholar] [CrossRef] [PubMed]
  34. Burstone, C.J. The Integumental Profile. Am. J. Orthod. 1958, 44, 1–25. [Google Scholar] [CrossRef]
  35. Muzj, E. Musical and Architectural Proportions in the Anatomy of the Facial System. An Anthropometric Approach. Angle Orthod. 1982, 52, 177–210. [Google Scholar] [PubMed]
  36. Subtelny, J.D. The Soft Tissue Profile, Growth and Treatment Changes. Angle Orthod. 1961, 31, 105–122. [Google Scholar]
  37. Ngan, P.; Hägg, U.; Yiu, C.; Merwin, D.; Wei, S.H. Soft Tissue and Dentoskeletal Profile Changes Associated with Maxillary Expansion and Protraction Headgear Treatment. Am. J. Orthod. Dentofac. Orthop. 1996, 109, 38–49. [Google Scholar] [CrossRef]
  38. Ruf, S.; Pancherz, H. Dentoskeletal Effects and Facial Profile Changes in Young Adults Treated with the Herbst Appliance. Angle Orthod. 1999, 69, 239–246. [Google Scholar] [CrossRef] [PubMed]
  39. Godt, A.; Müller, A.; Kalwitzki, M.; Göz, G. Angles of Facial Convexity in Different Skeletal Classes. Eur. J. Orthod. 2007, 29, 648–653. [Google Scholar] [CrossRef] [PubMed]
  40. Mazurkiewicz, P.; Oblizajek, M.; Rzeszowska, J.; Sobolewska, E.; Ey-Chmielewska, H.; Szymańska-Kozula, R. Determining the Occlusal Plane: A Literature Review. Cranio J. Craniomandib. Pract. 2022, 40, 341–347. [Google Scholar] [CrossRef]
  41. Lin, Y.; Guo, R.; Hou, L.; Fu, Z.; Li, W. Stability of Maxillary Protraction Therapy in Children with Class III Malocclusion: A Systematic Review and Meta-Analysis. Clin. Oral Investig. 2018, 22, 2639–2652. [Google Scholar] [CrossRef]
  42. Kazimierczak, N.; Kazimierczak, W.; Serafin, Z.; Nowicki, P.; Lemanowicz, A.; Nadolska, K.; Janiszewska-Olszowska, J. Correlation Analysis of Nasal Septum Deviation and Results of AI-Driven Automated 3D Cephalometric Analysis. J. Clin. Med. 2023, 12, 6621. [Google Scholar] [CrossRef] [PubMed]
  43. Kazimierczak, N.; Kazimierczak, W.; Serafin, Z.; Nowicki, P.; Jankowski, T.; Jankowska, A.; Janiszewska-Olszowska, J. Skeletal Facial Asymmetry: Reliability of Manual and Artificial Intelligence-Driven Analysis. Dentomaxillofacial Radiol. 2024, 53, 52–59. [Google Scholar] [CrossRef] [PubMed]
  44. Yassir, Y.A.; Salman, A.R.; Nabbat, S.A. The Accuracy and Reliability of WebCeph for Cephalometric Analysis. J. Taibah Univ. Med. Sci. 2022, 17, 57–66. [Google Scholar] [CrossRef]
  45. Silva, T.P.; Pinheiro, M.C.R.; Freitas, D.Q.; Gaêta-Araujo, H.; Oliveira-Santos, C. Assessment of Accuracy and Reproducibility of Cephalometric Identification Performed by 2 Artificial Intelligence-Driven Tracing Applications and Human Examiners. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. 2024, 137, 431–440. [Google Scholar] [CrossRef] [PubMed]
  46. Khalid, R.F.; Azeez, S.M. Comparison of Cephalometric Measurements of On-Screen Images by CephX Software and Hard-Copy Printouts by Conventional Manual Tracing. J. Hunan Univ. Nat. Sci. 2022, 49, 68–74. [Google Scholar] [CrossRef]
  47. de Queiroz Tavares Borges Mesquita, G.; Vieira, W.A.; Vidigal, M.T.C.; Travençolo, B.A.N.; Beaini, T.L.; Spin-Neto, R.; Paranhos, L.R.; de Brito Júnior, R.B. Artificial Intelligence for Detecting Cephalometric Landmarks: A Systematic Review and Meta-Analysis. J. Digit. Imaging 2023, 36, 1158–1179. [Google Scholar] [CrossRef] [PubMed]
  48. Schwendicke, F.; Chaurasia, A.; Arsiwala, L.; Lee, J.H.; Elhennawy, K.; Jost-Brinkmann, P.G.; Demarco, F.; Krois, J. Deep Learning for Cephalometric Landmark Detection: Systematic Review and Meta-Analysis. Clin. Oral Investig. 2021, 25, 4299–4309. [Google Scholar] [CrossRef] [PubMed]
  49. Londono, J.; Ghasemi, S.; Hussain Shah, A.; Fahimipour, A.; Ghadimi, N.; Hashemi, S.; Khurshid, Z.; Dashti, M. Evaluation of Deep Learning and Convolutional Neural Network Algorithms Accuracy for Detecting and Predicting Anatomical Landmarks on 2D Lateral Cephalometric Images: A Systematic Review and Meta-Analysis. Saudi Dent. J. 2023, 35, 487–497. [Google Scholar] [CrossRef] [PubMed]
  50. Jihed, M.; Dallel, I.; Tobji, S.; Amor, A. Ben The Impact of Artificial Intelligence on Contemporary Orthodontic Treatment Planning—A Systematic Review and Meta-Analysis. Sch. J. Dent. Sci. 2022, 9, 70–87. [Google Scholar] [CrossRef]
  51. Junaid, N.; Khan, N.; Ahmed, N.; Abbasi, M.S.; Das, G.; Maqsood, A.; Ahmed, A.R.; Marya, A.; Alam, M.K.; Heboyan, A. Development, Application, and Performance of Artificial Intelligence in Cephalometric Landmark Identification and Diagnosis: A Systematic Review. Healthcare 2022, 10, 2454. [Google Scholar] [CrossRef]
  52. Rauniyar, S.; Jena, S.; Sahoo, N.; Mohanty, P.; Dash, B.P. Artificial Intelligence and Machine Learning for Automated Cephalometric Landmark Identification: A Meta-Analysis Previewed by a Systematic Review. Cureus 2023, 15, e40934. [Google Scholar] [CrossRef] [PubMed]
  53. Serafin, M.; Baldini, B.; Cabitza, F.; Carrafiello, G.; Baselli, G.; Del Fabbro, M.; Sforza, C.; Caprioglio, A.; Tartaglia, G.M. Accuracy of Automated 3D Cephalometric Landmarks by Deep Learning Algorithms: Systematic Review and Meta-Analysis. Radiol. Med. 2023, 128, 544–555. [Google Scholar] [CrossRef] [PubMed]
  54. Waller, J.; O’connor, A.; Rafaat, E.; Amireh, A.; Dempsey, J.; Martin, C.; Umair, M. Applications and Challenges of Artificial Intelligence in Diagnostic and Interventional Radiology. Pol. J. Radiol. 2022, 87, 113–117. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Comparison of superimposed cephalometric landmarks on sample patient: (A) CephX, (B) AudaxCeph, (C) WebCeph.
Figure 1. Comparison of superimposed cephalometric landmarks on sample patient: (A) CephX, (B) AudaxCeph, (C) WebCeph.
Jcm 13 03733 g001
Figure 2. Graphs presenting sample CA parameters with excellent concordance among the 3 programs. (A)—facial axis; (B)—SNB. Mean values, 95% confidence intervals (CIs), ranges.
Figure 2. Graphs presenting sample CA parameters with excellent concordance among the 3 programs. (A)—facial axis; (B)—SNB. Mean values, 95% confidence intervals (CIs), ranges.
Jcm 13 03733 g002
Figure 3. Graphs presenting sample CA parameters with poor concordance among the 3 programs. (A)—facial convexity angle; (B)—occlusal plane; (C)—lower incisor to occlusal plane; (D)—lower incisor to the mandible. Mean values, 95% confidence intervals (CIs), ranges.
Figure 3. Graphs presenting sample CA parameters with poor concordance among the 3 programs. (A)—facial convexity angle; (B)—occlusal plane; (C)—lower incisor to occlusal plane; (D)—lower incisor to the mandible. Mean values, 95% confidence intervals (CIs), ranges.
Jcm 13 03733 g003
Figure 4. Graphs presenting sample CA parameters with poor concordance among the 3 programs. (A)—mandibular arc; (B)—MAND. 1 to APo; (C)—/I to NA; (D)—/I to NB. Mean values, 95% confidence intervals (CIs), ranges.
Figure 4. Graphs presenting sample CA parameters with poor concordance among the 3 programs. (A)—mandibular arc; (B)—MAND. 1 to APo; (C)—/I to NA; (D)—/I to NB. Mean values, 95% confidence intervals (CIs), ranges.
Jcm 13 03733 g004
Table 1. Eligibility criteria.
Table 1. Eligibility criteria.
CriteriaDescription
Inclusion CriteriaPatients aged 12–20 years
Lateral cephalograms obtained during orthodontic treatment planning
Good image quality without artifacts
Exclusion CriteriaPoor image quality
Presence of artifacts or asymmetries
Significant double borders of the mandible
Table 2. Categorized list of the assessed parameters as mentioned in the analyzed software.
Table 2. Categorized list of the assessed parameters as mentioned in the analyzed software.
No.Parameter (Unified Name)CephXWebCephAudaxCeph
Downs
1Facial AngleFacial AngleFH-N-PogFacial Angle
2Angle ConvexityAngle Convexity-N-A-Pg
3A-B PlaneA-B PlaneNANPg/AB
4Mandibular PlaneFH-GoGnMandibular PlaneFH/ML′
5Y-AxisFH-S-GnY-axisFH/Y
6Occlusal PlaneFH-Occ PlaneCant of the Occlusal PlaneFH/OcP
7Upper Incisor to Lower IncisorUI to LIInterincisal AngleInterincisal Angle
8Lower Incisor to Occlusal PlaneLI to Occ PlaneIncisor Occlusal Plane Angle−1/OcP
9Lower Incisor to Mandibular PlaneLI to MandIncisor Mandibular Plane Angle−1/ML′
10Upper Incisor to A-PogUI to A-PogUpper Incisor to A-Pog Line+1/APg
Ricketts
11Maxillary DepthFH to N-ANAMaxillary Depth
12Maxillary HeightN-PTV to A ptNANA
13SN to Palatal PlaneSN TO PALATAL PLANENANA
14Facial DepthFH to N-PogFacial DepthFacial Depth
15Facial AxisNa-Ba to PTV-Gn ptFacial AxisFacial Axis
16Facial TaperNa-Gn-GoFacial TaperFacial Taper
17Mandibular Plane (Ricketts)FH-GoGnMandibular Plane Angle (Ricketts)FMA
18Corpus LengthXi to PmNACorpus Axis
19Mandibular ArcDC-Xi to Xi-PmMandibular ArcMandibular Arc
20Point A ConvexityA to N-PogConvexity of Point AConvexity
21Lower Facial HeightANS-Xi-PogDenture Height (Lower Facial Height)Lower Facial Height
23Maxillary Incisor to A-PoMAX.1 to APoNANA
24Upper Molar to PTVMAX.6 to PTVUpper Molar to PtVNA
25Mandibular Incisor to A-PoMAND. 1 to APoL1 to A-Pog (mm)Lower 1 to APg (mm)
26Hinge Axis AngleDC-Go-LINANA
27Maxillary Incisor to Mandibular IncisorMAX.1 to MAND.1Intercisal AngleInterincisal Angle
28OverjetOverjetNANA
29OverbiteOverbiteNANA
31Upper Lip to E-LineUpper Lip to E-LineNANA
32Lower Lip to E-LineLower Lip to E-LineLower Lip to E-PlaneLi/E-Line
Steiner
33SNASNASNAAngle SNA
34SNBSNBSNBAngle SNB
35ANBANBANBANB
36Maxillary Incisor to NA (deg)I/to NAU1 to NA (deg)+1/NA
37Maxillary Incisor to NA (mm)I/to NAU1 to NA (mm)+1i/NA
38Mandibular Incisor to NB (deg)/I to NBL1 to NB (deg)−1/NB
39Mandibular Incisor to NB (mm)/I to NBL1 to NB (mm)−1i/NB
40Interincisal AngleInterincisal AngleInterincisal AngleInterincisal Angle
41Occlusal Plane to SNOcc to SNOcclusal Plane to SN AngleSN/OcP
42Mandibular Plane to SNGOGN-SNMandibular Plane Angle (Go-Gn to SN)SN/GoGn
43Pogonion to NBPog to NBNAPg/NB
NA—not assessed.
Table 3. Summary of the mean results of analyses performed by all three selected CA platforms.
Table 3. Summary of the mean results of analyses performed by all three selected CA platforms.
ParameterSoftwareMeanSDMedianMinMaxQ1Q3p
FACIAL ANGLECephX-A89.543.6489.0982.68101.2087.0891.62p = 0.001 *
B.A > C
AudaXCeph-B89.463.4689.2681.30103.3187.5591.05
WebCeph-C88.773.4389.0381.7899.8986.4990.56
ANGLE CONVEXITY (DOWNS)CephX-A176.3210.41174.00161.78222.66171.13179.30p < 0.001 *
A > C > B
AudaXCeph-B7.184.536.280.6619.683.779.69
WebCeph-C7.994.517.880.5020.434.1010.96
MAND. PLANE (DOWNS)CephX-A24.905.6123.2313.6337.3421.2129.57p < 0.001 *
A > B.C
AudaXCeph-B22.495.7922.0112.1234.3217.3027.68
WebCeph-C22.056.3121.0110.5533.2517.1327.44
Y-AXISCephX58.134.0058.1849.5265.3056.0361.06p = 0.064
AudaXCeph57.743.9057.6448.2866.2555.4160.21
WebCeph58.023.9358.2048.7565.0355.6761.61
OCCLUSAL PLANECephX-A42.8072.235.530.13179.931.6513.98p = 0.01 *
A > C
AudaXCeph-B6.114.015.380.3217.613.218.38
WebCeph-C5.864.055.270.0916.652.508.48
UI to LICephX-A130.0611.10129.31103.04167.22123.03135.18p = 0.029 *
B > A
AudaXCeph-B131.3911.25130.48106.21162.23123.94138.08
WebCeph-C130.2910.96129.28105.95160.59123.05137.04
LI to Occ PL.CephX-A69.238.3367.9355.17105.4963.2672.47p < 0.001 *
A > B.C
AudaXCeph-B20.317.3021.702.3236.0217.2225.68
WebCeph-C20.627.5220.761.8836.0816.0027.19
LI to MANDCephX-A87.137.9388.6056.63102.3383.6992.10p < 0.001 *
A > C.B
AudaXCeph-B6.984.726.350.0921.003.539.73
WebCeph-C6.874.487.390.1318.852.799.56
UI to A-PogCephX-A6.192.976.170.3114.713.888.38p < 0.001 *
A > B > C
AudaXCeph-B5.942.835.730.0312.213.998.48
WebCeph-C5.482.745.460.8213.893.667.58
FACIAL DEPTHCephX-A89.553.4289.4282.6898.5387.1091.91p = 0.003 *
A.B > C
AudaXCeph-B89.353.0789.3081.3097.5487.4591.15
WebCeph-C88.763.1589.0681.7897.5386.5090.60
FACIAL AXISCephX-A89.255.3990.2977.27102.2284.1992.59p < 0.001 *
B > A.C
AudaXCeph-B90.145.4490.9277.84103.7585.1693.68
WebCeph-C89.175.2790.1878.0599.9084.4592.83
FACIAL TAPERCephX-A66.914.7467.0855.1474.2064.0870.52p < 0.001 *
C > A.B
AudaXCeph-B66.905.0666.7953.2275.0764.1070.57
WebCeph-C72.236.6172.4153.4384.5069.3876.56
MAND. PLANE (RICKETTS)CephX-A24.545.6522.6113.6337.3420.9129.02p < 0.001 *
B > A > C
AudaXCeph-B23.755.7422.9513.5035.8118.7528.96
WebCeph-C21.896.2320.7610.5533.2516.7027.44
MAND. ARCCephX-A28.057.8228.3515.9469.5123.8531.75p < 0.001 *
C > B > A
AudaXCeph-B33.068.4532.3813.7648.6727.5339.69
WebCeph-C37.765.2139.1326.5146.2735.1241.85
A pt. CONVEXITYCephX-A3.462.143.340.179.282.264.42p = 0.022 *
B > A
AudaXCeph-B3.684.222.850.2729.821.784.27
WebCeph-C3.601.963.660.248.552.184.86
LOW.FACE.HEIGHTCephX-A41.155.2640.2029.3851.2537.0545.31p < 0.001 *
C > B > A
AudaXCeph-B43.566.6742.7132.3556.6837.7148.96
WebCeph-C44.236.0843.1431.4355.6639.1749.21
MAND.1 to APoCephX-A2.131.791.480.117.790.803.04p < 0.001 *
B > C.A
AudaXCeph-B3.315.262.640.0438.421.193.96
WebCeph-C2.111.731.730.077.000.743.26
MAX.1 to MAND.1CephX-A130.0611.10129.31103.04167.22123.03135.18p = 0.029 *
B > A
AudaXCeph-B131.3911.25130.48106.21162.23123.94138.08
WebCeph-C130.2910.96129.28105.95160.59123.05137.04
LOWER LIP to E-LINECephX2.602.111.900.029.781.044.12p = 0.981
AudaXCeph2.902.172.580.158.670.914.28
WebCeph2.672.012.500.036.981.084.13
SNACephX-A82.893.4983.2174.4889.7481.0685.23p = 0.039 *
A > B
AudaXCeph-B82.383.6782.7173.9089.7179.7384.83
WebCeph-C82.703.5083.1073.8489.4880.3385.22
SNBCephX-A79.594.2980.0069.1690.1377.3681.75p = 0.035 *
A > C.B
AudaXCeph-B79.324.3679.3768.7090.3676.9881.79
WebCeph-C79.094.1979.4867.4389.6577.3482.40
ANBCephX-A4.082.254.060.289.972.385.56p = 0.008 *
C > B
AudaXCeph-B3.852.183.830.3110.082.255.35
WebCeph-C4.352.194.560.038.802.735.79
I/to NA (deg)CephX-A23.328.3822.005.7644.3818.7128.05p < 0.001 *
A.C > B
AudaXCeph-B20.838.8020.511.1941.2416.0425.43
WebCeph-C21.998.4521.230.5539.5017.9227.02
I/to NA (mm)CephX-A4.052.344.060.138.902.465.56p = 0.011 *
A.B > C
AudaXCeph-B4.644.783.590.0633.652.415.94
WebCeph-C3.482.153.090.347.821.854.98
/I to NB (deg)CephX22.856.4624.652.1031.1419.1427.23p = 0.118
AudaXCeph24.036.8825.606.5134.9219.6829.18
WebCeph23.586.8824.684.2135.4319.4528.45
/I to NB (mm)CephX-A3.892.173.850.058.762.485.18p < 0.001 *
B > C > A
AudaXCeph-B4.982.884.620.1916.663.186.44
WebCeph-C4.452.334.070.0310.692.866.10
I/to/ICephX-A130.0611.10129.31103.04167.22123.03135.18p = 0.029 *
B > A
AudaXCeph-B131.3911.25130.48106.21162.23123.94138.08
WebCeph-C130.2910.96129.28105.95160.59123.05137.04
Occ to SNCephX-A12.595.1011.663.7127.118.9615.08p < 0.001 *
B > C > A
AudaXCeph-B14.794.8714.315.4924.3611.2317.98
WebCeph-C14.044.6813.402.4223.0310.6117.53
GOGN-SNCephX-A36.226.3535.5925.6847.2930.9441.98p < 0.001 *
A > B > C
AudaXCeph-B30.406.9130.4018.4841.9124.6536.83
WebCeph-C29.336.8528.2516.4142.2023.7635.33
p—Friedman test + post hoc analysis (paired Wilcoxon tests with Bonferroni correction). * statistically significant (p < 0.05).
Table 4. Summary of the mean results of analyses performed by two of the selected platforms (CephX and AudaXCepx).
Table 4. Summary of the mean results of analyses performed by two of the selected platforms (CephX and AudaXCepx).
ParameterMeasurementMeanSDMedianMinMaxQ1Q3p
A-B PLANECephX6.843.437.10.1415.354.239.68p < 0.001 *
AudaXCeph9.795.419.430.9523.645.814.06
MAX DEPTHCephX91.892.9791.4786.5499.1389.9894.02p = 0.004 *
AudaXCeph91.312.9991.4183.8497.4689.5293.06
CORPUS LENGTHCephX73.776.573.4339.6788.5370.2278.12p < 0.001 *
AudaXCeph72.1437.6967.8859.11391.4764.3670.01
POG to NBCephX1.961.591.7406.690.623.01p < 0.001 *
AudaXCeph2.461.662.210.186.241.243.47
p—Friedman test + post hoc analysis (paired Wilcoxon tests with Bonferroni correction). * statistically significant (p < 0.05).
Table 5. Results of the concordance analysis of all three CA platforms.
Table 5. Results of the concordance analysis of all three CA platforms.
ParameterICC95% CIAgreement (Cicchetti)Agreement (Koo and Li)
FACIAL ANGLE0.9100.8640.943ExcellentExcellent
ANGLE CONVEXITY (DOWNS)0.075−0.0730.254PoorPoor
MAND. PLANE (DOWNS)0.9030.8540.939ExcellentExcellent
Y AXIS0.9270.8890.954ExcellentExcellent
OCCLUSAL PLANE0.000−0.1370.172PoorPoor
UI to LI0.9420.9130.963ExcellentExcellent
LI to Occ PL.0.000−0.1370.172PoorPoor
LI to MAND0.3880.2220.553PoorPoor
UI to A-Pog0.9220.8810.951ExcellentExcellent
FACIAL DEPTH0.8980.8450.937ExcellentGood
FACIAL AXIS0.9750.9600.984ExcellentExcellent
FACIAL TAPER0.7890.6920.865ExcellentGood
MAND. PLANE (RICKETTS)0.9010.8490.938ExcellentExcellent
MAND. ARC0.5100.3480.659FairFair
A pt. CONVEXITY0.5320.3730.676FairFair
LOW.FACE.HEIGHT0.9470.9180.967ExcellentExcellent
MAND. 1 to APo0.4160.2460.582FairPoor
MAX.1 to MAND.10.9420.9130.963ExcellentExcellent
LOWER LIP to E-LINE0.7990.7050.871ExcellentGood
SNA0.8970.8430.935ExcellentGood
SNB0.9680.9500.981ExcellentExcellent
ANB0.9080.8600.943ExcellentExcellent
I/to NA (deg)0.9180.8730.949ExcellentExcellent
I/to NA (mm)0.4940.3310.646FairPoor
/I to NB (deg)0.9380.9030.961ExcellentExcellent
/I to NB (mm)0.7450.6330.834GoodFair
I/to/I0.9420.9130.963ExcellentExcellent
Occ to SN0.9030.8530.940ExcellentExcellent
GOGN-SN0.9310.8930.957ExcellentExcellent
ICC—interclass correlation coefficient, CI—confidence interval.
Table 6. Results of the concordance analysis of two of the selected platforms (CephX and AudaxCeph).
Table 6. Results of the concordance analysis of two of the selected platforms (CephX and AudaxCeph).
ParameterICC95% CIAgreement (Cicchetti)Agreement (Koo and Li)
A-B PLANE0.7980.6980.868ExcellentGood
MAX DEPTH0.8440.7640.899ExcellentGood
CORPUS LENGTH0.123−0.1060.339PoorPoor
POG to NB0.9450.9140.965ExcellentExcellent
ICC—interclass correlation coefficient, CI—confidence interval.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kazimierczak, W.; Gawin, G.; Janiszewska-Olszowska, J.; Dyszkiewicz-Konwińska, M.; Nowicki, P.; Kazimierczak, N.; Serafin, Z.; Orhan, K. Comparison of Three Commercially Available, AI-Driven Cephalometric Analysis Tools in Orthodontics. J. Clin. Med. 2024, 13, 3733. https://doi.org/10.3390/jcm13133733

AMA Style

Kazimierczak W, Gawin G, Janiszewska-Olszowska J, Dyszkiewicz-Konwińska M, Nowicki P, Kazimierczak N, Serafin Z, Orhan K. Comparison of Three Commercially Available, AI-Driven Cephalometric Analysis Tools in Orthodontics. Journal of Clinical Medicine. 2024; 13(13):3733. https://doi.org/10.3390/jcm13133733

Chicago/Turabian Style

Kazimierczak, Wojciech, Grzegorz Gawin, Joanna Janiszewska-Olszowska, Marta Dyszkiewicz-Konwińska, Paweł Nowicki, Natalia Kazimierczak, Zbigniew Serafin, and Kaan Orhan. 2024. "Comparison of Three Commercially Available, AI-Driven Cephalometric Analysis Tools in Orthodontics" Journal of Clinical Medicine 13, no. 13: 3733. https://doi.org/10.3390/jcm13133733

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop