Editorial Type:
Article Category: Research Article
 | 
Online Publication Date: 21 Aug 2024

Evaluation of automated photograph-cephalogram image integration using artificial intelligence models

,
,
,
,
,
, and
Page Range: 595 – 601
DOI: 10.2319/010124-1.1
Save
Download PDF

ABSTRACT

Objectives

To develop and evaluate an automated method for combining a digital photograph with a lateral cephalogram.

Materials and Methods

A total of 985 digital photographs were collected and soft tissue landmarks were manually detected. Then 2500 lateral cephalograms were collected, and corresponding soft tissue landmarks were manually detected. Using the images and landmark identification information, two different artificial intelligence (AI) models—one for detecting soft tissue on photographs and the other for identifying soft tissue on cephalograms—were developed using different deep-learning algorithms. The digital photographs were rotated, scaled, and shifted to minimize the squared sum of distances between the soft tissue landmarks identified by the two different AI models. As a validation process, eight soft tissue landmarks were selected on digital photographs and lateral cephalometric radiographs from 100 additionally collected validation subjects. Paired t-tests were used to compare the accuracy of measures obtained between the automated and manual image integration methods.

Results

The validation results showed statistically significant differences between the automated and manual methods on the upper lip and soft tissue B point. Otherwise, no statistically significant difference was found.

Conclusions

Automated photograph-cephalogram image integration using AI models seemed to be as reliable as manual superimposition procedures.

INTRODUCTION

Since facial profile changes can accompany orthodontic treatment, photograph-cephalogram image integration has commonly been used in various clinical orthodontic situations, such as in planning treatment,1–3 predicting treatment outcomes,4–7 and predicting facial growth.8,9 Authors of these studies typically analyze and predict soft tissue changes from lateral cephalometric radiographs; however, to present these changes more realistically, visualization of the changes through photographs is necessary.

Authors of all previous prediction studies used manual integration results to match the soft tissue appearance in the digital photograph with that in the radiograph. Such manual image integration is a laborious procedure that heavily depends on clinician skill in aligning the photograph and cephalogram using soft tissue curves and landmarks. However, interest is growing in applying artificial intelligence (AI) models in clinical orthodontic practice. For example, to reduce manpower burden and increase objectivity, various approaches have been used to improve the accuracy of automatic identification of cephalometric landmarks, and authors of a recent study showed that the performance of AI models was comparable with that of human experts.10–13 Currently, although some commercial cephalometric programs can provide orthodontic clinicians with a kind of automated photograph-cephalogram image overlay function, it is difficult to find an explicitly reported method to implement such automated photograph-cephalogram image integration.

The purpose of this study was to develop an automated method for combining digital photography and lateral cephalography through the application of AI models. To evaluate the clinical applicability of this automatic integration method, its image integration accuracy was compared with that of a traditional manual integration method.

MATERIALS AND METHODS

All training, test, and validation photographs and cephalograms were collected from the picture-aided communication system server (Infinitt Healthcare Co Ltd, Seoul, Korea) at Seoul National University Dental Hospital, Seoul, Korea. The institutional review boards at Seoul National University Dental Hospital (ERI 19007) and Seoul National University School of Dentistry (S-D20200015) approved the research protocol.

Training Two Different AI Models

To develop an automated method for combining a photograph and cephalogram, two different AI models were created: one for detecting soft tissue profiles on photographs and the other for identifying soft tissue curves on cephalograms.

A total of 985 digital photographs were collected to create an AI model capable of automatically identifying soft tissue landmarks on profile photographs. To train the AI model, a total of 18 soft tissue landmarks, from glabella to the cervical point (Table 1), were manually defined on these photographs by an examiner (JHM). Subsequently, this labeling information was used to train a residual neural network (ResNet), which is a deep-learning method frequently used for image identification problems.14

Table 1. A Total of 32 Soft Tissue Landmarks Was Used to Train the AI to Detect Landmarks in Lateral Cephalometric Images. Eighteen Soft Tissue Landmarks on Digital Photographs, Marked With a Symbol ■, Were Used to Train the AI. To Validate the Integration Methods, Eight Landmarks Marked With Were Selected and Used for Comparisons Between Automated and Manual Integration Methods
Table 1.

Then 2500 lateral cephalograms were collected, and a total of 32 soft tissue landmarks, from glabella to the soft tissue terminal point (Table 1, first column), were manually detected on these cephalograms by another examiner (SJL) to create an AI model that could identify soft tissue landmarks on lateral cephalograms. This AI model was based on the you-only-look-once version 3 (YOLO-v3) algorithm, which is a deep-learning method developed for real-time object detection.15,16 YOLO-v3 was reported to demonstrate higher accuracy than other machine-learning methods in automated cephalometric analyses and the detection of multiple cephalometric landmarks.11–13,15

Automated Integration Method

Figure 1 demonstrates the automated integration method. Initially, the two AI models automatically identified landmarks in digital photographs and lateral cephalometric images. Once the 18 landmarks which are common to both modalities were identified, the digital photograph was rotated, scaled, and shifted to minimize the squared sum of distances between the corresponding soft tissue landmarks identified by the two different AI models.

Figure 1.Figure 1.Figure 1.
Figure 1. Experimental design.

Citation: The Angle Orthodontist 94, 6; 10.2319/010124-1.1

In mathematical terms, the automated integration by the AI was equivalent to finding the transformation condition T that minimizes the root mean square (RMS) argument where and represent the position of the ith soft tissue landmark automatically identified in the digital photograph and lateral cephalometric radiograph, respectively, n is the total number of soft tissue landmarks identified in the digital photograph, and ‖ ‖ stands for the Euclidean distance measure calculated in millimeters (Figure 1).

Manual Integration Method

The manual procedure for combining photographs and cephalograms is a common practice used by orthodontic clinicians. In contrast with computer-based methods that rely on multiple soft tissue landmarks, human clinicians can easily draw and connect a smooth outline of soft tissue without much difficulty.

To facilitate the manual integration procedure, custom-made software was developed in the Python programming language (Python Software Foundation, Wilmington, Del) to move, rotate, and resize the soft tissue outline as needed. Figure 2 illustrates the manual integration of soft tissue profiles obtained from the lateral cephalometric radiograph and digital photograph. Three guidelines were used: the soft tissue outline from glabella to the cervical point, based on landmarks identified by the AI model; the Frankfort horizontal plane; and the sella-nasion plane. The manual image integrations were performed by a different examiner (MGK).

Figure 2.Figure 2.Figure 2.
Figure 2. Manual integration procedures for soft tissue profiles obtained from the lateral cephalometric radiograph and digital photograph. The soft tissue outlines from glabella to the cervical point, the Frankfort horizontal plane, and the sella-nasion plane were used as guiding lines to rotate, resize, and translate.

Citation: The Angle Orthodontist 94, 6; 10.2319/010124-1.1

Validation via Comparisons Between Automated and Manual Integration Methods

To validate the automated image integration method according to the common validation protocol,17,18 new data not used during the training or learning procedures were collected from 100 randomly selected additional subjects who had undergone both lateral cephalometric radiographs and digital profile photographs for diagnostic purposes. Individuals whose images were used in the training of either of the two AI models were excluded from the validation dataset.

For each of these 100 validation subjects, eight soft tissue landmarks (glabella, soft tissue nasion, pronasale, subnasale, upper lip, lower lip, soft tissue point B, and soft tissue pogonion) were manually identified on both lateral cephalometric images and digital photographs by a different human examiner (SJC). Custom-made digitizing software developed in Python was used to record x and y coordinates of each landmark relative to the Cartesian coordinate system. For each manual integration, the coordinates of the soft tissue landmarks on the digital photograph were transformed according to the transformation implemented by the AI model when it transformed the digital photograph to match the soft tissue profile of the lateral cephalometric image (Figure 3). Then RMS values in millimeters were measured between the coordinates of the soft tissue landmarks obtained from the lateral cephalometric image and the coordinates of the transformed soft tissue landmarks on the digital photograph, both involving eight landmarks that were manually identified. These values were used to compare the accuracy between the two image integration methods.

Figure 3.Figure 3.Figure 3.
Figure 3. To compare the accuracy of the two methods, eight soft tissue landmarks were selected from digital photographs and lateral cephalometric radiographs of the 100 validation subjects. The distance of the soft tissue landmarks between the transformed coordinates of the digital photograph (white) and the lateral cephalometric radiograph (red) was calculated.

Citation: The Angle Orthodontist 94, 6; 10.2319/010124-1.1

Paired t-tests were conducted to compare the accuracy of the measured RMS for each landmark and to compare the pooled RMS values. Statistical significance was set at P < .05. All statistical analyses were carried out using R language.19

RESULTS

The paired t-test results for each soft tissue landmark showed statistically significant differences in RMS values between the automated and manual image integration methods for the upper lip (0.37 mm, P < .01) and soft tissue B point (0.52 mm, P < .01). The other six landmarks did not show a statistically significant difference.

The pooled mean RMS values measured from the automated and manual soft tissue integration methods were 2.97 ± 0.95 mm and 2.92 ± 1.04 mm, respectively, and did not show a statistically significant difference (Table 2).

Table 2. To Compare Accuracy Between the Automated and Manual Image Integration Methods, Eight Soft Tissue Landmarks Were Selected on the Digital Photographs and Lateral Cephalometric Radiographs for the 100 Validation Subjects. For Each Integration Method, the Root Mean Square (RMS) Values in Millimeters Were Calculated Between the Position of a Given Soft Tissue Landmark in the Lateral Cephalometric Image and Its Transformed Position in the Digital Photograph, and Paired t-Tests Were Conducted
Table 2.

DISCUSSION

The purpose of this study was to develop and evaluate an automated method for combining digital photographs with lateral cephalograms. To develop this automated integration method, 18 soft tissue landmarks on digital photographs and lateral cephalometric images were automatically identified by two different AI models, and the images were then transformed to minimize the distances between these soft tissue landmarks in the two image types. We are the first to focus on developing an automated photograph-cephalogram image integration method and to evaluate it in comparison with the traditional manual process. The results of the present study showed that applying multiple AI models and the least squares concept might be as reliable as manual superimposition procedures.

The accuracy of the automatic identification of soft tissue landmarks could affect the accuracy of image integration in the later step. When developing an accurate AI model, not only the characteristics of the target variables but also the AI data size should be considered.20 However, no solid guidance exists to determine appropriate sample sizes when developing an AI model.21 Moon et al. (2020)1 suggested that at least 2300 cephalograms would be necessary to develop an automatic landmark identification AI model that would be as accurate as human examiners. Accordingly, in the present study, we collected 2500 cephalograms, manually identified cephalometric landmarks, and developed an AI model using the manually labeled data. Additionally, to estimate the number of digital photographs required, a preliminary study was conducted using 200 digital photographs as learning data. The results of this preliminary study demonstrated a failure rate of approximately 10% in the detection of 18 soft tissue landmarks in the test data. According to the sample size estimation guidance by Moon et al.,11 985 photographs would be needed to increase the accuracy of the AI model to an acceptable level that was characterized by a human interexaminer difference of 1.5 mm.11,17

Considering that combining photography and cephalometry is often necessary with particular care in several clinical areas of orthodontics, it was unexpected that only a few studies on this procedure could be found in a literature review.2,3 The study of Dvortsin et al. (2011)2 appears to have been the first study to focus on the issue of photograph-cephalogram image integration. The authors suggested a method for reorientating the lateral cephalogram to the natural head position (NHP) according to a standardized photograph taken at NHP.2 However, this method was too heavily dependent upon the NHP photograph, and it might not be feasible to obtain this in clinical orthodontic practice. Unlike cephalometric radiography, for which a tripod head stabilizer or cephalostat is used, photography does not generally involve a head holder that can fix head orientation in a repeatable manner. Consequently, profile photographs may have elongated and/or foreshortened interlandmark distances in addition to random intersubject variability. Therefore, a more sophisticated method for photograph-cephalogram image integration or a standardized method for overlaying two images might be necessary to develop an automated process.

Manual image integration is commonly performed by overlaying two images according to their profile outlines. Like manual procedures, the photograph-cephalogram image integration method proposed by Wang et al. (2018)3 used profile outlines. Their method appeared to provide an automatic solution for image integration using a hierarchical contour detection algorithm to achieve image congruence of soft tissue outlines of the forehead on the lateral cephalogram and photograph.3 However, this method seemed to be a bit theoretical. In practice, its application would still require corrections to control for curve deviation. Even more seriously, image integration accuracy could be vulnerable to the heavy reliance on a short outline span limited to the forehead region.

The image integration method used in the present study depended on multiple soft tissue landmarks that were automatically detected on both photographs and cephalograms. The idea of depending on multiple landmarks (MLs) and seeking a least sums of squares solution was inspired by studies on automated cephalometric superimposition.12,22 Authors of these cephalometric image superimposition studies used MLs located on the cranial base to align two cephalographic images so that the distance between each landmark was minimized. This ML superimposition method is known to have better accuracy than the conventional sella-nasion-line superimposition method. The ML superimposition produced results like those of Bjork’s superimposition method, especially in the evaluation of growth changes in growing children.12,22 While the ML superimposition method overlayed two cephalometric images automatically, in the present study, we aligned and overlaid images from two different modalities: digital photography and cephalography. However, the basic idea of using MLs and pursuing the least squares solutions was the same.

To facilitate three-dimensional (3D) visualization during orthodontic diagnosis, several methods for overlaying 3D tomographic images and digital dental models have been introduced.23–25 While these 3D image integration methods are likely to gain popularity in the near future, it is also true that 3D images are not routinely obtained during a patient’s first visit, with planar two-dimensional photographs and lateral cephalograms being more commonly used in clinical orthodontic practice.26 It may be some time before 3D image integration tools become routinely accessible.

A limitation of this study was that the method was dependent upon the landmark-based approach, and a relatively large number of 18 landmarks needed to be found on both photographs and cephalograms. However, these limitations might no longer be a significant barrier for today’s practice environment since most contemporary commercial cephalometric software programs provide automatic landmark detection functions. Therefore, the automatic image integration method of the present study might be compatible with the current clinical environment. In addition, since the photograph-cephalogram overlay method of the present study uses 18 soft tissue landmarks, deviations or aberrations in some of the landmarks would not cause a significant impact on the pooled mean RMS values or the quality of the whole image integration. Even the statistically significant RMS values shown in Table 2 might not be clinically significant if each landmark is positioned within the profile curves. Applying the simple and intuitive least squares concept might be another advantageous feature of this method. The simple least squares idea using multiple soft tissue landmarks that forms the focus of this study could be easily implemented in most cephalometric programs currently available for use in clinical orthodontic practice.

CONCLUSIONS

  • The automated photograph-cephalogram image integration method using AI models seemed as reliable as manual superimposition procedures.

  • The method may be particularly compatible with contemporary cephalometric programs providing automatic landmark identification functions.

ACKNOWLEDGMENTS

This study was supported by a grant (03-2022-0046) from the Seoul National University Dental Hospital Research Fund.

DISCLOSURE

All authors of this study declare that they have no conflict of interest.

REFERENCES

  • 1.

    Tugran M, Baka ZM. Esthetic evaluation of profile photographs showing various sagittal and vertical patterns. Am J Orthod Dentofacial Orthop. 2021;159:281291.

  • 2.

    Dvortsin DP, Ye Q, Pruim GJ, Dijkstra PU, Ren Y. Reliability of the integrated radiograph-photograph method to obtain natural head position in cephalometric diagnosis. Angle Orthod. 2011;81:889894.

  • 3.

    Wang S, Li H, Zou B, Zhang W. A novel contour-based registration of lateral cephalogram and profile photograph. Comput Med Imaging Graph. 2018;63:923.

  • 4.

    Park JA, Moon JH, Lee JM, et al. Does artificial intelligence predict orthognathic surgical outcomes better than conventional linear regression methods? Angle Orthod. 2024. doi:10.2319/111423-756.1

  • 5.

    Cho SJ, Moon JH, Ko DY, et al. Orthodontic treatment outcome predictive performance differences between artificial intelligence and conventional methods. Angle Orthod. 2024. doi:10.2319/111823-767.1

  • 6.

    Suh HY, Lee HJ, Lee YS, Eo SH, Donatelli RE, Lee SJ. Predicting soft tissue changes after orthognathic surgery: the sparse partial least squares method. Angle Orthod. 2019;89:910916.

  • 7.

    Kim K, Lee SJ, Eo SH, Cho SJ, Lee JW. Modified partial least squares method implementing mixed-effect model. Commun Stat Appl Methods. 2023;30:6573.

  • 8.

    Moon JH, Shin HK, Lee JM, et al. Comparison of individualized facial growth prediction models based on the partial least squares and artificial intelligence. Angle Orthod. 2024;94:207215.

  • 9.

    Moon JH, Kim MG, Hwang HW, Cho SJ, Donatelli RE, Lee SJ. Evaluation of an individualized facial growth prediction model based on the multivariate partial least squares method. Angle Orthod. 2022;92:705713.

  • 10.

    Hwang HW, Moon JH, Kim MG, Donatelli RE, Lee SJ. Evaluation of automated cephalometric analysis based on the latest deep learning method. Angle Orthod. 2021;91:329335.

  • 11.

    Moon JH, Hwang HW, Yu Y, Kim MG, Donatelli RE, Lee SJ. How much deep learning is enough for automatic identification to be reliable? Angle Orthod. 2020;90:823830.

  • 12.

    Moon JH, Hwang HW, Lee SJ. Evaluation of an automated superimposition method for computer-aided cephalometrics. Angle Orthod. 2020;90:390396.

  • 13.

    Hwang HW, Park JH, Moon JH, et al. Automated identification of cephalometric landmarks: Part 2- Might it be better than human? Angle Orthod. 2020;90:6976.

  • 14.

    He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2016:770778.

  • 15.

    Park JH, Hwang HW, Moon JH, et al. Automated identification of cephalometric landmarks: part 1—comparisons between the latest deep-learning methods YOLOV3 and SSD. Angle Orthod. 2019;89:903909.

  • 16.

    Redmon J, Farhadi A. YOLOv3: an incremental improvement. arXiv:1804.02767. 2018.

  • 17.

    Moon JH, Lee JM, Park JA, Suh H, Lee SJ. Reliability statistics every orthodontist should know. Semin Orthod. 2024;30:4549.

  • 18.

    Donatelli RE, Lee SJ. How to test validity in orthodontic research: a mixed dentition analysis example. Am J Orthod Dentofacial Orthop. 2015;147:272279.

  • 19.

    R Development Core Team. R: a language and environment for statistical computing. 2024.

    Vienna, Austria
    .

  • 20.

    Yu W, Lee SJ, Cho H. Partial least squares regression trees for multivariate response data with multicollinear predictors. IEEE Access. 2024;12:3663636644.

  • 21.

    Lee JM, Moon JH, Park JA, Kim JH, Lee SJ. Factors influencing the development of artificial intelligence in orthodontics. Orthod Craniofac Res. 2024. doi:10.1111/ocr.12806.

  • 22.

    Kim MG, Moon JH, Hwang HW, Cho SJ, Donatelli RE, Lee SJ. Evaluation of an automated superimposition method based on multiple landmarks for growing patients. Angle Orthod. 2022;92:226232.

  • 23.

    Zou B, Kim JH, Kim SH , et al. Accuracy of a surface-based fusion method when integrating digital models and the cone beam computed tomography scans with metal artifacts. Sci Rep. 2022;12:8034.

  • 24.

    Xiao Z, Liu Z, Gu Y. Integration of digital maxillary dental casts with 3D facial images in orthodontic patients. Angle Orthod. 2020;90:397404.

  • 25.

    Sauppe S, Abkai C, Hourfar J, Ludwig B, Ulrici J, Hell E. Automatic fusion of lateral cephalograms and digital volume tomography data-perspective for combining two modalities in the future. Dentomaxillofac Radiol. 2015;44:20150073.

  • 26.

    Dehesa-Santos A, Park JA, Lee SJ, Iglesias-Linares A. East Asian and Southern European craniofacial Class III phenotype: two sides of the same coin? Clin Oral Investig. 2024;28:84.

Copyright: © 2024 by The EH Angle Education and Research Foundation, Inc.
Figure 1.
Figure 1.

Experimental design.


Figure 2.
Figure 2.

Manual integration procedures for soft tissue profiles obtained from the lateral cephalometric radiograph and digital photograph. The soft tissue outlines from glabella to the cervical point, the Frankfort horizontal plane, and the sella-nasion plane were used as guiding lines to rotate, resize, and translate.


Figure 3.
Figure 3.

To compare the accuracy of the two methods, eight soft tissue landmarks were selected from digital photographs and lateral cephalometric radiographs of the 100 validation subjects. The distance of the soft tissue landmarks between the transformed coordinates of the digital photograph (white) and the lateral cephalometric radiograph (red) was calculated.


Contributor Notes

 Private Practice, Cheonan, Korea.
 PhD Graduate Student, Seoul National University, Seoul, Korea.
 Research Scientist, AI Research Center, DDH Inc, Seoul, Korea.
 Clinical Lecturer, Department of Orthodontics, Seoul National University Dental Hospital, Seoul, Korea.
 Professor, Department of Orthodontics and Dental Research Institute, Seoul National University School of Dentistry, Seoul, Korea.
Corresponding author: Dr Shin-Jae Lee, Professor, Department of Orthodontics and Dental Research Institute, Seoul National University School of Dentistry, Jongro-Gu, Seoul 03080, Korea (e-mail: nonext.shinjae@gmail.com)
Received: 01 Jan 2024
Accepted: 01 Jul 2024
  • Download PDF