Abstract
To properly repair and maintain implants, which are bone tissue implants that replace natural tooth roots, it is crucial to accurately identify their brand and specification. Deep learning has demonstrated outstanding capabilities in analysis, such as image identification and classification, by learning the inherent rules and degrees of representation of data models. The purpose of this study is to evaluate deep learning algorithms and their supporting application software for their ability to recognize and categorize three dimensional (3D) Cone Beam Computed Tomography (CBCT) images of dental implants. By using CBCT technology, the 3D imaging data of 27 implants of various sizes and brands were obtained. Following manual processing, the data were transformed into a data set that had 13,500 two-dimensional data. Nine deep learning algorithms including GoogleNet, InceptionResNetV2, InceptionV3, ResNet50, ResNet50V2, ResNet101, ResNet101V2, ResNet152 and ResNet152V2 were used to perform the data. Accuracy rates, confusion matrix, ROC curve, AUC, number of model parameters and training times were used to assess the efficacy of these algorithms. These 9 deep learning algorithms achieved training accuracy rates of 100%, 99.3%, 89.3%, 99.2%, 99.1%, 99.5%, 99.4%, 99.5%, 98.9%, test accuracy rates of 98.3%, 97.5%, 94.8%, 85.4%, 92.5%, 80.7%, 93.6%, 93.2%, 99.3%, area under the curve (AUC) values of 1.00, 1.00, 1.00, 1.00, 1.00, 1.00, 1.00, 1.00, 1.00. When used to identify implants, all nine algorithms perform satisfactorily, with ResNet152V2 achieving the highest test accuracy, classification accuracy, confusion matrix area under the curve, and receiver operating characteristic curve area under the curve area. The results showed that the ResNet152V2 has the best classification effect on identifying implants. The artificial intelligence identification system and application software based on this algorithm can efficiently and accurately identify the brands and specifications of 27 classified implants through processed 3D CBCT images in vitro, with high stability and low recognition cost.
Similar content being viewed by others
Explore related subjects
Find the latest articles, discoveries, and news in related topics.Introduction
For more than 40 years, dental implant system has been used as a safe and predictable treatment for replacing natural tooth roots in bone tissue, with a high survival rate of more than 90%1,2,3. The widespread use of implants makes long-term use prone to biological and mechanical complications, such as screw loosening and fracture, poor implant stability, peri-implant mucositis and peri-implanttitis4,5,6, etc. In this case, for the dentist to effectively perform necessary repairs and maintenance, they must first be able to identify specifics about the implants, such as the implant brand, the implant’s length, the diameter of the implant's post, and the type of abutment that was used.
With the continuous development of implant systems, more than 220 implant brands have been marketed since the beginning of the twentieth century7. Due to the increasing number of brands and types of implants worldwide, it is not easy to correctly identify implant systems8. In addition, when changing clinics for treatment or when the original treatment clinic closes, some patients lose their original medical records, including information about the implants. In this case, it is challenging for the doctor to determine the type of implant through an oral examination and imaging results relying solely on expertise and experience, thus making it hard for the patient to obtain precise and prompt treatment. Therefore, a method is required that can quickly and accurately identify implants with limited data.
Artificial intelligence (AI) refers to the study of machine simulation of human intelligence, thought behavior and its laws of a discipline technology. Recent advances in Internet technology have fueled AI’s rapid development and broadened its scope of possible applications. Artificial intelligence has found applications in medical imaging, diagnostic aid, etc.9. One of the cornerstones of artificial intelligence is deep learning, which creates neural networks to analyze and learn like the human brain. This technique allows computers to learn from data samples by training neural networks that mimic the human brain's mechanisms for recognizing patterns in information, such as text, sound and images. In deep learning research, convolutional neural networks (CNN) are most often used with excellent performance in image detection, classification, segmentation, and other analysis. It has also been successfully applied to many fields of dentistry. For example, in the identification of caries, oral dysplasia, oral cancer and other diseases have shown significant advantages10,11,12,13,14. It has been shown that using two-dimensional medical image like panoramic and apical radiographs, deep learning algorithms for implant identification are fairly accurate, and their implant identification and categorization is more accurate than that of the participating dentists15,16,17. Chaurasia A’s systematic review also showed, deep Learning models showed high accuracy in identifying and classifying dental implant systems using panoramic and periapical radiographic images18. However, CBCT images can avoid distortion and anamorphic of the image, while still accurately reflecting information related to the implant, making them preferable to traditional two-dimensional images. To our knowledge, there are no study using 3D CBCT images to identify implant brands and specifications by deep learning algorithms, which may be inextricably linked to the difficulty of processing and transferring 3D data. The purpose of this study is to evaluate deep learning algorithms and their supporting application software for their ability to recognize and categorize 3D CBCT images of dental implants.
Materials and methods
Study design
In this study, 3D CBCT images of implants in the model in-vitro were processed to obtain a two-dimensional dataset. Additionally, 9 deep learning algorithms were used to train the recognition models for the oral implant brands and specifications, such as length and diameter, with the goal of determining the best classification algorithm by comparing each algorithm's classification performance metrics.
Material preparation
27 implant systems of various sizes from 9 brands were prepared in total (Table 1) contains details on each of them. Each implant was placed separately into the maxillary model, which used PMMA resin material and are shown Low-density images during imaging. There was no use of dental prosthetic components such as cover screws, healing abutments, or loaded prosthetics. And all CT image data were completed by the oral radiologist using CBCT (NewTom VGi), and a total of 9 sets of 3D images were taken.
(Scanning conditions: tube voltage 110 kv, tube current 3.77 mA, exposure time 3.6 s, layer thickness 0.3 mm, exposure parameters are automatic.)
Data set preparation
Four dentists took the two-dimensional printscreen in coronal and sagittal sections, perpendicular to the central long axis of the implant. While the central long axis of the implant remained unchanged in the CBCT image (Fig. 1A). Fixing the crop window size during printscreen will result in a 180*180 pixel resolution. Random rotations were performed on the coronal and sagittal positions, and two-dimensional printscreen with different rotation angles were taken by this method to fully obtain the implant information. 500 printscreen were taken for each implant, 250 each in coronal (Fig. 1B) and sagittal (Fig. 1C) positions, and manually identified by category according to the brand and specifications provided by the manufacturer, resulting in a final two-dimensional image dataset of 13,500 images. Once that was done, the dataset was randomly divided into a training dataset with 10,800 sheets and a test dataset with 2700 sheets.
Deep learning algorithms
In this study, we classify the dataset using 9 deep learning algorithms, including GoogleNet, InceptionResNetV2, InceptionV3, ResNet50, ResNet50V2, ResNet101, ResNet101V2, ResNet152 and ResNet152V2. GoogLeNet is a 22-layer deep network, and its primary feature is an architecture designed to make better use of the network's available computational resources. To increase the network's depth and breadth, however, it requires a large amount of computation19. In order to simplify the intricate calculation, InceptionV2 introduces BN layer on InceptionV1 (GoogleNet) to avoid gradient disappearance and uses multiple small-scale convolutions instead of one large-scale convolution to reduce the number of parameters and boost computational efficiency at the same time; After the category fully connected layer, we add an LSR layer and swap out the 7 × 7 convolutional kernels for three 3 × 3 ones, Inception V3 improves upon Inception V2's computation efficiency and network classification accuracy20. Since increasing the depth of the network may cause gradient explosion and disappearance, based on these problems, He et al.21 proposed the ResNet network, a residual network, which can form a very deep network after stacking, making it possible to train very deep neural networks (NN), but the deepening of the number of layers of the network leads to its low operational efficiency. Based on the idea of whether the inception network can be as efficient when the network is deeper and wider, Sergey Ioffe et al., invented a new model, InceptionResNetV2, which is the resultant network architecture combining the Inception architecture with Residual Connections, with far better performance than the same expensive Inception network without Residual Connections. It is not only more computationally efficient than ResNet, but also has a richer network hierarchy than the Inception architecture21,22. Additionally, we have posted the code for the 9 deep learning algorithms' test and training datasets at https://github.com/AIdental/27-/blob/main/first%20gitHub.txt.
Evaluation indicators
We selected accuracy rates, training time, model parameters, confusion matrix, and receiver operating characteristic (ROC) curve as the evaluation indicators for this study, in order to fully assess the classification accuracy and performance of the algorithm. With these indicators, the efficacy of this experimental algorithm in implants identification can be evaluated.
Accuracy rates
The degree to which the anticipated value and the actual value coincide under certain experimental circumstances is referred as the accuracy rates. The accuracy rates in this study are the proportion of properly categorized implant samples relative to the total number of implant samples and reflect the classification accuracy of the algorithm.
Confusion matrix
In machine learning, the confusion matrix is an error matrix that is frequently used to visually evaluate the classification accuracy of supervised learning algorithms, and the size of the confusion matrix is a square matrix of (n, n)(n: implant type). In the image accuracy evaluation, it is mostly employed to compare the categorization result to the actual value, and the accuracy of the classification result can be shown inside the matrix. The confusion matrix is determined by comparing the position and classification of each measured image elements to its corresponding value in the classified image.
Receiver operating characteristic (ROC) curve
Area Under Curve (AUC) is defined as the area under the ROC curve and is a performance metric that measures the quality of a classifier. As a rule, the classifier corresponding to a bigger AUC is better, hence we frequently utilize the AUC value as the model's evaluation criterion when the ROC curve is ambiguous about which classifier is superior. When AUC = 1: it represents perfect classifier, 0.5 < AUC < 1, better than random classifier, 0 < AUC < 0.5, worse than random classifier. The horizontal coordinate of the ROC curve indicates 1-specificity and the vertical coordinate indicates sensitivity, as shown in Eqs. (1) and (2) below.
In this study, true positive (TP) reflects the amount of implant-type samples that were accurately predicted. False negative (FN) refers to the number of samples actually predicted into other implant types for that implant type. True negative (TN), indicates the number of samples actually predicted correctly for other implant types. And False positive (FP), reflects the number of samples actually predicted into that implant for other implant types.
Results
Accuracy rates, training times, number of model parameters (Table 2)
The batchsize and epoch trained by the 9 algorithm models tested in this study were set to 64 and 50, respectively, and in the 27 classification implant recognition, after weights were pre-trained, the training set accuracy and test set accuracy of all nine algorithms surpassed 80%.Among them, the test accuracy rate of ResNet152V2 test is 99.3%, which is the highest among the 9 algorithms and has the best classification accuracy (Table 2).
Confusion matrix
Another analysis results in the implant multiclassification confusion matrix based on a dataset using 9 deep learning models (Fig. 2), and comparing the position of each measured image element with the corresponding position in the classified image yields the highest classification accuracy for the ResNet152V2 algorithm model.
ROC curve
The closer the ROC curve is to the upper left corner of the graph, the higher the sensitivity, the lower the 1-specificity, i.e., the higher the sensitivity, the lower the false positive rate, and the larger the AUC area, the better the performance of the diagnostic method. Therefore, according to results of the ROC curves for the 9 algorithms, ResNet152V2 has the best performance, and the AUC area (Fig. 3).
Implant 27 classification prediction
Using the optimal ResNet152V2 in conjunction with PyQt5, this study has created an application for implant 27 classification prediction, which has been tested in Affliated Stomatological Hospital of Nanchang University. As an example, the implant test image (testPic.PNG) in Fig. 4A can be entered into the program to predict the corresponding implant 27 classification probability results. The graph of the results(Fig. 4B) is obtained after inputting implant test image(Fig. 4A) into the application we developed, the probability of the corresponding prediction category decreases from front to back and the implant is most likely ABT 4.2–11.5 mm implant.The application recognizes the procedure in just a few seconds and is quite effective.
Discussion
With the widespread use of implant restorative modalities, their maintenance and repair work is increasing, and the lack of implant-related data for various reasons makes it very difficult for dentists to identify implants through intraoral and imaging examinations. This work aims to discover a system that can accurately and rapidly identify implants using limited data. Lee et al.22 stated that the accuracy of using the appropriate deep learning model for implant classification from medical imaging data exceeded that of the participating dentists, and the feasibility of deep learning in implant identification has been initially verified.
Sukegawa et al.23,24 reported that convolutional neural networks were employed in earlier research for implant classification and identification, but they relied on image data from apical and panoramic and did not yet have a method for processing and using CBCT 3D images for implant identification. By manually intercepting the CBCT images of the implants in various levels and directions, the challenge of processing and transferring 3D data was solved in this study. The 3D images were successfully transformed into 2D data sets and trained for recognition. The sagittal and coronal were successfully located using this manual interception technique, and the resulting 2D images can prevent anatomical structure overlap and image distortion25.The effective processing and application of CBCT 3D images has improved the accuracy of the deep learning algorithm mode for identifying and classifying the brand and specifications of dental implants, which can assist dentists in their professional job more effectively. Additionally, this study used a total of 27 classifications and 9 brands of implants with different diameters and lengths, increasing the generalizability of the deep learning algorithm and application software in comparison to the experiment using only 12 implants in the study by Sukegawa et al.24. At the same time, the experiment obtained a two-dimensional data set of 13,500 images by processing 27 classified implant 3D images, which further enriched the database of implant image data and paved the way for future research on more classified implant recognition.
The nine algorithms used in this study are based on three currently popular architectures: Inception, ResNet, and a combination of Inception and ResNet, Which neural network architectures perform the best for current image categorization tasks. GoogLeNet (InceptionV1) and ResNet demonstrated excellent multi-category image classification and target detection performance in the ImageNet competition in 2014, and the ILSVRC in 2015, respectively. The 9 algorithms also achieved better results in this study, with accuracy rates above 80% in all cases. Compared to Jae-Hong Lee et al., who used only the InceptionV3 algorithm and Shintaro Sukegawa et al., who used only the ResNet system variant structure for implant recognition, we used the most algorithms and used a more comprehensive classification algorithm20,24.
The experimental results confirmed the accuracy of all nine deep learning algorithms for implant system identification and classification using CBCT images, especially the ResNet152V2 showed the highest classification accuracy, best accuracy in confusion matrix, and the best classification performance for the ROC curve. Although this algorithm has a large number of model parameters and a long training time, the time difference between the application developed based on it and other algorithms used for identification is short when used in clinical applications. This elapsed time is negligible when considered from the perspective of clinical applications. Considering that the best results were achieved using this algorithm for the identification of implants on CBCT images, it can better assist the dentist in identifying the implant system. This may be due to ResNet contains 12 layers of network levels, which is more favourable to picture feature extraction and better network expression. It also rebuilt a residual network's fundamental unit, making it easier to train and more effective in generalization. In addition, the findings of Figs. 3, 4 indicate that the deep learning system accurately identifies all types of implants on CBCT images. In particular, ABT 4.2–11.5 mm, Astra 5.0–9.0 mm, Nobel-CC 5.0–11.0 mm, SIC 4.5–13.0 mm, and SNUC 4.3–8.0 mm implants were best identified and basically not misclassified. The Straumann 5.0–10.0 mm implant had the worst recognition result. This result may be the consequence of a tiny change in diameter, length, and general morphology between this implant and other implants of the same brand, making it more difficult to identify than other implants.
In addition, our ultimate goal is to use deep learning algorithms for implant identification in clinical care. Through the comprehensive consideration of various evaluation indicators, the recognition effect of ResNet152V2 is optimal, therefore we use this algorithm to further develop an application for implant 27 classification that can be operated by doctors and can be augmented for reinforcement learning, and has been initially tested in Affliated Stomatological Hospital of Nanchang University. In the test process, we can directly capture and input images according to the method in this study for implants that have not completed the implantation process in other hospitals, or do not have a prosthetic components. However, for implants with a prosthesis or healing screw, only the root portion of the implant should be captured in CBCT to complete the acquisition of input images. However, the accuracy of implant identification is decreased in this way.
This study also has some limitations. First of all, the data collected in this study were from implants placed in vitro model, and the influence of background structures such as bone and the Suprastructure of the implant on the experimental results was not considered. Although we have applied the developed application to the clinic for preliminary trials, the accuracy has decreased, and there is still a lack of specific data to support this. In addition, our processing of CBCT images is still limited to two-dimensional interception and has not completed three-dimensional transformation. Therefore, in future studies, we should also address the 3D transformation of CBCT images and and collect CBCT images of patients with implants with superstructure as a dataset to complete the optimization of the application by training the dataset. Secondly, in the experimental design, we selected implants with different diameters and lengths, mainly to verify whether different types of implants of the same brand can be effectively identified. However, it is not comprehensive enough to include all types of implants. Future research will further solve this problem.Thirdly, due to the limitation of time and conditions, the CBCT images used for identification in this study were all taken by the same equipment, while the spatial resolution and magnification of different CBCT equipment differed.Therefore, in future studies, we should also conduct multicenter studies to collect more image data using different models of CBCT machines to establish a more complete and comprehensive database for better generalization of experimental results. Finally, in this study, the algorithm used in the currently developed implant 27 classification prediction is the optimal one, while the correct rate of classification may decrease with the increase of implant dataset in the future. We should use more algorithms to train the dataset and further optimize the application to make the operation easier and the classification recognition more accurate for the physician to obtain the missing implant information to assist the dentist’s clinical work.
Conclusion
In this study, the artificial intelligence recognition system built on ResNet152V2 and the preliminary developed application software can efficiently and accurately identify the brand and specification of 27 classified implants by the processed CBCT 3D images in vitro with high stability, low recognition cost and wide applicability. The program can assist the dentist in implant identification, saving a great deal of time and increasing accuracy, which can serve as a reference for the replacement and repair of implants in the future. At the same time, the dataset of this study solves the problem of processing and transferring 3D images, further enriching the database of different types of implant images, and paving the way for future promotion of the application and for studies related to the identification of more kinds of implants.
Data availability
The datasets generated and analysed during the current study are available in the [Github] repository, [You can click this weblink to datasets: https://github.com/AIdental/27-/blob/main/first%20gitHub.txt].
References
Albrektsson, T., Branemark, P. I., Hansson, H. A. & Lindstrom, J. Osseointegrated titanium implants. Requirements for ensuring a long-lasting, direct bone-to-implant anchorage in man. Acta Orthop. Scand. 52, 155–170. https://doi.org/10.3109/17453678108991776 (1981).
Pjetursson, B. E., Thoma, D., Jung, R., Zwahlen, M. & Zembic, A. A systematic review of the survival and complication rates of implant-supported fixed dental prostheses (FDPs) after a mean observation period of at least 5 years. Clin. Oral Implants Res. 23(Suppl 6), 22–38. https://doi.org/10.1111/j.1600-0501.2012.02546.x (2012).
Swierkot, K., Lottholz, P., Flores-de-Jacoby, L. & Mengel, R. Mucositis, peri-implantitis, implant success, and survival of implants in patients with treated generalized aggressive periodontitis: 3-to 16-year results of a prospective long-term cohort study. J. Periodontol. 83, 1213–1225. https://doi.org/10.1902/jop.2012.110603 (2012).
Staedt, H. et al. Potential risk factors for early and late dental implant failure: A retrospective clinical study on 9080 implants. Int. J. Implant. Dent. 6, 81. https://doi.org/10.1186/s40729-020-00276-w (2020).
Tsigarida, A. et al. Peri-implant diseases and biologic complications at implant-supported fixed dental prostheses in partially edentulous patients. J. Prosthodont. 29, 429–435. https://doi.org/10.1111/jopr.13165 (2020).
Lee, K. Y. et al. Clinical study on screw loosening in dental implant prostheses: A 6-year retrospective study. J. Korean Assoc. Oral Maxillofac. Surg. 46, 133–142. https://doi.org/10.5125/jkaoms.2020.46.2.133 (2020).
Jokstad, A. et al. Quality of dental implants. Int. Dent. J. 53, 409–443. https://doi.org/10.1111/j.1875-595x.2003.tb00918.x (2003).
Nuzzolese, E., Lusito, S., Solarino, B. & Vella, G. D. Radiographic dental implants recognition for geographic evaluation in human identification. J. Forens. Odontostomatol. 26, 8–11 (2008).
Lee, J. H., Kim, D. H. & Jeong, S. N. Diagnosis of cystic lesions using panoramic and cone beam computed tomographic images based on deep learning neural network. Oral Dis. 26, 152–158. https://doi.org/10.1111/odi.13223 (2020).
Lee, J. G. et al. Deep learning in medical imaging: General overview. Korean J. Radiol. 18, 570–584. https://doi.org/10.3348/kjr.2017.18.4.570 (2017).
Jader, G., Oliveira, L. & Pithon, M. Automatic segmenting teeth in X-ray images: Trends, a novel data set, benchmarking and future perspectives. Expert Syst. Appl. 107, 15–31. https://doi.org/10.1016/j.eswa.2018.04.001 (2018).
Bayraktar, Y. & Ayan, E. Diagnosis of interproximal caries lesions with deep convolutional neural network in digital bitewing radiographs. Clin. Oral Investig. 26, 623–632. https://doi.org/10.1007/s00784-021-04040-1 (2022).
Camalan, S. et al. Convolutional neural network-based clinical predictors of oral dysplasia: Class activation map analysis of deep learning results. Cancers (Basel) 13, 1291. https://doi.org/10.3390/cancers13061291 (2021).
Jubair, F. et al. A novel lightweight deep convolutional neural network for early detection of oral cancer. Oral Dis. 28, 1123–1130. https://doi.org/10.1111/odi.13825 (2021).
Lee, J. H. & Jeong, S. N. Efficacy of deep convolutional neural network algorithm for the identification and classification of dental implant systems, using panoramic and periapical radiographs: A pilot study. Med. (Baltim.) 99, e20787. https://doi.org/10.1097/MD.0000000000020787 (2020).
Kim, J. E. et al. Transfer learning via deep neural networks for implant fixture system classification using periapical radiographs. J. Clin. Med. 9, 1117. https://doi.org/10.3390/jcm9041117 (2020).
Lee, J. H., Kim, Y. T. & Jeong, S. N. A performance comparison between automated deep learning and dental professionals in classification of dental implant systems from dental imaging: A multi-center study. Diagn. (Basel) 10, 910. https://doi.org/10.3390/diagnostics10110910 (2020).
Chaurasia, A., Namachivayam, A., Koca-Ünsal, R. B. & Lee, J. H. Deep-learning performance in identifying and classifying dental implant systems from dental imaging: A systematic review and meta-analysis. J. Periodontal. Implant. Sci. https://doi.org/10.5051/jpis.2300160008 (2023).
Szegedy, C. et al. Going deeper with convolutions. In 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 1–9 (2015). https://doi.org/10.1109/CVPR.2015.7298594.
Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J. & Wojna, Z. Rethinking the inception architecture for computer vision. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2818–2826 (2016). https://doi.org/10.1109/CVPR.2016.308.
He, K. M., Zhang, X. Y., Ren, S. Q. & Sun, J. Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)770–778 (2016). https://doi.org/10.1109/CVPR.2016.90.
Szegedy, C., Ioffe, S., Vanhoucke, V. & Alemi, A. A. Inception-v4, Inception-ResNet and the impact of residual connections on learning. In AAAI’17: Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence 4278–4284 (2016). https://doi.org/10.48550/arXiv.1602.07261.
Sukegawa, S. et al. Deep neural networks for dental implant system classification. Biomolecules 10, 984. https://doi.org/10.3390/biom10070984 (2020).
Sukegawa, S. et al. Multi-task deep learning model for classification of dental implant brand and treatment stage using dental panoramic radiograph images. Biomolecules 11, 815. https://doi.org/10.3390/biom11060815 (2021).
Scarfe, W. C. & Farman, A. G. What is cone-beam CT and how does it work?. Dent. Clin. N. Am. 52, 707–730. https://doi.org/10.1016/j.cden.2008.05.005 (2008).
Acknowledgements
The authors thank Dr. Peng Zhang, for her significant contribution to this study.
Funding
This work was supported by the financial support from the National Natural Science Foundation of China (NOS. 82160194 and 81960492), Natural Science Foundation of Jiangxi Province (20192BAB205054) and the Key Research and Development Program of Jiangxi Province (20212BBG73022). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Author information
Authors and Affiliations
Contributions
S.O., S.H.: these authors contributed equally to this work and share first authorship. Data preparation, writing, methodology, review and editing. D.S.: data curation, formal analysis. H.W.: data curation, formal analysis, writing. J.C.: data curation, writing, review and editing. Y.C.: data preparation. D.Y.: data preparation. H.O.: data curation, formal analysis, methodology, review and editing. L.L.: methodology, review and editing. H.O. and L.L. are as correspondence author. Last, all authors reviewed the manuscript.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Ou-yang, S., Han, S., Sun, D. et al. The preliminary in vitro study and application of deep learning algorithm in cone beam computed tomography image implant recognition. Sci Rep 13, 18467 (2023). https://doi.org/10.1038/s41598-023-45757-1
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-023-45757-1
- Springer Nature Limited
We’re sorry, something doesn't seem to be working properly.
Please try refreshing the page. If that doesn't work, please contact support so we can address the problem.