Comparative analysis of diagnostic accuracy in endodontic assessments: dental students vs. artificial intelligence
-
Abubaker Qutieshat
Abubaker QutieshatAdult Restorative Dentistry, Oman Dental College, Muscat, Oman; Restorative Dentistry, Dundee Dental Hospital and School, University of Dundee, Dundee, UKSearch for this author in:
, Alreem Al Rusheidi
Alreem Al RusheidiAdult Restorative Dentistry, Oman Dental College, Muscat, OmanSearch for this author in:Samiya Al GhammariAdult Restorative Dentistry, Oman Dental College, Muscat, OmanSearch for this author in:Abdulghani AlarabiAdult Restorative Dentistry, Oman Dental College, Muscat, OmanSearch for this author in:Abdurahman SalemDental Technology, School of Health & Society, University of Bolton, Greater Manchester, UKSearch for this author in:Maja ZelihicForbes School of Business and Technology, University of Arizona Global Campus, Chandler, AZ, USASearch for this author in:
Abstract
Objectives
This study evaluates the comparative diagnostic accuracy of dental students and artificial intelligence (AI), specifically a modified ChatGPT 4, in endodontic assessments related to pulpal and apical conditions. The findings are intended to offer insights into the potential role of AI in augmenting dental education.
Methods
Involving 109 dental students divided into junior (54) and senior (55) groups, the study compared their diagnostic accuracy against ChatGPT’s across seven clinical scenarios. Juniors had the American Association of Endodontists (AEE) terminology assistance, while seniors relied on prior knowledge. Accuracy was measured against a gold standard by experienced endodontists, using statistical analysis including Kruskal-Wallis and Dwass-Steel-Critchlow-Fligner tests.
Results
ChatGPT achieved significantly higher accuracy (99.0 %) compared to seniors (79.7 %) and juniors (77.0 %). Median accuracy was 100.0 % for ChatGPT, 85.7 % for seniors, and 82.1 % for juniors. Statistical tests indicated significant differences between ChatGPT and both student groups (p<0.001), with no notable difference between the student cohorts.
Conclusions
The study reveals AI’s capability to outperform dental students in diagnostic accuracy regarding endodontic assessments. This underscores AIs potential as a reference tool that students could utilize to enhance their understanding and diagnostic skills. Nevertheless, the potential for overreliance on AI, which may affect the development of critical analytical and decision-making abilities, necessitates a balanced integration of AI with human expertise and clinical judgement in dental education. Future research is essential to navigate the ethical and legal frameworks for incorporating AI tools such as ChatGPT into dental education and clinical practices effectively.
Introduction
The 2022 release of OpenAI’s ChatGPT generated widespread interest for its human-like responses. It raised anticipation similar to transformative technologies like mobile devices and the World Wide Web. However, concerns about its implications for human capabilities, employment, ethics, and regulations emerged early on. Despite uncertainties, ChatGPT’s potential as a diagnostic tool in dental education sparked optimism and scrutiny.
Artificial intelligence (AI) refers to the ability of an integrated platform to obtain, process, and implement skills and knowledge acquired through education or experience that are usually linked to human intelligence. It is made up of a neural network architecture that mimics human thinking and has been applied in dentistry to diagnose dental diseases, plan treatments, make clinical decisions, and predict prognoses. There are two main types of AI in healthcare: machine learning and deep learning. Machine learning requires significant human intervention for identifying and selecting relevant features from data, a process crucial for the algorithm’s learning and prediction accuracy. In contrast, deep learning minimizes this need by autonomously discovering patterns and features directly from large datasets, though it still necessitates human expertise in crafting network architectures and managing extensive training data [1].
In recent years, AI has emerged as a powerful tool in healthcare, potentially revolutionizing clinical practice and improving patient outcomes. One area where it has shown significant promise is in aiding clinicians in making accurate diagnoses. AI algorithms can analyse large amounts of medical data, such as medical images, patient records, and laboratory results, and provide insights that can help clinicians make more informed decisions [2].
Developing diagnostic acumen is a crucial element of a dental student’s training, enabling them to transition into competent practitioners. This development necessitates exposure to a wide array of clinical cases, constructive feedback from faculty, and self-evaluation exercises. In recent years, the healthcare industry, including dentistry, has been experiencing significant technological advancements. AI, which has demonstrated its efficacy in improving accuracy, reducing costs, and minimizing medical errors, is emerging as a transformative force within contemporary dentistry. Its role extends beyond that of a simple tool; it is becoming an integral part of an advanced technological ecosystem that is shaping the future of dental practice. Notably, the trend towards integrating AI into healthcare and dentistry is supported by substantial investments from both the private and public sectors [3, 4].
In the medical field, AI has demonstrated the potential to assist clinicians in making more accurate diagnoses. For example, one study found that AI-based algorithms could accurately detect and diagnose diabetic retinopathy, a common cause of blindness, with sensitivity and specificity comparable to or better than human experts [5]. In medical imaging, AI algorithms have been shown to accurately identify and diagnose various medical conditions, including skin cancer [6], lung cancer [7], and breast cancer [8], with high sensitivity and specificity. Additionally, AI can enhance clinical workflows, reduce diagnostic errors, and improve patient outcomes by managing medical data, identifying high-risk patients for personalized care, and supporting student training in history-taking skills [9], [10], [11]. It also advances medico-dental diagnostics through the detection and classification of pathologies, rating attractiveness, and predicting age and gender; deep learning models specifically excel in identifying abnormalities in medical images and have significantly improved dentists’ diagnostic accuracy, sensitivity, and specificity [12], [13], [14].
Dental radiology benefits significantly from AI, automating image reading and enabling early diagnosis across various areas, such as dental emergencies and prosthetic planning. Through learning from experience and adapting to new data, AI improves the standardization process quality [4, 15]. Deep learning models diagnose conditions like caries, bone loss, periapical lesions, malocclusion, and anomalies radiographically [16], [17], [18], [19], [20], [21], [22]. Studies highlight AI’s efficacy in enhancing dental records by accurately assigning tooth numbers on radiographs, reducing processing time [13], identifying cysts and tumours from panoramic radiographs [23], [24], [25], and detecting proximal caries with reliability, though without depth differentiation [20].
AI has shown potential in revolutionizing dentistry, providing early and accurate detection of dental caries and addressing the limitations of traditional diagnostic methods [23, 26]. These AI-based models have been found to be highly accurate, cost-effective, and reduce false negatives, aiding dentists in clinical practice and assisting non-dental professionals in detecting dental caries [3, 27].
In dental education, virtual patient simulations play a critical role, particularly during times of shortage of real patients. The integration of AI in virtual patient simulations can enhance the realism of the experience and provide an interactive platform for students to practice on a range of clinical cases. AI-powered chatbots have been shown to be beneficial in dental education, providing students with personalized learning experiences while improving patient safety [28].
AI has also shown potential in diagnosing orofacial pain and Temporomandibular Joint Disorders (TMDs). Research has demonstrated the possibility of creating an AI-based system to support non-specialists in early TMD detection, potentially leading to faster referrals to specialized medical centres [29, 30]. The proposed system is intended to assist general dentists in making a primary diagnosis, avoiding suboptimal recognition and management in interdisciplinary treatment.
AI has been proposed for evaluating cellular and structural atypia of oral squamous cell carcinoma (SCC), offering a more advanced diagnosis and making it a suitable method for developing AI in this field [31]. In implantology, researchers have developed a method for the accurate detection of dental implants and peri-implant tissue using a deep learning model [32]. AI utilizing Convolutional Neural Network (CNN) has demonstrated remarkable accuracy in the computer-assisted classification of odontogenic cysts in digital images of histological tissue sections [33].
In endodontics, various applications of AI have been demonstrated, including the diagnosis of root canal pathology, the determination of working length measurements, and the prediction of retreatment success with very high accuracy [34, 35]. The use of AI in this field has the potential to perform diagnostic and prognostic predictions with decision-making ability, improving dental care. This can be particularly valuable in areas with a lack of highly qualified personnel.
Furthermore, AI allows for the integration of multiple and heterogeneous information domains, such as dental history, clinical data, and sociodemographic information. AI has been used to diagnose periapical diseases, discover root fractures, evaluate root morphology, determine working length, and trace the apical opening, as well as predict retreatment outcomes [35]. A systematic review found that AI models can be a valuable asset for dentists with less experience and non-specialists, providing them with a more accurate and reliable diagnostic tool to support their decision-making [34].
In a qualitative study, several enablers were identified to encourage the implementation of AI in dental diagnostics [36]. One of the main enablers was the ability to generate an automated diagnostic report after interaction with AI findings, saving time for reporting. Another enabler was the expectation that AI would increase diagnostic accuracy. The use of AI as a second independent opinion was also seen as a way to increase trust in the diagnosis made by the dentist.
While dentistry still lacks a standard level of quality and current diagnostic techniques can be dependent on the operator’s subjective perception and prone to errors, AI can help provide a unique treatment plan for each patient, from planning to treatment and follow-up. AI can learn automatically and improve its performance over time, helping clinicians improve their diagnoses, save time, and increase profitability. The development of AI-based diagnostic techniques can help overcome the limitations of manual methods. However, it is important to balance the use of AI with intelligence, objectivity, and common sense and have an appropriate learning curve.
The objective of this study is to compare the diagnostic accuracy of dental students with that of artificial intelligence, specifically a modified ChatGPT 4, in endodontic assessments, including pulpal and apical diagnoses. By doing so, this research aims to shed light on the potential of AI to serve as a reference tool that may enhance the diagnostic process within dental education. The findings are expected to contribute valuable insights into the role of AI technologies in augmenting educational outcomes and guiding future investigations in this evolving field.
Materials and methods
The study involved two distinct groups of dental students. The first group consisted of 54 junior students, defined as third-year students who had just begun their clinical exposure by treating patients. The second group comprised 55 senior students, defined as fifth-year students who are nearing graduation. To aid in the diagnosis, junior students were provided with a sheet listing terminology definitions according to the American Association of Endodontists (AAE) [37], while senior students did not receive this aid and had to rely on their accumulated knowledge.
The AI that was used is a modified ChatGPT 4 (15 Mar 2023 version) that was trained with the pulpal and apical terminology definitions according to the AEE. This modification involved fine-tuning the model with a curated dataset comprising pulpal and apical terminology definitions, case scenarios and diagnostic criteria as defined by the AEE. The fine-tuning process utilized a supervised learning approach, where the model was trained to identify and apply these specialized dental terms in the context of diagnosing pulpal and apical conditions accurately. This tailored training aimed to enhance the model’s proficiency in dental diagnostics beyond its general capabilities.
Each student was tasked with diagnosing seven distinct clinical scenarios that emulated real-life patient cases. These scenarios were presented in a uniform manner. Students made pulpal and apical diagnoses for each case without the assistance of the AI tool. The investigators then used ChatGPT to diagnose the same seven clinical scenarios 60 times. The students’ diagnoses were compared to the ChatGPT diagnoses and a gold standard diagnosis determined by a panel of experienced endodontists.
Furthermore, to explore students’ self-awareness of their diagnostic performance, both senior and junior students were asked to predict the number of mistakes they believed they made prior to receiving feedback on their actual performance. To facilitate this, a predefined scale was introduced for the prediction task, where ‘0 mistakes’ corresponded to a score of 1, ‘1–3 mistakes’ to a score of 2, ‘4–6 mistakes’ to a score of 3, ‘7–9 mistakes’ to a score of 4, and ‘more than 10 mistakes’ to a score of 5.
The primary outcome measure was the accuracy of the pulpal and apical diagnoses made by the students. This was calculated as the number of correctly diagnosed cases divided by the total number of cases. The secondary outcome measure was the confidence level of the diagnosis with and without the use of ChatGPT.
The data collected were analysed using descriptive statistics and inferential statistics (SPSS software, Version 26, IBM Corp., Armonk, NY, USA), including non-parametric one-way ANOVA (Kruskal-Wallis test) to compare the medians of the three groups (senior students, junior students, and ChatGPT). In cases where the Kruskal-Wallis test revealed significant differences, Dwass-Steel-Critchlow-Fligner pairwise comparisons were performed as a post-hoc test to identify which specific groups had significant differences in their medians.
Results
In this study, we compared the diagnostic accuracy of pulpal and apical diagnoses between senior students, junior students, and the ChatGPT tool. The results showed that ChatGPT had a higher mean accuracy (99.0 %) compared to senior students (79.7 %) and junior students (77.0 %). Similarly, ChatGPT had a higher median accuracy (100.0 %) than senior students (85.7 %) and junior students (82.1 %). When looking at the spread of the data, ChatGPT had a much lower standard deviation (0.0254) than senior students (0.122) and junior students (0.155), indicating greater consistency in ChatGPT’s performance (Table 1). The interquartile range (IQR) showed that ChatGPT had the narrowest spread (0.00) compared to senior students (0.0714) and junior students (0.214) (Figure 1). The Shapiro-Wilk test revealed that the data for all three groups were not normally distributed (p<0.001).
Descriptive statistics of diagnostic accuracy for senior students, junior students, and ChatGPT.
Group | n | Mean | Median | SD | IQR | Min | Max |
---|---|---|---|---|---|---|---|
Senior students | 55 | 0.797 | 0.857 | 0.122 | 0.0714 | 0.429 | 0.929 |
Junior students | 54 | 0.77 | 0.821 | 0.155 | 0.214 | 0.429 | 0.929 |
ChatGPT | 55 | 0.99 | 1 | 0.0254 | 0 | 0.929 | 1 |
-
n, number of participants. Diagnostic accuracy is reported as proportions (multiplied by 100 % for percentage). SD, standard deviation; IQR, interquartile range.

Box plot of diagnostic accuracy for senior students, junior students, and ChatGPT. The boxes represent the interquartile range with the median line dividing the boxes. Whiskers extend to the most extreme non-outlier values, and points labelled with numbers indicate outliers within the data. The y-axis represents diagnostic accuracy as a proportion of correctly diagnosed cases, with values ranging from 0 to 1.0, where 1.0 signifies 100 % accuracy. ChatGPT demonstrates higher median accuracy and a narrower IQR compared to both senior and junior students.
The Kruskal-Wallis test reveals a statistically significant difference among the medians of the three groups being compared (χ2=107, df=2, p<0.001), with a large effect size (ε2) of 0.655. the Dwass-Steel-Critchlow-Fligner pairwise comparisons reveal no significant difference between the senior and junior groups (p=0.829), but both the senior and junior groups have statistically significant differences in their medians when compared to the ChatGPT group (p<0.001). Overall, these findings suggest that ChatGPT outperformed both senior and junior students in terms of diagnostic accuracy, consistency, and data spread.
The perceived difficulty level of the clinical scenarios, as reported by both junior and senior students, was moderately challenging on average. Senior students reported a mean difficulty level of 2.76 out of 5 (SD 0.60), while junior students reported a slightly higher mean difficulty level of 2.83 out of 5 (SD 0.94).
Additionally, both groups predicted their number of mistakes using a predefined scale (0 mistakes=1, 1–3 mistakes=2, 4–6 mistakes=3, 7–9 mistakes=4, and more than 10 mistakes=5). Senior students’ mean prediction was 2.49 (SD 0.71), indicating they expected to make between one and three mistakes on average. Similarly, junior students’ mean prediction was 2.44 (SD 1.17), also suggesting they anticipated making between one and three mistakes on average.
These findings imply that both junior and senior students perceived the clinical scenarios as moderately challenging and anticipated making a relatively small number of mistakes in their diagnoses. However, the junior students’ responses exhibited a higher degree of variability, as shown by the larger standard deviations for both perceived difficulty level and predicted number of mistakes.
Discussion
Our investigation reveals that the application of AI in dental education, particularly through a specifically adapted version of ChatGPT 4, notably surpasses the diagnostic accuracy of both senior and junior dental students in endodontic assessments. This outcome suggests that integrating AI into dental curricula may have the potential to enhance students’ diagnostic capabilities. More than just boosting diagnostic precision, this integration promises to reinforce students’ confidence and consistency in clinical evaluations, which are essential components of clinical decision-making.
Senior students’ slightly higher accuracy compared to junior students can be attributed to their educational and clinical experience. Senior students typically spend more time in dental school and are exposed to more clinical cases, which may have given them a better understanding of pulpal and apical diagnoses. However, the small difference in accuracy between the two groups suggests that, at least in the context of the clinical scenarios presented in this study, both junior and senior students have a comparable level of competence in diagnosing pulpal and apical conditions. The small difference in accuracy may also indicate that junior students’ foundational knowledge and skills are sufficient for them to perform adequately in this particular diagnostic task.
The higher degree of variability in junior students’ perceived difficulty level and predicted number of mistakes, as evidenced by larger standard deviations, is worth noting. This variation may reflect differences in individual junior students’ exposure to clinical cases, confidence level, or mastery of relevant diagnostic terminology and concepts. While senior students performed slightly better in this study, the small difference between the two groups suggests that both junior and senior students have a comparable level of diagnostic competence. The higher variability among junior students, on the other hand, emphasises the importance of continuing education and clinical exposure to improve diagnostic skills and confidence.
ChatGPT performed significantly better in pulpal and apical diagnoses than both junior and senior students. This outstanding performance demonstrates ChatGPT’s potential value as an aid for dental students seeking more accurate diagnoses. Integrating ChatGPT into educational and clinical settings could provide several benefits, including serving as a supportive resource for students, encouraging active learning and critical thinking, and aiding in the standardisation of the diagnostic process across various levels of education and experience.
Harvard Business Review’s article cited Nobel Prize Winner Daniel Kahneman, who pointed out deep flaws within human reasoning, betting on the superiority of AI’s “reasoning” with the human capability to adjust being at the core of this problem [38]. Concerns have been raised, however, about the potential overreliance on AI tools such as ChatGPT, which could impede the development of critical analytical and decision-making skills and inadvertently lead to a false sense of confidence among dental students and recent graduates [39]. To alleviate these concerns, it is critical to emphasise the importance of using AI tools in a balanced manner, positioning them as supplementary resources that supplement, rather than replace, critical thinking and clinical judgement.
The current study’s findings contrast with those of Huh’s (2023) study [40], which compared ChatGPT’s performance in a parasitology examination to that of medical students. The disparities in results can be attributed to a variety of factors, including differences in subject matter, examination formats, and the specific versions or modifications of ChatGPT used in each study.
The findings of our study are consistent with broader trends and conclusions highlighted in the literature over the last two decades, which investigate clinical and educational aspects of dentistry in relation to practical applications of AI [34], [35], [36]. There is a pressing need for dental curricula to be updated to include AI applications, ensuring that both educators and students are adequately prepared to use these technologies in their practice [41, 42].
However, it is critical to highlight the ethical and legal implications of AI integration in dental education, emphasising the importance of reaching an agreement on the safe and responsible use of these tools. While our study demonstrates the potential benefits of using ChatGPT as an assistive tool in dental diagnosis, it also emphasises the importance of taking a cautious and balanced approach to incorporating AI in dental education and practice [43], ensuring that its potential is fully realised without jeopardising students’ clinical reasoning abilities or any ethical or legal considerations.
While the absence of dental X-rays in the clinical scenarios is a limitation, it is important to note that the improvement in diagnostic accuracy observed with ChatGPT remains meaningful and clinically relevant. Despite the lack of dental radiographs, ChatGPT demonstrated a significant improvement in diagnostic accuracy when compared to both senior and junior students. This suggests that AI-based tools like ChatGPT have the potential to improve clinical decision-making and patient outcomes even when diagnostic information is limited. The incorporation of dental radiographs into future research could provide more robust evidence supporting the efficacy of AI tools in improving diagnostic precision in dentistry. This will be especially apparent with the upcoming release of the multimodal ChatGPT, which will be able to interpret and analyse images submitted to it [44]. Moreover, the research was conducted in a single dental school, which restricts the generalizability of the results. Further studies involving multiple dental schools and diverse populations are necessary to validate and expand upon our findings in order to confirm the broader applicability of ChatGPT as an effective diagnostic aid in dental education.
Conclusions
Our investigation contributes to the growing body of knowledge regarding the application of AI in dental education, emphasizing diagnostic accuracy. The study demonstrates that AI, through a specifically adapted version of ChatGPT 4, surpasses both senior and junior dental students in endodontic diagnosis accuracy. The standout diagnostic performance of ChatGPT highlights its value as a significant educational tool. By incorporating ChatGPT into dental curricula, students could greatly improve their diagnostic skills, leading to increased accuracy, confidence, and consistency in clinical evaluations, which, in turn, may positively influence patient care outcomes.
However, it is crucial to consider ChatGPT and similar AI technologies as supplements to traditional educational content. They should serve to enrich the critical thinking and clinical judgement skills developed through conventional training. Embracing AI in this supportive capacity is expected to elevate diagnostic proficiency and contribute to reducing diagnostic errors, ultimately improving healthcare delivery. Future research should delve into the practical, ethical, and legal aspects of weaving AI tools like ChatGPT into dental education and clinical practice. Identifying optimal ways to integrate these technologies for the benefit of students and patients alike will be key as the healthcare education landscape continues to evolve.
-
Research ethics: The study was approved by the Ethics Committee of Oman Dental College (ODC-2022-AE-191).
-
Informed consent: Informed consent was obtained from all individuals included in this study.
-
Author contributions: All authors have made substantial contributions to the work, draughted or revised the manuscript critically, approved the final version for publication, and agree to be accountable for all aspects of its accuracy and integrity.
-
Competing interests: The authors state no conflict of interest.
-
Research funding: None declared.
-
Data availability: The raw data can be obtained on request from the corresponding author.
References
1. Russell, SJ. Artificial intelligence a modern approach. New Jersey: Pearson Education, Inc.; 2010.Search in Google Scholar
2. Mupparapu, M, Wu, C-W, Chen, Y-C. Artificial intelligence, machine learning, neural networks, and deep learning: futuristic concepts for new dental diagnosis. Quintessence Int 2018;49:687–8. https://doi.org/10.3290/j.qi.a41107.Search in Google Scholar PubMed
3. Chifor, R, Arsenescu, T, Dascalu-Rusu, LM, Badea, AF. Automated diagnosis using artificial intelligence a step forward for preventive dentistry: a systematic review. Rom J Stomatol 2022;68:106–15. https://doi.org/10.37897/rjs.2022.3.7.Search in Google Scholar
4. De Angelis, F, Pranno, N, Franchina, A, Di Carlo, S, Brauner, E, Ferri, A, et al.. Artificial intelligence: a new diagnostic software in dentistry: a preliminary performance diagnostic study. Int J Environ Res Publ Health 2022;19:1728. https://doi.org/10.3390/ijerph19031728.Search in Google Scholar PubMed PubMed Central
5. Ting, DSW, Cheung, CY-L, Lim, G, Tan, GSW, Quang, ND, Gan, A, et al.. Development and validation of a deep learning system for diabetic retinopathy and related eye diseases using retinal images from multiethnic populations with diabetes. JAMA 2017;318:2211–23. https://doi.org/10.1001/jama.2017.18152.Search in Google Scholar PubMed PubMed Central
6. Esteva, A, Kuprel, B, Novoa, RA, Ko, J, Swetter, SM, Blau, HM, et al.. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017;542:115–8. https://doi.org/10.1038/nature21056.Search in Google Scholar PubMed PubMed Central
7. Ardila, D, Kiraly, AP, Bharadwaj, S, Choi, B, Reicher, JJ, Peng, L, et al.. End-to-end lung cancer screening with three-dimensional deep learning on low-dose chest computed tomography. Nat Med 2019;25:954–61. https://doi.org/10.1038/s41591-019-0447-x.Search in Google Scholar PubMed
8. Kooi, T, Litjens, G, Van Ginneken, B, Gubern-Mérida, A, Sánchez, CI, Mann, R, et al.. Large scale deep learning for computer aided detection of mammographic lesions. Med Image Anal 2017;35:303–12. https://doi.org/10.1016/j.media.2016.07.007.Search in Google Scholar PubMed
9. Jiang, F, Jiang, Y, Zhi, H, Dong, Y, Li, H, Ma, S, et al.. Artificial intelligence in healthcare: past, present and future. Stroke Vasc Neurol 2017;2:230–43. https://doi.org/10.1136/svn-2017-000101.Search in Google Scholar PubMed PubMed Central
10. Churpek, MM, Yuen, TC, Winslow, C, Meltzer, DO, Kattan, MW, Edelson, DP. Multicenter comparison of machine learning methods and conventional regression for predicting clinical deterioration on the wards. Crit Care Med 2016;44:368. https://doi.org/10.1097/ccm.0000000000001571.Search in Google Scholar PubMed PubMed Central
11. Maicher, KR, Stiff, A, Scholl, M, White, M, Fosler-Lussier, E, Schuler, W, et al.. Artificial intelligence in virtual standardized patients: combining natural language understanding and rule based dialogue management to improve conversational fidelity. Med Teach 2022;45:1–7. https://doi.org/10.1080/0142159x.2022.2130216.Search in Google Scholar PubMed
12. Patcas, R, Bornstein, MM, Schätzle, MA, Timofte, R. Artificial intelligence in medico-dental diagnostics of the face: a narrative review of opportunities and challenges. Clin Oral Invest 2022;26:6871–9. https://doi.org/10.1007/s00784-022-04724-2.Search in Google Scholar PubMed PubMed Central
13. Kabir, T, Lee, C-T, Chen, L, Jiang, X, Shams, S. A comprehensive artificial intelligence framework for dental diagnosis and charting. BMC Oral Health 2022;22:1–13. https://doi.org/10.1186/s12903-022-02514-6.Search in Google Scholar PubMed PubMed Central
14. Ezhov, M, Gusarev, M, Golitsyna, M, Yates, JM, Kushnerev, E, Tamimi, D, et al.. Clinically applicable artificial intelligence system for dental diagnosis with CBCT. Sci Rep 2021;11:15006. https://doi.org/10.1038/s41598-021-94093-9.Search in Google Scholar PubMed PubMed Central
15. Futyma-Gąbka, K, Różyło-Kalinowska, I. The use of artificial intelligence in radiological diagnosis and detection of dental caries: a systematic review. J Stomatol 2021;74:262–6. https://doi.org/10.5114/jos.2021.111664.Search in Google Scholar
16. Ngoc, V, Viet, DH, Anh, LK, Minh, DQ, Nghia, LL, Loan, HK, et al.. Periapical lesion diagnosis support system based on X-ray images using machine learning technique. World J Dent 2021;12:190. https://doi.org/10.5005/jp-journals-10015-1820.Search in Google Scholar
17. Alevizakos, V, Bekes, K, Steffen, R, von See, C. Artificial intelligence system for training diagnosis and differentiation with molar incisor hypomineralization (MIH) and similar pathologies. Clin Oral Invest 2022;26:6917–23. https://doi.org/10.1007/s00784-022-04646-z.Search in Google Scholar PubMed
18. Alotaibi, G, Awawdeh, M, Farook, FF, Aljohani, M, Aldhafiri, RM, Aldhoayan, M. Artificial intelligence (AI) diagnostic tools: utilizing a convolutional neural network (CNN) to assess periodontal bone level radiographically – a retrospective study. BMC Oral Health 2022;22:399. https://doi.org/10.1186/s12903-022-02436-3.Search in Google Scholar PubMed PubMed Central
19. Bayraktar, Y, Ayan, E. Diagnosis of interproximal caries lesions with deep convolutional neural network in digital bitewing radiographs. Clin Oral Investig 2022;26:623–32. https://doi.org/10.1007/s00784-021-04040-1.Search in Google Scholar PubMed PubMed Central
20. García-Cañas, Á, Bonfanti-Gris, M, Paraíso-Medina, S, Martínez-Rus, F, Pradíes, G. Diagnosis of interproximal caries lesions in bitewing radiographs using a deep convolutional neural network-based software. Caries Res 2022;56:503–11. https://doi.org/10.1159/000527491.Search in Google Scholar PubMed
21. Ma, J, Xue, C, Bai, D. Progress on application of artificial intelligence technology in orthodontic diagnosis and treatment. J Prev Treat Stomatol Dis 2022;30:278–82.Search in Google Scholar
22. Schönewolf, J, Meyer, O, Engels, P, Schlickenrieder, A, Hickel, R, Gruhn, V, et al.. Artificial intelligence-based diagnostics of molar-incisor-hypomineralization (MIH) on intraoral photographs. Clin Oral Invest 2022;26:5923–30. https://doi.org/10.1007/s00784-022-04552-4.Search in Google Scholar PubMed PubMed Central
23. Lee, J-H, Kim, D-H, Jeong, S-N, Choi, S-H. Detection and diagnosis of dental caries using a deep learning-based convolutional neural network algorithm. J Dent 2018;77:106–11. https://doi.org/10.1016/j.jdent.2018.07.015.Search in Google Scholar PubMed
24. Kwon, O, Yong, T-H, Kang, S-R, Kim, JE, Huh, KH, Heo, MS, et al.. Automatic diagnosis for cysts and tumors of both jaws on panoramic radiographs using a deep convolution neural network. Dentomaxillofac Radiol 2020;49:20200185. https://doi.org/10.1259/dmfr.20200185.Search in Google Scholar PubMed PubMed Central
25. Başaran, M, Çelik, Ö, Bayrakdar, IS, Bilgir, E, Orhan, K, Odabaş, A, et al.. Diagnostic charting of panoramic radiography using deep-learning artificial intelligence system. Oral Radiol 2022;38:363–9. https://doi.org/10.1007/s11282-021-00572-0.Search in Google Scholar PubMed
26. Musri, N, Christie, B, Ichwan, SJA, Cahyanto, A. Deep learning convolutional neural network algorithms for the early detection and diagnosis of dental caries on periapical radiographs: a systematic review. Imag Sci Dent 2021;51:237. https://doi.org/10.5624/isd.20210074.Search in Google Scholar PubMed PubMed Central
27. Khanagar, SB, Alfouzan, K, Awawdeh, M, Alkadi, L, Albalawi, F, Alfadley, A. Application and performance of artificial intelligence technology in detection, diagnosis and prediction of dental caries (DC) – a systematic review. Diagnostics 2022;12:1083. https://doi.org/10.3390/diagnostics12051083.Search in Google Scholar PubMed PubMed Central
28. Suárez, A, Adanero, A, Díaz-Flores García, V, Freire, Y, Algar, J. Using a virtual patient via an artificial intelligence chatbot to develop dental students’ diagnostic skills. Int J Environ Res Publ Health 2022;19:8735. https://doi.org/10.3390/ijerph19148735.Search in Google Scholar PubMed PubMed Central
29. Kreiner, M, Viloria, J. A novel artificial neural network for the diagnosis of orofacial pain and temporomandibular disorders. J Oral Rehabil 2022;49:884–9. https://doi.org/10.1111/joor.13350.Search in Google Scholar PubMed
30. Reda, B, Contardo, L, Prenassi, M, Guerra, E, Derchi, G, Marceglia, S. Artificial intelligence to support early diagnosis of temporomandibular disorders: a preliminary case study. J Oral Rehabil 2023;50:31–8. https://doi.org/10.1111/joor.13383.Search in Google Scholar PubMed
31. Oya, K, Kokomoto, K, Nozaki, K, Toyosawa, S. Oral squamous cell carcinoma diagnosis in digitized histological images using convolutional neural network. J Dent Sci 2023;18:322–9. https://doi.org/10.1016/j.jds.2022.08.017.Search in Google Scholar PubMed PubMed Central
32. Jang, WS, Kim, S, Yun, PS, Jang, HS, Seong, YW, Yang, HS, et al.. Accurate detection for dental implant and peri-implant tissue by transfer learning of faster R-CNN: a diagnostic accuracy study. BMC Oral Health 2022;22:1–7. https://doi.org/10.1186/s12903-022-02539-x.Search in Google Scholar PubMed PubMed Central
33. Bittencourt, MAV, de Sá Mafra, PH, Julia, RS, Travençolo, B, Silva, P, Blumenberg, C, et al.. Accuracy of computer-aided image analysis in the diagnosis of odontogenic cysts: a systematic review. Med Oral Patol Oral Cir Bucal 2021;26:e368. https://doi.org/10.4317/medoral.24238.Search in Google Scholar PubMed PubMed Central
34. Boreak, N. Effectiveness of artificial intelligence applications designed for endodontic diagnosis, decision-making, and prediction of prognosis: a systematic review. J Contemp Dent Pract 2020;21:926–34. https://doi.org/10.5005/jp-journals-10024-2894.Search in Google Scholar
35. Karobari, MI, Adil, AH, Basheer, SN, Murugesan, S, Savadamoorthi, KS, Mustafa, M, et al.. Evaluation of the diagnostic and prognostic accuracy of artificial intelligence in endodontic dentistry: a comprehensive review of literature. Comput Math Methods Med 2023;2023:1–9. https://doi.org/10.1155/2023/7049360.Search in Google Scholar PubMed PubMed Central
36. Müller, A, Mertens, SM, Göstemeyer, G, Krois, J, Schwendicke, F. Barriers and enablers for artificial intelligence in dental diagnostics: a qualitative study. J Clin Med 2021;10:1612. https://doi.org/10.3390/jcm10081612.Search in Google Scholar PubMed PubMed Central
37. American Association of Endodontists. AAE endodontic diagnosis. Chicago, IL: American Association of Endodontists Colleagues for Excellence Newsletter; 2013. Available from: https://www.aae.org/specialty/wp-content/uploads/sites/2/2017/07/endodonticdiagnosisfall2013.pdf [Accessed 12 Apr 2024].Search in Google Scholar
38. Agrawal, A, Gans, J, Goldfarb, A. The economics of artificial intelligence: an agenda. Chicago and London: University of Chicago Press; 2019.10.7208/chicago/9780226613475.001.0001Search in Google Scholar
39. Tolsgaard, MG, Pusic, MV, Sebok-Syer, SS, Gin, B, Svendsen, MB, Syer, MD, et al.. The fundamentals of artificial intelligence in medical education research: AMEE guide no. 156. Med Teach 2023;45:565–73. https://doi.org/10.1080/0142159x.2023.2180340.Search in Google Scholar
40. Huh, S. Are ChatGPT’s knowledge and interpretation ability comparable to those of medical students in Korea for taking a parasitology examination?: a descriptive study. J Educ Eval Health Prof 2023;20:1. https://doi.org/10.3352/jeehp.2023.20.01.Search in Google Scholar
41. Brickley, M, Shepherd, J, Armstrong, R. Neural networks: a new technique for development of decision support systems in dentistry. J Dent 1998;26:305–9. https://doi.org/10.1016/s0300-5712(97)00027-4.Search in Google Scholar PubMed
42. Thurzo, A, Strunga, M, Urban, R, Surovková, J, Afrashtehfar, KI. Impact of artificial intelligence on dental education: a review and guide for curriculum update. Educ Sci 2023;13:150. https://doi.org/10.3390/educsci13020150.Search in Google Scholar
43. Morreel, S, Mathysen, D, Verhoeven, V. Aye, AI! ChatGPT passes multiple-choice family medicine exam. Med Teach 2023;45:1. https://doi.org/10.1080/0142159x.2023.2187684.Search in Google Scholar
44. Lund, BD, Wang, T. Chatting about ChatGPT: how may AI and GPT impact academia and libraries? Libr Hi Tech News 2023;40:26–9. https://doi.org/10.1108/lhtn-01-2023-0009.Search in Google Scholar
© 2024 the author(s), published by De Gruyter, Berlin/Boston
This work is licensed under the Creative Commons Attribution 4.0 International License.
Articles in the same Issue
- Frontmatter
- Editorial
- The growing threat of hijacked journals
- Review
- Effects of SNAPPS in clinical reasoning teaching: a systematic review with meta-analysis of randomized controlled trials
- Mini Review
- Diagnostic value of D-dimer in differentiating multisystem inflammatory syndrome in Children (MIS-C) from Kawasaki disease: systematic literature review and meta-analysis
- Opinion Papers
- Masquerade of authority: hijacked journals are gaining more credibility than original ones
- FRAMED: a framework facilitating insight problem solving
- Algorithms in medical decision-making and in everyday life: what’s the difference?
- Original Articles
- Computerized diagnostic decision support systems – a comparative performance study of Isabel Pro vs. ChatGPT4
- Comparative analysis of diagnostic accuracy in endodontic assessments: dental students vs. artificial intelligence
- Assessing the Revised Safer Dx Instrument® in the understanding of ambulatory system design changes for type 1 diabetes and autism spectrum disorder in pediatrics
- The Big Three diagnostic errors through reflections of Japanese internists
- SASAN: ground truth for the effective segmentation and classification of skin cancer using biopsy images
- Computable phenotype for diagnostic error: developing the data schema for application of symptom-disease pair analysis of diagnostic error (SPADE)
- Development of a disease-based hospital-level diagnostic intensity index
- HbA1c and fasting plasma glucose levels are equally related to incident cardiovascular risk in a high CVD risk population without known diabetes
- Short Communications
- Can ChatGPT-4 evaluate whether a differential diagnosis list contains the correct diagnosis as accurately as a physician?
- Analysis of thicknesses of blood collection needle by scanning electron microscopy reveals wide heterogeneity
- Letters to the Editor
- For any disease a human can imagine, ChatGPT can generate a fake report
- The dilemma of epilepsy diagnosis in Pakistan
- The Japanese universal health insurance system in the context of diagnostic equity
- Case Report – Lessons in Clinical Reasoning
- Lessons in clinical reasoning – pitfalls, myths, and pearls: a case of tarsal tunnel syndrome caused by an intraneural ganglion cyst
Articles in the same Issue
- Frontmatter
- Editorial
- The growing threat of hijacked journals
- Review
- Effects of SNAPPS in clinical reasoning teaching: a systematic review with meta-analysis of randomized controlled trials
- Mini Review
- Diagnostic value of D-dimer in differentiating multisystem inflammatory syndrome in Children (MIS-C) from Kawasaki disease: systematic literature review and meta-analysis
- Opinion Papers
- Masquerade of authority: hijacked journals are gaining more credibility than original ones
- FRAMED: a framework facilitating insight problem solving
- Algorithms in medical decision-making and in everyday life: what’s the difference?
- Original Articles
- Computerized diagnostic decision support systems – a comparative performance study of Isabel Pro vs. ChatGPT4
- Comparative analysis of diagnostic accuracy in endodontic assessments: dental students vs. artificial intelligence
- Assessing the Revised Safer Dx Instrument® in the understanding of ambulatory system design changes for type 1 diabetes and autism spectrum disorder in pediatrics
- The Big Three diagnostic errors through reflections of Japanese internists
- SASAN: ground truth for the effective segmentation and classification of skin cancer using biopsy images
- Computable phenotype for diagnostic error: developing the data schema for application of symptom-disease pair analysis of diagnostic error (SPADE)
- Development of a disease-based hospital-level diagnostic intensity index
- HbA1c and fasting plasma glucose levels are equally related to incident cardiovascular risk in a high CVD risk population without known diabetes
- Short Communications
- Can ChatGPT-4 evaluate whether a differential diagnosis list contains the correct diagnosis as accurately as a physician?
- Analysis of thicknesses of blood collection needle by scanning electron microscopy reveals wide heterogeneity
- Letters to the Editor
- For any disease a human can imagine, ChatGPT can generate a fake report
- The dilemma of epilepsy diagnosis in Pakistan
- The Japanese universal health insurance system in the context of diagnostic equity
- Case Report – Lessons in Clinical Reasoning
- Lessons in clinical reasoning – pitfalls, myths, and pearls: a case of tarsal tunnel syndrome caused by an intraneural ganglion cyst