Next Article in Journal
Modeling Data Sovereignty in Public Cloud—A Comparison of Existing Solutions
Previous Article in Journal
YoloSortC3D: A YOLOv8, DeepSort, and C3D-Based Method for Sheep Rumination Behavior Recognition
Previous Article in Special Issue
Abutment Tooth Formation Simulator for Naked-Eye Stereoscopy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Clinical, Research, and Educational Applications of ChatGPT in Dentistry: A Narrative Review

by
Francesco Puleio
1,
Giorgio Lo Giudice
1,*,
Angela Mirea Bellocchio
1,
Ciro Emiliano Boschetti
2 and
Roberto Lo Giudice
3
1
Department of Biomedical and Dental Sciences and Morphofunctional Imaging, Università degli Studi di Messina, 98121 Messina, Italy
2
Multidisciplinary Department of Medical-Surgical and Dental Specialties, Università degli studi della Campania “Luigi Vanvitelli”, 81100 Naples, Italy
3
Department of Human Pathology of Adults and Developmental Age, Università degli Studi di Messina, 98121 Messina, Italy
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(23), 10802; https://doi.org/10.3390/app142310802
Submission received: 10 October 2024 / Revised: 18 November 2024 / Accepted: 20 November 2024 / Published: 21 November 2024
(This article belongs to the Special Issue Digital Dentistry and Oral Health)

Abstract

:
Artificial intelligence (AI), specifically Generative Pre-trained Transformer (GPT) technology, has revolutionized various fields, including medicine and dentistry. The AI model ChatGPT, developed by OpenAI, mimics human language on a large scale, generating coherent and contextually appropriate responses. ChatGPT serves as an auxiliary resource for diagnosis and decision-making across various medical disciplines. This comprehensive narrative review aims to explore how ChatGPT can assist the dental sector, highlighting its potential to enhance various aspects of the discipline. This review includes a literature search on the application of ChatGPT in dentistry, with a focus on the differences between the free version, ChatGPT 3.5, and the more advanced subscription-based version, ChatGPT 4. Specifically, ChatGPT has proven to be effective in enhancing user interaction, providing fast and accurate information and improving the accessibility of knowledge. However, despite these advantages, several limitations are identified, including concerns regarding the accuracy of responses in complex scenarios, ethical considerations surrounding its use, and the need for improved training to handle highly specialized queries. In conclusion, while ChatGPT offers numerous benefits in terms of efficiency and scalability, further research and development are needed to address these limitations, particularly in areas requiring greater precision, ethical oversight, and specialized expertise.

1. Introduction

Artificial intelligence (AI) is a field of computer science that involves developing algorithms capable of performing tasks that would normally require human intelligence [1]. These tasks include recognizing patterns and images, making decisions based on complex data, understanding and translating languages, and the ability to respond to questions or voice commands [2,3].
One of the most significant developments in AI application in recent years has been the creation and release of ChatGPT by OpenAI https://openai.com (accessed on 1 October 2024) (San Francisco, CA, USA). ChatGPT is a large language model based on Generative Pre-trained Transformer (GPT) technology, which uses deep learning techniques to produce text that can mimic human style on a large scale. ChatGPT was launched as a trial version in November 2022 and later made publicly available, gaining immediate attention for its ability to generate coherent and contextually appropriate responses across various fields of knowledge and everyday life [4]. The technology behind ChatGPT, and more generally all GPT models, operates through what is known as transfer learning. Initially, the model is pre-trained on a vast dataset of publicly available texts; this phase is not specific to a task but aims to develop a general understanding of language. Subsequently, it undergoes a refinement process, during which the model is further trained on a more specific text corpus (for example, medical texts) to make its responses more appropriate to certain applications [5,6]. Furthermore, the AI algorithm uses machine learning (ML) and natural language processing techniques to identify meaningful patterns and understand user needs [7]. On the other end, the user applies prompt engineering, a process that guides AI systems to produce desired results. Specifically, it involves writing, refining, and optimizing the starting data (prompts) to encourage AI systems to create specific outputs.
Generative AI systems are designed to generate specific outputs based on the quality of the instructions provided, thus following the rule that quality instructions determine quality results, minimizing bias. The objective is to create queries that help generative AI models understand the language and every nuance to reach a specific goal. Of course, creating unbiased prompts is crucial for ensuring the fairness and effectiveness of the tool. This can be achieved by keeping the prompt clear and specific, providing sufficient context and balancing specificity and generality. Knowing if the training dataset is diverse enough would also reduce the risk of biases.
The applications of AI are numerous, from autonomous vehicle driving to customizing online shopping experiences, advanced robotics, and financial analysis [8]. AI has shown remarkable potential by being applied across all branches of the medical sector: in 2019, a team of researchers from MIT (https://www.mit.edu accessed on 1 October 2024) and Harvard Medical School (https://hms.harvard.edu accessed on 1 October 2024) developed an AI algorithm that could predict the risk of developing breast cancer up to five years before it clinically manifests, using mammograms and analyzing subtle patterns in breast tissue that are invisible to the human eye [9]. Different artificial intelligence models were developed using ultrasound data, patient histories, and medical records to diagnose polycystic ovary syndrome, allowing for earlier and more accurate diagnosis compared to traditional methods based solely on clinical and hormonal analysis [10]. Systems have been developed that integrate AI to assist doctors in choosing the most effective cancer treatments based on a vast database of the medical literature, clinical guidelines, and patient records; these systems have been used in various clinical settings to personalize oncology therapies [11]. ChatGPT can serve as an auxiliary resource for diagnosis and decision-making across various disciplines, such as cardiology, radiology, and urology, or the psychiatric field [12,13]. These examples demonstrate how AI is already a fundamental component in modern medical practice, improving diagnostic accuracy, optimizing treatments, and saving lives through quicker identification of critical conditions. The integration of these advanced technologies in the healthcare sector represents one of the most promising frontiers for the medicine of the future [7].
Similarly, like all fields of medicine, dentistry is also undergoing a digital revolution driven by continuous technological progress and the use of AI among various new tools. Oral and maxillofacial radiology benefits from AI integration in automatic tooth and bone segmentation, identification of dental implant systems, artifact reduction, and pathology detection [14,15,16,17,18]. It has also been shown to be helpful in improving digital impression precision, in assisting in clinical photographic diagnosis, and, generally speaking, in the fields of 3D digital dentistry (i.e., CAD/CAM, 3D printing, 3D scanning) [19,20,21,22,23,24].
Regarding GPTs, while real-world applications are under scrutiny, there are several initiatives aimed at evaluating their precision, practicality, and potential uses, taking into account their versatility. Therefore, this comprehensive narrative review aims to explore how ChatGPT can assist the entire dental sector, analyzing the potential of this technology and the ongoing studies on its use for clinical, research, and educational purposes.

2. Materials and Methods

In this review, a systematic approach was adopted to identify relevant studies. The literature search was conducted using the following electronic databases: Ovid MEDLINE, PubMed, and Web of Science. No searches of other databases or gray literature were performed. The search was conducted up to November 2024. The following search string was used: “ChatGPT” AND “Dentistry”.

2.1. Eligibility Criteria

The full texts of all possibly relevant research papers were chosen, considering as inclusion criteria only peer-reviewed articles that specifically discussed the applications of ChatGPT in all possible dental applications.
Clinical trials, observational studies, reviews, case reports, and case series were considered adequate study designs to ensure the inclusion of a broad range of evidence.
The exclusion criteria that were considered were as follows:
  • Opinion pieces, editorial commentaries, and letters to editor.
  • Studies focused on a specific medical specialty.
  • Papers not focused on AI language models.
  • Studies not in the English language.
  • Papers without available full text.

2.2. Data Extraction and Review Process

Two reviewers (F.P. and G.L.G.) performed data extraction using a standardized extraction form. Findings were compared and arranged thematically. The extracted data included study characteristics, objectives, key findings, and conclusions. Any discrepancies between the two reviewers were resolved through discussion. In cases where agreement could not be reached, a third senior reviewer (R.L.G.) was consulted to ensure an unbiased resolution. A flow chart of paper selection and screening is shown in Figure 1.

3. Results

The search engines yielded a total of 405 results. Duplicate studies were excluded, resulting in 268 remaining studies. Among these, 69 articles were excluded because they did not fit into the study design inclusion criteria. Analyzing the abstracts of 199 studies, 64 were discarded because they did not align with this review’s objective or they did not pertain to ChatGPT or any other AI program, and 45 were excluded because they related to a specific field of medicine (cardiology, orthopedic surgery, neurology, etc.). Finally, 90 articles underwent full-text examination. Each research study was then categorized according to the topic covered (Table 1).

4. Discussion

Currently, ChatGPT is available in a free version as ChatGPT 3.5 and a more advanced version that requires a subscription as ChatGPT 4. Technological differences between the two versions can affect the output from the chatbox. Hirosawa T. et al. assessed the accuracy of differential diagnoses generated by ChatGPT 3.5 and ChatGPT 4, comparing the results with those from human doctors to evaluate their alignment with the correct final diagnosis. The analysis revealed that ChatGPT 4 achieved higher rates of correct diagnosis compared to ChatGPT 3.5 [114].
Using one version or the other for any type of application could therefore yield different results. The advantages of version 4 are listed below:
  • It can process a greater number of words simultaneously compared to ChatGPT 3.5, enabling more extensive conversations and context as well as improved text comprehension.
  • It provides more accurate responses thanks to a larger and more diverse database.
  • It successfully incorporates contextual details and generates responses that are consistent not just with the immediate input but with an entire conversation.
  • It reduces the number of plausible but incorrect responses (known as “hallucinations”), providing a crucial improvement in the trustworthiness of the generated results.
Although version 4 manages to reduce the phenomenon of hallucinations, this issue is still significantly present, potentially compromising the reliability of the responses [115]. One of the major inherent limitations of this technology is indeed its tendency to produce outputs that are not based on truth or objective reality but rather plausible responses. ChatGPT 4 defines artificial hallucination as “the production of information that seems plausible but is incorrect or fabricated. It occurs when the model generates a response based on patterns it has learned during training, even though the information may not be factually accurate or may not exist in real-world knowledge. This can be particularly challenging in situations where accuracy is critical, such as medical advice or factual reporting. Efforts are continuously made to reduce the frequency of these hallucinations by improving the training data and the models’ algorithms.” Hence, the human element remains crucial in analyzing the chatbot’s response and in evaluating its truthfulness, as it cannot be given complete autonomy.
Despite this inherent limitation in GPT technology, the chatbot has proven useful in all fields of medicine as a support tool. This narrative review assesses the areas where ChatGPT can aid the dental sector, regardless of the version used. The literature review has highlighted various areas of dentistry where ChatGPT proves to be particularly effective.

4.1. Applications in Dental Research

Uribe and Maldupa systematically assessed the use of generative AI in dental research, showing how from 2018 to 2024, the use of words signaling the employment of GPTs in dental research abstracts increased from 47.1 per 10,000 papers to 224.2 per 10,000 papers [75]. Due to the language focus of LLM technology, it has significant potential for application in writing, reviewing, and editing research.
Several authors showed how this tool is effective in increasing productivity, inclusivity, and overall help and at the same time raised strong research ethics concerns, often promoting the need of strict guidelines [70,73,74]. Nonetheless, ChatGPT streamlines the synthesis of the extensive dental literature, facilitating systematic reviews by suggesting PICO questions and meta-analyses by identifying gaps and suggesting future research directions, thus enhancing knowledge discovery and evidence-based practice, or generating figures, tables, and other visual elements to summarize data [71,72]. ChatGPT currently does not have direct access to databases of scientific articles such as PubMed, Embase, Web of Science, or Google Scholar; the articles to be included in the review need to be manually added into the chatbox. Hence, the inclusion phase in the review remains a task that requires the operator’s choice and not that of the AI. On the other hand, the subsequent phases of screening and eligibility can be performed by AI.
The use of ChatGPT in research proves useful in correcting sentences with syntactic or grammatical errors and translating them into other languages or in setting up the bibliography according to a well-defined format. However, it cannot be used to generate bibliographic references. Numerous studies suggest that references generated by ChatGPT may indeed be fictitious [4,116,117,118]. In the submission phase, ChatGPT seems to effectively distinguish between predatory and legitimate dental journals, with accuracy rates of 92.5% and 71%, respectively [69].

4.2. Clinical Applications

4.2.1. Diagnostics and Radiology

Although it is not possible to delegate complete diagnostic autonomy, ChatGPT can assist clinicians by providing information on possible diagnoses if symptom descriptions are inputted. ChatGPT-4 demonstrates potential as an auxiliary tool for diagnosing typical and mildly atypical presentations of common diseases, can reliably provide trustworthy and transparent recommendations according to current consensus guidelines, and can clarify doubts [27,28,29,30]. When looking for treatment recommendations, LLMs generally aligned with the current literature up to 75% of the time, although data sources were frequently missing. Both GPT-4 and clinician reviews proposed procedures that might lead to overtreatment. Among the models, GPT-4 demonstrated the highest overall accuracy [26]. However, its performance has been shown to decline with greater atypicality [31].
The performance of the GPT program in describing and providing differential diagnoses for radiolucent lesions in panoramic radiographs is variable and at this stage limited in its use for clinical application [32]. When inputting into the ChatGPT 4 chatbox “If I provide you with images of intraoral X-rays, or orthopantomograms, would you be able to diagnose pathologies such as tooth decay, periodontitis, or other oral pathologies?” the chatbot responds: “I’m not able to diagnose medical conditions or interpret medical images like intraoral X-rays. It’s important to consult a qualified healthcare professional, such as a dentist or radiologist, who can provide accurate diagnoses and appropriate treatment recommendations for tooth decay, periodontitis, or other oral pathologies.” It is not possible to use the ChatGPT 4 chatbox directly as a diagnostic tool. Within the OpenAI GPT platform, it is possible to find numerous pre-trained AIs, or to train a new one, to overcome this limitation.

4.2.2. Traumatology

Regarding dental traumatology, AI and LLMs have been limitedly applied. Ozden et al. stated in their research that although ChatGPT and Google Bard are potential knowledge resources, their consistency and accuracy in responding to dental trauma queries remain limited, with correct answers to 57.5% of the questions [64,65].

4.2.3. Oral and Maxillofacial Surgery

Employment of ChatGPT as an intelligent virtual assistant in oral surgery has been tested, with accuracy results of 71.7%, and response consistency from moderate to almost perfect and good proficiency as a consultant on the subject [38,119]. It was shown to provide valuable information and guidance during orthognathic surgery consultations [39]. Moreover, it was able to assist in patient communication for third molar extraction, dental implantology, and after-surgery follow-ups [41,42,44]. However, Çoban, E., and B. Altay expressed their concerns about potential bias toward specific dental implant brands in these answers [42].
While Isisk et al. described high accuracy and quality in ChatGPT Plus’s responses in the field of oral and maxillofacial surgery, except for the questions requiring a detailed response or a comment, Balel et al. used a modified version named ScholarGPT. The results excelled in oral and maxillofacial surgery questions, providing more consistent and high-quality responses compared to ChatGPT [40,43]. These results suggest that GPT models based on academic databases can provide more accurate and reliable information. Scholar GPT can in fact access a wide range of scholarly resources, including Google Scholar, PubMed, JSTOR, and Arxiv, and it has built-in critical reading skills to help analyze and interpret research papers and PDFs.

4.2.4. Prosthodontics

Freire et al. evaluated the performance of ChatGPT in generating answers about removable dental prostheses (RDPs) and tooth-supported fixed dental prostheses (FDPs). Technical questions were formulated regarding the various stages of prosthesis fabrication, including impressions, vertical dimension, trial phases, occlusal registration, and cementation. The answers showed a reliability of 25.6%, demonstrating that currently, ChatGPT has a limited ability to generate answers related to RDPs and tooth-supported FDPs [62]. The prosthodontist cannot rely on the chatbot to obtain information on correctly making a prosthesis yet.

4.2.5. Periodontology

Regarding periodontal diseases, LLMs seem to consistently offer accurate guidance in most responses and can classify periodontitis at a reasonable level [58,59,61].
Danesh et al., on the other hand, reported that ChatGPT 3.5 and ChatGPT 4 answered 57.9% and 73.6% of in-service questions correctly. They concluded that while ChatGPT 4 showed a higher proficiency compared to ChatGPT 3.5, both chatbot models leave considerable room for misinformation with their responses relating to periodontology [60].

4.2.6. Endodontics

On the quality of endodontic information provided, ChatGPT outperforms dental students in diagnostic accuracy regarding endodontic assessments, and its 3.5 version seems to provide more credible information on topics related to endodontics compared to Google Bard and Bing, although it only gives 60% correct answers [33,34,35,36]. Suarez et al. simulated patient–physician questions related to the field of endodontics. The answers generated by ChatGPT showed high consistency (85.44%) and an average accuracy of 57.33%. However, significant differences in accuracy were observed based on question difficulty, with lower accuracy for easier questions [37].

4.2.7. Orthodontics and Pediatrics

LLMs seem to show great potential in supporting evidence-based orthodontic choices, data digitization, and treatment assistance with a high level of accuracy and completeness to the general orthodontic questions [46,48,49]. Dursun and Bilici Geçer in their work explained how generally accurate, moderately reliable, and moderate- to good-quality answers to questions about clear aligners were given [47]. On the other hand, Abu Arqub et al. demonstrated that ChatGPT has a limited ability to provide completely accurate and reliable responses regarding aligners: 58% of the AI-generated answers were scored as objectively true, 18% were selected facts, 9% were minimal facts, and 15% were false [45].
AI-powered orthodontic workflow was described by Surokova et al., who concluded that it enhances dental practice with precise, personalized treatment and at the same time raises new ethical and legal issues [50].
Rokhshad et al., in two separate studies on special needs and pediatric patients, demonstrated that chatbots showed acceptable consistency for the former group but lower accuracy than dentists for the latter [57,63].

4.2.8. Patients’ Communication and Self-Education

As this technology becomes widely and freely available, it is to be expected that patients will try to inform themselves using these tools, as they began to do when internet search engines became widespread.
Incerti Parenti et al. in his research regarding obstructive sleep apnea showed how ChatGPT-3.5 delivers higher quality and more comprehensive information compared to Google Search, although its responses are less readable [54].
This AI model excels in customizing patient communications, delivering understandable explanations of dental conditions and treatments, postprocedural instructions, and responses to frequently asked questions, thereby enhancing patient engagement and compliance. This can improve patient understanding and compliance with treatment. This possible application has been confirmed in the fields of orthognathic surgery, orthodontics, oral cancer, pediatrics, and postoperative care with overall good results [51,52,53,55,56]. Vassis et al. rated as “neither deficient nor satisfactory” the information received from ChatGPT, with GPT-4 achieving higher scores than GPT-3.5. At the same time, the patients perceived the GPT-generated information as more useful and more comprehensive and experienced less nervousness when reading the GPT-generated information, preferring in 80% of cases the AI-generated information over the standard text [55].

4.3. Administrative Applications

ChatGPT can help create precise and exhaustive patients’ documentation. By entering patient data, clinical findings, and treatment plans, specialists can utilize ChatGPT’s capabilities for text generation to formulate comprehensive reports [68]. This method boosts the quality of documentation, ensuring that critical information is effectively recorded and may support seamless communication among healthcare providers, possibly overcoming language barriers in doctor–patient interaction. While no specific result regarding administrative application in dentistry resulted from the research selection, Eggmann et al. underlined the potential of LLMs in supporting health care professionals in routine written communications and record keeping, potentially enhancing the quality of patient care and saving time, thus reducing costs [25]. However, the input of sensitive health-related information raises privacy and cybersecurity concerns.

4.4. Educational Enhancements

ChatGPT can be effectively employed to improve the dentistry educational field for both students and teachers.

4.4.1. The Students’ Perspective

Abdaljaleel et al. conducted the TAME-ChatGPT (Technology Acceptance Model Edited to Assess ChatGPT Adoption) survey among university students for educational purposes [102]. The results showed that the employment of this LLM is supported due to increased accuracy and speed, ease of use, perceived usefulness, low perceived risks, and low “anxiety” levels among the participants, suggesting a readiness to adopt ChatGPT despite the recognized concerns. The study found that students generally had low “anxiety” about potential drawbacks of ChatGPT, such as declining critical thinking, overdependence on technology, and diminished originality. Instead, they viewed ChatGPT as a valuable educational tool with low perceived “anxiety” about academic integrity or skill development issues. Students’ positive attitude towards the use of LLMs and usefulness is also confirmed by several other authors [103,104,105,106].

4.4.2. The Teachers’ Perspective

A worldwide survey of educators by Uribe et al. showed a prevalent positive yet cautious view towards AI chatbot integration in dental curricula, underscoring the need for clear implementation guidelines [113]. GPTs can generate always-changing questions for exams, especially at the knowledge and comprehension levels, making them useful for large-scale exams provided that supervision on the output is always ensured [107].
Moreover, ChatGPT may generate scenarios, questions, answers, and explanations in real time, building more engaging learning for students. This type of content can simulate a realistic conversation between students and virtual patients or clinical scenarios, offering the chance to exercise in a controlled environment. de Vries et al. included ChatGPT actively in an entire course, deeming it extremely useful also in summarizing general overarching principles, which are sometimes spread over more than one chapter, allowing students to reach the learning goals [109].
Educators can use GPTs to check if tasks were fulfilled using GPTs, adjusting their teaching and assessments to benefit learners while addressing potential misuse [108,112]. Moreover, GPTs can supplement humans in grading paper and essays [110,111].

4.4.3. Mastery and Expertise Tests

Several authors tested LLMs’ capabilities against licensing exams and specialist boards. Their good proficiency was underlined in multiple=choice questions and short answers rather than thematic, open ones or mathematical analysis. The results exceeded the cut-off scores and showed they performed exceptionally well in specific subjects, in one case performing better than licensed dentists, as described by Revilla-Léon et Al., with the exception of the Iranian Endodontics Specialist Board, where the tool employed obtained a score of 40/100 [90,91,92,93,95,96,97,98,99,100,101]. Another underwhelming result was achieved in oral and maxillofacial surgery. Specifically, in Jeong et al.’s study, the students’ overall accuracy rate was 81.2%, while that of the chatbots varied from 50.0% to 65.4%. ChatGPT Plus achieved a higher accuracy rate for basic knowledge than the students (93.8% vs. 78.7%). However, all chatbots performed poorly in image interpretation, with accuracy rates below 35.0% [94]. On the topic of image recognition, Morishita et al. showed how the new ChatGPT-4V’s image recognition feature exhibited limitations, especially in handling image-intensive and complex clinical practical questions, and is not yet fully suitable as an educational support tool for dental students at its current stage [97].
Tests were also performed to assess the proficiency against patients’ oral health-related queries, in-service exam questions, oral lesion diagnosis, and recognized assessments in several dental subjects with good outcomes [78,79,80,81,82,83,84,85,87]. Regarding oral and maxillofacial surgery, Puladi et al. highlighted that classic diseases are underrepresented, and the current literature on large language models in this field lacks sufficient evidence [86].
In the dispute between GPT use vs. literature research methodology, Kavadella et al. showed how students using ChatGPT to compose a learning assignment performed significantly better than their fellow students who used the literature research methodology, while in Saravia-Rojas et al.’s work, higher values through the traditional method than with ChatGPT were reached regarding final scores and scores for the criteria of utilization of evidence, evaluation of arguments, and generation of alternatives. Nonetheless, the students highly valued the experience of using ChatGPT for academic tasks [88,89].
Shortcomings of this technology in the educational field have been stressed by authors regarding cognitive and general bias, inaccuracies, privacy issues, and the risk of overreliance [76,77].

4.5. Ethical and Practical Considerations

Legal issues related to the use of ChatGPT are troubling due to unclear accountability and copyright matters [116,120,121]. Fighting plagiarism and academic incorrectness requires the creation of solid ethical principles that harmonize the usefulness of AI with personal interests. Researchers who use ChatGPT for writing academic papers should openly disclose its use.
There are numerous problems related to patient privacy, which could be compromised [67,121]. When sensitive patient information is entered into the ChatGPT chatbox, it is unclear how these data are stored or for how long. Even if it is unintentional, data may be stored on the AI’s servers, increasing the risk of unauthorized access. These data could be used to further train the AI model. This means that sensitive information could be used in ways that were not intended or authorized, including sharing with third parties. Any computer system, including those hosting ChatGPT, can be vulnerable to security breaches. In the event of a breach, sensitive patient information could be exposed to bad actors. To protect patients’ data, the use of offline LLMs such as META LLaMA or any kind of anonymization could be a solution. Nonetheless, offline LLMs should be proposed as standard, since anonymization still raises possible concerns due to the data being sent and processed to external servers anyway [68]. This use also lacks compliance with privacy regulations: using AI to manage patient data may not comply with privacy regulations such as the GDPR in Europe or HIPAA in the United States, if the AI is not properly configured to protect and manage data in compliance with these laws. To mitigate these risks, it is essential that any use of AI, such as ChatGPT, in a clinical context is carefully evaluated and managed [122,123].
At present (November 2024), ChatGPT is accessible at no cost in its GPT-4o mini version and for the use of custom GPTs. Monthly fees are charged for updated and multimodal builds (also known as GPT-4, GPT-4o, and DALL-E) and to create customized GPT versions. Yet, there exists a concern that ChatGPT and comparable artificial intelligence systems could become prohibitively costly to access as their popularity increases. This might lead to a disproportionate allocation of resources among researchers across different disciplines [66,124]. Various publications have even credited the chatbot with authorship [125,126]. Nevertheless, attributing authorship carries significant ethical implications for scholarly articles and has emerged as a pressing issue in academic journals. A recent article in Nature declared that an AI chatbot should not be recognized as author of a scientific paper, as it cannot be accountable for the claims made in the article [127].

4.6. Future Directions

As AI technologies continue to advance, ChatGPT is expected to become more integrated into dental practice, offering advanced diagnostic tools, personalized treatment options, and real-time predictive analytics.
It could become an even more effective tool to assist dentists in preliminary diagnoses and in providing personalized advice to patients; it could be trained with dentistry-specific datasets to recognize symptoms, suggest diagnostic tests, and offer preliminary treatment options based on the most up-to-date practices. As GPTs become more advanced, if a version specifically trained for dental offices were integrated into the management software, it could help with patient management, appointment scheduling, and even managing prescriptions and communications with other healthcare professionals. Despite notable successes, AI technology and specifically models like ChatGPT have numerous limitations. The quality and reliability of responses may be inconsistent, particularly on topics with a wide variety of opinions or less documented knowledge. Furthermore, these models can generate plausible but incorrect responses, a phenomenon known as “hallucination” in the generated texts [115].
These limitations raise significant concerns, especially in critical fields such as medicine, where the accuracy of information is crucial. To overcome these obstacles, AI developers are working on multiple fronts: improving learning algorithms to reduce the risk of errors, increasing the size and quality of training datasets to better cover the breadth and depth of human knowledge, and developing new methodologies for validating the responses generated to ensure their reliability. Furthermore, the implementation of transparency systems in AI answers is another crucial field of development, which will help build confidence in its use in professional systems and contexts, including dentistry and medicine in general.

4.7. Alternative AI Models to ChatGPT

There are several alternative language models to ChatGPT, each with unique characteristics: Anthropic’s Claude, Google’s Gemini (formerly known as Bard), Microsoft’s Copilot, Meta’s Llama, and BERT (and its derivatives such as ALBERT, RoBERTa, DistilBERT, SpanBERT, and DeBERTa) are just few of the available language models that can be used depending on the task to be performed. While GPT-4 is a generative model based on a unidirectional (auto-regressive) transformer architecture, BERT, RoBERTa, and ALBERT use a bidirectional architecture. This difference allows bidirectional models to better understand the context of words by simultaneously considering the words to the left and right of a given word. GPT-4, on the other hand, generates the text one word at a time sequentially, but it is optimized to produce coherent and fluent texts. Some models like Copilot are based on GPT-4 and enhanced with data coming from proprietary search engines like Bing or Google search. Generally speaking, every model is optimized for a variety of natural language processing tasks, each with unique approaches to transformer architectures and training techniques. Therefore, the choice of the best model depends on the usage context and specific application needs. Studies comparing various versions of the same model or different language models are increasingly appearing in scholarly publications, highlighting the growing interest in these models in dentistry and in other specific fields of medicine [49,71,98,128,129].

4.8. Limitations

At the time of conducting this review, there is limited original research on the application of ChatGPT in dentistry.
The most significant limitation in comparing the selected research results lies in the methods used. Different LLMs, or different versions of the same LLM, vary in the quality of their responses. This issue, along with the automatic “contextual refinement” performed by LLMs, makes comparisons between papers difficult, if not impossible, unless each material and methods section specifies the exact version and build of the software used and details any contextual refinement or prompt engineering applied, if any. Another limitation is the model’s tendency to generate “hallucinations”, that is, responses that seem plausible but are actually incorrect or invented. This phenomenon represents a significant challenge and can lead to erroneous diagnoses or inaccurate medical advice if the model is used without human supervision. To address this issue, various methods can be employed. One is fine-tuning on curated data, using high-quality datasets to ensure reliable information. Retrieval-augmented generation (RAG) combines the model’s generative abilities with external sources for more accurate responses. Model calibration adjusts prediction confidence levels for reliability. Prompt engineering guides models to better outputs by crafting precise questions and instructions. Human-in-the-loop feedback involves humans correcting hallucinations, aiding the model’s learning. Post-processing techniques filter out hallucinated content for more accurate final responses.
ChatGPT is not designed to interpret medical images such as intraoral X-rays or orthopantomographs, which limits its usefulness as a direct diagnostic tool. Another critical aspect is the necessity for constant system updates to keep the knowledge up-to-date with recent developments in the medical field, thereby avoiding the dissemination of outdated or incorrect information. The lack of direct access to scientific databases limits its ability to provide responses based on the updated medical literature, making the experts’ intervention indispensable to ensure the accuracy of the information used and generated by the model.
Prompt engineering, while a powerful tool for guiding AI systems, does come with its own set of challenges that might be seen as limitations to AI’s use and to study comparison. The end-user should be aware of the training datasets used, since their quality may impact the performance of the model or may present biases. The model may not behave as intended due to prompt formulation, or communication barriers may rise when it does not fully understand the intent or the subtle distinctions behind the prompts. It can be challenging to obtain the exact output you want from the AI model in the first attempt.
“Contextual understanding” or “contextual refinement” is when LLMs refine their answers using the context from earlier interactions to provide more accurate and relevant responses. If research studies do not account for this automatic adjustment, the results may be biased. Therefore, determining the best way to initiate a prompt, and in the end, understanding and evaluating the reasoning behind the generated responses can be tricky. These challenges highlight the complexity of prompt engineering and the need for careful consideration and ongoing refinement to achieve optimal, comparable results.

5. Conclusions

The increasing number of studies published in the literature over the past two years on ChatGPT, and large language models in general, show a real interest in this tool and its application in diverse areas of dentistry:
  • Dental research: It has been shown to be useful in study designing, abstract generation, draft correction, syntax error correction, translations, and reference formatting. At present, ChatGPT cannot be trusted to generate bibliographic references from a text due to frequent “hallucinations”, errors, and/or outdated information.
  • Clinical application: ChatGPT can be used, only if properly trained by the operator, to assist the diagnostic workflow. However, it is not yet possible to use it directly as a diagnostic tool, and answers should not be accepted without question, primarily because there is a chance that an outdated version of the model might be in use.
  • Administrative applications: Through inputted clinical findings and treatment plans, the tool is able to generate high-quality reports, spot trends, and manage patient follow-up appointments.
  • Educational enhancements: By simulating a realistic conversation, it is capable of assisting students in generating scenarios, questions, answers, and explanations in a quick and engaging way. It can compose ever-changing tests and grade them.
  • Ethical and practical considerations: The processing and storage of patients’ sensitive data raises significant privacy concerns. This is a crucial aspect to consider in the deployment of such tools in a clinical setting.
The results of this narrative review show embryonic results with little clinical application with patients; thus, ChatGPT is still far from being used on a daily basis. Nonetheless, ChatGPT stands as a pivotal innovation in dentistry, promising to enhance research capabilities, clinical accuracy, administrative efficiency, and educational processes. Effective integration, coupled with stringent ethical practices and ongoing validation, is essential for realizing the full potential of AI in dentistry.

Author Contributions

Conceptualization, F.P. and A.M.B.; methodology, G.L.G.; validation, A.M.B.; formal analysis, F.P.; writing—original draft preparation, F.P. and G.L.G.; writing—review and editing, G.L.G. and C.E.B.; project administration, R.L.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially funded by the European Union (NextGenerationEU), through the MUR_PNRR project SAMOTHRACE.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chen, M.; Decary, M. Artificial intelligence in healthcare: An essential guide for health leaders. Healthc. Manag. Forum 2020, 33, 10–18. [Google Scholar] [CrossRef] [PubMed]
  2. Aggarwal, A.; Tam, C.C.; Wu, D.; Li, X.; Qiao, S. Artificial Intelligence-Based Chatbots for Promoting Health Behavioral Changes: Systematic Review. J. Med. Internet Res. 2023, 25, E40789. [Google Scholar] [CrossRef] [PubMed]
  3. Wailthare, S.; Gaikwad, T.; Khadse, K.; Dubey, P. Artificial intelligence-based chat-bot. Int. Res. J. Eng. Technol. 2018, 5, 2305–2306. [Google Scholar]
  4. Eysenbach, G. The Role of ChatGPT, Generative Language Models, and Artificial Intelligence in Medical Education: A Conversation With ChatGPT and a Call for Papers. JMIR Med. Educ. 2023, 9, E46885. [Google Scholar] [CrossRef]
  5. Galvao Gomes da Silva, J.; Kavanagh, D.J.; Belpaeme, T.; Taylor, L.; Beeson, K.; Andrade, J. Experiences of a Motivational Interview Delivered by a Robot: Qualitative Study. J. Med. Internet Res. 2018, 20, E116. [Google Scholar] [CrossRef] [PubMed]
  6. Stephens, T.N.; Joerin, A.; Rauws, M.; Werk, L.N. Feasibility of pediatric obesity and prediabetes treatment support through Tess, the AI behavioral coaching chatbot. Transl. Behav. Med. 2019, 9, 440–447. [Google Scholar] [CrossRef]
  7. Milne-Ives, M.; de Cock, C.; Lim, E.; Shehadeh, M.H.; de Pennington, N.; Mole, G.; Normando, E.; Meinert, E. The Effectiveness of Artificial Intelligence Conversational Agents in Health Care: Systematic Review. J. Med. Internet Res. 2020, 22, E20346. [Google Scholar] [CrossRef]
  8. Krishnan, C.; Gupta, A.; Gupta, A.; Singh, G. Impact of Artificial Intelligence-Based Chatbots on Customer Engagement and Business Growth. In Deep Learning for Social Media Data Analytics; Serrano-Estrada, L., Saxena, A., Biswas, A., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2022; pp. 195–210. [Google Scholar]
  9. Yala, A.; Lehman, C.; Schuster, T.; Portnoi, T.; Barzilay, R. A Deep Learning Mammography-based Model for Improved Breast Cancer Risk Prediction. Radiology 2019, 292, 60–66. [Google Scholar] [CrossRef]
  10. Verma, P.; Maan, P.; Gautam, R.; Arora, T. Unveiling the Role of Artificial Intelligence (AI) in Polycystic Ovary Syndrome (PCOS) Diagnosis: A Comprehensive Review. Reprod. Sci. 2024, 31, 2901–2915. [Google Scholar] [CrossRef]
  11. Paranjape, K.; Schinkel, M.; Nannan Panday, R.; Car, J.; Nanayakkara, P. Introducing Artificial Intelligence Training in Medical Education. JMIR Med. Educ. 2019, 5, E16048. [Google Scholar] [CrossRef]
  12. Dave, T.; Athaluri, S.A.; Singh, S. ChatGPT in medicine: An overview of its applications, advantages, limitations, future prospects, and ethical considerations. Front. Artif. Intell. 2023, 6, 1169595. [Google Scholar] [CrossRef] [PubMed]
  13. Prada, P.; Perroud, N.; Thorens, G. Artificial intelligence and psychiatry: Questions from psychiatrists to ChatGPT. Rev. Med. Suisse 2023, 19, 532–536. [Google Scholar] [CrossRef]
  14. Yang, J.; Xie, Y.; Liu, L.; Xia, B.; Cao, Z.; Guo, C. Automated Dental Image Analysis by Deep Learning on Small Dataset. In Proceedings of the 2018 IEEE 42nd Annual Computer Software and Applications Conference (COMPSAC), Tokyo, Japan, 23–27 July 2018; pp. 492–497. [Google Scholar] [CrossRef]
  15. Schwendicke, F.; Golla, T.; Dreher, M.; Krois, J. Convolutional neural networks for dental image diagnostics: A scoping review. J. Dent. 2019, 91, 103226. [Google Scholar] [CrossRef] [PubMed]
  16. Lee, J.H.; Jeong, S.N. Efficacy of deep convolutional neural network algorithm for the identification and classification of dental implant systems, using panoramic and periapical radiographs: A pilot study. Medicine 2020, 99, E20787. [Google Scholar] [CrossRef] [PubMed]
  17. Qiu, B.; Guo, J.; Kraeima, J.; Glas, H.H.; Borra, R.J.H.; Witjes, M.J.H.; van Ooijen, P.M.A. Automatic segmentation of the mandible from computed tomography scans for 3D virtual surgical planning using the convolutional neural network. Phys. Med. Biol. 2019, 64, 175020. [Google Scholar] [CrossRef]
  18. Bonny, T.; Al Nassan, W.; Obaideen, K.; Al Mallahi, M.N.; Mohammad, Y.; El-Damanhoury, H.M. Contemporary Role and Applications of Artificial Intelligence in Dentistry. F1000Res 2023, 12, 1179. [Google Scholar] [CrossRef]
  19. Kattadiyil, M.T.; Mursic, Z.; AlRumaih, H.; Goodacre, C.J. Intraoral scanning of hard and soft tissues for partial removable dental prosthesis fabrication. J. Prosthet. Dent. 2014, 112, 444–448. [Google Scholar] [CrossRef]
  20. Engels, P.; Meyer, O.; Schonewolf, J.; Schlickenrieder, A.; Hickel, R.; Hesenius, M.; Gruhn, V.; Kuhnisch, J. Automated detection of posterior restorations in permanent teeth using artificial intelligence on intraoral photographs. J. Dent. 2022, 121, 104124. [Google Scholar] [CrossRef]
  21. Li, H.; Sakai, T.; Tanaka, A.; Ogura, M.; Lee, C.; Yamaguchi, S.; Imazato, S. Interpretable AI Explores Effective Components of CAD/CAM Resin Composites. J. Dent. Res. 2022, 101, 1363–1371. [Google Scholar] [CrossRef]
  22. Rojek, I.; Mikolajewski, D.; Dostatni, E.; Macko, M. AI-Optimized Technological Aspects of the Material Used in 3D Printing Processes for Selected Medical Applications. Materials 2020, 13, 5437. [Google Scholar] [CrossRef]
  23. Ng, W.L.; Chan, A.; Ong, Y.S.; Chua, C.K. Deep learning for fabrication and maturation of 3D bioprinted tissues and organs. Virtual Phys. Prototyp. 2020, 15, 340–358. [Google Scholar] [CrossRef]
  24. Liu, Y.; Shang, X.; Shen, Z.; Hu, B.; Wang, Z.; Xiong, G. 3D Deep Learning for 3D Printing of Tooth Model. In Proceedings of the 2019 IEEE International Conference on Service Operations and Logistics, and Informatics (SOLI), Zhengzhou, China, 6–8 November 2019; pp. 274–279. [Google Scholar] [CrossRef]
  25. Eggmann, F.; Weiger, R.; Zitzmann, N.; Blatz, M. Implications of large language models such as ChatGPT for dental medicine. J. Esthet. Restor. Dent. 2023, 35, 1098–1102. [Google Scholar] [CrossRef] [PubMed]
  26. Baig, Z.; Lawrence, D.; Ganhewa, M.; Cirillo, N. Accuracy of Treatment Recommendations by Pragmatic Evidence Search and Artificial Intelligence: An Exploratory Study. Diagnostics 2024, 14, 527. [Google Scholar] [CrossRef]
  27. Islam, A.; Banerjee, A.; Wati, S.M.; Banerjee, S.; Shrivastava, D.; Srivastava, K.C. Utilizing Artificial Intelligence Application for Diagnosis of Oral Lesions and Assisting Young Oral Histopathologist in Deriving Diagnosis from Provided Features—A Pilot study. J. Pharm. Bioallied Sci. 2024, 16, S1136–S1139. [Google Scholar] [CrossRef]
  28. Mohammad-Rahimi, H.; Khoury, Z.; Alamdari, M.I.; Rokhshad, R.; Motie, P.; Parsa, A.; Tavares, T.; Sciubba, J.; Price, J.; Sultan, A. Performance of AI chatbots on controversial topics in oral medicine, pathology, and radiology. Oral. Surg. Oral. Med. Oral. Pathol. Oral. Radiol. 2024, 137, 508–514. [Google Scholar] [CrossRef]
  29. Russe, M.F.; Rau, A.; Ermer, M.A.; Rothweiler, R.; Wenger, S.; Kloble, K.; Schulze, R.K.W.; Bamberg, F.; Schmelzeisen, R.; Reisert, M.; et al. A content-aware chatbot based on GPT 4 provides trustworthy recommendations for Cone-Beam CT guidelines in dental imaging. Dentomaxillofacial Radiol. 2024, 53, 109–114. [Google Scholar] [CrossRef] [PubMed]
  30. Sahu, P.K.; Benjamin, L.A.; Singh Aswal, G.; Williams-Persad, A. ChatGPT in research and health professions education: Challenges, opportunities, and future directions. Postgrad. Med. J. 2023, 100, 50–55. [Google Scholar] [CrossRef]
  31. Shikino, K.; Shimizu, T.; Otsuka, Y.; Tago, M.; Takahashi, H.; Watari, T.; Sasaki, Y.; Iizuka, G.; Tamura, H.; Nakashima, K.; et al. Evaluation of ChatGPT-Generated Differential Diagnosis for Common Diseases With Atypical Presentation: Descriptive Research. JMIR Med. Educ. 2024, 10, E58758. [Google Scholar] [CrossRef] [PubMed]
  32. Silva, T.P.; Andrade-Bortoletto, M.F.S.; Ocampo, T.S.C.; Alencar-Palha, C.; Bornstein, M.M.; Oliveira-Santos, C.; Oliveira, M.L. Performance of a commercially available Generative Pre-trained Transformer (GPT) in describing radiolucent lesions in panoramic radiographs and establishing differential diagnoses. Clin. Oral. Investig. 2024, 28, 204. [Google Scholar] [CrossRef]
  33. Mohammad-Rahimi, H.; Ourang, S.A.; Pourhoseingholi, M.A.; Dianat, O.; Dummer, P.M.H.; Nosrat, A. Validity and reliability of artificial intelligence chatbots as public sources of information on endodontics. Int. Endod. J. 2024, 57, 305–314. [Google Scholar] [CrossRef] [PubMed]
  34. Ourang, S.A.; Sohrabniya, F.; Mohammad-Rahimi, H.; Dianat, O.; Aminoshariae, A.; Nagendrababu, V.; Dummer, P.M.H.; Duncan, H.F.; Nosrat, A. Artificial intelligence in endodontics: Fundamental principles, workflow, and tasks. Int. Endod. J. 2024, 57, 1546–1565. [Google Scholar] [CrossRef]
  35. Qutieshat, A.; Al Rusheidi, A.; Al Ghammari, S.; Alarabi, A.; Salem, A.; Zelihic, M. Comparative analysis of diagnostic accuracy in endodontic assessments: Dental students vs. artificial intelligence. Diagnosis 2024, 11, 259–265. [Google Scholar] [CrossRef]
  36. Snigdha, N.T.; Batul, R.; Karobari, M.I.; Adil, A.H.; Dawasaz, A.A.; Hameed, M.S.; Mehta, V.; Noorani, T.Y. Assessing the Performance of ChatGPT 3.5 and ChatGPT 4 in Operative Dentistry and Endodontics: An Exploratory Study. Hum. Behav. Emerg. Tech. 2024, 2024, 8. [Google Scholar] [CrossRef]
  37. Suarez, A.; Diaz-Flores Garcia, V.; Algar, J.; Gomez Sanchez, M.; Llorente de Pedro, M.; Freire, Y. Unveiling the ChatGPT phenomenon: Evaluating the consistency and accuracy of endodontic question answers. Int. Endod. J. 2024, 57, 108–113. [Google Scholar] [CrossRef]
  38. Acar, A.H. Can natural language processing serve as a consultant in oral surgery? J. Stomatol. Oral. Maxillofac. Surg. 2024, 125, 101724. [Google Scholar] [CrossRef]
  39. Alten, A.; Gündeş, E.; Tuncer, E.; Kozanoğlu, E.; Akalın, B.E.; Emekli, U. Integrating artificial intelligence in orthognathic surgery: A case study of ChatGPT’s role in enhancing physician-patient consultations for dentofacial deformities. J. Plast. Reconstr. Aesthet. Surg. 2023, 87, 405–407. [Google Scholar] [CrossRef] [PubMed]
  40. Balel, Y. ScholarGPT’s performance in oral and maxillofacial surgery. J. Stomatol. Oral. Maxillofac. Surg. 2024, 102114. [Google Scholar] [CrossRef] [PubMed]
  41. Cai, Y.; Zhao, R.; Zhao, H.; Li, Y.; Gou, L. Exploring the use of ChatGPT/GPT-4 for patient follow-up after oral surgeries. Int. J. Oral. Maxillofac. Surg. 2024, 53, 867–872. [Google Scholar] [CrossRef]
  42. Çoban, E.; Altay, B. ChatGPT May Help Inform Patients in Dental Implantology. Int. J. Oral. Maxillofac. Implant. 2024, 39, 203–208. [Google Scholar] [CrossRef] [PubMed]
  43. Isik, G.; Kafadar-Gurbuz, I.; Elgun, F.; Kara, R.U.; Berber, B.; Ozgul, S.; Gunbay, T. Is Artificial Intelligence a Useful Tool for Clinical Practice of Oral and Maxillofacial Surgery? J. Craniofacial Surg. 2024, 10–97. [Google Scholar]
  44. Jacobs, T.; Shaari, A.; Gazonas, C.B.; Ziccardi, V.B. Is ChatGPT an Accurate and Readable Patient Aid for Third Molar Extractions? J. Oral. Maxillofac. Surg. 2024, 82, 1239–1245. [Google Scholar] [CrossRef]
  45. Abu Arqub, S.; Al-Moghrabi, D.; Allareddy, V.; Upadhyay, M.; Vaid, N.; Yadav, S. Content analysis of AI-generated (ChatGPT) responses concerning orthodontic clear aligners. Angle Orthod. 2024, 94, 263–272. [Google Scholar] [CrossRef]
  46. Daraqel, B.; Wafaie, K.; Mohammed, H.; Cao, L.; Mheissen, S.; Liu, Y.; Zheng, L. The performance of artificial intelligence models in generating responses to general orthodontic questions: ChatGPT vs Google Bard. Am. J. Orthod. Dentofac. Orthop. 2024, 165, 652–662. [Google Scholar] [CrossRef]
  47. Dursun, D.; Bilici Geçer, R. Can artificial intelligence models serve as patient information consultants in orthodontics? BMC Med. Inf. Inform. Decis. Mak. 2024, 24, 211. [Google Scholar] [CrossRef]
  48. Lima, N.G.M.; Costa, L.; Santos, P.B. ChatGPT in orthodontics: Limitations and possibilities. Australas. Orthod. J. 2024, 40, 19–21. [Google Scholar] [CrossRef]
  49. Makrygiannakis, M.A.; Giannakopoulos, K.; Kaklamanos, E.G. Evidence-based potential of generative artificial intelligence large language models in orthodontics: A comparative study of ChatGPT, Google Bard, and Microsoft Bing. Eur. J. Orthod. 2024, cjae017. [Google Scholar] [CrossRef]
  50. Surovková, J.; Haluzová, S.; Strunga, M.; Urban, R.; Lifková, M.; Thurzo, A. The New Role of the Dental Assistant and Nurse in the Age of Advanced Artificial Intelligence in Telehealth Orthodontic Care with Dental Monitoring: Preliminary Report. Appl. Sci.-Basel 2023, 13, 16. [Google Scholar] [CrossRef]
  51. Batool, I.; Naved, N.; Kazmi, S.M.R.; Umer, F. Leveraging Large Language Models in the delivery of post-operative dental care: A comparison between an embedded GPT model and ChatGPT. BDJ Open 2024, 10, 48. [Google Scholar] [CrossRef]
  52. Gugnani, N.; Pandit, I.K.; Gupta, M.; Gugnani, S.; Kathuria, S. Parental concerns about oral health of children: Is ChatGPT helpful in finding appropriate answers? J. Indian. Soc. Pedod. Prev. Dent. Apr. Jun. 2024, 42, 104–111. [Google Scholar] [CrossRef]
  53. Hassona, Y.; Alqaisi, D.; Al-Haddad, A.; Georgakopoulou, E.A.; Malamos, D.; Alrashdan, M.S.; Sawair, F. How good is ChatGPT at answering patients’ questions related to early detection of oral (mouth) cancer? Oral. Surg. Oral. Med. Oral. Pathol. Oral. Radiol. 2024, 138, 269–278. [Google Scholar] [CrossRef] [PubMed]
  54. Incerti Parenti, S.; Bartolucci, M.L.; Biondi, E.; Maglioni, A.; Corazza, G.; Gracco, A.; Alessandri-Bonetti, G. Online Patient Education in Obstructive Sleep Apnea: ChatGPT versus Google Search. Healthcare 2024, 12, 1781. [Google Scholar] [CrossRef] [PubMed]
  55. Vassis, S.; Powell, H.; Petersen, E.; Barkmann, A.; Noeldeke, B.; Kristensen, K.D.; Stoustrup, P. Large-Language Models in Orthodontics: Assessing Reliability and Validity of ChatGPT in Pretreatment Patient Education. Cureus 2024, 16, E68085. [Google Scholar] [CrossRef]
  56. Yurdakurban, E.; Topsakal, K.G.; Duran, G.S. A comparative analysis of AI-based chatbots: Assessing data quality in orthognathic surgery related patient information. J. Stomatol. Oral. Maxillofac. Surg. 2024, 125, 101757. [Google Scholar] [CrossRef]
  57. Rokhshad, R.; Zhang, P.; Mohammad-Rahimi, H.; Pitchika, V.; Entezari, N.; Schwendicke, F. Accuracy and consistency of chatbots versus clinicians for answering pediatric dentistry questions: A pilot study. J. Dent. 2024, 144, 104938. [Google Scholar] [CrossRef]
  58. Alan, R.; Alan, B.M. Utilizing ChatGPT-4 for Providing Information on Periodontal Disease to Patients: A DISCERN Quality Analysis. Cureus 2023, 15, E46213. [Google Scholar] [CrossRef]
  59. Babayiğit, O.; Tastan Eroglu, Z.; Ozkan Sen, D.; Ucan Yarkac, F. Potential Use of ChatGPT for Patient Information in Periodontology: A Descriptive Pilot Study. Cureus 2023, 15, E48518. [Google Scholar] [CrossRef]
  60. Danesh, A.; Pazouki, H.; Danesh, F.; Danesh, A.; Vardar-Sengul, S. Artificial intelligence in dental education: ChatGPT’s performance on the periodontic in-service examination. J. Periodontol. 2024, 95, 682–687. [Google Scholar] [CrossRef]
  61. Tastan Eroglu, Z.; Babayigit, O.; Ozkan Sen, D.; Ucan Yarkac, F. Performance of ChatGPT in classifying periodontitis according to the 2018 classification of periodontal diseases. Clin. Oral. Investig. 2024, 28, 407. [Google Scholar] [CrossRef] [PubMed]
  62. Freire, Y.; Laorden, A.S.; Perez, J.O.; Sanchez, M.G.; Garcia, V.D.-F.; Suarez, A. ChatGPT performance in prosthodontics: Assessment of accuracy and repeatability in answer generation. J. Prosthet. Dent. 2024, 131, 659.e651–659.e656. [Google Scholar] [CrossRef] [PubMed]
  63. Rokhshad, R.; Fadul, M.; Zhai, G.; Carr, K.; Jackson, J.G.; Zhang, P. A Comparative Analysis of Responses of Artificial Intelligence Chatbots in Special Needs Dentistry. Pediatr. Dent. 2024, 46, 337–344. [Google Scholar] [PubMed]
  64. Khan, M.K. Novel applications of artificial intelligence, machine learning, and deep learning-based modalities in dental traumatology: An overview of evidence-based literature. MRIMS J. Health Sci. 2024, 12, 223–227. [Google Scholar] [CrossRef]
  65. Ozden, I.; Gokyar, M.; Ozden, M.E.; Sazak Ovecoglu, H. Assessment of artificial intelligence applications in responding to dental trauma. Dent. Traumatol. 2024, 40, 722–729. [Google Scholar] [CrossRef]
  66. Alhaidry, H.M.; Fatani, B.; Alrayes, J.O.; Almana, A.M.; Alfhaed, N.K. ChatGPT in Dentistry: A Comprehensive Review. Cureus 2023, 15, e38317. [Google Scholar] [CrossRef]
  67. de Souza, L.L.; Pontes, H.A.R.; Martins, M.D.; Fonesca, F.P.; Corrêa, F.; Coracin, F.L.; Khurram, S.A.; Hagag, A.; Santos-Silva, A.R.; Vargas, P.A.; et al. ChatGPT and dentistry: A step toward the future. Gen. Dent. 2024, 72, 72–77. [Google Scholar]
  68. Huang, H.Y.; Zheng, O.; Wang, D.D.; Yin, J.Y.; Wang, Z.J.; Ding, S.X.; Yin, H.; Xu, C.; Yang, R.J.; Zheng, Q.; et al. ChatGPT for shaping the future of dentistry: The potential of multi-modal large language model. Int. J. Oral. Sci. 2023, 15, 13. [Google Scholar] [CrossRef] [PubMed]
  69. Al-Moghrabi, D.; Abu Arqub, S.; Maroulakos, M.P.; Pandis, N.; Fleming, P.S. Can ChatGPT identify predatory biomedical and dental journals? A cross-sectional content analysis. J. Dent. 2024, 142, 104840. [Google Scholar] [CrossRef]
  70. Bagde, H.; Dhopte, A.; Alam, M.K.; Basri, R. A systematic review and meta-analysis on ChatGPT and its utilization in medical and dental research. Heliyon 2023, 9, E23050. [Google Scholar] [CrossRef]
  71. Demir, G.B.; Sukut, Y.; Duran, G.S.; Topsakal, K.G.; Gorgulu, S. Enhancing systematic reviews in orthodontics: A comparative examination of GPT-3.5 and GPT-4 for generating PICO-based queries with tailored prompts and configurations. Eur. J. Orthod. 2024, 46, cjae011. [Google Scholar] [CrossRef]
  72. Fatani, B. ChatGPT for Future Medical and Dental Research. Cureus 2023, 15, E37285. [Google Scholar] [CrossRef] [PubMed]
  73. George Pallivathukal, R.; Kyaw Soe, H.H.; Donald, P.M.; Samson, R.S.; Hj Ismail, A.R. ChatGPT for Academic Purposes: Survey Among Undergraduate Healthcare Students in Malaysia. Cureus 2024, 16, E53032. [Google Scholar] [CrossRef] [PubMed]
  74. Tiwari, A.; Kumar, A.; Jain, S.; Dhull, K.S.; Sajjanar, A.; Puthenkandathil, R.; Paiwal, K.; Singh, R. Implications of ChatGPT in Public Health Dentistry: A Systematic Review. Cureus 2023, 15, E40367. [Google Scholar] [CrossRef]
  75. Uribe, S.E.; Maldupa, I. Estimating the use of ChatGPT in dental research publications. J. Dent. 2024, 149, 105275. [Google Scholar] [CrossRef] [PubMed]
  76. Claman, D.; Sezgin, E. Artificial Intelligence in Dental Education: Opportunities and Challenges of Large Language Models and Multimodal Foundation Models. JMIR Med. Educ. 2024, 10, E52346. [Google Scholar] [CrossRef] [PubMed]
  77. Roganović, J. Familiarity with ChatGPT Features Modifies Expectations and Learning Outcomes of Dental Students. Int. Dent. J. 2024, 74, 1456–1462. [Google Scholar] [CrossRef]
  78. Albagieh, H.; Alzeer, Z.O.; Alasmari, O.N.; Alkadhi, A.A.; Naitah, A.N.; Almasaad, K.F.; Alshahrani, T.S.; Alshahrani, K.S.; Almahmoud, M.I. Comparing Artificial Intelligence and Senior Residents in Oral Lesion Diagnosis: A Comparative Study. Cureus 2024, 16, E51584. [Google Scholar] [CrossRef] [PubMed]
  79. Ali, K.; Barhom, N.; Tamimi, F.; Duggal, M. ChatGPT-A double-edged sword for healthcare education? Implications for assessments of dental students. Eur. J. Dent. Educ. 2024, 28, 206–211. [Google Scholar] [CrossRef]
  80. Aminoshariae, A.; Nosrat, A.; Nagendrababu, V.; Dianat, O.; Mohammad-Rahimi, H.; O’Keefe, A.; Setzer, F. Artificial Intelligence in Endodontic Education. J. Endod. 2024, 50, 562–578. [Google Scholar] [CrossRef]
  81. Giannakopoulos, K.; Kavadella, A.; Salim, A.A.; Stamatopoulos, V.; Kaklamanos, E.G. Evaluation of the Performance of Generative AI Large Language Models ChatGPT, Google Bard, and Microsoft Bing Chat in Supporting Evidence-Based Dentistry: Comparative Mixed Methods Study. J. Med. Internet Res. 2023, 25, 15. [Google Scholar] [CrossRef]
  82. Kunzle, P.; Paris, S. Performance of large language artificial intelligence models on solving restorative dentistry and endodontics student assessments. Clin. Oral Investig. 2024, 28, 575. [Google Scholar] [CrossRef]
  83. Li, C.; Zhang, J.; Abdul-Masih, J.; Zhang, S.; Yang, J. Performance of ChatGPT and Dental Students on Concepts of Periodontal Surgery. Eur. J. Dent. Educ. 2024. [Google Scholar] [CrossRef]
  84. Molena, K.F.; Macedo, A.P.; Ijaz, A.; Carvalho, F.K.; Gallo, M.J.D.; Wanderley Garcia de Paula, E.S.F.; de Rossi, A.; Mezzomo, L.A.; Mugayar, L.R.F.; Queiroz, A.M. Assessing the Accuracy, Completeness, and Reliability of Artificial Intelligence-Generated Responses in Dentistry: A Pilot Study Evaluating the ChatGPT Model. Cureus 2024, 16, E65658. [Google Scholar] [CrossRef] [PubMed]
  85. Praveen, G.; Poornima, U.L.S.; Akkaloori, A.; Bharathi, V. ChatGPT as a Tool for Oral Health Education: A Systematic Evaluation of ChatGPT Responses to Patients’ Oral Health-related Queries. J. Nat. Sci. Med. Jul. Sep. 2024, 7, 154–157. [Google Scholar] [CrossRef]
  86. Puladi, B.; Gsaxner, C.; Kleesiek, J.; Hölzle, F.; Röhrig, R.; Egger, J. The impact and opportunities of large language models like ChatGPT in oral and maxillofacial surgery: A narrative review. Int. J. Oral Maxillofac. Surg. 2024, 53, 78–88. [Google Scholar] [CrossRef]
  87. Sabri, H.; Saleh, M.H.A.; Hazrati, P.; Merchant, K.; Misch, J.; Kumar, P.S.; Wang, H.L.; Barootchi, S. Performance of three artificial intelligence (AI)-based large language models in standardized testing; implications for AI-assisted dental education. J. Periodontal Res. 2024. [Google Scholar] [CrossRef]
  88. Kavadella, A.; Silva, M.; Kaklamanos, E.G.; Stamatopoulos, V.; Giannakopoulos, K.; Kavadella, A. Evaluation of ChatGPT’s Real-Life Implementation in Undergraduate Dental Education: Mixed Methods Study. JMIR Med. Educ. 2024, 10, 14. [Google Scholar] [CrossRef]
  89. Saravia-Rojas, M.A.; Camarena-Fonseca, A.R.; Leon-Manco, R.; Geng-Vivanco, R. Artificial intelligence: ChatGPT as a disruptive didactic strategy in dental education. J. Dent. Educ. 2024, 88, 872–876. [Google Scholar] [CrossRef]
  90. Chau, R.C.W.; Thu, K.M.; Yu, O.Y.; Hsung, R.T.C.; Lo, E.C.M.; Lam, W.Y.H. Performance of Generative Artificial Intelligence in Dental Licensing Examinations. Int. Dent. J. 2024, 74, 616–621. [Google Scholar] [CrossRef] [PubMed]
  91. Dashti, M.; Ghasemi, S.; Ghadimi, N.; Hefzi, D.; Karimian, A.; Zare, N.; Fahimipour, A.; Khurshid, Z.; Chafjiri, M.M.; Ghaedsharaf, S. Performance of ChatGPT 3.5 and 4 on U.S. dental examinations: The INBDE, ADAT, and DAT. Imaging Sci. Dent. 2024, 54, 271–275. [Google Scholar] [CrossRef] [PubMed]
  92. Farajollahi, M.; Modaberi, A. Can ChatGPT pass the “Iranian Endodontics Specialist Board” exam? Iran. Endod. J. 2023, 18, 192. [Google Scholar]
  93. Fuchs, A.; Trachsel, T.; Weiger, R.; Eggmann, F. ChatGPT’s performance in dentistry and allergyimmunology assessments: A comparative study. Swiss Dent. J. 2023, 134, 1–17. [Google Scholar] [CrossRef]
  94. Jeong, H.; Han, S.S.; Yu, Y.; Kim, S.; Jeon, K.J. How well do large language model-based chatbots perform in oral and maxillofacial radiology? Dentomaxillofac Radiol. 2024, 53, 390–395. [Google Scholar] [CrossRef]
  95. Jin, H.K.; Lee, H.E.; Kim, E. Performance of ChatGPT-3.5 and GPT-4 in national licensing examinations for medicine, pharmacy, dentistry, and nursing: A systematic review and meta-analysis. BMC Med. Educ. 2024, 24, 1013. [Google Scholar] [CrossRef]
  96. Kim, W.; Kim, B.C.; Yeom, H.G. Performance of Large Language Models on the Korean Dental Licensing Examination: A Comparative Study. Int. Dent. J. 2024, 5, 5. [Google Scholar] [CrossRef]
  97. Morishita, M.; Fukuda, H.; Muraoka, K.; Nakamura, T.; Hayashi, M.; Yoshioka, I.; Ono, K.; Awano, S. Evaluating GPT-4V’s performance in the Japanese national dental examination: A challenge explored. J. Dent. Sci. 2024, 19, 1595–1600. [Google Scholar] [CrossRef]
  98. Ohta, K.; Ohta, S. The Performance of GPT-3.5, GPT-4, and Bard on the Japanese National Dentist Examination: A Comparison Study. Cureus 2023, 15, E50369. [Google Scholar] [CrossRef]
  99. Revilla-León, M.; Barmak, B.A.; Sailer, I.; Kois, J.C.; Att, W. Performance of an Artificial Intelligence-Based Chatbot (ChatGPT) Answering the European Certification in Implant Dentistry Exam. Int. J. Prosthodont. 2024, 37, 221–224. [Google Scholar] [CrossRef]
  100. Song, E.S.; Lee, S.P. Comparative Analysis of the Response Accuracies of Large Language Models in the Korean National Dental Hygienist Examination Across Korean and English Questions. Int. J. Dent. Hyg. 2024. [Google Scholar] [CrossRef]
  101. Takagi, S.; Koda, M.; Watari, T. The Performance of ChatGPT-4V in Interpreting Images and Tables in the Japanese Medical Licensing Exam. JMIR Med. Educ. 2024, 10, E54283. [Google Scholar] [CrossRef]
  102. Abdaljaleel, M.; Barakat, M.; Alsanafi, M.; Salim, N.A.; Abazid, H.; Malaeb, D.; Mohammed, A.H.; Hassan, B.A.R.; Wayyes, A.M.; Farhan, S.S.; et al. Author Correction: A multinational study on the factors influencing university students’ attitudes and usage of ChatGPT. Sci. Rep. 2024, 14, 8281. [Google Scholar] [CrossRef]
  103. Alnaim, N.; AlSanad, D.S.; Albelali, S.; Almulhem, M.; Almuhanna, A.F.; Attar, R.W.; Alsahli, M.; Albagmi, S.; Bakhshwain, A.M.; Almazrou, S.; et al. Effectiveness of ChatGPT in remote learning environments: An empirical study with medical students in Saudi Arabia. Nutr. Health 2024, 16, 2601060241273596. [Google Scholar] [CrossRef]
  104. Kurt Demirsoy, K.; Buyuk, S.K.; Bicer, T. How reliable is the artificial intelligence product large language model ChatGPT in orthodontics? Angle Orthod. 2024, 94, 602–607. [Google Scholar] [CrossRef]
  105. Rahad, K.; Martin, K.; Amugo, I.; Ferguson, S.; Curtis, A.; Davis, A.; Gangula, P.; Wang, Q. ChatGPT to Enhance Learning in Dental Education at a Historically Black Medical College. Dent. Res. Oral. Health 2024, 7, 8–14. [Google Scholar] [CrossRef]
  106. Sallam, M.; Salim, N.A.; Barakat, M.; Al-Tammemi, A.B. ChatGPT applications in medical, dental, pharmacy, and public health education: A descriptive study highlighting the advantages and limitations. Narra J. 2023, 3, E103. [Google Scholar] [CrossRef]
  107. Ahmed, W.M.; Azhari, A.A.; Alfaraj, A.; Alhamadani, A.; Zhang, M.; Lu, C.T. The Quality of AI-Generated Dental Caries Multiple Choice Questions: A Comparative Analysis of ChatGPT and Google Bard Language Models. Heliyon 2024, 10, e28198. [Google Scholar] [CrossRef]
  108. Brondani, M.; Alves, C.; Ribeiro, C.; Braga, M.M.; Garcia, R.C.M.; Ardenghi, T.; Pattanaporn, K. Artificial intelligence, ChatGPT, and dental education: Implications for reflective assignments and qualitative research. J. Dent. Educ. 2024. [Google Scholar] [CrossRef]
  109. de Vries, T.J.; Schoenmaker, T.; Peferoen, L.A.N.; Krom, B.P.; Bloemena, E. Design and evaluation of an immunology and pathology course that is tailored to today’s dentistry students. Front. Oral. Health 2024, 5, 1386904. [Google Scholar] [CrossRef]
  110. Quah, B.; Zheng, L.; Sng, T.J.H.; Yong, C.W.; Islam, I. Reliability of ChatGPT in automated essay scoring for dental undergraduate examinations. BMC Med. Educ. 2024, 24, 962. [Google Scholar] [CrossRef]
  111. Shamim, M.S.; Zaidi, S.J.A.; Rehman, A. The Revival of Essay-Type Questions in Medical Education: Harnessing Artificial Intelligence and Machine Learning. JCPSP J. Coll. Physicians Surg. Pak. 2024, 34, 595–599. [Google Scholar] [CrossRef]
  112. Shete, A.; Shete, M.; Chavan, M.; Channe, P.; Sapkal, R.; Buva, K. Evaluation of ChatGPT as a New Assessment Tool in Dental Education. J. Indian. Acad. Oral. Med. Radiol. Jul. Sep. 2024, 36, 259–263. [Google Scholar]
  113. Uribe, S.E.; Maldupa, I.; Kavadella, A.; El Tantawi, M.; Chaurasia, A.; Fontana, M.; Marino, R.; Innes, N.; Schwendicke, F. Artificial intelligence chatbots and large language models in dental education: Worldwide survey of educators. Eur. J. Dent. Educ. 2024, 28, 865–876. [Google Scholar] [CrossRef]
  114. Hirosawa, T.; Kawamura, R.; Harada, Y.; Mizuta, K.; Tokumasu, K.; Kaji, Y.; Suzuki, T.; Shimizu, T. ChatGPT-Generated Differential Diagnosis Lists for Complex Case-Derived Clinical Vignettes: Diagnostic Accuracy Evaluation. JMIR Med. Inform. 2023, 11, E48808. [Google Scholar] [CrossRef]
  115. Alkaissi, H.; McFarlane, S.I. Artificial Hallucinations in ChatGPT: Implications in Scientific Writing. Cureus 2023, 15, E35179. [Google Scholar] [CrossRef]
  116. Biswas, S. ChatGPT and the Future of Medical Writing. Radiology 2023, 307, e223312. [Google Scholar] [CrossRef]
  117. Thorp, H.H. ChatGPT is fun, but not an author. Science 2023, 379, 313. [Google Scholar] [CrossRef]
  118. Haman, M.; Skolnik, M. Using ChatGPT to conduct a literature review. Acc. Res. 2024, 31, 1244–1246. [Google Scholar] [CrossRef]
  119. Suárez, A.; Jiménez, J.; de Pedro, M.L.; Andreu-Vázquez, C.; García, V.D.F.; Sánchez, M.G.; Freire, Y. Beyond the Scalpel: Assessing ChatGPT’s potential as an auxiliary intelligent virtual assistant in oral surgery. Comp. Struct. Biotechnol. J. 2024, 24, 46–52. [Google Scholar] [CrossRef]
  120. Stokel-Walker, C.; Van Noorden, R. What ChatGPT and generative AI mean for science. Nature 2023, 614, 214–216. [Google Scholar] [CrossRef]
  121. Mijwil, M.; Mohammad, A.; Ahmed Hussein, A. ChatGPT: Exploring the Role of Cybersecurity in the Protection of Medical Information. Mesopotamian J. CyberSecurity 2023, 2023, 18–21. [Google Scholar] [CrossRef]
  122. Hasal, M.; Nowaková, J.; Ahmed Saghair, K.; Abdulla, H.; Snášel, V.; Ogiela, L. Chatbots: Security, privacy, data protection, and social aspects. Concurr. Comput. Pract. Exp. 2021, 33, E6426. [Google Scholar] [CrossRef]
  123. Gerke, S.; Minssen, T.; Cohen, G. Chapter 12—Ethical and legal challenges of artificial intelligence-driven healthcare. In Artificial Intelligence in Healthcare; Bohr, A., Memarzadeh, K., Eds.; Academic Press: Cambridge, MA, USA, 2020; pp. 295–336. [Google Scholar]
  124. Anderson, N.; Belavy, D.L.; Perle, S.M.; Hendricks, S.; Hespanhol, L.; Verhagen, E.; Memon, A.R. AI did not write this manuscript, or did it? Can we trick the AI text detector into generated texts? The potential future of ChatGPT and AI in Sports & Exercise Medicine manuscript generation. BMJ Open Sport. Exerc. Med. 2023, 9, E001568. [Google Scholar] [CrossRef]
  125. ChatGPT Generative Pre-trained Transformer; Zhavoronkov, A. Rapamycin in the context of Pascal’s Wager: Generative pre-trained transformer perspective. Oncoscience 2022, 9, 82–84. [Google Scholar] [CrossRef] [PubMed]
  126. O’Connor, S. Open artificial intelligence platforms in nursing education: Tools for academic progress or abuse? Nurse Educ. Pract. 2023, 66, 103537. [Google Scholar] [CrossRef] [PubMed]
  127. Stokel-Walker, C. ChatGPT listed as author on research papers: Many scientists disapprove. Nature 2023, 613, 620–621. [Google Scholar] [CrossRef] [PubMed]
  128. Gomez-Cabello, C.A.; Borna, S.; Pressman, S.M.; Haider, S.A.; Forte, A.J. Large Language Models for Intraoperative Decision Support in Plastic Surgery: A Comparison between ChatGPT-4 and Gemini. Medicina 2024, 60, 957. [Google Scholar] [CrossRef]
  129. Rossettini, G.; Rodeghiero, L.; Corradi, F.; Cook, C.; Pillastrini, P.; Turolla, A.; Castellini, G.; Chiappinotto, S.; Gianola, S.; Palese, A. Comparative accuracy of ChatGPT-4, Microsoft Copilot and Google Gemini in the Italian entrance test for healthcare sciences degrees: A cross-sectional study. BMC Med. Educ. 2024, 24, 694. [Google Scholar] [CrossRef]
Figure 1. Papers selection and screening flow chart.
Figure 1. Papers selection and screening flow chart.
Applsci 14 10802 g001
Table 1. Studies included in this review.
Table 1. Studies included in this review.
TopicSubtopicAuthors (Year) [Reference]Conclusions
Administrative applications Eggmann, F., et al. (2023) [25]ChatGPT can streamline administrative workflows and aid in clinical decision support, given comprehensive, unbiased data. However, it raises privacy and cybersecurity concerns and lacks reliability and up-to-date knowledge compared to traditional search engines, especially for health queries.
Clinical applicationsDiagnostics and radiologyBaig, Z., et al. (2024) [26]AI programs’ treatment recommendations generally matched the current literature with up to 75% agreement, though data sources were often missing, except for Bard. Both GPT-4 and clinician reviews suggested procedures potentially leading to overtreatment. GPT-4 had the highest overall accuracy.
Clinical applicationsDiagnostics and radiologyIslam, A., et al. (2024) [27]The proficiency of ChatGPT in handling intricate reasoning queries within pathology demonstrated a noteworthy level of relational accuracy. Consequently, its text output created coherent links between elements, producing meaningful responses. This suggests that scholars or students can rely on this program to address reasoning-based inquiries.
Clinical applicationsDiagnostics and radiologyMohammad-Rahimi, H., et al. (2024) [28]GPT-4 excelled in providing high-quality information on controversial dental topics. However, developers should incorporate scientific citation authenticators to validate citations due to the high incidence of fabricated references.
Clinical applicationsDiagnostics and radiologyRusse, M.F.D.M., et al. (2024) [29]A content-aware chatbot using GPT-4 reliably provided recommendations according to current consensus guidelines. The responses were deemed trustworthy and transparent and therefore facilitate the integration of artificial intelligence into clinical decision-making.
Clinical applicationsDiagnostics and radiologySahu, P.K., et al. (2023) [30]A content-aware chatbot using GPT-4 reliably followed current guidelines, providing trustworthy and transparent recommendations. This supports AI’s integration into clinical decision-making.
Clinical applicationsDiagnostics and radiologyShikino, K., et al. (2024) [31]ChatGPT-4 demonstrates potential as an auxiliary tool for diagnosing typical and mildly atypical presentations of common diseases. However, its performance declines with greater atypicality.
Clinical applicationsDiagnostics and radiologySilva, T.P., et al. (2024) [32]The performance of the GPT program in describing and providing differential diagnoses for radiolucent lesions in panoramic radiographs is variable and at this stage limited in its use for clinical application.
Clinical applicationsEndodonticsMohammad-Rahimi, H., et al. (2024) [33]GPT-3.5 provided more credible information on topics related to endodontics compared to Google Bard and Bing.
Clinical applicationsEndodonticsOurang, S.A., et al. (2024) [34]The paper reviews AI concepts in endodontics, focusing on machine learning for diagnosis and computer vision for dental image interpretation. It emphasizes the need for rigorous validation and ethical transparency. AI has significant potential to enhance endodontic research, education, and patient care with interdisciplinary collaboration.
Clinical applicationsEndodonticsQutieshat, A., et al. (2024) [35]The study reveals AI’s capability to outperform dental students in diagnostic accuracy regarding endodontic assessments.
Clinical applicationsEndodonticsSnigdha, N.T., et al. (2024) [36]The results showed no statistically significant differences between the two versions, indicating comparable response accuracy.
Clinical applicationsEndodonticsSuarez, A., et al. (2024) [37]The answers generated by ChatGPT showed high consistency (85.44%). ChatGPT achieved an average accuracy of 57.33%. However, significant differences in accuracy were observed based on question difficulty, with lower accuracy for easier questions.
Clinical applicationsOral and maxillofacial surgeryAcar, A.H. (2024) [38]ChatGPT excels in answering oral surgery-related questions with superior accuracy, completeness, and clarity, making it a valuable tool for detailed information.
Clinical applicationsOral and maxillofacial surgeryAlten, A., et al. (2023) [39]ChatGPT-4 can provide valuable information and guidance during orthognathic surgery consultations, but it cannot replace direct medical consultation.
Clinical applicationsOral and maxillofacial surgeryBalel, Y. (2024) [40]Scholar GPT excelled in oral and maxillofacial surgery questions, providing more consistent and high-quality responses compared to ChatGPT. Models using academic databases offer more accurate and reliable information.
Clinical applicationsOral and maxillofacial surgeryCai, Y., et al. (2024) [41]ChatGPT/GPT-4 excelled in medical knowledge accuracy and recommendation rationality while also accurately sensing and providing reassurance about patient emotions. It can be used for patient follow-up after oral surgeries but should be supervised by healthcare professionals to consider current limitations.
Clinical applicationsOral and maxillofacial surgeryÇoban, E., and B. Altay (2024) [42]The AI platform can educate patients about dental implantology and treatment procedures, but there is concern about potential bias toward specific dental implant brands.
Clinical applicationsOral and maxillofacial surgeryIsik, G., et al. [43]The study outcomes emphasized high accuracy and quality in ChatGPT Plus’s responses except for the questions requiring a detailed response or a comment.
Clinical applicationsOral and maxillofacial surgeryJacobs, T., et al. (2024) [44]AI was able to provide mostly accurate responses, and content was closely aligned with AAOMS guidelines. However, responses were too complex for the average third molar extraction patient and were deficient in citations and references.
Clinical applicationsOral and maxillofacial surgerySuarez, A., et al. (2024) [37]Final grade accuracy was found to be 71.7%, and consistency of the experts’ grading across iterations ranged from moderate to almost perfect.
Clinical applicationsOrthodonticsAbu Arqub, S., et al. (2024) [45]The accuracy of ChatGPT’s responses was generally insufficient, often missing relevant literature citations. Additionally, its capability to provide up-to-date and precise information was limited.
Clinical applicationsOrthodonticsDaraqel, B. a. b., et al. (2024) [46]Both ChatGPT- and Google Bard-generated responses were rated with a high level of accuracy and completeness to the general orthodontic questions posed. However, acquiring answers was generally faster using the Google Bard model.
Clinical applicationsOrthodonticsDursun, D., and R. Bilici Geçer (2024) [47]All chatbot models provided generally accurate, moderately reliable, and moderate- to good-quality answers to questions the clear aligners.
Clinical applicationsOrthodonticsLima, N.G.M., et al. (2024) [48]AI improves patient communication, diagnosis support, data digitization, and treatment assistance. ChatGPT aids in care, billing, and health information access but may provide nonsensical responses and poses privacy risks.
Clinical applicationsOrthodonticsMakrygiannakis, M.A., et al. (2024) [49]LLMs hold promise for evidence-based orthodontics, but their limitations can lead to incorrect decisions if not used carefully. They cannot replace orthodontists’ critical thinking and expertise.
Clinical applicationsOrthodonticsSurovková, J., et al. (2023) [50]The paper introduces an AI-powered orthodontic workflow, highlighting new responsibilities for orthodontic nurses and assessing its use over three years with Dental Monitoring. It concludes that AI enhances dental practice with precise, personalized treatment but raises new ethical and legal issues.
Clinical applicationsPatients’ communication and self-educationBatool, I., et al. (2024) [51]Embedded GPT model showed better results as compared to ChatGPT in providing postoperative dental care emphasizing the benefits of embedding and prompt engineering.
Clinical applicationsPatients’ communication and self-educationGugnani, N., et al. (2024) [52]Overall, the responses were found to be complete and logical and in clear language, with only some inadequacies being reported in a few of the answers.
Clinical applicationsPatients’ communication and self-educationHassona, Y. (2024) [53]ChatGPT is an attractive and potentially useful resource for informing patients about early detection of oral cancer. Nevertheless, concerns do exist about readability and actionability of the offered information.
Clinical applicationsPatients’ communication and self-educationIncerti Parenti, S., et al. (2024) [54]The study suggests that while ChatGPT-3.5 can be a valuable tool for patient education, efforts to improve readability are necessary to ensure accessibility and utility for all patients.
Clinical applicationsPatients’ communication and self-educationVassis, S., et al. (2024) [55]Although patients generally prefer AI-generated information regarding the side effects of orthodontic treatment, the tested prompts fall short of providing thoroughly satisfactory and high-quality education to patients.
Clinical applicationsPatients’ communication and self-educationYurdakurban, E., et al. (2024) [56]AI-based chatbots with a variety of features have usually provided answers with high quality, reliability, and difficult readability to questions.
Clinical applicationsPediatricsRokhshad, R., et al. (2024) [57]In the pilot study, chatbots showed lower accuracy than dentists. Chatbots may not be recommended yet for clinical pediatric dentistry.
Clinical applicationsPeriodontologyAlan, R., and B.M. Alan (2023) [58]Consistently offered accurate guidance in most responses.
Clinical applicationsPeriodontologyBabayiğit, O., et al. (2023) [59]While ChatGPT may not offer absolute precision without expert supervision, it can still serve as a valuable resource for periodontologists, with some risk of inaccuracies.
Clinical applicationsPeriodontologyDanesh, A., et al. (2024) [60]While ChatGPT 4 showed a higher proficiency compared to ChatGPT 3.5, both chatbot models leave considerable room for misinformation with their responses relating to periodontology.
Clinical applicationsPeriodontologyTastan Eroglu, Z., et al. (2024) [61]The present performance of ChatGPT in the classification of periodontitis exhibited a reasonable level. However, it is expected that additional improvements would increase its effectiveness and broaden its range of functionalities.
Clinical applicationsProsthodonticsFreire, Y., et al. (2024) [62]The results show that currently, ChatGPT has limited ability to generate answers related to RDPs and tooth-supported FDPs.
Clinical applicationsSpecial needsRokhshad, R., et al. (2024) [63]Chatbots exhibit acceptable consistency in responding to questions related to special needs dentistry and better accuracy in responding to true/false questions than diagnostic questions.
Clinical applicationsTraumatologyKhan, M.K. (2024) [64]AI and its subsets have been applied in a very limited number of fields of dental traumatology. However, the findings from the literature were found favorable and promising.
Clinical applicationsTraumatologyOzden, I., et al. (2024) [65]Although ChatGPT and Google Bard are potential knowledge resources, their consistency and accuracy in responding to dental trauma queries remain limited.
Comprehensive Alhaidry, H.M., et al. (2023) [66]AI has greatly advanced dentistry, particularly in research. ChatGPT can transform dental and healthcare systems, but caution and policies are needed to mitigate hazards, and continuous monitoring is recommended due to ethical concerns and improper reference generation.
Comprehensive de Souza, L.L., et al. (2024) [67]Integrating ChatGPT in dentistry can be highly beneficial, but it is crucial to address ethical considerations, accuracy, and privacy concerns.
Comprehensive Huang, H.Y., et al. (2023) [68]While LLMs offer significant potential benefits, the challenges, such as data privacy, data quality, and model bias, need further study.
Dental research Al-Moghrabi, D., et al. (2024) [69]ChatGPT may effectively distinguish between predatory and legitimate journals, with accuracy rates of 92.5% and 71%, respectively.
Dental research Bagde, H., et al. (2023) [70]ChatGPT has the ability to provide appropriate solutions to questions in the medical and dentistry areas, but researchers and doctors should cautiously assess its responses because they might not always be dependable.
Dental research Demir, G.B., et al. (2024) [71]Both ChatGPT 3.5 and 4 can be pivotal tools for generating PICO-driven queries in orthodontics when optimally configured. However, the precision required in medical research necessitates a judicious and critical evaluation of LLM-generated outputs, advocating for a circumspect integration into scientific investigations.
Dental research Fatani, B. (2023) [72]ChatGPT can help find and summarize academic papers, generate drafts, and translate content, streamlining and simplifying academic writing. However, its use in scientific writing should be regulated and monitored due to ethical considerations.
Dental research George Pallivathukal, R., et al. (2024) [73]The study aids in creating guidelines for implementing GAI chatbots in healthcare education, emphasizing benefits and risks, and informing AI developers and educators about ChatGPT’s potential in academia.
Dental research Tiwari, A., et al. (2023) [74]Studies show ChatGPT helps in scientific and dental research but should not be solely relied on due to ethical concerns and the need for review.
Dental research Uribe, S.E., and I. Maldupa (2024) [75]GenAI can potentially increase productivity and inclusivity, but it raises concerns such as bias, inaccuracy, and distortion of academic incentives. Therefore, the findings support the need for clear AI guidelines and standards for academic publishing.
Educational enhancements Claman, D., and E. Sezgin (2024) [76]LLMs can enhance dental education by offering personalized feedback, case scenarios, and educational content. However, they also present challenges like bias, inaccuracies, privacy issues, and the risk of overreliance.
Educational enhancements Roganović, J. (2024) [77]A majority of students in the cohort were reluctant to use ChatGPT. Furthermore, familarity (reading) with ChatGPT features appears to alter the expectations and enhance learning performance of students, suggesting an AI description-related cognitive bias.
Educational enhancementsGeneral expertiseAlbagieh, H., et al. (2024) [78]No significant difference was found in response scores. However, residents showed low agreement, while LLMs showed high agreement. Dentists should leverage AI for diagnosis and treatment.
Educational enhancementsGeneral expertiseAli, K., et al. (2024) [79]Generative AI can transform virtual learning. Healthcare educators should adapt to its benefits for learners while mitigating dishonest use.
Educational enhancementsGeneral expertiseAminoshariae, A., et al. (2024) [80]AI in endodontic education will support clinical and didactic teaching through individualized feedback; enhanced, augmented, and virtually generated training aids; automated detection and diagnosis; treatment planning and decision support; and AI-based student progress evaluation and personalized education.
Educational enhancementsGeneral expertiseGiannakopoulos, K., et al. (2023) [81]Although LLMs show promise in evidence-based dentistry, their limitations can lead to harmful decisions if not used carefully. They should complement, not replace, a dentist’s critical thinking and expertise.
Educational enhancementsGeneral expertiseKunzle, P., and S. Paris (2024) [82]Overall, there are large performance differences among LLMAs. Only the ChatGPT-4 models achieved a success ratio that could be used with caution to support the dental academic curriculum.
Educational enhancementsGeneral expertiseLi, C., et al. (2024) [83]For periodontal surgery exams, ChatGPT’s accuracy was not as high as students’, but it shows potential in assisting with the curriculum and helping with clinical letters and reviews.
Educational enhancementsGeneral expertiseMolena, K.F., et al. (2024) [84]ChatGPT initially demonstrated good accuracy and completeness, which was further improved with machine learning (ML) over time. However, some inaccurate answers and references persisted.
Educational enhancementsGeneral expertisePraveen, G., et al. (2024) [85]ChatGPT generated clear, scientifically accurate and relevant, comprehensive, and consistent responses to diverse oral health-related queries despite some significant limitations.
Educational enhancementsGeneral expertisePuladi, B., et al. (2024) [86]Classic OMS diseases are underrepresented. The current literature related to LLMs in OMS has a limited evidence level.
Educational enhancementsGeneral expertiseSabri, H., et al. (2024) [87]ChatGPT-4 performed well on AAP in-service exam questions, outperforming Gemini and ChatGPT-3.5. While it shows potential as an educational tool in periodontics and oral implantology, limitations like processing image-based inquiries, inconsistent responses, and not reaching absolute accuracy must be considered.
Educational enhancementsGPT vs. literature researchKavadella, A., et al. (2024) [88]Students using ChatGPT for their learning assignments performed significantly better in the knowledge examination than their fellow students who used the literature research methodology.
Educational enhancementsGPT vs. literature researchSaravia-Rojas, M.A., et al. (2024) [89]Dental students highly valued the experience of using ChatGPT for academic tasks. Nonetheless, the traditional method of searching for scientific articles yield higher scores.
Educational enhancementsLicensing exam/mastery testsChau, R.C.W., et al. (2024) [90]The newer version of GenAI has shown good proficiency in answering multiple-choice questions from dental licensing examinations.
Educational enhancementsLicensing exam/mastery testsDashti, M., et al. (2024) [91]Both ChatGPT 3.5 and 4 effectively handled knowledge-based, case history, and comprehension questions, with ChatGPT 4 being more reliable and surpassing the performance of 3.5. ChatGPT 4’s perfect score in comprehension questions underscores its trainability in specific subjects. However, both versions exhibited weaker performance in mathematical analysis.
Educational enhancementsLicensing exam/mastery testsFarajollahi, M., and A. Modaberi (2023) [92]Out of 100 questions asked from ChatGPT, a score of 40 was obtained.
Educational enhancementsLicensing exam/mastery testsFuchs, A., et al. (2023) [93]The performance disparity between SFLEDM and EEAACI assessments highlights ChatGPT’s varying proficiency due to differences in training data. Priming can help, but healthcare use must be cautious due to inherent risks.
Educational enhancementsLicensing exam/mastery testsJeong, H., et al. (2024) [94]The performance of chatbots in oral and maxillofacial radiology was unsatisfactory. Further training using specific, relevant data derived solely from reliable sources is required.
Educational enhancementsLicensing exam/mastery testsJin, H.K., et al. (2024) [95]The accuracy levels ranged from 36 to 77% for ChatGPT-3.5 and from 64.4 to 100% for GPT-4. Additionally, in the context of health licensing examinations, the ChatGPT models exhibited greater proficiency in the following order: pharmacy, medicine, dentistry, and nursing.
Educational enhancementsLicensing exam/mastery testsKim, W., et al. (2024) [96]Using the KDLE as a benchmark, the study demonstrates that although LLMs have not yet reached human-level performance in overall scores, both Claude3-Opus and ChatGPT-4 exceed the cut-off scores and perform exceptionally well in specific subjects.
Educational enhancementsLicensing exam/mastery testsMorishita, M., et al. (2024) [97]While innovative, ChatGPT-4V’s image recognition feature exhibited limitations, especially in handling image-intensive and complex clinical practical questions, and is not yet fully suitable as an educational support tool for dental students at its current stage.
Educational enhancementsLicensing exam/mastery testsOhta, K., and Ohta, S. (2023) [98]GPT-4 achieved the highest overall score in the JNDE, followed by Bard and GPT-3.5. However, only Bard surpassed the passing score for essential questions.
Educational enhancementsLicensing exam/mastery testsRevilla-León, M., et al. (2024) [99]The AI-based chatbot tested not only passed the exam but performed better than licensed dentists.
Educational enhancementsLicensing exam/mastery testsSong, E.S., and S.P. Lee (2024) [100]GPT-4 shows great potential for medical education and standardized testing, especially in English. However, performance varies across subjects and languages, highlighting the need for diverse and localized training data to improve effectiveness.
Educational enhancementsLicensing exam/mastery testsTakagi, S., et al. (2024) [101]ChatGPT-4V successfully passed the 117th JMLE, demonstrating proficiency in handling questions, including image- and table-based questions.
Educational enhancementsThe students’ perspectiveAbdaljaleel, M., et al. (2024) [102]The study validated “TAME-ChatGPT” as a useful tool for assessing ChatGPT adoption among university students.
Educational enhancementsThe students’ perspectiveAlnaim, N., et al. (2024) [103]Despite challenges and varied perceptions based on gender and education level, the overwhelmingly positive attitudes toward ChatGPT underscore its potential as a valuable tool in medical education.
Educational enhancementsThe students’ perspectiveKurt Demirsoy, K., et al. (2024) [104]ChatGPT has significant potential in terms of usability for patient information and education in the field of orthodontics if it is developed and necessary updates are made.
Educational enhancementsThe students’ perspectiveRahad, K., et al. (2024) [105]The results showed that ChatGPT can assist in dental essay writing and generate relevant content for dental students, in addition to other benefits.
Educational enhancementsThe students’ perspectiveSallam, M., et al. (2023) [106]ChatGPT has potential in medical, dental, pharmacy, and public health education by improving personalized learning, clinical reasoning, and the understanding of complex concepts. However, concerns include data privacy, biased and inaccurate content, and risks to critical thinking and communication skills, highlighting the need for proper guidelines.
Educational enhancementsThe teachers’ perspectiveAhmed, W.M., et al. (2024) [107]ChatGPT and Bard can generate numerous questions about dental caries, especially at the knowledge and comprehension levels, making them useful for large-scale exams. However, educators need to review and adapt these questions to ensure they meet their learning objectives.
Educational enhancementsThe teachers’ perspectiveBrondani, M., et al. (2024) [108]Instructors could usually tell if reflections were generated by ChatGPT or students. However, the thematic analysis content from ChatGPT matched that of qualitative researchers.
Educational enhancementsThe teachers’ perspectivede Vries, T.J., et al. (2024) [109]These methods proved to be appropriate and logical choices for reaching the learning goals of the course.
Educational enhancementsThe teachers’ perspectiveQuah, B., et al. (2024) [110]The study shows the potential of ChatGPT for essay marking. However, an appropriate rubric design is essential for optimal reliability.
Educational enhancementsThe teachers’ perspectiveShamim, M.S., et al. (2024) [111]AI and ML technologies can potentially supplement human grading in the assessment of essays.
Educational enhancementsThe teachers’ perspectiveShete, A., et al. (2024) [112]Instead of treating artificial intelligence as a threat, dental educators need to adapt teaching and assessments in dental education for the benefit of learners while mitigating its dishonest use.
Educational enhancementsThe teachers’ perspectiveUribe, S.E., et al. (2024) [113]A positive yet cautious view towards AI chatbot integration in dental curricula is prevalent, underscoring the need for clear implementation guidelines.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Puleio, F.; Lo Giudice, G.; Bellocchio, A.M.; Boschetti, C.E.; Lo Giudice, R. Clinical, Research, and Educational Applications of ChatGPT in Dentistry: A Narrative Review. Appl. Sci. 2024, 14, 10802. https://doi.org/10.3390/app142310802

AMA Style

Puleio F, Lo Giudice G, Bellocchio AM, Boschetti CE, Lo Giudice R. Clinical, Research, and Educational Applications of ChatGPT in Dentistry: A Narrative Review. Applied Sciences. 2024; 14(23):10802. https://doi.org/10.3390/app142310802

Chicago/Turabian Style

Puleio, Francesco, Giorgio Lo Giudice, Angela Mirea Bellocchio, Ciro Emiliano Boschetti, and Roberto Lo Giudice. 2024. "Clinical, Research, and Educational Applications of ChatGPT in Dentistry: A Narrative Review" Applied Sciences 14, no. 23: 10802. https://doi.org/10.3390/app142310802

APA Style

Puleio, F., Lo Giudice, G., Bellocchio, A. M., Boschetti, C. E., & Lo Giudice, R. (2024). Clinical, Research, and Educational Applications of ChatGPT in Dentistry: A Narrative Review. Applied Sciences, 14(23), 10802. https://doi.org/10.3390/app142310802

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop