1
|
Fawaz P, El Sayegh P, Vande Vannet B. Artificial intelligence in revolutionizing orthodontic practice. World J Methodol 2025; 15:100598. [DOI: 10.5662/wjm.v15.i3.100598] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/21/2024] [Revised: 10/07/2024] [Accepted: 12/18/2024] [Indexed: 03/06/2025] Open
Abstract
This analytical research paper explores the transformative impact of artificial intelligence (AI) in orthodontics, with a focus on its objectives: Identifying current applications, evaluating benefits, addressing challenges, and projecting future developments. AI, a subset of computer science designed to simulate human intelligence, has seen rapid integration into orthodontic practice. The paper examines AI technologies such as machine learning, deep learning, natural language processing, computer vision, and robotics, which are increasingly used to analyze patient data, assist with diagnosis and treatment planning, automate routine tasks, and improve patient communication. AI systems offer precise malocclusion diagnoses, predict treatment outcomes, and customize treatment plans by leveraging dental imagery. They also streamline image analysis, improve diagnostic accuracy, and enhance patient engagement through personalized communication. The objectives include evaluating the benefits of AI in terms of efficiency, accuracy, and personalized care, while acknowledging the challenges like data quality, algorithm transparency, and practical implementation. Despite these hurdles, AI presents promising prospects in advanced imaging, predictive analytics, and clinical decision-making. In conclusion, AI holds the potential to revolutionize orthodontic practices by improving operational efficiency, diagnostic precision and patient outcomes. With collaborative efforts to overcome challenges, AI could play a pivotal role in advancing orthodontic care.
Collapse
Affiliation(s)
- Paul Fawaz
- Faculty of Dentistry, Department of Orthodontics, University Lorraine, Nancy 54000, France
| | - Patrick El Sayegh
- Faculty of Dentistry, Saint Joseph University of Beirut, Beirouth 11042020, Lebanon
| | - Bart Vande Vannet
- Faculty of Dentistry, Department of Orthodontics, University Lorraine, Nancy 54000, France
- Institut Jean Lamour, Campus Artem (403), University Lorraine, Nancy 54000, France
| |
Collapse
|
2
|
Kula B, Kula A, Bagcier F, Alyanak B. Artificial intelligence solutions for temporomandibular joint disorders: Contributions and future potential of ChatGPT. Korean J Orthod 2025; 55:131-141. [PMID: 40104855 PMCID: PMC11922634 DOI: 10.4041/kjod24.106] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2024] [Revised: 10/25/2024] [Accepted: 12/09/2024] [Indexed: 03/20/2025] Open
Abstract
Objective This study aimed to evaluate the reliability and usefulness of information generated by Chat Generative Pre-Trained Transformer (ChatGPT) on temporomandibular joint disorders (TMD). Methods We asked ChatGPT about the diseases specified in the TMD classification and scored the responses using Likert reliability and usefulness scales, the modified DISCERN (mDISCERN) scale, and the Global Quality Scale (GQS). Results The highest Likert scores for both reliability and usefulness were for masticatory muscle disorders (mean ± standard deviation [SD]: 6.0 ± 0), and the lowest scores were for inflammatory disorders of the temporomandibular joint (mean ± SD: 4.3 ± 0.6 for reliability, 4.0 ± 0 for usefulness). The median Likert reliability score indicates that the responses are highly reliable. The median Likert usefulness score was 5 (4-6), indicating that the responses were moderately useful. A comparative analysis was performed, and no statistically significant differences were found in any subject for either reliability or usefulness (P = 0.083-1.000). The median mDISCERN score was 4 (3-5) for the two raters. A statistically significant difference was observed in the mean mDISCERN scores between the two raters (P = 0.046). The GQS scores indicated a moderate to high quality (mean ± SD: 3.8 ± 0.8 for rater 1, 4.0 ± 0.5 for rater 2). No statistically significant correlation was found between mDISCERN and GQS scores (r = -0.006, P = 0.980). Conclusions Although ChatGPT-4 has significant potential, it can be used as an additional source of information regarding TMD for patients and clinicians.
Collapse
Affiliation(s)
- Betul Kula
- Department of Orthodontics, Istanbul Galata University, Istanbul, Türkiye
| | - Ahmet Kula
- Department of Prosthodontics, Uskudar University, Istanbul, Türkiye
| | - Fatih Bagcier
- Physical Medicine and Rehabilitation Clinic, Basaksehir Cam and Sakura City Hospital, Istanbul, Türkiye
| | - Bulent Alyanak
- Department of Physical Medicine and Rehabilitation, Golcuk Necati Celik State Hospital, Kocaeli, Türkiye
| |
Collapse
|
3
|
Zhou X, Chen Y, Abdulghani EA, Zhang X, Zheng W, Li Y. Performance in answering orthodontic patients' frequently asked questions: Conversational artificial intelligence versus orthodontists. J World Fed Orthod 2025:S2212-4438(25)00012-8. [PMID: 40140287 DOI: 10.1016/j.ejwf.2025.02.001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/09/2024] [Revised: 02/11/2025] [Accepted: 02/11/2025] [Indexed: 03/28/2025]
Abstract
OBJECTIVES Can conversational artificial intelligence (AI) help alleviate orthodontic patients' general doubts? This study aimed to investigate the performance of conversational AI in answering frequently asked questions (FAQs) from orthodontic patients, with comparison to orthodontists. MATERIALS AND METHODS Thirty FAQs were selected covering the pre-, during-, and postorthodontic treatment stages. Each question was respectively answered by AI (Chat Generative Pretrained Transformer [ChatGPT]-4) and two orthodontists (Ortho. A and Ortho. B), randomly drawn out of a panel. Their responses to the 30 FAQs were ranked by four raters, randomly selected from another panel of orthodontists, resulting in 120 rankings. All the participants were Chinese, and all the questions and answers were conducted in Chinese. RESULTS Among the 120 rankings, ChatGPT was ranked first in 61 instances (50.8%), second in 35 instances (29.2%), and third in 24 instances (20.0%). Furthermore, the mean rank of ChatGPT was 1.69 ± 0.79, significantly better than that of Ortho. A (2.23 ± 0.79, P < 0.001) and Ortho. B (2.08 ± 0.79, P < 0.05). No significant difference was found between the two orthodontist groups. Additionally, the Spearman correlation coefficient between the average ranking of ChatGPT and the inter-rater agreement was 0.69 (P < 0.001), indicating a strong positive correlation between the two variables. CONCLUSIONS Overall, the conversational AI ChatGPT-4 may outperform orthodontists in addressing orthodontic patients' FAQs, even in a non-English language. In addition, ChatGPT tends to perform better when responding to questions with answers widely accepted among orthodontic professionals, and vice versa.
Collapse
Affiliation(s)
- Xinlianyi Zhou
- State Key Laboratory of Oral Diseases, National Center for Stomatology, National Clinical Research Center for Oral Diseases, Department of Orthodontics, West China Hospital of Stomatology, Sichuan University, Chengdu, China
| | - Yao Chen
- State Key Laboratory of Oral Diseases, National Center for Stomatology, National Clinical Research Center for Oral Diseases, Department of Orthodontics, West China Hospital of Stomatology, Sichuan University, Chengdu, China
| | - Ehab A Abdulghani
- State Key Laboratory of Oral Diseases, National Center for Stomatology, National Clinical Research Center for Oral Diseases, Department of Orthodontics, West China Hospital of Stomatology, Sichuan University, Chengdu, China; Department of Orthodontics and Dentofacial Orthopedics, College of Dentistry, Thamar University, Dhamar, Yemen
| | - Xu Zhang
- State Key Laboratory of Oral Diseases, National Center for Stomatology, National Clinical Research Center for Oral Diseases, Department of Orthodontics, West China Hospital of Stomatology, Sichuan University, Chengdu, China
| | - Wei Zheng
- State Key Laboratory of Oral Diseases, National Center for Stomatology, National Clinical Research Center for Oral Diseases, Department of Oral and Maxillofacial Surgery, West China Hospital of Stomatology, Sichuan University, Chengdu, China.
| | - Yu Li
- State Key Laboratory of Oral Diseases, National Center for Stomatology, National Clinical Research Center for Oral Diseases, Department of Orthodontics, West China Hospital of Stomatology, Sichuan University, Chengdu, China.
| |
Collapse
|
4
|
Li C, Zhang J, Abdul-Masih J, Zhang S, Yang J. Performance of ChatGPT and Dental Students on Concepts of Periodontal Surgery. EUROPEAN JOURNAL OF DENTAL EDUCATION : OFFICIAL JOURNAL OF THE ASSOCIATION FOR DENTAL EDUCATION IN EUROPE 2025; 29:36-43. [PMID: 39446672 DOI: 10.1111/eje.13047] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/13/2024] [Revised: 08/16/2024] [Accepted: 10/03/2024] [Indexed: 10/26/2024]
Abstract
INTRODUCTION As a large language model, chat generative pretrained transformer (ChatGPT) has provided a valuable tool for various medical scenarios with its interactive dialogue-based interface. However, there is a lack of studies on ChatGPT's effectiveness in handling dental tasks. This study aimed to compare the knowledge and comprehension abilities of ChatGPT-3.5/4 with that of dental students about periodontal surgery. MATERIALS AND METHODS A total of 134 dental students participated in this study. We designed a questionnaire consisting of four questions about the inclination for ChatGPT, 25 multiple-choice, and one open-ended question. As the comparison of ChatGPT-3.5 and 4, the question about the inclination was removed, and the rest was the same. The response time of ChatGPT-3.5 and 4 as well as the comparison of ChatGPT-3.5 and 4' performances with dental students were measured. Regarding students' feedback on the open-ended question, we also compared the outcomes of ChatGPT-4' and teacher's review. RESULTS On average, ChatGPT-3.5 and 4 required 3.63 ± 1.18 s (95% confidence interval [CI], 3.14, 4.11) and 12.49 ± 7.29 s (95% CI, 9.48, 15.50), respectively (p < 0.001) for each multiple-choice question. For these 25 questions, the accuracy was 21.51 ± 2.72, 14 and 20 for students, ChatGPT-3.5 and 4, respectively. Furthermore, the outcomes of ChatGPT-4's review were consistent with that of teacher's review. CONCLUSIONS For dental examinations related to periodontal surgery, ChatGPT's accuracy was not yet comparable to that of the students. Nevertheless, ChatGPT shows promise in assisting students with the curriculum and helping practitioners with clinical letters and reviews of students' textual descriptions.
Collapse
Affiliation(s)
- Chen Li
- State Key Laboratory of Oral Diseases & National Center for Stomatology & National Clinical Research Center for Oral Diseases & Department of Periodontics, West China Hospital of Stomatology, Sichuan University, Chengdu, Sichuan, China
| | - Jinmei Zhang
- Periodontics, University of Iowa College of Dentistry, Iowa City, Iowa, USA
| | - John Abdul-Masih
- Iowa Institute of Oral Health Research, University of Iowa College of Dentistry, Iowa City, Iowa, USA
| | - Sihan Zhang
- State Key Laboratory of Oral Diseases & National Center for Stomatology & National Clinical Research Center for Oral Diseases & Department of Periodontics, West China Hospital of Stomatology, Sichuan University, Chengdu, Sichuan, China
| | - Jingmei Yang
- State Key Laboratory of Oral Diseases & National Center for Stomatology & National Clinical Research Center for Oral Diseases & Department of Periodontics, West China Hospital of Stomatology, Sichuan University, Chengdu, Sichuan, China
| |
Collapse
|
5
|
Zhang S, Chu Q, Li Y, Liu J, Wang J, Yan C, Liu W, Wang Y, Zhao C, Zhang X, Chen Y. Evaluation of large language models under different training background in Chinese medical examination: a comparative study. Front Artif Intell 2024; 7:1442975. [PMID: 39697797 PMCID: PMC11652508 DOI: 10.3389/frai.2024.1442975] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/10/2024] [Accepted: 11/06/2024] [Indexed: 12/20/2024] Open
Abstract
BackgroundRecently, Large Language Models have shown impressive potential in medical services. However, the aforementioned research primarily discusses the performance of LLMs developed in English within English-speaking medical contexts, ignoring the LLMs developed under different linguistic environments with respect to their performance in the Chinese clinical medicine field.ObjectiveThrough a comparative analysis of three LLMs developed under different training background, we firstly evaluate their potential as medical service tools in a Chinese language environment. Furthermore, we also point out the limitations in the application of Chinese medical practice.MethodUtilizing the APIs provided by three LLMs, we conducted an automated assessment of their performance in the 2023 CMLE. We not only examined the accuracy of three LLMs across various question, but also categorized the reasons for their errors. Furthermore, we performed repetitive experiments on selected questions to evaluate the stability of the outputs generated by the LLMs.ResultThe accuracy of GPT-4, ERNIE Bot, and DISC-MedLLM in CMLE are 65.2, 61.7, and 25.3%. In error types, the knowledge errors of GPT-4 and ERNIE Bot account for 52.2 and 51.7%, while hallucinatory errors account for 36.4 and 52.6%. In the Chinese text generation experiment, the general LLMs demonstrated high natural language understanding ability and was able to generate clear and standardized Chinese texts. In repetitive experiments, the LLMs showed a certain output stability of 70%, but there were still cases of inconsistent output results.ConclusionGeneral LLMs, represented by GPT-4 and ERNIE Bot, demonstrate the capability to meet the standards of the CMLE. Despite being developed and trained in different linguistic contexts, they exhibit excellence in understanding Chinese natural language and Chinese clinical knowledge, highlighting their broad potential application in Chinese medical practice. However, these models still show deficiencies in mastering specialized knowledge, addressing ethical issues, and maintaining the outputs stability. Additionally, there is a tendency to avoid risk when providing medical advice.
Collapse
Affiliation(s)
- Siwen Zhang
- School of Medical Device, Shenyang Pharmaceutical University, Shenyang, China
| | - Qi Chu
- Department of Clinical Laboratory, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences, Beijing, China
| | - Yujun Li
- School of Pharmacy, Shenyang Pharmaceutical University, Shenyang, China
| | - Jialu Liu
- School of Pharmacy, Shenyang Pharmaceutical University, Shenyang, China
| | - Jiayi Wang
- School of Pharmacy, Shenyang Pharmaceutical University, Shenyang, China
| | - Chi Yan
- School of Pharmacy, Shenyang Pharmaceutical University, Shenyang, China
| | - Wenxi Liu
- School of Pharmaceutical Engineering, Shenyang Pharmaceutical University, Shenyang, China
| | - Yizhen Wang
- School of Pharmaceutical Engineering, Shenyang Pharmaceutical University, Shenyang, China
| | - Chengcheng Zhao
- School of Pharmacy, Shenyang Pharmaceutical University, Shenyang, China
| | - Xinyue Zhang
- School of Pharmacy, Shenyang Pharmaceutical University, Shenyang, China
| | - Yuwen Chen
- School of Business Administration, Shenyang Pharmaceutical University, Shenyang, China
| |
Collapse
|
6
|
Zheng J, Ding X, Pu JJ, Chung SM, Ai QYH, Hung KF, Shan Z. Unlocking the Potentials of Large Language Models in Orthodontics: A Scoping Review. Bioengineering (Basel) 2024; 11:1145. [PMID: 39593805 PMCID: PMC11591942 DOI: 10.3390/bioengineering11111145] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2024] [Revised: 10/30/2024] [Accepted: 11/11/2024] [Indexed: 11/28/2024] Open
Abstract
(1) Background: In recent years, large language models (LLMs) such as ChatGPT have gained significant attention in various fields, including dentistry. This scoping review aims to examine the current applications and explore potential uses of LLMs in the orthodontic domain, shedding light on how they might improve dental healthcare. (2) Methods: We carried out a comprehensive search in five electronic databases, namely PubMed, Scopus, Embase, ProQuest and Web of Science. Two authors independently screened articles and performed data extraction according to the eligibility criteria, following the PRISMA-ScR guideline. The main findings from the included articles were synthesized and analyzed in a narrative way. (3) Results: A total of 706 articles were searched, and 12 papers were eventually included. The applications of LLMs include improving diagnostic and treatment efficiency in orthodontics as well as enhancing communication with patients. (4) Conclusions: There is emerging research in countries worldwide on the use of LLMs in orthodontics, suggesting an upward trend in their acceptance within this field. However, the potential application of LLMs remains in its early stage, with a noticeable lack of extensive studies and tailored products to address specific clinical needs.
Collapse
Affiliation(s)
- Jie Zheng
- Department of Biomedical Sciences, City University of Hong Kong, Hong Kong, China;
| | - Xiaoqian Ding
- Division of Paediatric Dentistry and Orthodontics, Faculty of Dentistry, The University of Hong Kong, Hong Kong, China; (X.D.); (S.M.C.)
| | - Jingya Jane Pu
- Division of Oral and Maxillofacial Surgery, Faculty of Dentistry, The University of Hong Kong, Hong Kong, China;
| | - Sze Man Chung
- Division of Paediatric Dentistry and Orthodontics, Faculty of Dentistry, The University of Hong Kong, Hong Kong, China; (X.D.); (S.M.C.)
| | - Qi Yong H. Ai
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong, China;
| | - Kuo Feng Hung
- Applied Oral Science & Community Dental Care, Faculty of Dentistry, The University of Hong Kong, Hong Kong, China
| | - Zhiyi Shan
- Division of Paediatric Dentistry and Orthodontics, Faculty of Dentistry, The University of Hong Kong, Hong Kong, China; (X.D.); (S.M.C.)
| |
Collapse
|
7
|
Mohammed MH, Omer ZQ, Aziz BB, Abdulkareem JF, Mahmood TMA, Kareem FA, Mohammad DN. Convolutional Neural Network-Based Deep Learning Methods for Skeletal Growth Prediction in Dental Patients. J Imaging 2024; 10:278. [PMID: 39590742 PMCID: PMC11595330 DOI: 10.3390/jimaging10110278] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2024] [Revised: 10/27/2024] [Accepted: 10/31/2024] [Indexed: 11/28/2024] Open
Abstract
This study aimed to predict the skeletal growth maturation using convolutional neural network-based deep learning methods using cervical vertebral maturation and the lower 2nd molar calcification level so that skeletal maturation can be detected from orthopantomography using multiclass classification. About 1200 cephalometric radiographs and 1200 OPGs were selected from patients seeking treatment in dental centers. The level of skeletal maturation was detected by CNN using the multiclass classification method, and each image was identified as a cervical vertebral maturation index (CVMI); meanwhile, the chronological age was estimated from the level of the 2nd molar calcification. The model's final result demonstrates a high degree of accuracy with which each stage and gender can be predicted. Cervical vertebral maturation reported high accuracy in males (98%), while females showed high accuracy of 2nd molar calcification. CNN multiclass classification is an accurate method to detect the level of maturation, whether from cervical maturation or the calcification of the lower 2nd molar, and the calcification level of the lower 2nd molar is a reliable method to trust in the growth level, so the traditional OPG is enough for this purpose.
Collapse
Affiliation(s)
- Miran Hikmat Mohammed
- Department of Basic Sciences, College of Dentistry, University of Sulaimani, Sulaimaniyah 46001, Iraq;
| | - Zana Qadir Omer
- Department of POP, College of Dentistry, Hawler Medical University, Erbil 44001, Iraq;
| | - Barham Bahroz Aziz
- Department of Prosthodontics, College of Dentistry, University of Sulaimani, Sulaimaniyah 46001, Iraq; (B.B.A.); (J.F.A.)
| | - Jwan Fateh Abdulkareem
- Department of Prosthodontics, College of Dentistry, University of Sulaimani, Sulaimaniyah 46001, Iraq; (B.B.A.); (J.F.A.)
| | | | - Fadil Abdullah Kareem
- Department of Pedodontics and Community Oral Health, College of Dentistry, University of Sulaimani, Sulaimaniyah 46001, Iraq
| | - Dena Nadhim Mohammad
- Department of Oral Diagnosis, College of Dentistry, University of Sulaimani, Sulaimaniyah 46001, Iraq;
| |
Collapse
|
8
|
Uribe SE, Maldupa I, Kavadella A, El Tantawi M, Chaurasia A, Fontana M, Marino R, Innes N, Schwendicke F. Artificial intelligence chatbots and large language models in dental education: Worldwide survey of educators. EUROPEAN JOURNAL OF DENTAL EDUCATION : OFFICIAL JOURNAL OF THE ASSOCIATION FOR DENTAL EDUCATION IN EUROPE 2024; 28:865-876. [PMID: 38586899 DOI: 10.1111/eje.13009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/25/2023] [Revised: 02/15/2024] [Accepted: 03/18/2024] [Indexed: 04/09/2024]
Abstract
INTRODUCTION Interest is growing in the potential of artificial intelligence (AI) chatbots and large language models like OpenAI's ChatGPT and Google's Gemini, particularly in dental education. To explore dental educators' perceptions of AI chatbots and large language models, specifically their potential benefits and challenges for dental education. MATERIALS AND METHODS A global cross-sectional survey was conducted in May-June 2023 using a 31-item online-questionnaire to assess dental educators' perceptions of AI chatbots like ChatGPT and their influence on dental education. Dental educators, representing diverse backgrounds, were asked about their use of AI, its perceived impact, barriers to using chatbots, and the future role of AI in this field. RESULTS 428 dental educators (survey views = 1516; response rate = 28%) with a median [25/75th percentiles] age of 45 [37, 56] and 16 [8, 25] years of experience participated, with the majority from the Americas (54%), followed by Europe (26%) and Asia (10%). Thirty-one percent of respondents already use AI tools, with 64% recognising their potential in dental education. Perception of AI's potential impact on dental education varied by region, with Africa (4[4-5]), Asia (4[4-5]), and the Americas (4[3-5]) perceiving more potential than Europe (3[3-4]). Educators stated that AI chatbots could enhance knowledge acquisition (74.3%), research (68.5%), and clinical decision-making (63.6%) but expressed concern about AI's potential to reduce human interaction (53.9%). Dental educators' chief concerns centred around the absence of clear guidelines and training for using AI chatbots. CONCLUSION A positive yet cautious view towards AI chatbot integration in dental curricula is prevalent, underscoring the need for clear implementation guidelines.
Collapse
Affiliation(s)
- Sergio E Uribe
- Department of Conservative Dentistry and Oral Health, Riga Stradins University, Riga, Latvia
- Faculty of Dentistry, Universidad de Valparaiso, Valparaíso, Chile
- Baltic Biomaterials Centre of Excellence, Headquarters at Riga Technical University, Riga, Latvia
- ITU/WHO Focus Group AI on Health, Topic Group Dental, Geneva, Switzerland
| | - Ilze Maldupa
- Department of Conservative Dentistry and Oral Health, Riga Stradins University, Riga, Latvia
| | - Argyro Kavadella
- School of Dentistry, European University Cyprus, Nicosia, Cyprus
| | - Maha El Tantawi
- Faculty of Dentistry, Alexandria University, Alexandria, Egypt
| | - Akhilanand Chaurasia
- ITU/WHO Focus Group AI on Health, Topic Group Dental, Geneva, Switzerland
- Department of Oral Medicine & Radiology, King George's Medical University, Lucknow, Uttar Pradesh, India
| | - Margherita Fontana
- Department of Cariology, Restorative Sciences and Endodontics, School of Dentistry, University of Michigan, Ann Arbor, Michigan, USA
| | - Rodrigo Marino
- Melbourne Dental School, The University of Melbourne, Melbourne, Victoria, Australia
| | - Nicola Innes
- School of Dentistry, College of Biomedical & Life Sciences, Cardiff University, Cardiff, UK
| | - Falk Schwendicke
- ITU/WHO Focus Group AI on Health, Topic Group Dental, Geneva, Switzerland
- Department of Conservative Dentistry and Periodontology, Ludwig-Maximilians-University Munich, Munich, Germany
| |
Collapse
|
9
|
Surendran A, Daigavane P, Shrivastav S, Kamble R, Sanchla AD, Bharti L, Shinde M. The Future of Orthodontics: Deep Learning Technologies. Cureus 2024; 16:e62045. [PMID: 38989357 PMCID: PMC11234326 DOI: 10.7759/cureus.62045] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2024] [Accepted: 06/09/2024] [Indexed: 07/12/2024] Open
Abstract
Deep learning has emerged as a revolutionary technical advancement in modern orthodontics, offering novel methods for diagnosis, treatment planning, and outcome prediction. Over the past 25 years, the field of dentistry has widely adopted information technology (IT), resulting in several benefits, including decreased expenses, increased efficiency, decreased need for human expertise, and reduced errors. The transition from preset rules to learning from real-world examples, particularly machine learning (ML) and artificial intelligence (AI), has greatly benefited the organization, analysis, and storage of medical data. Deep learning, a type of AI, enables robots to mimic human neural networks, allowing them to learn and make decisions independently without the need for explicit programming. Its ability to automate cephalometric analysis and enhance diagnosis through 3D imaging has revolutionized orthodontic operations. Deep learning models have the potential to significantly improve treatment outcomes and reduce human errors by accurately identifying anatomical characteristics on radiographs, thereby expediting analytical processes. Additionally, the use of 3D imaging technologies such as cone-beam computed tomography (CBCT) can facilitate precise treatment planning, allowing for comprehensive examinations of craniofacial architecture, tooth movements, and airway dimensions. In today's era of personalized medicine, deep learning's ability to customize treatments for individual patients has propelled the field of orthodontics forward tremendously. However, it is essential to address issues related to data privacy, model interpretability, and ethical considerations before orthodontic practices can use deep learning in an ethical and responsible manner. Modern orthodontics is evolving, thanks to the ability of deep learning to deliver more accurate, effective, and personalized orthodontic treatments, improving patient care as technology develops.
Collapse
Affiliation(s)
- Aathira Surendran
- Department of Orthodontics & Dentofacial Orthopaedics, Sharad Pawar Dental College & Hospital, Wardha, IND
| | - Pallavi Daigavane
- Department of Orthodontics & Dentofacial Orthopaedics, Sharad Pawar Dental College & Hospital, Wardha, IND
| | - Sunita Shrivastav
- Department of Orthodontics & Dentofacial Orthopaedics, Sharad Pawar Dental College & Hospital, Wardha, IND
| | - Ranjit Kamble
- Department of Orthodontics & Dentofacial Orthopaedics, Sharad Pawar Dental College & Hospital, Wardha, IND
| | - Abhishek D Sanchla
- Department of Orthodontics & Dentofacial Orthopaedics, Sharad Pawar Dental College & Hospital, Wardha, IND
| | - Lovely Bharti
- Department of Orthodontics & Dentofacial Orthopaedics, Sharad Pawar Dental College & Hospital, Wardha, IND
| | - Mrudula Shinde
- Department of Orthodontics & Dentofacial Orthopaedics, Sharad Pawar Dental College & Hospital, Wardha, IND
| |
Collapse
|
10
|
Kılınç DD, Mansız D. Examination of the reliability and readability of Chatbot Generative Pretrained Transformer's (ChatGPT) responses to questions about orthodontics and the evolution of these responses in an updated version. Am J Orthod Dentofacial Orthop 2024; 165:546-555. [PMID: 38300168 DOI: 10.1016/j.ajodo.2023.11.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2023] [Revised: 11/01/2023] [Accepted: 11/01/2023] [Indexed: 02/02/2024]
Abstract
INTRODUCTION This study aimed to assess the reliability and readability of Chatbot Generative Pretrained Transformer (ChatGPT) responses to questions about orthodontics and the evolution of these responses in an updated version. METHODS Frequently asked questions about orthodontics by laypeople on Web sites were determined using the Google Search Tool. These questions were asked to both ChatGPT's March 23 version and May 24 version on April 20, 2023, and July 12, 2023, respectively. Responses were assessed for readability and reliability using the Flesch-Kincaid and DISCERN tests. RESULTS The mean DISCERN value for general questions was 2.96 ± 0.05, 3.04 ± 0.06, 2.38 ± 0.27, and 2.82 ± 0.31 for treatment-related questions; the mean Flesch-Kincaid Reading Ease score for general questions was 29.28 ± 8.22, 25.12 ± 7.39, 47.67 ± 10.77, and 41.60 ± 9.54 for treatment-related questions; mean Flesch-Kincaid Grade Level for general questions was 14.52 ± 1.48 and 14.04 ± 1.25 and 11.90 ± 2.08 and 11.41 ± 1.88 for treatment-related questions; in first and second evaluations respectively (P = 0.001). CONCLUSIONS In the second evaluation, the reliability of the answers given to general questions and treatment-related questions increased. However, in both evaluations, the reliability of the answers was found to be moderate according to the DISCERN tool. On the second evaluation, Flesch Reading Ease Scores for both general questions and treatment-related questions decreased, meaning that the readability of the new response texts became more difficult. Flesch-Kincaid Grade Level results were found at the college graduate level in the first and second evaluations for general questions and at the high school level in the first and second evaluations for treatment-related questions.
Collapse
Affiliation(s)
- Delal Dara Kılınç
- Department of Orthodontics, School of Dental Medicine, Bahçeşehir University, Istanbul, Turkey.
| | - Duygu Mansız
- Department of Orthodontics, Faculty of Dentistry, Istanbul Aydin University, Istanbul, Turkey
| |
Collapse
|
11
|
Beaulieu-Jones BR, Berrigan MT, Shah S, Marwaha JS, Lai SL, Brat GA. Evaluating capabilities of large language models: Performance of GPT-4 on surgical knowledge assessments. Surgery 2024; 175:936-942. [PMID: 38246839 PMCID: PMC10947829 DOI: 10.1016/j.surg.2023.12.014] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2023] [Revised: 12/09/2023] [Accepted: 12/15/2023] [Indexed: 01/23/2024]
Abstract
BACKGROUND Artificial intelligence has the potential to dramatically alter health care by enhancing how we diagnose and treat disease. One promising artificial intelligence model is ChatGPT, a general-purpose large language model trained by OpenAI. ChatGPT has shown human-level performance on several professional and academic benchmarks. We sought to evaluate its performance on surgical knowledge questions and assess the stability of this performance on repeat queries. METHODS We evaluated the performance of ChatGPT-4 on questions from the Surgical Council on Resident Education question bank and a second commonly used surgical knowledge assessment, referred to as Data-B. Questions were entered in 2 formats: open-ended and multiple-choice. ChatGPT outputs were assessed for accuracy and insights by surgeon evaluators. We categorized reasons for model errors and the stability of performance on repeat queries. RESULTS A total of 167 Surgical Council on Resident Education and 112 Data-B questions were presented to the ChatGPT interface. ChatGPT correctly answered 71.3% and 67.9% of multiple choice and 47.9% and 66.1% of open-ended questions for Surgical Council on Resident Education and Data-B, respectively. For both open-ended and multiple-choice questions, approximately two-thirds of ChatGPT responses contained nonobvious insights. Common reasons for incorrect responses included inaccurate information in a complex question (n = 16, 36.4%), inaccurate information in a fact-based question (n = 11, 25.0%), and accurate information with circumstantial discrepancy (n = 6, 13.6%). Upon repeat query, the answer selected by ChatGPT varied for 36.4% of questions answered incorrectly on the first query; the response accuracy changed for 6/16 (37.5%) questions. CONCLUSION Consistent with findings in other academic and professional domains, we demonstrate near or above human-level performance of ChatGPT on surgical knowledge questions from 2 widely used question banks. ChatGPT performed better on multiple-choice than open-ended questions, prompting questions regarding its potential for clinical application. Unique to this study, we demonstrate inconsistency in ChatGPT responses on repeat queries. This finding warrants future consideration including efforts at training large language models to provide the safe and consistent responses required for clinical application. Despite near or above human-level performance on question banks and given these observations, it is unclear whether large language models such as ChatGPT are able to safely assist clinicians in providing care.
Collapse
Affiliation(s)
- Brendin R Beaulieu-Jones
- Department of Surgery, Beth Israel Deaconess Medical Center, Boston, MA; Department of Biomedical Informatics, Harvard Medical School, Boston, MA. https://twitter.com/bratogram
| | | | - Sahaj Shah
- Geisinger Commonwealth School of Medicine, Scranton, PA
| | - Jayson S Marwaha
- Division of Colorectal Surgery, National Taiwan University Hospital, Taipei, Taiwan
| | - Shuo-Lun Lai
- Division of Colorectal Surgery, National Taiwan University Hospital, Taipei, Taiwan
| | - Gabriel A Brat
- Department of Surgery, Beth Israel Deaconess Medical Center, Boston, MA; Department of Biomedical Informatics, Harvard Medical School, Boston, MA.
| |
Collapse
|
12
|
Cil G, Dogan K. The efficacy of artificial intelligence in urology: a detailed analysis of kidney stone-related queries. World J Urol 2024; 42:158. [PMID: 38483582 PMCID: PMC10940482 DOI: 10.1007/s00345-024-04847-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2023] [Accepted: 01/24/2024] [Indexed: 03/17/2024] Open
Abstract
PURPOSE The study aimed to assess the efficacy of OpenAI's advanced AI model, ChatGPT, in diagnosing urological conditions, focusing on kidney stones. MATERIALS AND METHODS A set of 90 structured questions, compliant with EAU Guidelines 2023, was curated by seasoned urologists for this investigation. We evaluated ChatGPT's performance based on the accuracy and completeness of its responses to two types of questions [binary (true/false) and descriptive (multiple-choice)], stratified into difficulty levels: easy, moderate, and complex. Furthermore, we analyzed the model's learning and adaptability capacity by reassessing the initially incorrect responses after a 2 week interval. RESULTS The model demonstrated commendable accuracy, correctly answering 80% of binary questions (n:45) and 93.3% of descriptive questions (n:45). The model's performance showed no significant variation across different question difficulty levels, with p-values of 0.548 for accuracy and 0.417 for completeness, respectively. Upon reassessment of initially 12 incorrect responses (9 binary to 3 descriptive) after two weeks, ChatGPT's accuracy showed substantial improvement. The mean accuracy score significantly increased from 1.58 ± 0.51 to 2.83 ± 0.93 (p = 0.004), underlining the model's ability to learn and adapt over time. CONCLUSION These findings highlight the potential of ChatGPT in urological diagnostics, but also underscore areas requiring enhancement, especially in the completeness of responses to complex queries. The study endorses AI's incorporation into healthcare, while advocating for prudence and professional supervision in its application.
Collapse
Affiliation(s)
- Gökhan Cil
- Department of Urology, Bagcilar Training and Research Hospital, University of Health Sciences, Istanbul, Turkey.
| | - Kazim Dogan
- Department of Urology, Faculty of Medicine, Istinye University, Istanbul, Turkey
| |
Collapse
|
13
|
Li J, Dada A, Puladi B, Kleesiek J, Egger J. ChatGPT in healthcare: A taxonomy and systematic review. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 245:108013. [PMID: 38262126 DOI: 10.1016/j.cmpb.2024.108013] [Citation(s) in RCA: 61] [Impact Index Per Article: 61.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/07/2023] [Revised: 12/29/2023] [Accepted: 01/08/2024] [Indexed: 01/25/2024]
Abstract
The recent release of ChatGPT, a chat bot research project/product of natural language processing (NLP) by OpenAI, stirs up a sensation among both the general public and medical professionals, amassing a phenomenally large user base in a short time. This is a typical example of the 'productization' of cutting-edge technologies, which allows the general public without a technical background to gain firsthand experience in artificial intelligence (AI), similar to the AI hype created by AlphaGo (DeepMind Technologies, UK) and self-driving cars (Google, Tesla, etc.). However, it is crucial, especially for healthcare researchers, to remain prudent amidst the hype. This work provides a systematic review of existing publications on the use of ChatGPT in healthcare, elucidating the 'status quo' of ChatGPT in medical applications, for general readers, healthcare professionals as well as NLP scientists. The large biomedical literature database PubMed is used to retrieve published works on this topic using the keyword 'ChatGPT'. An inclusion criterion and a taxonomy are further proposed to filter the search results and categorize the selected publications, respectively. It is found through the review that the current release of ChatGPT has achieved only moderate or 'passing' performance in a variety of tests, and is unreliable for actual clinical deployment, since it is not intended for clinical applications by design. We conclude that specialized NLP models trained on (bio)medical datasets still represent the right direction to pursue for critical clinical applications.
Collapse
Affiliation(s)
- Jianning Li
- Institute for Artificial Intelligence in Medicine, University Hospital Essen (AöR), Girardetstraße 2, 45131 Essen, Germany
| | - Amin Dada
- Institute for Artificial Intelligence in Medicine, University Hospital Essen (AöR), Girardetstraße 2, 45131 Essen, Germany
| | - Behrus Puladi
- Institute of Medical Informatics, University Hospital RWTH Aachen, Pauwelsstraße 30, 52074 Aachen, Germany; Department of Oral and Maxillofacial Surgery, University Hospital RWTH Aachen, Pauwelsstraße 30, 52074 Aachen, Germany
| | - Jens Kleesiek
- Institute for Artificial Intelligence in Medicine, University Hospital Essen (AöR), Girardetstraße 2, 45131 Essen, Germany; TU Dortmund University, Department of Physics, Otto-Hahn-Straße 4, 44227 Dortmund, Germany
| | - Jan Egger
- Institute for Artificial Intelligence in Medicine, University Hospital Essen (AöR), Girardetstraße 2, 45131 Essen, Germany; Center for Virtual and Extended Reality in Medicine (ZvRM), University Hospital Essen, University Medicine Essen, Hufelandstraße 55, 45147 Essen, Germany.
| |
Collapse
|
14
|
Abi-Rafeh J, Xu HH, Kazan R, Tevlin R, Furnas H. Large Language Models and Artificial Intelligence: A Primer for Plastic Surgeons on the Demonstrated and Potential Applications, Promises, and Limitations of ChatGPT. Aesthet Surg J 2024; 44:329-343. [PMID: 37562022 DOI: 10.1093/asj/sjad260] [Citation(s) in RCA: 15] [Impact Index Per Article: 15.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2023] [Revised: 08/02/2023] [Accepted: 08/04/2023] [Indexed: 08/12/2023] Open
Abstract
BACKGROUND The rapidly evolving field of artificial intelligence (AI) holds great potential for plastic surgeons. ChatGPT, a recently released AI large language model (LLM), promises applications across many disciplines, including healthcare. OBJECTIVES The aim of this article was to provide a primer for plastic surgeons on AI, LLM, and ChatGPT, including an analysis of current demonstrated and proposed clinical applications. METHODS A systematic review was performed identifying medical and surgical literature on ChatGPT's proposed clinical applications. Variables assessed included applications investigated, command tasks provided, user input information, AI-emulated human skills, output validation, and reported limitations. RESULTS The analysis included 175 articles reporting on 13 plastic surgery applications and 116 additional clinical applications, categorized by field and purpose. Thirty-four applications within plastic surgery are thus proposed, with relevance to different target audiences, including attending plastic surgeons (n = 17, 50%), trainees/educators (n = 8, 24.0%), researchers/scholars (n = 7, 21%), and patients (n = 2, 6%). The 15 identified limitations of ChatGPT were categorized by training data, algorithm, and ethical considerations. CONCLUSIONS Widespread use of ChatGPT in plastic surgery will depend on rigorous research of proposed applications to validate performance and address limitations. This systemic review aims to guide research, development, and regulation to safely adopt AI in plastic surgery.
Collapse
|
15
|
Hatia A, Doldo T, Parrini S, Chisci E, Cipriani L, Montagna L, Lagana G, Guenza G, Agosta E, Vinjolli F, Hoxha M, D’Amelio C, Favaretto N, Chisci G. Accuracy and Completeness of ChatGPT-Generated Information on Interceptive Orthodontics: A Multicenter Collaborative Study. J Clin Med 2024; 13:735. [PMID: 38337430 PMCID: PMC10856539 DOI: 10.3390/jcm13030735] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2024] [Revised: 01/21/2024] [Accepted: 01/25/2024] [Indexed: 02/12/2024] Open
Abstract
Background: this study aims to investigate the accuracy and completeness of ChatGPT in answering questions and solving clinical scenarios of interceptive orthodontics. Materials and Methods: ten specialized orthodontists from ten Italian postgraduate orthodontics schools developed 21 clinical open-ended questions encompassing all of the subspecialities of interceptive orthodontics and 7 comprehensive clinical cases. Questions and scenarios were inputted into ChatGPT4, and the resulting answers were evaluated by the researchers using predefined accuracy (range 1-6) and completeness (range 1-3) Likert scales. Results: For the open-ended questions, the overall median score was 4.9/6 for the accuracy and 2.4/3 for completeness. In addition, the reviewers rated the accuracy of open-ended answers as entirely correct (score 6 on Likert scale) in 40.5% of cases and completeness as entirely correct (score 3 n Likert scale) in 50.5% of cases. As for the clinical cases, the overall median score was 4.9/6 for accuracy and 2.5/3 for completeness. Overall, the reviewers rated the accuracy of clinical case answers as entirely correct in 46% of cases and the completeness of clinical case answers as entirely correct in 54.3% of cases. Conclusions: The results showed a high level of accuracy and completeness in AI responses and a great ability to solve difficult clinical cases, but the answers were not 100% accurate and complete. ChatGPT is not yet sophisticated enough to replace the intellectual work of human beings.
Collapse
Affiliation(s)
- Arjeta Hatia
- Orthodontics Postgraduate School, Department of Medical Biotechnologies, University of Siena, 53100 Siena, Italy; (T.D.); (L.C.)
| | - Tiziana Doldo
- Orthodontics Postgraduate School, Department of Medical Biotechnologies, University of Siena, 53100 Siena, Italy; (T.D.); (L.C.)
| | - Stefano Parrini
- Oral Surgery Postgraduate School, Department of Medical Biotechnologies, University of Siena, 53100 Siena, Italy;
| | - Elettra Chisci
- Orthodontics Postgraduate School, University of Ferrara, 44121 Ferrara, Italy
| | - Linda Cipriani
- Orthodontics Postgraduate School, Department of Medical Biotechnologies, University of Siena, 53100 Siena, Italy; (T.D.); (L.C.)
| | - Livia Montagna
- Orthodontics Postgraduate School, University of Cagliari, 09121 Cagliari, Italy;
| | - Giuseppina Lagana
- Orthodontics Postgraduate School, “Sapienza” University of Rome, 00185 Rome, Italy;
| | - Guia Guenza
- Orthodontics Postgraduate School, University of Milano, 20019 Milan, Italy
| | - Edoardo Agosta
- Orthodontics Postgraduate School, University of Torino, 10024 Turin, Italy
| | - Franceska Vinjolli
- Orthodontics Postgraduate School, University of Roma Tor Vergata, 00133 Rome, Italy;
| | - Meladiona Hoxha
- Orthodontics Postgraduate School, “Cattolica” University of Rome, 00168 Rome, Italy;
| | - Claudio D’Amelio
- Orthodontics Postgraduate School, University of Chieti, 66100 Chieti, Italy;
| | - Nicolò Favaretto
- Orthodontics Postgraduate School, University of Trieste, 34100 Trieste, Italy
| | - Glauco Chisci
- Oral Surgery Postgraduate School, Department of Medical Biotechnologies, University of Siena, 53100 Siena, Italy;
| |
Collapse
|
16
|
Rahad K, Martin K, Amugo I, Ferguson S, Curtis A, Davis A, Gangula P, Wang Q. ChatGPT to Enhance Learning in Dental Education at a Historically Black Medical College. DENTAL RESEARCH AND ORAL HEALTH 2024; 7:8-14. [PMID: 38404561 PMCID: PMC10887427 DOI: 10.26502/droh.0069] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 02/27/2024]
Abstract
The recent rise of powerful large language model (LLM)-based AI tools, exemplified by ChatGPT and Bard, poses a great challenge to contemporary dental education. It simultaneously offers a unique resource that potentially complements today's teaching and learning, where existing widely available learning resources have often fallen short. Although the LLM tools will shape both the clinical and educational aspects of dentistry profoundly, the didactic curricula, which primarily rely on lecture-based courses where instructors impart knowledge through presentations and discussions, need to be upgraded urgently. In this paper, we used dental course materials, syllabi, and textbooks adopted currently in the School of Dentistry (SOD) at Meharry Medical College to assess the potential utility and effectiveness of ChatGPT in dental education. We collected the responses of the chatbot to questions as well as students' interactions with it for assessment. Our results showed that ChatGPT can assist in dental essay writing and generate relevant content for dental students, in addition to other benefits. The limitations of ChatGPT were also discussed in the paper.
Collapse
Affiliation(s)
- Khandoker Rahad
- Department of Computer Science & Data Science, School of Applied Computational Sciences, Meharry Medical College, Nashville, TN, USA
| | - Kianna Martin
- Department of ODS & Research, School of Dentistry, Meharry Medical College, Nashville, TN, USA
| | - Ihunna Amugo
- Department of ODS & Research, School of Dentistry, Meharry Medical College, Nashville, TN, USA
| | - Shania Ferguson
- Department of ODS & Research, School of Dentistry, Meharry Medical College, Nashville, TN, USA
| | - Angela Curtis
- Department of ODS & Research, School of Dentistry, Meharry Medical College, Nashville, TN, USA
| | - Anniya Davis
- Department of ODS & Research, School of Dentistry, Meharry Medical College, Nashville, TN, USA
| | - Pandu Gangula
- Department of ODS & Research, School of Dentistry, Meharry Medical College, Nashville, TN, USA
| | - Qingguo Wang
- Department of Computer Science & Data Science, School of Applied Computational Sciences, Meharry Medical College, Nashville, TN, USA
| |
Collapse
|
17
|
Kazimierczak N, Kazimierczak W, Serafin Z, Nowicki P, Nożewski J, Janiszewska-Olszowska J. AI in Orthodontics: Revolutionizing Diagnostics and Treatment Planning-A Comprehensive Review. J Clin Med 2024; 13:344. [PMID: 38256478 PMCID: PMC10816993 DOI: 10.3390/jcm13020344] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2023] [Revised: 12/29/2023] [Accepted: 01/05/2024] [Indexed: 01/24/2024] Open
Abstract
The advent of artificial intelligence (AI) in medicine has transformed various medical specialties, including orthodontics. AI has shown promising results in enhancing the accuracy of diagnoses, treatment planning, and predicting treatment outcomes. Its usage in orthodontic practices worldwide has increased with the availability of various AI applications and tools. This review explores the principles of AI, its applications in orthodontics, and its implementation in clinical practice. A comprehensive literature review was conducted, focusing on AI applications in dental diagnostics, cephalometric evaluation, skeletal age determination, temporomandibular joint (TMJ) evaluation, decision making, and patient telemonitoring. Due to study heterogeneity, no meta-analysis was possible. AI has demonstrated high efficacy in all these areas, but variations in performance and the need for manual supervision suggest caution in clinical settings. The complexity and unpredictability of AI algorithms call for cautious implementation and regular manual validation. Continuous AI learning, proper governance, and addressing privacy and ethical concerns are crucial for successful integration into orthodontic practice.
Collapse
Affiliation(s)
- Natalia Kazimierczak
- Kazimierczak Private Medical Practice, Dworcowa 13/u6a, 85-009 Bydgoszcz, Poland
| | - Wojciech Kazimierczak
- Kazimierczak Private Medical Practice, Dworcowa 13/u6a, 85-009 Bydgoszcz, Poland
- Department of Radiology and Diagnostic Imaging, Collegium Medicum, Nicolaus Copernicus University in Torun, Jagiellońska 13-15, 85-067 Bydgoszcz, Poland
| | - Zbigniew Serafin
- Department of Radiology and Diagnostic Imaging, Collegium Medicum, Nicolaus Copernicus University in Torun, Jagiellońska 13-15, 85-067 Bydgoszcz, Poland
| | - Paweł Nowicki
- Kazimierczak Private Medical Practice, Dworcowa 13/u6a, 85-009 Bydgoszcz, Poland
| | - Jakub Nożewski
- Department of Emeregncy Medicine, University Hospital No 2 in Bydgoszcz, Ujejskiego 75, 85-168 Bydgoszcz, Poland
| | | |
Collapse
|
18
|
Rahim A, Khatoon R, Khan TA, Syed K, Khan I, Khalid T, Khalid B. Artificial intelligence-powered dentistry: Probing the potential, challenges, and ethicality of artificial intelligence in dentistry. Digit Health 2024; 10:20552076241291345. [PMID: 39539720 PMCID: PMC11558748 DOI: 10.1177/20552076241291345] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2024] [Accepted: 09/27/2024] [Indexed: 11/16/2024] Open
Abstract
Introduction Healthcare amelioration is exponential to technological advancement. In the recent era of automation, the consolidation of artificial intelligence (AI) in dentistry has rendered transformation in oral healthcare from a hardware-centric approach to a software-centric approach, leading to enhanced efficiency and improved educational and clinical outcomes. Objectives The aim of this narrative overview is to extend the succinct of the major events and innovations that led to the creation of modern-day AI and dentistry and the applicability of the former in dentistry. This article also prompts oral healthcare workers to endeavor a liable and optimal approach for effective incorporation of AI technology into their practice to promote oral health by exploring the potentials, constraints, and ethical considerations of AI in dentistry. Methods A comprehensive approach for searching the white and grey literature was carried out to collect and assess the data on AI, its use in dentistry, and the associated challenges and ethical concerns. Results AI in dentistry is still in its evolving phase with paramount applicabilities relevant to risk prediction, diagnosis, decision-making, prognosis, tailored treatment plans, patient management, and academia as well as the associated challenges and ethical concerns in its implementation. Conclusion The upsurging advancements in AI have resulted in transformations and promising outcomes across all domains of dentistry. In futurity, AI may be capable of executing a multitude of tasks in the domain of oral healthcare, at the level of or surpassing the ability of mankind. However, AI could be of significant benefit to oral health only if it is utilized under responsibility, ethicality and universality.
Collapse
Affiliation(s)
- Abid Rahim
- Sardar Begum Dental College, Gandhara University, Peshawar, Pakistan
| | - Rabia Khatoon
- Sardar Begum Dental College, Gandhara University, Peshawar, Pakistan
| | - Tahir Ali Khan
- Sardar Begum Dental College, Gandhara University, Peshawar, Pakistan
| | - Kawish Syed
- Sardar Begum Dental College, Gandhara University, Peshawar, Pakistan
| | - Ibrahim Khan
- Sardar Begum Dental College, Gandhara University, Peshawar, Pakistan
| | - Tamsal Khalid
- Sardar Begum Dental College, Gandhara University, Peshawar, Pakistan
| | - Balaj Khalid
- Syed Babar Ali School of Science and Engineering, Lahore University of Management Sciences, Lahore, Pakistan
| |
Collapse
|
19
|
Rahad K, Martin K, Amugo I, Ferguson S, Curtis A, Davis A, Gangula P, Wang Q. ChatGPT to enhance learning in dental education at a historically black medical college. RESEARCH SQUARE 2023:rs.3.rs-3546693. [PMID: 37986988 PMCID: PMC10659452 DOI: 10.21203/rs.3.rs-3546693/v2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 03/12/2024]
Abstract
The recent rise of powerful large language model (LLM)-based AI tools, exemplified by ChatGPT and Bard, poses a great challenge to contemporary dental education while simultaneously offering a unique resource and approach that potentially complements today's teaching and learning, where existing widely available learning resources have often fallen short. Although both the clinical and educational aspects of dentistry will be shaped profoundly by the LLM tools, the didactic curricula, which primarily rely on lecture-based courses where instructors impart knowledge through presentations and discussions, need to be upgraded urgently. In this paper, we used dental course materials, syllabi, and textbooks adopted currently in the School of Dentistry (SOD) at Meharry Medical College to assess the potential utility and effectiveness of ChatGPT in dental education. We collected the responses of the chatbot to questions as well as students' interactions with it for assessment. Our results showed that ChatGPT can assist in dental essay writing and generate relevant content for dental students, in addition to other benefits. The limitations of ChatGPT were also discussed in the paper.
Collapse
|
20
|
Rahad K, Martin K, Amugo I, Ferguson S, Curtis A, Davis A, Gangula P, Wang Q. ChatGPT to enhance learning in dental education at a historically black medical college. RESEARCH SQUARE 2023:rs.3.rs-3546693. [PMID: 37986988 PMCID: PMC10659452 DOI: 10.21203/rs.3.rs-3546693/v1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/22/2023]
Abstract
The recent rise of powerful large language model (LLM)-based AI tools, exemplified by ChatGPT and Bard, poses a great challenge to contemporary dental education while simultaneously offering a unique resource and approach that potentially complements today's teaching and learning, where existing widely available learning resources have often fallen short. Although both the clinical and educational aspects of dentistry will be shaped profoundly by the LLM tools, the didactic curricula, which primarily rely on lecture-based courses where instructors impart knowledge through presentations and discussions, need to be upgraded urgently. In this paper, we used dental course materials, syllabi, and textbooks adopted currently in the School of Dentistry (SOD) at Meharry Medical College to assess the potential utility and effectiveness of ChatGPT in dental education. We collected the responses of the chatbot to questions as well as students' interactions with it for assessment. Our results showed that ChatGPT can assist in dental essay writing and generate relevant content for dental students, in addition to other benefits. The limitations of ChatGPT were also discussed in the paper.
Collapse
|
21
|
Ahmad PN, Liu Y, Khan K, Jiang T, Burhan U. BIR: Biomedical Information Retrieval System for Cancer Treatment in Electronic Health Record Using Transformers. SENSORS (BASEL, SWITZERLAND) 2023; 23:9355. [PMID: 38067736 PMCID: PMC10708614 DOI: 10.3390/s23239355] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/28/2023] [Revised: 10/25/2023] [Accepted: 10/29/2023] [Indexed: 12/18/2023]
Abstract
The rapid growth of electronic health records (EHRs) has led to unprecedented biomedical data. Clinician access to the latest patient information can improve the quality of healthcare. However, clinicians have difficulty finding information quickly and easily due to the sheer data mining volume. Biomedical information retrieval (BIR) systems can help clinicians find the information required by automatically searching EHRs and returning relevant results. However, traditional BIR systems cannot understand the complex relationships between EHR entities. Transformers are a new type of neural network that is very effective for natural language processing (NLP) tasks. As a result, transformers are well suited for tasks such as machine translation and text summarization. In this paper, we propose a new BIR system for EHRs that uses transformers for predicting cancer treatment from EHR. Our system can understand the complex relationships between the different entities in an EHR, which allows it to return more relevant results to clinicians. We evaluated our system on a dataset of EHRs and found that it outperformed state-of-the-art BIR systems on various tasks, including medical question answering and information extraction. Our results show that Transformers are a promising approach for BIR in EHRs, reaching an accuracy and an F1-score of 86.46%, and 0.8157, respectively. We believe that our system can help clinicians find the information they need more quickly and easily, leading to improved patient care.
Collapse
Affiliation(s)
- Pir Noman Ahmad
- School of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001, China
| | - Yuanchao Liu
- School of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001, China
| | - Khalid Khan
- Department of Computing Science and Mathematics, University of Stirling, Stirling FK9 4LA, UK
| | - Tao Jiang
- School of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001, China
| | - Umama Burhan
- Department of Computing Science and Mathematics, University of Stirling, Stirling FK9 4LA, UK
| |
Collapse
|
22
|
Čverha M, Varga I, Trenčanská T, Šufliarsky B, Thurzo A. The Evolution of Robin Sequence Treatment Based on the Biomimetic Interdisciplinary Approach: A Historical Review. Biomimetics (Basel) 2023; 8:536. [PMID: 37999177 PMCID: PMC10669884 DOI: 10.3390/biomimetics8070536] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2023] [Revised: 11/01/2023] [Accepted: 11/08/2023] [Indexed: 11/25/2023] Open
Abstract
The Robin sequence is a congenital anomaly characterized by a triad of features: micrognathia, glossoptosis, and airway obstruction. This comprehensive historical review maps the evolution of approaches and appliances for its treatment from the past to the current modern possibilities of an interdisciplinary combination of modern engineering, medicine, materials, and computer science combined approach with emphasis on designing appliances inspired by nature and individual human anatomy. Current biomimetic designs are clinically applied, resulting in appliances that are more efficient, comfortable, sustainable, and safer than legacy traditional designs. This review maps the treatment modalities that have been used for patients with a Robin sequence over the years. Early management of the Robin sequence focused primarily on airway maintenance and feeding support, while current management strategies involve both nonsurgical and surgical interventions and biomimetic biocompatible personalized appliances. The goal of this paper was to provide a review of the evolution of management strategies for patients with the Robin sequence that led to the current interdisciplinary biomimetic approaches impacting the future of Robin Sequence treatment with biomimetics at the forefront.
Collapse
Affiliation(s)
- Martin Čverha
- Clinic of Pediatric Otorhinolaryngology of the Medical Faculty Comenius University in Bratislava and National Institute of Children’s Diseases, 83101 Bratislava, Slovakia;
| | - Ivan Varga
- Institute of Histology and Embryology, Faculty of Medicine, Comenius University in Bratislava, 81372 Bratislava, Slovakia;
| | - Tereza Trenčanská
- Clinic of Pediatric Otorhinolaryngology of the Medical Faculty Comenius University in Bratislava and National Institute of Children’s Diseases, 83101 Bratislava, Slovakia;
| | - Barbora Šufliarsky
- Department of Oral and Maxillofacial Surgery, Faculty of Medicine, Comenius University in Bratislava and University Hospital, 81372 Bratislava, Slovakia;
| | - Andrej Thurzo
- Department of Orthodontics, Regenerative and Forensic Dentistry, Faculty of Medicine, Comenius University in Bratislava, 81102 Bratislava, Slovakia
| |
Collapse
|
23
|
Veseli E, Tovani-Palone MR, Veseli A, Kastrati L. Should ChatGPT Have Some Applicability in the Management of Emergency Dental Care for Emigrant Adults and Children? J Contemp Dent Pract 2023; 24:819-820. [PMID: 38238266 DOI: 10.5005/jp-journals-10024-3576] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/23/2024]
Abstract
How to cite this article: Veseli E, Tovani-Palone MR, Veseli A, et al. Should ChatGPT Have Some Applicability in the Management of Emergency Dental Care for Emigrant Adults and Children? J Contemp Dent Pract 2023;24(11):819-820. Keywords: Artificial intelligence, Dental care, Dentistry, Emigrants and immigrants, Public health.
Collapse
Affiliation(s)
- Enis Veseli
- Department of Prosthodontics, Dental School, Faculty of Medicine, University of Pristina, Pristina, Kosovo, Orcid: https://orcid.org/0000-0002-7553-378X
| | - Marcos Roberto Tovani-Palone
- Department of Research Analytics, Saveetha Dental College and Hospitals, Saveetha Institute of Medical and Technical Sciences, Chennai, Tamil Nadu, India, e-mail: , Orcid: https://orcid.org/0000-0003-1149-2437
| | - Argjira Veseli
- University of Zagreb, School of Dental Medicine, Dental Science, Zagreb, Croatia, Orcid: https://orcid.org/0009-0004-0001-029X
| | - Lum Kastrati
- Department of Diabetes, Endocrinology, Nutritional Medicine and Metabolism, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland, Orcid: https://orcid.org/0000-0001-5759-1479
| |
Collapse
|
24
|
Pascadopoli M, Zampetti P, Nardi MG, Pellegrini M, Scribante A. Smartphone Applications in Dentistry: A Scoping Review. Dent J (Basel) 2023; 11:243. [PMID: 37886928 PMCID: PMC10605491 DOI: 10.3390/dj11100243] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Revised: 10/05/2023] [Accepted: 10/19/2023] [Indexed: 10/28/2023] Open
Abstract
This scoping review aims to investigate the latest literature concerning the use of smartphone applications (apps) in the prevention, management, and monitoring of oral diseases. Smartphone applications are software programs that are designed to run on smartphones. Nowadays, smartphones are regularly used by people of all ages, and mobile health apps (MHAs) represent an important means of spreading information related to oral health, which is the state of the mouth and teeth, including the gums and other tissues. Several apps have been designed to promote prevention, diagnosis, and therapeutic adherence monitoring. This scoping review considered randomized clinical trials, cross-sectional studies, before-after (pre-post) studies with no control group, and observational studies. Once the inclusion and exclusion criteria had been defined, a preliminary confined search was performed on PubMed and Scopus; key terms from the collected articles were selected to design a search strategy, and then a search of all the included articles' reference lists was run for further research. Studies were excluded if they did not fulfill the inclusion criteria. The preferred reporting items for scoping reviews (PRISMA-ScR) consensus was followed. The risk of bias was evaluated by providing a qualitative analysis of the clinical studies via the National Heart, Lung, and Blood Institute (NHLBI) Quality Assessment of Controlled Intervention Studies, Observational Cohort Studies, and Cross-Sectional Studies (NHLBI, NIH). A total of 21 studies were included in this review. As it is clear from the studies selected, the literature indicates that MHAs are effective in improving oral hygiene in adolescents and children and reducing the dental plaque index, including in patients undergoing orthodontic treatment. MHAs are also able to reduce the symptoms of patients affected by obstructive sleep apnea-hypopnea syndrome (OSAHS) and improve the swallowing-related quality of life of elderly patients. MHAs are furthermore recommended to decrease dental anxiety among patients, both during dental procedures and the post-operative period. MHAs are useful to spread knowledge about traumatic dental injuries among non-oral health professionals and to monitor dental erosion and awake bruxism. MHAs' clinical outcomes might have been influenced by the demographic features of the subjects involved. Further studies considering a longer follow-up period and larger samples are needed. In conclusion, MHAs can be considered a useful tool to monitor oral disease and increase patients' quality of life related to oral health.
Collapse
Affiliation(s)
- Maurizio Pascadopoli
- Unit of Orthodontics and Pediatric Dentistry, Section of Dentistry, Department of Clinical, Surgical, Diagnostic and Pediatric Sciences, University of Pavia, 27100 Pavia, Italy; (M.P.); (P.Z.); (A.S.)
| | - Paolo Zampetti
- Unit of Orthodontics and Pediatric Dentistry, Section of Dentistry, Department of Clinical, Surgical, Diagnostic and Pediatric Sciences, University of Pavia, 27100 Pavia, Italy; (M.P.); (P.Z.); (A.S.)
| | - Maria Gloria Nardi
- Unit of Orthodontics and Pediatric Dentistry, Section of Dentistry, Department of Clinical, Surgical, Diagnostic and Pediatric Sciences, University of Pavia, 27100 Pavia, Italy; (M.P.); (P.Z.); (A.S.)
| | - Matteo Pellegrini
- Maxillofacial Surgery and Dental Unit, Fondazione IRCCS Cà Granda Ospedale Maggiore Policlinico, 20122 Milan, Italy
- Department of Biomedical, Surgical and Dental Sciences, University of Milan, Via della Commenda 10, 20122 Milan, Italy
| | - Andrea Scribante
- Unit of Orthodontics and Pediatric Dentistry, Section of Dentistry, Department of Clinical, Surgical, Diagnostic and Pediatric Sciences, University of Pavia, 27100 Pavia, Italy; (M.P.); (P.Z.); (A.S.)
- Unit of Dental Hygiene, Section of Dentistry, Department of Clinical, Surgical, Diagnostic and Pediatric Sciences, University of Pavia, 27100 Pavia, Italy
| |
Collapse
|
25
|
Liu J, Zhang C, Shan Z. Application of Artificial Intelligence in Orthodontics: Current State and Future Perspectives. Healthcare (Basel) 2023; 11:2760. [PMID: 37893833 PMCID: PMC10606213 DOI: 10.3390/healthcare11202760] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/24/2023] [Revised: 10/11/2023] [Accepted: 10/16/2023] [Indexed: 10/29/2023] Open
Abstract
In recent years, there has been the notable emergency of artificial intelligence (AI) as a transformative force in multiple domains, including orthodontics. This review aims to provide a comprehensive overview of the present state of AI applications in orthodontics, which can be categorized into the following domains: (1) diagnosis, including cephalometric analysis, dental analysis, facial analysis, skeletal-maturation-stage determination and upper-airway obstruction assessment; (2) treatment planning, including decision making for extractions and orthognathic surgery, and treatment outcome prediction; and (3) clinical practice, including practice guidance, remote care, and clinical documentation. We have witnessed a broadening of the application of AI in orthodontics, accompanied by advancements in its performance. Additionally, this review outlines the existing limitations within the field and offers future perspectives.
Collapse
Affiliation(s)
- Junqi Liu
- Division of Paediatric Dentistry and Orthodontics, Faculty of Dentistry, The University of Hong Kong, Hong Kong SAR, China;
| | - Chengfei Zhang
- Division of Restorative Dental Sciences, Faculty of Dentistry, The University of Hong Kong, Hong Kong SAR, China;
| | - Zhiyi Shan
- Division of Paediatric Dentistry and Orthodontics, Faculty of Dentistry, The University of Hong Kong, Hong Kong SAR, China;
| |
Collapse
|
26
|
Fawaz P, Sayegh PE, Vannet BV. What is the current state of artificial intelligence applications in dentistry and orthodontics? JOURNAL OF STOMATOLOGY, ORAL AND MAXILLOFACIAL SURGERY 2023; 124:101524. [PMID: 37270174 DOI: 10.1016/j.jormas.2023.101524] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/22/2023] [Revised: 05/08/2023] [Accepted: 05/31/2023] [Indexed: 06/05/2023]
Abstract
BACKGROUND The use of Artificial Intelligence (AI) in the medical field has the potential to bring about significant improvements in patient care and outcomes. AI is being used in dentistry and more specifically in orthodontics through the development of diagnostic imaging tools, the development of treatment planning tools, and the development of robotic surgery. The aim of this study is to present the latest emerging AI softwares and applications in dental field to benefit from. TYPES OF STUDIES REVIEWED Search strategies were conducted in three electronic databases, with no date limits in the following databases up to April 30, 2023: MEDLINE, PUBMED, and GOOGLE® SCHOLAR for articles related to AI in dentistry & orthodontics. No inclusion and exclusion criteria were used for the selection of the articles. Most of the articles included (n = 79) are reviews of the literature, retro/prospective studies, systematic reviews and meta-analyses, and observational studies. RESULTS The use of AI in dentistry and orthodontics is a rapidly growing area of research and development, with the potential to revolutionize the field and bring about significant improvements in patient care and outcomes; this can save clinicians' chair-time and push for more individualized treatment plans. Results from the various studies reported in this review are suggestive that the accuracy of AI-based systems is quite promising and reliable. PRACTICAL IMPLICATIONS AI application in the healthcare field has proven to be efficient and helpful for the dentist to be more precise in diagnosis and clinical decision-making. These systems can simplify the tasks and provide results in quick time which can save dentists time and help them perform their duties more efficiently. These systems can be of greater aid and can be used as auxiliary support for dentists with lesser experience.
Collapse
Affiliation(s)
- Paul Fawaz
- Academic Lecturer & Researcher at the Orthodontic department Université de Lorraine, Nancy, France.
| | | | - Bart Vande Vannet
- Clinical and Academical responsable of the Orthodontic department at Université de Lorraine, Nancy, France.
| |
Collapse
|
27
|
Giansanti D. Bridging the Gap: Exploring Opportunities, Challenges, and Problems in Integrating Assistive Technologies, Robotics, and Automated Machines into the Health Domain. Healthcare (Basel) 2023; 11:2462. [PMID: 37685498 PMCID: PMC10487463 DOI: 10.3390/healthcare11172462] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2023] [Revised: 08/11/2023] [Accepted: 08/31/2023] [Indexed: 09/10/2023] Open
Abstract
The field of healthcare is continually evolving and advancing due to new technologies and innovations [...].
Collapse
Affiliation(s)
- Daniele Giansanti
- National Centre for Innovative Technologies in Public Health, Italian National Institute of Health, 00161 Rome, Italy
| |
Collapse
|
28
|
Watters C, Lemanski MK. Universal skepticism of ChatGPT: a review of early literature on chat generative pre-trained transformer. Front Big Data 2023; 6:1224976. [PMID: 37680954 PMCID: PMC10482048 DOI: 10.3389/fdata.2023.1224976] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Accepted: 07/10/2023] [Indexed: 09/09/2023] Open
Abstract
ChatGPT, a new language model developed by OpenAI, has garnered significant attention in various fields since its release. This literature review provides an overview of early ChatGPT literature across multiple disciplines, exploring its applications, limitations, and ethical considerations. The review encompasses Scopus-indexed publications from November 2022 to April 2023 and includes 156 articles related to ChatGPT. The findings reveal a predominance of negative sentiment across disciplines, though subject-specific attitudes must be considered. The review highlights the implications of ChatGPT in many fields including healthcare, raising concerns about employment opportunities and ethical considerations. While ChatGPT holds promise for improved communication, further research is needed to address its capabilities and limitations. This literature review provides insights into early research on ChatGPT, informing future investigations and practical applications of chatbot technology, as well as development and usage of generative AI.
Collapse
Affiliation(s)
- Casey Watters
- Faculty of Law, Bond University, Gold Coast, QLD, Australia
| | | |
Collapse
|
29
|
Beaulieu-Jones BR, Shah S, Berrigan MT, Marwaha JS, Lai SL, Brat GA. Evaluating Capabilities of Large Language Models: Performance of GPT4 on Surgical Knowledge Assessments. MEDRXIV : THE PREPRINT SERVER FOR HEALTH SCIENCES 2023:2023.07.16.23292743. [PMID: 37502981 PMCID: PMC10371188 DOI: 10.1101/2023.07.16.23292743] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/29/2023]
Abstract
Background Artificial intelligence (AI) has the potential to dramatically alter healthcare by enhancing how we diagnosis and treat disease. One promising AI model is ChatGPT, a large general-purpose language model trained by OpenAI. The chat interface has shown robust, human-level performance on several professional and academic benchmarks. We sought to probe its performance and stability over time on surgical case questions. Methods We evaluated the performance of ChatGPT-4 on two surgical knowledge assessments: the Surgical Council on Resident Education (SCORE) and a second commonly used knowledge assessment, referred to as Data-B. Questions were entered in two formats: open-ended and multiple choice. ChatGPT output were assessed for accuracy and insights by surgeon evaluators. We categorized reasons for model errors and the stability of performance on repeat encounters. Results A total of 167 SCORE and 112 Data-B questions were presented to the ChatGPT interface. ChatGPT correctly answered 71% and 68% of multiple-choice SCORE and Data-B questions, respectively. For both open-ended and multiple-choice questions, approximately two-thirds of ChatGPT responses contained non-obvious insights. Common reasons for inaccurate responses included: inaccurate information in a complex question (n=16, 36.4%); inaccurate information in fact-based question (n=11, 25.0%); and accurate information with circumstantial discrepancy (n=6, 13.6%). Upon repeat query, the answer selected by ChatGPT varied for 36.4% of inaccurate questions; the response accuracy changed for 6/16 questions. Conclusion Consistent with prior findings, we demonstrate robust near or above human-level performance of ChatGPT within the surgical domain. Unique to this study, we demonstrate a substantial inconsistency in ChatGPT responses with repeat query. This finding warrants future consideration and presents an opportunity to further train these models to provide safe and consistent responses. Without mental and/or conceptual models, it is unclear whether language models such as ChatGPT would be able to safely assist clinicians in providing care.
Collapse
Affiliation(s)
- Brendin R Beaulieu-Jones
- Department of Surgery, Beth Israel Deaconess Medical Center, Boston, MA
- Department of Biomedical Informatics, Harvard Medical School, Boston, MA
| | - Sahaj Shah
- Geisinger Commonwealth School of Medicine, Scranton, PA
| | | | - Jayson S Marwaha
- Division of Colorectal Surgery, National Taiwan University Hospital, Taipei, Taiwan
| | - Shuo-Lun Lai
- Division of Colorectal Surgery, National Taiwan University Hospital, Taipei, Taiwan
| | - Gabriel A Brat
- Department of Surgery, Beth Israel Deaconess Medical Center, Boston, MA
- Department of Biomedical Informatics, Harvard Medical School, Boston, MA
| |
Collapse
|
30
|
Abstract
ChatGPT is an artificial intelligence (AI) chatbot developed by OpenAI and it first became available to the public in November 2022. ChatGPT can assist in finding academic papers on the web and summarizing them. This chatbot has the potential to be applied in scientific writing, it has the ability to generate automated drafts, summarize articles, and translate content from several languages. This in turn can make academic writing faster and less challenging. However, due to ethical considerations, its use in scientific writing should be regulated and carefully monitored. Few papers have discussed the use of ChatGPT in scientific research writing. This review aims to discuss all the relevant published papers that discuss the use of ChatGPT in medical and dental research.
Collapse
Affiliation(s)
- Bader Fatani
- Dentistry, College of Dentistry, King Saud University, Riyadh, SAU
| |
Collapse
|
31
|
Alhaidry HM, Fatani B, Alrayes JO, Almana AM, Alfhaed NK. ChatGPT in Dentistry: A Comprehensive Review. Cureus 2023; 15:e38317. [PMID: 37266053 PMCID: PMC10230850 DOI: 10.7759/cureus.38317] [Citation(s) in RCA: 17] [Impact Index Per Article: 8.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 04/29/2023] [Indexed: 06/03/2023] Open
Abstract
Chat generative pre-trained transformer (ChatGPT) is an artificial intelligence chatbot that uses natural language processing that can respond to human input in a conversational manner. ChatGPT has numerous applications in the health care system including dentistry; it is used in diagnoses and for assessing disease risk and scheduling appointments. It also has a role in scientific research. In the dental field, it has provided many benefits such as detecting dental and maxillofacial abnormalities on panoramic radiographs and identifying different dental restorations. Therefore, it helps in decreasing the workload. But even with these benefits, one should take into consideration the risks and limitations of this chatbot. Few articles mentioned the use of ChatGPT in dentistry. This comprehensive review represents data collected from 66 relevant articles using PubMed and Google Scholar as databases. This review aims to discuss all relevant published articles on the use of ChatGPT in dentistry.
Collapse
Affiliation(s)
- Hind M Alhaidry
- Advanced General Dentistry, Prince Sultan Military Medical City, Riyadh, SAU
| | - Bader Fatani
- Dentistry, College of Dentistry, King Saud University, Riyadh, SAU
| | - Jenan O Alrayes
- Dentistry, College of Dentistry, King Saud University, Riyadh, SAU
| | | | - Nawaf K Alfhaed
- Dentistry, College of Dentistry, King Saud University, Riyadh, SAU
| |
Collapse
|