1
|
Ryhänen J, Wong GC, Anttila T, Chung KC. Overview of artificial intelligence in hand surgery. J Hand Surg Eur Vol 2025; 50:738-751. [PMID: 40035151 DOI: 10.1177/17531934251322723] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 03/05/2025]
Abstract
Artificial intelligence has evolved significantly since its inception, becoming a powerful tool in medicine. This paper provides an overview of the core principles, applications and future directions of artificial intelligence in hand surgery. Artificial intelligence has shown promise in improving diagnostic accuracy, predicting outcomes and assisting in patient education. However, despite its potential, its application in hand surgery is still nascent, with most studies being retrospective and limited by small sample sizes. To harness the full potential of artificial intelligence in hand surgery and support broader adoption, more robust, large-scale studies are needed. Collaboration among researchers, through data sharing and federated learning, is essential for advancing artificial intelligence from experimental to clinically validated tools, ultimately enhancing patient care and clinical workflows.
Collapse
Affiliation(s)
- Jorma Ryhänen
- Musculoskeletal and Plastic Surgery, Department of Hand Surgery, University of Helsinki and Helsinki University Hospital, Helsinki, Finland
| | - Gordon C Wong
- Section of Plastic Surgery, Michigan Medicine, Ann Arbor, MI, USA
| | - Turkka Anttila
- Musculoskeletal and Plastic Surgery, Department of Hand Surgery, University of Helsinki and Helsinki University Hospital, Helsinki, Finland
| | - Kevin C Chung
- Section of Plastic Surgery, Department of Surgery, Michigan Medicine, Ann Arbor, MI, USA
| |
Collapse
|
2
|
Archer T, Archer S, Shah SS. Musculoskeletal fracture detection: artificial intelligence and machine learning-based diagnostic advantages and pitfalls. Eur Radiol 2025; 35:3649-3651. [PMID: 39638943 DOI: 10.1007/s00330-024-11250-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2024] [Revised: 11/16/2024] [Accepted: 11/22/2024] [Indexed: 12/07/2024]
Affiliation(s)
- Truman Archer
- Texas Tech University Health Sciences Center School of Medicine, Lubbock, TX, USA.
| | - Sawyer Archer
- Texas Tech University Health Sciences Center School of Medicine, Lubbock, TX, USA
| | | |
Collapse
|
3
|
Loeffen DV, Zijta FM, Boymans TA, Wildberger JE, Nijssen EC. AI for fracture diagnosis in clinical practice: Four approaches to systematic AI-implementation and their impact on AI-effectiveness. Eur J Radiol 2025; 187:112113. [PMID: 40252277 DOI: 10.1016/j.ejrad.2025.112113] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/19/2024] [Revised: 12/16/2024] [Accepted: 04/12/2025] [Indexed: 04/21/2025]
Abstract
PURPOSE Artificial Intelligence (AI) has been shown to enhance fracture-detection-accuracy, but the most effective AI-implementation in clinical practice is less well understood. In the current study, four approaches to AI-implementation are evaluated for their impact on AI-effectiveness. MATERIALS AND METHODS Retrospective single-center study based on all consecutive, around-the-clock radiographic examinations for suspected fractures, and accompanying clinical-practice radiologist-diagnoses, between January and March 2023. These image-sets were independently analysed by a dedicated bone-fracture-detection-AI. Findings were combined with radiologist clinical-practice diagnoses to simulate the four AI-implementation methods deemed most relevant to clinical workflows: AI-standalone (radiologist-findings not consulted); AI-problem-solving (AI-findings consulted when radiologist in doubt); AI-triage (radiologist-findings consulted when AI in doubt); and AI-safety net (AI-findings consulted when radiologist diagnosis negative). Reference-standard diagnoses were established by two senior musculoskeletal-radiologists (by consensus in cases of disagreement). Radiologist- and radiologist + AI diagnoses were compared for false negatives (FN), false positives (FP) and their clinical consequences. Experience-level-subgroups radiologists-in-training-, non-musculoskeletal-radiologists, and dedicated musculoskeletal-radiologists were analysed separately. RESULTS 1508 image-sets were included (1227 unique patients; 40 radiologist-readers). Radiologist results were: 2.7 % FN (40/1508), 28 with clinical consequences; 1.2 % FP (18/1508), 2 received full-fracture treatments (11.1 %). All AI-implementation methods changed overall FN and FP with statistical significance (p < 0.001): AI-standalone 1.5 % FN (23/1508; 11 consequences), 6.8 % FP (103/1508); AI-problem-solving 3.2 % FN (48/1508; 31 consequences), 0.6 % FP (9/1508); AI-triage 2.1 % FN (32/1508; 18 consequences), 1.7 % FP (26/1508); AI-safety net 0.07 % FN (1/1508; 1 consequence), 7.6 % FP (115/1508). Subgroups show similar trends, except AI-triage increased FN for all subgroups except radiologists-in-training. CONCLUSION Implementation methods have a large impact on AI-effectiveness. These results suggest AI should not be considered for problem-solving or triage at this time; AI standalone performs better than either and may be a source of assistance where radiologists are unavailable. Best results were obtained implementing AI as safety net, which eliminates missed fractures with serious clinical consequences; even though false positives are increased, unnecessary treatments are limited.
Collapse
Affiliation(s)
- Daan V Loeffen
- Department of Radiology and Nuclear Medicine, Maastricht University Medical Centre, the Netherlands; CARIM School for Cardiovascular Diseases, Maastricht University, the Netherlands
| | - Frank M Zijta
- Department of Radiology and Nuclear Medicine, Maastricht University Medical Centre, the Netherlands; CARIM School for Cardiovascular Diseases, Maastricht University, the Netherlands; CAPHRI Care and Public Health Research Institute, Maastricht University, the Netherlands
| | - Tim A Boymans
- CAPHRI Care and Public Health Research Institute, Maastricht University, the Netherlands; Department of Orthopaedic Surgery, Maastricht University Medical Centre, the Netherlands
| | - Joachim E Wildberger
- Department of Radiology and Nuclear Medicine, Maastricht University Medical Centre, the Netherlands; CARIM School for Cardiovascular Diseases, Maastricht University, the Netherlands
| | - Estelle C Nijssen
- Department of Radiology and Nuclear Medicine, Maastricht University Medical Centre, the Netherlands; CARIM School for Cardiovascular Diseases, Maastricht University, the Netherlands.
| |
Collapse
|
4
|
Oftring ZS, Deutsch K, Tolks D, Jungmann F, Kuhn S. Novel Blended Learning on Artificial Intelligence for Medical Students: Qualitative Interview Study. JMIR MEDICAL EDUCATION 2025; 11:e65220. [PMID: 40418795 DOI: 10.2196/65220] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/09/2024] [Revised: 03/03/2025] [Accepted: 04/06/2025] [Indexed: 05/28/2025]
Abstract
Background Artificial intelligence (AI) systems are becoming increasingly relevant in everyday clinical practice, with Food and Drug Administration-approved AI solutions now available in many specialties. This development has far-reaching implications for doctors and the future medical profession, highlighting the need for both practicing physicians and medical students to acquire the knowledge, skills, and attitudes necessary to effectively use and evaluate these technologies. Currently, however, there is limited experience with AI-focused curricular training and continuing education. Objective This paper first introduces a novel blended learning curriculum including one module on AI for medical students in Germany. Second, this paper presents findings from a qualitative postcourse evaluation of students' knowledge and attitudes toward AI and their overall perception of the course. Methods Clinical-year medical students can attend a 5-day elective course called "Medicine in the Digital Age," which includes one dedicated AI module alongside 4 others on digital doctor-patient communication; digital health applications and smart devices; telemedicine; and virtual/augmented reality and robotics. After course completion, participants were interviewed in semistructured small group interviews. The interview guide was developed deductively from existing evidence and research questions compiled by our group. A subset of interview questions focused on students' knowledge, skills, and attitudes regarding medical AI, and their overall course assessment. Responses were analyzed using Mayring's qualitative content analysis. This paper reports on the subset of students' statements about their perception and attitudes toward AI and the elective's general evaluation. Results We conducted a total of 18 group interviews, in which all 35 (100%) participants (female=11, male=24) from 3 consecutive course runs participated. This produced a total of 214 statements on AI, which were assigned to the 3 main categories "Areas of Application," "Future Work," and "Critical Reflection." The findings indicate that students have a nuanced and differentiated understanding of AI. Additionally, 610 statements concerned the elective's overall assessment, demonstrating great learning benefits and high levels of acceptance of the teaching concept. All 35 students would recommend the elective to peers. Conclusions The evaluation demonstrated that the AI module effectively generates competences regarding AI technology, fosters a critical perspective, and prepares medical students to engage with the technology in a differentiated manner. The curriculum is feasible, beneficial, and highly accepted among students, suggesting it could serve as a teaching model for other medical institutions. Given the growing number and impact of medical AI applications, there is a pressing need for more AI-focused curricula and further research on their educational impact.
Collapse
Affiliation(s)
- Zoe S Oftring
- Institute for Digital Medicine, Philipps University Marburg and University Clinic Giessen & Marburg, Baldingerstrasse 1, Marburg, 35042, Germany, 49 (0)6421 ext 58
- Department of Paediatrics, University Clinic Giessen & Marburg, Marburg, Germany
| | - Kim Deutsch
- Institute of Educational Science, Johannes Gutenberg University, Mainz, Germany
| | - Daniel Tolks
- Institute of Anatomy, Rostock University Medical Centre, Rostock, Germany
- Professorship in Health Management, International University of Applied Science, Hamburg, Germany
| | - Florian Jungmann
- Xcare Group Radiology, Nuclear Medicine and Radiotherapy, Saarlouis, Germany
| | - Sebastian Kuhn
- Institute for Digital Medicine, Philipps University Marburg and University Clinic Giessen & Marburg, Baldingerstrasse 1, Marburg, 35042, Germany, 49 (0)6421 ext 58
| |
Collapse
|
5
|
Smith ME, Zalesky CC, Lee S, Gottlieb M, Adhikari S, Goebel M, Wegman M, Garg N, Lam SH. Artificial Intelligence in Emergency Medicine: A Primer for the Nonexpert. J Am Coll Emerg Physicians Open 2025; 6:100051. [PMID: 40034198 PMCID: PMC11874537 DOI: 10.1016/j.acepjo.2025.100051] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/27/2023] [Revised: 12/15/2024] [Accepted: 01/02/2025] [Indexed: 03/05/2025] Open
Abstract
Artificial intelligence (AI) is increasingly being utilized to augment the practice of emergency medicine due to rapid technological advances and breakthroughs. AI applications have been used to enhance triage systems, predict disease-specific risk, estimate staffing needs, forecast patient decompensation, and interpret imaging findings in the emergency department setting. This article aims to help readers without formal training become informed end-users of AI in emergency medicine. The authors will briefly discuss the principles and key terminology of AI, the reasons for its rising popularity, its potential applications in the emergency department setting, and its limitations. Additionally, resources for further self-studying will also be provided.
Collapse
Affiliation(s)
- Moira E. Smith
- Department of Emergency Medicine, University of Virginia, Charlottesville, Virginia, USA
| | - C. Christopher Zalesky
- Department of Anesthesia, Division of Critical Care, Emory University School of Medicine, Atlanta, Georgia, USA
| | - Sangil Lee
- Department of Emergency Medicine, University of Iowa Carver College of Medicine, Iowa City, Iowa, USA
| | - Michael Gottlieb
- Emergency Ultrasound Division, Department of Emergency Medicine, Rush University Medical Center, Chicago, Illinois, USA
| | - Srikar Adhikari
- Department of Emergency Medicine, University of Arizona, Tucson, Arizona, USA
| | - Mat Goebel
- Department of Emergency Medicine, Mercy Medical Center - Trinity Health of New England, Springfield, Massachusetts, USA
| | - Martin Wegman
- Department of Emergency Medicine, Orange Park Medical Center, Orange Park, Florida, USA
| | - Nidhi Garg
- Department of Emergency Medicine, South Shore University Hospital/Northwell Health, Donald and Barbara Zucker School of Medicine at Hofstra/Northwell, Hempstead, New York, USA
| | - Samuel H.F. Lam
- Section of Emergency Medicine, Children's Hospital Colorado, University of Colorado School of Medicine, Aurora, Colorado, USA
| |
Collapse
|
6
|
Hembroff G, Klochko C, Craig J, Changarnkothapeecherikkal H, Loi RQ. Improved Automated Quality Control of Skeletal Wrist Radiographs Using Deep Multitask Learning. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025; 38:838-849. [PMID: 39187704 PMCID: PMC11950583 DOI: 10.1007/s10278-024-01220-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/06/2024] [Revised: 07/17/2024] [Accepted: 07/29/2024] [Indexed: 08/28/2024]
Abstract
Radiographic quality control is an integral component of the radiology workflow. In this study, we developed a convolutional neural network model tailored for automated quality control, specifically designed to detect and classify key attributes of wrist radiographs including projection, laterality (based on the right/left marker), and the presence of hardware and/or casts. The model's primary objective was to ensure the congruence of results with image requisition metadata to pass the quality assessment. Using a dataset of 6283 wrist radiographs from 2591 patients, our multitask-capable deep learning model based on DenseNet 121 architecture achieved high accuracy in classifying projections (F1 Score of 97.23%), detecting casts (F1 Score of 97.70%), and identifying surgical hardware (F1 Score of 92.27%). The model's performance in laterality marker detection was lower (F1 Score of 82.52%), particularly for partially visible or cut-off markers. This paper presents a comprehensive evaluation of our model's performance, highlighting its strengths, limitations, and the challenges encountered during its development and implementation. Furthermore, we outline planned future research directions aimed at refining and expanding the model's capabilities for improved clinical utility and patient care in radiographic quality control.
Collapse
Affiliation(s)
- Guy Hembroff
- Department of Applied Computing, Michigan Technological University, 1400 Townsend Drive, Houghton, MI, 49931, USA.
| | - Chad Klochko
- Department of Radiology, Division of Musculoskeletal Radiology, Henry Ford Hospital, 2799 West Grand Boulevard, Detroit, MI, 48202, USA
| | - Joseph Craig
- Department of Radiology, Division of Musculoskeletal Radiology, Henry Ford Hospital, 2799 West Grand Boulevard, Detroit, MI, 48202, USA
| | | | - Richard Q Loi
- Department of Radiology, Division of Musculoskeletal Radiology, Henry Ford Hospital, 2799 West Grand Boulevard, Detroit, MI, 48202, USA
| |
Collapse
|
7
|
Dubreucq Guerif E, Agut S, Rousseau A, Bompard R, Goulet H. Evaluation of the use of artificial intelligence in the detection of appendicular skeletal fractures in adult patients consulting in an emergency department, a retrospective study. Eur J Emerg Med 2025; 32:144-146. [PMID: 40009538 DOI: 10.1097/mej.0000000000001193] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/28/2025]
Affiliation(s)
| | | | - Alexandra Rousseau
- Unité de Recherche Clinique (URC) de l'Est Parisien Hôpital Saint-Antoine, Assistance Publique-Hôpitaux de Paris, Paris, France
| | | | | |
Collapse
|
8
|
Jaillat A, Cyteval C, Baron Sarrabere MP, Ghomrani H, Maman Y, Thouvenin Y, Pastor M. Added value of artificial intelligence for the detection of pelvic and hip fractures. Jpn J Radiol 2025:10.1007/s11604-025-01754-0. [PMID: 40038216 DOI: 10.1007/s11604-025-01754-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2024] [Accepted: 02/13/2025] [Indexed: 03/06/2025]
Abstract
PURPOSE To assess the added value of artificial intelligence (AI) for radiologists and emergency physicians in the radiographic detection of pelvic fractures. MATERIALS & METHODS In this retrospective study, one junior radiologist reviewed 940 X-rays of patients admitted to emergency for a fall with suspicion of pelvic fracture between March 2020 and June 2021. The radiologist analyzed the X-rays alone and then using an AI system (BoneView). In a random sample of 100 exams, the same procedure was repeated alongside five other readers (three radiologists and two emergency physicians with 3-30 years of experience). The reference diagnosis was based on the patient's full set of medical imaging exams and medical records in the months following emergency admission. RESULTS A total of 633 confirmed pelvic fractures (64.8% from hip and 35.2% from pelvic ring) in 940 patients and 68 pelvic fractures (60% from hip and 40% from pelvic ring) in the 100-patient sample were included. In the whole dataset, the junior radiologist achieved a significant sensitivity improvement with AI assistance (Se-PELVIC = 77.25% to 83.73%; p < 0.001, Se-HIP 93.24 to 96.49%; p < 0.001 and Se-PELVIC RING 54.60% to 64.50%; p < 0.001). However, there was a significant decrease in specificity with AI assistance (Spe-PELVIC = 95.24% to 93.25%; p = 0.005 and Spe-HIP = 98.30% to 96.90%; p = 0.005). In the 100-patient sample, the two emergency physicians obtained an improvement in fracture detection sensitivity across the pelvic area + 14.70% (p = 0.0011) and + 10.29% (p < 0.007) respectively without a significant decrease in specificity. For hip fractures, E1's sensitivity increased from 59.46% to 70.27% (p = 0.04), and E2's sensitivity increased from 78.38% to 86.49% (p = 0.08). For pelvic ring fractures, E1's sensitivity increased from 12.90% to 32.26% (p = 0.012), and E2's sensitivity increased from 19.35% to 32.26% (p = 0.043). CONCLUSION AI improved the diagnostic performance for emergency physicians and radiologists with limited experience in pelvic fracture screening.
Collapse
Affiliation(s)
- Anthony Jaillat
- Osteoarticular Medical Imaging Section, Department of Medical Imaging, Lapeyronie University Hospital, Montpellier, France
| | - Catherine Cyteval
- Osteoarticular Medical Imaging Section, Department of Medical Imaging, Lapeyronie University Hospital, Montpellier, France
| | - Marie-Pierre Baron Sarrabere
- Osteoarticular Medical Imaging Section, Department of Medical Imaging, Lapeyronie University Hospital, Montpellier, France
| | - Hamza Ghomrani
- Emergency Department, Lapeyronie University Hospital, Montpellier, France
| | - Yoav Maman
- Emergency Department, Lapeyronie University Hospital, Montpellier, France
| | - Yann Thouvenin
- Osteoarticular Medical Imaging Section, Department of Medical Imaging, Lapeyronie University Hospital, Montpellier, France
| | - Maxime Pastor
- Osteoarticular Medical Imaging Section, Department of Medical Imaging, Lapeyronie University Hospital, Montpellier, France.
| |
Collapse
|
9
|
Xie Y, Chen X, Yang H, Wang H, Zhou H, Lu L, Zhang J, Liu P, Ye Z. Integrating blockchain technology with artificial intelligence for the diagnosis of tibial plateau fractures. Eur J Trauma Emerg Surg 2025; 51:119. [PMID: 39984717 DOI: 10.1007/s00068-025-02793-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2024] [Accepted: 02/06/2025] [Indexed: 02/23/2025]
Abstract
PURPOSE The application of artificial intelligence (AI) in healthcare has seen widespread implementation, with numerous studies highlighting the development of robust algorithms. However, limited attention has been given to the secure utilization of raw data for medical model training, and its subsequent impact on clinical decision-making and real-world applications. This study aims to assess the feasibility and effectiveness of an advanced diagnostic model that integrates blockchain technology and AI for the identification of tibial plateau fractures (TPFs) in emergency settings. METHOD In this study, blockchain technology was utilized to construct a distributed database for trauma orthopedics, images collected from three independent hospitals for model training, testing, and internal validation. Then, a distributed network combining blockchain and deep learning was developed for the detection of TPFs, with model parameters aggregated across multiple nodes to enhance accuracy. The model's performance was comprehensively evaluated using metrics including accuracy, sensitivity, specificity, F1 score, and the area under the receiver operating characteristic curve (AUC). In addition, the performance of the centralized model, the distributed AI model, clinical orthopedic attending physicians, and AI-assisted attending physicians was tested on an external validation dataset. RESULTS In the testing set, the accuracy of our distributed model was 0.9603 [95% CI (0.9598, 0.9605)] and the AUC was 0.9911 [95% CI (0.9893, 0.9915)] for TPF detection. In the external validation set, the accuracy reached 0.9636 [95% CI (0.9388, 0.9762)], was slightly higher than that of the centralized YOLOv8n model at 0.9632 [95% CI (0.9387, 0.9755)] (p > 0.05), and exceeded the orthopedic physician at 0.9291 [95% CI (0.9002, 0.9482)] and radiology attending physician at 0.9175 [95% CI (0.8891, 0.9393)], with a statistically significant difference (p < 0.05). Additionally, the centralized model (4.99 ± 0.01 min) had shorter diagnosis times compared to the orthopedic attending physician (25.45 ± 1.92 min) and the radiology attending physician (26.21 ± 1.20 min), with a statistically significant difference (p < 0.05). CONCLUSION The model based on the integration of blockchain technology and AI can realize safe, collaborative, and convenient assisted diagnosis of TPF. Through the aggregation of training parameters by decentralized algorithms, it can achieve model construction without data leaving the hospital and may exert clinical application value in the emergency settings.
Collapse
Affiliation(s)
- Yi Xie
- Department of Orthopedics Surgery, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
- Laboratory of Intelligent Medicine, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
| | - Xiaoliang Chen
- People's Hospital of Ningxia Hui Autonomous Region, Ningxia Medical University, Yinchuan, China
| | - Huiwen Yang
- Laboratory of Intelligent Medicine, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
- Department of Otorhinolaryngology, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
| | - Honglin Wang
- Department of Orthopedics Surgery, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
- Laboratory of Intelligent Medicine, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
| | - Hong Zhou
- Department of Orthopedics Surgery, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
- Laboratory of Intelligent Medicine, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
| | - Lin Lu
- Department of Orthopedics Surgery, Renmin Hospital of Wuhan University, Wuhan, China
| | - Jiayao Zhang
- Department of Orthopedics Surgery, Fuzhou University Affiliated Provincial Hospital, Fuzhou, China.
| | - Pengran Liu
- Department of Orthopedics Surgery, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China.
- Laboratory of Intelligent Medicine, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China.
| | - Zhewei Ye
- Department of Orthopedics Surgery, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China.
- Laboratory of Intelligent Medicine, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China.
| |
Collapse
|
10
|
Binh LN, Nhu NT, Nhi PTU, Son DLH, Bach N, Huy HQ, Le NQK, Kang JH. Impact of deep learning on pediatric elbow fracture detection: a systematic review and meta-analysis. Eur J Trauma Emerg Surg 2025; 51:115. [PMID: 39976732 DOI: 10.1007/s00068-025-02779-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/28/2024] [Accepted: 01/25/2025] [Indexed: 05/10/2025]
Abstract
OBJECTIVES Pediatric elbow fractures are a common injury among children. Recent advancements in artificial intelligence (AI), particularly deep learning (DL), have shown promise in diagnosing these fractures. This study systematically evaluated the performance of DL models in detecting pediatric elbow fractures. MATERIALS AND METHODS A comprehensive search was conducted in PubMed (Medline), EMBASE, and IEEE Xplore for studies published up to October 20, 2023. Studies employing DL models for detecting elbow fractures in patients aged 0 to 16 years were included. Key performance metrics, including sensitivity, specificity, and area under the curve (AUC), were extracted. The study was registered in PROSPERO (ID: CRD42023470558). RESULTS The search identified 22 studies, of which six met the inclusion criteria for the meta-analysis. The pooled sensitivity of DL models for pediatric elbow fracture detection was 0.93 (95% CI: 0.91-0.96). Specificity values ranged from 0.84 to 0.92 across studies, with a pooled estimate of 0.89 (95% CI: 0.85-0.92). The AUC ranged from 0.91 to 0.99, with a pooled estimate of 0.95 (95% CI: 0.93-0.97). Further analysis highlighted the impact of preprocessing techniques and the choice of model backbone architecture on performance. CONCLUSION DL models demonstrate exceptional accuracy in detecting pediatric elbow fractures. For optimal performance, we recommend leveraging backbone architectures like ResNet, combined with manual preprocessing supervised by radiology and orthopedic experts.
Collapse
Affiliation(s)
- Le Nguyen Binh
- College of Medicine, Taipei Medical University, Taipei, 11031, Taiwan
- Department of Orthopedics and Trauma, Cho Ray Hospital, Ho Chi Minh City, Vietnam
- AIBioMed Research Group, Taipei Medical University, Taipei, 11031, Taiwan
- SBH Ortho Clinic, Ho Chi Minh City, Vietnam
| | - Nguyen Thanh Nhu
- College of Medicine, Taipei Medical University, Taipei, 11031, Taiwan
- Faculty of Medicine, Can Tho University of Medicine and Pharmacy, Can Tho, 94117, Vietnam
| | - Pham Thi Uyen Nhi
- Ho Chi Minh City Hospital of Dermato-Venereology, Ho Chi Minh City, Vietnam
| | - Do Le Hoang Son
- Department of Orthopedics and Trauma, Cho Ray Hospital, Ho Chi Minh City, Vietnam
- Faculty of Medicine, Can Tho University of Medicine and Pharmacy, Can Tho, 94117, Vietnam
| | - Nguyen Bach
- Faculty of Medicine, Can Tho University of Medicine and Pharmacy, Can Tho, 94117, Vietnam
- Department of Orthopedics, University Medical Center Ho Chi Minh City, 201 Nguyen Chi Thanh Street, District 5, Ho Chi Minh City, Vietnam
| | - Hoang Quoc Huy
- Faculty of Medicine, Can Tho University of Medicine and Pharmacy, Can Tho, 94117, Vietnam
- Department of Orthopedics, University Medical Center Ho Chi Minh City, 201 Nguyen Chi Thanh Street, District 5, Ho Chi Minh City, Vietnam
| | - Nguyen Quoc Khanh Le
- AIBioMed Research Group, Taipei Medical University, Taipei, 11031, Taiwan.
- In-Service Master Program in Artificial Intelligence in Medicine, College of Medicine, Taiwan and AIBioMed Research Group, Taipei Medical University, Taipei, 11031, Taiwan.
- Translational Imaging Research Center, Taipei Medical University Hospital, Taipei, 11031, Taiwan.
| | - Jiunn-Horng Kang
- College of Medicine, Taipei Medical University, Taipei, 11031, Taiwan.
- Department of Physical Medicine and Rehabilitation, School of Medicine, College of Medicine, Taipei Medical University, Taipei, 11031, Taiwan.
- Department of Physical Medicine and Rehabilitation, Taipei Medical University Hospital, Taipei, 11031, Taiwan.
- Graduate Institute of Nanomedicine and Medical Engineering, College of Biomedical Engineering, Taipei Medical University, 250 Wuxing Street, Xinyi District, Taipei, 11031, Taiwan.
| |
Collapse
|
11
|
Sharifi G, Hajibeygi R, Zamani SAM, Easa AM, Bahrami A, Eshraghi R, Moafi M, Ebrahimi MJ, Fathi M, Mirjafari A, Chan JS, Dixe de Oliveira Santo I, Anar MA, Rezaei O, Tu LH. Diagnostic performance of neural network algorithms in skull fracture detection on CT scans: a systematic review and meta-analysis. Emerg Radiol 2025; 32:97-111. [PMID: 39680295 DOI: 10.1007/s10140-024-02300-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2024] [Accepted: 11/08/2024] [Indexed: 12/17/2024]
Abstract
BACKGROUND AND AIM The potential intricacy of skull fractures as well as the complexity of underlying anatomy poses diagnostic hurdles for radiologists evaluating computed tomography (CT) scans. The necessity for automated diagnostic tools has been brought to light by the shortage of radiologists and the growing demand for rapid and accurate fracture diagnosis. Convolutional Neural Networks (CNNs) are a potential new class of medical imaging technologies that use deep learning (DL) to improve diagnosis accuracy. The objective of this systematic review and meta-analysis is to assess how well CNN models diagnose skull fractures on CT images. METHODS PubMed, Scopus, and Web of Science were searched for studies published before February 2024 that used CNN models to detect skull fractures on CT scans. Meta-analyses were conducted for area under the receiver operating characteristic curve (AUC), sensitivity, specificity, and accuracy. Egger's and Begg's tests were used to assess publication bias. RESULTS Meta-analysis was performed for 11 studies with 20,798 patients. Pooled average AUC for implementing pre-training for transfer learning in CNN models within their training model's architecture was 0.96 ± 0.02. The pooled averages of the studies' sensitivity and specificity were 1.0 and 0.93, respectively. The accuracy was obtained 0.92 ± 0.04. Studies showed heterogeneity, which was explained by differences in model topologies, training models, and validation techniques. There was no significant publication bias detected. CONCLUSION CNN models perform well in identifying skull fractures on CT scans. Although there is considerable heterogeneity and possibly publication bias, the results suggest that CNNs have the potential to improve diagnostic accuracy in the imaging of acute skull trauma. To further enhance these models' practical applicability, future studies could concentrate on the utility of DL models in prospective clinical trials.
Collapse
Affiliation(s)
- Guive Sharifi
- Skull base Research Center, Loghman Hakim Hospital, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Ramtin Hajibeygi
- Tehran University of Medical Sciences, School of Medicine, Tehran, Iran
| | | | - Ahmed Mohamedbaqer Easa
- Department of Radiology Technology, Collage of Health and Medical Technology, Al-Ayen Iraqi University, Thi-Qar, 64001, Iraq
| | | | | | - Maral Moafi
- Cell Biology and Anatomical Sciences, School of Medicine, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Mohammad Javad Ebrahimi
- Cell Biology and Anatomical Sciences, School of Medicine, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Mobina Fathi
- Skull base Research Center, Loghman Hakim Hospital, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Arshia Mirjafari
- Department of Radiological Sciences, University of California, Los Angeles, CA, USA
- College of Osteopathic Medicine of The Pacific, Western University of Health Sciences, Pomona, CA, USA
| | - Janine S Chan
- Keck School of Medicine of USC, Los Angeles, CA, USA
| | | | | | - Omidvar Rezaei
- Skull base Research Center, Loghman Hakim Hospital, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Long H Tu
- Department of Radiology and Biomedical Imaging, Yale School of Medicine, CT, USA.
| |
Collapse
|
12
|
Oude Nijhuis KD, Barvelink B, Prijs J, Zhao Y, Liao Z, Jaarsma RL, IJpma FFA, Colaris JW, Doornberg JN, Wijffels MME, Machine Learning Consortium. An open source convolutional neural network to detect and localize distal radius fractures on plain radiographs. Eur J Trauma Emerg Surg 2025; 51:26. [PMID: 39820574 PMCID: PMC11742337 DOI: 10.1007/s00068-024-02731-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Collaborators] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/19/2024] [Accepted: 09/29/2024] [Indexed: 01/19/2025]
Abstract
PURPOSE Distal radius fractures (DRFs) are often initially assessed by junior doctors under time constraints, with limited supervision, risking significant consequences if missed. Convolutional Neural Networks (CNNs) can aid in diagnosing fractures. This study aims to internally and externally validate an open source algorithm for the detection and localization of DRFs. METHODS Patients from a level 1 trauma center from Adelaide, Australia that presented between 2016 and 2020 with wrist trauma were retrospectively included. Radiographs were reviewed confirming the presence or absence of a fracture, as well as annotating radius, ulna, and fracture location. An internal validation dataset from the same hospital was created. An external validation set was created with two other level 1 trauma centers, from Groningen and Rotterdam, the Netherlands. Three surgeons reviewed both sets for DRFs. RESULTS The algorithm was trained on 659 radiographs. The internal validation set included 190 patients, showing an accuracy of 87% and an AUC of 0.93 for DRF detection. The external validation set consisted of 188 patients, with an accuracy and AUC were 82% and 0.88 respectively. Radial and ulnar bone segmentation on the internal validation was excellent with an AP50 of 99 and 98, but moderate for fracture segmentation with an AP50 of 29. For external validation the AP50 was 92, 89 and 25 for radius, ulna, and fracture respectively. CONCLUSION This open-source algorithm effectively detects DRFs with high accuracy and localizes them with moderate accuracy. It can assist clinicians in diagnosing suspected DRFs and is the first radiograph-based CNN externally validated on patients from multiple hospitals.
Collapse
Affiliation(s)
- Koen D Oude Nijhuis
- Department of Orthopaedic Surgery, University Medical Centre Groningen and Groningen University, Hanzeplein 1, 9713PZ, Groningen, the Netherlands.
- Department of Trauma Surgery, University Medical Centre Groningen and Groningen University, Hanzeplein 1, 9713PZ, Groningen, the Netherlands.
| | - Britt Barvelink
- Department of Orthopaedics and Sports Medicine, Erasmus University Medical Centre, Rotterdam, The Netherlands
| | - Jasper Prijs
- Department of Orthopaedic Surgery, University Medical Centre Groningen and Groningen University, Hanzeplein 1, 9713PZ, Groningen, the Netherlands
- Department of Trauma Surgery, University Medical Centre Groningen and Groningen University, Hanzeplein 1, 9713PZ, Groningen, the Netherlands
- Department of Orthopaedic Surgery, Flinders University and Medical Centre, Adelaide, South Australia, Australia
| | - Yang Zhao
- Australian Institute for Machine Learning, Adelaide, Australia
| | - Zhibin Liao
- Australian Institute for Machine Learning, Adelaide, Australia
| | - Ruurd L Jaarsma
- Department of Orthopaedic Surgery, Flinders University and Medical Centre, Adelaide, South Australia, Australia
| | - Frank F A IJpma
- Department of Trauma Surgery, University Medical Centre Groningen and Groningen University, Hanzeplein 1, 9713PZ, Groningen, the Netherlands
| | - Joost W Colaris
- Department of Orthopaedics and Sports Medicine, Erasmus University Medical Centre, Rotterdam, The Netherlands
| | - Job N Doornberg
- Department of Orthopaedic Surgery, University Medical Centre Groningen and Groningen University, Hanzeplein 1, 9713PZ, Groningen, the Netherlands
- Department of Orthopaedic Surgery, Flinders University and Medical Centre, Adelaide, South Australia, Australia
| | - Mathieu M E Wijffels
- Trauma Research Unit, Department of Surgery, Erasmus MC, University Medical Centre Rotterdam, Rotterdam, the Netherlands
| | | |
Collapse
Collaborators
Hans van Luit, Olga Canta, Sanne Hoeksema, Charlotte L E Laane, Kaan Aksakal, Haras Mhmud, Paul Jutte, Max Gordon, Wouter Mallee, Andrew Duckworth, Niels Schep, Ran Hendrickx, Anne-Eva Bulstra, Lente Dankelman, Max Reijman,
Collapse
|
13
|
Luan A, von Rabenau L, Serebrakian AT, Crowe CS, Do BH, Eberlin KR, Chang J, Pridgen BC. Machine Learning-Aided Diagnosis Enhances Human Detection of Perilunate Dislocations. Hand (N Y) 2025:15589447241308603. [PMID: 39815415 PMCID: PMC11736725 DOI: 10.1177/15589447241308603] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 01/18/2025]
Abstract
BACKGROUND Perilunate/lunate injuries are frequently misdiagnosed. We hypothesize that utilization of a machine learning algorithm can improve human detection of perilunate/lunate dislocations. METHODS Participants from emergency medicine, hand surgery, and radiology were asked to evaluate 30 lateral wrist radiographs for the presence of a perilunate/lunate dislocation with and without the use of a machine learning algorithm, which was used to label the lunate. Human performance with and without the machine learning tool was evaluated using sensitivity, specificity, accuracy, and F1 score. RESULTS A total of 137 participants were recruited, with 55 respondents from emergency medicine, 33 from radiology, and 49 from hand surgery. Thirty-nine participants were attending physicians or fellows, and 98 were residents. Use of the machine learning tool improved specificity from 88% to 94%, accuracy from 89% to 93%, and F1 score from 0.89 to 0.92. When stratified by training level, attending physicians and fellows had an improvement in specificity from 93% to 97%. For residents, use of the machine learning tool resulted in improved accuracy from 86% to 91% and specificity from 86% to 93%. The performance of surgery and radiology residents improved when assisted by the tool to achieve similar accuracy to attendings, and their assisted diagnostic performance reaches levels similar to that of the fully automated artificial intelligence tool. CONCLUSIONS Use of a machine learning tool improves resident accuracy for radiographic detection of perilunate dislocations, and improves specificity for all training levels. This may help to decrease misdiagnosis of perilunate dislocations, particularly when subspecialist evaluation is delayed.
Collapse
Affiliation(s)
- Anna Luan
- Stanford University, CA, USA
- Massachusetts General Hospital, Boston, USA
| | | | | | | | | | | | | | - Brian C. Pridgen
- University of Washington, Seattle, USA
- The Buncke Clinic, San Francisco, CA, USA
| |
Collapse
|
14
|
Zeng J, Zou F, Chen H, Liang D. Texture analysis combined with machine learning in radiographs of the knee joint: potential to identify tibial plateau occult fractures. Quant Imaging Med Surg 2025; 15:502-514. [PMID: 39838981 PMCID: PMC11744106 DOI: 10.21037/qims-24-799] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2024] [Accepted: 11/01/2024] [Indexed: 01/23/2025]
Abstract
Background Missed or delayed diagnosis of occult fractures of tibial plateau may cause adverse effects on patients. The objective of this study was to evaluate the diagnostic performance of texture analysis (TA) of knee joint radiographs combined with machine learning (ML) in identifying patients at risk of tibial plateau occult fractures. Methods A total of 169 patients with negative fracture on knee X-ray films from 2018 to 2022 who were diagnosed with occult tibial plateau fractures or no fractures by subsequent magnetic resonance imaging (MRI) examination were retrospectively enrolled. The X-ray images of the patient's knee joint were used for texture feature extraction. A total of 9 ML feature selection methods (including 6 mainstream methods and 3 methods provided by MaZda software) combined with 3 classification methods were used to build the best diagnostic model. The performance of each model was evaluated by accuracy, F1-value, and area under the curve (AUC). Results The least absolute shrinkage and selection operator (LASSO) method had the best performance of the 6 mainstream methods, with an accuracy of 0.81, an F1 value of 0.80, and an AUC of 0.920, all of which were higher than those of the other five methods (accuracy range: 0.65-0.80, F1 score range: 0.61-0.79, AUC range: 0.722-0.895). Among the three feature selection models in MaZda software, the most ideal method for accuracy measurement was the MI method, reaching 0.77. In the measurement of the F1 value and AUC, MaZda's best method was Fisher, reaching 0.78 and 0.888, respectively. All indicators were lower than those of the LASSO method. The combination of LASSO and support vector machine (SVM) yielded the best classification performance, while the performance of the combination of LASSO and logistic regression was slightly inferior, but the difference was not statistically significant. Conclusions TA of knee joint radiography combined with ML has achieved high performance in identifying patients at risk of occult fractures of the tibial plateau. Considering both the model performance and computational complexity, the LASSO feature selection method combined with the logistic regression classifier yielded the best classification performance in this process.
Collapse
Affiliation(s)
- Ju Zeng
- Department of Medical Imaging, Sichuan Orthopedic Hospital, Chengdu, China
| | - Fenghua Zou
- School of Management and Economics, University of Electronic Science and Technology of China, Chengdu, China
| | - Haoxi Chen
- School of Management and Economics, University of Electronic Science and Technology of China, Chengdu, China
| | - Decui Liang
- School of Management and Economics, University of Electronic Science and Technology of China, Chengdu, China
| |
Collapse
|
15
|
Bongrand P. Should Artificial Intelligence Play a Durable Role in Biomedical Research and Practice? Int J Mol Sci 2024; 25:13371. [PMID: 39769135 PMCID: PMC11676049 DOI: 10.3390/ijms252413371] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2024] [Revised: 11/26/2024] [Accepted: 12/09/2024] [Indexed: 01/11/2025] Open
Abstract
During the last decade, artificial intelligence (AI) was applied to nearly all domains of human activity, including scientific research. It is thus warranted to ask whether AI thinking should be durably involved in biomedical research. This problem was addressed by examining three complementary questions (i) What are the major barriers currently met by biomedical investigators? It is suggested that during the last 2 decades there was a shift towards a growing need to elucidate complex systems, and that this was not sufficiently fulfilled by previously successful methods such as theoretical modeling or computer simulation (ii) What is the potential of AI to meet the aforementioned need? it is suggested that recent AI methods are well-suited to perform classification and prediction tasks on multivariate systems, and possibly help in data interpretation, provided their efficiency is properly validated. (iii) Recent representative results obtained with machine learning suggest that AI efficiency may be comparable to that displayed by human operators. It is concluded that AI should durably play an important role in biomedical practice. Also, as already suggested in other scientific domains such as physics, combining AI with conventional methods might generate further progress and new applications, involving heuristic and data interpretation.
Collapse
Affiliation(s)
- Pierre Bongrand
- Laboratory Adhesion and Inflammation (LAI), Inserm UMR 1067, Cnrs Umr 7333, Aix-Marseille Université UM 61, 13009 Marseille, France
| |
Collapse
|
16
|
Xing P, Zhang L, Wang T, Wang L, Xing W, Wang W. A deep learning algorithm that aids visualization of femoral neck fractures and improves physician training. Injury 2024; 55:111997. [PMID: 39504732 DOI: 10.1016/j.injury.2024.111997] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/12/2024] [Revised: 09/26/2024] [Accepted: 10/26/2024] [Indexed: 11/08/2024]
Abstract
PURPOSE Missed fractures are the most common radiologic error in clinical practice, and erroneous classification could lead to inappropriate treatment and unfavorable prognosis. Here, we developed a fully automated deep learning model to detect and classify femoral neck fractures using plain radiographs, and evaluated its utility for diagnostic assistance and physician training. METHODS 1527 plain pelvic and hip radiographs obtained between April 2014 and July 2023 at our Hospital were selected for the model training and evaluation. Faster R-CNN was used to locate the femoral neck. DenseNet-121 was used for Garden classification of the femoral neck fracture, while an additional segmentation method used to visualize the probable fracture area. The model was assessed by the area under the receiver operating characteristic curve (AUC). The accuracy, sensitivity, and specificity for clinicians fracture detection in the diagnostic assistance and physician training experiments were determined. RESULTS The accuracy of the model for fracture detection was 94.1 %. The model achieved AUCs of 0.99 for no femoral neck fractures, 0.94 for Garden I/II fractures, and 0.99 for Garden III/IV fractures. In the diagnostic assistance study, the emergency physicians had an average accuracy of 86.33 % unaided and 92.03 % aided, sensitivity of 85.94 % unaided and 91.78 % aided, and specificity of 87.88 % unaided and 93.13 % aided in detecting fractures. In the physician training study, the accuracy, sensitivity, and specificity of the trainees for fracture classification were 81.83 %, 77.28 %, and 84.85 %, respectively, before training, compared with 90.65 %, 88.31 %, and 92.21 %, respectively, after training. CONCLUSIONS The model represents a valuable tool for physicians to better visualize fractures and improve training outcomes, indicating deep learning algorithms as a promising approach to improve clinical practice and medical education.
Collapse
Affiliation(s)
- Pengyi Xing
- Department of Radiology, The 989th Hospital of the PLA Joint Logistics Support Force, Luoyang, Henan Province, China
| | - Li Zhang
- Department of Gastroenterology and Endocrinology, The 989th Hospital of the PLA Joint Logistics Support Force, Luoyang, Henan Province, China
| | - Tiegong Wang
- Department of Orthopedics Trauma, Shanghai Changhai Hospital, Naval Military Medical University, Shanghai, China
| | - Lipeng Wang
- Department of Orthopedic Surgery, Zhongshan Hospital, Fudan University, Shanghai, China
| | - Wanting Xing
- Department of Radiology, The 989th Hospital of the PLA Joint Logistics Support Force, Luoyang, Henan Province, China
| | - Wei Wang
- Department of Radiology, The 989th Hospital of the PLA Joint Logistics Support Force, Luoyang, Henan Province, China; Department of Radiology, General hospital of Central Theater Command, Wuhan, Hubei Province, China.
| |
Collapse
|
17
|
Gan K, Liu Y, Zhang T, Xu D, Lian L, Luo Z, Li J, Lu L. Deep Learning Model for Automatic Identification and Classification of Distal Radius Fracture. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:2874-2882. [PMID: 38862852 PMCID: PMC11612100 DOI: 10.1007/s10278-024-01144-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/05/2024] [Revised: 05/13/2024] [Accepted: 05/15/2024] [Indexed: 06/13/2024]
Abstract
Distal radius fracture (DRF) is one of the most common types of wrist fractures. We aimed to construct a model for the automatic segmentation of wrist radiographs using a deep learning approach and further perform automatic identification and classification of DRF. A total of 2240 participants with anteroposterior wrist radiographs from one hospital between January 2015 and October 2021 were included. The outcomes were automatic segmentation of wrist radiographs, identification of DRF, and classification of DRF (type A, type B, type C). The Unet model and Fast-RCNN model were used for automatic segmentation. The DenseNet121 model and ResNet50 model were applied to DRF identification of DRF. The DenseNet121 model, ResNet50 model, VGG-19 model, and InceptionV3 model were used for DRF classification. The area under the curve (AUC) with 95% confidence interval (CI), accuracy, precision, and F1-score was utilized to assess the effectiveness of the identification and classification models. Of these 2240 participants, 1440 (64.3%) had DRF, of which 701 (48.7%) were type A, 278 (19.3%) were type B, and 461 (32.0%) were type C. Both the Unet model and the Fast-RCNN model showed good segmentation of wrist radiographs. For DRF identification, the AUCs of the DenseNet121 model and the ResNet50 model in the testing set were 0.941 (95%CI: 0.926-0.965) and 0.936 (95%CI: 0.913-0.955), respectively. The AUCs of the DenseNet121 model (testing set) for classification type A, type B, and type C were 0.96, 0.96, and 0.96, respectively. The DenseNet121 model may provide clinicians with a tool for interpreting wrist radiographs.
Collapse
Affiliation(s)
- Kaifeng Gan
- Department of Orthopaedics, the Affiliated LiHuiLi Hospital of Ningbo University, No. 57 Xingning Road, Yinzhou District, Ningbo, 315211, Zhejiang, China
| | - Yunpeng Liu
- Ningbo University of Technology, Ningbo, 315100, Zhejiang, China
| | - Ting Zhang
- Department of Orthopaedics, the Affiliated LiHuiLi Hospital of Ningbo University, No. 57 Xingning Road, Yinzhou District, Ningbo, 315211, Zhejiang, China
| | - Dingli Xu
- Health Science Center, Ningbo University, Ningbo, 315000, Zhejiang, China
| | - Leidong Lian
- Health Science Center, Ningbo University, Ningbo, 315000, Zhejiang, China
| | - Zhe Luo
- Health Science Center, Ningbo University, Ningbo, 315000, Zhejiang, China
| | - Jin Li
- Department of Orthopaedics, the Affiliated LiHuiLi Hospital of Ningbo University, No. 57 Xingning Road, Yinzhou District, Ningbo, 315211, Zhejiang, China
| | - Liangjie Lu
- Department of Orthopaedics, the Affiliated LiHuiLi Hospital of Ningbo University, No. 57 Xingning Road, Yinzhou District, Ningbo, 315211, Zhejiang, China.
| |
Collapse
|
18
|
Mao J, Du Y, Xue J, Hu J, Mai Q, Zhou T, Zhou Z. Automated detection and classification of mandibular fractures on multislice spiral computed tomography using modified convolutional neural networks. Oral Surg Oral Med Oral Pathol Oral Radiol 2024; 138:803-812. [PMID: 39384413 DOI: 10.1016/j.oooo.2024.07.010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/13/2023] [Revised: 07/19/2024] [Accepted: 07/26/2024] [Indexed: 10/11/2024]
Abstract
OBJECTIVE To evaluate the performance of convolutional neural networks (CNNs) for the automated detection and classification of mandibular fractures on multislice spiral computed tomography (MSCT). STUDY DESIGN MSCT data from 361 patients with mandibular fractures were retrospectively collected. Two experienced maxillofacial surgeons annotated the images as ground truth. Fractures were detected utilizing the following models: YOLOv3, YOLOv4, Faster R-CNN, CenterNet, and YOLOv5-TRS. Fracture sites were classified by the following models: AlexNet, GoogLeNet, ResNet50, original DenseNet-121, and modified DenseNet-121. The performance was evaluated for accuracy, sensitivity, specificity, and area under the curve (AUC). AUC values were compared using the Z-test and P values <.05 were considered to be statistically significant. RESULTS Of all of the detection models, YOLOv5-TRS obtained the greatest mean accuracy (96.68%). Among all of the fracture subregions, body fractures were the most reliably detected (with accuracies of 88.59%-99.01%). For classification models, the AUCs for body fractures were higher than those of condyle and angle fractures, and they were all above 0.75, with the highest AUC at 0.903. Modified DenseNet-121 had the best overall classification performance with a mean AUC of 0.814. CONCLUSIONS The modified CNN-based models demonstrated high reliability for the diagnosis of mandibular fractures on MSCT.
Collapse
Affiliation(s)
- Jingjing Mao
- Ningxia Medical University, Ningxia Key Laboratory of Oral Disease Research, Yinchuan, P.R. China
| | - Yuhu Du
- College of Computer Science and Engineering, North Minzu University, Yinchuan, P.R. China
| | - Jiawen Xue
- Ningxia Medical University, Ningxia Key Laboratory of Oral Disease Research, Yinchuan, P.R. China
| | - Jingjing Hu
- Department of Oral and Maxillofacial Surgery, Guyuan People's Hospital, Guyuan, P.R. China
| | - Qian Mai
- Department of Stomatology, The First People's Hospital of Yinchuan, Yinchuan, P.R. China
| | - Tao Zhou
- College of Computer Science and Engineering, North Minzu University, Yinchuan, P.R. China
| | - Zhongwei Zhou
- Department of Oral and Maxillofacial Surgery, General Hospital of Ningxia Medical University, Yinchuan, P.R. China; Institution of Medical Sciences, General Hospital of Ningxia Medical University, Yinchuan, P.R. China.
| |
Collapse
|
19
|
Ibanez V, Jucker D, Ebert LC, Franckenberg S, Dobay A. Classification of rib fracture types from postmortem computed tomography images using deep learning. Forensic Sci Med Pathol 2024; 20:1208-1214. [PMID: 37968549 PMCID: PMC11790768 DOI: 10.1007/s12024-023-00751-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/05/2023] [Indexed: 11/17/2023]
Abstract
Human or time resources can sometimes fall short in medical image diagnostics, and analyzing images in full detail can be a challenging task. With recent advances in artificial intelligence, an increasing number of systems have been developed to assist clinicians in their work. In this study, the objective was to train a model that can distinguish between various fracture types on different levels of hierarchical taxonomy and detect them on 2D-image representations of volumetric postmortem computed tomography (PMCT) data. We used a deep learning model based on the ResNet50 architecture that was pretrained on ImageNet data, and we used transfer learning to fine-tune it to our specific task. We trained our model to distinguish between "displaced," "nondisplaced," "ad latus," "ad longitudinem cum contractione," and "ad longitudinem cum distractione" fractures. Radiographs with no fractures were correctly predicted in 95-99% of cases. Nondisplaced fractures were correctly predicted in 80-86% of cases. Displaced fractures of the "ad latus" type were correctly predicted in 17-18% of cases. The other two displaced types of fractures, "ad longitudinem cum contractione" and "ad longitudinem cum distractione," were correctly predicted in 70-75% and 64-75% of cases, respectively. The model achieved the best performance when the level of hierarchical taxonomy was high, while it had more difficulties when the level of hierarchical taxonomy was lower. Overall, deep learning techniques constitute a reliable solution for forensic pathologists and medical practitioners seeking to reduce workload.
Collapse
Affiliation(s)
- Victor Ibanez
- Forensic Machine Learning Technology Center, Zurich Institute of Forensic Medicine, University of Zurich, Winterthurerstrasse 190/52, CH-8057, Zurich, Switzerland
| | - Dario Jucker
- Zurich Institute of Forensic Medicine, 3D Centre Zurich, University of Zurich, Winterthurerstrasse 190/52, CH-8057, Zurich, Switzerland
| | - Lars C Ebert
- Zurich Institute of Forensic Medicine, 3D Centre Zurich, University of Zurich, Winterthurerstrasse 190/52, CH-8057, Zurich, Switzerland
| | - Sabine Franckenberg
- Diagnostic and Interventional Radiology, University Hospital Zurich, Rämistrasse 100, 8091, Zurich, Switzerland
- Zurich Institute of Forensic Medicine, 3D Centre Zurich, University of Zurich, Winterthurerstrasse 190/52, CH-8057, Zurich, Switzerland
| | - Akos Dobay
- Forensic Machine Learning Technology Center, Zurich Institute of Forensic Medicine, University of Zurich, Winterthurerstrasse 190/52, CH-8057, Zurich, Switzerland.
| |
Collapse
|
20
|
Oude Nijhuis KD, Dankelman LHM, Wiersma JP, Barvelink B, IJpma FFA, Verhofstad MHJ, Doornberg JN, Colaris JW, Wijffels MME. AI for detection, classification and prediction of loss of alignment of distal radius fractures; a systematic review. Eur J Trauma Emerg Surg 2024; 50:2819-2831. [PMID: 38981869 PMCID: PMC11666746 DOI: 10.1007/s00068-024-02557-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/05/2024] [Accepted: 05/14/2024] [Indexed: 07/11/2024]
Abstract
PURPOSE Early and accurate assessment of distal radius fractures (DRFs) is crucial for optimal prognosis. Identifying fractures likely to lose threshold alignment (instability) in a cast is vital for treatment decisions, yet prediction tools' accuracy and reliability remain challenging. Artificial intelligence (AI), particularly Convolutional Neural Networks (CNNs), can evaluate radiographic images with high performance. This systematic review aims to summarize studies utilizing CNNs to detect, classify, or predict loss of threshold alignment of DRFs. METHODS A literature search was performed according to the PRISMA. Studies were eligible when the use of AI for the detection, classification, or prediction of loss of threshold alignment was analyzed. Quality assessment was done with a modified version of the methodologic index for non-randomized studies (MINORS). RESULTS Of the 576 identified studies, 15 were included. On fracture detection, studies reported sensitivity and specificity ranging from 80 to 99% and 73-100%, respectively; the AUC ranged from 0.87 to 0.99; the accuracy varied from 82 to 99%. The accuracy of fracture classification ranged from 60 to 81% and the AUC from 0.59 to 0.84. No studies focused on predicting loss of thresholds alignement of DRFs. CONCLUSION AI models for DRF detection show promising performance, indicating the potential of algorithms to assist clinicians in the assessment of radiographs. In addition, AI models showed similar performance compared to clinicians. No algorithms for predicting the loss of threshold alignment were identified in our literature search despite the clinical relevance of such algorithms.
Collapse
Affiliation(s)
- Koen D Oude Nijhuis
- Department of Orthopedic Surgery, Groningen, Groningen University Medical Centre, Groningen, The Netherlands.
- Department of Surgery, Groningen, University Medical Centre, Groningen, The Netherlands.
| | - Lente H M Dankelman
- Trauma Research Unit Department of Surgery, Erasmus MC, University Medical Center Rotterdam, P.O. Box 2040, Rotterdam, 3000 CA, The Netherlands.
- Department of Orthopedic Surgery, Hand and Arm Center, Massachusetts General Hospital, Boston MA, Harvard Medical School, Boston MA, The Netherlands.
| | - Jort P Wiersma
- Department of Orthopedic Surgery, Groningen, Groningen University Medical Centre, Groningen, The Netherlands
- University Medical Center, Utrecht, The Netherlands
| | - Britt Barvelink
- Department of Orthopedics and Sports Medicine, Erasmus University Medical Centre, Rotterdam, The Netherlands
| | - Frank F A IJpma
- Department of Surgery, Groningen, University Medical Centre, Groningen, The Netherlands
| | - Michael H J Verhofstad
- Trauma Research Unit Department of Surgery, Erasmus MC, University Medical Center Rotterdam, P.O. Box 2040, Rotterdam, 3000 CA, The Netherlands
| | - Job N Doornberg
- Department of Orthopedic Surgery, Groningen, Groningen University Medical Centre, Groningen, The Netherlands
- Department of Surgery, Groningen, University Medical Centre, Groningen, The Netherlands
- Department of Orthopaedic and Trauma Surgery, Flinders University and Flinders Medical Centre, Adelaide, Australia
| | - Joost W Colaris
- Department of Orthopedics and Sports Medicine, Erasmus University Medical Centre, Rotterdam, The Netherlands
| | - Mathieu M E Wijffels
- Trauma Research Unit Department of Surgery, Erasmus MC, University Medical Center Rotterdam, P.O. Box 2040, Rotterdam, 3000 CA, The Netherlands
| |
Collapse
|
21
|
Zhang JY, Yang JM, Wang XM, Wang HL, Zhou H, Yan ZN, Xie Y, Liu PR, Hao ZW, Ye ZW. Application and Prospects of Deep Learning Technology in Fracture Diagnosis. Curr Med Sci 2024; 44:1132-1140. [PMID: 39551854 DOI: 10.1007/s11596-024-2928-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/23/2024] [Accepted: 08/18/2024] [Indexed: 11/19/2024]
Abstract
Artificial intelligence (AI) is an interdisciplinary field that combines computer technology, mathematics, and several other fields. Recently, with the rapid development of machine learning (ML) and deep learning (DL), significant progress has been made in the field of AI. As one of the fastest-growing branches, DL can effectively extract features from big data and optimize the performance of various tasks. Moreover, with advancements in digital imaging technology, DL has become a key tool for processing high-dimensional medical image data and conducting medical image analysis in clinical applications. With the development of this technology, the diagnosis of orthopedic diseases has undergone significant changes. In this review, we describe recent research progress on DL in fracture diagnosis and discuss the value of DL in this field, providing a reference for better integration and development of DL technology in orthopedics.
Collapse
Affiliation(s)
- Jia-Yao Zhang
- Department of Orthopedics, Fuzhou University Affiliated Provincial Hospital, Fuzhou, 350013, China
- Department of Orthopedics, Fujian Provincial Hospital, Fuzhou, 350013, China
- Intelligent Medical Laboratory, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China
| | - Jia-Ming Yang
- Intelligent Medical Laboratory, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China
- Department of Orthopedics, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China
| | - Xin-Meng Wang
- Department of Biochemistry and Molecular Biology, School of Basic Medical Sciences, Dali University, Dali, 671000, China
| | - Hong-Lin Wang
- Intelligent Medical Laboratory, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China
- Department of Orthopedics, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China
| | - Hong Zhou
- Intelligent Medical Laboratory, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China
- Department of Orthopedics, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China
| | - Zi-Neng Yan
- Intelligent Medical Laboratory, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China
- Department of Orthopedics, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China
| | - Yi Xie
- Intelligent Medical Laboratory, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China
- Department of Orthopedics, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China
| | - Peng-Ran Liu
- Intelligent Medical Laboratory, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China.
- Department of Orthopedics, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China.
| | - Zhi-Wei Hao
- School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan, 430074, China.
| | - Zhe-Wei Ye
- Intelligent Medical Laboratory, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China.
- Department of Orthopedics, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, 430022, China.
| |
Collapse
|
22
|
Ennab M, Mcheick H. Enhancing interpretability and accuracy of AI models in healthcare: a comprehensive review on challenges and future directions. Front Robot AI 2024; 11:1444763. [PMID: 39677978 PMCID: PMC11638409 DOI: 10.3389/frobt.2024.1444763] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2024] [Accepted: 09/27/2024] [Indexed: 12/17/2024] Open
Abstract
Artificial Intelligence (AI) has demonstrated exceptional performance in automating critical healthcare tasks, such as diagnostic imaging analysis and predictive modeling, often surpassing human capabilities. The integration of AI in healthcare promises substantial improvements in patient outcomes, including faster diagnosis and personalized treatment plans. However, AI models frequently lack interpretability, leading to significant challenges concerning their performance and generalizability across diverse patient populations. These opaque AI technologies raise serious patient safety concerns, as non-interpretable models can result in improper treatment decisions due to misinterpretations by healthcare providers. Our systematic review explores various AI applications in healthcare, focusing on the critical assessment of model interpretability and accuracy. We identify and elucidate the most significant limitations of current AI systems, such as the black-box nature of deep learning models and the variability in performance across different clinical settings. By addressing these challenges, our objective is to provide healthcare providers with well-informed strategies to develop innovative and safe AI solutions. This review aims to ensure that future AI implementations in healthcare not only enhance performance but also maintain transparency and patient safety.
Collapse
|
23
|
Luan A, Maan Z, Lin KY, Yao J. Application of Machine Learning to Osteoporosis and Osteopenia Screening Using Hand Radiographs. J Hand Surg Am 2024:S0363-5023(24)00432-5. [PMID: 39556066 DOI: 10.1016/j.jhsa.2024.09.008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/06/2024] [Revised: 08/15/2024] [Accepted: 09/10/2024] [Indexed: 11/19/2024]
Abstract
PURPOSE Fragility fractures associated with osteoporosis and osteopenia are a common cause of morbidity and mortality. Current methods of diagnosing low bone mineral density require specialized dual x-ray absorptiometry (DXA) scans. Plain hand radiographs may have utility as an alternative screening tool, although optimal diagnostic radiographic parameters are unknown, and measurement is prone to human error. The aim of the present study was to develop and validate an artificial intelligence algorithm to screen for osteoporosis and osteopenia using standard hand radiographs. METHODS A cohort of patients with both a DXA scan and a plain hand radiograph within 12 months of one another was identified. Hand radiographs were labeled as normal, osteopenia, or osteoporosis based on corresponding DXA hip T-scores. A deep learning algorithm was developed using the ResNet-50 framework and trained to predict the presence of osteoporosis or osteopenia on hand radiographs using labeled images. The results from the algorithm were validated using a separate balanced validation set, with the calculation of sensitivity, specificity, accuracy, and receiver operating characteristic curve using definitions from corresponding DXA scans as the reference standard. RESULTS There was a total of 687 images in the normal category, 607 images in the osteopenia category, and 130 images in the osteoporosis category for a total of 1,424 images. When predicting low bone density (osteopenia or osteoporosis) versus normal bone density, sensitivity was 88.5%, specificity was 65.4%, overall accuracy was 80.8%, and the area under the curve was 0.891, at the standard threshold of 0.5. If optimizing for both sensitivity and specificity, at a threshold of 0.655, the model achieved a sensitivity of 84.6% at a specificity of 84.6%. CONCLUSIONS The findings represent a possible step toward more accessible, cost-effective, automated diagnosis and therefore earlier treatment of osteoporosis/osteopenia. TYPE OF STUDY/LEVEL OF EVIDENCE Diagnostic II.
Collapse
Affiliation(s)
- Anna Luan
- Division of Plastic and Reconstructive Surgery, Department of Surgery, Stanford University School of Medicine, Stanford, CA
| | - Zeshaan Maan
- Division of Plastic and Reconstructive Surgery, Department of Surgery, Stanford University School of Medicine, Stanford, CA
| | - Kun-Yi Lin
- Department of Orthopedics, Tri-Service General Hospital, National Defense Medical Center, Taipei, Taiwan, Republic of China
| | - Jeffrey Yao
- Department of Orthopaedic Surgery, Stanford University School of Medicine, Stanford, CA; Robert A. Chase Hand and Upper Limb Center, Department of Orthopaedic Surgery, Stanford University Medical Center, Redwood City, CA.
| |
Collapse
|
24
|
Zhao T, Meng X, Wang Z, Hu Y, Fan H, Han J, Zhu N, Niu F. Diagnostic evaluation of blunt chest trauma by imaging-based application of artificial intelligence. Am J Emerg Med 2024; 85:35-43. [PMID: 39213808 DOI: 10.1016/j.ajem.2024.08.019] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/19/2024] [Accepted: 08/12/2024] [Indexed: 09/04/2024] Open
Abstract
Artificial intelligence (AI) is becoming increasingly integral in clinical practice, such as during imaging tasks associated with the diagnosis and evaluation of blunt chest trauma (BCT). Due to significant advances in imaging-based deep learning, recent studies have demonstrated the efficacy of AI in the diagnosis of BCT, with a focus on rib fractures, pulmonary contusion, hemopneumothorax and others, demonstrating significant clinical progress. However, the complicated nature of BCT presents challenges in providing a comprehensive diagnosis and prognostic evaluation, and current deep learning research concentrates on specific clinical contexts, limiting its utility in addressing BCT intricacies. Here, we provide a review of the available evidence surrounding the potential utility of AI in BCT, and additionally identify the challenges impeding its development. This review offers insights on how to optimize the role of AI in the diagnostic evaluation of BCT, which can ultimately enhance patient care and outcomes in this critical clinical domain.
Collapse
Affiliation(s)
- Tingting Zhao
- The Department of Radiology, Tianjin University Tianjin Hospital, 406 Jiefang Southern Road, Tianjin, China; Graduate School, Tianjin University, Tianjin, China
| | - Xianghong Meng
- The Department of Radiology, Tianjin University Tianjin Hospital, 406 Jiefang Southern Road, Tianjin, China; Graduate School, Tianjin University, Tianjin, China.
| | - Zhi Wang
- The Department of Radiology, Tianjin University Tianjin Hospital, 406 Jiefang Southern Road, Tianjin, China; Graduate School, Tianjin University, Tianjin, China.
| | - Yongcheng Hu
- The Department of Radiology, Tianjin University Tianjin Hospital, 406 Jiefang Southern Road, Tianjin, China
| | - Hongxing Fan
- The Department of Radiology, Tianjin University Tianjin Hospital, 406 Jiefang Southern Road, Tianjin, China; Graduate School, Tianjin Medical University, Tianjin, China
| | - Jun Han
- The Department of Radiology, Tianjin University Tianjin Hospital, 406 Jiefang Southern Road, Tianjin, China; Graduate School, Tianjin University, Tianjin, China
| | - Nana Zhu
- The Department of Radiology, Tianjin University Tianjin Hospital, 406 Jiefang Southern Road, Tianjin, China; Graduate School, Tianjin Medical University, Tianjin, China
| | - Feige Niu
- The Department of Radiology, Tianjin University Tianjin Hospital, 406 Jiefang Southern Road, Tianjin, China; Graduate School, Tianjin Medical University, Tianjin, China
| |
Collapse
|
25
|
Mohseni A, Ghotbi E, Kazemi F, Shababi A, Jahan SC, Mohseni A, Shababi N. Artificial Intelligence in Radiology: What Is Its True Role at Present, and Where Is the Evidence? Radiol Clin North Am 2024; 62:935-947. [PMID: 39393852 DOI: 10.1016/j.rcl.2024.03.008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/13/2024]
Abstract
The integration of artificial intelligence (AI) in radiology has brought about substantial advancements and transformative potential in diagnostic imaging practices. This study presents an overview of the current research on the application of AI in radiology, highlighting key insights from recent studies and surveys. These recent studies have explored the expected impact of AI, encompassing machine learning and deep learning, on the work volume of diagnostic radiologists. The present and future role of AI in radiology holds great promise for enhancing diagnostic capabilities, improving workflow efficiency, and ultimately, advancing patient care.
Collapse
Affiliation(s)
- Alireza Mohseni
- Johns Hopkins University School of Medicine, 600 N. Wolfe Street / Phipps 446, Baltimore, MD 21287, USA.
| | - Elena Ghotbi
- Johns Hopkins University School of Medicine, 600 N. Wolfe Street / Phipps 446, Baltimore, MD 21287, USA
| | - Foad Kazemi
- Johns Hopkins University School of Medicine, 600 N. Wolfe Street / Phipps 446, Baltimore, MD 21287, USA
| | - Amirali Shababi
- School of Medicine, Iran University of Medical Sciences, Hemat Highway next to Milad Tower 14535, Tehran, Iran
| | - Shayan Chashm Jahan
- Department of Computer Science, University of Maryland, 8125 Paint Branch Drive College Park, MD 20742, USA
| | - Anita Mohseni
- Azad University Tehran Medical Branch, Danesh, Shariati Street, Tehran, Iran 19395/1495
| | - Niloufar Shababi
- Johns Hopkins University School of Medicine, 600 N. Wolfe Street / Phipps 446, Baltimore, MD 21287, USA
| |
Collapse
|
26
|
Broggi G, Mazzucchelli M, Salzano S, Barbagallo GMV, Certo F, Zanelli M, Palicelli A, Zizzo M, Koufopoulos N, Magro G, Caltabiano R. The emerging role of artificial intelligence in neuropathology: Where are we and where do we want to go? Pathol Res Pract 2024; 263:155671. [PMID: 39490225 DOI: 10.1016/j.prp.2024.155671] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/21/2024] [Revised: 09/11/2024] [Accepted: 10/20/2024] [Indexed: 11/05/2024]
Abstract
The field of neuropathology, a subspecialty of pathology which studies the diseases affecting the nervous system, is experiencing significant changes due to advancements in artificial intelligence (AI). Traditionally reliant on histological methods and clinical correlations, neuropathology is now experiencing a revolution due to the development of AI technologies like machine learning (ML) and deep learning (DL). These technologies enhance diagnostic accuracy, optimize workflows, and enable personalized treatment strategies. AI algorithms excel at analyzing histopathological images, often revealing subtle morphological changes missed by conventional methods. For example, deep learning models applied to digital pathology can effectively differentiate tumor grades and detect rare pathologies, leading to earlier and more precise diagnoses. Progress in neuroimaging is another helpful tool of AI, as enhanced analysis of MRI and CT scans supports early detection of neurodegenerative diseases. By identifying biomarkers and progression patterns, AI aids in timely therapeutic interventions, potentially slowing disease progression. In molecular pathology, AI's ability to analyze complex genomic data helps uncover the genetic and molecular basis of neuropathological conditions, facilitating personalized treatment plans. AI-driven automation streamlines routine diagnostic tasks, allowing pathologists to focus on complex cases, especially in settings with limited resources. This review explores AI's integration into neuropathology, highlighting its current applications, benefits, challenges, and future directions.
Collapse
Affiliation(s)
- Giuseppe Broggi
- Department of Medical and Surgical Sciences and Advanced Technologies "G.F. Ingrassia", Anatomic Pathology, University of Catania, Catania 95123, Italy.
| | - Manuel Mazzucchelli
- Department of Medical and Surgical Sciences and Advanced Technologies "G.F. Ingrassia", Anatomic Pathology, University of Catania, Catania 95123, Italy
| | - Serena Salzano
- Department of Medical and Surgical Sciences and Advanced Technologies "G.F. Ingrassia", Anatomic Pathology, University of Catania, Catania 95123, Italy
| | | | - Francesco Certo
- Department of Neurological Surgery, Policlinico "G. Rodolico-S. Marco" University Hospital, Catania 95121, Italy
| | - Magda Zanelli
- Pathology Unit, Azienda USL-IRCCS di Reggio Emilia, Reggio Emilia 42123, Italy
| | - Andrea Palicelli
- Pathology Unit, Azienda USL-IRCCS di Reggio Emilia, Reggio Emilia 42123, Italy
| | - Maurizio Zizzo
- Surgical Oncology Unit, Azienda USL-IRCCS di Reggio Emilia, Reggio Emilia 42123, Italy
| | - Nektarios Koufopoulos
- Second Department of Pathology, Medical School, National and Kapodistrian University of Athens, Attikon University Hospital, Athens 15772, Greece
| | - Gaetano Magro
- Department of Medical and Surgical Sciences and Advanced Technologies "G.F. Ingrassia", Anatomic Pathology, University of Catania, Catania 95123, Italy
| | - Rosario Caltabiano
- Department of Medical and Surgical Sciences and Advanced Technologies "G.F. Ingrassia", Anatomic Pathology, University of Catania, Catania 95123, Italy
| |
Collapse
|
27
|
Yang TH, Sun YN, Li RS, Horng MH. The Detection and Classification of Scaphoid Fractures in Radiograph by Using a Convolutional Neural Network. Diagnostics (Basel) 2024; 14:2425. [PMID: 39518391 PMCID: PMC11545356 DOI: 10.3390/diagnostics14212425] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/16/2024] [Revised: 10/13/2024] [Accepted: 10/23/2024] [Indexed: 11/16/2024] Open
Abstract
OBJECTIVE Scaphoid fractures, particularly occult and non-displaced fractures, are difficult to detect using traditional X-ray methods because of their subtle appearance and variability in bone density. This study proposes a two-stage CNN approach to detect and classify scaphoid fractures using anterior-posterior (AP) and lateral (LA) X-ray views for more accurate diagnosis. METHODS This study emphasizes the use of multi-view X-ray images (AP and LA views) to improve fracture detection and classification. The multi-view fusion module helps integrate information from both views to enhance detection accuracy, particularly for occult fractures that may not be visible in a single view. The proposed method includes two stages, which are stage 1: detect the scaphoid bone using Faster RCNN and a Feature Pyramid Network (FPN) for region proposal and small object detection. The detection accuracy for scaphoid localization is 100%, with Intersection over Union (IoU) scores of 0.8662 for AP views and 0.8478 for LA views. And stage 2: perform fracture classification using a ResNet backbone and FPN combined with a multi-view fusion module to combine features from both AP and LA views. This stage achieves a classification accuracy of 89.94%, recall of 87.33%, and precision of 90.36%. RESULTS The proposed model performs well in both scaphoid bone detection and fracture classification. The multi-view fusion approach significantly improves recall and accuracy in detecting fractures compared to single-view approaches. In scaphoid detection, both AP and LA views achieved 100% detection accuracy. In fracture detection, using multi-view fusion, the accuracy for AP views reached 87.16%, and for LA views, it reached 83.83%. CONCLUSIONS The multi-view fusion model effectively improves the detection of scaphoid fractures, particularly in cases of occult and non-displaced fractures. The model provides a reliable, automated approach to assist clinicians in detecting and diagnosing scaphoid fractures more efficiently.
Collapse
Affiliation(s)
- Tai-Hua Yang
- Department of Biomedical Engineering, National Cheng Kung University, Tainan 701, Taiwan
- Department of Orthopedic Surgery, College of Medicine, National Cheng Kung University Hospital, National Cheng Kung University, Tainan 704, Taiwan
| | - Yung-Nien Sun
- Department of Computer Science and Information Engineering, National Cheng Kung University, Tainan 701, Taiwan
| | - Rong-Shiang Li
- Department of Computer Science and Information Engineering, National Cheng Kung University, Tainan 701, Taiwan
| | - Ming-Huwi Horng
- Department of Computer Science and Information Engineering, National Cheng Kung University, Tainan 701, Taiwan
| |
Collapse
|
28
|
Anderson PG, Tarder-Stoll H, Alpaslan M, Keathley N, Levin DL, Venkatesh S, Bartel E, Sicular S, Howell S, Lindsey RV, Jones RM. Deep learning improves physician accuracy in the comprehensive detection of abnormalities on chest X-rays. Sci Rep 2024; 14:25151. [PMID: 39448764 PMCID: PMC11502915 DOI: 10.1038/s41598-024-76608-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2024] [Accepted: 10/15/2024] [Indexed: 10/26/2024] Open
Abstract
Chest X-rays are the most commonly performed medical imaging exam, yet they are often misinterpreted by physicians. Here, we present an FDA-cleared, artificial intelligence (AI) system which uses a deep learning algorithm to assist physicians in the comprehensive detection and localization of abnormalities on chest X-rays. We trained and tested the AI system on a large dataset, assessed generalizability on publicly available data, and evaluated radiologist and non-radiologist physician accuracy when unaided and aided by the AI system. The AI system accurately detected chest X-ray abnormalities (AUC: 0.976, 95% bootstrap CI: 0.975, 0.976) and generalized to a publicly available dataset (AUC: 0.975, 95% bootstrap CI: 0.971, 0.978). Physicians showed significant improvements in detecting abnormalities on chest X-rays when aided by the AI system compared to when unaided (difference in AUC: 0.101, p < 0.001). Non-radiologist physicians detected abnormalities on chest X-ray exams as accurately as radiologists when aided by the AI system and were faster at evaluating chest X-rays when aided compared to unaided. Together, these results show that the AI system is accurate and reduces physician errors in chest X-ray evaluation, which highlights the potential of AI systems to improve access to fast, high-quality radiograph interpretation.
Collapse
Affiliation(s)
- Pamela G Anderson
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA.
| | | | - Mehmet Alpaslan
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
| | - Nora Keathley
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
| | - David L Levin
- Department of Radiology, Stanford University School of Medicine, 453 Quarry Rd, Palo Alto, CA, 94305, USA
| | - Srivas Venkatesh
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
| | - Elliot Bartel
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
| | - Serge Sicular
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
- The Mount Sinai Hospital, 1 Gustave L. Levy Place, New York, NY, 10029, USA
| | - Scott Howell
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
| | - Robert V Lindsey
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
| | - Rebecca M Jones
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
| |
Collapse
|
29
|
Nally DM, Kearns EC, Dalli J, Moynagh N, Hanley K, Neary P, Cahill RA. Patient public perspectives on digital colorectal cancer surgery (DALLAS). EUROPEAN JOURNAL OF SURGICAL ONCOLOGY 2024:108705. [PMID: 39532576 DOI: 10.1016/j.ejso.2024.108705] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2024] [Revised: 09/15/2024] [Accepted: 09/23/2024] [Indexed: 11/16/2024]
Abstract
INTRODUCTION The importance of patient perspectives is increasingly appreciated in clinical practice and academia with formal engagement processes developing worldwide. Digital surgery encompasses intraoperative patient data (including surgical video) analysis and so requires public-patient involvement (PPI). METHODS Engagement events were conducted based on NIHR and GRIPP2 LF guidelines. Following informative talks on digital surgery, invited patients and patient relatives were split into focus groups regarding 1) Research; 2) Data; 3) Industry Involvement; and 4) Artificial Intelligence in surgery. Scribed feedback was thematically analysed by two researchers independently. A pre and post event survey was sought voluntarily. RESULTS 36 participant perspectives were analysed. In general, patients were enthusiastic about having a voice in surgical research and sharing their journey, with most groups concluding that capturing this was most appropriate after treatment recovery. The use of patient data for surgical development (i.e. research and education) was endorsed unanimously for the purpose of future patient benefit when responsibly and transparently managed and the value of industry was acknowledged. From 30 pre/post surveys (all p > 0.05), participants afforded the greatest data (including video) ownership claim to the surgical team (52 %/48 %) versus patients (32 %/24 %) and the hospital (12 %/24 %). While most (73 %/80 %) agreed that AI should be applied in surgical care, most felt the surgeon most valuable (93 %/80 %) with participants disagreeing that AI should make diagnoses (57 %/64 %) or treat patients (70 %/70 %) without human input. CONCLUSION Patients capably represent stable views and expectations that can strengthen modern and evolving surgical development involving data privacy, ownership and management.
Collapse
Affiliation(s)
- Deirdre M Nally
- UCD Centre for Precision Surgery, University College Dublin, Ireland
| | - Emma C Kearns
- UCD Centre for Precision Surgery, University College Dublin, Ireland
| | - Jeffrey Dalli
- UCD Centre for Precision Surgery, University College Dublin, Ireland
| | - Niamh Moynagh
- UCD Centre for Precision Surgery, University College Dublin, Ireland
| | - Kate Hanley
- Department of Surgery, Mater Misericordiae University Hospital, Dublin, Ireland
| | - Peter Neary
- Department of Academic Surgery, University College Cork, Cork, Ireland; Department of Colorectal Surgery, University Hospital Waterford, Waterford City, Ireland
| | - Ronan A Cahill
- UCD Centre for Precision Surgery, University College Dublin, Ireland; Department of Surgery, Mater Misericordiae University Hospital, Dublin, Ireland.
| |
Collapse
|
30
|
Breu R, Avelar C, Bertalan Z, Grillari J, Redl H, Ljuhar R, Quadlbauer S, Hausner T. Artificial intelligence in traumatology. Bone Joint Res 2024; 13:588-595. [PMID: 39417424 PMCID: PMC11484119 DOI: 10.1302/2046-3758.1310.bjr-2023-0275.r3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 10/19/2024] Open
Abstract
Aims The aim of this study was to create artificial intelligence (AI) software with the purpose of providing a second opinion to physicians to support distal radius fracture (DRF) detection, and to compare the accuracy of fracture detection of physicians with and without software support. Methods The dataset consisted of 26,121 anonymized anterior-posterior (AP) and lateral standard view radiographs of the wrist, with and without DRF. The convolutional neural network (CNN) model was trained to detect the presence of a DRF by comparing the radiographs containing a fracture to the inconspicuous ones. A total of 11 physicians (six surgeons in training and five hand surgeons) assessed 200 pairs of randomly selected digital radiographs of the wrist (AP and lateral) for the presence of a DRF. The same images were first evaluated without, and then with, the support of the CNN model, and the diagnostic accuracy of the two methods was compared. Results At the time of the study, the CNN model showed an area under the receiver operating curve of 0.97. AI assistance improved the physician's sensitivity (correct fracture detection) from 80% to 87%, and the specificity (correct fracture exclusion) from 91% to 95%. The overall error rate (combined false positive and false negative) was reduced from 14% without AI to 9% with AI. Conclusion The use of a CNN model as a second opinion can improve the diagnostic accuracy of DRF detection in the study setting.
Collapse
Affiliation(s)
- Rosmarie Breu
- Orthopedic Hospital Vienna-Speising, Vienna, Austria
- AUVA Trauma Hospital Lorenz Böhler, Vienna, Austria
- Ludwig Boltzmann Institute for Traumatology, the Research Center in Cooperation with AUVA, Vienna, Austria
| | | | | | - Johannes Grillari
- Ludwig Boltzmann Institute for Traumatology, the Research Center in Cooperation with AUVA, Vienna, Austria
- Institute of Molecular Biotechnology, University of Natural Resources and Life Sciences, Vienna, Austria
- Austrian Cluster for Tissue Regeneration, Vienna, Austria
| | - Heinz Redl
- Ludwig Boltzmann Institute for Traumatology, the Research Center in Cooperation with AUVA, Vienna, Austria
- Austrian Cluster for Tissue Regeneration, Vienna, Austria
| | - Richard Ljuhar
- ImageBiopsy Lab, Vienna, Austria
- Institute of Molecular Biotechnology, University of Natural Resources and Life Sciences, Vienna, Austria
| | | | - Thomas Hausner
- AUVA Trauma Hospital Lorenz Böhler, Vienna, Austria
- Ludwig Boltzmann Institute for Traumatology, the Research Center in Cooperation with AUVA, Vienna, Austria
- Austrian Cluster for Tissue Regeneration, Vienna, Austria
- Department for Orthopedic Surgery and Traumatology, Paracelsus Medical University, Salzburg, Austria
| |
Collapse
|
31
|
Husarek J, Hess S, Razaeian S, Ruder TD, Sehmisch S, Müller M, Liodakis E. Artificial intelligence in commercial fracture detection products: a systematic review and meta-analysis of diagnostic test accuracy. Sci Rep 2024; 14:23053. [PMID: 39367147 PMCID: PMC11452402 DOI: 10.1038/s41598-024-73058-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2024] [Accepted: 09/12/2024] [Indexed: 10/06/2024] Open
Abstract
Conventional radiography (CR) is primarily utilized for fracture diagnosis. Artificial intelligence (AI) for CR is a rapidly growing field aimed at enhancing efficiency and increasing diagnostic accuracy. However, the diagnostic performance of commercially available AI fracture detection solutions (CAAI-FDS) for CR in various anatomical regions, their synergy with human assessment, as well as the influence of industry funding on reported accuracy are unknown. Peer-reviewed diagnostic test accuracy (DTA) studies were identified through a systematic review on Pubmed and Embase. Diagnostic performance measures were extracted especially for different subgroups such as product, type of rater (stand-alone AI, human unaided, human aided), funding, and anatomical region. Pooled measures were obtained with a bivariate random effects model. The impact of rater was evaluated with comparative meta-analysis. Seventeen DTA studies of seven CAAI-FDS analyzing 38,978 x-rays with 8,150 fractures were included. Stand-alone AI studies (n = 15) evaluated five CAAI-FDS; four with good sensitivities (> 90%) and moderate specificities (80-90%) and one with very poor sensitivity (< 60%) and excellent specificity (> 95%). Pooled sensitivities were good to excellent, and specificities were moderate to good in all anatomical regions (n = 7) apart from ribs (n = 4; poor sensitivity / moderate specificity) and spine (n = 4; excellent sensitivity / poor specificity). Funded studies (n = 4) had higher sensitivity (+ 5%) and lower specificity (-4%) than non-funded studies (n = 11). Sensitivity did not differ significantly between stand-alone AI and human AI aided ratings (p = 0.316) but specificity was significantly higher the latter group (p < 0.001). Sensitivity was significant lower in human unaided compared to human AI aided respectively stand-alone AI ratings (both p ≤ 0.001); specificity was higher in human unaided ratings compared to stand-alone AI (p < 0.001) and showed no significant differences AI aided ratings (p = 0.316). The study demonstrates good diagnostic accuracy across most CAAI-FDS and anatomical regions, with the highest performance achieved when used in conjunction with human assessment. Diagnostic accuracy appears lower for spine and rib fractures. The impact of industry funding on reported performance is small.
Collapse
Affiliation(s)
- Julius Husarek
- Department of Orthopaedic Surgery and Traumatology, Bern University Hospital, Inselspital, University of Bern, Bern, Switzerland
- University of Bern, Bern, Switzerland
- Faculty of Medicine, Medical University of Sofia, Sofia, Bulgaria
| | - Silvan Hess
- Department of Orthopaedic Surgery and Traumatology, Bern University Hospital, Inselspital, University of Bern, Bern, Switzerland
| | - Sam Razaeian
- Department for Trauma, Hand and Reconstructive Surgery, Saarland University, Kirrberger Str. 100, 66421, Homburg, Germany
| | - Thomas D Ruder
- Interventional and Pediatric Radiology, Inselspital, Bern University Hospital, University Institute of Diagnostic, University of Bern, Bern, Switzerland
| | - Stephan Sehmisch
- Department of Trauma Surgery, Hannover Medical School, Carl-Neuberg-Straße 1, 30625, Hannover, Germany
| | - Martin Müller
- Department of Emergency Medicine, Bern University Hospital, Inselspital, University of Bern, Bern, Switzerland
| | - Emmanouil Liodakis
- Department for Trauma, Hand and Reconstructive Surgery, Saarland University, Kirrberger Str. 100, 66421, Homburg, Germany.
| |
Collapse
|
32
|
Ma D, Wang Y, Zhang X, Su D, Ma M, Qian B, Yang X, Gao J, Wu Y. 3D U-Net Neural Network Architecture-Assisted LDCT to Acquire Vertebral Morphology Parameters: A Vertebral Morphology Comprehensive Analysis in a Chinese Population. Calcif Tissue Int 2024; 115:362-372. [PMID: 39017691 DOI: 10.1007/s00223-024-01255-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/12/2024] [Accepted: 07/01/2024] [Indexed: 07/18/2024]
Abstract
To evaluate the feasibility of acquiring vertebral height from chest low-dose computed tomography (LDCT) images using an artificial intelligence (AI) system based on 3D U-Net vertebral segmentation technology and the correlation and features of vertebral morphology with sex and age of the Chinese population. Patients who underwent chest LDCT between September 2020 and April 2023 were enrolled. The Altman and Pearson's correlation analyses were used to compare the correlation and consistency between the AI software and manual measurement of vertebral height. The anterior height (Ha), middle height (Hm), posterior height (Hp), and vertebral height ratios (VHRs) (Ha/Hp and Hm/Hp) were measured from T1 to L2 using an AI system. The VHR is the ratio of Ha to Hp or the ratio of Hm to Hp of the vertebrae, which can reflect the shape of the anterior wedge and biconcave vertebrae. Changes in these parameters, particularly the VHR, were analysed at different vertebral levels in different age and sex groups. The results of the AI methods were highly consistent and correlated with manual measurements. The Pearson's correlation coefficients were 0.855, 0.919, and 0.846, respectively. The trend of VHRs showed troughs at T7 and T11 and a peak at T9; however, Hm/Hp showed slight fluctuations. Regarding the VHR, significant sex differences were found at L1 and L2 in all age bands. This innovative study focuses on vertebral morphology for opportunistic analysis in the mainland Chinese population and the distribution tendency of vertebral morphology with ageing using a chest LDCT aided by an AI system based on 3D U-Net vertebral segmentation technology. The AI system demonstrates the potential to automatically perform opportunistic vertebral morphology analyses using LDCT scans obtained during lung cancer screening. We advocate the use of age-, sex-, and vertebral level-specific criteria for the morphometric evaluation of vertebral osteoporotic fractures for a more accurate diagnosis of vertebral fractures and spinal pathologies.
Collapse
Affiliation(s)
- Duoshan Ma
- Department of Radiology, The First Affiliated Hospital of Zhengzhou University, 1 Jianshe East Road, Zhengzhou, 450052, China
| | - Yan Wang
- Department of Radiology, The First Affiliated Hospital of Zhengzhou University, 1 Jianshe East Road, Zhengzhou, 450052, China
| | - Xinxin Zhang
- Department of Radiology, The First Affiliated Hospital of Zhengzhou University, 1 Jianshe East Road, Zhengzhou, 450052, China
| | - Danyang Su
- Department of Radiology, The First Affiliated Hospital of Zhengzhou University, 1 Jianshe East Road, Zhengzhou, 450052, China
| | - Mengze Ma
- Medical 3D Printing Center, The First Affiliated Hospital of Zhengzhou University, 1 Jianshe East Road, Zhengzhou, 450052, China
| | - Baoxin Qian
- Dongsheng Science and Technology Park, Room A206, B2, Huiying Medical Technology Co, Ltd, HaiDian District, Beijing City, 100192, China
| | - Xiaopeng Yang
- Medical 3D Printing Center, The First Affiliated Hospital of Zhengzhou University, 1 Jianshe East Road, Zhengzhou, 450052, China
| | - Jianbo Gao
- Department of Radiology, The First Affiliated Hospital of Zhengzhou University, 1 Jianshe East Road, Zhengzhou, 450052, China
| | - Yan Wu
- Medical 3D Printing Center, The First Affiliated Hospital of Zhengzhou University, 1 Jianshe East Road, Zhengzhou, 450052, China.
| |
Collapse
|
33
|
Keller M, Rohner M, Honigmann P. The potential benefit of artificial intelligence regarding clinical decision-making in the treatment of wrist trauma patients. J Orthop Surg Res 2024; 19:579. [PMID: 39294720 PMCID: PMC11411868 DOI: 10.1186/s13018-024-05063-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/03/2024] [Accepted: 09/07/2024] [Indexed: 09/21/2024] Open
Abstract
PURPOSE The implementation of artificial intelligence (AI) in health care is gaining popularity. Many publications describe powerful AI-enabled algorithms. Yet there's only scarce evidence for measurable value in terms of patient outcomes, clinical decision-making or socio-economic impact. Our aim was to investigate the significance of AI in the emergency treatment of wrist trauma patients. METHOD Two groups of physicians were confronted with twenty realistic cases of wrist trauma patients and had to find the correct diagnosis and provide a treatment recommendation. One group was assisted by an AI-enabled application which detects and localizes distal radius fractures (DRF) with near-to-perfect precision while the other group had no help. Primary outcome measurement was diagnostic accuracy. Secondary outcome measurements were required time, number of added CT scans and senior consultations, correctness of the treatment, subjective and objective stress levels. RESULTS The AI-supported group was able to make a diagnosis without support (no additional CT, no senior consultation) in significantly more of the cases than the control group (75% vs. 52%, p = 0.003). The AI-enhanced group detected DRF with superior sensitivity (1.00 vs. 0.96, p = 0.06) and specificity (0.99 vs. 0.93, p = 0.17), used significantly less additional CT scans to reach the correct diagnosis (14% vs. 28%, p = 0.02) and was subjectively significantly less stressed (p = 0.05). CONCLUSION The results indicate that physicians can diagnose wrist trauma more accurately and faster when aided by an AI-tool that lessens the need for extra diagnostic procedures. The AI-tool also seems to lower physicians' stress levels while examining cases. We anticipate that these benefits will be amplified in larger studies as skepticism towards the new technology diminishes.
Collapse
Affiliation(s)
- Marco Keller
- Hand and Peripheral Nerve Surgery, Department of Orthopaedic Surgery and Traumatology, Kantonsspital Baselland (Bruderholz, Liestal, Laufen), Bruderholz, Switzerland.
- Medical Additive Manufacturing Research Group (MAM), Department of Biomedical Engineering, University of Basel, Allschwil, Switzerland.
- Hand and Peripheral Nerve Surgery, Department of Orthopaedic Surgery, Traumatology and Hand Surgery, Spital Limmattal, Schlieren, Switzerland.
| | - Meret Rohner
- Hand and Peripheral Nerve Surgery, Department of Orthopaedic Surgery and Traumatology, Kantonsspital Baselland (Bruderholz, Liestal, Laufen), Bruderholz, Switzerland
- Medical Additive Manufacturing Research Group (MAM), Department of Biomedical Engineering, University of Basel, Allschwil, Switzerland
- Medical Faculty, University of Basel, Basel, Switzerland
| | - Philipp Honigmann
- Hand and Peripheral Nerve Surgery, Department of Orthopaedic Surgery and Traumatology, Kantonsspital Baselland (Bruderholz, Liestal, Laufen), Bruderholz, Switzerland
- Medical Additive Manufacturing Research Group (MAM), Department of Biomedical Engineering, University of Basel, Allschwil, Switzerland
- Department of Biomedical Engineering and Physics, Amsterdam UMC, University of Amsterdam, Meibergdreef 9, Amsterdam, The Netherlands
| |
Collapse
|
34
|
Yari A, Fasih P, Hosseini Hooshiar M, Goodarzi A, Fattahi SF. Detection and classification of mandibular fractures in panoramic radiography using artificial intelligence. Dentomaxillofac Radiol 2024; 53:363-371. [PMID: 38652576 PMCID: PMC11358630 DOI: 10.1093/dmfr/twae018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/26/2024] [Revised: 04/11/2024] [Accepted: 04/19/2024] [Indexed: 04/25/2024] Open
Abstract
OBJECTIVES This study evaluated the performance of the YOLOv5 deep learning model in detecting different mandibular fracture types in panoramic images. METHODS The dataset of panoramic radiographs with mandibular fractures was divided into training, validation, and testing sets, with 60%, 20%, and 20% of the images, respectively. An equal number of control images without fractures were also distributed among the datasets. The YOLOv5 algorithm was trained to detect six mandibular fracture types based on the anatomical location including symphysis, body, angle, ramus, condylar neck, and condylar head. Performance metrics of accuracy, precision, sensitivity (recall), specificity, dice coefficient (F1 score), and area under the curve (AUC) were calculated for each class. RESULTS A total of 498 panoramic images containing 673 fractures were collected. The accuracy was highest in detecting body (96.21%) and symphysis (95.87%), and was lowest in angle (90.51%) fractures. The highest and lowest precision values were observed in detecting symphysis (95.45%) and condylar head (63.16%) fractures, respectively. The sensitivity was highest in the body (96.67%) fractures and was lowest in the condylar head (80.00%) and condylar neck (81.25%) fractures. The highest specificity was noted in symphysis (98.96%), body (96.08%), and ramus (96.04%) fractures, respectively. The dice coefficient and AUC were highest in detecting body fractures (0.921 and 0.942, respectively), and were lowest in detecting condylar head fractures (0.706 and 0.812, respectively). CONCLUSION The trained algorithm achieved promising results in detecting most fracture types, particularly in body and symphysis regions indicating machine learning potential as a diagnostic aid for clinicians.
Collapse
Affiliation(s)
- Amir Yari
- Department of Oral and Maxillofacial Surgery, School of Dentistry, Kashan University of Medical Sciences, Kashan, 8715973474, Iran
| | - Paniz Fasih
- Department of Prosthodontics, School of Dentistry, Kashan University of Medical Sciences, Kashan, 8715973474, Iran
| | - Mohammad Hosseini Hooshiar
- Department of Periodontics, School of Dentistry, Tehran University of Medical Sciences, Tehran, 1439955991, Iran
| | - Ali Goodarzi
- Department of Oral and Maxillofacial Surgery, School of Dentistry, Shiraz University of Medical Sciences, Shiraz, 7195615878, Iran
| | - Seyedeh Farnaz Fattahi
- Department of Prosthodontics, School of Dentistry, Shiraz University of Medical Sciences, Shiraz, 7195615878, Iran
| |
Collapse
|
35
|
Yao Y, Lin C, Chen H, Lin H, Hsiung W, Wang S, Sun Y, Tang Y, Chou P. Development and validation of deep learning models for identifying the brand of pedicle screws on plain spine radiographs. JOR Spine 2024; 7:e70001. [PMID: 39291095 PMCID: PMC11406509 DOI: 10.1002/jsp2.70001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/09/2024] [Revised: 07/18/2024] [Accepted: 08/18/2024] [Indexed: 09/19/2024] Open
Abstract
Background In spinal revision surgery, previous pedicle screws (PS) may need to be replaced with new implants. Failure to accurately identify the brand of PS-based instrumentation preoperatively may increase the risk of perioperative complications. This study aimed to develop and validate an optimal deep learning (DL) model to identify the brand of PS-based instrumentation on plain radiographs of spine (PRS) using anteroposterior (AP) and lateral images. Methods A total of 529 patients who received PS-based instrumentation from seven manufacturers were enrolled in this retrospective study. The postoperative PRS were gathered as ground truths. The training, validation, and testing datasets contained 338, 85, and 106 patients, respectively. YOLOv5 was used to crop out the screws' trajectory, and the EfficientNet-b0 model was used to develop single models (AP, Lateral, Merge, and Concatenated) based on the different PRS images. The ensemble models were different combinations of the single models. Primary outcomes were the models' performance in accuracy, sensitivity, precision, F1-score, kappa value, and area under the curve (AUC). Secondary outcomes were the relative performance of models versus human readers and external validation of the DL models. Results The Lateral model had the most stable performance among single models. The discriminative performance was improved by the ensemble method. The AP + Lateral ensemble model had the most stable performance, with an accuracy of 0.9434, F1 score of 0.9388, and AUC of 0.9834. The performance of the ensemble models was comparable to that of experienced orthopedic surgeons and superior to that of inexperienced orthopedic surgeons. External validation revealed that the Lat + Concat ensemble model had the best accuracy (0.9412). Conclusion The DL models demonstrated stable performance in identifying the brand of PS-based instrumentation based on AP and/or lateral images of PRS, which may assist orthopedic spine surgeons in preoperative revision planning in clinical practice.
Collapse
Affiliation(s)
- Yu‐Cheng Yao
- School of MedicineNational Yang Ming Chiao Tung UniversityTaipeiTaiwan
- Department of Orthopedics and TraumatologyTaipei Veterans General HospitalTaipeiTaiwan
| | - Cheng‐Li Lin
- Department of Orthopaedic Surgery, National Cheng Kung University Hospital, College of MedicineNational Cheung Kung UniversityTainanTaiwan
| | - Hung‐Hsun Chen
- Program of Artificial Intelligence and Information SecurityFu Jen Catholic UniversityNew Taipei CityTaiwan
| | - Hsi‐Hsien Lin
- School of MedicineNational Yang Ming Chiao Tung UniversityTaipeiTaiwan
- Department of Orthopedics and TraumatologyTaipei Veterans General HospitalTaipeiTaiwan
| | - Wei Hsiung
- School of MedicineNational Yang Ming Chiao Tung UniversityTaipeiTaiwan
- Department of Orthopedics and TraumatologyTaipei Veterans General HospitalTaipeiTaiwan
- Department of OrthopedicsShin Kong Wu Ho‐Su Memorial HospitalTaipeiTaiwan
| | - Shih‐Tien Wang
- School of MedicineNational Yang Ming Chiao Tung UniversityTaipeiTaiwan
- Department of Orthopedics and TraumatologyTaipei Veterans General HospitalTaipeiTaiwan
- Kinmen HospitalMinistry of Health and WelfareKinmenTaiwan
| | - Ying‐Chou Sun
- School of MedicineNational Yang Ming Chiao Tung UniversityTaipeiTaiwan
- Department of RadiologyTaipei Veterans General HospitalTaipeiTaiwan
- Department of Medical Imaging and Radiological TechnologyYuanpei University of Medical TechnologyHsinchuTaiwan
| | - Yu‐Hsuan Tang
- Department of Medical Imaging and Radiological TechnologyYuanpei University of Medical TechnologyHsinchuTaiwan
| | - Po‐Hsin Chou
- School of MedicineNational Yang Ming Chiao Tung UniversityTaipeiTaiwan
- Department of Orthopedics and TraumatologyTaipei Veterans General HospitalTaipeiTaiwan
| |
Collapse
|
36
|
Marullo G, Ulrich L, Antonaci FG, Audisio A, Aprato A, Massè A, Vezzetti E. Classification of AO/OTA 31A/B femur fractures in X-ray images using YOLOv8 and advanced data augmentation techniques. Bone Rep 2024; 22:101801. [PMID: 39324016 PMCID: PMC11422035 DOI: 10.1016/j.bonr.2024.101801] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/26/2024] [Revised: 08/20/2024] [Accepted: 09/05/2024] [Indexed: 09/27/2024] Open
Abstract
Femur fractures are a significant worldwide public health concern that affects patients as well as their families because of their high frequency, morbidity, and mortality. When employing computer-aided diagnostic (CAD) technologies, promising results have been shown in the efficiency and accuracy of fracture classification, particularly with the growing use of Deep Learning (DL) approaches. Nevertheless, the complexity is further increased by the need to collect enough input data to train these algorithms and the challenge of interpreting the findings. By improving on the results of the most recent deep learning-based Arbeitsgemeinschaft für Osteosynthesefragen and Orthopaedic Trauma Association (AO/OTA) system classification of femur fractures, this study intends to support physicians in making correct and timely decisions regarding patient care. A state-of-the-art architecture, YOLOv8, was used and refined while paying close attention to the interpretability of the model. Furthermore, data augmentation techniques were involved during preprocessing, increasing the dataset samples through image processing alterations. The fine-tuned YOLOv8 model achieved remarkable results, with 0.9 accuracy, 0.85 precision, 0.85 recall, and 0.85 F1-score, computed by averaging the values among all the individual classes for each metric. This study shows the proposed architecture's effectiveness in enhancing the AO/OTA system's classification of femur fractures, assisting physicians in making prompt and accurate diagnoses.
Collapse
Affiliation(s)
- Giorgia Marullo
- Department of Management, Production, and Design, Politecnico di Torino, C.so Duca degli Abruzzi, 24, Torino 10129, Italy
| | - Luca Ulrich
- Department of Management, Production, and Design, Politecnico di Torino, C.so Duca degli Abruzzi, 24, Torino 10129, Italy
| | - Francesca Giada Antonaci
- Department of Management, Production, and Design, Politecnico di Torino, C.so Duca degli Abruzzi, 24, Torino 10129, Italy
| | - Andrea Audisio
- Pediatric Orthopaedics and Traumatology, Regina Margherita Children's Hospital, Torino 10126, Italy
| | - Alessandro Aprato
- Department of Surgical Sciences, University of Turin, Torino 10124, Italy
| | - Alessandro Massè
- Department of Surgical Sciences, University of Turin, Torino 10124, Italy
| | - Enrico Vezzetti
- Department of Management, Production, and Design, Politecnico di Torino, C.so Duca degli Abruzzi, 24, Torino 10129, Italy
| |
Collapse
|
37
|
Franco PN, Maino C, Mariani I, Gandola DG, Sala D, Bologna M, Talei Franzesi C, Corso R, Ippolito D. Diagnostic performance of an AI algorithm for the detection of appendicular bone fractures in pediatric patients. Eur J Radiol 2024; 178:111637. [PMID: 39053306 DOI: 10.1016/j.ejrad.2024.111637] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2024] [Revised: 07/15/2024] [Accepted: 07/16/2024] [Indexed: 07/27/2024]
Abstract
PURPOSE To evaluate the diagnostic performance of an Artificial Intelligence (AI) algorithm, previously trained using both adult and pediatric patients, for the detection of acute appendicular fractures in the pediatric population on conventional X-ray radiography (CXR). MATERIALS AND METHODS In this retrospective study, anonymized extremities CXRs of pediatric patients (age <17 years), with or without fractures, were included. Six hundred CXRs (maintaining the positive-for-fracture and negative-for-fracture balance) were included, grouping them per body part (shoulder/clavicle, elbow/upper arm, hand/wrist, leg/knee, foot/ankle). Follow-up CXRs and/or second-level imaging were considered as reference standard. A deep learning algorithm interpreted CXRs for fracture detection on a per-patient, per-radiograph, and per-location level, and its diagnostic performance values were compared with the reference standard. AI diagnostic performance was computed by using cross-tables, and 95 % confidence intervals [CIs] were obtained by bootstrapping. RESULTS The final cohort included 312 male and 288 female with a mean age of 8.9±4.5 years. Three undred CXRs (50 %) were positive for fractures, according to the reference standard. For all fractures, the AI tool showed a per-patient 91.3% (95%CIs = 87.6-94.3) sensitivity, 76.7% (71.5-81.3) specificity, and 84% (82.1-86.0) accuracy. In the per-radiograph analysis the AI tool showed 85% (81.9-87.8) sensitivity, 88.5% (86.3-90.4) specificity, and 87.2% (85.7-89.6) accuracy. In the per-location analysis, the AI tool identified 606 bounding boxes: 472 (77.9 %) were correct, 110 (18.1 %) incorrect, and 24 (4.0 %) were not-overlapping. CONCLUSION The AI algorithm provides good overall diagnostic performance for detecting appendicular fractures in pediatric patients.
Collapse
Affiliation(s)
- Paolo Niccolò Franco
- Department of Diagnostic Radiology, Fondazione IRCCS San Gerardo dei Tintori, Via Pergolesi 33, 20900 Monza, MB, Italy
| | - Cesare Maino
- Department of Diagnostic Radiology, Fondazione IRCCS San Gerardo dei Tintori, Via Pergolesi 33, 20900 Monza, MB, Italy.
| | - Ilaria Mariani
- Department of Diagnostic Radiology, Fondazione IRCCS San Gerardo dei Tintori, Via Pergolesi 33, 20900 Monza, MB, Italy
| | - Davide Giacomo Gandola
- Department of Diagnostic Radiology, Fondazione IRCCS San Gerardo dei Tintori, Via Pergolesi 33, 20900 Monza, MB, Italy
| | - Davide Sala
- Synbrain The Human-Machine Cooperation - AI/ML solutions, Corso Milano 23, 20900 Monza, MB, Italy
| | - Marco Bologna
- Synbrain The Human-Machine Cooperation - AI/ML solutions, Corso Milano 23, 20900 Monza, MB, Italy
| | - Cammillo Talei Franzesi
- Department of Diagnostic Radiology, Fondazione IRCCS San Gerardo dei Tintori, Via Pergolesi 33, 20900 Monza, MB, Italy
| | - Rocco Corso
- Department of Diagnostic Radiology, Fondazione IRCCS San Gerardo dei Tintori, Via Pergolesi 33, 20900 Monza, MB, Italy
| | - Davide Ippolito
- Department of Diagnostic Radiology, Fondazione IRCCS San Gerardo dei Tintori, Via Pergolesi 33, 20900 Monza, MB, Italy; Department of Medicine and Surgery - University of Milano Bicocca, Via Cadore 33, 20090 Monza, MB, Italy
| |
Collapse
|
38
|
Kutbi M. Artificial Intelligence-Based Applications for Bone Fracture Detection Using Medical Images: A Systematic Review. Diagnostics (Basel) 2024; 14:1879. [PMID: 39272664 PMCID: PMC11394268 DOI: 10.3390/diagnostics14171879] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/15/2024] [Revised: 08/19/2024] [Accepted: 08/26/2024] [Indexed: 09/15/2024] Open
Abstract
Artificial intelligence (AI) is making notable advancements in the medical field, particularly in bone fracture detection. This systematic review compiles and assesses existing research on AI applications aimed at identifying bone fractures through medical imaging, encompassing studies from 2010 to 2023. It evaluates the performance of various AI models, such as convolutional neural networks (CNNs), in diagnosing bone fractures, highlighting their superior accuracy, sensitivity, and specificity compared to traditional diagnostic methods. Furthermore, the review explores the integration of advanced imaging techniques like 3D CT and MRI with AI algorithms, which has led to enhanced diagnostic accuracy and improved patient outcomes. The potential of Generative AI and Large Language Models (LLMs), such as OpenAI's GPT, to enhance diagnostic processes through synthetic data generation, comprehensive report creation, and clinical scenario simulation is also discussed. The review underscores the transformative impact of AI on diagnostic workflows and patient care, while also identifying research gaps and suggesting future research directions to enhance data quality, model robustness, and ethical considerations.
Collapse
Affiliation(s)
- Mohammed Kutbi
- College of Computing and Informatics, Saudi Electronic University, Riyadh 13316, Saudi Arabia
| |
Collapse
|
39
|
Nowroozi A, Salehi MA, Shobeiri P, Agahi S, Momtazmanesh S, Kaviani P, Kalra MK. Artificial intelligence diagnostic accuracy in fracture detection from plain radiographs and comparing it with clinicians: a systematic review and meta-analysis. Clin Radiol 2024; 79:579-588. [PMID: 38772766 DOI: 10.1016/j.crad.2024.04.009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/08/2024] [Revised: 04/09/2024] [Accepted: 04/15/2024] [Indexed: 05/23/2024]
Abstract
PURPOSE Fracture detection is one of the most commonly used and studied aspects of artificial intelligence (AI) in medicine. In this systematic review and meta-analysis, we aimed to summarize available literature and data regarding AI performance in fracture detection on plain radiographs and various factors affecting it. METHODS We systematically reviewed studies evaluating AI algorithms in detecting bone fractures in plain radiographs, combined their performance using meta-analysis (a bivariate regression approach), and compared it with that of clinicians. We also analyzed the factors potentially affecting algorithm performance using meta-regression. RESULTS Our analysis included 100 studies. In 83 studies with confusion matrices, AI algorithms showed a sensitivity of 91.43% and a specificity of 92.12% (Area under the summary receiver operator curve = 0.968). After adjustment and false discovery rate correction, tibia/fibula (excluding ankle) fractures were associated with higher (7.0%, p=0.004) AI sensitivity, while more recent publications (5.5%, p=0.003) and Xception architecture (6.6%, p<0.001) were associated with higher specificity. Clinicians and AI showed similar specificity in fracture identification, although AI leaned to higher sensitivity (7.6%, p=0.07). Radiologists, on the other hand, were more specific than AI overall and in several subgroups, and more sensitive to hip fractures before FDR correction. CONCLUSIONS Currently available AI aids could result in a significant improvement in care where radiologists are not readily available. Moreover, identifying factors affecting algorithm performance could guide AI development teams in their process of optimizing their products.
Collapse
Affiliation(s)
- A Nowroozi
- School of Medicine, Tehran University of Medical Sciences, Tehran, Iran
| | - M A Salehi
- School of Medicine, Tehran University of Medical Sciences, Tehran, Iran
| | - P Shobeiri
- School of Medicine, Tehran University of Medical Sciences, Tehran, Iran
| | - S Agahi
- School of Medicine, Tehran University of Medical Sciences, Tehran, Iran
| | - S Momtazmanesh
- School of Medicine, Tehran University of Medical Sciences, Tehran, Iran
| | - P Kaviani
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, MA 02114, USA
| | - M K Kalra
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, MA 02114, USA.
| |
Collapse
|
40
|
Dasegowda G, Sato JY, Elton DC, Garza-Frias E, Schultz T, Bridge CP, Bizzo BC, Kalra MK, Dreyer KJ. No code machine learning: validating the approach on use-case for classifying clavicle fractures. Clin Imaging 2024; 112:110207. [PMID: 38838448 DOI: 10.1016/j.clinimag.2024.110207] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Revised: 04/24/2024] [Accepted: 05/23/2024] [Indexed: 06/07/2024]
Abstract
PURPOSE We created an infrastructure for no code machine learning (NML) platform for non-programming physicians to create NML model. We tested the platform by creating an NML model for classifying radiographs for the presence and absence of clavicle fractures. METHODS Our IRB-approved retrospective study included 4135 clavicle radiographs from 2039 patients (mean age 52 ± 20 years, F:M 1022:1017) from 13 hospitals. Each patient had two-view clavicle radiographs with axial and anterior-posterior projections. The positive radiographs had either displaced or non-displaced clavicle fractures. We configured the NML platform to automatically retrieve the eligible exams using the series' unique identification from the hospital virtual network archive via web access to DICOM Objects. The platform trained a model until the validation loss plateaus. Once the testing was complete, the platform provided the receiver operating characteristics curve and confusion matrix for estimating sensitivity, specificity, and accuracy. RESULTS The NML platform successfully retrieved 3917 radiographs (3917/4135, 94.7 %) and parsed them for creating a ML classifier with 2151 radiographs in the training, 100 radiographs for validation, and 1666 radiographs in testing datasets (772 radiographs with clavicle fracture, 894 without clavicle fracture). The network identified clavicle fracture with 90 % sensitivity, 87 % specificity, and 88 % accuracy with AUC of 0.95 (confidence interval 0.94-0.96). CONCLUSION A NML platform can help physicians create and test machine learning models from multicenter imaging datasets such as the one in our study for classifying radiographs based on the presence of clavicle fracture.
Collapse
Affiliation(s)
- Giridhar Dasegowda
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, USA; Mass General Brigham AI, USA
| | - James Yuichi Sato
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, USA; Mass General Brigham AI, USA
| | - Daniel C Elton
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, USA; Mass General Brigham AI, USA
| | - Emiliano Garza-Frias
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, USA; Mass General Brigham AI, USA
| | - Thomas Schultz
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, USA; Mass General Brigham AI, USA
| | - Christopher P Bridge
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, USA; Mass General Brigham AI, USA
| | - Bernardo C Bizzo
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, USA; Mass General Brigham AI, USA
| | - Mannudeep K Kalra
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, USA; Mass General Brigham AI, USA.
| | - Keith J Dreyer
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, USA; Mass General Brigham AI, USA
| |
Collapse
|
41
|
Petrella RJ. The AI Future of Emergency Medicine. Ann Emerg Med 2024; 84:139-153. [PMID: 38795081 DOI: 10.1016/j.annemergmed.2024.01.031] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/03/2023] [Revised: 01/23/2024] [Accepted: 01/24/2024] [Indexed: 05/27/2024]
Abstract
In the coming years, artificial intelligence (AI) and machine learning will likely give rise to profound changes in the field of emergency medicine, and medicine more broadly. This article discusses these anticipated changes in terms of 3 overlapping yet distinct stages of AI development. It reviews some fundamental concepts in AI and explores their relation to clinical practice, with a focus on emergency medicine. In addition, it describes some of the applications of AI in disease diagnosis, prognosis, and treatment, as well as some of the practical issues that they raise, the barriers to their implementation, and some of the legal and regulatory challenges they create.
Collapse
Affiliation(s)
- Robert J Petrella
- Emergency Departments, CharterCARE Health Partners, Providence and North Providence, RI; Emergency Department, Boston VA Medical Center, Boston, MA; Emergency Departments, Steward Health Care System, Boston and Methuen, MA; Harvard Medical School, Boston, MA; Department of Chemistry and Chemical Biology, Harvard University, Cambridge, MA; Department of Medicine, Brigham and Women's Hospital, Boston, MA.
| |
Collapse
|
42
|
Zhu X, Liu D, Liu L, Guo J, Li Z, Zhao Y, Wu T, Liu K, Liu X, Pan X, Qi L, Zhang Y, Cheng L, Chen B. Fully Automatic Deep Learning Model for Spine Refracture in Patients with OVCF: A Multi-Center Study. Orthop Surg 2024; 16:2052-2065. [PMID: 38952050 PMCID: PMC11293932 DOI: 10.1111/os.14155] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/11/2024] [Revised: 06/06/2024] [Accepted: 06/09/2024] [Indexed: 07/03/2024] Open
Abstract
BACKGROUND The reaserch of artificial intelligence (AI) model for predicting spinal refracture is limited to bone mineral density, X-ray and some conventional laboratory indicators, which has its own limitations. Besides, it lacks specific indicators related to osteoporosis and imaging factors that can better reflect bone quality, such as computed tomography (CT). OBJECTIVE To construct a novel predicting model based on bone turn-over markers and CT to identify patients who were more inclined to suffer spine refracture. METHODS CT images and clinical information of 383 patients (training set = 240 cases of osteoporotic vertebral compression fractures (OVCF), validation set = 63, test set = 80) were retrospectively collected from January 2015 to October 2022 at three medical centers. The U-net model was adopted to automatically segment ROI. Three-dimensional (3D) cropping of all spine regions was used to achieve the final ROI regions including 3D_Full and 3D_RoiOnly. We used the Densenet 121-3D model to model the cropped region and simultaneously build a T-NIPT prediction model. Diagnostics of deep learning models were assessed by constructing ROC curves. We generated calibration curves to assess the calibration performance. Additionally, decision curve analysis (DCA) was used to assess the clinical utility of the predictive models. RESULTS The performance of the test model is comparable to its performance on the training set (dice coefficients of 0.798, an mIOU of 0.755, an SA of 0.767, and an OS of 0.017). Univariable and multivariable analysis indicate that T_P1NT was an independent risk factor for refracture. The performance of predicting refractures in different ROI regions showed that 3D_Full model exhibits the highest calibration performance, with a Hosmer-Lemeshow goodness-of-fit (HL) test statistic exceeding 0.05. The analysis of the training and test sets showed that the 3D_Full model, which integrates clinical and deep learning results, demonstrated superior performance with significant improvement (p-value < 0.05) compared to using clinical features independently or using only 3D_RoiOnly. CONCLUSION T_P1NT was an independent risk factor of refracture. Our 3D-FULL model showed better performance in predicting high-risk population of spine refracture than other models and junior doctors do. This model can be applicable to real-world translation due to its automatic segmentation and detection.
Collapse
Affiliation(s)
- Xuetao Zhu
- Department of Orthopaedic SurgeryQilu Hospital of Shandong University, Cheeloo College of Medicine of Shandong UniversityJinanP. R. China
| | - Dejian Liu
- Department of Orthopaedic SurgeryQilu Hospital of Shandong University, Cheeloo College of Medicine of Shandong UniversityJinanP. R. China
| | - Lian Liu
- Department of Emergency SurgeryQilu Hospital of Shandong University, Cheeloo College of Medicine of Shandong UniversityJinanP. R. China
| | - Jingxuan Guo
- Department of anesthesiologyAffiliated Hospital of Shandong University of Traditional Chinese MedicineJinanChina
| | - Zedi Li
- Department of Orthopaedic SurgeryQilu Hospital of Shandong University, Cheeloo College of Medicine of Shandong UniversityJinanP. R. China
| | - Yixiang Zhao
- Department of Orthopaedic SurgeryYantaishan HospitalYantaiChina
| | - Tianhao Wu
- Department of Hepatopancreatobiliary SurgeryGraduate School of Dalian Medical UniversityDalianChina
| | - Kaiwen Liu
- Department of Orthopaedic SurgeryQilu Hospital of Shandong University, Cheeloo College of Medicine of Shandong UniversityJinanP. R. China
| | - Xinyu Liu
- Department of Orthopaedic SurgeryQilu Hospital of Shandong University, Cheeloo College of Medicine of Shandong UniversityJinanP. R. China
| | - Xin Pan
- Department of Orthopaedic SurgeryQilu Hospital of Shandong University, Cheeloo College of Medicine of Shandong UniversityJinanP. R. China
| | - Lei Qi
- Department of Orthopaedic SurgeryQilu Hospital of Shandong University, Cheeloo College of Medicine of Shandong UniversityJinanP. R. China
| | - Yuanqiang Zhang
- Department of Orthopaedic SurgeryQilu Hospital of Shandong University, Cheeloo College of Medicine of Shandong UniversityJinanP. R. China
| | - Lei Cheng
- Department of Orthopaedic SurgeryQilu Hospital of Shandong University, Cheeloo College of Medicine of Shandong UniversityJinanP. R. China
| | - Bin Chen
- Department of Orthopaedic SurgeryQilu Hospital of Shandong University, Cheeloo College of Medicine of Shandong UniversityJinanP. R. China
| |
Collapse
|
43
|
Hsiung W, Lin HY, Lin HH, Yao YC, Wang ST, Chang MC, Chou PH. MRI-based lesion quality score assessing ossification of the posterior longitudinal ligament of the cervical spine. Spine J 2024; 24:1162-1169. [PMID: 38365006 DOI: 10.1016/j.spinee.2024.02.007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/03/2023] [Revised: 01/14/2024] [Accepted: 02/06/2024] [Indexed: 02/18/2024]
Abstract
BACKGROUND CONTEXT No method currently exists for MRI-based determination of ossification of the posterior longitudinal ligament (OPLL) of the cervical spine using objective criteria. PURPOSE The purpose of this study was to develop an MRI-based score to determine whether a lesion represents a cervical OPLL lesion and to establish the objective diagnostic value. STUDY DESIGN Retrospective cohort in a single medical institution. PATIENT SAMPLE Thirty-five patients undergoing surgery for OPLL (Group A) and 99 patients undergoing cervical disc arthroplasty for soft disc herniation (Group B) between 2011 and 2020 were retrospectively included. All OPLL lesions on unenhanced MRI scan were correlated with a corresponding CT scan. Demographics were comparable between the two groups. OUTCOME MEASURES (PHYSIOLOGIC MEASURES) Using unenhanced magnetic resonance imaging (MRI), the T1- and T2- lesion quality (LQ) scores were calculated. Receiver operating characteristic (ROC) analysis was performed to calculate the area-under-the-curve (AUC) of both LQ scores as a predictor of the presence of OPLL. Computed tomography (CT)-based Hounsfield unit (HU) values of OPLL lesions were obtained and compared with both LQ scores. The LQ scores for MRI scanners from different manufacturers were compared using Student's t test to confirm the validity of the LQ score by scanner type. METHODS The regions of interest for signal intensity (SI) were defined as the darkest site of the lesion and the cerebrospinal fluid (CSF) at the cerebellomedullary cistern. The T1 and T2 LQ scores were measured as the ratio of the SI at the darkest site of the lesion divided by the SI of the CSF. RESULTS The T1 and T2 LQ scores in Group A were significantly lower than those in Group B (p<.001). ROC analysis determined that T1 and T2 LQ scores of 0.46 and 0.07, respectively, could distinguish the presence of OPLL with an accuracy of 0.93 and 0.89, respectively (p<.001). When the T1 LQ score of the lesion is <0.46, a diagnosis of OPLL may be suspected with 100% sensitivity and 92.3% specificity. The HU of the lesion had a moderate negative correlation with the T1 LQ score (r=-0.665, p<.0001). Both LQ scores were unaffected by manufacturer type. CONCLUSIONS This study found a correlation between the MRI-based T1 LQ scores and CT-based HU value for identifying OPLL lesions. Additional studies will be needed to validate that the T1 LQ score from the unenhanced MRI scan can identify cervical OPLL.
Collapse
Affiliation(s)
- Wei Hsiung
- School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan; Department of Orthopedics and Traumatology, Taipei Veterans General Hospital, Taipei, Taiwan; Department of Orthopedic Surgery, Shin Kong Wu Huo-Shih Memorial Hospital
| | - Han-Ying Lin
- School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan; Department of Radiology, Taipei Veterans General Hospital, Taipei, Taiwan
| | - Hsi-Hsien Lin
- School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan; Department of Orthopedics and Traumatology, Taipei Veterans General Hospital, Taipei, Taiwan
| | - Yu-Cheng Yao
- School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan; Department of Orthopedics and Traumatology, Taipei Veterans General Hospital, Taipei, Taiwan
| | - Shih-Tien Wang
- School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan; Department of Orthopedics and Traumatology, Taipei Veterans General Hospital, Taipei, Taiwan; Kinmen Hospital, Ministry of Health and Welfare, Taiwan
| | - Ming-Chau Chang
- School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan; Department of Orthopedics and Traumatology, Taipei Veterans General Hospital, Taipei, Taiwan
| | - Po-Hsin Chou
- School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan; Department of Orthopedics and Traumatology, Taipei Veterans General Hospital, Taipei, Taiwan.
| |
Collapse
|
44
|
Talebi S, Gai S, Sossin A, Zhu V, Tong E, Mofrad MRK. Deep Learning for Perfusion Cerebral Blood Flow (CBF) and Volume (CBV) Predictions and Diagnostics. Ann Biomed Eng 2024; 52:1568-1575. [PMID: 38402314 PMCID: PMC11082011 DOI: 10.1007/s10439-024-03471-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2023] [Accepted: 02/06/2024] [Indexed: 02/26/2024]
Abstract
Dynamic susceptibility contrast magnetic resonance perfusion (DSC-MRP) is a non-invasive imaging technique for hemodynamic measurements. Various perfusion parameters, such as cerebral blood volume (CBV) and cerebral blood flow (CBF), can be derived from DSC-MRP, hence this non-invasive imaging protocol is widely used clinically for the diagnosis and assessment of intracranial pathologies. Currently, most institutions use commercially available software to compute the perfusion parametric maps. However, these conventional methods often have limitations, such as being time-consuming and sensitive to user input, which can lead to inconsistent results; this highlights the need for a more robust and efficient approach like deep learning. Using the relative cerebral blood volume (rCBV) and relative cerebral blood flow (rCBF) perfusion maps generated by FDA-approved software, we trained a multistage deep learning model. The model, featuring a combination of a 1D convolutional neural network (CNN) and a 2D U-Net encoder-decoder network, processes each 4D MRP dataset by integrating temporal and spatial features of the brain for voxel-wise perfusion parameters prediction. An auxiliary model, with similar architecture, but trained with truncated datasets that had fewer time-points, was designed to explore the contribution of temporal features. Both qualitatively and quantitatively evaluated, deep learning-generated rCBV and rCBF maps showcased effective integration of temporal and spatial data, producing comprehensive predictions for the entire brain volume. Our deep learning model provides a robust and efficient approach for calculating perfusion parameters, demonstrating comparable performance to FDA-approved commercial software, and potentially mitigating the challenges inherent to traditional techniques.
Collapse
Affiliation(s)
- Salmonn Talebi
- Departments of Bioengineering and Mechanical Engineering, University of California, 208A Stanley Hall #1762, Berkeley, CA, 94720-1762, USA
| | - Siyu Gai
- Departments of Electrical Engineering and Computer Science, University of California, Berkeley, California, USA
| | - Aaron Sossin
- Department of Bioinformatics, Stanford School of Medicine, Stanford University, Stanford, California, USA
| | - Vivian Zhu
- Department of Bioinformatics, Stanford School of Medicine, Stanford University, Stanford, California, USA
| | - Elizabeth Tong
- Department of Radiology, Stanford School of Medicine, Stanford University, 725 Welch Rd Rm 1860, Palo Alto, Stanford, CA, 94304, USA.
| | - Mohammad R K Mofrad
- Departments of Bioengineering and Mechanical Engineering, University of California, 208A Stanley Hall #1762, Berkeley, CA, 94720-1762, USA.
| |
Collapse
|
45
|
Pridgen B, von Rabenau L, Luan A, Gu AJ, Wang DS, Langlotz C, Chang J, Do B. Automatic Detection of Perilunate and Lunate Dislocations on Wrist Radiographs Using Deep Learning. Plast Reconstr Surg 2024; 153:1138e-1141e. [PMID: 37467052 DOI: 10.1097/prs.0000000000010928] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/21/2023]
Abstract
SUMMARY Delayed or missed diagnosis of perilunate or lunate dislocations can lead to significant morbidity. Advances in computer vision provide an opportunity to improve diagnostic performance. In this study, a deep learning algorithm was used for detection of perilunate and lunate dislocations on lateral wrist radiographs. A total of 435 lateral wrist radiographs were labeled as normal or pathologic (perilunate or lunate dislocation). The lunate in each radiograph was segmented with a rectangular bounding box. Images were partitioned into training and test sets. Two neural networks, consisting of an object detector followed by an image classifier, were applied in series. First, the object detection module was used to localize the lunate. Next, the image classifier performed a binary classification for normal or pathologic. The accuracy, sensitivity, and specificity of the overall system were evaluated. A receiver operating characteristic curve and the associated area under the curve were used to demonstrate the overall performance of the computer vision algorithm. The lunate object detector was 97.0% accurate at identifying the lunate. Accuracy was 98.7% among the subgroup of normal wrist radiographs and 91.3% among the subgroup of wrist radiographs with perilunate/lunate dislocations. The perilunate/lunate dislocation classifier had a sensitivity (recall) of 93.8%, a specificity of 93.3%, and an accuracy of 93.4%. The area under the curve was 0.986. The authors have developed a proof-of-concept computer vision system for diagnosis of perilunate/lunate dislocations on lateral wrist radiographs. This novel deep learning algorithm has potential to improve clinical sensitivity to ultimately prevent delayed or missed diagnosis of these injuries.
Collapse
Affiliation(s)
- Brian Pridgen
- From the Division of Plastic Surgery, Department of Surgery
- The Buncke Clinic
| | | | - Anna Luan
- From the Division of Plastic Surgery, Department of Surgery
| | | | - David S Wang
- Department of Radiology, Stanford University School of Medicine
| | - Curtis Langlotz
- Department of Radiology, Stanford University School of Medicine
| | - James Chang
- From the Division of Plastic Surgery, Department of Surgery
| | - Bao Do
- Department of Radiology, Stanford University School of Medicine
| |
Collapse
|
46
|
Kim J, Seok J. ctGAN: combined transformation of gene expression and survival data with generative adversarial network. Brief Bioinform 2024; 25:bbae325. [PMID: 38980369 PMCID: PMC11232285 DOI: 10.1093/bib/bbae325] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/19/2024] [Revised: 05/29/2024] [Accepted: 06/21/2024] [Indexed: 07/10/2024] Open
Abstract
Recent studies have extensively used deep learning algorithms to analyze gene expression to predict disease diagnosis, treatment effectiveness, and survival outcomes. Survival analysis studies on diseases with high mortality rates, such as cancer, are indispensable. However, deep learning models are plagued by overfitting owing to the limited sample size relative to the large number of genes. Consequently, the latest style-transfer deep generative models have been implemented to generate gene expression data. However, these models are limited in their applicability for clinical purposes because they generate only transcriptomic data. Therefore, this study proposes ctGAN, which enables the combined transformation of gene expression and survival data using a generative adversarial network (GAN). ctGAN improves survival analysis by augmenting data through style transformations between breast cancer and 11 other cancer types. We evaluated the concordance index (C-index) enhancements compared with previous models to demonstrate its superiority. Performance improvements were observed in nine of the 11 cancer types. Moreover, ctGAN outperformed previous models in seven out of the 11 cancer types, with colon adenocarcinoma (COAD) exhibiting the most significant improvement (median C-index increase of ~15.70%). Furthermore, integrating the generated COAD enhanced the log-rank p-value (0.041) compared with using only the real COAD (p-value = 0.797). Based on the data distribution, we demonstrated that the model generated highly plausible data. In clustering evaluation, ctGAN exhibited the highest performance in most cases (89.62%). These findings suggest that ctGAN can be meaningfully utilized to predict disease progression and select personalized treatments in the medical field.
Collapse
Affiliation(s)
- Jaeyoon Kim
- School of Electrical and Computer Engineering, Korea University, 145 Anam-ro, Seongbuk-gu, Seoul, 02841, Korea
| | - Junhee Seok
- School of Electrical and Computer Engineering, Korea University, 145 Anam-ro, Seongbuk-gu, Seoul, 02841, Korea
| |
Collapse
|
47
|
Hansen V, Jensen J, Kusk MW, Gerke O, Tromborg HB, Lysdahlgaard S. Deep learning performance compared to healthcare experts in detecting wrist fractures from radiographs: A systematic review and meta-analysis. Eur J Radiol 2024; 174:111399. [PMID: 38428318 DOI: 10.1016/j.ejrad.2024.111399] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2023] [Revised: 01/29/2024] [Accepted: 02/26/2024] [Indexed: 03/03/2024]
Abstract
OBJECTIVE To perform a systematic review and meta-analysis of the diagnostic accuracy of deep learning (DL) algorithms in the diagnosis of wrist fractures (WF) on plain wrist radiographs, taking healthcare experts consensus as reference standard. METHODS Embase, Medline, PubMed, Scopus and Web of Science were searched in the period from 1 Jan 2012 to 9 March 2023. Eligible studies were patients with wrist radiographs for radial and ulnar fractures as the target condition, studies using DL algorithms based on convolutional neural networks (CNN), and healthcare experts consensus as the minimum reference standard. Studies were assessed with a modified QUADAS-2 tool, and we applied a bivariate random-effects model for meta-analysis of diagnostic test accuracy data. RESULTS Our study was registered at PROSPERO with ID: CRD42023431398. We included 6 unique studies for meta-analysis, with a total of 33,026 radiographs. CNN performance compared to reference standards for the included articles found a summary sensitivity of 92% (95% CI: 80%-97%) and a summary specificity of 93% (95% CI: 76%-98%). The generalized bivariate I-squared statistic indicated considerable heterogeneity between the studies (81.90%). Four studies had one or more domains at high risk of bias and two studies had concerns regarding applicability. CONCLUSION The diagnostic accuracy of CNNs was comparable to that of healthcare experts in wrist radiographs for investigation of WF. There is a need for studies with a robust reference standard, external data-set validation and investigation of diagnostic performance of healthcare experts aided with CNNs. CLINICAL RELEVANCE STATEMENT DL matches healthcare experts in diagnosing WFs, which potentially benefits patient diagnosis.
Collapse
Affiliation(s)
- V Hansen
- Department of Radiology and Nuclear Medicine, Hospital of South West Jutland, University Hospital of Southern Denmark, Esbjerg, Denmark
| | - J Jensen
- Department of Radiology, Odense University Hospital, Odense, Denmark; Research and Innovation Unit of Radiology, University of Southern Denmark, Odense, Denmark
| | - M W Kusk
- Department of Radiology and Nuclear Medicine, Hospital of South West Jutland, University Hospital of Southern Denmark, Esbjerg, Denmark; Department of Regional Health Research, Faculty of Health Sciences, University of Southern Denmark, Odense, Denmark; Imaging Research Initiative Southwest (IRIS), Hospital of South West Jutland, University Hospital of Southern Denmark, Esbjerg, Denmark; Radiography and Diagnostic Imaging, School of Medicine, University College Dublin, Belfield 4, Dublin, Ireland
| | - O Gerke
- Department of Nuclear Medicine, Odense University Hospital, Odense, Denmark; Department of Clinical Research, University of Southern Denmark, Odense, Denmark
| | - H B Tromborg
- Department of Clinical Research, University of Southern Denmark, Odense, Denmark; Department of Orthopedic Surgery, Odense University Hospital, Odense, Denmark
| | - S Lysdahlgaard
- Department of Radiology and Nuclear Medicine, Hospital of South West Jutland, University Hospital of Southern Denmark, Esbjerg, Denmark; Department of Regional Health Research, Faculty of Health Sciences, University of Southern Denmark, Odense, Denmark; Imaging Research Initiative Southwest (IRIS), Hospital of South West Jutland, University Hospital of Southern Denmark, Esbjerg, Denmark.
| |
Collapse
|
48
|
Mert S, Stoerzer P, Brauer J, Fuchs B, Haas-Lützenberger EM, Demmer W, Giunta RE, Nuernberger T. Diagnostic power of ChatGPT 4 in distal radius fracture detection through wrist radiographs. Arch Orthop Trauma Surg 2024; 144:2461-2467. [PMID: 38578309 PMCID: PMC11093861 DOI: 10.1007/s00402-024-05298-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/09/2024] [Accepted: 03/27/2024] [Indexed: 04/06/2024]
Abstract
Distal radius fractures rank among the most prevalent fractures in humans, necessitating accurate radiological imaging and interpretation for optimal diagnosis and treatment. In addition to human radiologists, artificial intelligence systems are increasingly employed for radiological assessments. Since 2023, ChatGPT 4 has offered image analysis capabilities, which can also be used for the analysis of wrist radiographs. This study evaluates the diagnostic power of ChatGPT 4 in identifying distal radius fractures, comparing it with a board-certified radiologist, a hand surgery resident, a medical student, and the well-established AI Gleamer BoneView™. Results demonstrate ChatGPT 4's good diagnostic accuracy (sensitivity 0.88, specificity 0.98, diagnostic power (AUC) 0.93), surpassing the medical student (sensitivity 0.98, specificity 0.72, diagnostic power (AUC) 0.85; p = 0.04) significantly. Nevertheless, the diagnostic power of ChatGPT 4 lags behind the hand surgery resident (sensitivity 0.99, specificity 0.98, diagnostic power (AUC) 0.985; p = 0.014) and Gleamer BoneView™(sensitivity 1.00, specificity 0.98, diagnostic power (AUC) 0.99; p = 0.006). This study highlights the utility and potential applications of artificial intelligence in modern medicine, emphasizing ChatGPT 4 as a valuable tool for enhancing diagnostic capabilities in the field of medical imaging.
Collapse
Affiliation(s)
- Sinan Mert
- Division of Hand, Plastic and Aesthetic Surgery, LMU University Hospital, LMU Munich, 80336, München, Germany.
| | - Patrick Stoerzer
- Division of Hand, Plastic and Aesthetic Surgery, LMU University Hospital, LMU Munich, 80336, München, Germany
| | - Johannes Brauer
- Division of Hand, Plastic and Aesthetic Surgery, LMU University Hospital, LMU Munich, 80336, München, Germany
| | - Benedikt Fuchs
- Division of Hand, Plastic and Aesthetic Surgery, LMU University Hospital, LMU Munich, 80336, München, Germany
| | | | - Wolfram Demmer
- Division of Hand, Plastic and Aesthetic Surgery, LMU University Hospital, LMU Munich, 80336, München, Germany
| | - Riccardo E Giunta
- Division of Hand, Plastic and Aesthetic Surgery, LMU University Hospital, LMU Munich, 80336, München, Germany
| | - Tim Nuernberger
- Division of Hand, Plastic and Aesthetic Surgery, LMU University Hospital, LMU Munich, 80336, München, Germany
| |
Collapse
|
49
|
Fu T, Viswanathan V, Attia A, Zerbib-Attal E, Kosaraju V, Barger R, Vidal J, Bittencourt LK, Faraji N. Assessing the Potential of a Deep Learning Tool to Improve Fracture Detection by Radiologists and Emergency Physicians on Extremity Radiographs. Acad Radiol 2024; 31:1989-1999. [PMID: 37993303 DOI: 10.1016/j.acra.2023.10.042] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/16/2023] [Revised: 10/23/2023] [Accepted: 10/25/2023] [Indexed: 11/24/2023]
Abstract
RATIONALE AND OBJECTIVES To evaluate the standalone performance of a deep learning (DL) based fracture detection tool on extremity radiographs and assess the performance of radiologists and emergency physicians in identifying fractures of the extremities with and without the DL aid. MATERIALS AND METHODS The DL tool was previously developed using 132,000 appendicular skeletal radiographs divided into 87% training, 11% validation, and 2% test sets. Stand-alone performance was evaluated on 2626 de-identified radiographs from a single institution in Ohio, including at least 140 exams per body region. Consensus from three US board-certified musculoskeletal (MSK) radiologists served as ground truth. A multi-reader retrospective study was performed in which 24 readers (eight each of emergency physicians, non-MSK radiologists, and MSK radiologists) identified fractures in 186 cases during two independent sessions with and without DL aid, separated by a one-month washout period. The accuracy (area under the receiver operating curve), sensitivity, specificity, and reading time were compared with and without model aid. RESULTS The model achieved a stand-alone accuracy of 0.986, sensitivity of 0.987, and specificity of 0.885, and high accuracy (> 0.95) across stratification for body part, age, gender, radiographic views, and scanner type. With DL aid, reader accuracy increased by 0.047 (95% CI: 0.034, 0.061; p = 0.004) and sensitivity significantly improved from 0.865 (95% CI: 0.848, 0.881) to 0.955 (95% CI: 0.944, 0.964). Average reading time was shortened by 7.1 s (27%) per exam. When stratified by physician type, this improvement was greater for emergency physicians and non-MSK radiologists. CONCLUSION The DL tool demonstrated high stand-alone accuracy, aided physician diagnostic accuracy, and decreased interpretation time.
Collapse
Affiliation(s)
- Tianyuan Fu
- University Hospitals Cleveland Medical Center, Case Western Reserve University, Cleveland, Ohio, USA (T.F., V.V., V.K., R.B., L.K.B., N.F.).
| | - Vidya Viswanathan
- University Hospitals Cleveland Medical Center, Case Western Reserve University, Cleveland, Ohio, USA (T.F., V.V., V.K., R.B., L.K.B., N.F.)
| | - Alexandre Attia
- Azmed, 10 Rue d'Uzès, 75002, Paris, France (A.A., E.Z.A., J.V.)
| | | | - Vijaya Kosaraju
- University Hospitals Cleveland Medical Center, Case Western Reserve University, Cleveland, Ohio, USA (T.F., V.V., V.K., R.B., L.K.B., N.F.)
| | - Richard Barger
- University Hospitals Cleveland Medical Center, Case Western Reserve University, Cleveland, Ohio, USA (T.F., V.V., V.K., R.B., L.K.B., N.F.)
| | - Julien Vidal
- Azmed, 10 Rue d'Uzès, 75002, Paris, France (A.A., E.Z.A., J.V.)
| | - Leonardo K Bittencourt
- University Hospitals Cleveland Medical Center, Case Western Reserve University, Cleveland, Ohio, USA (T.F., V.V., V.K., R.B., L.K.B., N.F.)
| | - Navid Faraji
- University Hospitals Cleveland Medical Center, Case Western Reserve University, Cleveland, Ohio, USA (T.F., V.V., V.K., R.B., L.K.B., N.F.)
| |
Collapse
|
50
|
Cheng CT, Kuo LW, Ouyang CH, Hsu CP, Lin WC, Fu CY, Kang SC, Liao CH. Development and evaluation of a deep learning-based model for simultaneous detection and localization of rib and clavicle fractures in trauma patients' chest radiographs. Trauma Surg Acute Care Open 2024; 9:e001300. [PMID: 38646620 PMCID: PMC11029226 DOI: 10.1136/tsaco-2023-001300] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/23/2024] Open
Abstract
Purpose To develop a rib and clavicle fracture detection model for chest radiographs in trauma patients using a deep learning (DL) algorithm. Materials and methods We retrospectively collected 56 145 chest X-rays (CXRs) from trauma patients in a trauma center between August 2008 and December 2016. A rib/clavicle fracture detection DL algorithm was trained using this data set with 991 (1.8%) images labeled by experts with fracture site locations. The algorithm was tested on independently collected 300 CXRs in 2017. An external test set was also collected from hospitalized trauma patients in a regional hospital for evaluation. The receiver operating characteristic curve with area under the curve (AUC), accuracy, sensitivity, specificity, precision, and negative predictive value of the model on each test set was evaluated. The prediction probability on the images was visualized as heatmaps. Results The trained DL model achieved an AUC of 0.912 (95% CI 87.8 to 94.7) on the independent test set. The accuracy, sensitivity, and specificity on the given cut-off value are 83.7, 86.8, and 80.4, respectively. On the external test set, the model had a sensitivity of 88.0 and an accuracy of 72.5. While the model exhibited a slight decrease in accuracy on the external test set, it maintained its sensitivity in detecting fractures. Conclusion The algorithm detects rib and clavicle fractures concomitantly in the CXR of trauma patients with high accuracy in locating lesions through heatmap visualization.
Collapse
Affiliation(s)
- Chi-Tung Cheng
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital Linkou, Taoyuan, Taiwan
- Department of medicine, Chang Gung university, Taoyuan, Taiwan
| | - Ling-Wei Kuo
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital Linkou, Taoyuan, Taiwan
- Department of medicine, Chang Gung university, Taoyuan, Taiwan
| | - Chun-Hsiang Ouyang
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital Linkou, Taoyuan, Taiwan
- Department of medicine, Chang Gung university, Taoyuan, Taiwan
| | - Chi-Po Hsu
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital Linkou, Taoyuan, Taiwan
- Department of medicine, Chang Gung university, Taoyuan, Taiwan
| | - Wei-Cheng Lin
- Department of Electrical Engineering, Chang Gung University, Taoyuan, Taiwan
| | - Chih-Yuan Fu
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital Linkou, Taoyuan, Taiwan
- Department of medicine, Chang Gung university, Taoyuan, Taiwan
| | - Shih-Ching Kang
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital Linkou, Taoyuan, Taiwan
- Department of medicine, Chang Gung university, Taoyuan, Taiwan
| | - Chien-Hung Liao
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital Linkou, Taoyuan, Taiwan
- Department of medicine, Chang Gung university, Taoyuan, Taiwan
| |
Collapse
|