1
|
Wang H, Wang X, Du Y, Wang Y, Bai Z, Wu D, Tang W, Zeng H, Tao J, He J. Prediction of lymph node metastasis in papillary thyroid carcinoma using non-contrast CT-based radiomics and deep learning with thyroid lobe segmentation: A dual-center study. Eur J Radiol Open 2025; 14:100639. [PMID: 40093877 PMCID: PMC11908562 DOI: 10.1016/j.ejro.2025.100639] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2025] [Revised: 02/10/2025] [Accepted: 02/19/2025] [Indexed: 03/19/2025] Open
Abstract
Objectives This study aimed to develop a predictive model for lymph node metastasis (LNM) in papillary thyroid carcinoma (PTC) patients by deep learning radiomic (DLRad) and clinical features. Methods This study included 271 thyroid lobes from 228 PTC patients who underwent preoperative neck non-contrast CT at Center 1 (May 2021-April 2024). LNM status was confirmed via postoperative pathology, with each thyroid lobe labeled accordingly. The cohort was divided into training (n = 189) and validation (n = 82) cohorts, with additional temporal (n = 59 lobes, Center 1, May-August 2024) and external (n = 66 lobes, Center 2) test cohorts. Thyroid lobes were manually segmented from the isthmus midline, ensuring interobserver consistency (ICC ≥ 0.8). Deep learning and radiomics features were selected using LASSO algorithms to compute DLRad scores. Logistic regression identified independent predictors, forming DLRad, clinical, and combined models. Model performance was evaluated using AUC, calibration, decision curves, and the DeLong test, compared against radiologists' assessments. Results Independent predictors of LNM included age, gender, multiple nodules, tumor size group, and DLRad. The combined model demonstrated superior diagnostic performance with AUCs of 0.830 (training), 0.799 (validation), 0.819 (temporal test), and 0.756 (external test), outperforming the DLRad model (AUCs: 0.786, 0.730, 0.753, 0.642), clinical model (AUCs: 0.723, 0.745, 0.671, 0.660), and radiologist evaluations (AUCs: 0.529, 0.606, 0.620, 0.503). It also achieved the lowest Brier scores (0.167, 0.184, 0.175, 0.201) and the highest net benefit in decision-curve analysis at threshold probabilities > 20 %. Conclusions The combined model integrating DLRad and clinical features exhibits good performance in predicting LNM in PTC patients.
Collapse
Affiliation(s)
- Hao Wang
- Department of Radiology, The Fourth Affiliated Hospital of Nanjing Medical University, Nanjing 210031, PR China
| | - Xuan Wang
- Department of Radiology, Zhongda Hospital Southeast University (JiangBei), Nanjing 210048, PR China
| | - Yusheng Du
- Department of Radiology, The Fourth Affiliated Hospital of Nanjing Medical University, Nanjing 210031, PR China
| | - You Wang
- Department of Radiology, The Fourth Affiliated Hospital of Nanjing Medical University, Nanjing 210031, PR China
| | - Zhuojie Bai
- Department of Radiology, The Fourth Affiliated Hospital of Nanjing Medical University, Nanjing 210031, PR China
| | - Di Wu
- Department of Radiology, Zhongda Hospital Southeast University (JiangBei), Nanjing 210048, PR China
| | - Wuliang Tang
- Department of Radiology, Zhongda Hospital Southeast University (JiangBei), Nanjing 210048, PR China
| | - Hanling Zeng
- Department of General Surgery, The Fourth Affiliated Hospital of Nanjing Medical University, Nanjing 210031, PR China
| | - Jing Tao
- Department of General Surgery, The Fourth Affiliated Hospital of Nanjing Medical University, Nanjing 210031, PR China
| | - Jian He
- Department of Nuclear Medicine, Nanjing Drum Tower Hospital, The Affiliated Hospital of Medicine school, Nanjing University, Nanjing 210008, PR China
| |
Collapse
|
2
|
Musa A, Prasad R, Hernandez M. Addressing cross-population domain shift in chest X-ray classification through supervised adversarial domain adaptation. Sci Rep 2025; 15:11383. [PMID: 40181036 PMCID: PMC11968948 DOI: 10.1038/s41598-025-95390-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/08/2025] [Accepted: 03/20/2025] [Indexed: 04/05/2025] Open
Abstract
Medical image analysis, empowered by artificial intelligence (AI), plays a crucial role in modern healthcare diagnostics. However, the effectiveness of machine learning models hinges on their ability to generalize to diverse patient populations, presenting domain shift challenges. This study explores the domain shift problem in chest X-ray classification, focusing on cross-population variations, especially in underrepresented groups. We analyze the impact of domain shifts across three population datasets acting as sources using a Nigerian chest X-ray dataset acting as the target. Model performance is evaluated to assess disparities between source and target populations, revealing large discrepancies when the models trained on a source were applied to the target domain. To address with the evident domain shift among the populations, we propose a supervised adversarial domain adaptation (ADA) technique. The feature extractor is first trained on the source domain using a supervised loss function in ADA. The feature extractor is then frozen, and an adversarial domain discriminator is introduced to distinguish between the source and target domains. Adversarial training fine-tunes the feature extractor, making features from both domains indistinguishable, thereby creating domain-invariant features. The technique was evaluated on the Nigerian dataset, showing significant improvements in chest X-ray classification performance. The proposed model achieved a 90.08% accuracy and a 96% AUC score, outperforming existing approaches such as multi-task learning (MTL) and continual learning (CL). This research highlights the importance of developing domain-aware models in AI-driven healthcare, offering a solution to cross-population domain shift challenges in medical imaging.
Collapse
Affiliation(s)
- Aminu Musa
- Deparment of Computer Science, African University of Science and Technology, Abuja, 900107, Nigeria.
- Department of Computer Science, Federal University Dutse, Dutse, Nigeria.
| | - Rajesh Prasad
- Deparment of Computer Science, African University of Science and Technology, Abuja, 900107, Nigeria
- Department of Computer Science and Engineering, Ajay Kumar Garg Engineering College, Ghaziabad, 201015, India
| | - Monica Hernandez
- Deparment of Computer Science, University of Zaragoza, Zaragoza, 50018, Spain
| |
Collapse
|
3
|
Kaur P, Mahajan P. Detection of brain tumors using a transfer learning-based optimized ResNet152 model in MR images. Comput Biol Med 2025; 188:109790. [PMID: 39951980 DOI: 10.1016/j.compbiomed.2025.109790] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2024] [Revised: 01/28/2025] [Accepted: 01/31/2025] [Indexed: 02/17/2025]
Abstract
Brain tumors are incredibly harmful and can drastically reduce life expectancy. Most researchers use magnetic resonance (MR) scans to detect tumors because they can provide detailed images of the affected area. Recently, AI-based deep learning methods have emerged to enhance diagnostic accuracy through efficient data processing. This study investigates the effectiveness of deep transfer learning techniques for accurate brain tumor diagnosis. A preprocessing pipeline is used to enhance the image quality. This pipeline includes morphological operations such as erosion and dilation for shape refinement, Gaussian blurring for noise reduction, and thresholding for image cropping. Principal Component Analysis (PCA) is applied for dimensionality reduction, and data augmentation enriches the dataset. The dataset is partitioned into training (80 %) and testing (20 %). Pretrained ResNet152 and GoogleNet extract meaningful features from the images. These extracted features are then classified using conventional machine learning classifiers: Support Vector Machine (SVM), K-Nearest Neighbors (KNN), Classification and Regression Trees (CART), and Gaussian Naive Bayes (GNB). This study compares the performance of two pre-trained models for medical image analysis. Performance metrics such as accuracy, sensitivity, recall, and F1-Score evaluate the final classification results. ResNet152 outperforms GoogleNet, achieving a 98.53 % accuracy, an F1 score of 97.4 %, and a sensitivity of 96.52 %. This study highlights integrating deep learning and traditional machine-learning techniques in medical image analysis for effective brain tumor detection.
Collapse
Affiliation(s)
- Prabhpreet Kaur
- Department of Computer Engineering and Technology, Guru Nanak Dev University, Amritsar, India.
| | - Priyanka Mahajan
- Department of Computer Engineering and Technology, Guru Nanak Dev University, Amritsar, India.
| |
Collapse
|
4
|
Zhao F, Liu M, Xiang M, Li D, Jiang X, Jin X, Lin C, Wang R. Unsupervised and Self-supervised Learning in Low-Dose Computed Tomography Denoising: Insights from Training Strategies. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025; 38:902-930. [PMID: 39231886 PMCID: PMC11950483 DOI: 10.1007/s10278-024-01213-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/13/2024] [Revised: 06/30/2024] [Accepted: 07/01/2024] [Indexed: 09/06/2024]
Abstract
In recent years, X-ray low-dose computed tomography (LDCT) has garnered widespread attention due to its significant reduction in the risk of patient radiation exposure. However, LDCT images often contain a substantial amount of noises, adversely affecting diagnostic quality. To mitigate this, a plethora of LDCT denoising methods have been proposed. Among them, deep learning (DL) approaches have emerged as the most effective, due to their robust feature extraction capabilities. Yet, the prevalent use of supervised training paradigms is often impractical due to the challenges in acquiring low-dose and normal-dose CT pairs in clinical settings. Consequently, unsupervised and self-supervised deep learning methods have been introduced for LDCT denoising, showing considerable potential for clinical applications. These methods' efficacy hinges on training strategies. Notably, there appears to be no comprehensive reviews of these strategies. Our review aims to address this gap, offering insights and guidance for researchers and practitioners. Based on training strategies, we categorize the LDCT methods into six groups: (i) cycle consistency-based, (ii) score matching-based, (iii) statistical characteristics of noise-based, (iv) similarity-based, (v) LDCT synthesis model-based, and (vi) hybrid methods. For each category, we delve into the theoretical underpinnings, training strategies, strengths, and limitations. In addition, we also summarize the open source codes of the reviewed methods. Finally, the review concludes with a discussion on open issues and future research directions.
Collapse
Affiliation(s)
- Feixiang Zhao
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Ouhai District, Wenzhou, 325000, Zhejiang, China
- College of Nuclear Technology and Automation Engineering, Chengdu University of Technology, 1 East Third Road, Chengdu, 610059, Sichuan, China
| | - Mingzhe Liu
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Ouhai District, Wenzhou, 325000, Zhejiang, China
- College of Computer Science and Cyber Security, Chengdu University of Technology, 1 East Third Road, Chengdu, 610059, Sichuan, China
| | - Mingrong Xiang
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Ouhai District, Wenzhou, 325000, Zhejiang, China.
- School of Information Technology, Deakin University, Melbourne Burwood Campus, 221 Burwood Hwy, Melbourne, 3125, Victoria, Australia.
| | - Dongfen Li
- College of Computer Science and Cyber Security, Chengdu University of Technology, 1 East Third Road, Chengdu, 610059, Sichuan, China
| | - Xin Jiang
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Ouhai District, Wenzhou, 325000, Zhejiang, China
| | - Xiance Jin
- Department of Radiotherapy Center, The first Affiliated Hospital of Wenzhou Medical University, Ouhai District, Wenzhou, 325000, Zhejiang, China
| | - Cai Lin
- Department of Burn, Wound Repair and Regenerative Medicine Center, The first Affiliated Hospital of Wenzhou Medical University, Ouhai District, Wenzhou, 325000, Zhejiang, China
| | - Ruili Wang
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Ouhai District, Wenzhou, 325000, Zhejiang, China
- School of Mathematical and Computational Science, Massey University, SH17, Albany, 0632, Auckland, New Zealand
| |
Collapse
|
5
|
Brodsky V, Ullah E, Bychkov A, Song AH, Walk EE, Louis P, Rasool G, Singh RS, Mahmood F, Bui MM, Parwani AV. Generative Artificial Intelligence in Anatomic Pathology. Arch Pathol Lab Med 2025; 149:298-318. [PMID: 39836377 DOI: 10.5858/arpa.2024-0215-ra] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 12/20/2024] [Indexed: 01/22/2025]
Abstract
CONTEXT.— Generative artificial intelligence (AI) has emerged as a transformative force in various fields, including anatomic pathology, where it offers the potential to significantly enhance diagnostic accuracy, workflow efficiency, and research capabilities. OBJECTIVE.— To explore the applications, benefits, and challenges of generative AI in anatomic pathology, with a focus on its impact on diagnostic processes, workflow efficiency, education, and research. DATA SOURCES.— A comprehensive review of current literature and recent advancements in the application of generative AI within anatomic pathology, categorized into unimodal and multimodal applications, and evaluated for clinical utility, ethical considerations, and future potential. CONCLUSIONS.— Generative AI demonstrates significant promise in various domains of anatomic pathology, including diagnostic accuracy enhanced through AI-driven image analysis, virtual staining, and synthetic data generation; workflow efficiency, with potential for improvement by automating routine tasks, quality control, and reflex testing; education and research, facilitated by AI-generated educational content, synthetic histology images, and advanced data analysis methods; and clinical integration, with preliminary surveys indicating cautious optimism for nondiagnostic AI tasks and growing engagement in academic settings. Ethical and practical challenges require rigorous validation, prompt engineering, federated learning, and synthetic data generation to help ensure trustworthy, reliable, and unbiased AI applications. Generative AI can potentially revolutionize anatomic pathology, enhancing diagnostic accuracy, improving workflow efficiency, and advancing education and research. Successful integration into clinical practice will require continued interdisciplinary collaboration, careful validation, and adherence to ethical standards to ensure the benefits of AI are realized while maintaining the highest standards of patient care.
Collapse
Affiliation(s)
- Victor Brodsky
- From the Department of Pathology and Immunology, Washington University School of Medicine in St Louis, St Louis, Missouri (Brodsky)
| | - Ehsan Ullah
- the Department of Surgery, Health New Zealand, Counties Manukau, New Zealand (Ullah)
| | - Andrey Bychkov
- the Department of Pathology, Kameda Medical Center, Kamogawa City, Chiba Prefecture, Japan (Bychkov)
- the Department of Pathology, Nagasaki University, Nagasaki, Japan (Bychkov)
| | - Andrew H Song
- the Department of Pathology, Brigham and Women's Hospital, Boston, Massachusetts (Song, Mahmood)
| | - Eric E Walk
- Office of the Chief Medical Officer, PathAI, Boston, Massachusetts (Walk)
| | - Peter Louis
- the Department of Pathology and Laboratory Medicine, Rutgers Robert Wood Johnson Medical School, New Brunswick, New Jersey (Louis)
| | - Ghulam Rasool
- the Department of Oncologic Sciences, Morsani College of Medicine and Department of Electrical Engineering, University of South Florida, Tampa (Rasool)
- the Department of Machine Learning, Moffitt Cancer Center and Research Institute, Tampa, Florida (Rasool)
- Department of Machine Learning, Neuro-Oncology, Moffitt Cancer Center and Research Institute, Tampa, Florida (Rasool)
| | - Rajendra S Singh
- Dermatopathology and Digital Pathology, Summit Health, Berkley Heights, New Jersey (Singh)
| | - Faisal Mahmood
- the Department of Pathology, Brigham and Women's Hospital, Boston, Massachusetts (Song, Mahmood)
| | - Marilyn M Bui
- Department of Machine Learning, Pathology, Moffitt Cancer Center and Research Institute, Tampa, Florida (Bui)
| | - Anil V Parwani
- the Department of Pathology, The Ohio State University, Columbus (Parwani)
| |
Collapse
|
6
|
van Slobbe R, Herrmannova D, Boeke DJ, Lima-Walton ES, Abu-Hanna A, Vagliano I. Multimodal convolutional neural networks for the prediction of acute kidney injury in the intensive care. Int J Med Inform 2025; 196:105815. [PMID: 39914070 DOI: 10.1016/j.ijmedinf.2025.105815] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2024] [Revised: 01/20/2025] [Accepted: 01/26/2025] [Indexed: 02/28/2025]
Abstract
Increased monitoring of health-related data for ICU patients holds great potential for the early prediction of medical outcomes. Research on whether the use of clinical notes and concepts from knowledge bases can improve the performance of prediction models is limited. We investigated the effects of combining clinical variables, clinical notes, and clinical concepts. We focus on the early prediction of Acute Kidney Injury (AKI) in the intensive care unit (ICU). AKI is a sudden reduction in kidney function measured by increased serum creatinine (SCr) or decreased urine output. AKI may occur in up to 30% of ICU stays. We developed three models based on convolutional neural networks using data from the Medical Information Mart for Intensive Care (MIMIC) database. The models used clinical variables, free-text notes, and concepts from the Elsevier H-Graph. Our models achieved good predictive performance (AUROC 0.73-0.90). These models were assessed both when using Scr and urine output as predictors and when omitting them. When Scr and urine output were used as predictors, models that included clinical notes and concepts together with clinical variables performed on par with models that only used clinical variables. When excluding SCr and urine output, predictive performance improved by combining multiple modalities. The models that used only clinical variables were externally validated on the eICU dataset and transported fairly to the new population (AUROC 0.68-0.77). Our in-depth comparison of modalities and text representations may further guide researchers and practitioners in applying multimodal models for predicting AKI and inspire them to investigate multimodality and contextualized embeddings for other tasks. Our models can support clinicians to promptly recognize and treat deteriorating AKI patients and may improve patient outcomes in the ICU.
Collapse
Affiliation(s)
| | | | - D J Boeke
- Elsevier B.V., Amsterdam, the Netherlands
| | | | - A Abu-Hanna
- Department of Medical Informatics, Amsterdam UMC, University of Amsterdam, Amsterdam, the Netherlands; Amsterdam Public Health research institute, Amsterdam, the Netherlands
| | - I Vagliano
- Department of Medical Informatics, Amsterdam UMC, University of Amsterdam, Amsterdam, the Netherlands; Amsterdam Public Health research institute, Amsterdam, the Netherlands.
| |
Collapse
|
7
|
Yuan Z, Wang J, Xu Y, Xu M. CC-TransXNet: a hybrid CNN-transformer network for automatic segmentation of optic cup and optic disk from fundus images. Med Biol Eng Comput 2025; 63:1027-1044. [PMID: 39602064 DOI: 10.1007/s11517-024-03244-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/03/2024] [Accepted: 11/07/2024] [Indexed: 11/29/2024]
Abstract
Accurate segmentation of the optic disk (OD) and optic cup (OC) regions of the optic nerve head is a critical step in glaucoma diagnosis. Existing architectures based on convolutional neural networks (CNNs) still suffer from insufficient global information and poor generalization ability to small sample datasets. Besides, advanced transformer-based models, although capable of capturing global image features, perform poorly in medical image segmentation due to numerous parameters and insufficient local spatial information. To address the above two problems, we propose an innovative W-shaped hybrid network framework, CC-TransXNet, which combines the advantages of CNN and transformer. Firstly, by employing TransXNet and improved ResNet as feature extraction modules, the network considers local and global features to enhance its generalization ability. Secondly, the convolutional block attention module (CBAM) is introduced in the residual structure to improve the ability to recognize the OD and OC by applying attention in both the channel and spatial dimensions. Thirdly, the Contextual Attention (CoT) self-attention mechanism is used in the skip connection to adaptively allocate attention to the contextual information, further enhancing the segmentation's accuracy. We conducted experiments on four publicly available datasets (REFUGE 2, RIM-ONE DL, GAMMA, and Drishti-GS). Compared with the traditional U-Net, CNN, and transformer-based networks, our proposed CC-TransXNet improves the segmentation accuracy and significantly enhances the generalization ability on small datasets. Moreover, CC-TransXNet effectively controls the number of parameters in the model through optimized design to avoid the risk of overfitting, proving its potential for efficient segmentation.
Collapse
Affiliation(s)
- Zhongzheng Yuan
- Department of Software Engineering, Harbin University of Science and Technology, Weihai, 264300, China
| | - Jinke Wang
- Department of Software Engineering, Harbin University of Science and Technology, Weihai, 264300, China.
| | - Yukun Xu
- Department of Software Engineering, Harbin University of Science and Technology, Weihai, 264300, China
| | - Min Xu
- Weihai Municipal Hospital, Affiliated to Shandong University, Weihai, 264299, China
| |
Collapse
|
8
|
Yilihamu EEY, Zeng FS, Shang J, Yang JT, Zhong H, Feng SQ. GPT4LFS (generative pre-trained transformer 4 omni for lumbar foramina stenosis): enhancing lumbar foraminal stenosis image classification through large multimodal models. Spine J 2025:S1529-9430(25)00165-2. [PMID: 40157428 DOI: 10.1016/j.spinee.2025.03.011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/11/2024] [Revised: 03/07/2025] [Accepted: 03/22/2025] [Indexed: 04/01/2025]
Abstract
BACKGROUND CONTEXT Lumbar foraminal stenosis (LFS) is a common spinal condition that requires accurate assessment. Current magnetic resonance imaging (MRI) reporting processes are often inefficient, and while deep learning has potential for improvement, challenges in generalization and interpretability limit its diagnostic effectiveness compared to physician expertise. PURPOSE The present study aimed to leverage a multimodal large language model to improve the accuracy and efficiency of LFS image classification, thereby enabling rapid and precise automated diagnosis, reducing the dependence on manually annotated data, and enhancing diagnostic efficiency. STUDY DESIGN/SETTING Retrospective study conducted from April 2017 to March 2023. PATIENT SAMPLE Sagittal T1-weighted MRI data for the lumbar spine were collected from 1,200 patients across three medical centers. A total of 810 patient cases were included in the final analysis, with data collected from seven different MRI devices. OUTCOME MEASURES Automated classification of LFS using the multi modal large language model. Accuracy, sensitivity, Specificity and Cohen's Kappa coefficient were calculated. METHODS An advanced multimodal fusion framework GPT4LFS was developed with the primary objective of integrating imaging data and natural language descriptions to comprehensively capture the complex LFS features. The model employed a pre-trained ConvNeXt as the image processing module for extracting high-dimensional imaging features. Concurrently, medical descriptive texts generated by the multimodal large language model GPT-4o and encoded and feature-extracted using RoBERTa were utilized to optimize the model's contextual understanding capabilities. The Mamba architecture was implemented during the feature fusion stage, effectively integrating imaging and textual features and thereby enhancing the performance of the classification task. Finally, the stability of the model's detection results was validated by evaluating classification task metrics, such as the accuracy, sensitivity, specificity, and Kappa coefficients. RESULTS The training set comprised 6,299 images from 635 patients, the internal test set included 820 images from 82 patients, and the external test set was composed of 930 images from 93 patients. The GPT4LFS model demonstrated an overall accuracy of 93.7%, sensitivity of 95.8%, and specificity of 94.5% in the internal test set (Kappa = 0.89,95% confidence interval (CI): 0.84-0.96, p<.001). In the external test set, the overall accuracy was 92.2%, with a sensitivity of 92.2% and a specificity of 97.4% (Kappa = 0.88, 95% CI: 0.84-0.89, p<.001). Both the internal and external test sets showed excellent consistency in the model. After the article is published, we will make the full code publicly available on GitHub. CONCLUSIONS Using the GPT4LFS model for LFS image categorization demonstrated accuracy and the capacity for feature description at a level commensurate with that of professional clinicians.
Collapse
Affiliation(s)
- Elzat Elham-Yilizati Yilihamu
- Shandong University, Orthopedic Research Center of Shandong University & Advanced Medical Research Institute, Jinan 250000, China.
| | - Fan-Shuo Zeng
- Department of Rehabilitation of the Second Hospital of Shandong University, Cheeloo College of Medicine, Shandong University, Jinan 250000, China.
| | - Jun Shang
- Renci Hospital of Xuzhou Medical University, Xuzhou 221000, China.
| | - Jin-Tao Yang
- Medical Research Department of Jiangsu Shiyu Intelligent Medical Technology Co., Nanjing 210000, China.
| | - Hai Zhong
- The Second Hospital of Shandong University, Cheeloo College of Medicine, Shandong University, Jinan 250000, China.
| | - Shi-Qing Feng
- The Second Hospital of Shandong University, Cheeloo College of Medicine, Shandong University, Jinan 250000, China.
| |
Collapse
|
9
|
Sarveswaran T, Rajangam V. An ensemble approach using multidimensional convolutional neural networks in wavelet domain for schizophrenia classification from sMRI data. Sci Rep 2025; 15:10257. [PMID: 40133457 PMCID: PMC11937485 DOI: 10.1038/s41598-025-93912-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2024] [Accepted: 03/10/2025] [Indexed: 03/27/2025] Open
Abstract
Schizophrenia is a complicated mental condition marked by disruptions in thought processes, perceptions, and emotional responses, which can cause severe impairment in everyday functioning. sMRI is a non-invasive neuroimaging technology that visualizes the brain's structure while providing precise information on its anatomy and potential problems. This paper investigates the role of multidimensional Convolutional Neural Network (CNN) architectures: 1D-CNN, 2D-CNN and 3D-CNN, using the DWT subbands of sMRI data. 1D-CNN involves energy features extracted from the CD subband of sMRI data. The sum of gradient magnitudes of CD subband, known as energy feature, highlights diagonal high frequency elements associated with schizophrenia. 2D-CNN uses the CH subband decomposed by DWT that enables feature extraction from horizontal high frequency coefficients of sMRI data. In the case of 3D-CNNs, the CV subband is used which leads to volumetric feature extraction from vertical high frequency coefficients. Feature extraction in DWT domain explores textural changes, edges, coarse and fine details present in sMRI data from which the multidimensional feature extraction is carried out for classification.Through maximum voting technique, the proposed model optimizes schizophrenia classification from the multidimensional CNN models. The generalization of the proposed model for the two datasets proves convincing in improving the classification accuracy. The multidimensional CNN architectures achieve an average accuracy of 93.2%, 95.8%, and 98.0%, respectively, while the proposed model achieves an average accuracy of 98.9%.
Collapse
Affiliation(s)
| | - Vijayarajan Rajangam
- Centre for Healthcare Advancement, Innovation and Research, School of Electronics Engineering, Vellore Institute of Technology, Chennai, India.
| |
Collapse
|
10
|
Lam NFD, Cai J, Ng KH. Artificial intelligence and its potential integration with the clinical practice of diagnostic imaging medical physicists: a review. Phys Eng Sci Med 2025:10.1007/s13246-025-01535-z. [PMID: 40126762 DOI: 10.1007/s13246-025-01535-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/30/2024] [Accepted: 03/12/2025] [Indexed: 03/26/2025]
Abstract
Current clinical practice in imaging medical physics is concerned with quality assurance, image processing and analysis, radiation dosimetry, risk assessment and radiation protection, and in-house training and research. Physicist workloads are projected to increase as medical imaging technologies become more sophisticated. Artificial intelligence (AI) is a rising technology with potential to assist medical physicists in their work. Exploration of AI integration into imaging medical physicist workloads is limited. In this review paper, we provide an overview of AI techniques, outline their potential usage in imaging medical physics, and discuss the limitations and challenges to clinical adoption of AI technologies.
Collapse
Affiliation(s)
| | - Jing Cai
- Department of Health Technology and Informatics, Hong Kong Polytechnic University, Hung Hom, Hong Kong
| | - Kwan Hoong Ng
- Department of Biomedical Imaging, Faculty of Medicine, Universiti Malaya, Kuala Lumpur, Malaysia.
- Faculty of Medicine and Health Sciences, UCSI University, Springhill, Negri Sembilan, Malaysia.
| |
Collapse
|
11
|
Sekkat H, Khallouqi A, Rhazouani OE, Halimi A. Automated Detection of Hydrocephalus in Pediatric Head Computed Tomography Using VGG 16 CNN Deep Learning Architecture and Based Automated Segmentation Workflow for Ventricular Volume Estimation. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025:10.1007/s10278-025-01482-x. [PMID: 40108068 DOI: 10.1007/s10278-025-01482-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/30/2024] [Revised: 02/23/2025] [Accepted: 03/11/2025] [Indexed: 03/22/2025]
Abstract
Hydrocephalus, particularly congenital hydrocephalus in infants, remains underexplored in deep learning research. While deep learning has been widely applied to medical image analysis, few studies have specifically addressed the automated classification of hydrocephalus. This study proposes a convolutional neural network (CNN) model based on the VGG16 architecture to detect hydrocephalus in infant head CT images. The model integrates an automated method for ventricular volume extraction, applying windowing, histogram equalization, and thresholding techniques to segment the ventricles from surrounding brain structures. Morphological operations refine the segmentation and contours are extracted for visualization and volume measurement. The dataset consists of 105 head CT scans, each with 60 slices covering the ventricular volume, resulting in 6300 slices. Manual segmentation by three trained radiologists served as the reference standard. The automated method showed a high correlation with manual measurements, with R2 values ranging from 0.94 to 0.99. The mean absolute percentage error (MAPE) ranged 3.99 to 11.13%, while the root mean square error (RRMSE) from 4.56 to 13.74%. To improve model robustness, the dataset was preprocessed, normalized, and augmented with rotation, shifting, zooming, and flipping. The VGG16-based CNN used pre-trained convolutional layers with additional fully connected layers for classification, predicting hydrocephalus or normal labels. Performance evaluation using a multi-split strategy (15 independent splits) achieved a mean accuracy of 90.4% ± 1.2%. This study presents an automated approach for ventricular volume extraction and hydrocephalus detection, offering a promising tool for clinical and research applications with high accuracy and reduced observer bias.
Collapse
Affiliation(s)
- Hamza Sekkat
- Sciences and Engineering of Biomedicals, Biophysics and Health Laboratory, Higher Institute of Health Sciences, Hassan 1st University, Settat, 26000, Morocco.
- Department of Radiotherapy, International Clinic of Settat, Settat, Morocco.
| | - Abdellah Khallouqi
- Sciences and Engineering of Biomedicals, Biophysics and Health Laboratory, Higher Institute of Health Sciences, Hassan 1st University, Settat, 26000, Morocco
- Department of Radiology, Public Hospital of Mediouna, Mediouna, Morocco
- Department of Radiology, Private Clinic Hay Mouhamadi, Casablanca, Morocco
| | - Omar El Rhazouani
- Sciences and Engineering of Biomedicals, Biophysics and Health Laboratory, Higher Institute of Health Sciences, Hassan 1st University, Settat, 26000, Morocco
| | - Abdellah Halimi
- Sciences and Engineering of Biomedicals, Biophysics and Health Laboratory, Higher Institute of Health Sciences, Hassan 1st University, Settat, 26000, Morocco
| |
Collapse
|
12
|
Lee J, Ham S, Kim N, Park HS. Development of a deep learning-based model for guiding a dissection during robotic breast surgery. Breast Cancer Res 2025; 27:34. [PMID: 40065440 PMCID: PMC11895239 DOI: 10.1186/s13058-025-01981-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/08/2025] [Accepted: 02/12/2025] [Indexed: 03/14/2025] Open
Abstract
BACKGROUND Traditional surgical education is based on observation and assistance in surgical practice. Recently introduced deep learning (DL) techniques enable the recognition of the surgical view and automatic identification of surgical landmarks. However, there was no previous studies have conducted to develop surgical guide for robotic breast surgery. To develop a DL model for guiding the dissection plane during robotic mastectomy for beginners and trainees. METHODS Ten surgical videos of robotic mastectomy procedures were recorded. Video frames taken at 1-s intervals were converted to PNG format. The ground truth was manually delineated by two experienced surgeons using ImageJ software. The evaluation metrics were the Dice similarity coefficient (DSC) and Hausdorff distance (HD). RESULTS A total of 8,834 images were extracted from ten surgical videos of robotic mastectomies performed between 2016 and 2020. Skin flap dissection during the robotic mastectomy console time was recorded. The median age and body mass index of the patients was 47.5 (38-52) years and 22.00 (19.30-29.52) kg/m2, respectively, and the median console time was 32 (21-48) min. Among the 8,834 images, 428 were selected and divided into training, validation, and testing datasets at a ratio of 7:1:2. Two experts determined that the DSC of our model was 0.828[Formula: see text]5.28 and 0.818[Formula: see text]6.96, while the HDs were 9.80[Formula: see text]2.57 and 10.32[Formula: see text]1.09. CONCLUSION DL can serve as a surgical guide for beginners and trainees, and can be used as a training tool to enhance surgeons' surgical skills.
Collapse
Affiliation(s)
- Jeea Lee
- Department of Surgery, Uijeongbu Eulji Medical Center, Eulji University, Uijeongbu-si, Gyeonggi-do, South Korea
- Department of Surgery, Graduate School of Medicine, Yonsei University College of Medicine, Seoul, South Korea
| | - Sungwon Ham
- Healthcare Readiness Institute for Unified Korea, Korea University Ansan Hospital, Korea University College of Medicine, 123 Jeokgeum-ro, Danwon-gu, Ansan City, 15355, Gyeonggi-do, South Korea
| | - Namkug Kim
- Department of Convergence Medicine, University of Ulsan College of Medicine, Asan Medical Center, Seoul, South Korea.
- Department of Radiology and Research Institute of Radiology, University of Ulsan College of Medicine, Asan Medical Center, Seoul, South Korea.
| | - Hyung Seok Park
- Department of Surgery, Yonsei University College of Medicine, 50-1, Yonsei-ro, Seodaemun-gu, Seoul, 03722, South Korea.
- Institute for Innovation in Digital Healthcare, Yonsei University, Seoul, South Korea.
| |
Collapse
|
13
|
Qi H, Ni A, Feng Y, Peng Y, Yang B, Li G, Wang J. MBL-TransUNet: Enhancing Mesostructure Segmentation of Textile Composite Images via Multi-Scale Feature Fusion and Boundary Guided Learning. MATERIALS (BASEL, SWITZERLAND) 2025; 18:1215. [PMID: 40141498 PMCID: PMC11943508 DOI: 10.3390/ma18061215] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/11/2025] [Revised: 03/04/2025] [Accepted: 03/06/2025] [Indexed: 03/28/2025]
Abstract
Accurate segmentation is essential for creating digital twins based on volumetric images for high fidelity composite material analysis. Conventional techniques typically require labor-intensive and time-consuming manual effort, restricting their practical use. This paper presents a deep learning model, MBL-TransUNet, to address challenges in accurate tow-tow boundary identification via a Boundary-guided Learning module. Fabrics exhibit periodic characteristics; therefore, a Multi-scale Feature Fusion module was integrated to capture both local details and global patterns, thereby enhancing feature fusion and facilitating the effective integration of information across multiple scales. Furthermore, BatchFormerV2 was used to improve generalization through cross-batch learning. Experimental results show that MBL-TransUNet outperforms TransUNet. MIoU improved by 2.38%. In the zero-shot experiment, MIoU increased by 4.23%. The model demonstrates higher accuracy and robustness compared to existing methods. Ablation studies confirm that integrating these modules achieves optimal segmentation performance.
Collapse
Affiliation(s)
- Hang Qi
- School of Material Science and Engineering, Wuhan University of Technology, 122 Luoshi Road, Hongshan District, Wuhan 430070, China; (H.Q.); (Y.F.); (Y.P.); (J.W.)
| | - Aiqing Ni
- State Key Laboratory of Advanced Technology for Materials Synthesis and Processing, Wuhan University of Technology, 122 Luoshi Road, Hongshan District, Wuhan 430070, China
| | - Yuwei Feng
- School of Material Science and Engineering, Wuhan University of Technology, 122 Luoshi Road, Hongshan District, Wuhan 430070, China; (H.Q.); (Y.F.); (Y.P.); (J.W.)
| | - Yunsong Peng
- School of Material Science and Engineering, Wuhan University of Technology, 122 Luoshi Road, Hongshan District, Wuhan 430070, China; (H.Q.); (Y.F.); (Y.P.); (J.W.)
- Luoyang Ship Material Research Institute, 169 Binhe South Road, Luolong District, Luoyang 471039, China;
| | - Bin Yang
- School of Material Science and Engineering, Wuhan University of Technology, 122 Luoshi Road, Hongshan District, Wuhan 430070, China; (H.Q.); (Y.F.); (Y.P.); (J.W.)
| | - Guo Li
- Luoyang Ship Material Research Institute, 169 Binhe South Road, Luolong District, Luoyang 471039, China;
| | - Jihui Wang
- School of Material Science and Engineering, Wuhan University of Technology, 122 Luoshi Road, Hongshan District, Wuhan 430070, China; (H.Q.); (Y.F.); (Y.P.); (J.W.)
| |
Collapse
|
14
|
Liu X, Li B, Vernooij MW, Wolvius EB, Roshchupkin GV, Bron EE. AI-based association analysis for medical imaging using latent-space geometric confounder correction. Med Image Anal 2025; 102:103529. [PMID: 40073582 DOI: 10.1016/j.media.2025.103529] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2024] [Revised: 01/09/2025] [Accepted: 02/24/2025] [Indexed: 03/14/2025]
Abstract
This study addresses the challenges of confounding effects and interpretability in artificial-intelligence-based medical image analysis. Whereas existing literature often resolves confounding by removing confounder-related information from latent representations, this strategy risks affecting image reconstruction quality in generative models, thus limiting their applicability in feature visualization. To tackle this, we propose a different strategy that retains confounder-related information in latent representations while finding an alternative confounder-free representation of the image data. Our approach views the latent space of an autoencoder as a vector space, where imaging-related variables, such as the learning target (t) and confounder (c), have a vector capturing their variability. The confounding problem is addressed by searching a confounder-free vector which is orthogonal to the confounder-related vector but maximally collinear to the target-related vector. To achieve this, we introduce a novel correlation-based loss that not only performs vector searching in the latent space, but also encourages the encoder to generate latent representations linearly correlated with the variables. Subsequently, we interpret the confounder-free representation by sampling and reconstructing images along the confounder-free vector. The efficacy and flexibility of our proposed method are demonstrated across three applications, accommodating multiple confounders and utilizing diverse image modalities. Results affirm the method's effectiveness in reducing confounder influences, preventing wrong or misleading associations, and offering a unique visual interpretation for in-depth investigations by clinical and epidemiological researchers. The code is released in the following GitLab repository: https://gitlab.com/radiology/compopbio/ai_based_association_analysis.
Collapse
Affiliation(s)
- Xianjing Liu
- Department of Radiology and Nuclear Medicine, Erasmus University Medical Center, Rotterdam, The Netherlands; Department of Oral and Maxillofacial Surgery, Erasmus University Medical Center, Rotterdam, The Netherlands.
| | - Bo Li
- Harvard Medical School, Boston, MA, USA
| | - Meike W Vernooij
- Department of Radiology and Nuclear Medicine, Erasmus University Medical Center, Rotterdam, The Netherlands; Department of Epidemiology, Erasmus University Medical Center, Rotterdam, The Netherlands
| | - Eppo B Wolvius
- Department of Oral and Maxillofacial Surgery, Erasmus University Medical Center, Rotterdam, The Netherlands
| | - Gennady V Roshchupkin
- Department of Radiology and Nuclear Medicine, Erasmus University Medical Center, Rotterdam, The Netherlands; Department of Epidemiology, Erasmus University Medical Center, Rotterdam, The Netherlands
| | - Esther E Bron
- Department of Radiology and Nuclear Medicine, Erasmus University Medical Center, Rotterdam, The Netherlands
| |
Collapse
|
15
|
Liu Y, Zhang S, Tang Y, Zhao X, He ZX. A multi-scale pyramid residual weight network for medical image fusion. Quant Imaging Med Surg 2025; 15:1793-1821. [PMID: 40160660 PMCID: PMC11948399 DOI: 10.21037/qims-24-851] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/26/2024] [Accepted: 01/09/2025] [Indexed: 04/02/2025]
Abstract
Background Due to the inherent limitations of imaging sensors, acquiring medical images that simultaneously provide functional metabolic information and detailed structural organization remains a significant challenge. Multi-modal image fusion has emerged as a critical technology for clinical diagnosis and surgical navigation, as it enables the integration of complementary information from different imaging modalities. However, existing deep learning (DL)-based fusion methods often face difficulties in effectively combining high-frequency detail information with low-frequency contextual information, which frequently leads to the degradation of high-frequency details. Therefore, there is a pressing need for a method that addresses these challenges, preserving both high- and low-frequency information while maintaining clear structural contours. In response to this issue, a novel convolutional neural network (CNN), named the multi-scale pyramid residual weight network (LYWNet), is proposed. The objective of this approach is to improve the fusion process by effectively integrating high- and low-frequency information, thereby enhancing the quality and accuracy of multimodal image fusion. This method aims to overcome the limitations of current fusion techniques and ensure the preservation of both functional and structural details, ultimately contributing to more precise clinical diagnoses and better surgical navigation outcomes. Methods We propose a novel CNN, LYWNet, designed to address these challenges. LYWNet is composed of three modules: (I) data preprocessing module: utilizes three convolutional layers to extract both deep and shallow features from the input images. (II) Feature extraction module: incorporates three identical multi-scale pyramid residual weight (LYW) blocks in series, each featuring three interactive branches to preserve high-frequency detail information effectively. (III) Image reconstruction module: utilizes a fusion algorithm based on feature distillation to ensure the effective integration of functional and anatomical information. The proposed image fusion algorithm enhances the interaction of contextual cues and retains the metabolic details from functional images while preserving texture details from anatomical images. Results The proposed LYWNet demonstrated its ability to retain high-frequency details during feature extraction, effectively combining them with low-frequency contextual information. The fusion results exhibited reduced differences between the fused image and the original images. The structural similarity (SSIM) and peak signal-to-noise ratio (PSNR) were 0.5592±0.0536 and 17.3594±1.0211, respectively, for single-photon emission computed tomography-magnetic resonance imaging (SPECT-MRI), 0.5195±0.0730 and 14.5324±1.7365 for PET-MRI; 0.5376±0.0442 and 13.9202±0.7265 for magnetic resonance imaging-computed tomography. Conclusions LYWNet excels at integrating high-frequency detail information and low-frequency contextual information, addressing the deficiencies of existing DL-based image fusion methods. This approach provides superior fused images that retain the functional metabolic information and anatomical texture, making it a valuable tool for clinical diagnosis and surgical navigation.
Collapse
Affiliation(s)
- Yiwei Liu
- Center for Biomedical Imaging Research, School of Biomedical Engineering, Tsinghua University, Beijing, China
| | - Shaoze Zhang
- Center for Biomedical Imaging Research, School of Biomedical Engineering, Tsinghua University, Beijing, China
| | - Yao Tang
- Department of Biomedical Engineering, Tsinghua University, Beijing, China
| | - Xihai Zhao
- Center for Biomedical Imaging Research, School of Biomedical Engineering, Tsinghua University, Beijing, China
| | - Zuo-Xiang He
- School of Clinical Medicine, Tsinghua University, Beijing, China
- Department of Nuclear Medicine, Beijing Tsinghua Changgung Hospital, Beijing, China
| |
Collapse
|
16
|
Deebani W, Aziz L, Aziz A, Basri WS, Alawad WM, Althubiti SA. Synergistic transfer learning and adversarial networks for breast cancer diagnosis: benign vs. invasive classification. Sci Rep 2025; 15:7461. [PMID: 40032913 DOI: 10.1038/s41598-025-90288-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2024] [Accepted: 02/11/2025] [Indexed: 03/05/2025] Open
Abstract
Current breast cancer diagnosis methods often face limitations such as high cost, time consumption, and inter-observer variability. To address these challenges, this research proposes a novel deep learning framework that leverages generative adversarial networks (GANs) for data augmentation and transfer learning to enhance breast cancer classification using convolutional neural networks (CNNs). The framework uses a two-stage augmentation approach. First, a conditional Wasserstein GAN (cWGAN) generates synthetic breast cancer images based on clinical data, enhancing training stability and enabling targeted feature incorporation. Second, traditional augmentation techniques (e.g., rotation, flipping, cropping) are applied to both original and synthetic images. A multi-scale transfer learning technique is also employed, integrating three pre-trained CNNs (DenseNet-201, NasNetMobile, ResNet-101) with a multi-scale feature enrichment scheme, allowing the model to capture features at various scales. The framework was evaluated on the BreakHis dataset, achieving an accuracy of 99.2% for binary classification and 98.5% for multi-class classification, significantly outperforming existing methods. This framework offers a more efficient, cost-effective, and accurate approach for breast cancer diagnosis. Future work will focus on generalizing the framework to clinical datasets and integrating it into diagnostic workflows.
Collapse
Affiliation(s)
- Wejdan Deebani
- Department of Mathematics, College of Science and Arts, King Abdul Aziz University, 21911, Rabigh, Saudi Arabia
| | - Lubna Aziz
- Department of Artificial Intelligence, FEST Iqra University Karachi, Karachi, Pakistan.
- Faculty of Computing, Universiti Teknologi Malaysia, Johor Bahru, Johor, Malaysia.
| | - Arshad Aziz
- Department of Artificial Intelligence, FEST Iqra University Karachi, Karachi, Pakistan
| | - Wael Sh Basri
- College of Business Administration, Management Information System, Northern Border University, Arar, Saudi Arabia
| | - Wedad M Alawad
- Department of Information Technology, College of Computer, Qassim University, Buraydah, 51452, Saudi Arabia
| | - Sara A Althubiti
- Department of Computer Science, College of Computer and Information Sciences, Majmaah University, 11952, Al-Majmaah, Saudi Arabia
| |
Collapse
|
17
|
Gravante G, Arosio AD, Curti N, Biondi R, Berardi L, Gandolfi A, Turri-Zanoni M, Castelnuovo P, Remondini D, Bignami M. Artificial intelligence and MRI in sinonasal tumors discrimination: where do we stand? Eur Arch Otorhinolaryngol 2025; 282:1557-1566. [PMID: 39719474 DOI: 10.1007/s00405-024-09169-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2024] [Accepted: 12/13/2024] [Indexed: 12/26/2024]
Abstract
BACKGROUND Artificial intelligence (AI) demonstrates high potential when applied to radiomic analysis of magnetic resonance imaging (MRI) to discriminate sinonasal tumors. This can enhance diagnostic suspicion beyond visual assessment alone and prior to biopsy, leading to expedite the diagnostic timeline and the treatment planning. The aim of the present work is to evaluate the current advancements and accuracy of this technology in this domain. METHODS A systematic literature review was conducted following PRISMA guidelines. Inclusion criteria comprised studies utilizing any machine learning approach applied to MRI of patients with sinonasal tumors. For each study, comprehensive data were gathered on the MRI protocols, feature extraction techniques, and classifiers employed to develop the AI model. The performance was assessed based on accuracy and area under the curve (AUC). RESULTS Fourteen studies, published between May 2017 and August 2024, were included. These studies were categorized into three groups: those examining both benign and malignant tumors, those investigating malignant tumor subpopulations, and those focusing on benign pathologies. All studies reported an AUC greater than 0.800, achieving AUC > 0.89 and accuracy > 0.81 when incorporating clinical-radiological variables. Notably, the best discrimination performance was observed in studies utilizing combined conventional MRI sequences, including T1-weighted, contrasted T1-weighted, and T2-weighted images. CONCLUSION The application of AI and radiomics in analyzing MRI scans presents significant promise for improving the discrimination of sinonasal tumors. Integrating clinical and radiological indicators enhances model performance, suggesting that future research should focus on larger patient cohorts and diverse AI methodologies to refine diagnostic accuracy and clinical utility.
Collapse
Affiliation(s)
- Giacomo Gravante
- Division of Otorhinolaryngology, Department of Biotechnology and Life Sciences, University of Insubria, Ospedale di Circolo, Varese, Italy.
- Unit of Otorhinolaryngology, Department of Biotechnology and Life Sciences, Ospedale di Circolo e Fondazione Macchi, University of Insubria, Via Guicciardini 9, Varese, 21100, Italy.
| | - Alberto Daniele Arosio
- Division of Otorhinolaryngology, Department of Biotechnology and Life Sciences, University of Insubria, Ospedale di Circolo, Varese, Italy
| | - Nico Curti
- Department of Physics and Astronomy, University of Bologna, Bologna, Italy
| | - Riccardo Biondi
- Data Science and Bioinformatics Laboratory, IRCCS Institute of Neurological Sciences of Bologna, Bologna, Italy
| | - Luigi Berardi
- Division of Otorhinolaryngology, Department of Biotechnology and Life Sciences, University of Insubria, Ospedale di Circolo, Varese, Italy
| | - Alberto Gandolfi
- Division of Otorhinolaryngology, Department of Biotechnology and Life Sciences, University of Insubria, Ospedale di Circolo, Varese, Italy
| | - Mario Turri-Zanoni
- Division of Otorhinolaryngology, Department of Biotechnology and Life Sciences, University of Insubria, Ospedale Sant'Anna, Como, Italy
- Head and Neck Surgery & Forensic Dissection Research Center (HNS&FDRc), Department of Biotechnology and Life Sciences, University of Insubria, Varese, Italy
| | - Paolo Castelnuovo
- Division of Otorhinolaryngology, Department of Biotechnology and Life Sciences, University of Insubria, Ospedale di Circolo, Varese, Italy
- Head and Neck Surgery & Forensic Dissection Research Center (HNS&FDRc), Department of Biotechnology and Life Sciences, University of Insubria, Varese, Italy
| | - Daniel Remondini
- Department of Physics and Astronomy, University of Bologna, Bologna, Italy
| | - Maurizio Bignami
- Division of Otorhinolaryngology, Department of Biotechnology and Life Sciences, University of Insubria, Ospedale di Circolo, Varese, Italy
- Head and Neck Surgery & Forensic Dissection Research Center (HNS&FDRc), Department of Biotechnology and Life Sciences, University of Insubria, Varese, Italy
| |
Collapse
|
18
|
Song B, Liang R. Integrating artificial intelligence with smartphone-based imaging for cancer detection in vivo. Biosens Bioelectron 2025; 271:116982. [PMID: 39616900 PMCID: PMC11789447 DOI: 10.1016/j.bios.2024.116982] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/13/2024] [Revised: 11/19/2024] [Accepted: 11/20/2024] [Indexed: 01/03/2025]
Abstract
Cancer is a major global health challenge, accounting for nearly one in six deaths worldwide. Early diagnosis significantly improves survival rates and patient outcomes, yet in resource-limited settings, the scarcity of medical resources often leads to late-stage diagnosis. Integrating artificial intelligence (AI) with smartphone-based imaging systems offers a promising solution by providing portable, cost-effective, and widely accessible tools for early cancer detection. This paper introduces advanced smartphone-based imaging systems that utilize various imaging modalities for in vivo detection of different cancer types and highlights the advancements of AI for in vivo cancer detection in smartphone-based imaging. However, these compact smartphone systems face challenges like low imaging quality and restricted computing power. The use of advanced AI algorithms to address the optical and computational limitations of smartphone-based imaging systems provides promising solutions. AI-based cancer detection also faces challenges. Transparency and reliability are critical factors in gaining the trust and acceptance of AI algorithms for clinical application, explainable and uncertainty-aware AI breaks the black box and will shape the future AI development in early cancer detection. The challenges and solutions for improving AI accuracy, transparency, and reliability are general issues in AI applications, the AI technologies, limitations, and potentials discussed in this paper are applicable to a wide range of biomedical imaging diagnostics beyond smartphones or cancer-specific applications. Smartphone-based multimodal imaging systems and deep learning algorithms for multimodal data analysis are also growing trends, as this approach can provide comprehensive information about the tissue being examined. Future opportunities and perspectives of AI-integrated smartphone imaging systems will be to make cutting-edge diagnostic tools more affordable and accessible, ultimately enabling early cancer detection for a broader population.
Collapse
Affiliation(s)
- Bofan Song
- Wyant College of Optical Sciences, The University of Arizona, Tucson, AZ, 85721, USA.
| | - Rongguang Liang
- Wyant College of Optical Sciences, The University of Arizona, Tucson, AZ, 85721, USA.
| |
Collapse
|
19
|
Shou Q, Zhao C, Shao X, Herting MM, Wang DJ. High resolution multi-delay arterial spin labeling with self-supervised deep learning denoising for pediatric choroid plexus perfusion MRI. Neuroimage 2025; 308:121070. [PMID: 39889809 DOI: 10.1016/j.neuroimage.2025.121070] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2024] [Revised: 11/21/2024] [Accepted: 01/29/2025] [Indexed: 02/03/2025] Open
Abstract
Choroid plexus (CP) is an important brain structure that produces cerebrospinal fluid (CSF). CP perfusion has been studied using multi-delay arterial spin labeling (MD-ASL) in adults but not in pediatric populations due to the challenge of small CP size in children. Here we present a high resolution (iso2 mm) MDASL protocol with 10-minute scan time and performed test-retest scans on 21 typically developing children aged 8 to 17 years. We further proposed a Transformer-based deep learning (DL) model with k-space weighted image average (KWIA) denoised images as reference for training the model. The performance of the model was evaluated by the SNR, bias and repeatability of the fitted perfusion parameters of the CP and gray matter. The proposed method was compared to several benchmark methods including KWIA, joint denoising and reconstruction with total generalized variation (TGV) regularization, as well as another self-supervised method termed Noise2Void. The results show that the proposed Transformer model with KWIA reference can effectively denoise multi-delay ASL images, not only improving the SNR for perfusion images of each delay, but also improving the SNR for the fitted perfusion maps for visualizing and quantifying CP perfusion in children. This may facilitate the use of MDASL in neurodevelopmental studies to characterize the development of CP and glymphatic system.
Collapse
Affiliation(s)
- Qinyang Shou
- Laboratory of Functional MRI Technology (LOFT), Stevens Neuroimaging and Informatics Institute, University of Southern California, Los Angeles, CA, United States
| | - Chenyang Zhao
- Laboratory of Functional MRI Technology (LOFT), Stevens Neuroimaging and Informatics Institute, University of Southern California, Los Angeles, CA, United States
| | - Xingfeng Shao
- Laboratory of Functional MRI Technology (LOFT), Stevens Neuroimaging and Informatics Institute, University of Southern California, Los Angeles, CA, United States
| | - Megan M Herting
- Department of Population and Public Health Sciences, Keck School of Medicine, University of Southern California, United States
| | - Danny Jj Wang
- Laboratory of Functional MRI Technology (LOFT), Stevens Neuroimaging and Informatics Institute, University of Southern California, Los Angeles, CA, United States.
| |
Collapse
|
20
|
Veturi YA, McNamara S, Kinder S, Clark CW, Thakuria U, Bearce B, Manoharan N, Mandava N, Kahook MY, Singh P, Kalpathy-Cramer J. EyeLiner: A Deep Learning Pipeline for Longitudinal Image Registration Using Fundus Landmarks. OPHTHALMOLOGY SCIENCE 2025; 5:100664. [PMID: 39877463 PMCID: PMC11773051 DOI: 10.1016/j.xops.2024.100664] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/01/2024] [Revised: 08/28/2024] [Accepted: 11/18/2024] [Indexed: 01/31/2025]
Abstract
Objective Detecting and measuring changes in longitudinal fundus imaging is key to monitoring disease progression in chronic ophthalmic diseases, such as glaucoma and macular degeneration. Clinicians assess changes in disease status by either independently reviewing or manually juxtaposing longitudinally acquired color fundus photos (CFPs). Distinguishing variations in image acquisition due to camera orientation, zoom, and exposure from true disease-related changes can be challenging. This makes manual image evaluation variable and subjective, potentially impacting clinical decision-making. We introduce our deep learning (DL) pipeline, "EyeLiner," for registering, or aligning, 2-dimensional CFPs. Improved alignment of longitudinal image pairs may compensate for differences that are due to camera orientation while preserving pathological changes. Design EyeLiner registers a "moving" image to a "fixed" image using a DL-based keypoint matching algorithm. Participants We evaluate EyeLiner on 3 longitudinal data sets: Fundus Image REgistration (FIRE), sequential images for glaucoma forecast (SIGF), and our internal glaucoma data set from the Colorado Ophthalmology Research Information System (CORIS). Methods Anatomical keypoints along the retinal blood vessels were detected from the moving and fixed images using a convolutional neural network and subsequently matched using a transformer-based algorithm. Finally, transformation parameters were learned using the corresponding keypoints. Main Outcome Measures We computed the mean distance (MD) between manually annotated keypoints from the fixed and the registered moving image. For comparison to existing state-of-the-art retinal registration approaches, we used the mean area under the curve (AUC) metric introduced in the FIRE data set study. Results EyeLiner effectively aligns longitudinal image pairs from FIRE, SIGF, and CORIS, as qualitatively evaluated through registration checkerboards and flicker animations. Quantitative results show that the MD decreased for this model after alignment from 321.32 to 3.74 pixels for FIRE, 9.86 to 2.03 pixels for CORIS, and 25.23 to 5.94 pixels for SIGF. We also obtained an AUC of 0.85, 0.94, and 0.84 on FIRE, CORIS, and SIGF, respectively, beating the current state-of-the-art SuperRetina (AUCFIRE = 0.76, AUCCORIS = 0.83, AUCSIGF = 0.74). Conclusions Our pipeline demonstrates improved alignment of image pairs in comparison to the current state-of-the-art methods on 3 separate data sets. We envision that this method will enable clinicians to align image pairs and better visualize changes in disease over time. Financial Disclosures Proprietary or commercial disclosure may be found in the Footnotes and Disclosures at the end of this article.
Collapse
Affiliation(s)
| | | | - Scott Kinder
- Department of Ophthalmology, University of Colorado Anschutz Medical Campus, Aurora, Colorado
| | | | - Upasana Thakuria
- Department of Ophthalmology, University of Colorado Anschutz Medical Campus, Aurora, Colorado
| | - Benjamin Bearce
- Department of Ophthalmology, University of Colorado Anschutz Medical Campus, Aurora, Colorado
| | - Niranjan Manoharan
- Department of Ophthalmology, University of Colorado Anschutz Medical Campus, Aurora, Colorado
| | - Naresh Mandava
- Department of Ophthalmology, University of Colorado Anschutz Medical Campus, Aurora, Colorado
| | - Malik Y. Kahook
- Department of Ophthalmology, University of Colorado Anschutz Medical Campus, Aurora, Colorado
| | - Praveer Singh
- Department of Ophthalmology, University of Colorado Anschutz Medical Campus, Aurora, Colorado
| | | |
Collapse
|
21
|
Firouzabadi SR, Tavanaei R, Mohammadi I, Alikhani A, Ansari A, Akhlaghpasand M, Hajikarimloo B, Yong RL, Margetis K. Performance of Machine Learning Models in Predicting BRAF Alterations Using Imaging Data in Low-Grade Glioma: A Systematic Review and Meta-Analysis. World Neurosurg 2025; 195:123742. [PMID: 39914655 DOI: 10.1016/j.wneu.2025.123742] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/21/2025] [Accepted: 01/24/2025] [Indexed: 02/28/2025]
Abstract
BACKGROUND Understanding the BRAF alterations preoperatively could remarkably assist in predicting tumor behavior, which leads to a more precise prognostication and management strategy. Recent advances in artificial intelligence (AI) have resulted in effective predictive models. Therefore, for the first time, this study aimed to review the performance of machine learning and deep learning models in predicting the BRAF alterations in low-grade gliomas (LGGs)using imaging data. METHODS PubMed/MEDLINE, Embase, and the Cochrane Library were systematically searched for studies published up to June 1, 2024, and evaluated the performance of AI models in predicting BRAF alterations in LGGs using imaging data. Pooled sensitivity, specificity, and area under the receiver operating curve (AUROC) were meta-analyzed. RESULTS A total of 6 studies with 951 patients were included in this systematic review. The pooled AUROC of internal validation cohorts for their best-performing model was 84.44%, with models detecting BRAF mutation, BRAF fusion, BRAF fusion from mutation, and BRAF wild type producing similar AUROCs of 90.75%, 84.59%, 82.33%, and 82%, respectively. The best-performing models had pooled sensitivities of 80.3%, 87.51%, and 74.14% and pooled specificities of 88.57%, 70.41%, and 83.98% for detection of BRAF fusion from mutation, BRAF fusion, and BRAF mutation, respectively. CONCLUSIONS AI models may perform relatively well in predicting BRAF alterations in LGG using imaging data and appear to be capable of high sensitivities and specificities. However, future studies with larger sample sizes implementing different machine learning or deep learning algorithms are required to reduce imprecision.
Collapse
Affiliation(s)
- Shahryar Rajai Firouzabadi
- Functional Neurosurgery Research Center, Shohada Tajrish Comprehensive Neurosurgical Center of Excellence, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Roozbeh Tavanaei
- Functional Neurosurgery Research Center, Shohada Tajrish Comprehensive Neurosurgical Center of Excellence, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Ida Mohammadi
- Functional Neurosurgery Research Center, Shohada Tajrish Comprehensive Neurosurgical Center of Excellence, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Alireza Alikhani
- Functional Neurosurgery Research Center, Shohada Tajrish Comprehensive Neurosurgical Center of Excellence, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Ali Ansari
- Student Research Committee, Shiraz University of Medical Sciences, Shiraz, Iran
| | - Mohammadhosein Akhlaghpasand
- Functional Neurosurgery Research Center, Shohada Tajrish Comprehensive Neurosurgical Center of Excellence, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Bardia Hajikarimloo
- Department of Neurological Surgery, University of Virginia, Charlottesville, Virginia, USA
| | - Raymund L Yong
- Department of Neurosurgery, Mount Sinai Hospital, Icahn School of Medicine, New York City, New York, USA
| | - Konstantinos Margetis
- Department of Neurosurgery, Mount Sinai Hospital, Icahn School of Medicine, New York City, New York, USA.
| |
Collapse
|
22
|
Sufian MA, Niu M. Hybrid deep learning for computational precision in cardiac MRI segmentation: Integrating Autoencoders, CNNs, and RNNs for enhanced structural analysis. Comput Biol Med 2025; 186:109597. [PMID: 39967188 DOI: 10.1016/j.compbiomed.2024.109597] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2024] [Revised: 12/07/2024] [Accepted: 12/17/2024] [Indexed: 02/20/2025]
Abstract
Recent advancements in cardiac imaging have been significantly enhanced by integrating deep learning models, offering transformative potential in early diagnosis and patient care. The research paper explores the application of hybrid deep learning methodologies, focusing on the roles of Autoencoders, Convolutional Neural Networks (CNNs), and Recurrent Neural Networks (RNNs) in enhancing cardiac image analysis. The study implements a comprehensive approach, combining traditional algorithms such as Sobel, Watershed, and Otsu's Thresholding with advanced deep learning models to achieve precise and accurate imaging outcomes. The Autoencoder model, developed for image enhancement and feature extraction, achieved a notable accuracy of 99.66% on the test data. Optimized for image recognition tasks, the CNN model demonstrated a high precision rate of 98.9%. The RNN model, utilized for sequential data analysis, showed a prediction accuracy of 98%, further underscoring the robustness of the hybrid framework. The research drew upon a diverse range of academic databases and pertinent publications within cardiac imaging and deep learning, focusing on peer-reviewed articles and studies published in the past five years. Models were implemented using the TensorFlow and Keras frameworks. The proposed methodology was evaluated in the clinical validation phase using advanced imaging protocols, including the QuickScan technique and balanced steady-state free precession (bSSFP) imaging. The validation metrics were promising: the Signal-to-Noise Ratio (SNR) was improved by 15%, the Contrast-to-Noise Ratio (CNR) saw an enhancement of 12%, and the ejection fraction (EF) analysis provided a 95% correlation with manually segmented data. These metrics confirm the efficacy of the models, showing significant improvements in image quality and diagnostic accuracy. The integration of adversarial defense strategies, such as adversarial training and model ensembling, have been analyzed to enhance model robustness against malicious inputs. The reliability and comparison of the model's ability have been investigated to maintain clinical integrity, even in adversarial attacks that could otherwise compromise segmentation outcomes. These findings indicate that integrating Autoencoders, CNNs, and RNNs within a hybrid deep-learning framework is promising for enhancing cardiac MRI segmentation and early diagnosis. The study contributes to the field by demonstrating the applicability of these advanced techniques in clinical settings, paving the way for improved patient outcomes through more accurate and timely diagnoses.
Collapse
Affiliation(s)
- Md Abu Sufian
- Shaanxi International Innovation Center for Transportation-Energy-Information Fusion and Sustainability, Chang'an University, Xi'an 710064, China; IVR Low-Carbon Research Institute, School of Energy and Electrical Engineering, Chang'an University, Xi'an 710064, China
| | - Mingbo Niu
- Shaanxi International Innovation Center for Transportation-Energy-Information Fusion and Sustainability, Chang'an University, Xi'an 710064, China; IVR Low-Carbon Research Institute, School of Energy and Electrical Engineering, Chang'an University, Xi'an 710064, China.
| |
Collapse
|
23
|
Qian X, Pei J, Han C, Liang Z, Zhang G, Chen N, Zheng W, Meng F, Yu D, Chen Y, Sun Y, Zhang H, Qian W, Wang X, Er Z, Hu C, Zheng H, Shen D. A multimodal machine learning model for the stratification of breast cancer risk. Nat Biomed Eng 2025; 9:356-370. [PMID: 39633027 DOI: 10.1038/s41551-024-01302-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/11/2023] [Accepted: 10/31/2024] [Indexed: 12/07/2024]
Abstract
Machine learning models for the diagnosis of breast cancer can facilitate the prediction of cancer risk and subsequent patient management among other clinical tasks. For the models to impact clinical practice, they ought to follow standard workflows, help interpret mammography and ultrasound data, evaluate clinical contextual information, handle incomplete data and be validated in prospective settings. Here we report the development and testing of a multimodal model leveraging mammography and ultrasound modules for the stratification of breast cancer risk based on clinical metadata, mammography and trimodal ultrasound (19,360 images of 5,216 breasts) from 5,025 patients with surgically confirmed pathology across medical centres and scanner manufacturers. Compared with the performance of experienced radiologists, the model performed similarly at classifying tumours as benign or malignant and was superior at pathology-level differential diagnosis. With a prospectively collected dataset of 191 breasts from 187 patients, the overall accuracies of the multimodal model and of preliminary pathologist-level assessments of biopsied breast specimens were similar (90.1% vs 92.7%, respectively). Multimodal models may assist diagnosis in oncology.
Collapse
Affiliation(s)
- Xuejun Qian
- School of Biomedical Engineering, ShanghaiTech University, Shanghai, China.
- State Key Laboratory of Advanced Medical Materials and Devices, ShanghaiTech University, Shanghai, China.
- Shanghai United Imaging Intelligence Co. Ltd., Shanghai, China.
| | - Jing Pei
- Department of Breast Surgery, The First Affiliated Hospital of Anhui Medical University, Hefei, China
- Department of General Surgery, The First Affiliated Hospital of Anhui Medical University, Hefei, China
| | - Chunguang Han
- Department of General Surgery, The First Affiliated Hospital of Anhui Medical University, Hefei, China
| | - Zhiying Liang
- School of Biomedical Engineering, ShanghaiTech University, Shanghai, China
| | - Gaosong Zhang
- Department of Ultrasound, The First Affiliated Hospital of Anhui Medical University, Hefei, China
| | - Na Chen
- Department of Ultrasound, Nanjing Hospital Affiliated to Nanjing Medical University, Nanjing, China
| | - Weiwei Zheng
- Department of Ultrasound, Xuancheng People's Hospital, Xuancheng, China
| | - Fanlun Meng
- Department of Breast Surgery, Fuyang Cancer Hospital, Fuyang, China
| | - Dongsheng Yu
- School of Biomedical Engineering, ShanghaiTech University, Shanghai, China
| | - Yixuan Chen
- School of Biomedical Engineering, ShanghaiTech University, Shanghai, China
| | - Yiqun Sun
- Department of Radiology, Fudan University Shanghai Cancer Center, Shanghai, China
| | - Hanqi Zhang
- Department of Ultrasound, The First Affiliated Hospital of Anhui Medical University, Hefei, China
| | - Wei Qian
- Department of Radiology, The Second Affiliated Hospital of Zhejiang University, Hangzhou, China
| | - Xia Wang
- Department of Radiology, The First Affiliated Hospital of Anhui Medical University, Hefei, China
| | - Zhuoran Er
- School of Biomedical Engineering, ShanghaiTech University, Shanghai, China
| | - Chenglu Hu
- School of Biomedical Engineering, ShanghaiTech University, Shanghai, China
| | - Hui Zheng
- Department of Ultrasound, The First Affiliated Hospital of Anhui Medical University, Hefei, China
| | - Dinggang Shen
- School of Biomedical Engineering, ShanghaiTech University, Shanghai, China
- State Key Laboratory of Advanced Medical Materials and Devices, ShanghaiTech University, Shanghai, China
- Shanghai United Imaging Intelligence Co. Ltd., Shanghai, China
- Shanghai Clinical Research and Trial Center, Shanghai, China
| |
Collapse
|
24
|
Chen T, Chen J, Liu H, Liu Z, Yu B, Wang Y, Zhao W, Peng Y, Li J, Yang Y, Wan H, Wang X, Zhang Z, Zhao D, Chen L, Chen L, Liao R, Liu S, Zeng G, Wen Z, Wang Y, Li X, Wang S, Miao H, Chen W, Zhu Y, Wang X, Ding C, Wang T, Li S, Zhang Y. Integration of longitudinal load-bearing tissue MRI radiomics and neural network to predict knee osteoarthritis incidence. J Orthop Translat 2025; 51:187-197. [PMID: 40144553 PMCID: PMC11937290 DOI: 10.1016/j.jot.2025.01.007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/16/2024] [Revised: 11/20/2024] [Accepted: 01/08/2025] [Indexed: 03/28/2025] Open
Abstract
Background Load-bearing structural degradation is crucial in knee osteoarthritis (KOA) progression, yet limited prediction models use load-bearing tissue radiomics for radiographic (structural) KOA incident. Purpose We aim to develop and test a Load-Bearing Tissue plus Clinical variable Radiomic Model (LBTC-RM) to predict radiographic KOA incidents. Study design Risk prediction study. Methods The 700 knees without radiographic KOA at baseline were included from Osteoarthritis Initiative cohort. We selected 2164 knee MRIs during 4-year follow-up. LBTC-RM, which integrated MRI features of meniscus, femur, tibia, femorotibial cartilage, and clinical variables, was developed in total development cohort (n = 1082, 542 cases vs. 540 controls) using neural network algorithm. Final predictive model was tested in total test cohort (n = 1082, 534 cases vs. 548 controls), which integrated data from five visits: baseline (n = 353, 191 cases vs. 162 controls), 3 years prior KOA (n = 46, 19 cases vs. 27 controls), 2 years prior KOA (n = 143, 77 cases vs. 66 controls), 1 year prior KOA (n = 220, 105 cases vs. 115 controls), and at KOA incident (n = 320, 156 cases vs. 164 controls). Results In total test cohort, LBTC-RM predicted KOA incident with AUC (95 % CI) of 0.85 (0.82-0.87); with LBTC-RM aid, performance of resident physicians for KOA prediction were improved, with specificity, sensitivity, and accuracy increasing from 50 %, 60 %, and 55 %-72 %, 73 %, and 72 %, respectively. The LBTC-RM output indicated an increased KOA risk (OR: 20.6, 95 % CI: 13.8-30.6, p < .001). Radiomic scores of load-bearing tissue raised KOA risk (ORs: 1.02-1.9) from 4-year prior KOA whereas 3-dimensional feature score of medial meniscus decreased the OR (0.99) of KOA incident at KOA confirmed. The 2-dimensional feature score of medial meniscus increased the ORs (1.1-1.2) of KOA symptom score from 2-year prior KOA. Conclusions We provided radiomic features of load-bearing tissue to improved KOA risk level assessment and incident prediction. The model has potential clinical applicability in predicting KOA incidents early, enabling physicians to identify high-risk patients before significant radiographic evidence appears. This can facilitate timely interventions and personalized management strategies, improving patient outcomes. The Translational Potential of this Article This study presents a novel approach integrating longitudinal MRI-based radiomics and clinical variables to predict knee osteoarthritis (KOA) incidence using machine learning. By leveraging deep learning for auto-segmentation and machine learning for predictive modeling, this research provides a more interpretable and clinically applicable method for early KOA detection. The introduction of a Radiomics Score System enhances the potential for radiomics as a virtual image-based biopsy tool, facilitating non-invasive, personalized risk assessment for KOA patients. The findings support the translation of advanced imaging and AI-driven predictive models into clinical practice, aiding early diagnosis, personalized treatment planning, and risk stratification for KOA progression. This model has the potential to be integrated into routine musculoskeletal imaging workflows, optimizing early intervention strategies and resource allocation for high-risk populations. Future validation across diverse cohorts will further enhance its clinical utility and generalizability.
Collapse
Affiliation(s)
- Tianyu Chen
- Hebei Medical University Clinical Medicine Postdoctoral Station (Hebei Medical University Third Hospital), Shijiazhuang, Hebei, 050051, People's Republic of China
- Department of Orthopaedic Surgery, NHC Key Laboratory of Intelligent Orthopaedic Equipment, Hebei Medical University Third Hospital, Shijiazhuang, Hebei, 050051, People's Republic of China
- Department of Orthopaedics, The Third Affiliated Hospital of Southern Medical University, Guangzhou, Guangdong, 510630, People's Republic of China
| | - Jian Chen
- Department of Orthopaedics, Chongqing University Three Gorges Hospital, Wanzhou, Chongqing, 404031, People's Republic of China
| | - Hao Liu
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Zhengrui Liu
- Department of Dermatology, Southern Medical University Affiliated Guangdong Provincial No. 2 People's Hospital, Guangzhou, Guangdong, 510310, People's Republic of China
| | - Bin Yu
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Yang Wang
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Wenbo Zhao
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Yinxiao Peng
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Jun Li
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Yun Yang
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Hang Wan
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Xing Wang
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Zhong Zhang
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Deng Zhao
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Lan Chen
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Lili Chen
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Ruyu Liao
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Shanhong Liu
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Guowei Zeng
- Department of Orthopaedics, Huizhou First Hospital, Guangdong Medical University, Huizhou, Guangdong, 516001, People's Republic of China
| | - Zhijia Wen
- Department of Orthopaedics, Huizhou First Hospital, Guangdong Medical University, Huizhou, Guangdong, 516001, People's Republic of China
| | - Yin Wang
- Department of Orthopaedics, Huizhou First Hospital, Guangdong Medical University, Huizhou, Guangdong, 516001, People's Republic of China
| | - Xu Li
- Department of Orthopaedics, Huizhou First Hospital, Guangdong Medical University, Huizhou, Guangdong, 516001, People's Republic of China
| | - Shengjie Wang
- Department of Orthopaedics, Huizhou First Hospital, Guangdong Medical University, Huizhou, Guangdong, 516001, People's Republic of China
| | - Haixiong Miao
- Department of Orthopaedics, Guangzhou Red Cross Hospital, Jinan University, Guangzhou, Guangdong, 510240, People's Republic of China
| | - Wei Chen
- Department of Orthopaedic Surgery, NHC Key Laboratory of Intelligent Orthopaedic Equipment, Hebei Medical University Third Hospital, Shijiazhuang, Hebei, 050051, People's Republic of China
| | - Yanbin Zhu
- Department of Orthopaedic Surgery, NHC Key Laboratory of Intelligent Orthopaedic Equipment, Hebei Medical University Third Hospital, Shijiazhuang, Hebei, 050051, People's Republic of China
| | - Xiaogang Wang
- Department of Orthopaedics, The Third Affiliated Hospital of Southern Medical University, Guangzhou, Guangdong, 510630, People's Republic of China
| | - Changhai Ding
- Clinical Research Centre, Zhujiang Hospital, Southern Medical University, Guangzhou, Guangdong, 510260, People's Republic of China
| | - Ting Wang
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
- Medical Research Center, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Shengfa Li
- Department of Orthopaedics, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, The Second Affiliated Chengdu Hospital of Chongqing Medical University, Chengdu, Sichuan, 610014, People's Republic of China
| | - Yingze Zhang
- Department of Orthopaedic Surgery, NHC Key Laboratory of Intelligent Orthopaedic Equipment, Hebei Medical University Third Hospital, Shijiazhuang, Hebei, 050051, People's Republic of China
| |
Collapse
|
25
|
Sahshong P, Chandra A, Mercado-Shekhar KP, Bhatt M. Deep denoising approach to improve shear wave phase velocity map reconstruction in ultrasound elastography. Med Phys 2025; 52:1481-1499. [PMID: 39714072 DOI: 10.1002/mp.17581] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2024] [Revised: 11/25/2024] [Accepted: 11/25/2024] [Indexed: 12/24/2024] Open
Abstract
BACKGROUND Measurement noise often leads to inaccurate shear wave phase velocity estimation in ultrasound shear wave elastography. Filtering techniques are commonly used for denoising the shear wavefields. However, these filters are often not sufficient, especially in fatty tissues where the signal-to-noise ratio (SNR) can be very low. PURPOSE The purpose of this study is to develop a deep learning approach for denoising shear wavefields in ultrasound shear wave elastography. This may lead to improved reconstruction of shear wave phase velocity image maps. METHODS The study addresses noise by transforming particle velocity data into a time-frequency representation. A neural network with encoder and decoder convolutional blocks effectively decomposes the input and extracts the signal of interest, improving the SNR in high-noise scenarios. The network is trained on simulated phantoms with elasticity values ranging from 3 to 60 kPa. A total of 1 85 570 samples with 80%-20 % $\%$ split were used for training and validation. The approach is tested on experimental phantom and ex-vivo goat liver tissue data. Performance was compared with the traditional filtering methods such as bandpass, median, and wavelet filtering. Kruskal-Wallis one-way analysis of variance was performed to check statistical significance. Multiple comparisons were performed using the Mann-Whitney U test and Holm-Bonferroni adjustment ofp - values $p-{\rm values}$ . RESULTS The results are evaluated using SNR and the percentage of pixels that can be reconstructed in the phase velocity maps. The SNR levels in experimental data improved from -2 to 9.9 dB levels to 15.6 to 30.3 dB levels. Kruskal-Wallis one-way analysis showed statistical significance (p < 0.05 $p<0.05$ ). Multiple comparisons with p-value corrections also showed statistically significant improvement when compared to the bandpass and wavelet filtering scheme (p < 0.05 $p<0.05$ ). Smoother phase velocity maps were reconstructed after denoising. The coefficient of variation is less than5 % $5\%$ in CIRS phantom and less than18 % $18\%$ in ex-vivo goat liver tissue. CONCLUSIONS The proposed approach demonstrates improvement in shear wave phase velocity image map reconstruction and holds promise that deep learning methods can be effectively utilized to extract true shear wave signal from measured noisy data.
Collapse
Affiliation(s)
- Phidakordor Sahshong
- Department of Electronics and Electrical Engineering, Indian Institute of Technology Guwahati, Assam, India
| | - Akash Chandra
- Department Of Biological Sciences And Engineering, Indian Institute of Technology, Gandhinagar, Gujarat, India
| | - Karla P Mercado-Shekhar
- Department Of Biological Sciences And Engineering, Indian Institute of Technology, Gandhinagar, Gujarat, India
| | - Manish Bhatt
- Department of Electronics and Electrical Engineering, Indian Institute of Technology Guwahati, Assam, India
| |
Collapse
|
26
|
Levent AE, Tanaka M, Kumawat C, Heng C, Nikolaos S, Latka K, Miyamoto A, Komatsubara T, Arataki S, Oda Y, Shinohara K, Uotani K. Review Article: Diagnostic Paradigm Shift in Spine Surgery. Diagnostics (Basel) 2025; 15:594. [PMID: 40075840 PMCID: PMC11899683 DOI: 10.3390/diagnostics15050594] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/23/2025] [Revised: 02/20/2025] [Accepted: 02/24/2025] [Indexed: 03/14/2025] Open
Abstract
Meticulous clinical examination is essential for spinal disorders to utilize the diagnostic methods and technologies that strongly support physicians and enhance clinical practice. A significant change in the approach to diagnosing spinal disorders has occurred in the last three decades, which has enhanced a more nuanced understanding of spine pathology. Traditional radiographic methods such as conventional and functional X-rays and CT scans are still the first line in the diagnosis of spinal disorders due to their low cost and accessibility. As more advanced imaging technologies become increasingly available worldwide, there is a constantly increasing trend in MRI scans for detecting spinal pathologies and making treatment decisions. Not only do MRI scans have superior diagnostic capabilities, but they also assist surgeons in performing meticulous preoperative planning, making them currently the most widely used diagnostic tool for spinal disorders. Positron Emission Tomography (PET) can help detect inflammatory lesions, infections, and tumors. Other advanced diagnostic tools such as CT/MRI fusion image, Functional Magnetic Resonance Imaging (fMRI), Upright and Kinetic MRI, magnetic resonance spectroscopy (MRS), diffusion-weighted imaging (DWI), and diffusion tensor imaging (DTI) could play an important role when it comes to detecting more special pathologies. However, some technical difficulties in the daily praxis and their high costs act as obstacles to their further spread. Integrating artificial intelligence and advancements in data analytics and virtual reality promises to enhance spinal procedures' precision, safety, and efficacy. As these technologies continue to develop, they will play a critical role in transforming spinal surgery. This paradigm shift emphasizes the importance of continuous innovation and adaptability in improving the diagnosis and treatment of spinal disorders.
Collapse
Affiliation(s)
- Aras Efe Levent
- Department of Orthopedic Surgery, Okayama Rosai Hospital, 1-10-25 Chikkomidorimachi, Minami Ward Okayama, Okayama 702-8055, Japan; (A.E.L.); (C.K.); (C.H.); (S.N.); (K.L.); (A.M.); (T.K.); (S.A.)
| | - Masato Tanaka
- Department of Orthopedic Surgery, Okayama Rosai Hospital, 1-10-25 Chikkomidorimachi, Minami Ward Okayama, Okayama 702-8055, Japan; (A.E.L.); (C.K.); (C.H.); (S.N.); (K.L.); (A.M.); (T.K.); (S.A.)
| | - Chetan Kumawat
- Department of Orthopedic Surgery, Okayama Rosai Hospital, 1-10-25 Chikkomidorimachi, Minami Ward Okayama, Okayama 702-8055, Japan; (A.E.L.); (C.K.); (C.H.); (S.N.); (K.L.); (A.M.); (T.K.); (S.A.)
- Department of Orthopedic Surgery, Sir Ganga Ram Hospital, Rajinder Nagar, New Delhi 110060, India
| | - Christian Heng
- Department of Orthopedic Surgery, Okayama Rosai Hospital, 1-10-25 Chikkomidorimachi, Minami Ward Okayama, Okayama 702-8055, Japan; (A.E.L.); (C.K.); (C.H.); (S.N.); (K.L.); (A.M.); (T.K.); (S.A.)
| | - Salamalikis Nikolaos
- Department of Orthopedic Surgery, Okayama Rosai Hospital, 1-10-25 Chikkomidorimachi, Minami Ward Okayama, Okayama 702-8055, Japan; (A.E.L.); (C.K.); (C.H.); (S.N.); (K.L.); (A.M.); (T.K.); (S.A.)
| | - Kajetan Latka
- Department of Orthopedic Surgery, Okayama Rosai Hospital, 1-10-25 Chikkomidorimachi, Minami Ward Okayama, Okayama 702-8055, Japan; (A.E.L.); (C.K.); (C.H.); (S.N.); (K.L.); (A.M.); (T.K.); (S.A.)
| | - Akiyoshi Miyamoto
- Department of Orthopedic Surgery, Okayama Rosai Hospital, 1-10-25 Chikkomidorimachi, Minami Ward Okayama, Okayama 702-8055, Japan; (A.E.L.); (C.K.); (C.H.); (S.N.); (K.L.); (A.M.); (T.K.); (S.A.)
| | - Tadashi Komatsubara
- Department of Orthopedic Surgery, Okayama Rosai Hospital, 1-10-25 Chikkomidorimachi, Minami Ward Okayama, Okayama 702-8055, Japan; (A.E.L.); (C.K.); (C.H.); (S.N.); (K.L.); (A.M.); (T.K.); (S.A.)
| | - Shinya Arataki
- Department of Orthopedic Surgery, Okayama Rosai Hospital, 1-10-25 Chikkomidorimachi, Minami Ward Okayama, Okayama 702-8055, Japan; (A.E.L.); (C.K.); (C.H.); (S.N.); (K.L.); (A.M.); (T.K.); (S.A.)
| | - Yoshiaki Oda
- Department of Orthopedic Surgery, Okayama University Hospital, Okayama 7000-8558, Japan; (Y.O.); (K.S.); (K.U.)
| | - Kensuke Shinohara
- Department of Orthopedic Surgery, Okayama University Hospital, Okayama 7000-8558, Japan; (Y.O.); (K.S.); (K.U.)
| | - Koji Uotani
- Department of Orthopedic Surgery, Okayama University Hospital, Okayama 7000-8558, Japan; (Y.O.); (K.S.); (K.U.)
| |
Collapse
|
27
|
P BV, Pg OP, Karrothu A, Gv S. A novel skin cancer detection architecture using tangent rat swarm optimization algorithm enabled DenseNet. NETWORK (BRISTOL, ENGLAND) 2025:1-31. [PMID: 40019033 DOI: 10.1080/0954898x.2025.2452274] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/05/2024] [Revised: 12/30/2024] [Accepted: 01/03/2025] [Indexed: 03/01/2025]
Abstract
This paper proposes a Tangent Rat Swarm Optimization (TRSO)-DenseNet for the detection of skin cancer to reduce the severity rate of cancer. Initially, the input image is pre-processed by employing a linear smoothing filter. The pre-processed image is transferred to skin lesion segmentation, where Mask-RCNN is utilized for segmenting the skin lesion. Then, image augmentation is performed using techniques such as vertical shifting, horizontal shifting, random rotation, brightness adjustment, blurring, and cropping. The augmented image is then fed into the feature extraction phase to identify statistical features, Haralick texture features, Convolutional Neural Network (CNN) features, Local Ternary Pattern (LTP), Histogram of Oriented Gradients (HOG), and Local Vector Pattern (LVP). Finally, the extracted features are fed into the skin cancer detection phase, where DenseNet is used to detect skin cancer. Here, DenseNet is structurally optimized by TRSO, which has the combination of the Tangent Search Algorithm (TSA) and Rat Swarm Optimizer (RSO). The TRSO-DenseNet model is implemented using MATLAB tool and analayzsed using the Society for Imaging Informatics in Medicine-International Skin Imaging Collaboration's (SIIM-ISIC) Melanoma Classification dataset. The Proposed model for skin cancer detection attained superior performance with an accuracy of 94.63%, TPR of 91.51%, and TNR of 92.46%.
Collapse
Affiliation(s)
| | - Om Prakash Pg
- Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Vaddeswaram
| | - Aravind Karrothu
- Department of Information Technology, GMR Institute of Technology, Rajam, India
| | | |
Collapse
|
28
|
Wu T, Long Q, Zeng L, Zhu J, Gao H, Deng Y, Han Y, Qu L, Yi W. Axillary lymph node metastasis in breast cancer: from historical axillary surgery to updated advances in the preoperative diagnosis and axillary management. BMC Surg 2025; 25:81. [PMID: 40016717 PMCID: PMC11869450 DOI: 10.1186/s12893-025-02802-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2024] [Accepted: 02/07/2025] [Indexed: 03/01/2025] Open
Abstract
Axillary lymph node status, which was routinely assessed by axillary lymph node dissection (ALND) until the 1990s, is a crucial factor in determining the stage, prognosis, and therapeutic strategy used for breast cancer patients. Axillary surgery for breast cancer patients has evolved from ALND to minimally invasive approaches. Over the decades, the application of noninvasive imaging techniques, machine learning approaches and emerging clinical prediction models for the detection of axillary lymph node metastasis greatly improves clinical diagnostic efficacy and provides optimal surgical selection. In this work, we summarize the historical axillary surgery and updated perspectives of axillary management for breast cancer patients.
Collapse
Affiliation(s)
- Tong Wu
- Department of General Surgery, The Second Xiangya Hospital, Central South University, Changsha, Hunan, 410011, China
- Clinical Research Center for Breast Disease in Hunan Province, Changsha, 410011, China
| | - Qian Long
- Department of General Surgery, The Second Xiangya Hospital, Central South University, Changsha, Hunan, 410011, China
- Clinical Research Center for Breast Disease in Hunan Province, Changsha, 410011, China
| | - Liyun Zeng
- Department of General Surgery, The Second Xiangya Hospital, Central South University, Changsha, Hunan, 410011, China
- Clinical Research Center for Breast Disease in Hunan Province, Changsha, 410011, China
| | - Jinfeng Zhu
- Department of General Surgery, The Second Xiangya Hospital, Central South University, Changsha, Hunan, 410011, China
- Clinical Research Center for Breast Disease in Hunan Province, Changsha, 410011, China
| | - Hongyu Gao
- Department of General Surgery, The Second Xiangya Hospital, Central South University, Changsha, Hunan, 410011, China
- Clinical Research Center for Breast Disease in Hunan Province, Changsha, 410011, China
| | - Yueqiong Deng
- Department of General Surgery, The Second Xiangya Hospital, Central South University, Changsha, Hunan, 410011, China
- Clinical Research Center for Breast Disease in Hunan Province, Changsha, 410011, China
| | - Yi Han
- Department of General Surgery, The Second Xiangya Hospital, Central South University, Changsha, Hunan, 410011, China
- Clinical Research Center for Breast Disease in Hunan Province, Changsha, 410011, China
| | - Limeng Qu
- Department of General Surgery, The Second Xiangya Hospital, Central South University, Changsha, Hunan, 410011, China.
- Clinical Research Center for Breast Disease in Hunan Province, Changsha, 410011, China.
| | - Wenjun Yi
- Department of General Surgery, The Second Xiangya Hospital, Central South University, Changsha, Hunan, 410011, China.
- Clinical Research Center for Breast Disease in Hunan Province, Changsha, 410011, China.
| |
Collapse
|
29
|
Cote MP, Gholipour A. Imaging Results in Data Usefully Analyzed by Artificial Intelligence Machine Learning. Arthroscopy 2025:S0749-8063(25)00145-8. [PMID: 40021066 DOI: 10.1016/j.arthro.2025.02.024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/17/2025] [Accepted: 02/19/2025] [Indexed: 03/03/2025]
Abstract
Many artificial intelligence (AI) machine learning (ML) papers focused on clinical outcomes use registry data inadequate for predictive modeling. In contrast, diagnostic imaging is an area where available information (pixels, etc.) can result in a reliable, clinically relevant, and accurate model. The use of deep learning for image analysis can reduce interobserver variability, and highlight subtle and meaningful features. AI augments, rather than replaces, clinical expertise, allowing faster, more consistent, and potentially more accurate diagnostic information. This is especially relevant when imaging data is abundant, as continuous model training can further refine diagnostic precision. An effective 3-step approach includes: 1) an efficient "detector" to determine where to look; 2) computational ability to focus on key features of the image and "blur out" background noise ("attention module"); and 3) interpreted key features ("explainability"). Next, the larger process of developing and employing a predictive model needs to be externally validated, to determine the extent to which these results will generalize outside of a single institution. Outside this setting, i.e., external validity, needs to be determined.
Collapse
Affiliation(s)
- Mark P Cote
- Department of Orthopaedic Surgery, Massachusetts General Hospital, Harvard Medical School, Boston, Massachusetts, U.S.A
| | - Alireza Gholipour
- Department of Orthopaedic Surgery, Massachusetts General Hospital, Harvard Medical School, Boston, Massachusetts, U.S.A
| |
Collapse
|
30
|
Ghasemi N, Rokhshad R, Zare Q, Shobeiri P, Schwendicke F. Artificial intelligence for osteoporosis detection on panoramic radiography: A systematic review and meta analysis. J Dent 2025; 156:105650. [PMID: 40010536 DOI: 10.1016/j.jdent.2025.105650] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2024] [Revised: 02/13/2025] [Accepted: 02/23/2025] [Indexed: 02/28/2025] Open
Abstract
INTRODUCTION Osteoporosis is a disease characterized by low bone mineral density and an increased risk of fractures. In dentistry, mandibular bone morphology, assessed for example on panoramic images, has been employed to detect osteoporosis. Artificial intelligence (AI) can aid in diagnosing bone diseases from radiographs. We aimed to systematically review, synthesize and appraise the available evidence supporting AI in detecting osteoporosis on panoramic radiographs. DATA Studies that used AI to detect osteoporosis on dental panoramic images were included. SOURCES On April 8, 2023, a first comprehensive search of electronic databases was conducted, including PubMed, Scopus, Embase, IEEE, arXiv, and Google Scholar (grey literature). This search was subsequently updated on October 6, 2024. STUDY SELECTION The Quality Assessment and Diagnostic Accuracy Tool-2 was employed to determine the risk of bias in the studies. Quantitative analyses involved meta-analyses of diagnostic accuracy measures, including sensitivity and specificity, yielding Diagnostic Odds Ratios (DOR) and synthesized positive likelihood ratios (LR+). The certainty of evidence was assessed using the Grading of Recommendations Assessment, Development, and Evaluation system. RESULTS A total of 24 studies were included. Accuracy ranged from 50% to 99%, sensitivity from 50% to 100%, and specificity from 38% to 100%. A minority of studies (n=10) had a low risk of bias in all domains, while the majority (n=18) showed low risk of applicability concerns. Pooled sensitivity was 87.92% and specificity 81.93%. DOR was 32.99, and L+ 4.87. Meta-regression analysis indicated that sample size had only a marginal impact on heterogeneity (R² = 0.078, p = 0.052), suggesting other study-level factors may contribute to variability. Egger's test suggested potential small-study effects (p < 0.001), indicating a risk of publication bias. CONCLUSION AI, particularly deep learning, showed high diagnostic accuracy in detecting osteoporosis on panoramic radiographs. The results indicate a strong potential for AI to enhance osteoporosis screening in dental settings. However, significant heterogeneity across studies and potential small-study effects highlight the need for further validation, standardization, and larger, well-powered studies to improve model generalizability. CLINICAL SIGNIFICANCE The application of AI in analyzing panoramic radiographs could transform osteoporosis screening in routine dental practice by providing early and accurate diagnosis. This has the potential to integrate osteoporosis detection seamlessly into dental workflows, improving patient outcomes and enabling timely referrals for medical intervention. Addressing issues of model validation and comparability is critical to translating these findings into widespread clinical use.
Collapse
Affiliation(s)
- Nikoo Ghasemi
- Department of Orthodontics and Dentofacial Orthopedics, School of Dentistry, Zanjan University of Medical Sciences, Zanjan, Iran
| | - Rata Rokhshad
- Topic Group Dental Diagnostics and Digital Dentistry, WHO Focus Group AI on Health, Berlin, Germany.
| | - Qonche Zare
- Department of oral and maxillofacial radiology, School of Dentistry, Hormozgan University of Medical Sciences, Bandar Abbas, Iran
| | - Parnian Shobeiri
- Department of Radiology, Memorial Sloan Kettering Cancer Center, New York, 10065, NY, United States
| | - Falk Schwendicke
- Clinic for Conservative Dentistry and Periodontology, LMU Klinikum, Munich, Germany
| |
Collapse
|
31
|
Bu S, Li Y, Liu G, Li Y. MAET-SAM: Magneto-Acousto-Electrical Tomography segmentation network based on the segment anything model. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2025; 22:585-603. [PMID: 40083283 DOI: 10.3934/mbe.2025022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/16/2025]
Abstract
Magneto-Acousto-Electrical Tomography (MAET) is a hybrid imaging method that combines advantages of ultrasound imaging and electrical impedance tomography to image the electrical conductivity of biological tissues. In practical applications, different tissue or disease organization display various conductivity traits. However, the conductivity map consists of overlapping signals measured at multiple locations, the reconstruction results are affected by noise, which results in blurred reconstruction boundaries, low contrast, and irregular artifact distributions. To improve the image resolution and reduce noise of MAET, a dataset of conductivity maps reconstructed from MAET was established, dubbed MAET-IMAGE. Based on this dataset, we proposed a MAET tomography segmentation network based on the Segment Anything Model (SAM), termed as MAET-SAM. Specifically, we froze the encoder weights of SAM to extract rich feature information of image and design, an adaptive decoder with no prompts. In the end, an end-to-end segmentation model for specific MAET images with MAET-IMAGE was proposed. Qualitative and quantitative experiments demonstrated that MAET-SAM outperformed traditional segmentation methods and segmentation models with initial weights in terms of MAET image segmentation performance, bringing new breakthroughs and advancements to the field of medical imaging analysis and clinical diagnosis.
Collapse
Affiliation(s)
- Shuaiyu Bu
- State Grid Beijing Electric Power Company, Beijing 100031, China
- Institute of Electrical Engineering, Chinese Academy of Sciences, Beijing 100190, China
- School of Electronic Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Yuanyuan Li
- Institute of Electrical Engineering, Chinese Academy of Sciences, Beijing 100190, China
- School of Electronic Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Guoqiang Liu
- Institute of Electrical Engineering, Chinese Academy of Sciences, Beijing 100190, China
- School of Electronic Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Yifan Li
- China Railway Communication and Signal Survey & Design Co., Beijing 100036, China
| |
Collapse
|
32
|
Zeng S, Liu Y, Duan X, Zhao X, Sun X, Zhang F. Artificial Intelligence in CT for Predicting Cervical Lymph Node Metastasis in Papillary Thyroid Cancer Patients: A Meta-analysis. Acad Radiol 2025:S1076-6332(25)00108-4. [PMID: 40000328 DOI: 10.1016/j.acra.2025.02.007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/06/2025] [Revised: 02/05/2025] [Accepted: 02/05/2025] [Indexed: 02/27/2025]
Abstract
PURPOSE This meta-analysis aims to evaluate the diagnostic performance of CT-based artificial intelligence (AI) in diagnosing cervical lymph node metastasis (LNM) of papillary thyroid cancer (PTC). METHODS A systematic search was conducted in PubMed, Embase, and Web of Science databases through December 2024, following PRISMA-DTA guidelines. Studies evaluating CT-based AI models for diagnosing cervical LNM in patients with pathologically confirmed PTC were included. The methodological quality was assessed using a modified QUADAS-2 tool. A bivariate random-effects model was used to calculate pooled sensitivity, specificity, and area under the curve (AUC). Heterogeneity was evaluated using I2 statistics, and meta-regression analyses were performed to explore potential sources of heterogeneity. RESULTS 17 studies comprising 1778 patients in internal validation sets and 4072 patients in external validation sets were included. In internal validation sets, AI demonstrated a sensitivity of 0.80 (95% CI: 0.71-0.86), specificity of 0.79 (95% CI: 0.73-0.84), and AUC of 0.86 (95% CI: 0.83-0.89). Radiologists suggested comparable performance with sensitivity of 0.77 (95% CI: 0.64-0.87), specificity of 0.79 (95% CI: 0.72-0.85), and AUC of 0.85 (95% CI: 0.81-0.88). Subgroup analyses revealed that deep learning methods outperformed machine learning in sensitivity (0.86 vs 0.72, P<0.05). No significant publication bias was found in internal validation sets for AI diagnosis (P=0.78). CONCLUSION CT-based AI showed comparable diagnostic performance to radiologists for detecting cervical LNM in PTC patients, with deep learning models showing superior sensitivity. AI could potentially serve as a valuable diagnostic support tool, though further prospective validation is warranted. Limitations include high heterogeneity among studies and insufficient external validation in diverse populations.
Collapse
Affiliation(s)
- Sixun Zeng
- School of Basic Medical Sciences, Chengdu University of Traditional Chinese Medicine, Chengdu, Sichuan, China (S.Z., Y.L., X.D., X.Z., F.Z.)
| | - Yingxian Liu
- School of Basic Medical Sciences, Chengdu University of Traditional Chinese Medicine, Chengdu, Sichuan, China (S.Z., Y.L., X.D., X.Z., F.Z.)
| | - Xinyi Duan
- School of Basic Medical Sciences, Chengdu University of Traditional Chinese Medicine, Chengdu, Sichuan, China (S.Z., Y.L., X.D., X.Z., F.Z.)
| | - Xin Zhao
- School of Basic Medical Sciences, Chengdu University of Traditional Chinese Medicine, Chengdu, Sichuan, China (S.Z., Y.L., X.D., X.Z., F.Z.)
| | - Xiangjuan Sun
- Hospital of Chengdu University of Traditional Chinese Medicine, Chengdu, Sichuan, China (X.S.)
| | - Fenghua Zhang
- School of Basic Medical Sciences, Chengdu University of Traditional Chinese Medicine, Chengdu, Sichuan, China (S.Z., Y.L., X.D., X.Z., F.Z.).
| |
Collapse
|
33
|
Yan Q, Zhang Y, Wei L, Liu X, Wang X. Assessment of anemia recovery using peripheral blood smears by deep semi-supervised learning. Ann Hematol 2025:10.1007/s00277-025-06254-9. [PMID: 39982510 DOI: 10.1007/s00277-025-06254-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2024] [Accepted: 02/10/2025] [Indexed: 02/22/2025]
Abstract
Monitoring anemia recovery is crucial for clinical intervention. Morphological assessment of red blood cells (RBCs) with peripheral blood smears (PBSs) provides additional information beyond routine blood tests. However, the PBS test is labor-intensive, reliant on manual analysis, and susceptible to variability in expert interpretations. Here we introduce a deep semi-supervised learning method, RBCMatch, to classify RBCs during anemia recovery. Using an acute hemolytic anemic mouse model, PBS images at four different time points during anemia recovery were acquired and segmented into 10,091 single RBC images, with only 5% annotated and used in model training. By employing the semi-supervised strategy Fixmatch, RBCMatch achieved an impressive average classification accuracy of 91.2% on the validation dataset and 87.5% on a held-out dataset, demonstrating its superior accuracy and robustness compared to supervised learning methods, especially when labeled samples are scarce. To characterize the anemia recovery process, principal components (PCs) of RBC embeddings were extracted and visualized. Our results indicated that RBC embeddings quantified the state of anemia recovery, and the second PC had a strong correlation with RBC count and hemoglobin concentration, demonstrating the model's ability to accurately depict RBC morphological changes during anemia recovery. Thus, this study provides a valuable tool for the automatic classification of RBCs and offers novel insights into the assessment of anemia recovery, with the potential to aid in clinical decision-making and prognosis analysis in the future.
Collapse
Affiliation(s)
- Qianming Yan
- Ministry of Education Key Laboratory of Bioinformatics, Center for Synthetic and Systems Biology, Beijing National Research Center for Information Science and Technology, Department of Automation, Tsinghua University, Beijing, China
| | - Yingying Zhang
- State Key Laboratory of Complex, Severe, and Rare Diseases, Haihe Laboratory of Cell Ecosystem, Department of Pathophysiology, Institute of Basic Medical Sciences, Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
- Department of Pathology, West China Hospital, Sichuan University, Chengdu, China
| | - Lei Wei
- Ministry of Education Key Laboratory of Bioinformatics, Center for Synthetic and Systems Biology, Beijing National Research Center for Information Science and Technology, Department of Automation, Tsinghua University, Beijing, China
| | - Xuehui Liu
- State Key Laboratory of Complex, Severe, and Rare Diseases, Haihe Laboratory of Cell Ecosystem, Department of Pathophysiology, Institute of Basic Medical Sciences, Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China.
| | - Xiaowo Wang
- Ministry of Education Key Laboratory of Bioinformatics, Center for Synthetic and Systems Biology, Beijing National Research Center for Information Science and Technology, Department of Automation, Tsinghua University, Beijing, China.
| |
Collapse
|
34
|
Anwar A, Rana S, Pathak P. Artificial intelligence in the management of metabolic disorders: a comprehensive review. J Endocrinol Invest 2025:10.1007/s40618-025-02548-x. [PMID: 39969797 DOI: 10.1007/s40618-025-02548-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/04/2024] [Accepted: 02/08/2025] [Indexed: 02/20/2025]
Abstract
This review explores the significant role of artificial intelligence (AI) in managing metabolic disorders like diabetes, obesity, metabolic dysfunction-associated fatty liver disease (MAFLD), and thyroid dysfunction. AI applications in this context encompass early diagnosis, personalized treatment plans, risk assessment, prevention, and biomarker discovery for early and accurate disease management. This review also delves into techniques involving machine learning (ML), deep learning (DL), natural language processing (NLP), computer vision, and reinforcement learning associated with AI and their application in metabolic disorders. The following study also enlightens the challenges and ethical considerations associated with AI implementation, such as data privacy, model interpretability, and bias mitigation. We have reviewed various AI-based tools utilized for the diagnosis and management of metabolic disorders, such as Idx, Guardian Connect system, and DreaMed for diabetes. Further, the paper emphasizes the potential of AI to revolutionize the management of metabolic disorders through collaborations among clinicians and AI experts, the integration of AI into clinical practice, and the necessity for long-term validation studies. The references provided in the paper cover a range of studies related to AI, ML, personalized medicine, metabolic disorders, and diagnostic tools in healthcare, including research on disease diagnostics, personalized therapy, chronic disease management, and the application of AI in diabetes care and nutrition.
Collapse
Affiliation(s)
- Aamir Anwar
- Department of Pharmacy, Amity University, Lucknow campus, 226010, Lucknow, Uttar Pradesh, India
| | - Simran Rana
- Department of Pharmacy, Amity University, Lucknow campus, 226010, Lucknow, Uttar Pradesh, India
| | - Priya Pathak
- Department of Pharmacy, Amity University, Lucknow campus, 226010, Lucknow, Uttar Pradesh, India.
| |
Collapse
|
35
|
Park JW, Ryu SM, Kim HS, Lee YK, Yoo JJ. Deep learning based screening model for hip diseases on plain radiographs. PLoS One 2025; 20:e0318022. [PMID: 39946371 PMCID: PMC11825046 DOI: 10.1371/journal.pone.0318022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2024] [Accepted: 01/08/2025] [Indexed: 02/16/2025] Open
Abstract
INTRODUCTION The interpretation of plain hip radiographs can vary widely among physicians. This study aimed to develop and validate a deep learning-based screening model for distinguishing normal hips from severe hip diseases on plain radiographs. METHODS Electronic medical records and plain radiograph from 2004 to 2012 were used to construct two patient groups: the hip disease group (those who underwent total hip arthroplasty) and normal group. A total of 1,726 radiographs (500 normal hip radiographs and 1,226 radiographs with hip diseases, respectively) were included and were allocated for training (320 and 783), validation (80 and 196), and test (100 and 247) groups. Four different models were designed-raw image for both training and test set, preprocessed image for training but raw image for the test set, preprocessed images for both sets, and change of backbone algorithm from DenseNet to EfficientNet. The deep learning models were compared in terms of accuracy, sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), F1-score, and area under the receiver operating characteristic curve (AUROC). RESULTS The mean age of the patients was 54.0 ± 14.8 years in the hip disease group and 49.8 ± 14.9 years in the normal group. The final model showed highest performance in both the internal test set (accuracy 0.96, sensitivity 0.96, specificity 0.97, PPV 0.99, NPV 0.99, F1-score 0.97, and AUROC 0.99) and the external validation set (accuracy 0.94, sensitivity 0.93, specificity 0.96, PPV 0.95, NPV 0.93, F1-score 0.94, and AUROC 0.98). In the gradcam image, while the first model depended on unrelated marks of radiograph, the second and third model mainly focused on the femur shaft and sciatic notch, respectively. CONCLUSION The deep learning-based model showed high accuracy and reliability in screening hip diseases on plain radiographs, potentially aiding physicians in more accurately diagnosing hip conditions.
Collapse
Affiliation(s)
- Jung-Wee Park
- Department of Orthopaedic Surgery, Seoul National University College of Medicine, Seoul, South Korea
- Department of Orthopaedic Surgery, Seoul National University Bundang Hospital, Seongnam, South Korea
| | - Seung Min Ryu
- Department of Orthopaedic Surgery, Seoul Medical Center, Seoul, Republic of Korea
| | - Hong-Seok Kim
- Department of Orthopaedic Surgery, Seoul National University College of Medicine, Seoul, South Korea
- Department of Orthopaedic Surgery, Seoul National University Hospital, Seoul, South Korea
| | - Young-Kyun Lee
- Department of Orthopaedic Surgery, Seoul National University College of Medicine, Seoul, South Korea
- Department of Orthopaedic Surgery, Seoul National University Bundang Hospital, Seongnam, South Korea
| | - Jeong Joon Yoo
- Department of Orthopaedic Surgery, Seoul National University College of Medicine, Seoul, South Korea
- Department of Orthopaedic Surgery, Seoul National University Hospital, Seoul, South Korea
| |
Collapse
|
36
|
Zhao T, Yue Y, Sun H, Li J, Wen Y, Yao Y, Qian W, Guan Y, Qi S. MAEMC-NET: a hybrid self-supervised learning method for predicting the malignancy of solitary pulmonary nodules from CT images. Front Med (Lausanne) 2025; 12:1507258. [PMID: 40012977 PMCID: PMC11861088 DOI: 10.3389/fmed.2025.1507258] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2024] [Accepted: 02/03/2025] [Indexed: 02/28/2025] Open
Abstract
Introduction Pulmonary granulomatous nodules (PGN) often exhibit similar CT morphological features to solid lung adenocarcinomas (SLA), making preoperative differentiation challenging. This study aims to address this diagnostic challenge by developing a novel deep learning model. Methods This study proposes MAEMC-NET, a model integrating generative (Masked AutoEncoder) and contrastive (Momentum Contrast) self-supervised learning to learn CT image representations of intra- and inter-solitary nodules. A generative self-supervised task of reconstructing masked axial CT patches containing lesions was designed to learn intra- and inter-slice image representations. Contrastive momentum is used to link the encoder in axial-CT-patch path with the momentum encoder in coronal-CT-patch path. A total of 494 patients from two centers were included. Results MAEMC-NET achieved an area under curve (95% Confidence Interval) of 0.962 (0.934-0.973). These results not only significantly surpass the joint diagnosis by two experienced chest radiologists (77.3% accuracy) but also outperform the current state-of-the-art methods. The model performs best on medical images with a 50% mask ratio, showing a 1.4% increase in accuracy compared to the optimal 75% mask ratio on natural images. Discussion The proposed MAEMC-NET effectively distinguishes between benign and malignant solitary pulmonary nodules and holds significant potential to assist radiologists in improving the diagnostic accuracy of PGN and SLA.
Collapse
Affiliation(s)
- Tianhu Zhao
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, China
| | - Yong Yue
- Department of Radiology, Shengjing Hospital of China Medical University, Shenyang, China
| | - Hang Sun
- School of Information Science and Engineering, Shenyang Ligong University, Shenyang, China
| | - Jingxu Li
- Department of Radiology, The First Affiliated Hospital of Guangzhou Medical University, Guangzhou, China
| | - Yanhua Wen
- Department of Radiology, The Fifth Affiliated Hospital of Guangzhou Medical University, Guangzhou, China
| | - Yudong Yao
- Department of Electrical and Computer Engineering, Stevens Institute of Technology, Hoboken, NJ, United States
| | - Wei Qian
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Yubao Guan
- Department of Radiology, The Fifth Affiliated Hospital of Guangzhou Medical University, Guangzhou, China
| | - Shouliang Qi
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, China
| |
Collapse
|
37
|
Kumar D, Mehta MA, Kotecha K, Kulkarni A. Computer-aided cholelithiasis diagnosis using explainable convolutional neural network. Sci Rep 2025; 15:4249. [PMID: 39905177 PMCID: PMC11794719 DOI: 10.1038/s41598-025-85798-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2024] [Accepted: 01/06/2025] [Indexed: 02/06/2025] Open
Abstract
Accurate and precise identification of cholelithiasis is essential for saving the lives of millions of people worldwide. Although several computer-aided cholelithiasis diagnosis approaches have been introduced in the literature, their use is limited because Convolutional Neural Network (CNN) models are black box in nature. Therefore, a novel approach for cholelithiasis classification using custom CNN with post-hoc model explanation is proposed. This paper presents multiple contributions. First, a custom CNN architecture is proposed to classify and predict cholelithiasis from ultrasound image. Second, a modified deep convolutional generative adversarial network is proposed to produce synthetic ultrasound images for better model generalization. Third, a hybrid visual explanation method is proposed by combining gradient-weighted class activation with local interpretable model agnostic explanation to generate a visual explanation using a heatmap. Fourth, an exhaustive performance analysis of the proposed approach on ultrasound images collected from three different Indian hospitals is presented to showcase its efficacy for computer-aided cholelithiasis diagnosis. Fifth, a team of radiologists evaluates and validates the prediction and respective visual explanations made using the proposed approach. The results reveal that the proposed cholelithiasis classification approach beats the performance of state-of-the-art pre-trained CNN and Vision Transformer models. The heatmap generated through the proposed hybrid explanation method offers detailed visual explanations to enhance transparency and trustworthiness in the medical domain.
Collapse
Affiliation(s)
- Dheeraj Kumar
- Department of Computer/IT Engineering, Gujarat Technological University, Ahmedabad, India.
- IT Department, Parul Institute of Engineering & Technology, Parul University, Vadodara, India.
| | - Mayuri A Mehta
- Department of Computer Engineering, Sarvajanik College of Engineering and Technology, Surat, India
| | - Ketan Kotecha
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, Symbiosis International University, Pune, India
- People's Friendship University of Russia Named After Patrice Lumumba (RUDN University), Moscow, Russian Federation
| | - Ambarish Kulkarni
- Computer Aided Engineering, School of Engineering, Swinburne University of Technology, Melbourne, Australia
| |
Collapse
|
38
|
Li Z, Lin J, Wang Y, Li J, Cao Y, Liu X, Wan W, Liu Q, Song X. Ultra-sparse reconstruction for photoacoustic tomography: Sinogram domain prior-guided method exploiting enhanced score-based diffusion model. PHOTOACOUSTICS 2025; 41:100670. [PMID: 39687486 PMCID: PMC11648917 DOI: 10.1016/j.pacs.2024.100670] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/03/2024] [Revised: 10/26/2024] [Accepted: 11/18/2024] [Indexed: 12/18/2024]
Abstract
Photoacoustic tomography, a novel non-invasive imaging modality, combines the principles of optical and acoustic imaging for use in biomedical applications. In scenarios where photoacoustic signal acquisition is insufficient due to sparse-view sampling, conventional direct reconstruction methods significantly degrade image resolution and generate numerous artifacts. To mitigate these constraints, a novel sinogram-domain priors guided extremely sparse-view reconstruction method for photoacoustic tomography boosted by enhanced diffusion model is proposed. The model learns prior information from the data distribution of sinograms under full-ring, 512-projections. In iterative reconstruction, the prior information serves as a constraint in least-squares optimization, facilitating convergence towards more plausible solutions. The performance of the method is evaluated using blood vessel simulation, phantoms, and in vivo experimental data. Subsequently, the transformation of the reconstructed sinograms into the image domain is achieved through the delay-and-sum method, enabling a thorough assessment of the proposed method. The results show that the proposed method demonstrates superior performance compared to the U-Net method, yielding images of markedly higher quality. Notably, for in vivo data under 32 projections, the sinogram structural similarity improved by ∼21 % over U-Net, and the image structural similarity increased by ∼51 % and ∼84 % compared to U-Net and delay-and-sum methods, respectively. The reconstruction in the sinogram domain for photoacoustic tomography enhances sparse-view imaging capabilities, potentially expanding the applications of photoacoustic tomography.
Collapse
Affiliation(s)
| | | | - Yiguang Wang
- School of Information Engineering, Nanchang University, Nanchang 330031, China
| | - Jiahong Li
- School of Information Engineering, Nanchang University, Nanchang 330031, China
| | - Yubin Cao
- School of Information Engineering, Nanchang University, Nanchang 330031, China
| | - Xuan Liu
- School of Information Engineering, Nanchang University, Nanchang 330031, China
| | - Wenbo Wan
- School of Information Engineering, Nanchang University, Nanchang 330031, China
| | - Qiegen Liu
- School of Information Engineering, Nanchang University, Nanchang 330031, China
| | - Xianlin Song
- School of Information Engineering, Nanchang University, Nanchang 330031, China
| |
Collapse
|
39
|
Pham TT, Brecheisen J, Wu CC, Nguyen H, Deng Z, Adjeroh D, Doretto G, Choudhary A, Le N. ItpCtrl-AI: End-to-end interpretable and controllable artificial intelligence by modeling radiologists' intentions. Artif Intell Med 2025; 160:103054. [PMID: 39689443 PMCID: PMC11757032 DOI: 10.1016/j.artmed.2024.103054] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2024] [Revised: 10/13/2024] [Accepted: 12/05/2024] [Indexed: 12/19/2024]
Abstract
Using Deep Learning in computer-aided diagnosis systems has been of great interest due to its impressive performance in the general domain and medical domain. However, a notable challenge is the lack of explainability of many advanced models, which poses risks in critical applications such as diagnosing findings in CXR. To address this problem, we propose ItpCtrl-AI, a novel end-to-end interpretable and controllable framework that mirrors the decision-making process of the radiologist. By emulating the eye gaze patterns of radiologists, our framework initially determines the focal areas and assesses the significance of each pixel within those regions. As a result, the model generates an attention heatmap representing radiologists' attention, which is then used to extract attended visual information to diagnose the findings. By allowing the directional input, our framework is controllable by the user. Furthermore, by displaying the eye gaze heatmap which guides the diagnostic conclusion, the underlying rationale behind the model's decision is revealed, thereby making it interpretable. In addition to developing an interpretable and controllable framework, our work includes the creation of a dataset, named Diagnosed-Gaze++, which aligns medical findings with eye gaze data. Our extensive experimentation validates the effectiveness of our approach in generating accurate attention heatmaps and diagnoses. The experimental results show that our model not only accurately identifies medical findings but also precisely produces the eye gaze attention of radiologists. The dataset, models, and source code will be made publicly available upon acceptance.
Collapse
Affiliation(s)
- Trong-Thang Pham
- AICV Lab, Department of EECS, University of Arkansas, AR 72701, USA.
| | - Jacob Brecheisen
- AICV Lab, Department of EECS, University of Arkansas, AR 72701, USA.
| | - Carol C Wu
- MD Anderson Cancer Center, Houston, TX 77079, USA.
| | - Hien Nguyen
- Department of ECE, University of Houston, TX 77204, USA.
| | - Zhigang Deng
- Department of CS, University of Houston, TX 77204, USA.
| | - Donald Adjeroh
- Department of CSEE, West Virginia University, WV 26506, USA.
| | | | - Arabinda Choudhary
- University of Arkansas for Medical Sciences, Little Rock, AR 72705, USA.
| | - Ngan Le
- AICV Lab, Department of EECS, University of Arkansas, AR 72701, USA.
| |
Collapse
|
40
|
Zhang X, Zhao J, Zong D, Ren H, Gao C. Taming vision transformers for clinical laryngoscopy assessment. J Biomed Inform 2025; 162:104766. [PMID: 39827999 DOI: 10.1016/j.jbi.2024.104766] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2024] [Revised: 12/09/2024] [Accepted: 12/26/2024] [Indexed: 01/22/2025]
Abstract
OBJECTIVE Laryngoscopy, essential for diagnosing laryngeal cancer (LCA), faces challenges due to high inter-observer variability and the reliance on endoscopist expertise. Distinguishing precancerous from early-stage cancerous lesions is particularly challenging, even for experienced practitioners, given their similar appearances. This study aims to enhance laryngoscopic image analysis to improve early screening/detection of cancer or precancerous conditions. METHODS We propose MedFormer, a laryngeal cancer classification method based on the Vision Transformer (ViT). To address data scarcity, MedFormer employs a customized transfer learning approach that leverages the representational power of pre-trained transformers. This method enables robust out-of-domain generalization by fine-tuning a minimal set of additional parameters. RESULTS MedFormer exhibits sensitivity-specificity values of 98%-89% for identifying precancerous lesions (leukoplakia) and 89%-97% for detecting cancer, surpassing CNN counterparts significantly. Additionally, when compared to the two selected ViT-based models, MedFormer also demonstrates superior performance. It also outperforms physician visual evaluations (PVE) in certain scenarios and matches PVE performance in all cases. Visualizations using class activation maps (CAM) and deformable patches demonstrate MedFormer's interpretability, aiding clinicians in understanding the model's predictions. CONCLUSION We highlight the potential of visual transformers in clinical laryngoscopic assessments, presenting MedFormer as an effective method for the early detection of laryngeal cancer.
Collapse
Affiliation(s)
- Xinzhu Zhang
- School of Computer Science and Technology, East China Normal University, North Zhongshan Road 3663, Shanghai, 200062, China
| | - Jing Zhao
- School of Computer Science and Technology, East China Normal University, North Zhongshan Road 3663, Shanghai, 200062, China.
| | - Daoming Zong
- School of Computer Science and Technology, East China Normal University, North Zhongshan Road 3663, Shanghai, 200062, China
| | - Henglei Ren
- Eye & ENT Hospital of Fudan University, Fenyang Road 83, Shanghai, 200000, China
| | - Chunli Gao
- Eye & ENT Hospital of Fudan University, Fenyang Road 83, Shanghai, 200000, China.
| |
Collapse
|
41
|
Wang H, Wu H, Wang Z, Yue P, Ni D, Heng PA, Wang Y. A Narrative Review of Image Processing Techniques Related to Prostate Ultrasound. ULTRASOUND IN MEDICINE & BIOLOGY 2025; 51:189-209. [PMID: 39551652 DOI: 10.1016/j.ultrasmedbio.2024.10.005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/08/2024] [Revised: 09/15/2024] [Accepted: 10/06/2024] [Indexed: 11/19/2024]
Abstract
Prostate cancer (PCa) poses a significant threat to men's health, with early diagnosis being crucial for improving prognosis and reducing mortality rates. Transrectal ultrasound (TRUS) plays a vital role in the diagnosis and image-guided intervention of PCa. To facilitate physicians with more accurate and efficient computer-assisted diagnosis and interventions, many image processing algorithms in TRUS have been proposed and achieved state-of-the-art performance in several tasks, including prostate gland segmentation, prostate image registration, PCa classification and detection and interventional needle detection. The rapid development of these algorithms over the past 2 decades necessitates a comprehensive summary. As a consequence, this survey provides a narrative review of this field, outlining the evolution of image processing methods in the context of TRUS image analysis and meanwhile highlighting their relevant contributions. Furthermore, this survey discusses current challenges and suggests future research directions to possibly advance this field further.
Collapse
Affiliation(s)
- Haiqiao Wang
- Medical UltraSound Image Computing (MUSIC) Lab, Smart Medical Imaging, Learning and Engineering (SMILE) Lab, School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen University, Shenzhen, China; Department of Computer Science and Engineering, The Chinese University of Hong Kong, Hong Kong, China
| | - Hong Wu
- Medical UltraSound Image Computing (MUSIC) Lab, Smart Medical Imaging, Learning and Engineering (SMILE) Lab, School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen University, Shenzhen, China
| | - Zhuoyuan Wang
- Medical UltraSound Image Computing (MUSIC) Lab, Smart Medical Imaging, Learning and Engineering (SMILE) Lab, School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen University, Shenzhen, China
| | - Peiyan Yue
- Medical UltraSound Image Computing (MUSIC) Lab, Smart Medical Imaging, Learning and Engineering (SMILE) Lab, School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen University, Shenzhen, China
| | - Dong Ni
- Medical UltraSound Image Computing (MUSIC) Lab, Smart Medical Imaging, Learning and Engineering (SMILE) Lab, School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen University, Shenzhen, China
| | - Pheng-Ann Heng
- Department of Computer Science and Engineering, The Chinese University of Hong Kong, Hong Kong, China
| | - Yi Wang
- Medical UltraSound Image Computing (MUSIC) Lab, Smart Medical Imaging, Learning and Engineering (SMILE) Lab, School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen University, Shenzhen, China.
| |
Collapse
|
42
|
Yin X, Lu Y, Cui Y, Zhou Z, Wen J, Huang Z, Yan Y, Yu J, Meng X. CT-based radiomics-deep learning model predicts occult lymph node metastasis in early-stage lung adenocarcinoma patients: A multicenter study. Chin J Cancer Res 2025; 37:12-27. [PMID: 40078558 PMCID: PMC11893343 DOI: 10.21147/j.issn.1000-9604.2025.01.02] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2024] [Accepted: 12/16/2024] [Indexed: 03/14/2025] Open
Abstract
Objective The neglect of occult lymph nodes metastasis (OLNM) is one of the pivotal causes of early non-small cell lung cancer (NSCLC) recurrence after local treatments such as stereotactic body radiotherapy (SBRT) or surgery. This study aimed to develop and validate a computed tomography (CT)-based radiomics and deep learning (DL) fusion model for predicting non-invasive OLNM. Methods Patients with radiologically node-negative lung adenocarcinoma from two centers were retrospectively analyzed. We developed clinical, radiomics, and radiomics-clinical models using logistic regression. A DL model was established using a three-dimensional squeeze-and-excitation residual network-34 (3D SE-ResNet34) and a fusion model was created by integrating seleted clinical, radiomics features and DL features. Model performance was assessed using the area under the curve (AUC) of the receiver operating characteristic (ROC) curve, calibration curves, and decision curve analysis (DCA). Five predictive models were compared; SHapley Additive exPlanations (SHAP) and Gradient-weighted Class Activation Mapping (Grad-CAM) were employed for visualization and interpretation. Results Overall, 358 patients were included: 186 in the training cohort, 48 in the internal validation cohort, and 124 in the external testing cohort. The DL fusion model incorporating 3D SE-Resnet34 achieved the highest AUC of 0.947 in the training dataset, with strong performance in internal and external cohorts (AUCs of 0.903 and 0.907, respectively), outperforming single-modal DL models, clinical models, radiomics models, and radiomics-clinical combined models (DeLong test: P<0.05). DCA confirmed its clinical utility, and calibration curves demonstrated excellent agreement between predicted and observed OLNM probabilities. Features interpretation highlighted the importance of textural characteristics and the surrounding tumor regions in stratifying OLNM risk. Conclusions The DL fusion model reliably and accurately predicts OLNM in early-stage lung adenocarcinoma, offering a non-invasive tool to refine staging and guide personalized treatment decisions. These results may aid clinicians in optimizing surgical and radiotherapy strategies.
Collapse
Affiliation(s)
- Xiaoyan Yin
- Department of Radiation Oncology, Shandong Cancer Hospital and Institute, Shandong First Medical University and Shandong Academy of Medical Science, Jinan 250117, China
| | - Yao Lu
- School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou 510006, China
| | - Yongbin Cui
- Department of Radiation Oncology, Shandong Cancer Hospital and Institute, Shandong First Medical University and Shandong Academy of Medical Science, Jinan 250117, China
| | - Zichun Zhou
- Department of Radiation Oncology, Shandong Cancer Hospital and Institute, Shandong First Medical University and Shandong Academy of Medical Science, Jinan 250117, China
| | - Junxu Wen
- Department of Radiation Oncology, Shandong Cancer Hospital and Institute, Shandong First Medical University and Shandong Academy of Medical Science, Jinan 250117, China
| | - Zhaoqin Huang
- Department of Radiology, Shandong Provincial Hospital, Shandong First Medical University, Jinan 250021, China
| | - Yuanyuan Yan
- Department of Radiology, Shandong Cancer Hospital and Institute, Shandong First Medical University and Shandong Academy of Medical Science, Jinan 250117, China
| | - Jinming Yu
- Department of Radiation Oncology, Shandong Cancer Hospital and Institute, Shandong First Medical University and Shandong Academy of Medical Science, Jinan 250117, China
| | - Xiangjiao Meng
- Department of Radiation Oncology, Shandong Cancer Hospital and Institute, Shandong First Medical University and Shandong Academy of Medical Science, Jinan 250117, China
| |
Collapse
|
43
|
Nazir A, Hussain A, Singh M, Assad A. A novel approach in cancer diagnosis: integrating holography microscopic medical imaging and deep learning techniques-challenges and future trends. Biomed Phys Eng Express 2025; 11:022002. [PMID: 39671712 DOI: 10.1088/2057-1976/ad9eb7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2024] [Accepted: 12/13/2024] [Indexed: 12/15/2024]
Abstract
Medical imaging is pivotal in early disease diagnosis, providing essential insights that enable timely and accurate detection of health anomalies. Traditional imaging techniques, such as Magnetic Resonance Imaging (MRI), Computer Tomography (CT), ultrasound, and Positron Emission Tomography (PET), offer vital insights into three-dimensional structures but frequently fall short of delivering a comprehensive and detailed anatomical analysis, capturing only amplitude details. Three-dimensional holography microscopic medical imaging provides a promising solution by capturing the amplitude (brightness) and phase (structural information) details of biological structures. In this study, we investigate the novel collaborative potential of Deep Learning (DL) and holography microscopic phase imaging for cancer diagnosis. The study comprehensively examines existing literature, analyzes advancements, identifies research gaps, and proposes future research directions in cancer diagnosis through the integrated Quantitative Phase Imaging (QPI) and DL methodology. This novel approach addresses a critical limitation of traditional imaging by capturing detailed structural information, paving the way for more accurate diagnostics. The proposed approach comprises tissue sample collection, holographic image scanning, preprocessing in case of imbalanced datasets, and training on annotated datasets using DL architectures like U-Net and Vision Transformer(ViT's). Furthermore, sophisticated concepts in DL, like the incorporation of Explainable AI (XAI) techniques, are suggested for comprehensive disease diagnosis and identification. The study thoroughly investigates the advantages of integrating holography imaging and DL for precise cancer diagnosis. Additionally, meticulous insights are presented by identifying the challenges associated with this integration methodology.
Collapse
Affiliation(s)
- Asifa Nazir
- Department of Computer Science and Engineering, Islamic University of Science and Technology, Awantipora, Pulwama, 192122, J&K, India
| | - Ahsan Hussain
- Department of Computer Science and Engineering, Islamic University of Science and Technology, Awantipora, Pulwama, 192122, J&K, India
| | - Mandeep Singh
- Department of Physics, Islamic University of Science and Technology, Awantipora, Kashmir, 192122, J&K, India †
| | - Assif Assad
- Department of Computer Science and Engineering, Islamic University of Science and Technology, Awantipora, Pulwama, 192122, J&K, India
| |
Collapse
|
44
|
Sorondo SM, Fereydooni A, Ho VT, Dossabhoy SS, Lee JT, Stern JR. Significant Radiation Reduction Using Cloud-Based AI Imaging in Manually Matched Cohort of Complex Aneurysm Repair. Ann Vasc Surg 2025; 114:24-29. [PMID: 39884499 DOI: 10.1016/j.avsg.2024.12.081] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/04/2024] [Revised: 12/23/2024] [Accepted: 12/28/2024] [Indexed: 02/01/2025]
Abstract
BACKGROUND Cloud-based, surgical augmented intelligence (Cydar Medical, Cambridge, United Kingdom) can be used for surgical planning and intraoperative imaging guidance during complex endovascular aortic procedures. We aim to evaluate radiation exposure, operative safety metrics, and postoperative renal outcomes following implementation of Cydar imaging guidance using a manually matched cohort of aortic procedures. METHODS We retrospectively reviewed our prospectively maintained database of endovascular aortic cases. Patients repaired using Cydar imaging were matched to patients who underwent a similar procedure without using Cydar. Matching was performed manually on a 1:1 basis using anatomy, device configuration, number of branches/fenestrations, and adjunctive procedures including in-situ laser fenestration. Radiation, contrast use, and other operative metrics were compared. Preoperative and postoperative maximum creatinine was compared to assess for acute kidney injury (AKI) based on risk, injury, failure, loss of kidney function, and end-stage kidney disease (RIFLE) criteria. RESULTS Hundred patients from 2012 to 2023 were identified: 50 cases (38 fenestrated endovascular aortic repairs, 2 thoracic endovascular aortic repairs, 3 octopus-type thoracoabdominal aortic aneurysm repair, 7 endovascular aneurysm repairs) where Cydar imaging was used, with suitable matches to 50 non-Cydar cases. Baseline characteristics including body mass index did not differ significantly between the 2 groups (27.8 ± 5.6 vs. 26.7 ± 6.1; P = 0.31). Radiation dose was significantly lower in the Cydar group (2529 ± 2256 vs. 3676 ± 2976 mGy; P < 0.03), despite there being no difference in fluoroscopy time (51 ± 29.4 vs. 58 ± 37.2 min; P = 0.37). Contrast volume (94 ± 37.4 vs. 93 ± 43.9 mL; P = 0.73), estimated blood loss (169 ± 223 vs. 193 ± 222 mL; P = 0.97), and procedure time (154 ± 78 vs. 165 ± 89.1 min) did not differ significantly. Additionally, Cydar versus non-Cydar patients did not show a significant difference between precreatinine and postcreatinine changes (0.13 ± 0.08 vs. 0.05 ± 0.07; P = 0.34). Only one patient in the non-Cydar group met RIFLE criteria for AKI postoperatively. CONCLUSION The use of cloud-based augmented intelligence imaging was associated with a significant reduction in radiation dose in a cohort of matched aortic procedures but did not appear to affect other parameters or renal function. Even with advanced imaging, surgeons should remain conscientious about radiation safety and administration of nephrotoxic contrast agents.
Collapse
Affiliation(s)
- Sabina M Sorondo
- Division of Vascular and Endovascular Surgery, Stanford University School of Medicine, Stanford, CA
| | - Arash Fereydooni
- Division of Vascular and Endovascular Surgery, Stanford University School of Medicine, Stanford, CA
| | - Vy T Ho
- Division of Vascular and Endovascular Surgery, Stanford University School of Medicine, Stanford, CA
| | - Shernaz S Dossabhoy
- Division of Vascular and Endovascular Surgery, Stanford University School of Medicine, Stanford, CA
| | - Jason T Lee
- Division of Vascular and Endovascular Surgery, Stanford University School of Medicine, Stanford, CA
| | - Jordan R Stern
- Division of Vascular & Endovascular Surgery, Weill Cornell Medicine, New York, NY.
| |
Collapse
|
45
|
Sandri E, Cerdá Olmedo G, Piredda M, Werner LU, Dentamaro V. Explanatory AI Predicts the Diet Adopted Based on Nutritional and Lifestyle Habits in the Spanish Population. Eur J Investig Health Psychol Educ 2025; 15:11. [PMID: 39997075 PMCID: PMC11854735 DOI: 10.3390/ejihpe15020011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2024] [Revised: 01/13/2025] [Accepted: 01/17/2025] [Indexed: 02/26/2025] Open
Abstract
This study used Explainable Artificial Intelligence (XAI) with SHapley Additive exPlanations (SHAP) to examine dietary and lifestyle habits in the Spanish population and identify key diet predictors. A cross-sectional design was used, employing the validated NutSo-HH scale to gather data on nutrition, lifestyle, and socio-demographic factors. The CatBoost method combined with SHAP was applied. The sample included 22,181 Spanish adults: 17,573 followed the Mediterranean diet, 1425 were vegetarians, 365 were vegans, and 1018 practiced intermittent fasting. Fish consumption was the strongest dietary indicator, with vegans abstaining and some vegetarians consuming it occasionally. Age influenced diet: younger individuals preferred vegan/vegetarian diets, while older adults adhered to the Mediterranean diet. Vegans and vegetarians consumed less junk food, and intermittent fasters were more physically active. The model effectively predicts the Mediterranean diet but struggles with others due to sample imbalance, highlighting the need for larger studies on plant-based and intermittent fasting diets.
Collapse
Affiliation(s)
- Elena Sandri
- Faculty of Medicine and Health Sciences, Catholic University of Valencia San Vicente Mártir, c/Quevedo, 2, 46001 Valencia, Spain; (E.S.); (G.C.O.)
- Doctoral School, Catholic University of Valencia San Vicente Mártir, c/Quevedo 2, 46001 Valencia, Spain
| | - Germán Cerdá Olmedo
- Faculty of Medicine and Health Sciences, Catholic University of Valencia San Vicente Mártir, c/Quevedo, 2, 46001 Valencia, Spain; (E.S.); (G.C.O.)
| | - Michela Piredda
- Department of Medicine and Surgery, Research Unit Nursing Science, Università Campus Bio-Medico di Roma, Via Alvaro del Portillo, 21-00128 Rome, Italy
| | - Lisa Ursula Werner
- Faculty of Teaching and Science of Education, Catholic University of Valencia San Vicente Mártir, c/Quevedo, 2, 46001 Valencia, Spain;
| | - Vincenzo Dentamaro
- Department of Computer Science, University of Bari Aldo Moro, Via E. Orabona, 4, 70125 Bari, Italy;
| |
Collapse
|
46
|
Tavanaei R, Akhlaghpasand M, Alikhani A, Hajikarimloo B, Ansari A, Yong RL, Margetis K. Performance of Radiomics-based machine learning and deep learning-based methods in the prediction of tumor grade in meningioma: a systematic review and meta-analysis. Neurosurg Rev 2025; 48:78. [PMID: 39849257 DOI: 10.1007/s10143-025-03236-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2024] [Revised: 01/10/2025] [Accepted: 01/15/2025] [Indexed: 01/25/2025]
Abstract
Currently, the World Health Organization (WHO) grade of meningiomas is determined based on the biopsy results. Therefore, accurate non-invasive preoperative grading could significantly improve treatment planning and patient outcomes. Considering recent advances in machine learning (ML) and deep learning (DL), this meta-analysis aimed to evaluate the performance of these models in predicting the WHO meningioma grade using imaging data. A systematic search was performed in PubMed/MEDLINE, Embase, and the Cochrane Library for studies published up to April 1, 2024, and reporting the performance metrics of the ML models in predicting of WHO meningioma grade using imaging studies. Pooled area under the receiver operating characteristics curve (AUROC), specificity, and sensitivity were estimated. Subgroup and meta-regression analyses were performed based on a number of potential influencing variables. A total of 32 studies with 15,365 patients were included in the present study. The overall pooled sensitivity, specificity, and AUROC of ML methods for prediction of tumor grade in meningioma were 85% (95% CI, 79-89%), 87% (95% CI, 81-91%), and 93% (95% CI, 90-95%), respectively. Both the type of validation and study cohort (training or test) were significantly associated with model performance. However, no significant association was found between the sample size or the type of ML method and model performance. The ML predictive models show a high overall performance in predicting the WHO meningioma grade using imaging data. Further studies on the performance of DL algorithms in larger datasets using external validation are needed.
Collapse
Affiliation(s)
- Roozbeh Tavanaei
- Functional Neurosurgery Research Center, Shohada Tajrish Comprehensive Neurosurgical Center of Excellence, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Mohammadhosein Akhlaghpasand
- Functional Neurosurgery Research Center, Shohada Tajrish Comprehensive Neurosurgical Center of Excellence, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Alireza Alikhani
- Functional Neurosurgery Research Center, Shohada Tajrish Comprehensive Neurosurgical Center of Excellence, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Bardia Hajikarimloo
- Department of Neurological Surgery, University of Virginia, Charlottesville, VA, USA
| | - Ali Ansari
- Student Research Committee, Shiraz University of Medical Sciences, Shiraz, Iran
| | - Raymund L Yong
- Department of Neurosurgery, Mount Sinai Hospital, Icahn School of Medicine, New York City, NY, USA
| | - Konstantinos Margetis
- Department of Neurosurgery, Mount Sinai Hospital, Icahn School of Medicine, New York City, NY, USA.
| |
Collapse
|
47
|
Qiong L, Chaofan L, Jinnan T, Liping C, Jianxiang S. Medical image segmentation based on frequency domain decomposition SVD linear attention. Sci Rep 2025; 15:2833. [PMID: 39843905 PMCID: PMC11754837 DOI: 10.1038/s41598-025-86315-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2024] [Accepted: 01/09/2025] [Indexed: 01/24/2025] Open
Abstract
Convolutional Neural Networks (CNNs) have achieved remarkable segmentation accuracy in medical image segmentation tasks. However, the Vision Transformer (ViT) model, with its capability of extracting global information, offers a significant advantage in contextual information compared to the limited receptive field of convolutional kernels in CNNs. Despite this, ViT models struggle to fully detect and extract high-frequency signals, such as textures and boundaries, in medical images. These high-frequency features are essential in medical imaging, as targets like tumors and pathological organs exhibit significant differences in texture and boundaries across different stages. Additionally, the high resolution of medical images leads to computational complexity in the self-attention mechanism of ViTs. To address these limitations, we propose a medical image segmentation network framework based on frequency domain decomposition using a Laplacian pyramid. This approach selectively computes attention features for high-frequency signals in the original image to enhance spatial structural information effectively. During attention feature computation, we introduce Singular Value Decomposition (SVD) to extract an effective representation matrix from the original image, which is then applied in the attention computation process for linear projection. This method reduces computational complexity while preserving essential features. We demonstrated the segmentation validity and superiority of our model on the Abdominal Multi-Organ Segmentation dataset and the Dermatological Disease dataset, and on the Synapse dataset our model achieved a score of 82.68 on the Dice metrics and 17.23 mm on the HD metrics. Experimental results indicate that our model consistently exhibits segmentation effectiveness and improved accuracy across various datasets.
Collapse
Affiliation(s)
- Liu Qiong
- School of Medical Imaging, Jiangsu Medical College, Yancheng, 224005, Jiangsu, China.
| | - Li Chaofan
- Affiliated Hospital 6 of Nantong University, Yancheng Third People's Hospital, Yancheng, 224001, Jiangsu, China
| | - Teng Jinnan
- Affiliated Hospital 6 of Nantong University, Yancheng Third People's Hospital, Yancheng, 224001, Jiangsu, China
| | - Chen Liping
- Affiliated Hospital 6 of Nantong University, Yancheng Third People's Hospital, Yancheng, 224001, Jiangsu, China
| | - Song Jianxiang
- Affiliated Hospital 6 of Nantong University, Yancheng Third People's Hospital, Yancheng, 224001, Jiangsu, China.
| |
Collapse
|
48
|
Albuquerque C, Henriques R, Castelli M. Deep learning-based object detection algorithms in medical imaging: Systematic review. Heliyon 2025; 11:e41137. [PMID: 39758372 PMCID: PMC11699422 DOI: 10.1016/j.heliyon.2024.e41137] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2024] [Revised: 12/04/2024] [Accepted: 12/10/2024] [Indexed: 01/06/2025] Open
Abstract
Over the past decade, Deep Learning (DL) techniques have demonstrated remarkable advancements across various domains, driving their widespread adoption. Particularly in medical image analysis, DL received greater attention for tasks like image segmentation, object detection, and classification. This paper provides an overview of DL-based object recognition in medical images, exploring recent methods and emphasizing different imaging techniques and anatomical applications. Utilizing a meticulous quantitative and qualitative analysis following PRISMA guidelines, we examined publications based on citation rates to explore into the utilization of DL-based object detectors across imaging modalities and anatomical domains. Our findings reveal a consistent rise in the utilization of DL-based object detection models, indicating unexploited potential in medical image analysis. Predominantly within Medicine and Computer Science domains, research in this area is most active in the US, China, and Japan. Notably, DL-based object detection methods have gotten significant interest across diverse medical imaging modalities and anatomical domains. These methods have been applied to a range of techniques including CR scans, pathology images, and endoscopic imaging, showcasing their adaptability. Moreover, diverse anatomical applications, particularly in digital pathology and microscopy, have been explored. The analysis underscores the presence of varied datasets, often with significant discrepancies in size, with a notable percentage being labeled as private or internal, and with prospective studies in this field remaining scarce. Our review of existing trends in DL-based object detection in medical images offers insights for future research directions. The continuous evolution of DL algorithms highlighted in the literature underscores the dynamic nature of this field, emphasizing the need for ongoing research and fitted optimization for specific applications.
Collapse
|
49
|
Lee S, Youn J, Kim H, Kim M, Yoon SH. CXR-LLaVA: a multimodal large language model for interpreting chest X-ray images. Eur Radiol 2025:10.1007/s00330-024-11339-6. [PMID: 39812665 DOI: 10.1007/s00330-024-11339-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/19/2024] [Revised: 10/25/2024] [Accepted: 12/04/2024] [Indexed: 01/16/2025]
Abstract
OBJECTIVE This study aimed to develop an open-source multimodal large language model (CXR-LLaVA) for interpreting chest X-ray images (CXRs), leveraging recent advances in large language models (LLMs) to potentially replicate the image interpretation skills of human radiologists. MATERIALS AND METHODS For training, we collected 592,580 publicly available CXRs, of which 374,881 had labels for certain radiographic abnormalities (Dataset 1) and 217,699 provided free-text radiology reports (Dataset 2). After pre-training a vision transformer with Dataset 1, we integrated it with an LLM influenced by the LLaVA network. Then, the model was fine-tuned, primarily using Dataset 2. The model's diagnostic performance for major pathological findings was evaluated, along with the acceptability of radiologic reports by human radiologists, to gauge its potential for autonomous reporting. RESULTS The model demonstrated impressive performance in test sets, achieving an average F1 score of 0.81 for six major pathological findings in the MIMIC internal test set and 0.56 for six major pathological findings in the external test set. The model's F1 scores surpassed those of GPT-4-vision and Gemini-Pro-Vision in both test sets. In human radiologist evaluations of the external test set, the model achieved a 72.7% success rate in autonomous reporting, slightly below the 84.0% rate of ground truth reports. CONCLUSION This study highlights the significant potential of multimodal LLMs for CXR interpretation, while also acknowledging the performance limitations. Despite these challenges, we believe that making our model open-source will catalyze further research, expanding its effectiveness and applicability in various clinical contexts. KEY POINTS Question How can a multimodal large language model be adapted to interpret chest X-rays and generate radiologic reports? Findings The developed CXR-LLaVA model effectively detects major pathological findings in chest X-rays and generates radiologic reports with a higher accuracy compared to general-purpose models. Clinical relevance This study demonstrates the potential of multimodal large language models to support radiologists by autonomously generating chest X-ray reports, potentially reducing diagnostic workloads and improving radiologist efficiency.
Collapse
Affiliation(s)
- Seowoo Lee
- Department of Radiology, Seoul National University College of Medicine, Seoul National University Hospital, Seoul, Republic of Korea
| | - Jiwon Youn
- AI Graduate School, Gwangju Institute of Science and Technology, Gwangju, Republic of Korea
| | - Hyungjin Kim
- Department of Radiology, Seoul National University College of Medicine, Seoul National University Hospital, Seoul, Republic of Korea
| | - Mansu Kim
- AI Graduate School, Gwangju Institute of Science and Technology, Gwangju, Republic of Korea.
| | - Soon Ho Yoon
- Department of Radiology, Seoul National University College of Medicine, Seoul National University Hospital, Seoul, Republic of Korea.
| |
Collapse
|
50
|
Wang Q, Huang T, Luo X, Luo X, Li X, Cao K, Li D, Shen L. An Efficient Acute Lymphoblastic Leukemia Screen Framework Based on Multi-Modal Deep Neural Network. Int J Lab Hematol 2025. [PMID: 39810306 DOI: 10.1111/ijlh.14424] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/06/2024] [Revised: 11/18/2024] [Accepted: 12/19/2024] [Indexed: 01/16/2025]
Abstract
BACKGROUND Acute lymphoblastic leukemia (ALL) is a leading cause of death among pediatric malignancies. Early diagnosis of ALL is crucial for minimizing misdiagnosis, improving survival rates, and ensuring the implementation of precise treatment plans for patients. METHODS In this study, we propose a multi-modal deep neural network-based framework for early and efficient screening of ALL. Both white blood cell (WBC) scattergrams and complete blood count (CBC) are employed for ALL detection. The dataset comprises medical data from 233 patients with ALL, 283 patients with infectious mononucleosis (IM), and 183 healthy controls (HCs). RESULTS The combination of CBC data with WBC scattergrams achieved an accuracy of 98.43% in fivefold cross-validation and a sensitivity of 96.67% in external validation, demonstrating the efficacy of our method. Additionally, the area under the curve (AUC) of this model surpasses 0.99, outperforming well-trained medical technicians. CONCLUSIONS To the best of our knowledge, this framework is the first to incorporate WBC scattergrams with CBC data for ALL screening, proving to be an efficient method with enhanced sensitivity and specificity. Integrating this framework into the screening procedure shows promise for improving the early diagnosis of ALL and reducing the burden on medical technicians. The code and dataset are available at https://github.com/cvi-szu/ALL-Screening.
Collapse
Affiliation(s)
- Qiuming Wang
- Computer Vision Institute, College of Computer Science and Software, Shenzhen University, China
| | - Tao Huang
- Department of Laboratory Medicine, Shenzhen Children's Hospital, Shenzhen, China
| | - Xiaojuan Luo
- Department of Laboratory Medicine, Shenzhen Children's Hospital, Shenzhen, China
| | - Xiaoling Luo
- Computer Vision Institute, College of Computer Science and Software, Shenzhen University, China
| | - Xuechen Li
- School of Electronic and Information Engineering, Wuyi University, China
| | - Ke Cao
- Department of Laboratory Medicine, Shenzhen Children's Hospital, Shenzhen, China
| | - Defa Li
- Department of Laboratory Medicine, Shenzhen Children's Hospital, Shenzhen, China
| | - Linlin Shen
- Computer Vision Institute, College of Computer Science and Software, Shenzhen University, China
- Guangdong Provincial Key Laboratory of Intelligent Information Processing, Shenzhen University, Shenzhen, China
| |
Collapse
|