Akbulut S, Colak C. Explainable artificial intelligence and ensemble learning for hepatocellular carcinoma classification: State of the art, performance, and clinical implications. World J Hepatol 2025; 17(11): 109494 [DOI: 10.4254/wjh.v17.i11.109494]
Corresponding Author of This Article
Sami Akbulut, MD, FACS, Professor, Surgery and Liver Transplantation, Inonu University Faculty of Medicine, Elazig Yolu 10 Km, Malatya 44280, Türkiye. akbulutsami@gmail.com
Research Domain of This Article
Transplantation
Article-Type of This Article
Review
Open-Access Policy of This Article
This article is an open-access article which was selected by an in-house editor and fully peer-reviewed by external reviewers. It is distributed in accordance with the Creative Commons Attribution Non Commercial (CC BY-NC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited and the use is non-commercial. See: http://creativecommons.org/licenses/by-nc/4.0/
Nov 27, 2025 (publication date) through Dec 4, 2025
Times Cited of This Article
Times Cited (0)
Journal Information of This Article
Publication Name
World Journal of Hepatology
ISSN
1948-5182
Publisher of This Article
Baishideng Publishing Group Inc, 7041 Koll Center Parkway, Suite 160, Pleasanton, CA 94566, USA
Share the Article
Akbulut S, Colak C. Explainable artificial intelligence and ensemble learning for hepatocellular carcinoma classification: State of the art, performance, and clinical implications. World J Hepatol 2025; 17(11): 109494 [DOI: 10.4254/wjh.v17.i11.109494]
World J Hepatol. Nov 27, 2025; 17(11): 109494 Published online Nov 27, 2025. doi: 10.4254/wjh.v17.i11.109494
Explainable artificial intelligence and ensemble learning for hepatocellular carcinoma classification: State of the art, performance, and clinical implications
Sami Akbulut, Cemil Colak
Sami Akbulut, Surgery and Liver Transplantation, Inonu University Faculty of Medicine, Malatya 44280, Türkiye
Sami Akbulut, Cemil Colak, Biostatistics and Medical Informatics, Inonu University Faculty of Medicine, Malatya 44280, Türkiye
Author contributions: Akbulut S and Colak C conceived the project and designed research, wrote the manuscript and reviewed final version
Conflict-of-interest statement: The authors declare no competing interests.
Open Access: This article is an open-access article that was selected by an in-house editor and fully peer-reviewed by external reviewers. It is distributed in accordance with the Creative Commons Attribution NonCommercial (CC BY-NC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited and the use is non-commercial. See: https://creativecommons.org/Licenses/by-nc/4.0/
Corresponding author: Sami Akbulut, MD, FACS, Professor, Surgery and Liver Transplantation, Inonu University Faculty of Medicine, Elazig Yolu 10 Km, Malatya 44280, Türkiye. akbulutsami@gmail.com
Received: May 13, 2025 Revised: June 13, 2025 Accepted: October 10, 2025 Published online: November 27, 2025 Processing time: 198 Days and 11.2 Hours
Abstract
Hepatocellular carcinoma (HCC) remains a leading cause of cancer-related mortality globally, necessitating advanced diagnostic tools to improve early detection and personalized targeted therapy. This review synthesizes evidence on explainable ensemble learning approaches for HCC classification, emphasizing their integration with clinical workflows and multi-omics data. A systematic analysis [including datasets such as The Cancer Genome Atlas, Gene Expression Omnibus, and the Surveillance, Epidemiology, and End Results (SEER) datasets] revealed that explainable ensemble learning models achieve high diagnostic accuracy by combining clinical features, serum biomarkers such as alpha-fetoprotein, imaging features such as computed tomography and magnetic resonance imaging, and genomic data. For instance, SHapley Additive exPlanations (SHAP)-based random forests trained on NCBI GSE14520 microarray data (n = 445) achieved 96.53% accuracy, while stacking ensembles applied to the SEER program data (n = 1897) demonstrated an area under the receiver operating characteristic curve of 0.779 for mortality prediction. Despite promising results, challenges persist, including the computational costs of SHAP and local interpretable model-agnostic explanations analyses (e.g., TreeSHAP requiring distributed computing for metabolomics datasets) and dataset biases (e.g., SEER’s Western population dominance limiting generalizability). Future research must address inter-cohort heterogeneity, standardize explainability metrics, and prioritize lightweight surrogate models for resource-limited settings. This review presents the potential of explainable ensemble learning frameworks to bridge the gap between predictive accuracy and clinical interpretability, though rigorous validation in independent, multi-center cohorts is critical for real-world deployment.
Core Tip: Explainable artificial intelligence (XAI) seeks to improve the interpretability and transparency of machine learning models in healthcare settings. In this context, Explainable Ensemble Learning, a fundamental strategy within XAI, integrates multiple models, including Random Forest, Extreme Gradient Boosting, and Stacking, to improve classification performance in hepatocellular carcinoma (HCC). Despite their high predictive accuracy, the inherent "black-box" feature of ensemble methods remains a barrier to clinical practice. XAI techniques—such as SHapley Additive exPlanations, Local Interpretable Model-agnostic Explanations, and Gradient-weighted Class Activation Mapping—clarify model predictions, fostering medical trust and interpretability. By combining clinical, genetic, and imaging data with XAI frameworks, diagnosis, staging, and prognosis of HCC can be improved, ultimately supporting transparent and reliable decision-making in healthcare. Future research should focus on model interpretability, data integration, and user-friendly clinical interfaces.