BPG is committed to discovery and dissemination of knowledge
Minireviews
Copyright: ©Author(s) 2026. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution-NonCommercial (CC BY-NC 4.0) license. No commercial re-use. See permissions. Published by Baishideng Publishing Group Inc.
World J Methodol. Jun 20, 2026; 16(2): 115059
Published online Jun 20, 2026. doi: 10.5662/wjm.v16.i2.115059
Biases of large language models in diagnosing Cushing’s syndrome
Christos Savvidis, Costas Liakopoulos, Ioannis Ilias
Christos Savvidis, Ioannis Ilias, Department of Endocrinology, Hippocration General Hospital of Athens, Athens GR-11527, Attikí, Greece
Costas Liakopoulos, Quality Assurance Specialist, Athens GR-11854, Attikí, Greece
Author contributions: Savvidis C, Liakopoulos C, and Ilias I researched the literature and wrote the draft, final version of the article, and they thoroughly reviewed and endorsed the final manuscript.
Conflict-of-interest statement: All the authors report no relevant conflicts of interest for this article.
Corresponding author: Ioannis Ilias, MD, PhD, Director, Department of Endocrinology, Hippocration General Hospital of Athens, No. 63 Evrou Street, Athens GR-11527, Attikí, Greece. iiliasmd@yahoo.com
Received: October 9, 2025
Revised: October 31, 2025
Accepted: January 5, 2026
Published online: June 20, 2026
Processing time: 198 Days and 22 Hours
Core Tip

Core Tip: The diagnosis of Cushing’s syndrome remains challenging due to its rarity and its resemblance to common metabolic disorders. Large language models and other artificial intelligence-capable systems are potential diagnostic tools for early detection and differential diagnosis; nevertheless, they are likely to strengthen both human cognitive and training data biases. Large language models are susceptible to biases in the textual data they are trained on, reflecting human cognitive biases, while traditional machine learning models are susceptible to biases in structured data, leading to spectrum and measurement bias. Spectrum bias, exclusion of demographic variables, and heterogeneity of the data undermine diagnostic validity and justice. These heuristics are similar to clinicians’ mental shortcuts - anchoring, availability, and framing - and share the same diagnostic bias. Transparency, data diversity, and clinically relevant predictors are required to build unbiased, interpretable artificial intelligence solutions for endocrine diagnosis.