Savvidis C, Liakopoulos C, Ilias I. Biases of large language models in diagnosing Cushing’s syndrome. World J Methodol 2026; 16(2): 115059 [DOI: 10.5662/wjm.v16.i2.115059]
Corresponding Author of This Article
Ioannis Ilias, MD, PhD, Director, Department of Endocrinology, Hippocration General Hospital of Athens, No. 63 Evrou Street, Athens GR-11527, Attikí, Greece. iiliasmd@yahoo.com
Research Domain of This Article
Endocrinology & Metabolism
Article-Type of This Article
Minireviews
Open-Access Policy of This Article
This article is an open-access article which was selected by an in-house editor and fully peer-reviewed by external reviewers. It is distributed in accordance with the Creative Commons Attribution Non Commercial (CC BY-NC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited and the use is non-commercial. See: http://creativecommons.org/licenses/by-nc/4.0/
Jun 20, 2026 (publication date) through Apr 23, 2026
Times Cited of This Article
Times Cited (0)
Journal Information of This Article
Publication Name
World Journal of Methodology
ISSN
2222-0682
Publisher of This Article
Baishideng Publishing Group Inc, 7041 Koll Center Parkway, Suite 160, Pleasanton, CA 94566, USA
Share the Article
Savvidis C, Liakopoulos C, Ilias I. Biases of large language models in diagnosing Cushing’s syndrome. World J Methodol 2026; 16(2): 115059 [DOI: 10.5662/wjm.v16.i2.115059]
Author contributions: Savvidis C, Liakopoulos C, and Ilias I researched the literature and wrote the draft, final version of the article, and they thoroughly reviewed and endorsed the final manuscript.
Conflict-of-interest statement: All the authors report no relevant conflicts of interest for this article.
Corresponding author: Ioannis Ilias, MD, PhD, Director, Department of Endocrinology, Hippocration General Hospital of Athens, No. 63 Evrou Street, Athens GR-11527, Attikí, Greece. iiliasmd@yahoo.com
Received: October 9, 2025 Revised: October 31, 2025 Accepted: January 5, 2026 Published online: June 20, 2026 Processing time: 198 Days and 22.2 Hours
Abstract
The diagnosis of endogenous Cushing’s syndrome (CS) can be complicated and often delayed, given its low incidence (estimated globally at 1.8 cases to 4.5 cases per million people per year) and its clinical features that mimic far more prevalent metabolic disorders, such as central obesity, hypertension, and glucose intolerance. In clinical practice, physicians rely on cognitive heuristics that are prone to error, contributing to diagnostic delays (on average around 34 months pass from symptom onset to diagnosis of CS). Large language models and machine learning algorithms could be potential decision-support tools for screening and differential diagnosis of CS. However, these systems are at risk of inheriting and even amplifying existing cognitive biases and data-driven distortions embedded in their training data. Machine learning models designed for CS could be vulnerable to methodological flaws, notably spectrum bias and the exclusion of clinically relevant demographic variables, demanding attention from the endocrine and medical informatics communities. This paper examines how cognitive and algorithmic biases intersect in diagnostic models for CS, highlighting parallels between human diagnostic heuristics (e.g., anchoring, availability, and framing) and data-driven distortions (e.g., spectrum and measurement bias) in artificial intelligence.
Core Tip: The diagnosis of Cushing’s syndrome remains challenging due to its rarity and its resemblance to common metabolic disorders. Large language models and other artificial intelligence-capable systems are potential diagnostic tools for early detection and differential diagnosis; nevertheless, they are likely to strengthen both human cognitive and training data biases. Large language models are susceptible to biases in the textual data they are trained on, reflecting human cognitive biases, while traditional machine learning models are susceptible to biases in structured data, leading to spectrum and measurement bias. Spectrum bias, exclusion of demographic variables, and heterogeneity of the data undermine diagnostic validity and justice. These heuristics are similar to clinicians’ mental shortcuts - anchoring, availability, and framing - and share the same diagnostic bias. Transparency, data diversity, and clinically relevant predictors are required to build unbiased, interpretable artificial intelligence solutions for endocrine diagnosis.