Published online Dec 15, 2025. doi: 10.4251/wjgo.v17.i12.114341
Revised: September 26, 2025
Accepted: October 27, 2025
Published online: December 15, 2025
Processing time: 86 Days and 0.2 Hours
As large language models increasingly permeate medical workflows, a recent study evaluating ChatGPT 4.0’s performance in addressing patient queries about endoscopic submucosal dissection and endoscopic mucosal resection offers critical insights into three domains: Performance parity, cost democratization, and clinical readiness. The findings highlight ChatGPT’s high accuracy, completeness, and comprehensibility, suggesting potential cost efficiency in patient education. Yet, cost-effectiveness alone does not ensure clinical utility. Notably, the study relied exclusively on text-based prompts, omitting multimodal data such as pho
Core Tip: As large language models increasingly permeate medical workflows, this study offers insight into 3 areas: Performance parity, cost democratization, and clinical readiness. Perhaps one potentially compelling finding was cost efficiency. Yet cost-effectiveness alone does not ensure clinical utility. Notably, the study relied exclusively on text-based prompts, omitting multimodal data such as photographs or scans. This is an important limitation in a domain like endoscopy, which often can be driven visually. Large language model performance can drop precipitously when deprived of image context. Without multimodal integration, artificial intelligence tools may inevitably fail to capture key diagnostic signals.
