BPG is committed to discovery and dissemination of knowledge
Minireviews
Copyright: ©Author(s) 2026.
World J Gastrointest Endosc. Apr 16, 2026; 18(4): 117976
Published online Apr 16, 2026. doi: 10.4253/wjge.v18.i4.117976
Table 1 Representative artificial intelligence models for the entire endoscopic ultrasound workflow
Application field
Ref.
Algorithm type
Sample size
Validation method
Application scenarios
Accuracy (%)
Sensitivity/specificity (%)
PPV (%)
NPV (%)
F1 score
AUC (95%CI)
Public availability/ reproducibility
Pre-proceduralSirtl et al[11], 2023Machine Learning218 patientsMulticenter retrospectiveBiliary sludge risk screening8497.9/9289.598.20.930.96 (0.92-0.98)Code: Not available; data: Not shared
Chen et al[14], 2024Predictive model> 1000 patientsMulticenter prospectiveEndoscopic resection outcome predictionNANANANANANAModel: Not released; data: Institutional
Intra-proceduralWu et al[18], 20235 DL models ensemble290 patientsRCTAnatomical landmark trackingNANANANANANACode: Not available; system: Proprietary (EUS-IREAD)
Li et al[19], 2025Automated reporting system114 patientsProspective trialStandard site documentation90.3NANANANANACode: Closed source; data: Not public
Rizzatti et al[20], 2025DCNN550 casesTechnical validationReal-time anatomical navigationNANANANANANANot publicly available
Image diagnosisHirai et al[23], 2022EfficientNetV2-L631 casesMulticenter retrospective5-class SEL differentiation89.398.8/67.685.494.20.910.94 (0.90-0.97)Code: Not shared; data: Multi-institutional (restricted)
Joo et al[24], 2024Random forestNASingle-centerGIST intervention prediction89.693.8/81.888.989.50.910.896 (0.86-0.93)Code: Not available; data: Not disclosed
Liu et al[32], 2022RetinaNet + VGGNetNASingle-centerDepth/origin classification82.580.2/90.687.1850.830.88 (0.84-0.92)Not publicly released
Zhang et al[33], 2025YOLOv8s-seg + MobileNetv2NASingle-centerLayer identification76.50NANANANANACode: Not available; model: Proprietary
Cui et al[39], 2024Joint-AI12 physicians RCTSolid lesion diagnosis90.091.0/88.089.589.80.90.93 (0.89-0.96)Code: Not shared; data: Anonymized, available on request
Orzan et al[42], 2024CNN + DeepLabv3+112 cases (1248 images)Single-centerdCCA detection and segmentation97.8100.0/94.496.81000.980.99 (0.97-1.00)Not publicly available
Men et al[46], 2025ResNet50554 cases (8738 images)Dual-centerIdentification of 3 colon tumors80.972.9/84.478.579.80.750.85 (0.81-0.89)Code: Not released; data: Multi-center (restricted)
Cytology diagnosisFujii et al[52], 2024Transformer45 cases (4059 images)Augmentation testMalignant cell detection88.291.4/85.086.790.10.890.954 (0.93-0.98)Code: GitHub; data: Partially open
Fang et al[53], 2025SSCNNNASemi-supervised frameworkCytology analysis with limited labels95.193.8/97.396.594.80.950.97 (0.94-0.99)Code: Not available; model: Internal use only
Table 2 Comparison of regulatory approaches to artificial intelligence-assisted endoscopic ultrasound in foreign countries
Feature
United States (FDA)
European Union (MDR/IVDR + AI Act)
Regulatory coreFunction-based (what the software does) tiered regulationDevice risk + AI system risk dual matrix regulation.
Classification logicCADt → CADe → CADx → CADa, with increasing riskClass I, IIa, IIb, III (device risk), plus the AI Act’s “high-risk” category for most medical AI
Primary pathways510(k) (substantial equivalence), De Novo (novel low-moderate risk), PMA (high risk)Self-certification (class I low risk), notified body conformity assessment (class IIa and above)
AdaptabilityPredetermined change control plans allow for iterative updates to cleared algorithms within defined bounds without new submissionUpdate processes under regulations are stricter; significant software changes typically require re-assessment/notification to the notified body
Clinical evidenceEmphasizes prospective, multi-center clinical trials to demonstrate safety and effectivenessEmphasizes clinical evaluation with comprehensive technical documentation and performance validation, aligned with GDPR
TransparencyRequires disclosure of algorithm performanceThe AI act emphasizes transparency, requiring high-risk AI systems to provide clear usage information and ensure outputs are interpretable and overseen by humans
Process flowchartsAI medical device concept → determine intended use and function: CADt, CADe, CADx, or CADa → assign FDA risk class → generate clinical and technical evidence → submit to FDA for review → FDA approval clearance → post-market surveillanceMedical AI product → determine device risk class per MDR/IVDR → determine AI system risk level per AI act (typically “high-risk”) → comply with both sets of requirements → undergo conformity Assessment by a notified body (for class IIa and above) → obtain CE marking
Table 3 Phased clinical integration roadmap
Phase
Primary goals
Key evidence required
Regulatory milestones
Phase 1 (1-2 years)Low-risk auxiliary tasks (e.g., automated measurement, structured reporting)Feasibility studies, human-machine interaction safety dataRegistration or simplified clearance (e.g., class II medical device)
Phase 2 (3-5 years)Moderate-risk diagnostic support (e.g., lesion detection and classification)Multicenter diagnostic accuracy studies, clinical utility assessmentModerate-risk device approval (requires performance monitoring plan)
Phase 3 (> 5 years)High-risk autonomous functions (e.g., puncture path planning, real-time complication alert)Large-scale RCTs: Real-world effectiveness dataHigh-risk device approval (requires a full post-market surveillance system)