BPG is committed to discovery and dissemination of knowledge
Review Open Access
Copyright ©The Author(s) 2025. Published by Baishideng Publishing Group Inc. All rights reserved.
World J Gastroenterol. Sep 28, 2025; 31(36): 111137
Published online Sep 28, 2025. doi: 10.3748/wjg.v31.i36.111137
Application of deep learning-based convolutional neural networks in gastrointestinal disease endoscopic examination
Yang-Yang Wang, School of Physics and Electronic Information, Yan’an University, Yan’an 716000, Shaanxi Province, China
Bin Liu, Department of Pharmacy, Hubei Cancer Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan 430079, Hubei Province, China
Ji-Han Wang, Yan’an Medical College, Yan’an University, Yan’an 716000, Shaanxi Province, China
ORCID number: Yang-Yang Wang (0000-0002-4753-7193); Bin Liu (0009-0009-0259-6059); Ji-Han Wang (0000-0003-1925-330X).
Co-first authors: Yang-Yang Wang and Bin Liu.
Author contributions: Wang YY and Liu B made equal contributions as co-first authors; Wang YY and Wang JH conceptualized and designed the study, and constructed figures presented in this manuscript; Wang YY searched and reviewed published articles, and wrote the original manuscript; Liu B made revisions to the revised manuscript; Wang JH reviewed the original manuscript, and provided the funding acquisition; all authors approved the submitted version.
Supported by Open Funds for Shaanxi Provincial Key Laboratory of Infection and Immune Diseases, No. 2023-KFMS-1.
Conflict-of-interest statement: All the authors report no relevant conflicts of interest for this article.
Open Access: This article is an open-access article that was selected by an in-house editor and fully peer-reviewed by external reviewers. It is distributed in accordance with the Creative Commons Attribution NonCommercial (CC BY-NC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited and the use is non-commercial. See: https://creativecommons.org/Licenses/by-nc/4.0/
Corresponding author: Ji-Han Wang, MD, PhD, Yan’an Medical College, Yan’an University, No. 580 Shengdi Road, Yan’an 716000, Shaanxi Province, China. jihanwang@yau.edu.cn
Received: June 24, 2025
Revised: August 7, 2025
Accepted: August 26, 2025
Published online: September 28, 2025
Processing time: 87 Days and 13.3 Hours

Abstract

Gastrointestinal (GI) diseases, including gastric and colorectal cancers, significantly impact global health, necessitating accurate and efficient diagnostic methods. Endoscopic examination is the primary diagnostic tool; however, its accuracy is limited by operator dependency and interobserver variability. Advancements in deep learning, particularly convolutional neural networks (CNNs), show great potential for enhancing GI disease detection and classification. This review explores the application of CNNs in endoscopic imaging, focusing on polyp and tumor detection, disease classification, endoscopic ultrasound, and capsule endoscopy analysis. We discuss the performance of CNN models with traditional diagnostic methods, highlighting their advantages in accuracy and real-time decision support. Despite promising results, challenges remain, including data availability, model interpretability, and clinical integration. Future directions include improving model generalization, enhancing explainability, and conducting large-scale clinical trials. With continued advancements, CNN-powered artificial intelligence systems could revolutionize GI endoscopy by enhancing early disease detection, reducing diagnostic errors, and improving patient outcomes.

Key Words: Gastrointestinal diseases; Endoscopic examination; Deep learning; Convolutional neural networks; Computer-aided diagnosis

Core Tip: This review summarizes the latest advances in the application of deep learning-based convolutional neural networks in gastrointestinal endoscopy. It highlights convolutional neural networks’ roles in lesion detection, classification, segmentation, and real-time decision support, emphasizing their potential to enhance diagnostic accuracy, reduce variability, and integrate into clinical workflows for improved patient outcomes.



INTRODUCTION

Gastrointestinal (GI) diseases, from gastric and colorectal cancers (CRC) to inflammatory bowel disease (IBD), represent a major global health burden[1,2]. The early and accurate diagnosis of these conditions is critical for improving patient prognosis and optimizing treatment outcomes[3]. Endoscopic examination is a cornerstone of GI diagnosis, offering direct visualization of the mucosal surface and enabling real-time biopsy and therapeutic intervention[4,5].

Despite its key role in clinical gastroenterology, conventional endoscopy has several limitations[6]. The diagnostic accuracy of conventional endoscopy is often influenced by the experience and observational skills of the endoscopist. This can lead to significant interobserver variability and a risk of missed lesions, particularly small or flat neoplasms[7]. The increasing demand for procedures and the volume of data also challenge the consistency and efficiency of diagnosis[8]. Thus, technological innovations are urgently needed to enhance the reliability of endoscopic evaluations. Artificial intelligence (AI), particularly deep learning (DL), has emerged as a transformative tool in the field of medical imaging[9,10]. Among DL techniques, convolutional neural networks (CNNs) are especially well-suited for analyzing complex endoscopic image data, outperforming traditional machine learning algorithms in various tasks like image recognition, classification, and segmentation[11-15]. Recent studies have increasingly explored integrating CNNs into GI endoscopy to develop computer-aided detection and diagnosis systems[11,16-18]. These AI-driven systems have demonstrated promising results in enhancing the detection of colorectal polyps and classifying gastric lesions, with performance often comparable to or exceeding that of experienced endoscopists[18-20].

This review provides a comprehensive overview of the current state and future prospects of CNN applications in GI endoscopy. This is a narrative review, and the literature was surveyed by searching recent publications, with a focus on high-impact studies and landmark papers, to ensure a broad and unbiased representation of the field. As depicted in Figure 1, we first introduce fundamental CNN principles, then systematically examine their applications across various GI diagnostic scenarios. Additionally, we discuss model performance, challenges, and future directions to advance the clinical utility of CNN-based systems in endoscopic practice.

Figure 1
Figure 1 Organizational framework of the review. The diagram illustrates the logical flow of the manuscript, beginning with the clinical background of gastrointestinal diseases. It then progresses to the fundamental principles of convolutional neural networks and their specific applications in endoscopic examination, categorized into three main domains. The review concludes by addressing the key challenges and future perspectives for the clinical translation of these technologies. GI: Gastrointestinal; CNNs: Convolutional neural networks; AI: Artificial intelligence.
FUNDAMENTALS OF CNN IN MEDICAL IMAGE ANALYSIS

CNNs are a class of DL models specifically designed for processing data with a grid-like topology, such as images. CNNs have revolutionized the field of computer vision and have shown outstanding performance in a wide range of medical imaging tasks, including classification, detection, segmentation, and anomaly recognition. Their success lies in their ability to automatically and adaptively learn spatial hierarchies of features through backpropagation, reducing the need for manual feature engineering.

Basic architecture of CNNs

The basic architecture of a CNN typically comprises the following key components: An input layer, convolutional layers, activation functions, pooling layers, fully connected layers, and an output layer (Figure 2). The input layer serves as the entry point of the network, where raw image data, such as endoscopic images for detecting GI diseases, is fed into the model. This input is usually represented as a multi-dimensional tensor, incorporating the width, height, and the number of color channels (e.g., RGB). The convolutional layers are the core building blocks of CNNs. They apply a series of learnable filters (kernels) that slide over the input image to produce feature maps. These filters are designed to detect low-level patterns such as edges, textures, and contours in early layers, as well as more complex features such as lesions or anatomical structures in deeper layers. Figure 3 provides a detailed view of key operations within the network: Panel A shows a convolution operation using a 3 × 3 kernel that produces a 4 × 4 output feature map, while panel B demonstrates max-pooling with a 2 × 2 filter and stride of 2, which reduces the spatial resolution to 2 × 2. These operations not only condense and refine visual information but also enable the network to become increasingly sensitive to clinically relevant patterns in deeper layers. Together, this architectural design allows CNNs to support real-time decision-making in endoscopy, improve diagnostic accuracy, and reduce variability in clinical interpretation.

Figure 2
Figure 2 The basic architecture of a convolutional neural network. The architecture includes an input layer for receiving endoscopic image data, multiple convolutional layers for extracting spatial features, pooling layers to downsample and retain salient regions, fully connected layers to integrate high-level representations, and an output layer for classification or regression. Non-linear activations such as rectified linear unit are used between layers to introduce complexity.
Figure 3
Figure 3 Convolution and pooling diagram. A: Convolution operation with a 3 × 3 kernel resulting in a 4 × 4 output; B: Max-pooling operation using a 2 × 2 filter and stride 2 × 2, reducing feature map dimensions to 2 × 2. These operations illustrate how convolutional neural networks extract features and reduce spatial resolution while preserving critical information.
Training CNNs for medical image analysis

CNNs are trained using large datasets of labeled images. During training, the network adjusts its weights to minimize a loss function (e.g., cross-entropy for classification tasks), typically using stochastic gradient descent or its variants. In medical imaging, acquiring large, annotated datasets can be challenging due to privacy concerns, labeling costs, and inter-observer variability[21]. To address this, researchers often use data augmentation (e.g., rotations, flipping, and noise addition) and transfer learning, where pre-trained CNNs (e.g., ResNet, VGG, and Inception) are fine-tuned on medical datasets[22,23].

Moreover, model evaluation requires careful selection of performance metrics beyond simple accuracy. Commonly used metrics in medical applications include sensitivity, specificity, precision, F1-score, and area under the receiver operating characteristic curve, as these metrics provide a highly comprehensive assessment of the clinical relevance of a model[24,25]. For segmentation tasks, Dice coefficient and intersection over union are widely used[26,27]. However, high accuracy alone is insufficient; models must also generalize across diverse populations and imaging conditions. To this end, validation strategies such as cross-validation, external validation, and prospective testing are employed, with external and real-time clinical validations being particularly critical[28,29]. Public datasets like Kvasir and GIANA support standardized benchmarking but are limited by variability in annotation and image quality[30,31]. Furthermore, consistent reporting by following AI-specific guidelines such as CONSORT-AI and TRIPOD-AI is vital for reproducibility and comparison across studies[32].

Why CNNs are suited for GI endoscopic image analysis

GI endoscopic images present complex visual features such as mucosal patterns, vascular structures, and subtle lesion morphology. CNNs excel in recognizing such hierarchical and spatial features because of their local receptive fields and weight-sharing mechanisms[33]. Unlike traditional machine learning algorithms that heavily rely on handcrafted features, CNNs automatically extract relevant features, enabling more objective and scalable diagnostic tools[34,35]. Furthermore, CNNs can be applied to both static images and dynamic video streams, making them particularly suitable for real-time applications in GI endoscopy[36]. Advanced architectures such as 3D CNNs or temporal convolutional networks have also been developed to handle sequential video data, capturing spatial-temporal information critical for detecting transient abnormalities or tracking lesion progression[37,38].

Common CNN architectures in medical imaging

As demonstrated in Table 1[39-50], several CNN architectures, originally developed for computer vision tasks, have been successfully adapted and applied to the domain of medical image analysis. These adaptations often involve fine-tuning pre-trained models or designing novel architectures tailored to the specific characteristics and challenges inherent in medical imaging data. The success of CNNs in medical applications highlights their robust feature extraction capabilities and their ability to learn complex patterns from image data. This has led to advancements in tasks such as disease detection, segmentation, and diagnosis across various medical modalities.

Table 1 Convolutional neural network architectures adapted for medical image analysis.
Ref.
Architecture
Key strengths
Limitations
Applications
[39]AlexNetIntroduced rectified linear unit, dropout, graphics processing unit accelerationOverfitting on small datasetsHistopathology image classification
[40]VGGNetUniform structure, easy to implementLarge number of parameters, memory-intensivePolyp detection, organ segmentation
[41]GoogLeNetMulti-scale feature extraction, fewer parameters than VGGComplex architecture, harder to modifyLesion classification, colonoscopy image analysis
[42]ResNetResidual connections solve vanishing gradientCan overfit if dataset is smallDetection of GI tumors, segmentation of ulcers
[43]U-NetExcellent for biomedical segmentation, works with few imagesLimited to segmentation tasksPolyp segmentation, mucosal layer delineation
[44]DenseNetStrong gradient flow, parameter-efficientComputationally intensiveEndoscopic image classification, disease grading
[45]Attention U-NetIncorporates attention for better focus on relevant regionsIncreased complexity and longer training timeIBD severity scoring; small bowel bleeding detection
[46]EfficientNetOptimized trade-off between accuracy and speedRequires careful scaling and tuningLightweight, mobile-compatible GI image classification
[47]Swin-CNNWindow-based attention + CNN; balances local and global featuresComplex design; tuning is more demandingGI endoscopy video anomaly detection; GI tumor recognition
[48]TransUNetCombines CNN for feature extraction and Transformer for context modelingResource intensive; slower trainingComputed tomography/magnetic resonance imaging organ segmentation; GI lesion boundary detection
[49]MedT (medical transformer)Pure transformer-based; excellent at long-range dependency modelingNot optimal for small datasets; data hungryIntestinal lesion segmentation; colorectal cancer prediction
[50]ConvNeXtCombines CNN stability with Transformer-like design; efficient trainingRelatively new; limited ecosystem maturityMulti-organ classification; tumor region detection

VGGNet, exemplified by VGG-16 (Figure 4), is notable for its depth and use of uniform 3 × 3 convolution kernels, enabling a simpler yet more powerful design than earlier models like AlexNet. Its success laid the foundation for deeper networks such as ResNet. U-Net (Figure 5), widely used in medical image segmentation, features a symmetric U-shaped architecture with a Contracting Path (encoder) that captures context through downsampling, and an Expansive Path (decoder) that restores spatial resolution for precise localization. Skip connections link encoder and decoder layers at corresponding levels, enhancing the network’s ability to learn both global and local features. This design makes U-Net especially effective in delineating fine structures in medical images.

Figure 4
Figure 4 VGGNet-16 architecture. VGGNet comprises 16 weight layers and is characterized by repeated use of small (3 × 3) convolution filters and a uniform layer design. This deep and consistent architecture enhances its ability to capture hierarchical features, and laid the foundation for deeper networks like ResNet.
Figure 5
Figure 5 The U-Net architecture, widely used in medical image segmentation. The left side is the contracting path (encoder), which reduces spatial resolution while increasing feature complexity. The right side is the expansive path (decoder), which restores spatial resolution. Skip connections link corresponding encoder-decoder levels to retain spatial detail and improve segmentation accuracy.
Comparative analysis of key CNN architectures

The choice of CNN architecture is critical for achieving optimal performance, and various models have been adopted for GI endoscopy tasks, each with its own strengths and weaknesses. A direct comparison of these architectures reveals important trade-offs between performance, computational complexity, and suitability for specific tasks. For classification tasks, models like ResNet and EfficientNet are commonly used. ResNet, with its residual connections, effectively mitigates the vanishing gradient problem, allowing for the training of very deep networks. It has been widely adopted for polyp and lesion classification due to its robust performance. However, its high number of parameters can lead to significant computational overhead. In contrast, EfficientNet utilizes a compound scaling method to uniformly scale network depth, width, and resolution, leading to a much more efficient model with fewer parameters and higher accuracy. This makes it particularly suitable for scenarios where computational resources are limited, such as real-time deployment on edge devices.

For segmentation tasks, architectures based on the U-Net family are dominant. The classic U-Net architecture, with its contracting and expanding paths and skip connections, is highly effective for segmenting lesions with high precision. Its strengths lie in its ability to fuse high-resolution features from the encoder with high-level features from the decoder, which is crucial for pixel-level prediction. The more recent TransUNet, which integrates Transformer-based self-attention mechanisms into the U-Net structure, aims to capture long-range dependencies more effectively than pure CNN-based models. While TransUNet may offer improved performance on certain complex segmentation tasks, it often comes with a higher computational cost and a greater demand for large-scale training data. The decision between these architectures often depends on the specific clinical application. For instance, a ResNet or EfficientNet might be sufficient for a classification system to quickly screen for polyps, while a U-Net or its variants would be necessary for precise boundary detection for surgical planning or automated lesion size measurement.

APPLICATIONS OF CNN IN GI ENDOSCOPY

CNNs have demonstrated transformative potential in GI endoscopy, where the accurate and timely interpretation of endoscopic images is critical for disease diagnosis and management[51]. By learning spatial hierarchies of visual features directly from data, CNNs have enabled automation of key diagnostic tasks, reducing interobserver variability and improving lesion detection accuracy[51]. Their core clinical applications in GI endoscopy can be categorized into three major domains: Lesion detection and classification, lesion segmentation and disease severity assessment, and real-time clinical assistance and workflow integration.

Lesion detection and classification

The detection and classification of lesions represent foundational applications of CNNs in GI endoscopy[52,53]. These tasks are critical for the early diagnosis of malignancies and the prevention of cancer progression, especially in the context of CRC and upper GI neoplasms. CNNs enable automatic extraction of hierarchical image features and recognition of subtle visual patterns in endoscopic images, which may be challenging even for expert clinicians to identify.

In colonoscopy, CNN-based systems have been particularly effective in identifying colorectal polyps[54-56]. Missed polyps, especially flat or diminutive ones, are a well-known contributor to interval CRCs. Studies have demonstrated that CNN models trained on large annotated image datasets can achieve high sensitivity and specificity, often surpassing generalist endoscopists in polyp detection. For example, CNNs have been employed to distinguish adenomatous from hyperplastic polyps in real time, facilitating immediate clinical decision-making regarding resection or surveillance. Additionally, CNNs have been trained to classify polyps based on size, morphology, and surface patterns, such as the narrow-band imaging international colorectal endoscopic or the Workgroup Serrated Polyps and Polyposis classifications, further aiding risk stratification[57]. In upper GI endoscopy, lesion detection is equally critical but often more complex because of the heterogeneous nature of mucosal appearance. CNNs have shown promising results in detecting early gastric cancer, which may present as subtle discoloration, textural changes, or shallow depressions[57]. Similarly, CNN models have been developed to identify esophageal squamous cell carcinoma and dysplastic Barrett’s esophagus, outperforming or matching the performance of experienced specialists[57,58]. Many of these models are trained using narrow-band imaging, blue-light imaging, or linked color imaging to enhance mucosal contrast and improve lesion visibility[59,60].

One of the major strengths of CNNs in classification tasks lies in their ability to learn directly from raw pixel data without requiring handcrafted features. This feature allows them to generalize across different imaging modalities, endoscope brands, and lighting conditions, provided they are trained on diverse and well-curated datasets. Some models are now being designed with attention mechanisms or multi-task architectures to simultaneously detect and classify lesions while highlighting salient image regions to improve interpretability[61,62]. Looking forward, the integration of CNNs with electronic health records, histopathology, and genetic data could enable more personalized risk prediction and lesion characterization[63]. Furthermore, federated learning and domain adaptation techniques may allow CNNs to perform consistently across institutions, even when local data cannot be shared due to privacy concerns. Ultimately, CNNs hold significant promise to standardize and enhance lesion detection and classification in GI endoscopy, reducing variability, improving outcomes, and enabling earlier intervention. Table 2 offers a comprehensive overview of studies that used CNNs for lesion detection and classification in GI disease[64-109].

Table 2 Application of convolutional neural networks for lesion detection and classification in gastrointestinal diseases.
Data
Architecture
Application
Key findings
Ref.
Endoscopic imagesDeep CNN with high-resolution endoscopic inputAids in detecting laterally spreading tumors in the colon using deep learning to enhance endoscopic visualization accuracyThe CNN model significantly improved the detection rate of laterally spreading tumors compared to conventional endoscopy, reducing oversight and potentially improving early colorectal cancer diagnosis. Accuracy increased notably in identifying flat lesions that are otherwise difficult to detect manually[64]
GI endoscopy imagesDeep CNN trained on large annotated image datasetDesigned to detect and classify a wide range of GI tract diseases in real-time using endoscopic imageryThe model achieved high diagnostic accuracy across multiple disease categories, outperforming traditional classifiers. It demonstrated strong potential for integration into clinical decision support tools, increasing consistency and efficiency in endoscopic diagnoses[65]
Wireless capsule endoscopy imagesCNN with real-time image processing pipelineSupports non-invasive detection of GI diseases using capsule endoscopy video sequencesAchieved expert-level accuracy in detecting abnormal lesions in small bowel images. This system allows for automated and accurate screening, significantly reducing time spent by clinicians reviewing lengthy capsule footage[66]
White-light endoscopyCNN model trained on early gastric cancer casesEarly detection of gastric cancer through white-light endoscopy imagery with AI assistanceThe model enabled early cancer identification with high sensitivity and specificity, improving upon the detection rates of endoscopists. Its implementation could lead to reduced gastric cancer mortality by facilitating prompt treatment decisions[67]
Endoscopic imagesCNN with multi-center training dataDetecting Helicobacter pylori infection from endoscopy imagesThe CNN demonstrated high performance and generalizability across institutions. It offers a non-invasive, fast diagnostic method, comparable to biopsy confirmation, enabling real-time feedback during endoscopic exams[68]
Gastroscopy imagesDeep dense CNN with channel attention mechanismAutomated classification of esophageal disease types (e.g., ulcers, neoplasms)By incorporating attention modules, the model excelled in distinguishing subtle texture differences among esophageal pathologies, showing great promise in supporting less experienced endoscopists[58]
Capsule endoscopy imagesCNN for object detectionAutomatic detection of small intestinal hookwormsThe proposed system achieved high precision in detecting parasitic lesions, helping clinicians quickly identify the presence of hookworm infections, which are often missed due to subtle image features[69]
Capsule endoscopy imagesCNN trained on hemorrhagic potential lesionsIdentification and differentiation of small bowel lesions with bleeding riskAccurately categorized lesions by bleeding risk, offering clinicians critical insight for prioritizing cases. The system reduces subjectivity in evaluating lesion severity[70]
Endoscopic imagesDeep CNN classifier for esophageal lesion typesDifferentiation of protruding esophageal lesions (e.g., GI stromal tumors, leiomyomas)CNN provided more accurate differential diagnosis than generalist endoscopists. Can support pre-biopsy decision-making, potentially reducing unnecessary interventions[71]
Endoscopic imagesTransfer CNN architecture (e.g., ResNet-based)Detection of early gastric cancerTransfer learning from pre-trained CNNs enabled rapid training on limited labeled data while maintaining strong detection performance. Highlights usefulness of transfer learning in GI imaging where large datasets are rare[72]
Endoscopic imagesFusion of multiple CNNs via fuzzy Minkowski distanceMulti-class classification of GI disordersThis fuzzy fusion method significantly boosted CNN decision reliability in classifying various GI tract diseases. Particularly effective in handling noisy or borderline cases by leveraging ensemble confidence[73]
Double-balloon enteroscopyCustom deep CNN with lesion region focusingDetecting and classifying small bowel lesions (e.g., ulcers, tumors) during double-balloon enteroscopyAchieved high sensitivity for subtle and rare lesions in the small intestine. Helped reduce diagnostic time and enabled detection of hard-to-access areas, improving double-balloon enteroscopy clinical utility[74]
Endoscopic imagesCNN with transfer learning (e.g., ResNet pretrained on ImageNet)Detecting GI abnormalities in multi-center datasetsTransfer learning approach enhanced classification accuracy with limited training data. It proved robust across image variations, useful in practical endoscopy deployments with different imaging devices[75]
GI tract imagesDeep CNN with region proposal and hierarchical labelingClassification of syndromes in the GI tract (e.g., inflammation, bleeding, neoplasms)Outperformed traditional classifiers with over 90% accuracy on several syndrome categories. System demonstrated early signs of suitability for autonomous decision-making in AI-assisted endoscopy[76]
Capsule endoscopy imagesCNN with spatial attention and region localizationDetection and precise localization of GI diseases in capsule footageThe model not only identified lesions but also marked their spatial location, reducing time required for manual review. Showed high generalization across bleeding, polyps, and ulcers[53]
Endoscopic imagesMulti-fusion CNN with spatial and feature fusionDiagnosis of complex GI disease states through multiple feature pathsFusion approach combined texture, edge, and color patterns from multiple CNN streams, improving classification of ambiguous or compound lesions. Enabled more granular diagnosis[77]
Capsule endoscopy imagesDeep multi-class CNN classifierDistinguishing ulcerative colitis, polyps, and dyed-lifted polypsThe model achieved strong performance on multi-label lesion classification, aiding clinicians in identifying inflammation patterns vs neoplastic changes from wireless capsule endoscopy images[56]
Endoscopic imagesDeep CNN for histological predictionDetection of chronic atrophic gastritisThe system outperformed general practitioners in identifying chronic atrophic gastritis on white-light images. This could enable earlier intervention for patients at risk of gastric cancer[78]
Endoscopic imagesCNN-based diagnostic prediction modelClassification of multiple GI disorders including gastritis and neoplasmsThe system offered a probabilistic diagnosis output that correlated well with clinical reports. Could serve as a second reader to reduce misdiagnosis[79]
Endoscopic imagesDeep CNN trained on villous atrophy casesDetection of duodenal villous atrophy, a hallmark of celiac diseaseModel enabled accurate identification of atrophic mucosal features that are often missed by the naked eye. Can improve non-invasive screening for celiac disease[80]
Endoscopic imagesCNN with lesion localization layerEarly gastric cancer classification and localizationThe system not only detected cancer but pinpointed lesion boundaries, assisting in targeted biopsies. Shown to reduce oversight in flat lesions[81]
Endoscopic imagesDeep CNN prospectively testedDiagnosis of chronic atrophic gastritisProspective validation showed significant improvement in diagnosis rate compared to conventional methods. Demonstrated practical deployment readiness[82]
Endoscopic imagesCNN trained and validated against multiple architecturesDetection of gastric mucosal lesionsCompared multiple CNN methods and showed the proposed model performed best in both sensitivity and specificity. Suggested a robust preprocessing pipeline[83]
Endoscopic imagesDeep CNN with classification and detection modulesBroad GI disease detection and classification, including ulcers and tumorsThe CNN model integrated multiple diagnostic modules, enabling both lesion classification and visual confirmation. It significantly outperformed traditional rule-based systems and provided explainable outputs to support clinical decisions[84]
Endoscopic imagesDeep CNN trained on annotated gastric polyp datasetsDetection of gastric polyps with high sensitivityThe model effectively identified polyps, including flat and small ones. It helped reduce miss rates in routine gastroscopy and can act as a real-time assistant for novice endoscopists[54]
Endoscopic imagesEnsemble machine learning with CNN and decision tree comparisonsGI disease classification from diverse image typesThough primarily a comparative study, CNNs were shown to outperform other models in most lesion categories. Highlighted CNNs’ suitability for complex image recognition tasks in GI settings[85]
Endoscopic imagesDeep convolutional architecture trained on expert-labeled dataDiagnosis of common gastric lesions including gastritis and erosionsDemonstrated high diagnostic accuracy and reduced inter-observer variability, suggesting CNN use can help standardize endoscopic assessments across institutions[86]
Capsule endoscopyDeep CNN with binary output modeBinary lesion detection (normal vs abnormal) in capsule endoscopyThe binary CNN model achieved excellent sensitivity in identifying abnormal frames, acting as an efficient pre-filter to aid clinicians reviewing thousands of capsule images[87]
Endoscopic imagesDeep CNN tested across multiple centersGastritis classification using AI comparable to gastroenterologistsAchieved expert-level performance in classifying gastritis subtypes. Multicenter validation proved its robustness across varied endoscopic systems and clinical environments[20]
Endoscopic imagesAttention-guided CNN with feature weightingGI disease classification enhanced with attention mechanismsAttention modules enabled the model to focus on key lesion areas, improving classification accuracy for overlapping or complex cases. Boosted explainability in predictions[62]
Wireless capsule endoscopy imagesPretrained CNN fine-tuned on wireless capsule endoscopy imagesGI tract disease classification from capsule endoscopyThe transfer learning-based CNN adapted well to wireless capsule endoscopy image features. It accurately classified disease regions, even under motion blur and illumination changes[88]
Endoscopic imagesGI-Net: CNN with anomaly detection capabilityMulti-label classification of anomalies in GI tractGI-Net offered near real-time classification across numerous disease types, showing strong performance in clinical pilot tests. Especially useful in large-scale screenings[89]
Endoscopic imagesDeep CNN trained on Helicobacter pylori-positive and negative imagesEvaluation of Helicobacter pylori infectionModel achieved diagnostic accuracy comparable to experienced gastroenterologists. Could be integrated into live endoscopy for on-the-spot infection prediction[90]
Gastric endoscopic imagesDeep CNN optimized for malignancy detectionClassification of gastric malignanciesCNN showed excellent discrimination between benign and malignant lesions. Real-time feedback capability demonstrated potential for improving early gastric cancer outcomes[91]
Endoscopic imagesCNN for cancer detection trained on large datasetsAutomatic gastric cancer detection from endoscopyHigh sensitivity and specificity achieved. Could act as an AI “second observer” for difficult-to-see lesions during live endoscopy[92]
Endoscopic imagesDeep learning model with Helicobacter pylori-specific feature extractionHelicobacter pylori diagnosis using AI interpretation of mucosal patternsEnabled real-time assessment of infection risk without biopsy, saving time and cost. Also reduced patient discomfort by avoiding invasive tests[93]
GI tract imagesEnsemble of deep CNN + texture feature classifiersGI abnormality detection from diverse image typesFusion of deep and handcrafted features improved robustness in challenging image conditions. Showed enhanced adaptability across multiple GI lesion types[94]
Capsule endoscopy imagesCNN tailored for parasitic structure detectionHookworm detection in small bowelModel achieved very high detection rate and reduced review time. Critical in endemic regions where hookworm infections are often overlooked[95]
Endoscopic imagesCNN optimized for mucosal pattern recognitionDiagnosis of Helicobacter pylori infection based on endoscopic imagesEnhanced early infection detection through image-only methods. Potentially replaces biopsy in low-resource settings[96]
Endoscopic imagesCNN using hierarchical feature fusionLesion detection in GI endoscopyFeature fusion helped distinguish similar lesions, improving accuracy. Suitable for deployment in routine screening environments[97]
Endoscopic imagesDeep CNN with blind-spot awarenessEarly gastric cancer detection in real-timeThe model avoided blind spots, a common issue in manual diagnosis, reducing miss rate for flat and depressed lesions. Demonstrated strong potential for real-time applications[98]
Capsule endoscopy imagesNeural network with small-bowel angiectasia trainingDetection of GI angiectasiaAchieved accurate identification of angiectasias with high sensitivity. Provided a non-invasive method to detect obscure GI bleeding causes, potentially reducing missed diagnoses during manual video review[99]
Capsule endoscopy imagesCNN-based ulcer detection systemAutomated ulcer detection in small bowelThe CNN model effectively distinguished ulcers from normal mucosa, even in poor lighting or with capsule motion. Enhanced early diagnosis of Crohn’s disease and non-steroidal anti-inflammatory drug-induced ulcers[100]
Endoscopic imagesCNN architecture tailored for Helicobacter pylori featuresEvaluation of Helicobacter pylori infection statusThe model achieved performance comparable to histology-based diagnosis. Useful for non-invasive, in-procedure assessment, aiding in same-visit treatment planning[101]
Endoscopic imagesCNN trained on gastric tumor categoriesGastric neoplasm classification (e.g., adenomas, carcinomas)Outperformed junior endoscopists in differentiating benign vs malignant neoplasms. Accelerated workflow and supported biopsy decision-making[102]
Endoscopic imagesDeep learning classifier with ensemble optimizationGI disease recognition in endoscopic imagesHandled a wide range of disease types with strong classification performance. Showed potential for integration into automated report generation systems[103]
Capsule endoscopy imagesDeep CNN with binary detection of ulcerationDetection of erosions and ulcersDemonstrated over 90% accuracy in detecting clinically relevant ulcerative lesions. Reduced time spent reviewing wireless capsule endoscopy footage by half[104]
Biopsy imagesDeep CNN for histopathological image classificationDetection of diseases (e.g., gastritis, carcinoma) in GI biopsy samplesModel classified histopathological lesions with high fidelity. Useful in screening large volumes of slides, especially in resource-constrained pathology labs[105]
Endoscopic imagesDeep CNN architecture for visual pathology classificationGastric pathology detection (e.g., erosions, polyps, cancer)Showed robust multi-label classification capacity across common gastric pathologies. Aided endoscopists in characterizing mucosal abnormalities with high sensitivity[106]
Barium X-ray radiographyCNN trained on radiograph featuresGastritis detection from double-contrast GI seriesBrought deep learning to a classic imaging modality. Enabled detection of gastritis without the need for endoscopy, useful in areas lacking access to scopes[107]
Endoscopic imagesCNN classifier for esophageal cancerEsophageal cancer diagnosis using endoscopyCNN matched the diagnostic accuracy of experienced clinicians. Rapid lesion classification suggested potential use in screening and early detection programs[108]
Capsule endoscopy imagesDeep learning classifier trained with expert inputSmall-bowel disease and variant classificationDemonstrated gastroenterologist-level accuracy across a wide range of small-bowel findings. Significantly improved efficiency and reliability in interpreting long capsule videos[109]

The reviewed literature in Table 2 highlights the versatility of these models, with applications spanning colonoscopy, gastroscopy, and capsule endoscopy. A key finding is the consistent achievement of high diagnostic accuracy and sensitivity, often surpassing that of generalist endoscopists, particularly for subtle lesions like laterally spreading tumors or small polyps. The table also reveals a trend toward using advanced architectures, including transfer learning from models like ResNet, attention mechanisms, and multi-fusion CNNs to improve performance, generalizability, and interpretability. Furthermore, many studies demonstrate the potential for real-time application and the ability to reduce inter-observer variability, supporting the integration of these AI systems into routine clinical workflows.

While this section has primarily focused on the application of CNNs, it is also important to consider that traditional machine learning methods such as support vector machines, random forests, and gradient boosting machines have also played an important role in the assisted diagnosis of GI diseases. These methods typically rely on manually extracted image features (e.g., texture, color, and shape features), which are then used by a classifier to categorize lesions. For example, some early studies used support vector machines to classify polyps in colonoscopy images or random forests to differentiate esophageal lesions[110,111]. However, compared to CNNs, these traditional methods are less automated in their feature extraction, relying heavily on feature engineering. This limits their performance and generalization capabilities when dealing with large-scale, diverse endoscopic image data. With the rapid development of DL, CNNs have become the dominant approach in this field due to their powerful automatic feature learning capabilities, demonstrating superior performance in most image classification tasks. Nevertheless, in certain specific scenarios, such as when data is limited or when higher model interpretability is required, traditional machine learning methods still hold unique value, and their integration with DL models (e.g., feature fusion) is a promising avenue for future research.

Lesion segmentation and disease severity assessment

Lesion segmentation and disease severity assessment are essential components of GI endoscopic analysis, especially in the management of neoplastic and inflammatory conditions[101,112-114]. While detection tasks aim to localize the presence of lesions, segmentation goes a step further by precisely delineating their boundaries, enabling accurate measurements of lesion size, shape, and extent. CNNs, particularly those based on fully convolutional architectures such as U-Net, SegNet, and DeepLab, have been widely adopted for this purpose in GI endoscopy[101,115-118].

In colorectal endoscopy, CNN-based segmentation models are instrumental for assessing polyps identified during colonoscopy. Accurate segmentation allows for quantifying polyp dimensions, estimating surface area, and distinguishing sessile or flat morphologies, which directly influence clinical decisions regarding resection strategy. In particular, U-Net and its 3D variants have demonstrated excellent performance in segmenting polyps from both 2D frames and volumetric colonoscopy data[119]. Furthermore, models trained with pixel-level annotations enable high-precision margin detection, critical for ensuring complete removal during endoscopic mucosal resection or submucosal dissection. In upper GI endoscopy, CNNs have been developed to segment early gastric cancer, esophageal neoplasms, and gastric ulcers. These models help differentiate subtle cancerous changes from inflamed or atrophic mucosa, which is often visually similar. CNNs have also been adapted to multimodal imaging inputs[120,121], such as combining white-light endoscopy with narrow-band imaging or magnifying chromoendoscopy, further improving segmentation accuracy in complex cases. For IBD, CNN-based models can assess disease severity by segmenting inflamed vs normal mucosa and scoring the presence of ulcerations, bleeding, and edema[99,114,122]. Automated scoring systems trained on standard indices, such as the Mayo endoscopic subscore or the Ulcerative Colitis Endoscopic Index of Severity, can classify disease activity levels with high concordance to expert reviewers[123]. This characteristic allows for objective, reproducible evaluation of mucosal healing, a key target in IBD therapy. Moreover, CNNs have been successfully applied to capsule endoscopy, where it is impractical to conduct a manual review of over 50000 frames per procedure[100,109,124]. Segmentation models in this context automatically identify bleeding sites, erosions, and vascular lesions, dramatically reducing review time. Some systems further incorporate anatomical localization, helping clinicians pinpoint the precise site of pathology within the small bowel. Table 3 Lists studies that used CNNs for lesion segmentation and severity assessment in GI diseases[125-136].

Table 3 Applications of convolutional neural networks for lesion segmentation and disease severity assessment in gastrointestinal diseases.
Data
Architecture
Application
Key findings
Ref.
Double-balloon endoscopy imagesDeep CNN with lesion segmentation and severity scoringAutomated detection and severity grading of Crohn’s ulcers from double-balloon endoscopyThe model accurately segmented ulcers and assessed their severity, enabling objective Crohn’s Disease monitoring. Outperformed traditional scoring methods and showed potential to reduce inter-observer variability[125]
Capsule endoscopy imagesCNN segmentation network trained on angiodysplasiasAutomated segmentation of vascular lesions in small bowel imagesProvided accurate lesion boundary identification for angiodysplasias, facilitating hemorrhage risk stratification. Supports quicker and more consistent diagnosis compared to manual review[126]
Colon capsule endoscopyCNN trained to detect blood and mucosal lesionsSimultaneous detection and segmentation of bleeding and mucosal abnormalitiesEnabled precise localization of lesions and bleeding points in colon capsule footage. Improved lesion coverage and reduced diagnostic delay[127]
Endoscopic imagesCNN trained to grade UC severityAutomated assessment of UC severity from colonoscopyModel achieved expert-level grading accuracy across multiple severity stages. Significantly reduced inter-observer bias, suggesting suitability for clinical trial endpoints[128]
Colonoscopy imagesCNN-based model for pattern recognitionDifferentiation of Crohn’s disease vs UC from colonoscopyThe system accurately distinguished UC and Crohn’s disease patterns, providing real-time decision support for disease type classification[114]
Endoscopic imagesDeep learning classifier for depth predictionPredicting submucosal invasion in gastric neoplasmsModel reliably estimated invasion depth, reducing unnecessary surgical intervention. Useful in pre-treatment risk stratification[129]
Endoscopic imagesCNN for multi-feature predictionPrediction of early gastric cancer, invasion depth, and differentiationAI model outperformed experts in cancer invasion depth and differentiation. Enabled non-invasive yet accurate diagnosis during endoscopy[130]
Capsule endoscopy imagesDeep neural network tailored to stricture detectionDetection of Crohn’s-related intestinal stricturesThe model improved the detection of strictures, which are often missed in manual review. Accelerates diagnosis and may guide therapeutic decisions[131]
Confocal laser endomicroscopy imagesCNN trained for mucosal healing assessmentConfirmation of mucosal healing in Crohn’s diseaseEnabled fine-grained assessment of healing vs inflammation. Provided high-resolution insight for assessing treatment efficacy[132]
Endoscopic imagesAI-assisted classification modelUC disease activity scoring using Mayo classificationProvided consistent and reproducible activity scores. Reduced assessment variability, improving clinical and research utility[133]
Endoscopic imagesCNN vs human graders comparisonGrading UC severity from colonoscopyCNN showed equal or superior performance to human reviewers in grading UC. Suggested for use in high-throughput settings or trials[134]
Capsule endoscopy videosCNN for quantitative feature extractionQuantitative analysis of celiac disease lesionsModel extracted and quantified villous atrophy and mucosal abnormalities. Offered an objective metric to monitor disease progression[135]
Conventional endoscopy imagesCNN trained on histological labelsPredicting invasion depth of gastric cancerCNN-assisted depth prediction provided decision support for therapy planning. Could reduce need for unnecessary biopsies[136]

Based on the studies outlined in Table 3, CNNs are instrumental in advancing from simple lesion detection to precise segmentation and objective disease severity assessment. The reviewed research demonstrates the use of segmentation-focused architectures to delineate lesion boundaries, enabling quantitative measurements of ulcers and vascular lesions. A significant trend is the development of models for automated severity grading of IBDs like ulcerative colitis and Crohn’s disease, with results showing high concordance to expert-level assessments and a reduction in inter-observer bias. This capability is further extended to predicting submucosal invasion in gastric neoplasms, which can aid in pre-treatment stratification and reduce unnecessary procedures. The applications in capsule endoscopy also show that CNNs are effectively used to segment and localize lesions, thereby dramatically reducing the manual review time for clinicians.

Real-time AI assistance and workflow integration

A major innovation in GI endoscopy is the integration of CNNs into real-time clinical workflows. Modern computer-aided detection and diagnosis systems powered by CNNs can process high-resolution endoscopic videos in real time, providing immediate feedback during procedures[133,137,138]. This includes flagging potential lesions, highlighting abnormal regions, and suggesting histological risk levels. These tools support endoscopists by improving the detection rate, particularly for subtle or diminutive lesions, and reducing cognitive fatigue during long procedures. Real-time assistance has been shown to increase the adenoma detection rate, a key quality indicator in colonoscopy, thereby directly contributing to CRC prevention[139,140]. Furthermore, the deployment of real-time CNNs in upper GI procedures, such as esophagogastroduodenoscopy and endoscopic ultrasound-guided interventions, is an increasingly common practice[141]. In endoscopic ultrasound, CNNs aid in identifying anatomical structures, differentiating cystic lesions from the solid ones, and supporting fine-needle aspiration targeting[142,143]. In capsule endoscopy, CNNs enable a rapid triage of massive image volumes, highlighting frames with suspected pathology for physician review[144,145]. This feature has been shown to dramatically reduce interpretation time from hours to minutes while maintaining diagnostic integrity. However, it is crucial to recognize that many studies in this domain are based on relatively small or imbalanced datasets, which can limit the robustness of the reported performance. The development of models that are validated on diverse, multi-center data remains a key challenge. CNNs are also being explored for robotic endoscopy and autonomous navigation systems, where real-time tissue recognition and decision-making will be essential for safe, automated procedures[146,147]. However, while these studies consistently report high performance, many are retrospective and lack prospective validation in real-world clinical workflows. Furthermore, the variability in endoscope imaging settings and lighting conditions can affect the reported results and needs to be addressed in future research. Achieving this requires specific hardware and software considerations. Real-time inference necessitates efficient processing, often leveraging dedicated hardware like graphics processing unit or specialized edge devices integrated directly into the endoscopy suite. Minimizing computational latency is paramount to ensure the AI’s feedback is synchronized with the endoscopist’s movements, thereby preventing workflow disruption and maximizing clinical utility. The field’s translational maturity is evidenced by the emergence of commercially available systems that have received regulatory approvals, such as the United States Food and Drug Administration and the Conformite Europeenne mark in Europe, for real-time polyp detection during colonoscopy. Prominent examples include the GI GeniusTM (Medtronic, MN, United States), CAD EYETM (Fujifilm, Japan), and CADDIETM (Odin Medical/Olympus, PA, United States) systems. Furthermore, professional societies like the European Society of GI Endoscopy and the American Society for GI Endoscopy have issued position statements and guidelines on the expected value and proper clinical integration of AI, which are crucial for defining performance metrics and fostering adoption[148,149]. Table 4 Lists representative studies demonstrating the use of CNN-based systems for real-time AI support and workflow integration in GI endoscopy. The integration of these models into commercial platforms and regulatory-approved devices marks an important step toward clinical translation[150-157].

Table 4 Convolutional neural network-based real-time artificial intelligence support and workflow integration in gastrointestinal endoscopy.
Data
Architecture
Application
Key findings
Ref.
Gastroscopy imagesReal-time anatomical classification with CNNReal-time anatomical recognition during gastroscopy proceduresCNN accurately classified anatomical positions in real-time, reducing mislabeling and improving procedural documentation. Enabled smoother workflow integration with minimal latency[150]
Gastroscopy videosFrame-wise CNN classificationAutomated disease detection during endoscopy video reviewThe system provided real-time frame classification, improving detection of GI abnormalities in long video sequences and assisting in efficient case triage[151]
White-light endoscopyCNN for real-time Helicobacter pylori statusImmediate assessment of Helicobacter pylori infection during endoscopyReal-time feedback from the CNN allowed on-the-spot therapeutic decision-making and eliminated delays caused by biopsy processing[113]
Esophagogastroduodenoscopy videosCNN-based detection system integrated with endoscopeReal-time detection of gastroesophageal varicesThe system was validated across multiple centers and significantly reduced miss rates of varices in real-time, enhancing early intervention[152]
Colonoscopic imagesDeep CNN classifier for anatomical site recognitionAutomated real-time anatomical classification of colon imagesCNN correctly labeled anatomical sites, reducing reliance on user memory and ensuring consistent documentation. Improved novice performance[153]
Upper GI endoscopyCNN detection assistant in randomised controlled trialDetection of gastric neoplasms during routine proceduresA deep learning-based system reduced miss rate of gastric neoplasms significantly in a randomized controlled trial. Validated real-world clinical benefit[154]
Upper GI anatomy imagesMulti-task CNN modelSimultaneous detection of anatomical landmarks and structuresModel performed both classification and segmentation, improving navigation and assisting less experienced users in orientation[155]
Esophagogastroduodenoscopy imagesCNN trained on large anatomical datasetReal-time classification of esophagogastroduodenoscopy anatomyCNN classified images with expert-level accuracy, aiding report generation and training in real-time[141]
GI endoscopy videosCNN vs global feature comparisonReal-time disease detection efficiency benchmarkingCNNs proved significantly more accurate than traditional global features. Demonstrated readiness for real-time clinical deployment[156]
GI tract videosCNN with automated reporting moduleEnd-to-end disease detection and report generationAutomatically generated reports based on real-time CNN classification, reducing documentation time and enhancing standardization[157]

The studies presented in Table 4 demonstrate a clear trend toward developing and validating systems for immediate application during procedures, encompassing tasks from real-time anatomical classification and recognition to automated disease detection and a real-time assessment of Helicobacter pylori status. A significant finding across these studies is the consistent improvement in clinical efficiency, including reduced documentation time, enhanced detection rates for subtle abnormalities, and improved procedural documentation. The table also underscores the expanding application of these real-time systems to various procedures, such as gastroscopy and esophagogastroduodenoscopy, marking a key step toward the practical, real-world clinical translation of CNN-based tools.

CHALLENGES AND PERSPECTIVES

In our view, a major barrier to the clinical implementation of CNNs in GI endoscopy lies in the limitations related to data and model performance[158,159]. High-quality, annotated endoscopic image datasets are essential for training effective CNN models, yet they remain scarce due to data privacy concerns, the high cost of expert labeling, and inconsistencies across institutions. As a result, models are often trained on single-center datasets and may not generalize well to broader patient populations or different equipment settings. Moreover, variations in image quality, lighting conditions, bowel preparation, and endoscopic device types further challenge the robustness of these models in real-world applications[159,160]. Even when trained on adequate datasets, CNNs may struggle to adapt to new clinical environments without careful domain adaptation and validation across diverse settings. To address this, techniques like domain adaptation and federated learning are being explored. Domain adaptation helps models perform on new data distributions, while federated learning enables collaborative training across multiple institutions without sharing sensitive patient data, thus improving generalizability while protecting privacy.

Another critical issue is the interpretability of CNN predictions[161,162]. This is because, despite these models matching or exceeding human-level performance, their “black-box” nature makes it difficult for clinicians to understand or trust the decision-making process. This suggests that the absence of explainable outputs may reduce user confidence and hinder clinical adoption. To enhance trust and clinical adoption, methods for explainable AI are crucial. Techniques such as Grad-CAM, layer-wise relevance propagation, and SHapley Additive exPlanations can provide visual or numerical explanations for model decisions, making the AI’s reasoning more transparent to clinicians. Additionally, the practical integration of CNNs into real-time endoscopic workflows presents technical hurdles, including computational latency, interface design, and the risk of workflow disruption. To overcome these barriers, future research must focus on multi-center data collaboration, the development of explainable AI techniques, and close interdisciplinary cooperation to ensure that CNN-based tools are not only accurate but also usable and trustworthy in clinical settings.

In recent years, Transformer-based models, originally developed for natural language processing, have also gained significant attention in medical image analysis. Architectures like the vision transformer[163] and its medical variants (e.g., MedT) have shown promising results in both classification and segmentation tasks. Unlike CNNs that rely on local convolutional operations, transformers utilize a self-attention mechanism to capture long-range dependencies across the entire image[164]. This capability makes them particularly well-suited for tasks where global context is crucial. While their application in GI endoscopy is still emerging, the performance of models like TransUNet, which combines a U-Net-like structure with a transformer encoder, demonstrates their potential to address some of the limitations of pure CNN models[165]. Future research is expected to further explore the integration of Transformer-based models to enhance the accuracy and robustness of AI-assisted diagnosis in GI endoscopy.

CONCLUSION

CNNs have demonstrated significant potential to advance GI endoscopy by enhancing the accuracy and efficiency of lesion detection and classification. These models can support clinicians by reducing diagnostic errors and improving the consistency of endoscopic assessments. However, several barriers hinder their widespread clinical adoption, including limited and imbalanced datasets, poor generalizability across clinical settings, lack of model interpretability, and technical issues in real-time deployment. Ethical and regulatory concerns, such as data privacy and bias, also require attention. To address these challenges, future research should focus on developing explainable and robust models, conducting external and prospective validations, and fostering collaborations across institutions. Standardized reporting and alignment with clinical needs will be crucial for translating CNN-based systems into practice. While CNNs are not yet fully integrated into clinical GI workflows, they represent a promising tool that could transform endoscopic diagnosis and patient care. With continued improvements in hardware and enhanced interpretability, CNN-based assistance is likely to become a standard component of next-generation intelligent endoscopy systems.

Footnotes

Provenance and peer review: Invited article; Externally peer reviewed.

Peer-review model: Single blind

Specialty type: Gastroenterology and hepatology

Country of origin: China

Peer-review report’s classification

Scientific Quality: Grade A, Grade A, Grade A, Grade B, Grade B

Novelty: Grade B, Grade B, Grade B, Grade B, Grade B

Creativity or Innovation: Grade B, Grade B, Grade B, Grade B, Grade B

Scientific Significance: Grade A, Grade B, Grade B, Grade B, Grade B

P-Reviewer: Hany M, MD, PhD, Professor, Egypt; Ou JR, MD, China; Slimi H, PhD, Associate Professor, Tunisia S-Editor: Wu S L-Editor: A P-Editor: Lei YY

References
1.  Rogler G, Singh A, Kavanaugh A, Rubin DT. Extraintestinal Manifestations of Inflammatory Bowel Disease: Current Concepts, Treatment, and Implications for Disease Management. Gastroenterology. 2021;161:1118-1132.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 86]  [Cited by in RCA: 464]  [Article Influence: 116.0]  [Reference Citation Analysis (0)]
2.  Wang Y, Huang Y, Chase RC, Li T, Ramai D, Li S, Huang X, Antwi SO, Keaveny AP, Pang M. Global Burden of Digestive Diseases: A Systematic Analysis of the Global Burden of Diseases Study, 1990 to 2019. Gastroenterology. 2023;165:773-783.e15.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 83]  [Cited by in RCA: 75]  [Article Influence: 37.5]  [Reference Citation Analysis (0)]
3.  Huang RJ, Hwang JH. Improving the Early Diagnosis of Gastric Cancer. Gastrointest Endosc Clin N Am. 2021;31:503-517.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 16]  [Cited by in RCA: 26]  [Article Influence: 6.5]  [Reference Citation Analysis (0)]
4.  Spiceland CM, Lodhia N. Endoscopy in inflammatory bowel disease: Role in diagnosis, management, and treatment. World J Gastroenterol. 2018;24:4014-4020.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in CrossRef: 46]  [Cited by in RCA: 82]  [Article Influence: 11.7]  [Reference Citation Analysis (0)]
5.  Moran CP, Neary B, Doherty GA. Endoscopic evaluation in diagnosis and management of inflammatory bowel disease. World J Gastrointest Endosc. 2016;8:723-732.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in CrossRef: 14]  [Cited by in RCA: 17]  [Article Influence: 1.9]  [Reference Citation Analysis (1)]
6.  Kim SH, Chun HJ. Capsule Endoscopy: Pitfalls and Approaches to Overcome. Diagnostics (Basel). 2021;11:1765.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 1]  [Cited by in RCA: 14]  [Article Influence: 3.5]  [Reference Citation Analysis (0)]
7.  Ferrari C, Tadros M. Enhancing the Quality of Upper Gastrointestinal Endoscopy: Current Indicators and Future Trends. Gastroenterol Insights. 2024;15:1-18.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 2]  [Cited by in RCA: 2]  [Article Influence: 1.0]  [Reference Citation Analysis (0)]
8.  Gimeno-García AZ, Hernández-Pérez A, Nicolás-Pérez D, Hernández-Guerra M. Artificial Intelligence Applied to Colonoscopy: Is It Time to Take a Step Forward? Cancers (Basel). 2023;15:2193.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 14]  [Reference Citation Analysis (0)]
9.  Ali H, Muzammil MA, Dahiya DS, Ali F, Yasin S, Hanif W, Gangwani MK, Aziz M, Khalaf M, Basuli D, Al-Haddad M. Artificial intelligence in gastrointestinal endoscopy: a comprehensive review. Ann Gastroenterol. 2024;37:133-141.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 4]  [Cited by in RCA: 3]  [Article Influence: 3.0]  [Reference Citation Analysis (0)]
10.  Lewis JR, Pathan S, Kumar P, Dias CC. AI in Endoscopic Gastrointestinal Diagnosis: A Systematic Review of Deep Learning and Machine Learning Techniques. IEEE Access. 2024;12:163764-163786.  [PubMed]  [DOI]  [Full Text]
11.  Mohan BP, Khan SR, Kassab LL, Ponnada S, Chandan S, Ali T, Dulai PS, Adler DG, Kochhar GS. High pooled performance of convolutional neural networks in computer-aided diagnosis of GI ulcers and/or hemorrhage on wireless capsule endoscopy images: a systematic review and meta-analysis. Gastrointest Endosc. 2021;93:356-364.e4.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 33]  [Cited by in RCA: 26]  [Article Influence: 6.5]  [Reference Citation Analysis (0)]
12.  Choi J, Shin K, Jung J, Bae HJ, Kim DH, Byeon JS, Kim N. Convolutional Neural Network Technology in Endoscopic Imaging: Artificial Intelligence for Endoscopy. Clin Endosc. 2020;53:117-126.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 45]  [Cited by in RCA: 35]  [Article Influence: 7.0]  [Reference Citation Analysis (1)]
13.  Ma H, Wang L, Chen Y, Tian L. Convolutional neural network-based artificial intelligence for the diagnosis of early esophageal cancer based on endoscopic images: A meta-analysis. Saudi J Gastroenterol. 2022;28:332-340.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in RCA: 2]  [Reference Citation Analysis (0)]
14.  Yan T, Wong PK, Qin YY. Deep learning for diagnosis of precancerous lesions in upper gastrointestinal endoscopy: A review. World J Gastroenterol. 2021;27:2531-2544.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in CrossRef: 18]  [Cited by in RCA: 7]  [Article Influence: 1.8]  [Reference Citation Analysis (0)]
15.  Zhao Y, Hu B, Wang Y, Yin X, Jiang Y, Zhu X. Identification of gastric cancer with convolutional neural networks: a systematic review. Multimed Tools Appl. 2022;81:11717-11736.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 7]  [Cited by in RCA: 10]  [Article Influence: 3.3]  [Reference Citation Analysis (0)]
16.  Weigt J, Repici A, Antonelli G, Afifi A, Kliegis L, Correale L, Hassan C, Neumann H. Performance of a new integrated computer-assisted system (CADe/CADx) for detection and characterization of colorectal neoplasia. Endoscopy. 2022;54:180-184.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 22]  [Cited by in RCA: 53]  [Article Influence: 17.7]  [Reference Citation Analysis (0)]
17.  Attallah O, Sharkas M. GASTRO-CADx: a three stages framework for diagnosing gastrointestinal diseases. PeerJ Comput Sci. 2021;7:e423.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 26]  [Cited by in RCA: 23]  [Article Influence: 5.8]  [Reference Citation Analysis (0)]
18.  Raju ASN, Venkatesh K, Gatla RK, Eid MM, Flah A, Slanina Z, Ghaly RNR. Expedited Colorectal Cancer Detection Through a Dexterous Hybrid CADx System With Enhanced Image Processing and Augmented Polyp Visualization. IEEE Access. 2025;13:17524-17553.  [PubMed]  [DOI]  [Full Text]
19.  Ali H, Sharif M, Yasmin M, Rehmani MH, Riaz F. A survey of feature extraction and fusion of deep learning for detection of abnormalities in video endoscopy of gastrointestinal-tract. Artif Intell Rev. 2020;53:2635-2707.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 14]  [Cited by in RCA: 16]  [Article Influence: 2.7]  [Reference Citation Analysis (0)]
20.  Mu G, Zhu Y, Niu Z, Li H, Wu L, Wang J, Luo R, Hu X, Li Y, Zhang J, Hu S, Li C, Ding S, Yu H. Expert-level classification of gastritis by endoscopy using deep learning: a multicenter diagnostic trial. Endosc Int Open. 2021;9:E955-E964.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 2]  [Cited by in RCA: 11]  [Article Influence: 2.8]  [Reference Citation Analysis (0)]
21.  Li M, Jiang Y, Zhang Y, Zhu H. Medical image analysis using deep learning algorithms. Front Public Health. 2023;11:1273253.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 5]  [Cited by in RCA: 59]  [Article Influence: 29.5]  [Reference Citation Analysis (0)]
22.  Escobar J, Sanchez K, Hinojosa C, Arguello H, Castillo S.   Accurate Deep Learning-based Gastrointestinal Disease Classification via Transfer Learning Strategy. In: 2021 XXIII Symposium on Image, Signal Processing and Artificial Vision (STSIVA); 2021 Sept 15-17; Popayán, Colombia. IEEE: NY, United States, 2021: 1-5.  [PubMed]  [DOI]  [Full Text]
23.  Su Q, Wang F, Chen D, Chen G, Li C, Wei L. Deep convolutional neural networks with ensemble learning and transfer learning for automated detection of gastrointestinal diseases. Comput Biol Med. 2022;150:106054.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 34]  [Cited by in RCA: 25]  [Article Influence: 8.3]  [Reference Citation Analysis (0)]
24.  Hicks SA, Strümke I, Thambawita V, Hammou M, Riegler MA, Halvorsen P, Parasa S. On evaluation metrics for medical applications of artificial intelligence. Sci Rep. 2022;12:5979.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 8]  [Cited by in RCA: 276]  [Article Influence: 92.0]  [Reference Citation Analysis (0)]
25.  Kaur G, Saini S. Comparative analysis of RMSE and MAP metrics for evaluating CNN and LSTM models. AIP Conf Proc. 2024;3121:040003.  [PubMed]  [DOI]  [Full Text]
26.  Listyalina L, Mustiadi I, Dharmawan DA.   Joint Dice and Intersection over Union Losses for Deep Optical Disc Segmentation. In: 2020 3rd International Conference on Biomedical Engineering (IBIOMED); 2020 Oct 6-8; Yogyakarta, Indonesia. IEEE: NY, United States, 2020: 49-54.  [PubMed]  [DOI]  [Full Text]
27.  Rahman MA, Wang Y.   Optimizing Intersection-Over-Union in Deep Neural Networks for Image Segmentation. In: Bebis G, Boyle R, Parvin B, Koracin D, Porikli F, Skaff S, Entezari A, Min J, Iwai D, Sadagic A, Scheidegger C, Isenberg T. Advances in Visual Computing. Switzerland: Springer, 2016: 234-244.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 205]  [Cited by in RCA: 102]  [Article Influence: 11.3]  [Reference Citation Analysis (0)]
28.  Bradshaw TJ, Huemann Z, Hu J, Rahmim A. A Guide to Cross-Validation for Artificial Intelligence in Medical Imaging. Radiol Artif Intell. 2023;5:e220232.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 1]  [Cited by in RCA: 58]  [Article Influence: 29.0]  [Reference Citation Analysis (0)]
29.  Sejuti ZA, Islam MS. A hybrid CNN-KNN approach for identification of COVID-19 with 5-fold cross validation. Sens Int. 2023;4:100229.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 12]  [Reference Citation Analysis (0)]
30.  Smedsrud PH, Thambawita V, Hicks SA, Gjestang H, Nedrejord OO, Næss E, Borgli H, Jha D, Berstad TJD, Eskeland SL, Lux M, Espeland H, Petlund A, Nguyen DTD, Garcia-Ceja E, Johansen D, Schmidt PT, Toth E, Hammer HL, de Lange T, Riegler MA, Halvorsen P. Kvasir-Capsule, a video capsule endoscopy dataset. Sci Data. 2021;8:142.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 157]  [Cited by in RCA: 53]  [Article Influence: 13.3]  [Reference Citation Analysis (1)]
31.  Jaspers TJM, Boers TGW, Kusters CHJ, Jong MR, Jukema JB, de Groof AJ, Bergman JJ, de With PHN, van der Sommen F. Robustness evaluation of deep neural networks for endoscopic image analysis: Insights and strategies. Med Image Anal. 2024;94:103157.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
32.  Ibrahim H, Liu X, Rivera SC, Moher D, Chan AW, Sydes MR, Calvert MJ, Denniston AK. Reporting guidelines for clinical trials of artificial intelligence interventions: the SPIRIT-AI and CONSORT-AI guidelines. Trials. 2021;22:11.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 14]  [Cited by in RCA: 56]  [Article Influence: 14.0]  [Reference Citation Analysis (0)]
33.  Taye MM. Theoretical Understanding of Convolutional Neural Network: Concepts, Architectures, Applications, Future Directions. Computation. 2023;11:52.  [PubMed]  [DOI]  [Full Text]
34.  Jia F, Lei Y, Lin J, Zhou X, Lu N. Deep neural networks: A promising tool for fault characteristic mining and intelligent diagnosis of rotating machinery with massive data. Mech Syst Signal Process. 2016;72-73:303-315.  [PubMed]  [DOI]  [Full Text]
35.  Shin HC, Roth HR, Gao M, Lu L, Xu Z, Nogues I, Yao J, Mollura D, Summers RM. Deep Convolutional Neural Networks for Computer-Aided Detection: CNN Architectures, Dataset Characteristics and Transfer Learning. IEEE Trans Med Imaging. 2016;35:1285-1298.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 3594]  [Cited by in RCA: 1928]  [Article Influence: 214.2]  [Reference Citation Analysis (0)]
36.  Du W, Rao N, Liu D, Jiang H, Luo C, Li Z, Gan T, Zeng B. Review on the Applications of Deep Learning in the Analysis of Gastrointestinal Endoscopy Images. IEEE Access. 2019;7:142053-142069.  [PubMed]  [DOI]  [Full Text]
37.  Lea C, Flynn MD, Vidal R, Reiter A, Hager GD.   Temporal Convolutional Networks for Action Segmentation and Detection. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2017 Jul 21-26; Honolulu, HI, United States. IEEE: NY, United States, 2017: 1003-1012.  [PubMed]  [DOI]  [Full Text]
38.  Tan J, Gao Y, Liang Z, Cao W, Pomeroy MJ, Huo Y, Li L, Barish MA, Abbasi AF, Pickhardt PJ. 3D-GLCM CNN: A 3-Dimensional Gray-Level Co-Occurrence Matrix-Based CNN Model for Polyp Classification via CT Colonography. IEEE Trans Med Imaging. 2020;39:2013-2024.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 75]  [Cited by in RCA: 50]  [Article Influence: 10.0]  [Reference Citation Analysis (0)]
39.  Igarashi S, Sasaki Y, Mikami T, Sakuraba H, Fukuda S. Anatomical classification of upper gastrointestinal organs under various image capture conditions using AlexNet. Comput Biol Med. 2020;124:103950.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 27]  [Cited by in RCA: 45]  [Article Influence: 9.0]  [Reference Citation Analysis (0)]
40.  Nouman Noor M, Nazir M, Khan SA, Song O, Ashraf I. Efficient Gastrointestinal Disease Classification Using Pretrained Deep Convolutional Neural Network. Electronics. 2023;12:1557.  [PubMed]  [DOI]  [Full Text]
41.  Yang J, Ou Y, Chen Z, Liao J, Sun W, Luo Y, Luo C. A Benchmark Dataset of Endoscopic Images and Novel Deep Learning Method to Detect Intestinal Metaplasia and Gastritis Atrophy. IEEE J Biomed Health Inform. 2023;27:7-16.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 14]  [Reference Citation Analysis (0)]
42.  Ye B, Shu Z, Wang B, Wang S, Fu Y, Zhang L, Qi B, Dwivedi AK, Liu S. Attention Mechanism Guided SE + ResNet-H Model for Gastrointestinal Endoscopy Image Classification. IEEE Trans Instrum Meas. 2024;73:1-13.  [PubMed]  [DOI]  [Full Text]
43.  Beeche C, Singh JP, Leader JK, Gezer S, Oruwari AP, Dansingani KK, Chhablani J, Pu J. Super U-Net: a modularized generalizable architecture. Pattern Recognit. 2022;128:108669.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 28]  [Cited by in RCA: 16]  [Article Influence: 5.3]  [Reference Citation Analysis (0)]
44.  Souaidi M, Lafraxo S, Kerkaou Z, El Ansari M, Koutti L. A Multiscale Polyp Detection Approach for GI Tract Images Based on Improved DenseNet and Single-Shot Multibox Detector. Diagnostics (Basel). 2023;13:733.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 4]  [Reference Citation Analysis (0)]
45.  Petit O, Thome N, Rambour C, Themyr L, Collins T, Soler L.   U-Net Transformer: Self and Cross Attention for Medical Image Segmentation. In: Machine Learning in Medical Imaging: 12th International Workshop, MLMI 2021, Held in Conjunction with MICCAI 2021, 2021 Sept 27; Strasbourg, France. Berlin: Springer, 2021: 267-276.  [PubMed]  [DOI]  [Full Text]
46.  Vezakis IA, Georgas K, Fotiadis D, Matsopoulos GK. EffiSegNet: Gastrointestinal Polyp Segmentation through a Pre-Trained EfficientNet-based Network with a Simplified Decoder. Annu Int Conf IEEE Eng Med Biol Soc. 2024;2024:1-4.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 4]  [Cited by in RCA: 2]  [Article Influence: 2.0]  [Reference Citation Analysis (0)]
47.  Zhang X, Yang S, Luo W, Gao L, Zhang W.   Video Compression Artifact Reduction by Fusing Motion Compensation and Global Context in a Swin-CNN Based Parallel Architecture. In: Thirty-Seventh AAAI Conference on Artificial Intelligence, AAAI 2023, 2023 Feb 7-14; Washington, DC, United States. AAAI Press: CA, United States, 2023: 3489-3497.  [PubMed]  [DOI]  [Full Text]
48.  Ji Z, Sun H, Yuan N, Zhang H, Sheng J, Zhang X, Ganchev I. BGRD-TransUNet: A Novel TransUNet-Based Model for Ultrasound Breast Lesion Segmentation. IEEE Access. 2024;12:31182-31196.  [PubMed]  [DOI]  [Full Text]
49.  Wagner SJ, Reisenbüchler D, West NP, Niehues JM, Zhu J, Foersch S, Veldhuizen GP, Quirke P, Grabsch HI, van den Brandt PA, Hutchins GGA, Richman SD, Yuan T, Langer R, Jenniskens JCA, Offermans K, Mueller W, Gray R, Gruber SB, Greenson JK, Rennert G, Bonner JD, Schmolze D, Jonnagaddala J, Hawkins NJ, Ward RL, Morton D, Seymour M, Magill L, Nowak M, Hay J, Koelzer VH, Church DN; TransSCOT consortium, Matek C, Geppert C, Peng C, Zhi C, Ouyang X, James JA, Loughrey MB, Salto-Tellez M, Brenner H, Hoffmeister M, Truhn D, Schnabel JA, Boxberg M, Peng T, Kather JN. Transformer-based biomarker prediction from colorectal cancer histology: A large-scale multicentric study. Cancer Cell. 2023;41:1650-1661.e4.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 48]  [Cited by in RCA: 99]  [Article Influence: 49.5]  [Reference Citation Analysis (0)]
50.  Kaur J, Kaur P. A systematic literature analysis of multi-organ cancer diagnosis using deep learning techniques. Comput Biol Med. 2024;179:108910.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 3]  [Reference Citation Analysis (0)]
51.  Bhardwaj P, Kumar S, Kumar Y. A Comprehensive Analysis of Deep Learning-Based Approaches for the Prediction of Gastrointestinal Diseases Using Multi-class Endoscopy Images. Arch Computat Methods Eng. 2023;30:4499-4516.  [PubMed]  [DOI]  [Full Text]
52.  Patil RM, Giripunje S.   Deep Learning-Based Detection and Classification of Gastrointestinal Tract Diseases in Endoscopy Images. In: 2024 2nd DMIHER International Conference on Artificial Intelligence in Healthcare, Education and Industry (IDICAIEI); 2024 Aug 16; Wardha, India. IEEE: NY, United States, 2024: 1-6.  [PubMed]  [DOI]  [Full Text]
53.  Bajhaiya D, Narayanan Unni S. Deep learning-enabled detection and localization of gastrointestinal diseases using wireless-capsule endoscopic images. Biomed Signal Process Control. 2024;93:106125.  [PubMed]  [DOI]  [Full Text]
54.  Durak S, Bayram B, Bakırman T, Erkut M, Doğan M, Gürtürk M, Akpınar B. Deep neural network approaches for detecting gastric polyps in endoscopic images. Med Biol Eng Comput. 2021;59:1563-1574.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 1]  [Cited by in RCA: 8]  [Article Influence: 2.0]  [Reference Citation Analysis (0)]
55.  Ozawa T, Ishihara S, Fujishiro M, Kumagai Y, Shichijo S, Tada T. Automated endoscopic detection and classification of colorectal polyps using convolutional neural networks. Therap Adv Gastroenterol. 2020;13:1756284820910659.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 105]  [Cited by in RCA: 76]  [Article Influence: 15.2]  [Reference Citation Analysis (0)]
56.  Malik H, Naeem A, Sadeghi-niaraki A, Naqvi RA, Lee S. Multi-classification deep learning models for detection of ulcerative colitis, polyps, and dyed-lifted polyps using wireless capsule endoscopy images. Complex Intell Syst. 2024;10:2477-2497.  [PubMed]  [DOI]  [Full Text]
57.  Gadi SR, Muralidharan SS, Glissen Brown JR. Colonoscopy Quality, Innovation, and the Assessment of New Technology. Tech Innov Gastrointest Endosc. 2024;26:177-192.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 2]  [Reference Citation Analysis (0)]
58.  Du W, Rao N, Dong C, Wang Y, Hu D, Zhu L, Zeng B, Gan T. Automatic classification of esophageal disease in gastroscopic images using an efficient channel attention deep dense convolutional neural network. Biomed Opt Express. 2021;12:3066-3081.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 30]  [Cited by in RCA: 21]  [Article Influence: 5.3]  [Reference Citation Analysis (0)]
59.  Eelbode T, Hassan C, Demedts I, Roelandt P, Coron E, Bhandari P, Neumann H, Pech O, Repici A, Maes F, Bisschops R. Tu1959 Bli and Lci Improve Polyp Detection and Delineation Accuracy for Deep Learning Networks. Gastrointest Endosc. 2019;89:AB632.  [PubMed]  [DOI]  [Full Text]
60.  Fonollà R, E. W. van der Zander Q, Schreuder RM, Masclee AAM, Schoon EJ, van der Sommen F, de With PHN. A CNN CADx System for Multimodal Classification of Colorectal Polyps Combining WL, BLI, and LCI Modalities. Appl Sci. 2020;10:5040.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 9]  [Cited by in RCA: 11]  [Article Influence: 2.2]  [Reference Citation Analysis (0)]
61.  Zhou P, Cao Y, Li M, Ma Y, Chen C, Gan X, Wu J, Lv X, Chen C. HCCANet: histopathological image grading of colorectal cancer using CNN based on multichannel fusion attention mechanism. Sci Rep. 2022;12:15103.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 1]  [Cited by in RCA: 18]  [Article Influence: 6.0]  [Reference Citation Analysis (0)]
62.  Lonseko ZM, Adjei PE, Du W, Luo C, Hu D, Zhu L, Gan T, Rao N. Gastrointestinal Disease Classification in Endoscopic Images Using Attention-Guided Convolutional Neural Networks. Appl Sci. 2021;11:11136.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 4]  [Cited by in RCA: 12]  [Article Influence: 3.0]  [Reference Citation Analysis (0)]
63.  Zhao J, Feng Q, Wu P, Lupu RA, Wilke RA, Wells QS, Denny JC, Wei WQ. Learning from Longitudinal Data in Electronic Health Record and Genetic Data to Improve Cardiovascular Event Prediction. Sci Rep. 2019;9:717.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 108]  [Cited by in RCA: 104]  [Article Influence: 17.3]  [Reference Citation Analysis (0)]
64.  Lin Y, Zhang X, Li F, Zhang R, Jiang H, Lai C, Yi L, Li Z, Wu W, Qiu L, Yang H, Guan Q, Wang Z, Deng L, Zhao Z, Lu W, Lun W, Dai J, He S, Bai Y. A deep neural network improves endoscopic detection of laterally spreading tumors. Surg Endosc. 2025;39:776-785.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 1]  [Reference Citation Analysis (0)]
65.  Uma R, Ganesh K, Sudheer P, Muthu Kumar M.   A Deep Learning Based Detection and Classification of Gastrointestinal Diseases. In: 2024 6th International Conference on Electrical, Control and Instrumentation Engineering (ICECIE); 2024 Nov 23-23; Pattaya, Thailand. IEEE: NY, United States, 2024: 1-7.  [PubMed]  [DOI]  [Full Text]
66.  El-ghany SA, Mahmood MA, Abd El-aziz AA. An Accurate Deep Learning-Based Computer-Aided Diagnosis System for Gastrointestinal Disease Detection Using Wireless Capsule Endoscopy Image Analysis. Appl Sci. 2024;14:10243.  [PubMed]  [DOI]  [Full Text]
67.  Zhou B, Rao X, Xing H, Ma Y, Wang F, Rong L. A convolutional neural network-based system for detecting early gastric cancer in white-light endoscopy. Scand J Gastroenterol. 2023;58:157-162.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 6]  [Reference Citation Analysis (0)]
68.  Lee CK, Hofer I, Gabel E, Baldi P, Cannesson M. Development and Validation of a Deep Neural Network Model for Prediction of Postoperative In-hospital Mortality. Anesthesiology. 2018;129:649-662.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 91]  [Cited by in RCA: 129]  [Article Influence: 21.5]  [Reference Citation Analysis (0)]
69.  Gan T, Yang Y, Liu S, Zeng B, Yang J, Deng K, Wu J, Yang L. Automatic Detection of Small Intestinal Hookworms in Capsule Endoscopy Images Based on a Convolutional Neural Network. Gastroenterol Res Pract. 2021;2021:5682288.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 4]  [Cited by in RCA: 7]  [Article Influence: 1.8]  [Reference Citation Analysis (0)]
70.  Mascarenhas Saraiva MJ, Afonso J, Ribeiro T, Ferreira J, Cardoso H, Andrade AP, Parente M, Natal R, Mascarenhas Saraiva M, Macedo G. Deep learning and capsule endoscopy: automatic identification and differentiation of small bowel lesions with distinct haemorrhagic potential using a convolutional neural network. BMJ Open Gastroenterol. 2021;8:e000753.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 3]  [Cited by in RCA: 25]  [Article Influence: 6.3]  [Reference Citation Analysis (0)]
71.  Zhang M, Zhu C, Wang Y, Kong Z, Hua Y, Zhang W, Si X, Ye B, Xu X, Li L, Heng D, Liu B, Tian S, Wu J, Dang Y, Zhang G. Differential diagnosis for esophageal protruded lesions using a deep convolution neural network in endoscopic images. Gastrointest Endosc. 2021;93:1261-1272.e2.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 3]  [Cited by in RCA: 7]  [Article Influence: 1.8]  [Reference Citation Analysis (0)]
72.  Sakai Y, Takemoto S, Hori K, Nishimura M, Ikematsu H, Yano T, Yokota H. Automatic detection of early gastric cancer in endoscopic images using a transferring convolutional neural network. Annu Int Conf IEEE Eng Med Biol Soc. 2018;2018:4138-4141.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 39]  [Cited by in RCA: 48]  [Article Influence: 8.0]  [Reference Citation Analysis (0)]
73.  Asif S, Qurrat-ul-ain. A Fuzzy Minkowski Distance-based fusion of convolutional neural networks for gastrointestinal disease detection. Appl Soft Comput. 2024;158:111595.  [PubMed]  [DOI]  [Full Text]
74.  Zhu Y, Lyu X, Tao X, Wu L, Yin A, Liao F, Hu S, Wang Y, Zhang M, Huang L, Wang J, Zhang C, Gong D, Jiang X, Zhao L, Yu H. A newly developed deep learning-based system for automatic detection and classification of small bowel lesions during double-balloon enteroscopy examination. BMC Gastroenterol. 2024;24:10.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
75.  Patel V, Patel K, Goel P, Shah M.   Classification of Gastrointestinal Diseases from Endoscopic Images Using Convolutional Neural Network with Transfer Learning. In: 2024 5th International Conference on Intelligent Communication Technologies and Virtual Mobile Networks (ICICV); 2024 Mar 11-12; Tirunelveli, India. IEEE: NY, United States, 2024: 504-508.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 2]  [Reference Citation Analysis (0)]
76.  Khan ZF, Ramzan M, Raza M, Khan MA, Iqbal K, Kim T, Cha J. Deep Convolutional Neural Networks for Accurate Classification of Gastrointestinal Tract Syndromes. Comput Mater Continua. 2024;78:1207-1225.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 4]  [Reference Citation Analysis (0)]
77.  Hossain T, Shamrat FMJM, Zhou X, Mahmud I, Mazumder MSA, Sharmin S, Gururajan R. Development of a multi-fusion convolutional neural network (MF-CNN) for enhanced gastrointestinal disease diagnosis in endoscopy image analysis. PeerJ Comput Sci. 2024;10:e1950.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
78.  Shi Y, Wei N, Wang K, Wu J, Tao T, Li N, Lv B. Deep learning-assisted diagnosis of chronic atrophic gastritis in endoscopy. Front Oncol. 2023;13:1122247.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 4]  [Reference Citation Analysis (0)]
79.  Sharma A, Kumar R, Garg P. Deep learning-based prediction model for diagnosing gastrointestinal diseases using endoscopy images. Int J Med Inform. 2023;177:105142.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 21]  [Reference Citation Analysis (0)]
80.  Scheppach MW, Rauber D, Stallhofer J, Muzalyova A, Otten V, Manzeneder C, Schwamberger T, Wanzl J, Schlottmann J, Tadic V, Probst A, Schnoy E, Römmele C, Fleischmann C, Meinikheim M, Miller S, Märkl B, Stallmach A, Palm C, Messmann H, Ebigbo A. Detection of duodenal villous atrophy on endoscopic images using a deep learning algorithm. Gastrointest Endosc. 2023;97:911-916.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 9]  [Reference Citation Analysis (0)]
81.  Ma L, Su X, Ma L, Gao X, Sun M. Deep learning for classification and localization of early gastric cancer in endoscopic images. Biomed Signal Process Control. 2023;79:104200.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
82.  Zhao Q, Chi T. Deep learning model can improve the diagnosis rate of endoscopic chronic atrophic gastritis: a prospective cohort study. BMC Gastroenterol. 2022;22:133.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 1]  [Cited by in RCA: 15]  [Article Influence: 5.0]  [Reference Citation Analysis (0)]
83.  Nam JY, Chung HJ, Choi KS, Lee H, Kim TJ, Soh H, Kang EA, Cho SJ, Ye JC, Im JP, Kim SG, Kim JS, Chung H, Lee JH. Deep learning model for diagnosing gastric mucosal lesions using endoscopic images: development, validation, and method comparison. Gastrointest Endosc. 2022;95:258-268.e10.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 6]  [Cited by in RCA: 24]  [Article Influence: 8.0]  [Reference Citation Analysis (0)]
84.  Hmoud Al-Adhaileh M, Mohammed Senan E, Alsaade FW, Aldhyani THH, Alsharif N, Abdullah Alqarni A, Uddin MI, Alzahrani MY, Alzain ED, Jadhav ME. Deep Learning Algorithms for Detection and Classification of Gastrointestinal Diseases. Complexity. 2021;2021:6170416.  [PubMed]  [DOI]  [Full Text]
85.  Naz J, Sharif M, Yasmin M, Raza M, Khan MA. Detection and Classification of Gastrointestinal Diseases using Machine Learning. Curr Med Imaging. 2021;17:479-490.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 23]  [Cited by in RCA: 15]  [Article Influence: 3.8]  [Reference Citation Analysis (0)]
86.  Zhang L, Zhang Y, Wang L, Wang J, Liu Y. Diagnosis of gastric lesions through a deep convolutional neural network. Dig Endosc. 2021;33:788-796.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 20]  [Cited by in RCA: 28]  [Article Influence: 7.0]  [Reference Citation Analysis (0)]
87.  Kim SH, Hwang Y, Oh DJ, Nam JH, Kim KB, Park J, Song HJ, Lim YJ. Efficacy of a comprehensive binary classification model using a deep convolutional neural network for wireless capsule endoscopy. Sci Rep. 2021;11:17479.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 1]  [Cited by in RCA: 6]  [Article Influence: 1.5]  [Reference Citation Analysis (0)]
88.  Yogapriya J, Chandran V, Sumithra MG, Anitha P, Jenopaul P, Suresh Gnana Dhas C. Gastrointestinal Tract Disease Classification from Wireless Endoscopy Images Using Pretrained Deep Learning Model. Comput Math Methods Med. 2021;2021:5940433.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 7]  [Cited by in RCA: 31]  [Article Influence: 7.8]  [Reference Citation Analysis (0)]
89.  Gamage C, Wijesinghe I, Chitraranjan C, Perera I.   GI-Net: Anomalies Classification in Gastrointestinal Tract through Endoscopic Imagery with Deep Learning. In: 2019 Moratuwa Engineering Research Conference (MERCon); 2019 Jul 3-5; Moratuwa, Sri Lanka. IEEE: NY, United States, 2019: 66-71.  [PubMed]  [DOI]  [Full Text]
90.  Zheng W, Zhang X, Kim JJ, Zhu X, Ye G, Ye B, Wang J, Luo S, Li J, Yu T, Liu J, Hu W, Si J. High Accuracy of Convolutional Neural Network for Evaluation of Helicobacter pylori Infection Based on Endoscopic Images: Preliminary Experience. Clin Transl Gastroenterol. 2019;10:e00109.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 64]  [Cited by in RCA: 62]  [Article Influence: 10.3]  [Reference Citation Analysis (0)]
91.  Lee JH, Kim YJ, Kim YW, Park S, Choi YI, Kim YJ, Park DK, Kim KG, Chung JW. Spotting malignancies from gastric endoscopic images using deep learning. Surg Endosc. 2019;33:3790-3797.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 45]  [Cited by in RCA: 62]  [Article Influence: 10.3]  [Reference Citation Analysis (0)]
92.  Hirasawa T, Aoyama K, Tanimoto T, Ishihara S, Shichijo S, Ozawa T, Ohnishi T, Fujishiro M, Matsuo K, Fujisaki J, Tada T. Application of artificial intelligence using a convolutional neural network for detecting gastric cancer in endoscopic images. Gastric Cancer. 2018;21:653-660.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 568]  [Cited by in RCA: 428]  [Article Influence: 61.1]  [Reference Citation Analysis (0)]
93.  Itoh T, Kawahira H, Nakashima H, Yata N. Deep learning analyzes Helicobacter pylori infection by upper gastrointestinal endoscopy images. Endosc Int Open. 2018;6:E139-E144.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 116]  [Cited by in RCA: 130]  [Article Influence: 18.6]  [Reference Citation Analysis (0)]
94.  Nadeem S, Tahir MA, Naqvi SSA, Zaid M.   Ensemble of Texture and Deep Learning Features for Finding Abnormalities in the Gastro-Intestinal Tract. In: Nguyen N, Pimenidis E, Khan Z, Trawiński B. Computational Collective Intelligence. ICCCI 2018; 2018 Sept 5-7; Bristol, United Kingdom. Switzerland: Springer, 2018: 469-479.  [PubMed]  [DOI]  [Full Text]
95.  He JY, Wu X, Jiang YG, Peng Q, Jain R. Hookworm Detection in Wireless Capsule Endoscopy Images With Deep Learning. IEEE Trans Image Process. 2018;27:2379-2392.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 102]  [Cited by in RCA: 80]  [Article Influence: 11.4]  [Reference Citation Analysis (0)]
96.  Shichijo S, Nomura S, Aoyama K, Nishikawa Y, Miura M, Shinagawa T, Takiyama H, Tanimoto T, Ishihara S, Matsuo K, Tada T. Application of Convolutional Neural Networks in the Diagnosis of Helicobacter pylori Infection Based on Endoscopic Images. EBioMedicine. 2017;25:106-111.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 235]  [Cited by in RCA: 181]  [Article Influence: 22.6]  [Reference Citation Analysis (0)]
97.  Zhu R, Zhang R, Xue D.   Lesion detection of endoscopy images based on convolutional neural network features. In: 2015 8th International Congress on Image and Signal Processing (CISP); 2015 Oct 14-16; Shenyang, China. IEEE: NY, United States, 2015: 372-376.  [PubMed]  [DOI]  [Full Text]
98.  Wu L, Zhou W, Wan X, Zhang J, Shen L, Hu S, Ding Q, Mu G, Yin A, Huang X, Liu J, Jiang X, Wang Z, Deng Y, Liu M, Lin R, Ling T, Li P, Wu Q, Jin P, Chen J, Yu H. A deep neural network improves endoscopic detection of early gastric cancer without blind spots. Endoscopy. 2019;51:522-531.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 194]  [Cited by in RCA: 157]  [Article Influence: 26.2]  [Reference Citation Analysis (0)]
99.  Leenhardt R, Vasseur P, Li C, Saurin JC, Rahmi G, Cholet F, Becq A, Marteau P, Histace A, Dray X; CAD-CAP Database Working Group. A neural network algorithm for detection of GI angiectasia during small-bowel capsule endoscopy. Gastrointest Endosc. 2019;89:189-194.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 129]  [Cited by in RCA: 148]  [Article Influence: 24.7]  [Reference Citation Analysis (1)]
100.  Alaskar H, Hussain A, Al-Aseem N, Liatsis P, Al-Jumeily D. Application of Convolutional Neural Networks for Automated Ulcer Detection in Wireless Capsule Endoscopy Images. Sensors (Basel). 2019;19:1265.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 113]  [Cited by in RCA: 81]  [Article Influence: 13.5]  [Reference Citation Analysis (0)]
101.  Shichijo S, Endo Y, Aoyama K, Takeuchi Y, Ozawa T, Takiyama H, Matsuo K, Fujishiro M, Ishihara S, Ishihara R, Tada T. Application of convolutional neural networks for evaluating Helicobacter pylori infection status on the basis of endoscopic images. Scand J Gastroenterol. 2019;54:158-163.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 75]  [Cited by in RCA: 65]  [Article Influence: 10.8]  [Reference Citation Analysis (0)]
102.  Cho BJ, Bang CS, Park SW, Yang YJ, Seo SI, Lim H, Shin WG, Hong JT, Yoo YT, Hong SH, Choi JH, Lee JJ, Baik GH. Automated classification of gastric neoplasms in endoscopic images using a convolutional neural network. Endoscopy. 2019;51:1121-1129.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 115]  [Cited by in RCA: 97]  [Article Influence: 16.2]  [Reference Citation Analysis (1)]
103.  Kahsaygebreslassie A, Yaecobgirmaygezahegn, Hagos MT, Achimibenthal, Pooja.   Automated Gastrointestinal Disease Recognition for Endoscopic Images. In: 2019 International Conference on Computing, Communication, and Intelligent Systems (ICCCIS); 2019 Oct 18-19; Greater Noida, India. IEEE: NY, United States, 2019: 312-316.  [PubMed]  [DOI]  [Full Text]
104.  Aoki T, Yamada A, Aoyama K, Saito H, Tsuboi A, Nakada A, Niikura R, Fujishiro M, Oka S, Ishihara S, Matsuda T, Tanaka S, Koike K, Tada T. Automatic detection of erosions and ulcerations in wireless capsule endoscopy images based on a deep convolutional neural network. Gastrointest Endosc. 2019;89:357-363.e2.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 163]  [Cited by in RCA: 175]  [Article Influence: 29.2]  [Reference Citation Analysis (0)]
105.  Srivastava A, Sengupta S, Kang S, Kant K, Khan M, Ali SA, Moore SR, Amadi BC, Kelly P, Syed S, Brown DE.   Deep Learning for Detecting Diseases in Gastrointestinal Biopsy Images. In: 2019 Systems and Information Engineering Design Symposium (SIEDS); 2019 Apr 26-26, Charlottesville, VA, United States. IEEE: NY, United States, 2019: 1-4.  [PubMed]  [DOI]  [Full Text]
106.  Khryashchev VV, Stepanova OA, Lebedev AA, Kashin SV, Kuvaev RO.   Deep Learning for Gastric Pathology Detection in Endoscopic Images. In: Proceedings of the 2019 3rd International Conference on Graphics and Signal Processing; 2019 Jun 1-3, Hong Kong. Association for Computing Machinery: NY, United States, 2019: 90-94.  [PubMed]  [DOI]  [Full Text]
107.  Togo R, Yamamichi N, Mabe K, Takahashi Y, Takeuchi C, Kato M, Sakamoto N, Ishihara K, Ogawa T, Haseyama M. Detection of gastritis by a deep convolutional neural network from double-contrast upper gastrointestinal barium X-ray radiography. J Gastroenterol. 2019;54:321-329.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 20]  [Cited by in RCA: 25]  [Article Influence: 4.2]  [Reference Citation Analysis (0)]
108.  Horie Y, Yoshio T, Aoyama K, Yoshimizu S, Horiuchi Y, Ishiyama A, Hirasawa T, Tsuchida T, Ozawa T, Ishihara S, Kumagai Y, Fujishiro M, Maetani I, Fujisaki J, Tada T. Diagnostic outcomes of esophageal cancer by artificial intelligence using convolutional neural networks. Gastrointest Endosc. 2019;89:25-32.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 240]  [Cited by in RCA: 273]  [Article Influence: 45.5]  [Reference Citation Analysis (0)]
109.  Ding Z, Shi H, Zhang H, Meng L, Fan M, Han C, Zhang K, Ming F, Xie X, Liu H, Liu J, Lin R, Hou X. Gastroenterologist-Level Identification of Small-Bowel Diseases and Normal Variants by Capsule Endoscopy Using a Deep-Learning Model. Gastroenterology. 2019;157:1044-1054.e5.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 248]  [Cited by in RCA: 209]  [Article Influence: 34.8]  [Reference Citation Analysis (0)]
110.  Jerebko AK, Malley JD, Franaszek M, Summers RM. Support vector machines committee classification method for computer-aided polyp detection in CT colonography. Acad Radiol. 2005;12:479-486.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 49]  [Cited by in RCA: 26]  [Article Influence: 1.3]  [Reference Citation Analysis (0)]
111.  Awad M, Motai Y, Näppi J, Yoshida H. A Clinical Decision Support Framework for Incremental Polyps Classification in Virtual Colonoscopy. Algorithms. 2010;3:1-20.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 13]  [Cited by in RCA: 13]  [Article Influence: 0.9]  [Reference Citation Analysis (0)]
112.  Raut V, Gunjan R, Shete VV, Eknath UD. Gastrointestinal tract disease segmentation and classification in wireless capsule endoscopy using intelligent deep learning model. Comput Methods Biomech Biomed Eng Imaging Vis. 2023;11:606-622.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 1]  [Reference Citation Analysis (0)]
113.  Shen Y, Chen A, Zhang X, Zhong X, Ma A, Wang J, Wang X, Zheng W, Sun Y, Yue L, Zhang Z, Zhang X, Lin N, Kim JJ, Du Q, Liu J, Hu W. Real-Time Evaluation of Helicobacter pylori Infection by Convolution Neural Network During White-Light Endoscopy: A Prospective, Multicenter Study (With Video). Clin Transl Gastroenterol. 2023;14:e00643.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 7]  [Cited by in RCA: 2]  [Article Influence: 1.0]  [Reference Citation Analysis (0)]
114.  Wang L, Chen L, Wang X, Liu K, Li T, Yu Y, Han J, Xing S, Xu J, Tian D, Seidler U, Xiao F. Development of a Convolutional Neural Network-Based Colonoscopy Image Assessment Model for Differentiating Crohn's Disease and Ulcerative Colitis. Front Med (Lausanne). 2022;9:789862.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in RCA: 7]  [Reference Citation Analysis (0)]
115.  Afify HM, Mohammed KK, Hassanien AE. An improved framework for polyp image segmentation based on SegNet architecture. Int J Imaging Syst Tech. 2021;31:1741-1751.  [PubMed]  [DOI]  [Full Text]
116.  Nezhad SA, Khatibi T, Sohrabi M. Combining CNNs and 2-D visualization method for GI tract lesions classification. Multimed Tools Appl. 2024;83:15825-15844.  [PubMed]  [DOI]  [Full Text]
117.  Sharma N, Gupta S, Koundal D, Alyami S, Alshahrani H, Asiri Y, Shaikh A. U-Net Model with Transfer Learning Model as a Backbone for Segmentation of Gastrointestinal Tract. Bioengineering (Basel). 2023;10:119.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in RCA: 16]  [Reference Citation Analysis (0)]
118.  Wang J, Liu X. Medical image recognition and segmentation of pathological slices of gastric cancer based on Deeplab v3+ neural network. Comput Methods Programs Biomed. 2021;207:106210.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 26]  [Cited by in RCA: 48]  [Article Influence: 12.0]  [Reference Citation Analysis (0)]
119.  Ye R, Wang R, Guo Y, Chen L. SIA-Unet: A Unet with Sequence Information for Gastrointestinal Tract Segmentation. In: Khanna S, Cao J, Bai Q, Xu G. PRICAI 2022: Trends in Artificial Intelligence. PRICAI 2022. Lect Notes Comput Sci. 2022;13629:316-326.  [PubMed]  [DOI]  [Full Text]
120.  Wang Y, Tiusaba L, Jacobs S, Saruwatari M, Ning B, Levitt M, Sandler AD, Nam SH, Kang JU, Cha J. Unsupervised and quantitative intestinal ischemia detection using conditional adversarial network in multimodal optical imaging. J Med Imaging (Bellingham). 2022;9:064502.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 2]  [Reference Citation Analysis (0)]
121.  Zhu C, Hua Y, Zhang M, Wang Y, Li W, Ding Y, She Q, Zhang W, Si X, Kong Z, Liu B, Chen W, Wu J, Dang Y, Zhang G. A Multimodal Multipath Artificial Intelligence System for Diagnosing Gastric Protruded Lesions on Endoscopy and Endoscopic Ultrasonography Images. Clin Transl Gastroenterol. 2023;14:e00551.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 7]  [Cited by in RCA: 10]  [Article Influence: 5.0]  [Reference Citation Analysis (0)]
122.  Lafraxo S, El Ansari M.   GastroNet: Abnormalities Recognition in Gastrointestinal Tract through Endoscopic Imagery using Deep Learning Techniques. In: 2020 8th International Conference on Wireless Networks and Mobile Communications (WINCOM); 2020 Oct 27-29, Reims, France. IEEE: NY. United States, 2020: 1-5.  [PubMed]  [DOI]  [Full Text]
123.  Jiang X, Luo X, Nan Q, Ye Y, Miao Y, Miao J. Application of deep learning in the diagnosis and evaluation of ulcerative colitis disease severity. Therap Adv Gastroenterol. 2023;16:17562848231215579.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Reference Citation Analysis (0)]
124.  Min JK, Kwak MS, Cha JM. Overview of Deep Learning in Gastrointestinal Endoscopy. Gut Liver. 2019;13:388-393.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 74]  [Cited by in RCA: 116]  [Article Influence: 23.2]  [Reference Citation Analysis (0)]
125.  Xie W, Hu J, Liang P, Mei Q, Wang A, Liu Q, Liu X, Wu J, Yang X, Zhu N, Bai B, Mei Y, Liang Z, Han W, Cheng M. Deep learning-based lesion detection and severity grading of small-bowel Crohn's disease ulcers on double-balloon endoscopy images. Gastrointest Endosc. 2024;99:767-777.e5.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 5]  [Cited by in RCA: 8]  [Article Influence: 8.0]  [Reference Citation Analysis (0)]
126.  Chu Y, Huang F, Gao M, Zou DW, Zhong J, Wu W, Wang Q, Shen XN, Gong TT, Li YY, Wang LF. Convolutional neural network-based segmentation network applied to image recognition of angiodysplasias lesion under capsule endoscopy. World J Gastroenterol. 2023;29:879-889.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in CrossRef: 1]  [Cited by in RCA: 5]  [Article Influence: 2.5]  [Reference Citation Analysis (0)]
127.  Mascarenhas M, Ribeiro T, Afonso J, Ferreira JPS, Cardoso H, Andrade P, Parente MPL, Jorge RN, Mascarenhas Saraiva M, Macedo G. Deep learning and colon capsule endoscopy: automatic detection of blood and colonic mucosal lesions using a convolutional neural network. Endosc Int Open. 2022;10:E171-E177.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 3]  [Cited by in RCA: 23]  [Article Influence: 7.7]  [Reference Citation Analysis (0)]
128.  Lo B, Liu Z, Bendtsen F, Igel C, Vind I, Burisch J. High Accuracy in Classifying Endoscopic Severity in Ulcerative Colitis Using Convolutional Neural Network. Am J Gastroenterol. 2022;117:1648-1654.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 17]  [Reference Citation Analysis (0)]
129.  Cho BJ, Bang CS, Lee JJ, Seo CW, Kim JH. Prediction of Submucosal Invasion for Gastric Neoplasms in Endoscopic Images Using Deep-Learning. J Clin Med. 2020;9:1858.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 43]  [Cited by in RCA: 36]  [Article Influence: 7.2]  [Reference Citation Analysis (0)]
130.  Wu L, Wang J, He X, Zhu Y, Jiang X, Chen Y, Wang Y, Huang L, Shang R, Dong Z, Chen B, Tao X, Wu Q, Yu H. Deep learning system compared with expert endoscopists in predicting early gastric cancer and its invasion depth and differentiation status (with videos). Gastrointest Endosc. 2022;95:92-104.e3.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 50]  [Cited by in RCA: 42]  [Article Influence: 14.0]  [Reference Citation Analysis (0)]
131.  Marin-Santos D, Contreras-fernandez JA, Perez-borrero I, Pallares-manrique H, Gegundez-arias ME. Automatic detection of crohn disease in wireless capsule endoscopic images using a deep convolutional neural network. Appl Intell. 2023;53:12632-12646.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
132.  Udristoiu AL, Stefanescu D, Gruionu G, Gruionu LG, Iacob AV, Karstensen JG, Vilman P, Saftoiu A. Deep Learning Algorithm for the Confirmation of Mucosal Healing in Crohn's Disease, Based on Confocal Laser Endomicroscopy Images. J Gastrointestin Liver Dis. 2021;30:59-65.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 6]  [Cited by in RCA: 14]  [Article Influence: 3.5]  [Reference Citation Analysis (0)]
133.  Ozawa T, Ishihara S, Fujishiro M, Saito H, Kumagai Y, Shichijo S, Aoyama K, Tada T. Novel computer-assisted diagnosis system for endoscopic disease activity in patients with ulcerative colitis. Gastrointest Endosc. 2019;89:416-421.e1.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 182]  [Cited by in RCA: 148]  [Article Influence: 24.7]  [Reference Citation Analysis (1)]
134.  Stidham RW, Liu W, Bishu S, Rice MD, Higgins PDR, Zhu J, Nallamothu BK, Waljee AK. Performance of a Deep Learning Model vs Human Reviewers in Grading Endoscopic Disease Severity of Patients With Ulcerative Colitis. JAMA Netw Open. 2019;2:e193963.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 116]  [Cited by in RCA: 176]  [Article Influence: 29.3]  [Reference Citation Analysis (0)]
135.  Zhou T, Han G, Li BN, Lin Z, Ciaccio EJ, Green PH, Qin J. Quantitative analysis of patients with celiac disease by video capsule endoscopy: A deep learning method. Comput Biol Med. 2017;85:1-6.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 95]  [Cited by in RCA: 103]  [Article Influence: 12.9]  [Reference Citation Analysis (0)]
136.  Zhu Y, Wang QC, Xu MD, Zhang Z, Cheng J, Zhong YS, Zhang YQ, Chen WF, Yao LQ, Zhou PH, Li QL. Application of convolutional neural network in the diagnosis of the invasion depth of gastric cancer based on conventional endoscopy. Gastrointest Endosc. 2019;89:806-815.e1.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 201]  [Cited by in RCA: 232]  [Article Influence: 38.7]  [Reference Citation Analysis (0)]
137.  Jiang Q, Yu Y, Ren Y, Li S, He X. A review of deep learning methods for gastrointestinal diseases classification applied in computer-aided diagnosis system. Med Biol Eng Comput. 2025;63:293-320.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 2]  [Reference Citation Analysis (0)]
138.  Mohan BP, Khan SR, Kassab LL, Ponnada S, Mohy-Ud-Din N, Chandan S, Dulai PS, Kochhar GS. Convolutional neural networks in the computer-aided diagnosis of Helicobacter pylori infection and non-causal comparison to physician endoscopists: a systematic review with meta-analysis. Ann Gastroenterol. 2021;34:20-25.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 8]  [Cited by in RCA: 8]  [Article Influence: 1.6]  [Reference Citation Analysis (0)]
139.  Liu P, Wang P, Glissen Brown JR, Berzin TM, Zhou G, Liu W, Xiao X, Chen Z, Zhang Z, Zhou C, Lei L, Xiong F, Li L, Liu X. The single-monitor trial: an embedded CADe system increased adenoma detection during colonoscopy: a prospective randomized study. Therap Adv Gastroenterol. 2020;13:1756284820979165.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 57]  [Cited by in RCA: 56]  [Article Influence: 11.2]  [Reference Citation Analysis (0)]
140.  Mohan BP, Facciorusso A, Khan SR, Chandan S, Kassab LL, Gkolfakis P, Tziatzios G, Triantafyllou K, Adler DG. Real-time computer aided colonoscopy versus standard colonoscopy for improving adenoma detection rate: A meta-analysis of randomized-controlled trials. EClinicalMedicine. 2020;29-30:100622.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 16]  [Cited by in RCA: 22]  [Article Influence: 4.4]  [Reference Citation Analysis (0)]
141.  Takiyama H, Ozawa T, Ishihara S, Fujishiro M, Shichijo S, Nomura S, Miura M, Tada T. Automatic anatomical classification of esophagogastroduodenoscopy images using deep convolutional neural networks. Sci Rep. 2018;8:7497.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 119]  [Cited by in RCA: 88]  [Article Influence: 12.6]  [Reference Citation Analysis (0)]
142.  Liu GS, Huang PY, Wen ML, Zhuang SS, Hua J, He XP. Application of endoscopic ultrasonography for detecting esophageal lesions based on convolutional neural network. World J Gastroenterol. 2022;28:2457-2467.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in RCA: 3]  [Reference Citation Analysis (0)]
143.  Kim YH, Kim GH, Kim KB, Lee MW, Lee BE, Baek DH, Kim DH, Park JC. Application of A Convolutional Neural Network in The Diagnosis of Gastric Mesenchymal Tumors on Endoscopic Ultrasonography Images. J Clin Med. 2020;9:3162.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 32]  [Cited by in RCA: 33]  [Article Influence: 6.6]  [Reference Citation Analysis (0)]
144.  Tsuboi A, Oka S, Aoyama K, Saito H, Aoki T, Yamada A, Matsuda T, Fujishiro M, Ishihara S, Nakahori M, Koike K, Tanaka S, Tada T. Artificial intelligence using a convolutional neural network for automatic detection of small-bowel angioectasia in capsule endoscopy images. Dig Endosc. 2020;32:382-390.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 75]  [Cited by in RCA: 103]  [Article Influence: 20.6]  [Reference Citation Analysis (0)]
145.  Aoki T, Yamada A, Aoyama K, Saito H, Fujisawa G, Odawara N, Kondo R, Tsuboi A, Ishibashi R, Nakada A, Niikura R, Fujishiro M, Oka S, Ishihara S, Matsuda T, Nakahori M, Tanaka S, Koike K, Tada T. Clinical usefulness of a deep learning-based system as the first screening on small-bowel capsule endoscopy reading. Dig Endosc. 2020;32:585-591.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 53]  [Cited by in RCA: 75]  [Article Influence: 15.0]  [Reference Citation Analysis (0)]
146.  Turan M, Almalioglu Y, Araujo H, Konukoglu E, Sitti M. Deep EndoVO: A recurrent convolutional neural network (RCNN) based visual odometry approach for endoscopic capsule robots. Neurocomputing. 2018;275:1861-1870.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 65]  [Cited by in RCA: 35]  [Article Influence: 5.0]  [Reference Citation Analysis (0)]
147.  Boškoski I, Orlandini B, Papparella LG, Matteo MV, De Siena M, Pontecorvi V, Costamagna G. Robotics and Artificial Intelligence in Gastrointestinal Endoscopy: Updated Review of the Literature and State of the Art. Curr Robot Rep. 2021;2:43-54.  [PubMed]  [DOI]  [Full Text]
148.  Messmann H, Bisschops R, Antonelli G, Libânio D, Sinonquel P, Abdelrahim M, Ahmad OF, Areia M, Bergman JJGHM, Bhandari P, Boskoski I, Dekker E, Domagk D, Ebigbo A, Eelbode T, Eliakim R, Häfner M, Haidry RJ, Jover R, Kaminski MF, Kuvaev R, Mori Y, Palazzo M, Repici A, Rondonotti E, Rutter MD, Saito Y, Sharma P, Spada C, Spadaccini M, Veitch A, Gralnek IM, Hassan C, Dinis-Ribeiro M. Expected value of artificial intelligence in gastrointestinal endoscopy: European Society of Gastrointestinal Endoscopy (ESGE) Position Statement. Endoscopy. 2022;54:1211-1231.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 83]  [Cited by in RCA: 75]  [Article Influence: 25.0]  [Reference Citation Analysis (0)]
149.  Berzin TM, Parasa S, Wallace MB, Gross SA, Repici A, Sharma P. Position statement on priorities for artificial intelligence in GI endoscopy: a report by the ASGE Task Force. Gastrointest Endosc. 2020;92:951-959.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 40]  [Cited by in RCA: 66]  [Article Influence: 13.2]  [Reference Citation Analysis (0)]
150.  Su X, Liu W, Jiang S, Gao X, Chu Y, Ma L. Deep learning-based anatomical position recognition for gastroscopic examination. Technol Health Care. 2024;32:39-48.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
151.  Zhang C, Xiong Z, Chen S, Ding A, Cao Y, Liu B, Liu X. Automated Disease Detection in Gastroscopy Videos Using Convolutional Neural Networks. Front Med (Lausanne). 2022;9:846024.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Reference Citation Analysis (0)]
152.  Chen M, Wang J, Xiao Y, Wu L, Hu S, Chen S, Yi G, Hu W, Xie X, Zhu Y, Chen Y, Yang Y, Yu H. Automated and real-time validation of gastroesophageal varices under esophagogastroduodenoscopy using a deep convolutional neural network: a multicenter retrospective study (with video). Gastrointest Endosc. 2021;93:422-432.e3.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 19]  [Cited by in RCA: 18]  [Article Influence: 4.5]  [Reference Citation Analysis (0)]
153.  Saito H, Tanimoto T, Ozawa T, Ishihara S, Fujishiro M, Shichijo S, Hirasawa D, Matsuda T, Endo Y, Tada T. Automatic anatomical classification of colonoscopic images using deep convolutional neural networks. Gastroenterol Rep (Oxf). 2021;9:226-233.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 16]  [Cited by in RCA: 10]  [Article Influence: 2.5]  [Reference Citation Analysis (0)]
154.  Wu L, Shang R, Sharma P, Zhou W, Liu J, Yao L, Dong Z, Yuan J, Zeng Z, Yu Y, He C, Xiong Q, Li Y, Deng Y, Cao Z, Huang C, Zhou R, Li H, Hu G, Chen Y, Wang Y, He X, Zhu Y, Yu H. Effect of a deep learning-based system on the miss rate of gastric neoplasms during upper gastrointestinal endoscopy: a single-centre, tandem, randomised controlled trial. Lancet Gastroenterol Hepatol. 2021;6:700-708.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 82]  [Cited by in RCA: 78]  [Article Influence: 19.5]  [Reference Citation Analysis (0)]
155.  Xu Z, Tao Y, Wenfang Z, Ne L, Zhengxing H, Jiquan L, Weiling H, Huilong D, Jianmin S. Upper gastrointestinal anatomy detection with multi-task convolutional neural networks. Healthc Technol Lett. 2019;6:176-180.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 7]  [Cited by in RCA: 15]  [Article Influence: 2.5]  [Reference Citation Analysis (0)]
156.  Pogorelov K, Riegler M, Eskeland SL, de Lange T, Johansen D, Griwodz C, Schmidt PT, Halvorsen P. Efficient disease detection in gastrointestinal videos – global features versus neural networks. Multimed Tools Appl. 2017;76:22493-22525.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 54]  [Cited by in RCA: 14]  [Article Influence: 1.8]  [Reference Citation Analysis (0)]
157.  Harzig P, Einfalt M, Lienhart R.   Automatic Disease Detection and Report Generation for Gastrointestinal Tract Examination. In: Proceedings of the 27th ACM International Conference on Multimedia; 2019 Oct 21-25, France. Association for Computing Machinery: NY, United States, 2019: 2573-2577.  [PubMed]  [DOI]  [Full Text]
158.  Hann A, Troya J, Fitting D. Current status and limitations of artificial intelligence in colonoscopy. United European Gastroenterol J. 2021;9:527-533.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 10]  [Cited by in RCA: 34]  [Article Influence: 8.5]  [Reference Citation Analysis (0)]
159.  Alzubaidi L, Zhang J, Humaidi AJ, Al-Dujaili A, Duan Y, Al-Shamma O, Santamaría J, Fadhel MA, Al-Amidie M, Farhan L. Review of deep learning: concepts, CNN architectures, challenges, applications, future directions. J Big Data. 2021;8:53.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 3068]  [Cited by in RCA: 1118]  [Article Influence: 279.5]  [Reference Citation Analysis (0)]
160.  Mohapatra S, Swarnkar T, Mishra M, Al-dabass D, Mascella R.   Deep learning in gastroenterology: a brief review. In: Nayak J, Naik B, Pelusi D, Das AK. Handbook of Computational Intelligence in Biomedical Engineering and Healthcare. Switzerland: Elsevier, 2021: 121-149.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 9]  [Cited by in RCA: 1]  [Article Influence: 0.3]  [Reference Citation Analysis (0)]
161.  Barić D, Fumić P, Horvatić D, Lipic T. Benchmarking Attention-Based Interpretability of Deep Learning in Multivariate Time Series Predictions. Entropy (Basel). 2021;23:143.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 9]  [Cited by in RCA: 7]  [Article Influence: 1.8]  [Reference Citation Analysis (0)]
162.  Zhang Q, Zhu S. Visual interpretability for deep learning: a survey. Frontiers Inf Technol Electronic Eng. 2018;19:27-39.  [PubMed]  [DOI]  [Full Text]
163.  Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, Uszkoreit J, Houlsby N.   An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. 2020 Preprint. Available from: arXiv:2010.11929.  [PubMed]  [DOI]  [Full Text]
164.  Wu Y, Ma Y, Liu J, Du J, Xing L. Self-Attention Convolutional Neural Network for Improved MR Image Reconstruction. Inf Sci (N Y). 2019;490:317-328.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 80]  [Cited by in RCA: 55]  [Article Influence: 9.2]  [Reference Citation Analysis (0)]
165.  Chen J, Lu Y, Yu Q, Luo X, Adeli E, Wang Y, Lu L, Yuille AL, Zhou Y.   TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation. 2021 Preprint. Available from: arXiv:2102.04306.  [PubMed]  [DOI]  [Full Text]