APP下载

Evolving role of artificial intelligence in gastrointestinal endoscopy

2021-01-15GulshanParasherMorganWongManmeetRawat

World Journal of Gastroenterology 2020年46期

Gulshan Parasher, Morgan Wong, Manmeet Rawat

Abstract Artificial intelligence (AI) is a combination of different technologies that enable machines to sense, comprehend, and learn with human-like levels of intelligence. AI technology will eventually enhance human capability, provide machines genuine autonomy, and reduce errors, and increase productivity and efficiency. AI seems promising, and the field is full of invention, novel applications; however, the limitation of machine learning suggests a cautious optimism as the right strategy. AI is also becoming incorporated into medicine to improve patient care by speeding up processes and achieving greater accuracy for optimal patient care. AI using deep learning technology has been used to identify, differentiate catalog images in several medical fields including gastrointestinal endoscopy. The gastrointestinal endoscopy field involves endoscopic diagnoses and prognostication of various digestive diseases using image analysis with the help of various gastrointestinal endoscopic device systems. AI-based endoscopic systems can reliably detect and provide crucial information on gastrointestinal pathology based on their training and validation. These systems can make gastroenterology practice easier, faster, more reliable, and reduce inter-observer variability in the coming years. However, the thought that these systems will replace human decision making replace gastrointestinal endoscopists does not seem plausible in the near future. In this review, we discuss AI and associated various technological terminologies, evolving role in gastrointestinal endoscopy, and future possibilities.

Key Words: Artificial intelligence; Gastrointestinal endoscopy; Machine learning; Endoscopic diagnosis; Gastric cancer; Gastrointestinal diseases

INTRODUCTION

Artificial intelligence (AI) is a combination of advanced computational technologies that allows machines to learn and perform with intelligence similar to human brain. More recently, AI is becoming an integral part of practice of medicine, helping provide accurate and optimal patient care. Electronic health records have enhanced and accelerated physicians’ capabilities for treating patients and providing crucial data for innovation and development by using machine learning (ML)[1,2]. AI is probably the best technology for rapid and reliably accurate means to treat the growing patient population.

AI using deep learning (DL) technology has been used to identify, differentiate and catalog images in several medical fields including gastrointestinal endoscopy[1,2]. The gastrointestinal endoscopy field involves endoscopic diagnoses and prognostication of various digestive diseases using image analysis with the help of various gastrointestinal endoscopic device systems[3]. As the field of AI in endoscopy is going to grow like several applications in every-day life around us. Gastroenterologists need to be acquainted with the emerging applications, advantages, and limitations of AI as it pertains to gastrointestinal endoscopy and should embrace it sooner than later.

Despite the overwhelming initial excitement, a lot more remains to be discovered in the field of AI and endoscopy as the technology evolves and becomes more sophisticated. In this review, we discuss AI and associated various technological terminologies, evolving role in gastrointestinal endoscopy, and future possibilities. We searched the internet using PubMed, Google Scholar, and Google for related abstracts and published articles in the English language using keywords: “artificial intelligence and endoscopy”, “machine learning”, “gastric cancer”, “computer-aided diagnosis and endoscopy”, “computer-aided detection”, “CAD”, “neural networks” and “endoscopy” for this review.

AI RELATED TERMINOLOGY

AI refers to a machine or computer’s ability to perform tasks with cognition, similar to those of humans, such as learning and problem solving that mimic the human mind[1,4]. The fundamental principle of this technology is “machine learning”, which employs teaching computer algorithms to learn and recognize patterns in the data automatically[4,5]. On the other hand, DL is based on the human brain’s biological neural network called multi-layered artificial neural networks (ANNs)[5,6]. DL can recognize patterns, analyze data, and make conclusions and decisions[6].

ML

ML is a subdivision of AI that "focuses on teaching computers how to learn from available data, and identify similarities or patterns and can predict data without the need for external human intervention or programming[4,5]. ML allows us to create algorithms (Figure 1). These algorithms learn from and make predictions on data. ML has its foundations in computational learning theory in pattern recognition[4,5,7].

DL

DL or deep structured learning is part of a broader family of ML that uses algorithms for machines to comprehend training images and speech to extract specific features. DL technology is ideal for quantifying images and providing classification. A flexible set of rules and techniques makes it possible for the neural network to teach itself to accomplish goals it was not trained to solve[4,6,7].

ANNs

A neural network usually refers to software, hardware, or both modeled after and is similar to aspects of the human brain. Neural networks are layers of interconnected artificial neurons. Different layers analyze and learn from the previous layer and together can engage and perform more complex decision making[4,6]. Several studies involving AI in medicine use ANN’s.

Convolutional neural networks

Convolutional neural networks (CNNs) are a specific class of deep neural networks, usually applied to visual image analysis including image recognition and classification. These require minimal preprocessing and are also known as space invariant ANNs based on their shared weights architecture and translational invariants characteristic[4,6,7]. CNN is designed to utilize large datasets to learn patterns in correlating images. CNN’s connect inputs through pattern recognition that are trained through datasets[4,6,7].

AI AND ENDOSCOPY

Over time the image recognition using AI with ML has dramatically improved and currently has multiple applications in medicine as far as diagnostic imaging is concerned[1-3]. AI-related software is being used in the histological classification of gastric biopsy and characterization of colorectal lesions using endocytoscopy, diabetic retinopathy, skin cancer classification[3,8,9]. AI-based systems for endoscopy primarily improve detection rates, help predict histology based neoplasticvsnon-neoplastic behavior and subsequently guide therapy in the form of endoscopic resectionvssurgical resection[8,9]. AI-based computer algorithms provide diagnostic information by analyzing surface micro topological patterns, color differences, micro-vascular and pit pattern, narrow band imaging (NBI), high-magnification and endocytoscopy appearance from still images and video frames. The results are further validated subsequently by using another test database or prospectivein vivoclinical trials[3,8,9].

Initial studies have demonstrated the ability of AI-based endoscopy software in diagnosing Barrett’s esophagus (BE), esophageal and stomach cancer, andHelicobacter pylori(H. pylori)-related gastritis, and colorectal polyps[10]. Endobrain (Olympus; Tokyo, Japan) is one of the initial AI-based endoscopic system available for commercial use. The rapid progress of various AI systems will allow for more advanced predictions.

AI in upper endoscopy

Gastrointestinal malignancies remain the leading cause of mortality and morbidity worldwide. A high-quality endoscopic examination for the recognition of gastrointestinal neoplasms is essential for the gastroenterologist to enhance the detection of gastrointestinal neoplasms and optimize the treatment plans. Besides neoplasms, AI has also been increasingly used in other gastrointestinal diseases, including gastroesophageal reflux disease (GERD), infection, inflammation, and hemorrhage[10-13].

AI in GERD diagnosis

Upper endoscopy is the diagnostic modality of choice for GERD and its complications including BE. AI is reliably able to depict the anatomical location of endoscopic images. Takiyamaet al[14]reported a CNN model that is able to examine upper endoscopic images with precise anatomical localization with area under the receiver operating characteristic curve (AUROC) of 1.00 for the larynx and esophagus, and AUROC of 0.99 for stomach and duodenum landmarks. AI can also accurately predict the diagnosis of GERD. Using an ANN model, Paceet al[15]were able to predict GERD’s diagnosis using only 45 clinical variables in 159 cases with 100% accuracy. Changes in intrapapillary capillary loops and incremental increase as identified by NBI using deep CNN’s have been proposed as a disease marker for GERD. Near focus-NBI containing AI model has been described as a novel method for the diagnosis of GERD[16].

Figure 1 Schematic representation of different aspects of machine learning approach. AI: Artificial intelligence; ML: Machine learning.

AI in BE and esophageal cancer

Early detection of BE neoplastic changes is essential, especially in promising curative endoscopic treatment such as endoscopic mucosal resection (EMR) and radiofrequency ablation[10,11]. van der Sommenet al[11]developed an automated algorithm to help in the differentiation of early neoplastic lesions in BE. This algorithm included specific textures, color filters, and ML from one hundred images from upper endoscopy, and helped in detection of early neoplastic lesions with a sensitivity and specificity of 83%. Another study by a similar group used a novel computer model with 60ex vivovolumetric laser endomicroscopy (VLE) images[10]. VLE is an advanced imaging technology utilizing an optical coherence-based probe that provides a microscopic scan of the esophageal wall layers. The study compared the performance of the computer algorithm in detecting VLE images containing BE neoplasiavsVLE image interpretation by experts. The histological diagnosis, correlated with the images, served as the reference standard. The algorithm showed the ability to diagnose BE neoplasia with a sensitivity of 90% and specificity of 93%[10]. Current VLE consoles in clinical use have built-in AI for dysplasia and cancer detection.

The critical goal of Barrett’s surveillance programs is early detection of dysplasia. ML model has been used to differentiate dysplastic and non-neoplastic lesions in various studies[10,11,17]. Computer-based analysis can be used to identify malignant tissue at the cellular level by microscopic visualization of the mucosal surface using endocytoscopy images[17]. An image analysis model was able to diagnose esophageal squamous dysplasia by high-resolution microendoscopy (HRME), exhibited a sensitivity of 87% and specificity of 97%[18]. Further cost-effective modifications of this model using tablet-interfaced HRME with full automation for real-time analysis resulted in diagnostic accuracy of esophageal squamous cell carcinoma with 95% sensitivity and 91% specificity[19].

Horieet al[20]used CNNs-mediated AI for esophageal cancer diagnosis. A total of 8428 endoscopic images were used to train the CNN, including both white-light and narrow-band images. The model was able to diagnose esophageal cancer with a sensitivity of 95%. The ability to detect cancers smaller than 10 mm was 100%. This model had a very high diagnostic accuracy of 98% in differentiating between superficial and advanced esophageal cancer[20]. AI-based models have also been used for disease prognostication. An ANN model with improved accuracy was also described by Satoet al[21],418 esophageal cancer cases were used to predict 1-year and 5-year survival rates. In summary AI-based technologies and endoscopic systems show great promising results in improved detection of BE related dysplasia and esophageal carcinoma and show current and future promise in reducing mortality and morbidity related to these disease states.

AI in H. pylori infection diagnosis

H. pyloriinfection is an established risk factor for the development of gastroduodenal ulcers and gastric cancer. AI-based research in this field has been focused primarily to aid in the endoscopic diagnosis ofH. pyloriinfections[12,22]. The AI-based neural network model was used to predictH. pyloriinfection using related gastric histologic features. Parameters of 84 images from 30 patients were used to train this model, and reported a sensitivity of 85.4% and specificity of 90.9% for the detection ofH. pyloriinfection[22]. The model successfully identified gastric atrophy and gastric intestinal metaplasia, and was helpful in prediction of the infection severity with above 80% accuracy[22].

A prospective pilot study with an AI-based model for detection ofH. pyloriinfections demonstrated better yield with blue laser imaging-bright and linked color imaging training (AUROCs of 0.96 and 0.95) than white light imaging training (0.66)[23]. Another CNN model used 596 endoscopic images to detect theH. pyloriinfections. This model revealed exiting results with 86.7% of sensitivity and specificity[24]. Shichijoet al[12]used endoscopic images to diagnoseH. pyloriinfection and compared the performance of a CNN model to 23 endoscopists. The CNN model exhibited better sensitivity (88.9%vs79.0%) and specificity (87.4%vs83.2%), a superior accuracy (87.7%vs82.4%), and lesser time for diagnosis (194 svs230 s)[12,24].

AI in gastric cancer

Gastric cancer is the fifth most common form of cancer and the third leading cause of cancer-related death worldwide[8]. Early-stage endoscopic detection of gastric cancer helps us to triage patients for minimally invasive endoscopic therapy, including EMR or endoscopic submucosal dissection[8,25,26]. Esophagogastroduodenoscopy (EGD) is the current standard procedure for the diagnoses of gastric cancer. EGD in the diagnoses of gastric carcinoma carries a false-negative rate as high as 4.6%-25.8%[25]. Early gastric cancer lesions show only subtle morphologic changes can be easily missed and are difficult to distinguish from back-ground mucosa with atrophic changes[8,25,26].

AI is currently being utilized primarily for the detection and characterization of gastrointestinal neoplasms. Kubotaet al[26]utilized a backpropagation model after 10-time cross-validation using 902 gastric cancer endoscopic images. The diagnostic accuracy of this model was found to be limited, 77.2%, 49.1%, 51.0%, and 55.3% for T1-4 staging[26]. Hirasawaet al[8]created and trained a CNN based model containing 13584 endoscopic images and subsequently analyzed 2296 images with 92.2% sensitivity for gastric cancer detection. The diagnostic efficiency of neoplasm with 6 mm or greater size was 98.6%[8]. This model was able to identify all invasive cancers, which can be extremely hard to diagnose from gastritis even by experts. Gastritis findings with erythema, atrophy, and intestinal metaplasia were most common reasons for misdiagnosis[8].

Zhuet al[25]developed and trained a CNN model to recognize gastric cancer (M/SM1vsdeeper than SM1) using 790 conventional endoscopic images and subsequently performed analysis with another two hundred and three images. This study demonstrated superior accuracy (89.2%) and specificity (95.6%) as far as depth of invasion of gastric cancer was concerned[25]. Another computer-aided diagnosis (CAD) system using a support vector machine (SVM) was also used to identify early gastric cancer using magnifying NBI images[27]. This study showed very high diagnosis rate [accuracy 96.3%, positive predictive value (PPV) 98.3%, sensitivity 96.7%, specificity 95%] and area concordance performance (accuracy 73.8%, PPV 75.3%, sensitivity 65.5%, specificity 80.8%)[27].

AI in small intestinal endoscopic imaging

Video capsule endoscopy (VCE) is an integral part of endoscopic detection of small intestinal endoscopy. VCE is a subjective, time consuming diagnostic study, and due to low-resolution frame images, the interpretation of capsule endoscopy images and disease diagnosis relies on the image reviewer. AI has been used for VCE diagnosis of angiodysplasia, celiac or intestinal hookworms[13,28,29]. Leenhardtet al[13]developed a CNN based gastrointestinal angiodysplasia detection model using six hundred control and six hundred angiodysplasia images to create 4166 small bowel capsule endoscopy videos. Overall diagnostic performance was high with a sensitivity of 100%, a specificity of 96%, with a PPV of 96%, and negative predictive value (NPV) of 100%[29]. Another study using CNN based model showed sensitivity and specificity of 100% for celiac disease diagnosis[29]. Seguíet al[30]described a CNN model for small-intestine motility characterization using motility events. The mean classification accuracy was 96% for different intestinal motility events (“turbid”, “bubbles”, “clear blob”, “wrinkles”, “wall”, and “undefined”). More important and recent clinical studies related to AI, VCE and small intestinal imaging are summarized in Table 1. Currently available published studies show promising diagnostic accuracy using CNN for small intestinal imaging. Further multi center validation study will provide impetus to its use in clinical practice.

Improvement in capsule endoscopy may include improved capsule with controlled locomotion and navigation. This may enhance mucosal visualization and help better delineate small intestinal pathology further. This may provide us with the ability to perform real time VCE assisted forceps biopsy or fine needle aspiration (FNA) of the target lesion as well as endoscopic therapy in future[31,32]. Few proposed improvement include self-propelled mechanism with external control or a video capsule endoscope externally propelled by magnetic force with navigational controls. Multiple such prototypes for different indications are being currently investigated to assess suitability in clinical practice.

AI and endoscopic ultrasound

CAD has been used to assist endosonographers in the endoscopic ultrasound (EUS) evaluation of pancreatic lesions[33,34]. The diagnosis of a pancreatic neoplasm in the setting of chronic pancreatitis (CP) is challenging. EUS imaging is unable to distinguish neoplastic from reactive changes reliably. To differentiate between pancreatic cancer (PC) and CP, CAD techniques were used to extract EUS image parameters. A total of 105 features were extracted from 262 PC and 126 CP patient EUS images. Sixteen of these features were used for classification using a SVM model with a sensitivity rate of 94%[33]. Another ANN model based on digital image analysis of EUS images accurately classified PC with an area under the curve of 0.93[34].

ANN-based approach to differentiate PC and mass-forming CP using vascularity parameters obtained from post-processing of contrast-enhanced EUS videos suggest that vascularity parameters can reliably differentiate between PC and CP and can be used in an automated CAD system with excellent results (94.64% sensitivity and 94.44% specificity)[35]. EUS elastography was used for the differential diagnosis of CP and PC using EUS elastography movies and calculating hue histograms of individual images, followed by extended neural network analysis. Multilayer perceptron neural networks (MNNs) were initially trained to differentiate benign or malignant cases. This approach resulted in an outstanding average testing performance of 95% and a superior training performance of 97%[36]. Cytologic diagnosis of PC can be challenging at times. Attempts to integrate computer models into the cytopathologic diagnosis of PC have been studied. MNN using segment images of cell clusters from FNA of pancreatic tumors was able to differentiate between benign and malignant cells. The accuracy of MNN based slide analysis was equivalent to the cytopathologist review. Interestingly, the MNN model’s predictions showed a sensitivity of 80% for inconclusive FNA results[37]. Future advances in technology can help yield a more definitive diagnosis.

AI based models also provide an emerging approach to evaluate the risk of PC in cystic neoplasms. Using computer-aided frameworks, AI algorithms can recognize pancreatic cysts with malignant potential that are at increased risk of developing into PC[38,39]. A study using ANN and personal health data was useful in predicting the risk of PC, with a sensitivity of 80.7% and a specificity of 80.7%[39]. This particular ANN based model was able to stratify patients into low, medium, and high cancer risks[39].

AI in lower endoscopy

Colorectal cancer (CRC) is the second leading cause of cancer-related deaths in the United States. Over the past three decades, the rates of colon cancer and colon cancerrelated mortality have decreased due to the introduction of colonoscopy and polypectomy of adenomatous polyps[40,41]. A large cohort study conducted in the United States has also demonstrated an approximately 70% reduction in deaths after the institution of colonoscopy[42]. Colonoscopy is the best screening modality currently for diagnosis of adenomatous polyp and CRC. The CRC miss rate on screening colonoscopies is as high as 4%-5%[43], with the interval CRC reported as 7.2%-9% despite adherence to current guidelines[44]. The rate of interval CRC development is primarily due to missed lesions, newer lesions, and inadequate endoscopic treatment[45]. The overall pooled miss rate of polyps in a meta-analysis is 22% for polyps of any given size[46]. Consequently, the importance of standardizing the qualityof colonoscopy and increasing the detection of adenomatous polyps remains an important goal.

Table 1 Recent clinical studies in artificial intelligence using central neural network and capsule endoscopy for small intestinal imaging

Adenoma detection rate (ADR) has become a key quality measure for evaluating an endoscopist’s ability to find adenomas. However, rates of adenoma detection vary among endoscopists (7%-53%)[43]. Higher ADRs are associated with lower CRC mortality, with every 1% increase in ADR reflecting an associated 3.0% decrease in risk of interval CRC[43,45,46]. Given an increased focus on ADR, AI, or CAD are becoming rapidly integrated technologies utilized throughout modern medicine. In Gastroenterology, the use of AI is now being utilized to increase polyp detection rates and distinguish between adenomatous polyps and their hyperplastic counterparts. The use of AI in colonoscopy allows for the standardization of colonoscopy regardless of the endoscopists’ skills.

The primary role of AI in colonoscopy is thought to center around increased polyp detection and histopathological polyp differentiation without modification of the colonoscope or procedure. AI allows for real-time automatic polyp detection to help assist the endoscopist in detecting adenomatous lesions and signaling the endoscopist’s attention. AI or CAD-based models are trained using thousands of colonoscopic images to help identify and differentiate between hyperplastic and adenomatous polyps. After its initial training, the software continues to improve its detection abilities over time without the need for reprogramming. The introduction of AI in polyp detection was first performed in 2003, with pilot studies noting > 90% accuracy in preliminary studies. Initial models noted acceptable performance ranges with sensitivities of 48%-90%. However, these databases were small based on dataset of images of fewer than 20 polyps[3,47].

AI in colon polyp identification

CAD endoscopy allows for real-time automatic polyp detection and would assist the endoscopist in detecting lesions reliably and consistently. In the past, multiple CAD systems have been introduced in colon polyp detection. Initial systems utilized handcrafted features and variables that allowed for the identification of polyps. However, the usage of hand-crafted algorithms presented multiple issues, including calculation speed, computer power, lower sensitivity, and false-positive rate resulting from the presence of artifacts such as blood vessels, stool, and folds. Considerable advancement using DL automatically searches and defines important features of images resulting in better outcomes[48]. Retrospective studies for polyp detection using AI have demonstrated sensitivities higher than 90%, with NPV of 95%-97%. However, these trials are thought to be small-scale and non-clinical investigations[49].

Recent open, non-blinded trials in patients randomized to a standard colonoscopyvsthose randomized to colonoscopy with CAD noted a significantly increased ADR 20.3%vs29.1% in the CAD group. An increase in the mean number of adenomas detected was noted, 0.53 in the CAD groupvs0.31 in the standard colonoscopy group, mostly due to increased detection of diminutive adenomas[50]. A meta-analysis of 2815 patients correlated with the above study and demonstrated an ADR in the CAD colonoscopy population of 32.9%vsthe standardized colonoscopy group of 20.8%[51].

AI in colon polyp optical pathology

Optical diagnosis of colorectal polyps may prove to be more cost-effective and time saving than traditional post-polypectomy pathology diagnosis. Optical diagnosis eliminates the need for a pathologist as the polyps do not need to be reviewed by a pathologist. The application of CAD to magnifying narrow-band imaging has recently attracted the most significant interest in AI and colonoscopy. Initial studies performed by Tischendorfet al[52]and Grosset al[53]demonstrated the diagnostic accuracy of neoplastic polyps, being 85.3% and 93.1%, respectively. These studies utilized similar algorithms based on nine vessel features, which included (length, brightness, perimeter) from magnifying NBI images. The software was then able to classify these images into two pathological predictions, neoplasticvsnon-neoplastic. Recently retrospective studies on newly developed CAD systems utilizing DL algorithms have demonstrated 96.3% sensitivity and 78.1% specificity in identifying neoplastic or hyperplastic polyps[54]. Magnifying chromoendoscopy with CAD systems recognizing quantitative analysis of pit stricture or texture analysis of the entire image. Studies by Takemuraet al[55]have demonstrated an overall accuracy of 98.5%.

However, despite advancements in CAD and AI, there has not been much success in utilizing these modalities with white light endoscopy to identify polyp pathology. Komedaet al[56]applied a DL algorithm under white light endoscopy, with accuracy limited to 75.1%. Future developments include proposed combination detection systems with a DL algorithm for polyp detection under white light and a secondary algorithm that predicts the pathology of the detected polyp has been performed[57].

Limitations of AI in gastrointestinal endoscopy

Limitations of AI-based approaches in endoscopy stems from the fact that computers interpret data differently from those assessed by humans. Significant challenges of AIbased systems computer learning algorithms include: (1) AI-based systems need an extensive database for training to compare imaging findings requiring thousands of images; (2) Lack of real abstract human-like thinking in challenging cases; and (3) Existing systems are research-based mostly and suffer from considerable selection bias. Further, many of these systems are created in single institutions and serve a unique patient population, and results may not be entirely reproducible in everyday endoscopy practice. Limitations may be the result of imaging datasets, trial designs, and outcome indicators. For example, standard endoscopic images suffer from artifacts such as bubbles, mucus and stool, light exposure, and visualization angle by the endoscope camera, and the legal responsibility of misdiagnosis or incorrect diagnosis by AI-based systems. The current framework does not account for the medicolegal aspect of AI trained computers and devices in making definitive diagnosis as far as detection of gastrointestinal pathology is concerned. As the use of AI increases in our day to day life and in medicine, further clarification of the role of AI related technology in medicine specially as it pertains to medicolegal responsibility is needed as far as diagnosis and therapy related to AI based systems is concerned. The producers of AI based medical systems should seek advice regarding the liability requirements including possibility of multiple serial claims arising from errors or service disruption affecting an AI based system. In addition, similar to the use of any other medical equipment product liability would apply in relation to hardware or software malfunction. The medical practitioner would remain liable for any negligent use of the systems and need to make sure that the systems are in good working order with proper maintenance and that there are proper product liability protections in place. In essence for any AI-related medical systems to be widely adopted for diagnosis or treatment, we need to make sure the system is affordable, makes the physician’s job precise and efficient and has suitable mechanisms in place for medicolegal responsibilities.

To make the existing endoscopy based AI systems more robust, further advances in AI technology will be required to provide a reliable and accurate platform before implementation in general practice. To facilitate real-time assessment during the endoscopic examination, training, or testing CNN with unprocessed video should be used to simulate everyday endoscopy situation and make the model more realistic. In addition, the patient risk stratification should be incorporated into current CNN based models. This will further enhance the performance of AI based models specially in high-risk populations and reduce misdiagnosis in the general population. Future multicenter randomized trials are required to test various AI models in gastrointestinal endoscopy. These trials should be focused on answering the basic question whether AI models can enhance physician performance and determine the detection rate in day to day clinical settings. In the end a strong multidisciplinary collaboration among physicians, computer scientists, and entrepreneurs is required to promote AI’s clinical use in medical practice.

CONCLUSION

AI-based endoscopic systems can reliably detect and provide crucial information on gastrointestinal pathology based on their training and validation. These systems will make gastroenterology practice easier, faster, more reliable, and reduce inter-observer variability in the coming years. The thought that these systems will replace human decision making replace gastrointestinal endoscopists does not seem plausible in the near future.

Essential uses of AI in endoscopy will range from early identification of cancers using imaging enhancement with neural networks, predicting prognosis, and treatment outcomes. AI has found increasing applications in several non-medical specialties. What makes the use somewhat limited in medicine is the lack of such systems’ ability to think in a human-like manner in challenging non-algorithmic approaches and diagnostic dilemmas.

Significant advances in technology make us wonder about the day when sophisticated and powerful AI-based systems will replace gastroenterologists. Currently that does not seem to be the reality. Gastroenterologists should welcome and embrace AI-assisted technologies in their practice as and when commercially available after thorough vetting in different validation studies.

ACKNOWLEDGEMENTS

We are thankful to Mr. Parasher N for his excellent technical inputs and editing help.