Introduction
Gastrointestinal endoscopy is integral to the diagnosis and management of colorectal polyps. Optical diagnosis using advanced endoscopic technologies such as high-definition, magnification and electronic virtual chromoendoscopy permit accurate prediction of histological characteristics of colorectal lesions based on endoscopic appearances and is increasingly utilized, and its implementation across the endoscopic community is on the rise [1].
Accurate optical diagnosis allows small/diminutive colorectal polyps (< 10 mm) to be either spared or removed and discarded without the need for formal histological assessment: the “resect and discard” strategy [2]. The incorporation of optical diagnosis of small/diminutive polyps has been endorsed by The American Society of Gastrointestinal Endoscopy (ASGE) as well as recent European Society of Gastrointestinal Endoscopy (ESGE) guidelines [1]. If implemented there would be fewer specimens sent for histological analysis with substantial cost savings and reduced risk to patients with fewer unnecessary polypectomies [2]
[3].
In addition, optical characterization of colorectal polyps can accurately identify malignant areas within lesions and identify lesion borders, improving correct patient management. Sessile serrated lesions (SSL) are regarded as subtle lesions that can be easily missed; however optical diagnosis training through the use of polyp classification systems such as SIMPLE [4] and BASIC [5] may facilitate enhanced detection and characterization of SSL. Given the rise of artificial intelligence and its ability to improve detection of colorectal polyps, the technology can support polyp characterization provided endoscopists are skilled in optical diagnosis.
Training will be central to correctly implement optical diagnosis in clinical practice, as recognized in a recent evidence-based consensus [6]. Many different training strategies have been proposed and reported. Among them, traditional didactic training, computer-based self-learning and ad hoc training in vivo. This has become increasingly relevant following the coronavirus disease 2019 (COVID-19) pandemic, which has had an adverse impact on endoscopy training for trainees, particularly hands-on training, with a reduction in procedures of up to 96 % [7]. Societies are now recommending trainees utilize alternative learning opportunities such as cognitive-based learning [8]. The aim of this systematic review and meta-analysis was to provide an overview of training in optical diagnosis of colorectal polyps and in view of the COVID-19 pandemic complete subgroup analysis of computer-based training.
Methods
Methodology of our analysis, inclusion criteria and reporting were in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) recommendations [9] and the Meta-analyses Of Observational Studies in Epidemiology (MOOSE) checklist [10]. This systematic review was registered in the International Prospective Register of Systematic Reviews (PROSPERO) with registration number CRD42020167486.
Search strategy
We performed an electronic database search of PUBMED/Medline, Cochrane and SCOPUS databases in addition to gray literature (scanning reference lists), to identify studies reporting training for the optical diagnosis of colorectal polyps between 1980 and October 2019. The search strategy is outlined below:
Search strategy in PUBMED/Medline:
(Optical Diagnosis OR optical biopsy OR optical characterization) AND (colorectal polyp OR colorectal adenoma OR colorectal hyperplastic OR colorectal sessile serrated lesion OR colorectal sessile serrated adenoma), (Optical Diagnosis OR optical biopsy OR optical characterization) AND (colorectal polyp OR colorectal adenoma OR colorectal hyperplastic OR colorectal sessile serrated lesion OR colorectal sessile serrated adenoma) AND (training OR education), (Education OR Training) AND Colonoscopy AND (Colorectal polyp OR Colorectal adenoma OR Colorectal hyperplastic OR Colorectal sessile serrated lesion).
Search strategy in the other databases followed a similar but simplified strategy.
Inclusion and exclusion criteria
Studies were deemed eligible for inclusion according to the PICO statement (P, endoscopists undergoing assessment of optical diagnosis accuracy of small/diminutive colorectal polyps; I, endoscopists receiving optical diagnosis training; C, optical diagnosis as compared with histological result as gold standard; O, pre-training vs. post-training accuracy of optical diagnosis).
Studies not reporting pre-training vs. post-training performance and not published in English language were excluded. Randomized-controlled trials, observational and cohort studies and abstracts were all included for analysis.
Study selection
Titles and abstracts of all identified articles were independently screened by two authors (SS/RC) to exclude studies not related to the topic or not meeting inclusion criteria. Potentially relevant studies were screened for eligibility by analysis of the full text. Disagreements between the two authors were referred to and discussed with the senior author (MI) and resolved with consensus.
Data extraction and quality assessment
A standardized form was used to extract the data from each study. Data extracted included: a) Author name; b) Year of publication; c) Country; d) Training method (didactic, computer-based); e) Number of participants; f) Setting of study (in vivo/ex vivo); g) Number and size of polyps; h) Number of correct histology predictions (accuracy); i) Training material; j) Endoscopic platform; and k) Duration of training.
The risk of bias of included studies was assessed using the Cochrane Collaboration’s tool [11]. Each study was assessed for risk of bias through study design in selection bias (random sequence generation and methods to conceal allocation), performance bias (blinding of participants), detection bias (blinding of outcomes), attrition bias (completeness of outcome data) and reporting bias (selective reporting). The overall quality of evidence was summarized using The Grading of Recommendations Assessment, Development and Evaluation (GRADE) approach, categorizing the evidence as very low, low, moderate or high-quality.
Study definitions
A small/diminutive colorectal polyp was defined as ≤ 10 mm in size. Optical diagnosis training was defined as an intervention designed to educate participants on optical diagnosis methodology. Experienced endoscopists were participants who are independent endoscopists who have completed endoscopy training but who are not considered experts. Trainee endoscopists were defined as participants who are practicing within a designated gastroenterology training program.
Study outcomes
The primary outcome of our systematic review and meta-analysis was optical diagnosis accuracy compared with histological analysis before and after training intervention. We aimed to complete subgroup analysis by endoscopists experience (experienced, trainee endoscopists), training method (didactic, computer-based) and polyp size (small/diminutive polyps).
Statistical analysis
A random effect meta-analysis was performed to investigate the effect of training on the accuracy of optical diagnosis. For each study, accuracy was compared between post-training and pre-training stages and expressed as a risk ratio (RR) with 95 % confidence interval (CI), wwhich were pooled using a random-effects Mantel-Haenszel model. Forest plots were generated for all studies, followed by level of experience (trainee vs experienced endoscopists), and then for the subgroups of didactic vs. computer-based training. We also calculated 95 % CIs to determine the variation in effect between studies [12]
[13]. Statistical heterogeneity was assessed using I2 statistics, with a value of 0 % to 40 % accepted as not important, 30 % to 60 % as moderate, > 60 % as substantial and > 90 % as considerable heterogeneity [11].
Publication bias was assessed by observing asymmetry in funnel plots and sensitivity analyses performed by excluding outliers and then by year of publication.
Meta-analyses were performed using RevMan v5.3 (Cochrane Collaboration, Copenhagen: The Nordic Cochrane Centre, The Cochrane Collaboration, 2014) and pooled effects of didactic and computer-based training subjected to a random-effects meta-regression model using Open Meta-Analyst (Brown University). P < 0.05 was considered statistically significant.
Results
Study selection and characteristics
The PRISMA [9] and MOOSE [10] checklist (Appendix 1) and flow chart ([Fig. 1]) were followed to ensure compliance. The literature search yielded 1237 results. After preliminary screening of titles and abstracts, 113 were selected for full text review. Of these, 16 papers [4]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
[23]
[24]
[25]
[26]
[27]
[28] (before 2010 n = 1, 2010–2015 n = 7, 2016–2019 n = 8) matched the selection criteria and were included in the systematic review. Of these studies, seven were from Europe [4]
[17]
[21]
[23]
[26]
[27]
[28], six from North America [14]
[16]
[18]
[19]
[20], 24] and three from Asia [15]
[22]
[25]. Among included studies, eight reported on didactic training [4]
[14]
[15]
[16]
[18]
[22]
[23]
[27], seven on computer-based self-training [17]
[19]
[20]
[21]
[24]
[25]
[26] and one on computer-based self-training vs. didactic training [28]. The majority of studies were observational in design (n = 14) and there were two randomized trials [18]
[28]. Overall 11 studies were based on NBI system [14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
[24]
[25], one on iScan [26] and one on BLI (Blue Light Imaging) [27], two on NBI and iScan [4]
[28] and one on high-definition white light (HDWL)/chromoendoscopy [23]. Sixteen studies reported pre-training and post-training values, two studies included trainees [4]
[28], eight included fully qualified/BCSP (Bowel Cancer Screening Programme) endoscopists [14]
[15]
[18]
[19]
[20]
[21]
[22]
[25] and six studies included both groups [16]
[17]
[23]
[24]
[26]
[27]. There were nine studies that only included small/diminutive colorectal polyps [4]
[15]
[17]
[19]
[20]
[25]
[26]
[27]
[28]. Study characteristics are presented in [Table 1].
Fig. 1 PRISMA flowchart. From: Moher D, Liberati A, Tetzlaff J, Altman DG, The PRISMA Group (2009). Preferred Reporting Items for Systematic Reviews and Meta-Analyses: The PRISMA Statement. PLoS Med 6(7): e1000097. doi:10.1371/journal.pmed1000097
Table 1
Studies assessed.
Author, year
|
Study design
|
Type of training
|
Type of participant
|
Number of participants
|
Training duration
|
Training material
|
Platform
|
Number of polyps post-training
|
Rogart, 2008
|
Prospective, observational
|
Didactic
|
Experienced
|
4
|
1 hour
|
Ex vivo, Image?
|
NBI
|
20 (still images)
|
Higashi, 2010
|
Prospective, observational
|
Didactic
|
Experienced
|
4
|
1 hour
|
Ex vivo, image-based
|
NBI
|
44 (still images)
|
Raghavendra, 2010
|
Prospective, observational
|
Didactic
|
Experienced and trainee
|
25
|
20 minutes
|
Ex vivo, image-based
|
NBI
|
25 (still images)
|
Ignjatovic, 2011
|
Prospective, observational
|
Computer-based
|
Experienced and trainee
|
14
|
15 minutes
|
Ex vivo, image-based
|
NBI
|
30 (still images)
|
Coe, 2012
|
Randomized-controlled trial
|
Didactic
|
Experienced
|
15
|
2x 1-hour sessions
|
Ex vivo, images and videos
|
NBI
|
774 in total (in vivo)
|
Rastogi, 2014
|
Prospective, observational
|
Computer-based
|
Experienced
|
10
|
20 minutes
|
Ex vivo, images
|
NBI
|
40 (video format)
|
Sinh, 2015
|
Prospective, observational
|
Computer-based
|
Experienced
|
15
|
20 minutes
|
Ex vivo, image-based
|
NBI
|
40 (video format)
|
IJspeert, 2016
|
Prospective, observational
|
Computer-based
|
Experienced
|
10
|
20 minutes
|
Ex vivo, images
|
NBI
|
45 (still images)
|
Sikong, 2016
|
Prospective, observational
|
Didactic
|
Experienced
|
10
|
3 × 1-hour sessions over 3 months
|
Ex vivo, image-based
|
NBI
|
130 (still images)
|
Basford, 2017
|
Prospective, observational
|
Didactic
|
Experienced and trainee
|
10
|
30 minutes
|
Ex vivo, image-based
|
HD WLE and chromoendoscopy
|
37 (still images)
|
Aihara, 2018
|
Prospective, observational
|
Computer-based
|
Experienced and trainee
|
8
|
10 minutes
|
Ex vivo, image-based
|
NBI
|
50 (still images)
|
Iacucci, 2018
|
Prospective, observational
|
Didactic
|
Trainee
|
6
|
1 hour
|
Ex vivo, images and videos
|
iScan OE and NBI
|
80 (videos)
|
Bae, 2019
|
Prospective, observational
|
Computer-based
|
Experienced
|
15
|
30 minutes, weekly feedback and interim interactive training
|
Ex vivo, image-based
|
NBI
|
80 (still images)
|
Basford, 2019
|
Prospective, observational
|
Computer-based
|
Experienced and trainees
|
14
|
20 minutes
|
Ex vivo, image-based
|
iScan and chromoendoscopy
|
30 (still images)
|
Subramaniam, 2019
|
Prospective, observational
|
Didactic
|
Experienced and trainees
|
10
|
4 hours
|
Ex vivo, image-based
|
BLI
|
45 (still images)
|
Smith, 2019
|
Randomized-controlled trial
|
Computer-based and didactic
|
Trainees
|
16
|
1 hour
|
Ex vivo, images and videos
|
iScan OE and NBI
|
78 (videos)
|
NBI, narrow-band imaging; HD, high definition; WLE, white light endoscopy; BLI, blue light imaging; OE, optical enhancement.
Overall training efficacy and overall efficacy of training methods
When pooling together all studies, polyp sizes (16 studies, 179 participants) [4]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
[23]
[24]
[25]
[26]
[27], assessing the efficacy of any method of optical diagnosis training, pooled pre-training accuracy was 70.3 % (6416/9131 correct diagnoses) vs. post-training accuracy was 81.6 % (7416/9213 correct diagnoses) (RR 1.17 95 % CI 1.09–1.24 P < 0.001) ([Fig. 2]). The 95 % prediction interval was 0.85–1.61. There was considerable heterogeneity in these studies (I2 = 94 %) without significant publication bias (Supplementary Fig. 1a).
Fig. 2 Forest plots for all studies assessing the effect of training on accuracy of optical diagnosis of colorectal polyps.
When pooling only studies describing computer-based training (8 studies, 94 participants) [17]
[19]
[20]
[21]
[24]
[25]
[26]
[28], pooled pre-training accuracy was 69.2 % (3125/4514 correct diagnoses) vs. post-training accuracy of 80.0 % (3611/4514 correct diagnoses) (RR 1.16 95 % CI 1.09–1.23 P < 0.001) ([Fig. 3a]). The 95 % prediction interval was 0.92–1.46. We detected substantial heterogeneity in this subgroup (I2 84 %) without significant publication bias.
Fig. 3 Forest plots for studies assessing the effect of A computer-based and B didactic training on accuracy of optical diagnosis of colorectal polyps.
When pooling only studies describing didactic training (nine studies, 85 participants) [4]
[14]
[15]
[16]
[18]
[22]
[27]
[28], pooled pre-training accuracy was 71.3 % (3291/4617 correct diagnoses) vs. post-training accuracy of 83.1 % (3905/7516 correct diagnoses) (RR 1.15 95 % CI 1.03–1.29 P < 0.001) ([Fig. 3b]). The 95 % prediction interval was 0.77–1.71. We detected considerable heterogeneity in this subgroup (I2 96 %) without significant publication bias (Supplementary Fig. 1b).
On meta-regression analysis, there was no significant difference in post-training accuracy between didactic and computer-based delivery methods (P = 0.798).
Experienced endoscopists
We subsequently selected only studies describing the efficacy of training in experienced endoscopists. Method of training for experienced endoscopists comprised of didactic training (seven studies, 49 participants) [14]
[15]
[16]
[18]
[22]
[23]
[27] and computer-based training (six studies, 64 participants) [17]
[19]
[20]
[21]
[25]
[26].
After pooling these studies, optical diagnosis training method on optical diagnosis in experienced endoscopists improved accuracy from 69.8 % (3771/5403 correct diagnoses) to 82.4 % (4521/5485 correct diagnosis) (RR 1.20 95 % CI 1.11–1.29 P < 0.001). The 95 % prediction interval was 0.87–1.65 and heterogeneity was considerable (I2 92 %).
On subgroup analysis by type of training, didactic training improved optical diagnosis accuracy from 72.2 % (1685/2333 correct diagnoses) to 83.9 % (2027/2415 correct diagnoses) RR 1.19 (95 % CI 1.03–1.36 P < 0.001), and computer-based training from 67.9 % (2086/3070 correct diagnoses) to 81.2 % (2494/3070 correct diagnoses) RR 1.21 (95 % CI 1.13–1.30 P < 0.001). The improvement was not significantly different between the two training methods (P = 0.792) ([Fig. 4a]).
Fig. 4a Forest plots for studies assessing the impact of training on the accuracy optical diagnosis of colorectal polyps in experienced endoscopists.
Fig. 4b Forest plots for studies assessing the impact of training on the accuracy optical diagnosis of colorectal polyps in trainees.
Trainee endoscopists
A total of eight studies [4]
[16]
[17]
[23]
[24]
[26]
[27]
[28] evaluated impact of training among in 58 trainee endoscopists. In detail, four studies [4]
[16]
[23]
[27] with 28 participants used didactic training, and three studies [17]
[24]
[26] with 22 participants evaluated computer-based training. Only one study compared didactic vs computer-based training [28].
After aggregating these studies involving trainee participants, optical diagnosis training improved accuracy from 69.6 % (2645/3803 correct diagnoses) to 78.8 % (2995/3803 correct diagnoses) (RR 1.14; 95 % CI 1.06–1.24, P < 0.001). The 95 % prediction interval was 0.90–1.44. Once again, we detected considerable heterogeneity (I2 89 %). On subgroup analysis didactic training improved accuracy from 68.1 % (1606/2359 correct diagnoses) to 79.6 % (1878/2359 correct diagnoses) (RR 1.18; 95 % CI 1.04–1.34, P < 0.001) and computer-based training from 72.0 % (1039/1444 correct diagnoses) to 77.4 % (1117/1444 correct diagnoses) (RR 1.09; 95 % CI 1.03–1.15, P < 0.001) ([Fig. 4b]). The improvement in accuracy did not differ significantly between the two training methods (P = 0.312).
Small/diminutive colorectal polyps
We selected studies that only included small/diminutive colorectal polyps in the assessment of training in optical diagnosis. A total of nine studies assessed the impact of optical diagnosis training on 104 endoscopists [4]
[15]
[17]
[19]
[20]
[25]
[26]
[27]
[28]. The pooled pre-training accuracy was 68.1 % (3549/5214) and post-training accuracy was 77.1 % (4022/5214) (RR 1.16 95 % CI 1.08–1.24 P < 0.001) (Supplementary Fig. 2). We detected substantial heterogeneity in these studies (I2 87 %).
Study quality and risk of bias
The summary of the Cochrane Collaborationʼs risk of bias tool is presented in [Fig. 5]. Only two randomized trials were included [18]
[28], both of which report how the randomization process took place but both were unable to conceal the allocation to participants. Participants in all studies included were not blinded to the intervention. Participants were blinded to the outcomes of training during the study process for most studies and all studies produced complete datasets without selective reporting.
Fig. 5 Risk of bias of studies included using Cochrane Collaboration’s risk of bias tool
Using the GRADE approach, the quality of evidence was downgraded by two points to low quality due to risk of bias and inconsistency from the heterogeneity in the studies.
Discussion
According to our systematic review and meta-analysis, training in the optical diagnosis of colorectal polyps was associated with improved diagnostic accuracy in both experienced and trainee endoscopists. Furthermore, when only considering small/diminutive colorectal polyps there was also a statistically significant improvement in diagnostic accuracy following training. These results are clinically relevant and important for the following considerations. First, the efficacy of training has been proved for both experienced and trainees in endoscopy. Indeed, trainees showed a similar post-training accuracy level as compared to experienced endoscopists, confirming the positive impact of training on optical diagnosis. Second, irrespective of the pre-training accuracy level, training programs resulted in uniformly elevated accuracy levels. Third, there was no statistically significant difference seen in the improvement in accuracy between computer-based and didactic training.
Endoscopy practice is going through major changes, which have accelerated in response to the COVID-19 pandemic. To minimize the risk of exposure to COVID-19, units are taking steps to lessen the footfall in procedural rooms, [29]. This will inevitably have a serious consequence on endoscopy training, particularly given that restrictions may be in place until 2022 [30]
[31]. It is recognized that during these unprecedented times that trainees seek training in alternative means, perhaps focusing on cognitive skills [30]. This increased importance in alternative educational resources such as computer-based and simulation based training in optical diagnosis would address the unmet need during this period and likely in the longer term too [7].
Our study confirms that computer-based training is effective in improving optical diagnosis accuracy and furthermore, there is no statistical difference between computer-based and didactic training which is further confirmed by a randomized trial [28]. Although the P values and lower bound 95 % CIs of our analyses indicate a statistically significant beneficial effect of training, the 95 % prediction intervals suggest that, while the “average” training course is likely to result in an improvement in predictive accuracy after optical diagnosis training, there are likely to be courses where no benefit is observed. This is also reflected in the I
2 values which indicate considerable heterogeneity in most analyses. As such there needs to be a focus on the validation, standardization and quality assurance of training and long-term studies to assess the retention of optical diagnosis skills after a training intervention.
Most studies investigating the effectiveness of training methods on the optical diagnosis of colorectal polyps were small observational studies with only two randomized trials. The studies included a baseline level of performance prior to training enabling us to fully ascertain the effectiveness of training. Including trainee and experienced endoscopists, the target of intervention was relevant to the clinical practice of endoscopy. The substantial heterogeneity of studies may be explained by the following differences: training intervention (number of sessions and duration), assessment methods (type of media used, endoscopic platform and whether material was repeated pre- and post-training), type of lesions included (some studies only included hyperplastic and adenomas [16]
[17]
[27]), washout period between pre-training assessment and training, and participant characteristics (pre-training accuracy levels were as low as 45 % in one study among experienced endoscopists [16] and as high as 88 % in some trainees [27]).
The systematic review and meta-analysis has limitations. First, all included studies were unblinded due to the nature of the intervention (i. e. training methods). This however is a common bias of many evaluation of technological improvements in gastrointestinal endoscopy. Second, participants were aware that their performance in assessments were being monitored and analyzed, which may introduce further bias as there may have been a subconscious increased effort during the study period. One possible solution to this would be tracking in vivo optical diagnosis accuracy for a prolonged period of time before and after training. However, this would be resource intensive and may still incur bias. Third, not all studies provided data to calculate sensitivity, specificity or negative predictive values (against the recommended ASGE PIVI threshold of > 90 %), nor was it possible to compare receiver operating characteristics curves pre-training and post-training. Given the large degree of heterogeneity, the inconsistency of results, the lack of randomized controlled trials and inherent bias the quality of evidence of the studies is considered to be low.
This systematic review and meta-analysis has uncovered important gaps in evidence, which will need to be addressed in future studies. Further study on optical diagnosis training is required to establish the optimum method. There is a lack of robust randomized trials comparing training modes with only one comparing didactic with computer-based training [28]. Another important consideration is the effectiveness of training across endoscopic platforms, while initial attempts at this suggest that training is effective across NBI and iScan [4]
[28], however there is yet to be a study that includes all of NBI, iScan and BLI. The issue of retention of optical diagnosis over time is one that has yet to be thoroughly explored. The majority of studies included in this systematic review and meta-analysis involve participants scoring media on the same day as training, however none have demonstrated that optical diagnosis skills can be retained over a prolonged period of time. This is essential for optical diagnosis to be incorporated in everyday practice. Future studies should reassess participants after a prolonged period of time such as 3 to 6 months or even longer and should include an in vivo assessment component to formally investigate whether ex vivo training can translate to in vivo performance over time.
A future training strategy may include a combination of methods with perhaps didactic training from experts to set a foundation of knowledge, further supported by computer-based training or artificial intelligence (AI) systems. Computer-based training could be periodically reinforced relatively easily. The ESGE suggest the most likely scenario will be as a “second reader” as opposed to a “stand alone” system [32]. Therefore, for AI to be utilized in everyday practice there is a need for endoscopists to up skill in optical diagnosis to be able to safely interpret and act upon the readings from AI systems.
Conclusions
In conclusion, training in optical diagnosis improves the accuracy of histology predictions of colorectal polyps, including small/diminutive polyps. The optimal method of training may include a combination of training methods augmented with continuous in vivo training, which may be provided by trainers or AI systems. Future studies need to focus on standardizing and validating training modules to enhance cognitive skills of endoscopists.