Next Article in Journal
Evaluation of 8-Channel Radiative Antenna Arrays for Human Head Imaging at 10.5 Tesla
Previous Article in Journal
SAR.IoT: Secured Augmented Reality for IoT Devices Management
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Colorectal Polyp Image Detection and Classification through Grayscale Images and Deep Learning

1
Department of Gastroenterology and Hepatology, Linkou Chang Gung Memorial Hospital and Chang Gung University College of Medicine, No. 5, Fuxing St., Guishan Dist., Taoyuan City 333, Taiwan
2
Department of Computer Science and Information Engineering, Fu-Jen Catholic University, 510 Chung Cheng Rd., Hsinchuang Dist., New Taipei City 242, Taiwan
3
Graduate Institute of Applied Science and Engineering, Fu-Jen Catholic University, 510 Chung Cheng Rd., Hsinchuang Dist., New Taipei City 242, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(18), 5995; https://doi.org/10.3390/s21185995
Submission received: 13 July 2021 / Revised: 29 August 2021 / Accepted: 30 August 2021 / Published: 7 September 2021
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Colonoscopy screening and colonoscopic polypectomy can decrease the incidence and mortality rate of colorectal cancer (CRC). The adenoma detection rate and accuracy of diagnosis of colorectal polyp which vary in different experienced endoscopists have impact on the colonoscopy protection effect of CRC. The work proposed a colorectal polyp image detection and classification system through grayscale images and deep learning. The system collected the data of CVC-Clinic and 1000 colorectal polyp images of Linkou Chang Gung Medical Hospital. The red-green-blue (RGB) images were transformed to 0 to 255 grayscale images. Polyp detection and classification were performed by convolutional neural network (CNN) model. Data for polyp detection was divided into five groups and tested by 5-fold validation. The accuracy of polyp detection was 95.1% for grayscale images which is higher than 94.1% for RGB and narrow-band images. The diagnostic accuracy, precision and recall rates were 82.8%, 82.5% and 95.2% for narrow-band images, respectively. The experimental results show that grayscale images achieve an equivalent or even higher accuracy of polyp detection than RGB images for lightweight computation. It is also found that the accuracy of polyp detection and classification is dramatically decrease when the size of polyp images small than 1600 pixels. It is recommended that clinicians could adjust the distance between the lens and polyps appropriately to enhance the system performance when conducting computer-assisted colorectal polyp analysis.

1. Introduction

Colorectal cancer is globally the third most prevalent cancer and mainly derived from pre-malignant colorectal polyps [1]. Colonoscopy screening and endoscopic resection of colorectal polyps can reduce the incidence and mortality of colorectal cancer [2]. Image-enhanced colonoscopy with contrast enhancement through dyes or optical strategies facilitates the detection and observation of colorectal lesions [3]. However, the accuracy of the detection and diagnosis of colorectal polyps varies depending on the experience of endoscopists, thus influencing the effectiveness of colonoscopy in reducing the incidence and mortality of colorectal cancer [3,4].
For detailed observation of the colorectal lesions, endoscopists can apply dyes and/or virtual chromoendoscopy [5]. Colorectal lesions are stained with dyes such as indigo carmine, crystal violet or methyl blue and are observed by colonoscopy with or without magnification. Virtual chromoendoscopy is an image-enhanced endoscopic technique based on optical strategies for capturing images. Currently, the most commonly used virtual chromoendoscopy systems in clinical practice are narrow-band imaging (NBI), flexible spectral imaging color enhancement (FICE), i-scan, and blue laser imaging (BLI) [5,6,7]. In NBI, optical filters are used to filter out red light and narrow the bandwidth to a range in which only blue and green light are visible [6]. Hemoglobin absorbs blue light, with a wavelength of 400–430 nm, to highlight the morphology of the surface capillaries and green light, with a wavelength of 525–555 nm, to penetrate and highlight deeper blood vessels. FICE and i-scan enable image capture for digital processing. In these methods, lesion images are enhanced and highlighted without being darkened like those in NBI. In the BLI system, lasers with a wavelength of 450 nm are used to irradiate phosphor and generate white light (WL), and blue lasers with a wavelength of 410 nm are employed for narrow-band image observations [7]. Blue light is absorbed by hemoglobin, thereby highlighting the morphology of the surface capillaries, and WL increases the brightness of the images to facilitate observations. Because the property of increased vascularity and change of blood vessel diameter in premalignant and malignant colorectal polyps, NBI is capable of identifying these lesions and predict the depth of invasion. Virtual chromoendoscopy has the advantage of rapid image conversion, without the influence of a time-consuming dyeing process and uneven dyeing. Virtual chromoendoscopy also prevents discomfort or cytotoxicity to the examinees caused by the dyes [6].
Currently, common magnifying NBI diagnostic classifications are the Sano, Hiroshima, Showa, and Japan narrow-band imaging expert team (JNET) classifications, and the most prevalent non-magnifying NBI diagnostic classification is the NBI international colorectal endoscopic (NICE) classification (Table 1 and Figure 1) [8,9]. In most of these classifications, the histopathology of polyps is determined based on the optical characteristics of lesion color, structure of capillaries, and morphology of capillary surface. The NICE classification, for instance, categorizes colorectal polyps as type 1, hyperplastic polyps; type 2, adenomas, intramucosal carcinoma, or superficial submucosal carcinoma; and type 3, deep submucosal carcinoma. Rectosigmoid diminutive (≤5 mm) hyperplastic polyps are considered as non-neoplastic polyps without malignant potential [8]. The recently developed Workgroup serrAted polypS and Polyposis classification can enhance the diagnostic accuracy of sessile serrated adenoma polyps [10].
In 2011, the American Society of Gastrointestinal Endoscopy (ASGE) proposed a guideline for colorectal polyp diagnosis through a new optical endoscopic technique [11]. For diminutive polyps of the sigmoid colon and the rectum, the diagnose-and-leave or resect-and-discard strategy can be adopted if the correct diagnosis is made based on real-time endoscopic optics. The diagnose-and-leave strategy is applicable when the endoscopic optical diagnosis is highly precise (i.e., a negative predictive value of 90% and above for adenoma). Pathological biopsy is not required for such polyps. When the resect-and-discard strategy is adopted to remove diminutive colorectal polyps identified by high-confidence optical diagnosis, clinicians must combine optical diagnosis of rectosigmoid diminutive polyps with histopathological diagnosis for colorectal polyps >5 mm to determine the post-polypectomy surveillance interval. All colorectal polyps are sent for pathologic interpretation to ensure a consistency of 90%. Cost-effectiveness analyses have revealed that employing the innovative endoscopic optical technology to diagnose colorectal polyps and adopting either the diagnosis-and-leave or resect-and-discard strategy are cost-effective [12,13,14,15,16,17]. Nevertheless, the current challenge in the use of innovative endoscopic optical technology to diagnose colorectal polyps is that not all endoscopists meet the standards recommended by the ASGE. In fact, of the endoscopists in community hospitals, only 25% could achieved a colorectal polyp diagnosis accuracy of 90% by using the NBI-assisted optical diagnosis method [18], with the consistency rate of postoperative polyp resection monitoring intervals being only 85% and 81% [4,19]. Therefore, identifying the effective use of computer-aided diagnostic systems in helping clinicians decide the polyp treatment methods is imperative.
Different computer-aided diagnostic systems for colonoscopy have been proposed [20,21,22]. The color, shape, texture, and spatial features of the images are analyzed through image processing techniques (e.g., noise removal, image sharpening, and image segmentation) and according to the optical characteristics captured using virtual chromoendoscopy (i.e., color of colorectal polyps, structure of capillaries, and morphology of capillary surface). The images are further classified or identified using grouping algorithms, and an identification accuracy and sensitivity of approximately 72–89% [4] and 90% [23], respectively. Through texture analysis and wavelet transformation, Hafner and Kwitt [23,24] computed the red-green-blue (RGB) and LAB models by using a color wavelet cross-cooccurrence matrix and analyzed the polyp classifications by applying K-nearest clustering and feature vectors. Gross et al. [25] employed nonlinear filters and Canny edge detectors for tumor contouring. Tamaki [26] extracted image feature vectors by using the scale-invariant feature transform and performed identification using support vector machine classification. However, determining a standard or single feature vector to classify different types of polyp images by using the said methods is difficult.
In most computer-aided diagnostic systems, analysis is performed based on the magnification of endoscopic images. However, in clinical practice, non-magnified high-resolution colonoscopy images are most commonly used mainly because magnification endoscopes are more expensive, difficult to operate, and unsuitable for routine practice. Therefore, they are not preferred by general hospitals and clinicians and only used in medical centers for clinical research purposes. Several studies have proposed the application of deep learning to computer-aided diagnostic systems for colonoscopy, published data that conducted based on magnification endoscopic images yielded a diagnostic accuracy over 90% [21,27]. However, studies using non-magnification colonoscopic images for colorectal polyps classified by deep learning system is lacking.
Deep learning has been increasing used in the medical field [28,29,30,31,32,33]. Many studies have applied a convolutional neural network (CNN) to the identification and classification of colorectal polyps [27,34]. Chen et al. [35] employed the deep neural network to differentiate between NBI images of neoplastic and hyperplastic polyps. Park et al. [36] used the CNN to classify the polyps in colonoscopy images. Ribeiro et al. [37] increased data sets through small patches and conducted polyp classification according to the features extracted using a CNN. Shin et al. [38] proposed the false-positive and offline learning method based on a region-based CNN (R-CNN), successfully improving the existing polyp locating method. By using Faster R-CNN, Ren et al. [39] developed a learning method by drawing a bounding box around a polyp. On the basis of the Faster R-CNN of the visual geometry group-16 model, Mo et al. [40] trained 16 colonoscopy images. Wang et al. [41] employed the SegNet model for examining colonoscopy images. Urban et al. [42] used VGG16 and VGG19 models to conduct analyses through 7-fold validation. Zheng et al. [43] used the you-only-look-once (YOLO) model to identify the polyp location in WL and NBI images. All these methods were adopted to locate polyps and assist clinicians for the classification of polyps. In most of these methods, training and testing were based on RGB images. However, because RGB images, expressed by three primary color layers, are multidimensional images, the computation cost involved is higher. In addition, the network frameworks were composed of a considerable number of layers, which makes the process time-consuming.
Currently, many medical images are processed into grayscale images, such as ultrasound, computed tomography (CT), and magnetic resonance (MR) images. Tan et al. [44] applied a gray-level cooccurrence matrix and CNN to CT images for polyp diagnosis. Zhang et al. [45] compress the three-channel color images of chest CT to grayscale images. A five-layer deep CNN with stochastic pooling is used to diagnose chest-based COVID-19. Xie et al. [46] mentioned that in deep learning, colors are not the key features influencing accurate image classification. They also discovered that in X-ray image classification, the speed and accuracy of processing grayscale images were considerably higher than those of processing RGB images. The grayscale method used was ITU-R Recommendation BT.601 [47]. Moreover, misjudgments were easily made in the presence of excessive intestinal wall textures, when polyps were too small (Figure 2), and when polyp textures were similar (Figure 3) to intestinal wall textures.
In this study, we proposed a colorectal polyp detection and classification method by using grayscale images and a lightweight deep learning framework, which provided the same or similar detection effect as RGB images but incurred lower computing costs. To increase the applicability of the proposed method in clinical diagnosis, we developed a precise, computer-aided diagnostic system compatible with mobile devices.
The rest of this study is organized as follows: Section 2 explains the materials and methods used in the colorectal polyp image detection and classification system; Section 3 details the experiments performed on the system, and Section 4 presents the discussion and conclusions.

2. Materials and Methods

In this study, we employed colonoscopy images provided by the CVC Clinic [48,49] and the Department of Gastroenterology and Hepatology, Chang Gung Medical Hospital (CGMH) (Table 2). The CVC Clinic data set comprises 612 continuous images extracted from 29 WL images, among which 592 images had one polyp and the remaining 30 images had more than two polyps. The image resolution was 384 × 288 pixels. For image marking, the upper, lower, left, and right boundaries of the white part were set as the four sides of the bounding box according to the binarized ground truth (GT) obtained from the CVC Clinic data set. The endoscope models used were CF-260AI/AL and CF-290AI/AL (Olympus Optical Co, Ltd., Tokyo, Japan), and the endoscope host employed was EVIS LUCERA (Olympus Medical Systems) (Figure 4 and Figure 5). In both WL and NBI images, the parts with polyps were extracted. In all, 1000 polyp image data were captured, and 2160 image data were extracted from the experimental data, including 80 images of the same polyp from different angles. First, we selected one WL and one NBI image each from the 900 WL and 900 NBI colonoscopy images, respectively (1800 colonoscopy images in total). The image resolution was 640 × 480 pixels. Each image in the data set was manually marked by clinicians, and the file format used was tagged image file format (TIFF). In addition, we selected another 360 polyp images for a polyp classification test to improve classification accuracy.
Figure 6 presents the architecture for the colorectal polyp detection and diagnosis system. The system is capable of polyp detection and classification. During colorectal polyp detection, the input colonoscopy images were preprocessed. The images in the data set were converted to grayscale images during preprocessing and applied for feature extraction and training to detect the location of polyps. The detected polyps were subsequently classified into neoplastic polyps (including adenomas and deep submucosal invasive cancer) and hyperplastic polyps.

2.1. Polyp Detection

The main polyp detection procedures involve the conversion of images to grayscale and labeling the location of polyps. During gray scaling, RGB images were converted to grayscale images with pixel values ranging from 0 to 255. RGB pixels were converted to gray pixels by using ITU-R BT.601 [47]:
Gray = 0.229 × R + 0.587 × G + 0.117 × B
where R, G, and B are the pixel values of the red, green, and blue colors in color images, respectively (Figure 7 and Figure 8).
Furthermore, we adopted a CNN model for feature extraction (Figure 9). The analytical process included convolution, batch normalization (BN), rectified linear unit (ReLU) function implementation, and max pooling operations. The input colonoscopy images were scaled to 128 × 128 pixels, and a convolution filter was used for feature map extraction. The mean and standard deviation of the feature map were calculated in batches to normalize each channel of the input feature map. Subsequently, two trainable parameters, normalized through scaling and translation, were added to the model to effectively transfer the extracted feature map to the next layer, thereby increasing the training speed and enhancing the network stability. Threshold values of the normalized feature map that were less than 0 were set to 0. Through the filter, the maximum value of the feature map was obtained, thus highlighting the features:
Feature   map ( x , y ) = p q I ( p , q ) K ( x p + 1 , y q + 1 )
where I is the input colonoscopy image; K denotes the filter; p and q are the horizontal and vertical axis coordinates of the original image, respectively; and x and y are the horizontal and vertical axis coordinates of the output feature map, respectively:
x i ^ = x i μ B σ B 2 + ε
y i = γ x i ^ + β B N γ , β ( x i )
where x i ^ is the normalized output value; xi is the input value; μB is the mean of the batches; σB2 is the standard deviation of the batches; ε is the minimum value and ε = 10−5 in this study; yi is the output value of the normalized value; γ is the scaling parameter; and β is the translation parameter. The ReLU function performs threshold operations on the features in the upper layer, changing all values less than 0 to 0, which can clarify the features and facilitate network training. Threshold values of the normalized feature map that were less than 0 were set to 0:
R e L U ( B N y ) = {   B N y , B N y 0     0 , B N y < 0  
where BNy is the output value after BN, namely the input value of the ReLU function; The max pooling operation performs dimension reduction on the features to extract their maximum value. Through the filter, the maximum value of the feature map was obtained, thus highlighting the features:
a j = m a x p ( a i n × n )
where ai is the input value; p is the pooling regions; n is the pooling size; and aj is the output value after the maximum value is obtained. Polyp detection was then conducted in the bounding box transform layer. The prediction result of the feature map was obtained in the last layer and was output through a 1 × 1 convolutional layer by using a pixel-based approach and converted to a bounding box mode for display [50]. There are eight groups of bounding boxes to locate the polyps: (16, 16), (32, 32), (24, 48), (48, 24), (60, 80), (108, 72), (216, 144), and (180, 180). The cutoff value of polyp detection is 0.2. The size of the predicted bounding box was adjusted by converting it to the range of a true-value bounding box. Table 3 displays the framework of the CNN model. Figure 10 and Figure 11 show the results of WL and NBI polyp detection, respectively, where the first column is the RGB image; the second column is the grayscale image; the red box is the prediction box; and the green box is the actual polyp location marked by the clinician, namely the GT.

2.2. Polyp Classification

According to the biopsy reports, we classified colorectal polyps into neoplastic polyps and hyperplastic polyps by using a CNN model (Table 4). The cutoff value of polyp classification is 0.5.

2.3. Statistical Analysis

Categorical data were presented as rates and proportions and 95% confidence intervals (CI) were calculated. The McNemar’s test was used to compare the accuracy of polyp detection between RGB and NBI images and grayscale images for deep learning system. A two-tailed p value less than 0.05 is considered statistically significant. The statistical analysis was performed by using IBM SPSS Statistics 22 (SPSS Inc., Chicago, IL, USA).

3. Experimental Results

We divided the experiment data into three parts: the CVC Clinic images and the WL images and NBI images provided by the Chang Gung Medical Foundation. The images were divided based on data sets and different light sources to verify whether the dissimilarity and compatibility of images differed due to different light sources. The RGB and grayscale images in each data set were separately trained. This system was implemented using software, i.e., MATLAB version 2020b (The MathWorks, Inc., Natick, MA, USA).

3.1. Polyp Detection Results

The data sets for polyp detection were divided into five groups (Table 5), and the data were input to the colorectal polyp detection system through a 5-fold cross validation for training and testing. During testing, calculations were performed based on the marked-up parts and the prediction box and according to the standard used in the 2015 endoscopic vision challenge held by the Medical Image Computing and Computer-Assisted Intervention Society (MICCAI-2015) [30,51]. Table 6, Table 7, Table 8, Table 9, Table 10, Table 11, Table 12, Table 13 and Table 14 presented the detection accuracy of our models for the tested data of the CGMH-WL, CGMH-NBI, and CGMH-WL+ CGMH-NBI images. The accuracy value of polyp detection is the number of correct detected polyp images divided by the total number of polyp images. The results indicated that the detection accuracy of grayscale images reached almost 95%, which was higher than of the RGB images when using WL image dataset (Table 6 and Table 7). However, there is no difference in the accuracy of polyp detection between NBI images and grayscale transformation images when using NBI image dataset (Table 8 and Table 9). Among RGB images, detection errors were noted in situations where polyp protrusions were nonobvious or had a texture similar to that of the intestinal wall, images had excessive noise, or the polyps were too small. Particularly, polyps with a size less than 1600 pixels (40 × 40 pixels) in the image were likely to be misjudged. Moreover, we conducted a deep learning for detection of polyp size divided into greater and smaller/equal than 0.5 cm (Table 12). The identification result of the gray scaled CNN exhibited no significant difference. The accuracy of the images with a polyp size both greater and smaller than 0.5 cm ranged between 94.4% and 96.5% (Table 13 and Table 14). That is, image size was the key factor affecting the detection accuracy. We used 900 normal colonoscopy images as the testing data to verify the performance of polyp detection. The accuracy of normal colonoscopy images detection was 870/900 = 96.56%. We found that some RGB polyp images that were not fully focused were likely to be undetected if the images were not transformed to grayscale. It is a common phenomenon in clinicians during colonoscopy screening and usually happened in the dynamic colonoscopy images. In addition, some unfocused NBI images may also be the possible cause of misjudgment for polyp detection.

3.2. Polyp Classification Results

We classified 2072 WL and NBI colorectal polyp images provided by the Chang Gung Medical Hospital according to the results of pathological interpretation. The dataset was randomly split into training, validation, and testing group. None of the image shared between groups. The number of polyp images used for classification training, verification, and testing was 1370, 342, and 360, respectively (Table 15). The verification and testing data comprised the images that were accurately detected, and polyp images different from the preceding ones were used as the verification data. The 5-fold cross validation was also used for training, testing, and validation of polyp classification. Table 16 presents the performance index in which true positive (TP) is the situation where the center point of the prediction box falls at any GT; false positive (FP) is the situation where the center point of the prediction box does not fall at any GT; and false negative (FN) is the situation where no prediction box appears. True negative (TN) was not included because all the images in the data set had at least one polyp, implying that they were not negative images. Table 17 and Table 18 present the confusion matrices for the classification of WL and NBI grayscale images, respectively. Table 19 presents the performance evaluation of polyp classification. Figure 12 shows the ROC curve and the AUC value is 0.93. The results indicated that NBI images facilitated more accurate polyp identification than WL images. During polyp classification, neoplastic polyps were likely to be misjudged as hyperplastic polyps when polyp textures were not obvious, polyps were smaller than 1800 pixels, or when the images had an excessively low resolution. However, hyperplastic polyps were likely to be misjudged as neoplastic polyps because of the WL around polyps in the images and when polyps were smaller than 1800 pixels.
In Table 18, three neoplastic polyp images that were less than 1800 pixels were classified as hyperplastic polyps. The NPV for adenoma by using the proposed system is 84% (31/37), which did not reach the threshold of ≥90% for implementation of the diagnose-and-leave strategy. However, if the polyp images less than 1800 pixels were excluded from the study, the NPV is 91% (31/34), which achieved the threshold set by ASGE for diagnose-and-leave policy.

4. Discussion and Conclusions

For polyp detection, the mean accuracy of grayscale images was 95.1%, which was higher than that of RGB/NBI images (94.1%). Detection errors in RGB images were caused when polyp protrusions were nonobvious or had a texture similar to that of the intestinal wall (Figure 13), images had excessive noise (Figure 14), or the polyps had an excessively small pixel size (Figure 15). Particularly, polyps with a size less than 1600 pixels (40 × 40 pixels) in the image were likely to be misjudged. Therefore, we recommend that in polyp detection, the length and width of polyp images must be at least 40 pixels in the lens regardless of the actual size of the polyps. This is likely to reduce misjudgment. We also found that when the colorectal polyps image is more complex, such as with more wrinkles or light spots, the use of color image deep learning for polyp detection may be more accurate. In addition, when the vascular lines of colorectal polyps were less obvious, such as unfocused or residual images, it would be better to use gray-scale image deep learning for polyp detection. Next step we plan to use a hybrid deep learning model that combines color and grayscale images for polyp detection in the dynamic colonoscopy videos. Maybe it can take the advantages of both methods as well as improve the detection accuracy.
In polyp classification, neoplastic polyps were likely to be misjudged as hyperplastic polyps when polyp textures were not obvious (Figure 16), images had an excessively low resolution, or when polyps were smaller than 1800 pixels (Figure 17). Hyperplastic polyps, however, were likely to be misjudged as neoplastic polyps because of the WL around polyps in the images and when polyps were smaller than 1800 pixels.
In this study, we employed a deep learning framework for colorectal polyp detection by using grayscale images. The marked polyp data in the public database of the CVC Clinic and the clinical data provided by the Chang Gung Medical Hospital were collected for verification and testing. Rather than only using an analytic method for single-source data, we collected colonoscopy images captured by different brands and models of colonoscopes. This study had the following characteristics: first, the use of grayscaled RGB images facilitated the computation of the deep learning model. Second, in the deep learning framework, we used three convolutional filters (3 × 3, 1 × 3, and 3 × 1) in the convolutional layer to extract features. Different from conventional CNN models where single-feature filters were used, the proposed approach was capable of extracting more features for computation. We also compared the result with those obtained using other models and confirmed that the proposed approach could effectively increase the polyp detection accuracy. Third, grayscale images could reduce the dimensions and size of RGB images. Image deviation and noise caused by the difference in image gradient and endoscope lens could also be avoided, which increases the accuracy of polyp detection and classification. The results of the experiment indicate that in the detection of diminutive polyps—those smaller than 0.5 cm—the accuracy considerably decreased when polyps were smaller than 1600 pixels. The differentiation between neoplastic polyps and hyperplastic polyps became difficult when polyps were smaller than 1800 pixels. Hence, we recommend that clinicians could adjust the distance between the lens and polyps to enhance the detection results when conducting computer-assisted colorectal polyp analysis, which ensures that the polyps in the images have a minimum size of 1600 pixels.
Grayscale images achieve an equivalent or even higher accuracy of polyp detection than RGB/NBI images for lightweight computation. When NBI images were larger than 1800 pixels, our model could achieve high NPV (>90%) for neoplastic polyps. The colorectal polyp detection technique developed in this study can help clinicians rapidly locate polyps, whereas the colorectal polyp classification technique can assist clinicians in determining whether polypectomy or surgery is required, which may effectively improve patient outcome, avoid unnecessary polypectomy, and reduce misdiagnosis rate of endoscopists. In addition, hardware equipment is no longer limited by computation costs. The incorporation of simple mobile devices for edge computing and telemedicine services through lightweight computation and the hand-held colorectal polyp detection technique based on 5G wireless communication technology can substantially reduce the cost of medical equipment and increase medical standards.

Author Contributions

Conceptualization, C.-M.H. and C.-C.H.; methodology, C.-C.H., Z.-M.H. and F.-Y.S.; writing—original draft preparation, C.-M.H. and C.-C.H.; writing—review and editing, C.-M.H., C.-C.H. and M.-L.C.; funding acquisition, C.-M.H. and T.-H.C. All authors have read and agreed to the published version of the manuscript.

Funding

Grant supporting this work was obtained from the Chang Gung Medical Research Program (CMRPG3H1541). The funders had no role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board of Chang Gung Medical Foundation (protocol code 201801533B0 and date of approval 16 October 2018).

Informed Consent Statement

Not applicable.

Acknowledgments

We are very grateful to the reviewers for their valuable comments that helped to improve the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global Cancer Statistics 2020: GLO-BOCAN Estimates of Incidence and Mortality Worldwide for 36 Cancers in 185 Countries. CA Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef]
  2. Zauber, A.G.; Winawer, S.J.; O’Brien, M.J.; Ballegooijen, M.; Hankey, B.F. Colonoscopic polypectomy and long-term pre-vention of colorectal-cancer deaths. N. Engl. J. Med. 2012, 366, 687–696. [Google Scholar] [CrossRef] [PubMed]
  3. Suzuki, H.; Yamamura, T.; Nakamura, M.; Hsu, C.M.; Su, M.Y.; Chen, T.H.; Chiu, C.T.; Hirooka, Y.; Goto, H. An International Study on the Diagnostic Accuracy of the Japan Narrow-Band Imaging Expert Team Classification for Colorectal Polyps Ob-served with Blue Laser Imaging. Digestion 2020, 101, 339–346. [Google Scholar] [CrossRef]
  4. Kuiper, T.; Marsman, W.A.; Jansen, J.M.; van Soest, E.J.; Haan, Y.C.; Bakker, G.J.; Fockens, P.; Dekker, E. Accuracy for Optical Diagnosis of Small Colorectal Polyps in Nonacademic Settings. Clin. Gastroenterol. Hepatol. 2012, 10, 1016–1020. [Google Scholar] [CrossRef]
  5. Subramanian, V.; Ragunath, K. Advanced Endoscopic Imaging: A Review of Commercially Available Technologies. Clin. Gastroenterol. Hepatol. 2014, 12, 368–376.e1. [Google Scholar] [CrossRef]
  6. Kaltenbach, T.; Sano, Y.; Friedland, S.; Soetikno, R.; American Gastroenterological, A. American Gastroenterological Association (AGA) Institute technology assessment on image-enhanced endoscopy. Gastroenterology 2008, 134, 327–340. [Google Scholar] [CrossRef]
  7. Yoshida, N.; Hisabe, T.; Inada, Y.; Kugai, M.; Yagi, N.; Hirai, F.; Yao, K.; Matsui, T.; Iwashita, A.; Kato, M.; et al. The ability of a novel blue laser imaging system for the diagnosis of invasion depth of colorectal neoplasms. J. Gastroenterol. 2013, 49, 73–80. [Google Scholar] [CrossRef]
  8. Hewett, D.; Kaltenbach, T.; Sano, Y.; Tanaka, S.; Saunders, B.P.; Ponchon, T.; Soetikno, R.; Rex, D.K. Validation of a Simple Classification System for Endoscopic Diagnosis of Small Colorectal Polyps Using Narrow-Band Imaging. Gastroenterology 2012, 143, 599–607.e1. [Google Scholar] [CrossRef]
  9. Hayashi, N.; Tanaka, S.; Hewett, D.G.; Kaltenbach, T.R.; Sano, Y.; Ponchon, T.; Saunders, B.P.; Rex, D.K.; Soetikno, R.M. Endoscopic prediction of deep submucosal invasive carcinoma: Validation of the narrow-band imaging international colo-rectal endoscopic (NICE) classification. Gastrointest. Endosc. 2013, 78, 625–632. [Google Scholar] [CrossRef]
  10. IJspeert, J.E.; Bastiaansen, B.A.; van Leerdam, M.E.; Meijer, G.A.; van Eeden, S.; Sanduleanu, S.; Schoon, E.J.; Bisseling, T.M.; Spaander, M.C.; van Lelyveld, N.; et al. Development and validation of the WASP classification system for optical diagnosis of adenomas, hyperplastic polyps and sessile serrated adenomas/polyps. Gut 2016, 65, 963–970. [Google Scholar] [CrossRef]
  11. Rex, D.K.; Kahi, C.; O’Brien, M.J.; Levin, T.; Pohl, H.; Rastogi, A.; Burgart, L.; Imperiale, T.; Ladabaum, U.; Cohen, J.; et al. The American Society for Gastrointestinal Endoscopy PIVI (Preservation and Incorporation of Valuable Endoscopic Innovations) on real-time endoscopic assessment of the histology of diminutive colorectal polyps. Gastrointest. Endosc. 2011, 73, 419–422. [Google Scholar] [CrossRef]
  12. Lieberman, D.; Moravec, M.; Holub, J.; Michaels, L.; Eisen, G. Polyp Size and Advanced Histology in Patients Undergoing Colonoscopy Screening: Implications for CT Colonography. Gastroenterology 2008, 135, 1100–1105. [Google Scholar] [CrossRef] [Green Version]
  13. Chiu, H.-M.; Chang, L.-C.; Shun, C.-T.; Wu, M.-S.; Wang, H.-P. Current management of diminutive colorectal polyps in Taiwan. Dig. Endosc. 2014, 26, 64–67. [Google Scholar] [CrossRef] [Green Version]
  14. Kang, Y.K. Diminutive and Small Colorectal Polyps: The Pathologist’s Perspective. Clin. Endosc. 2014, 47, 404–408. [Google Scholar] [CrossRef] [Green Version]
  15. Abu Dayyeh, B.K.; Thosani, N.; Konda, V.; Wallace, M.B.; Rex, D.K.; Chauhan, S.S.; Hwang, J.H.; Komanduri, S.; Manfredi, M.; Maple, J.T.; et al. ASGE Technology Committee systematic review and meta-analysis assessing the ASGE PIVI thresholds for adopting real-time endoscopic assessment of the histology of diminutive colorectal polyps. Gastrointest. Endosc. 2015, 81, 502.e1–502.e16. [Google Scholar] [CrossRef] [PubMed]
  16. Hassan, C.; Pickhardt, P.J.; Rex, D.K. A Resect and Discard Strategy Would Improve Cost-Effectiveness of Colorectal Cancer Screening. Clin. Gastroenterol. Hepatol. 2010, 8, 865–869.e3. [Google Scholar] [CrossRef]
  17. Kessler, W.R.; Imperiale, T.; Klein, R.W.; Wielage, R.C.; Rex, D.K. A quantitative assessment of the risks and cost savings of forgoing histologic examination of diminutive polyps. Laryngo-Rhino-Otol. 2011, 43, 683–691. [Google Scholar] [CrossRef]
  18. Ladabaum, U.; Fioritto, A.; Mitani, A.; Desai, M.; Kim, J.P.; Rex, D.K.; Imperiale, T.; Gunaratnam, N. Real-Time Optical Biopsy of Colon Polyps with Narrow Band Imaging in Community Practice Does Not Yet Meet Key Thresholds for Clinical Decisions. Gastroenterology 2013, 144, 81–91. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Paggi, S.; Rondonotti, E.; Amato, A.; Terruzzi, V.; Imperiali, G.; Mandelli, G.; Terreni, N.; Lenoci, N.; Spinzi, G.; Radaelli, F. Resect and discard strategy in clinical practice: A prospective cohort study. Endoscopy 2012, 44, 899–904. [Google Scholar] [CrossRef]
  20. Mori, Y.; Kudo, S.E.; Berzin, T.M.; Misawa, M.; Takeda, K. Computer-aided diagnosis for colonoscopy. Endoscopy 2017, 49, 813–819. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Takemura, Y.; Yoshida, S.; Tanaka, S.; Kawase, R.; Onji, K.; Oka, S.; Tamaki, T.; Raytchev, B.; Kaneda, K.; Yoshihara, M.; et al. Computer-aided system for predicting the histology of colorectal tumors by using narrow-band imaging magnifying colonoscopy (with video). Gastrointest. Endosc. 2012, 75, 179–185. [Google Scholar] [CrossRef] [PubMed]
  22. Kominami, Y.; Yoshida, S.; Tanaka, S.; Sanomura, Y.; Hirakawa, T.; Raytchev, B.; Tamaki, T.; Koide, T.; Kaneda, K.; Chayama, K. Computer-aided diagnosis of colorectal polyp histology by using a real-time image recognition system and narrow-band imaging magnifying colonoscopy. Gastrointest. Endosc. 2015, 83, 643–649. [Google Scholar] [CrossRef]
  23. Hafner, M.; Brunauer, L.; Payer, H.; Resch, R.; Wrba, F.; Gangl, A.; V’ecsei, A.; Uhl, A. Pit Pattern Classification of Zoom-endoscopical Colon Images using DCT and FFT. In Proceedings of the IEEE International Symposium on Computer-Based Medical Systems, Maribor, Slovenia, 20–22 June 2007; pp. 159–164. [Google Scholar]
  24. Häfner, M.; Kwitt, R.; Uhl, A.; Wrba, F.; Gangl, A.; Vécsei, A. Computer-assisted pit-pattern classification in different wavelet domains for supporting dignity assessment of colonic polyps. Pattern Recognit. 2009, 42, 1180–1191. [Google Scholar] [CrossRef] [Green Version]
  25. Gross, S.; Kennel, M.; Stehle, T.; Wulff, J.; Tischendorf, J.; Trautwein, C.; Aach, T. Polyp Segmentation in NBI Colonoscopy. In Bildverarbeitung für die Medizin; Meinzer, H.P., Handels, H., Tolxdorff, T., Eds.; Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  26. Tamaki, T.; Yoshimuta, J.; Takeda, T.; Raytchev, B.; Kaneda, K.; Yoshida, S.; Takemura, Y.; Tanaka, S. A System for Colorectal Tumor Classification in Magnifying Endoscopic NBI Images. In Proceedings of the 10th Asian Conference on Computer Vision, Queenstown, New Zealand, 8–12 November 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 452–463. [Google Scholar]
  27. Korbar, B.; Olofson, A.M.; Miraflor, A.P.; Nicka, C.M.; Suriawinata, M.A.; Torresani, L.; Suriawinata, A.A.; Hassanpour, S. Looking Under the Hood: Deep Neural Network Visualization to Interpret Whole-Slide Image Analysis Outcomes for Col-orectal Polyps. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 821–827. [Google Scholar]
  28. Liedlgruber, M.; Uhl, A. Computer-Aided Decision Support Systems for Endoscopy in the Gastrointestinal Tract: A Review. IEEE Rev. Biomed. Eng. 2011, 4, 73–88. [Google Scholar] [CrossRef]
  29. Suzuki, K. Overview of deep learning in medical imaging. Radiol. Phys. Technol. 2017, 10, 257–273. [Google Scholar] [CrossRef]
  30. Shen, D.; Wu, G.; Suk, H.-I. Deep Learning in Medical Image Analysis. Annu. Rev. Biomed. Eng. 2017, 19, 221–248. [Google Scholar] [CrossRef] [Green Version]
  31. Lu, L.; Zheng, Y.; Carneiro, G.; Yang, L. Deep Learning and Convolutional Neural Networks for Medical Image Computing. Precision Medicine, High Performance and Large-Scale Datasets; Springer: Cham, Switzerland, 2017. [Google Scholar]
  32. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.; van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [Green Version]
  33. Zhang, Y.-D.; Satapathy, S.C.; Guttery, D.S.; Górriz, J.M.; Wang, S.-H. Improved Breast Cancer Classification Through Combining Graph Convolutional Network and Convolutional Neural Network. Inf. Process. Manag. 2020, 58, 102439. [Google Scholar] [CrossRef]
  34. Liu, Q. Deep Learning Applied to Automatic Polyp Detection in Colonoscopy Images; College of Southeast Norway, University College of Southeast Norway: Kongsberg, Norway, 2017. [Google Scholar]
  35. Chen, P.-J.; Lin, M.-C.; Lai, M.-J.; Lin, J.-C.; Lu, H.H.-S.; Tseng, V.S. Accurate Classification of Diminutive Colorectal Polyps Using Computer-Aided Analysis. Gastroenterology 2018, 154, 568–575. [Google Scholar] [CrossRef]
  36. Park, S.Y. Colonoscopic polyp detection using convolutional neural networks. In Proceedings of the SPIE 9785, Medical Imaging 2016: Computer-Aided Diagnosis, San Diego, CA, USA, 27 February–3 March 2016. [Google Scholar]
  37. Ribeiro, E.; Uhl, A.; Hafner, M. Colonic Polyp Classification with Convolutional Neural Networks. In Proceedings of the 2016 IEEE 29th Inter-national Symposium on Computer-Based Medical Systems (CBMS), Belfast and Dublin, Ireland, 20–24 June 2016; pp. 253–258. [Google Scholar]
  38. Shin, Y.; Qadir, H.A.; Aabakken, L.; Bergsland, J.; Balasingham, I. Automatic Colon Polyp Detection Using Region Based Deep CNN and Post Learning Approaches. IEEE Access 2018, 6, 40950–40962. [Google Scholar] [CrossRef]
  39. Ren, S.; He, K.; Girshick, R.; Sun, R. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef] [Green Version]
  40. Mo, X.; Tao, K.; Wang, Q.; Wang, G. An Efficient Approach for Polyps Detection in Endoscopic Videos Based on Faster R-CNN. In Proceedings of the 24th IEEE International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; IEEE: Beijing, China, 2018; pp. 3928–3934. [Google Scholar]
  41. Wang, P.; Xiao, X.; Glissen Brown, J.R.; Berzin, T.M.; Tu, M.; Xiong, F.; Hu, X.; Liu, P.; Song, Y.; Zhang, D.; et al. Development and validation of a deep-learning algorithm for the detection of polyps during colon-oscopy. Nat. Biomed. Eng. 2018, 2, 741–748. [Google Scholar] [CrossRef]
  42. Urban, G.; Tripathi, P.; Alkayali, T.; Mittal, M.; Jalali, F.; Karnes, W.; Baldi, P. Deep Learning Localizes and Identifies Polyps in Real Time With 96% Accuracy in Screening Colonoscopy. Gastroenterology 2018, 155, 1069–1078.e8. [Google Scholar] [CrossRef]
  43. Zheng, Y.; Yu, R.; Jiang, Y.; Mak, T.W.C.; Wong, S.H.; Lau, J.Y.W.; Poon, C.C.Y. Localisation of Colorectal Polyps by Convolutional Neural Network Features Learnt from White Light and Narrow Band Endoscopic Images of Multiple Data-bases. In Proceedings of the 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Honolulu, HI, USA, 17–21 July 2018; pp. 4142–4145. [Google Scholar]
  44. Tan, J.; Gao, Y.; Cao, W.; Pomeroy, M.; Zhang, S.; Huo, Y.; Li, L.; Liang, Z. GLCM-CNN: Gray Level Co-occurrence Matrix based CNN Model for Polyp Diagnosis. In Proceedings of the 2019 IEEE EMBS International Conference on Biomedical & Health Informatics (BHI), Chicago, IL, USA, 19–22 May 2019; pp. 1–4. [Google Scholar] [CrossRef]
  45. Zhang, Y.-D.; Satapathy, S.C.; Liu, S.; Li, G.-R. A five-layer deep convolutional neural network with stochastic pooling for chest CT-based COVID-19 diagnosis. Mach. Vis. Appl. 2020, 32, 1–13. [Google Scholar] [CrossRef]
  46. Xie, Y.; Richmond, D. Pre-training on Grayscale ImageNet Improves Medical Image Classification. In Proceedings of the Computer Vision–ECCV 2018 Workshops, Munich, Germany, 8–14 September 2018; pp. 476–484. [Google Scholar]
  47. ITU. ITU-R Recommendations Retrieved BT.601. Available online: https://www.itu.int/rec/R-REC-BT.601/ (accessed on 11 April 2018).
  48. Bernal, J.; Sánchez, J.; Vilariño, F. Towards automatic polyp detection with a polyp appearance model. Pattern Recognit. 2012, 45, 3166–3182. [Google Scholar] [CrossRef]
  49. Bernal, J.; Sánchez, F.J.; Fernández-Esparrach, M.G.; Gil, D.; Rodríguez, C.; Vilariño, F. WM-DOVA maps for accurate polyp highlighting in colonoscopy: Validation vs. saliency maps from physicians. Comput. Med Imaging Graph. 2015, 43, 99–111. [Google Scholar] [CrossRef]
  50. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  51. Bernal, J.; Tajkbaksh, N.; Sánchez, F.J.; Matuszewski, B.J.; Chen, H.; Yu, L.; Angermann, Q.; Romain, O.; Rustad, B.; Balasingham, I.; et al. Comparative Validation of Polyp Detection Methods in Video Colonoscopy: Results from the MICCAI 2015 Endoscopic Vision Challenge. IEEE Trans. Med. Imaging 2017, 36, 1231–1249. [Google Scholar] [CrossRef]
Figure 1. Example image of NICE classification.
Figure 1. Example image of NICE classification.
Sensors 21 05995 g001
Figure 2. Excessive intestinal wall textures with small polyps.
Figure 2. Excessive intestinal wall textures with small polyps.
Sensors 21 05995 g002
Figure 3. Small texture of polyp and intestinal wall.
Figure 3. Small texture of polyp and intestinal wall.
Sensors 21 05995 g003
Figure 4. Colonoscopy image.
Figure 4. Colonoscopy image.
Sensors 21 05995 g004
Figure 5. Segmented polyp image.
Figure 5. Segmented polyp image.
Sensors 21 05995 g005
Figure 6. System architecture.
Figure 6. System architecture.
Sensors 21 05995 g006
Figure 7. Polyp image before gray scaling.
Figure 7. Polyp image before gray scaling.
Sensors 21 05995 g007
Figure 8. Polyp image after gray scaling.
Figure 8. Polyp image after gray scaling.
Sensors 21 05995 g008
Figure 9. Colorectal polyp detection system architecture.
Figure 9. Colorectal polyp detection system architecture.
Sensors 21 05995 g009
Figure 10. WL polyp detection.
Figure 10. WL polyp detection.
Sensors 21 05995 g010
Figure 11. NBI polyp detection.
Figure 11. NBI polyp detection.
Sensors 21 05995 g011
Figure 12. ROC curve for classification.
Figure 12. ROC curve for classification.
Sensors 21 05995 g012
Figure 13. Nonobvious polyp protrusions or similar to intestinal wall.
Figure 13. Nonobvious polyp protrusions or similar to intestinal wall.
Sensors 21 05995 g013
Figure 14. Excessive noise images.
Figure 14. Excessive noise images.
Sensors 21 05995 g014
Figure 15. Small pixel size polyp images.
Figure 15. Small pixel size polyp images.
Sensors 21 05995 g015
Figure 16. Unobvious polyp textures.
Figure 16. Unobvious polyp textures.
Sensors 21 05995 g016
Figure 17. Polyp image size below 1800 pixels.
Figure 17. Polyp image size below 1800 pixels.
Sensors 21 05995 g017
Table 1. NBI international colorectal endoscopic (NICE) classification.
Table 1. NBI international colorectal endoscopic (NICE) classification.
ClassificationType 1Type 2Type 3
ColorSame or lighter than backgroundBrowner than backgroundBrown to dark brown than background, sometimes patchy whiter areas
VesselsNone or isolated lacyBrown vessels surrounding white structuresAreas of disrupted or missing vessels
Surface patternDark or white spots of uniform size, or homogeneous absence of patternOval, tubular or branchedAmorphous or abscent surface pattern
DiagnosisHyperplastic polypAdenomaDeep submucosal invasive cancer
Table 2. Experiment data set.
Table 2. Experiment data set.
NameImage (N)ResolutionFormat
CVC-Clinic612384 × 288.tif
CGMH-WL1080640 × 480.tif
CGMH-NBI1080640 × 480.tif
Table 3. Colorectal polyp detection model architecture.
Table 3. Colorectal polyp detection model architecture.
LayersFiltersSize/StrideOutput
Input 128 × 128
Convolutional16(3 × 3 + 3 × 1 + 1 × 3)/1128 × 128
Batch Normalization 128 × 128
ReLu 128 × 128
Max pooling 2 × 2/264 × 64
Convolutional32(3 × 3 + 3 × 1 + 1 × 3)/164 × 64
Batch Normalization 64 × 64
ReLu 64 × 64
Max pooling 2×2/232 × 32
Convolutional64(3 × 3 + 3 × 1 + 1 × 3)/132 × 32
Batch Normalization 32 × 32
ReLu 32 × 32
Max pooling 2 × 2/216 × 16
Convolutional128(3 × 3 + 3 × 1 + 1 × 3)/116 × 16
Batch Normalization 16 × 16
ReLu 16 × 16
Convolutional128(3 × 3 + 3 × 1 + 1 × 3)/116 × 16
Batch Normalization 16 × 16
ReLu 16 × 16
Convolutional128(3 × 3 + 3 × 1 + 1 × 3)/116 × 16
Batch Normalization 16 × 16
ReLu 16 × 16
Convolutional241 × 1/116 × 16
Transform 16 × 16
Output
Table 4. Colorectal polyp classification model architecture.
Table 4. Colorectal polyp classification model architecture.
LayersFiltersSize/StrideOutput
Convolutional16(3 × 3 + 3 × 1 + 1 × 3)/1480 × 640
Batch Normalization 480 × 640
ReLu 480 × 640
Max pooling 2 × 2/2240 × 320
Convolutional16(3 × 3 + 3 × 1 + 1 × 3)/1240 × 320
Batch Normalization 240 × 320
ReLu 240 × 320
Max pooling 2 × 2/2120 × 160
Convolutional32(3 × 3 + 3 × 1 + 1 × 3)/1120 × 160
Batch Normalization 120 × 160
ReLu 120 × 160
Max pooling 2 × 2/340 × 53
Convolutional32(3 × 3 + 3 × 1 + 1 × 3)/140 × 53
Batch Normalization 40 × 53
ReLu 40 × 53
Max pooling 2 × 2/313 × 18
Convolutional64(3 × 3 + 3 × 1 + 1 × 3)/113 × 18
Batch Normalization 13 × 18
ReLu 13 × 18
Max pooling 2 × 2/34 × 6
Convolutional64(3 × 3 + 3 × 1 + 1 × 3)/14 × 6
Batch Normalization 4 × 6
ReLu 4 × 6
Max pooling 2 × 2/12 × 3
Fully Connected 2
Softmax
Table 5. Training and testing dataset of polyp detection.
Table 5. Training and testing dataset of polyp detection.
DatasetTraining#Testing#
CVC-Clinic612-
CGMH-WL-900
CGMH-NBI-900
Table 6. CGMH-WL testing dataset of polyp detection.
Table 6. CGMH-WL testing dataset of polyp detection.
RGBGrayscale
PolypNormalPolypNormal
Polyp8435785941
Accuracy% (95% CI)%93.7 (91.9~95.2)95.4 (93.9~96.7)
Table 7. Comparison of polyp detection between RGB and grayscale images with CGMH-WLtesting dataset.
Table 7. Comparison of polyp detection between RGB and grayscale images with CGMH-WLtesting dataset.
RGBGrayscalep Value
Polyp DetectedNo Polyp Detected
Polyp detected8430<0.001
No polyp detected1641
Table 8. CGMH-NBI testing dataset of polyp detection.
Table 8. CGMH-NBI testing dataset of polyp detection.
NBIGrayscale
PolypNormalPolypNormal
Polyp8505085347
Accuracy% (95% CI)%94.4 (92.7~95.8)94.8 (93.1~96.1)
Table 9. Comparison of polyp detection between NBI and grayscale images with CGMH-NBI testing dataset.
Table 9. Comparison of polyp detection between NBI and grayscale images with CGMH-NBI testing dataset.
NBIGrayscalep Value
Polyp DetectedNo Polyp Detected
Polyp detected85000.25
No polyp detected347
Table 10. CGMH-WL + CGMH-NBI testing dataset of polyp detection.
Table 10. CGMH-WL + CGMH-NBI testing dataset of polyp detection.
RGB + NBIGrayscale
PolypNormalPolypNormal
Polyp1693107171288
Accuracy% (95% CI)%94.1 (92.8~95.1)95.1 (94.0~96.1)
Table 11. Comparison of polyp detection between RGB + NBI and grayscale images with CGMH WL + CGMH-NBI testing dataset.
Table 11. Comparison of polyp detection between RGB + NBI and grayscale images with CGMH WL + CGMH-NBI testing dataset.
RGB + NBIGrayscalep Value
Polyp DetectedNo Polyp Detected
Polyp detected16930<0.001
No polyp detected1988
Table 12. The image number of polyps size greater and smaller than 0.5 cm.
Table 12. The image number of polyps size greater and smaller than 0.5 cm.
# of Polyp Size > 0.5 cm# of Polyp Size ≤ 0.5 cmSubtotal
WL313587900
NBI294606900
Total60711931800
Table 13. Polyp > 0.5 cm.
Table 13. Polyp > 0.5 cm.
WLNBI
PolypNormalPolypNormal
Polyp3021128113
Accuracy% (95% CI)%96.5 (93.8~98.2)95.6 (92.6~97.6)
Table 14. Polyp size ≤ 0.5 cm.
Table 14. Polyp size ≤ 0.5 cm.
WLNBI
PolypNormalPolypNormal
Polyp5573057234
Accuracy% (95% CI)%94.9 (92.8~96.5)94.4 (92.2~96.1)
Table 15. Polyp classification dataset.
Table 15. Polyp classification dataset.
Data TypeNeoplastic PolypsHyperplastic PolypsSubtotalPercentage
Training#68568513701370/1712 = 80%
Verification#26379342342/1712 = 20%
Subtotal8568561712100%
Testing#248112360360/360 = 100%
(Training#) + (Verification#) + (Testing#) = 2072 images
Table 16. Performance indices.
Table 16. Performance indices.
Accuracy (Acc) Acc = TP + FN TP + FP + TN + FN
Precision (Prec) Prec = TP TP + FP
Recall (Rec) Rec = TP TP + FN
F1-measure (F1) F 1 = 2 × Prec × Rec Prec + Rec
F2-measure (F2) F 2 = 5 × Prec × Rec 4 × Prec + Rec
Table 17. Confusion matrix of WL polyp classification.
Table 17. Confusion matrix of WL polyp classification.
Neoplastic (Predicted)Hyperplastic (Predicted)
Neoplastic (Actual)11014
Hyperplastic (Actual)3620
Table 18. Confusion matrix of NBI polyp classification.
Table 18. Confusion matrix of NBI polyp classification.
Neoplastic (Predicted)Hyperplastic (Predicted)
Neoplastic(Actual)1186
Hyperplastic(Actual)2531
Table 19. Performance index of polyp classification.
Table 19. Performance index of polyp classification.
Acc%Prec%Rec%F1%F2%
WL72.275.388.781.585.7
NBI82.882.595.288.492.3
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hsu, C.-M.; Hsu, C.-C.; Hsu, Z.-M.; Shih, F.-Y.; Chang, M.-L.; Chen, T.-H. Colorectal Polyp Image Detection and Classification through Grayscale Images and Deep Learning. Sensors 2021, 21, 5995. https://doi.org/10.3390/s21185995

AMA Style

Hsu C-M, Hsu C-C, Hsu Z-M, Shih F-Y, Chang M-L, Chen T-H. Colorectal Polyp Image Detection and Classification through Grayscale Images and Deep Learning. Sensors. 2021; 21(18):5995. https://doi.org/10.3390/s21185995

Chicago/Turabian Style

Hsu, Chen-Ming, Chien-Chang Hsu, Zhe-Ming Hsu, Feng-Yu Shih, Meng-Lin Chang, and Tsung-Hsing Chen. 2021. "Colorectal Polyp Image Detection and Classification through Grayscale Images and Deep Learning" Sensors 21, no. 18: 5995. https://doi.org/10.3390/s21185995

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop