Next Article in Journal
Self-Regulating Adaptive Controller for Oxygen Support to Severe Respiratory Distress Patients and Human Respiratory System Modeling
Next Article in Special Issue
Comparison of Machine Learning Models and the Fatty Liver Index in Predicting Lean Fatty Liver
Previous Article in Journal
Identification of Prognostic Biomarkers for Suppressing Tumorigenesis and Metastasis of Hepatocellular Carcinoma through Transcriptome Analysis
Previous Article in Special Issue
A Multiscale Polyp Detection Approach for GI Tract Images Based on Improved DenseNet and Single-Shot Multibox Detector
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Negative Samples for Improving Object Detection—A Case Study in AI-Assisted Colonoscopy for Polyp Detection

by
Alba Nogueira-Rodríguez
1,2,*,
Daniel Glez-Peña
1,2,
Miguel Reboiro-Jato
1,2 and
Hugo López-Fernández
1,2
1
CINBIO, Department of Computer Science, ESEI—Escuela Superior de Ingeniería Informática, Universidade de Vigo, 32004 Ourense, Spain
2
SING Research Group, Galicia Sur Health Research Institute (IIS Galicia Sur), SERGAS-UVIGO, 36213 Vigo, Spain
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(5), 966; https://doi.org/10.3390/diagnostics13050966
Submission received: 15 January 2023 / Accepted: 1 March 2023 / Published: 3 March 2023

Abstract

:
Deep learning object-detection models are being successfully applied to develop computer-aided diagnosis systems for aiding polyp detection during colonoscopies. Here, we evidence the need to include negative samples for both (i) reducing false positives during the polyp-finding phase, by including images with artifacts that may confuse the detection models (e.g., medical instruments, water jets, feces, blood, excessive proximity of the camera to the colon wall, blurred images, etc.) that are usually not included in model development datasets, and (ii) correctly estimating a more realistic performance of the models. By retraining our previously developed YOLOv3-based detection model with a dataset that includes 15% of additional not-polyp images with a variety of artifacts, we were able to generally improve its F1 performance in our internal test datasets (from an average F1 of 0.869 to 0.893), which now include such type of images, as well as in four public datasets that include not-polyp images (from an average F1 of 0.695 to 0.722).

1. Introduction

Computer-aided diagnosis (CAD) systems for polyp detection are now reality, thanks to the advances in and the application of deep learning (DL). So far, many randomized control trials (RCT) have already been performed [1,2,3,4,5,6], some of them associated with commercial systems [7], and many more are being developed. The growing number of reviews on this topic demonstrates that polyp detection is much more advanced than polyp characterization [5,7,8,9], and a recent meta-analysis of six RCTs showed that there is an increase in both adenoma and polyp detection rates with the utilization of artificial intelligence (AI)-assisted colonoscopy [5].
In the context of the PolyDeep project (https://www.polydeep.org, accessed on 11 January 2023), we developed and reported, in 2021, a real-time polyp detection model based on a YOLOv3 pre-trained with PASCAL VOC datasets [10]. The base model was fine-tuned using a private dataset containing 941 different polyps and 28,576 images annotated by expert endoscopists [11]. The performance of this model in a bounding box-based evaluation using still images (a test partition of our private dataset) was an F1 score of 0.88 (recall = 0.87, precision = 0.89), reaching a state-of-the-art level.
In follow-up work [12], we reported, in 2022, the results of testing this model on ten public colonoscopy image datasets, which, to the best of our knowledge, was the biggest systematic evaluation of a polyp detection model without retraining. We also analyzed these results in the context of the results from another 20 state-of-the-art publications using the same public datasets. We observed that our model had an average F1 score decay of 13.65% when tested on the ten public datasets. This number may seem high, but it must be taken into account the high degree of variability in the public datasets, as some of them contain not-polyp images and multiple-polyp images, resulting in a slightly different scenario compared to the type of images used to develop our base model (only polyp-containing images).
Since then, we have also released an updated version of our dataset and other complementary datasets through the PIBAdb Colorectal Polyp Image Cohort (https://www.iisgaliciasur.es/home/biobanco/cohorte-pibadb, accessed on 11 January 2023), which is publicly available at the biobank of the Galicia Sur Health Research Institute (IISGS). This image cohort was created with the aim of making public the data collected during the PolyDeep project, including, mainly, images with polyps located with bounding boxes annotated by expert endoscopists, but also other colon images without polyps and the videos of colonoscopies from which the images were extracted. Notably, one of the complementary datasets published includes ~14,000 not-polyp images (the “PIBAdb Not Polyp” image gallery). These images without polyps have been extracted automatically (one image per second) from annotated videos and they may contain artifacts (e.g., instrumental, on-screen information, etc.), be blurred or very close to mucosa. This follows the general trend of the latest public datasets available, which are larger and include images without polyps.
In light of the findings of our previous works, our efforts to keep improving our model have been data-centric. In this regard, some studies have already used not-polyp images during training to improve the performance of the model [13], and, therefore, in this work we assess the usefulness of the recent not-polyp images included in the PIBAdb database to improve the performance of our previous model. Our belief is that training using these images will allow us to improve the predictive positive value of the model (precision), while maintaining its sensitivity levels (recall).

Rationale (Hypothesis)

The rationale behind the inclusion of no-object images, or negative samples, in an object detection network is guided by the fact that, when such models are implemented in a real setting, the model could face out-of-distribution samples, that is, image types that have content that was not used during the model development.
A polyp detection network trained only with polyp-containing images (at least one polyp per image) uses the outside-boxes image content as the “negative” samples, learning not to predict bounding boxes there. This may be sufficient if all the possible negative content is there, but this is not the case in endoscopies, especially during the polyp-finding phase. The normal mucosa content behind a polyp is not the only negative content to learn from. There are many other scenarios (out-of-distribution) that will arise during the exploration (e.g., medical instruments, water jets, feces, blood, excessive proximity of the camera to the colon wall, blurred images, etc.) that could lead the network to identify parts of the image as something similar to a polyp, or, at least, more similar to a polyp than to normal mucosa, producing many false positives.
With this study, we aim at exploiting this hypothesis by retraining our model with such out-of-distribution images in order to improve its positive predictive value, while maintaining its sensitivity. Moreover, we will put special focus on testing the model with this kind of image at different proportions, since test datasets that do not include them are not able to capture the real performance of detection models in the real setting. In summary, we want to move out-of-distribution samples to in-distribution samples.

2. Materials and Methods

2.1. Our Base Model for Polyp Detection

In 2021, we reported the results of a real-time colorectal polyp detection model based on YOLOv3 [11]. This model was trained and evaluated with a private dataset built using 28,576 images with located polyps from the PIBAdb cohort, both with white light (WL) and narrow-band imaging (NBI) light. This dataset includes a subset of images automatically extracted at a rate of one image per second (16,691) and a subset of images manually selected by expert endoscopists (11,885).
For the model development, the private dataset was divided into three partitions: (i) a training partition (49%; 13,873 images) for model training during the development phase, (ii) a validation partition (21%; 6045 images) for model evaluation and selecting the best model during the development phase, and (iii) a test partition (30%; 8658 images) for best model evaluation after the development phase. The resulting model, pre-trained with the PASCAL VOC 2007 and 2012 challenges and fine-tuned with the dataset comprising both the training and validation partitions, achieved an F1 score of 0.88 (recall = 0.87, precision = 0.89) and an average precision (AP) of 0.87 evaluated on the test partition. As noted in the introduction, this model was later evaluated on ten public datasets, without retraining, in order to assess its generalization capabilities [12], showing an average F1 decay of 13.65%.

2.2. Public Datasets

Table 1 shows the most relevant details of the ten public colonoscopy image datasets selected in our previous study [12], which are also the subject of this study. Regarding polyp location in the dataset images, four dataset provide it as bounding boxes (KUMC [14], SUN [13], LDPolypVideo [15], and Kvasir-SEG [16]), while seven datasets use binary masks (CVC-ClinicDB [17], CVC-ColonDB [18,19], CVC-PolypHD [18,19], ETIS-Larib [20], CVC-ClinicVideoDB [21,22], PICCOLO [23], and Kvasir-SEG [16]). As the polyp detection model evaluated in this work uses bounding boxes, it was necessary to transform the binary masks into bounding boxes in six datasets (ClinicDB [17], CVC-ColonDB [18,19], CVC-PolypHD [18,19], ETIS-Larib [20], CVC-ClinicVideoDB [21,22], and PICCOLO [23]. The process followed to perform these transformations was explained in our previous study [12], and the conversion scripts are available at https://github.com/sing-group/public-datasets-to-voc (accessed on 11 January 2023).
Taking into account the characteristics of the different datasets, a high degree of variability can be observed with respect to the number of images included, and whether they contain not-polyp images or not [8,24]. Nevertheless, the trend in recent years is to create increasingly larger datasets with the presence of images without polyps.

2.3. Experimental Setup

The experiments consist of training new models using the original polyp image dataset of our base model and progressively injecting it with different percentages of not-polyp images. The original dataset partitions (i.e., training, validation, and test) were augmented with not-polyp images at rates around 2%, 5%, 10%, and 15% of the initial number of images in the partition.
Starting from the images of each original partition, where each polyp only belongs to one single partition, we progressively added each percentage of images without polyps. The image injection was performed incrementally for each percentage, ensuring that the same not-polyp image was in the same partition type (training, validation, or test) and that the images found in the lower percentage datasets were part of the higher percentage datasets, creating supersets for each partition and dataset. For example, all the images in the 2% training are also included in the 5% dataset, along with an additional 3% of new not-polyp images. The number of resulting images for each partition is shown in the following, Figure 1.
Figure 2 shows diverse examples of not-polyp images that have been used to train the new networks. As can be seen, these images may contain medical instruments, water jets, feces, blood, excessive proximity of the camera to the colon wall, blurred images, etc.
As in our previous works [11,12], the model development and evaluation was done using Compi [25,26] pipelines, which can be found at the GitHub repository: https://github.com/sing-group/polydeep-object-detection (accessed on 11 January 2023). Specifically, the model training was performed using the train.xml pipeline, while the model evaluation was performed using the text.xml pipeline of this repository.
Once retrained (see details in “Model development and selection”), we evaluated the performance of each new model in an (i) intra-dataset evaluation, using the test partition of the base model, as well as the not-polyp augmented test partitions, and in an (ii) inter-dataset evaluation, using the ten public colonoscopy imaging datasets selected in the previous study.

2.4. Evaluation Metrics

Since the main goal of a colorectal polyp detection model is to identify all the polyps shown on camera during a colonoscopy, we have selected recall (or sensitivity) as one of the evaluation measures. In addition, as the ability to not show detections in locations where there is no polyp is also important, because they can disturb or distract an endoscopist, we have also selected precision (or positive predictive value) as a second evaluation measure. Finally, since it acts as a summary of the two evaluation measures mentioned above, we have chosen F 1 as the main evaluation measure.
Given that the output of the detection model is a bounding box, in this work we define:
  • True Positive (TP): as a predicted bounding box that is over a true bounding box.
  • False Positive (FP): as a predicted bounding box that is not over a true bounding box.
  • False Negative (FN): as a true bounding box where there is no predicted bounding box over it.
Taking into account these definitions, the selected performance metrics (i.e., recall, precision, and F 1 ) are defined as follows:
recall ( or   sensitivity ) = TP TP + FN
precision ( or   positive   predictive   value ) = TP TP + FP
F 1 = 2 × precision · recall precision + recall = 2 × TP 2 × TP + FP + FN

3. Results and Discussion

The experiments are focused on obtaining the performance metrics of each model in a test partition with similar images with which it was trained (intra-dataset evaluation) and, also, evaluating the models with a test partition where the images belong to public datasets (inter-dataset evaluation). The results are completed with a frame analysis to see how false positive identifications are removed.

3.1. Model Development and Selection

Following the same strategy as in our previous works [11,12], each model was retrained again, starting from the initial pre-trained YOLOv3 and using the training partition for 50 epochs. The same training parameters were used for all the models. The best model is selected by the maximum AP achieved in the validation partition during those 50 epochs. Once the best model is selected, it is configured with the confidence score threshold that maximizes the F1 score in the validation partition. Table 2 shows the metrics of the selected models with the best AP achieved in the validation partition images during the training process. AP is equivalent to the Area Under the Precision-Recall Curve (AUPRC), being the average precision at various confidence score thresholds.
As a result, we have the original model (“Nogueira et al.”), plus four additional models (“Model injection p”, where p is the percentage of injection of not-polyp images into the dataset partitions used to develop the model).

3.2. Intra-Dataset Evaluation

First, we evaluated the models trained after the injection of not-polyp images using the test partition of the base model (Nogueira et al., i.e., without not-polyp images). The results are shown in Table 3.
Regarding recall, all the models show a slight fluctuation, ranging from 0.85 in the model trained with the addition of 10% not-polyp images, to 0.88 in the model trained with the addition of 15% not-polyp images. On the other hand, the precision increases as the models are trained with increasing numbers of not-polyp images, and thus achieves the highest value in the model trained with the addition of 15% of not-polyp images (0.91), which also obtains the best F1 score (0.895). In comparison with the base model, the models trained with the addition of 5% and 15% not-polyp images show a slightly better F1 score, whereas the 2% and 10% models a slightly lower one. In summary, the difference is small, and no model shows a clear improvement in the original test partition.
However, as we are looking for a test scenario that better reflects the out-of-distribution samples that the models will face in a real setting, mimicked by including not-polyp images in the test partition, we also performed the evaluation with the test partitions that include different proportions of not-polyp images (also from 2% up to 15%). Since the true polyp images remain the same in all the new not-polyp-containing test partitions, the recall of any given model will remain exactly the same as shown in Table 3, but the precision, and thus the F1, are expected to be affected. Concretely, Table 4 shows the different precision metrics obtained by testing all the models with the different test partitions.
As can be seen, precision is positively correlated with the number of not-polyp images used for training the different models, which evidences the benefit of training with not-polyp images, and it is inversely correlated with the number of not-polyp images in the test partition, which is a statistically expected effect.
Putting it all together, Figure 3 shows the recall and precision comparing two distant test partitions: the original one, which does not contain any not-polyp images, and the largest not-polyp containing test partition, which contains the addition of 15% of not-polyp images.
It can be observed from Figure 3 that the precision decay of the original model is the biggest one (red bars in the leftmost column of Figure 3). However, as we create models trained with an increasing number of not-polyp images, this precision recovers (dark red bars’ progression in Figure 3).
In order to compare all the models, Figure 4 shows the F1 score of all the models in both test scenarios (without not-polyp images and with the addition of 15% of not-polyp images). As it can be seen, the test partition that includes the addition of 15% of not-polyp images (dark bars in Figure 4) is more challenging, and each model attains a lower F1 score. Moreover, the addition of not-polyp images for training does not show a consistent benefit in the original test partition (light bars in Figure 4), whereas it does in the test partition that adds 15% of not-polyp images (dark bars in Figure 4). We draw attention to the lowest value obtained with the original model in the not-polyp test partition (from the original 0.88 to 0.85). Finally, in terms of the F1 score, the model trained with the highest number of not-polyp images showed the best F1 score in both test partitions (0.895 and 0.891).

3.3. Inter-Dataset Evaluation

Table 5 shows the performance of our five models, i.e., the baseline model presented in Nogueira et al. [11], plus the four new models trained with different percentages of not-polyp image injections, on the same public datasets analyzed in our previous work [12]. Globally, the average F1 score remains almost the same, with an F1 around 0.75 (see global averages in Table 5). As a result, we could discard an unexpected worsening in the public datasets.
However, when the testing results are divided into functions of whether the testing dataset contains not-polyp images or not, the results differ. In the case of the results obtained in the datasets that contain not-polyp images, the models trained with not-polyp images attain, on average, greater than or equal to the F1 scores of the original model, where the model with a 15% injection of not-polyp images is able to increase the F1 in the four datasets, raising the original average of 0.695 to 0.722. The behavior of the same models in the datasets that only contain polyp images is more heterogeneous, but a general decrease in the average F1 is observed, ranging from 0.749 to 0.799 in comparison with the original average of 0.800.
In order to visualize the performance changes among our four new models in each public dataset, Figure 5 compares the F1 performances obtained by our baseline model (X axis) and the F1 performances obtained by the four new models trained with not-polyp images (Y axis). The public datasets that contain not-polyp images are marked in light blue.
Focusing more on the relative variations, Figure 6 shows the relative change on F1 among all the public datasets between the models trained with not-polyp images with respect to the original model (trained without not-polyp images) in the same test dataset.
Again, we have grouped the results according to the presence or not of not-polyp images in the test partition. The left part of Figure 6 shows a general benefit in the four public datasets with not-polyp images (dark boxes). By contrast, regarding the performance in the public datasets without not-polyp images, a general worsening is observed (light boxes). When the results are disaggregated by model, the right part of Figure 6 shows that the better performance is always obtained in datasets that include not-polyp images for all the models. Finally, the model trained with 15% of additional not-polyp images attains the highest benefit in the not-polyp-containing datasets (dark box in the right-most column in Figure 6), being the only one showing a benefit of all of the four public datasets with not-polyp images.

3.4. Qualitative Analysis

To perform a qualitative analysis, the five models (the base model presented in Nogueira et al. plus the new four models trained with different percentages of not-polyp image injections) have been used to process two videos at a rate of 25 fps: (i) an 18 s video of a polyp (450 frames); and (ii) a 14 s video of normal mucosa (not-polyp) (350 frames). The two videos are provided as Supplementary Videos S1 and S2, respectively. Table 6 shows a global summary of the predictions made over the two videos.
Regarding the polyp video, in Table 6 we see a general recall of 90%, in terms of distinct frames that contain at least one emitted bounding box, and very small differences among all the models (see relative change, third column in Table 6). Regarding the not-polyp video, we see a drastic reduction in the total number of bounding boxes, with respect to the original model, which were all false positives (see relative change, last column in Table 6).
In order to give some illustrative examples within these two videos, Figure 7 contains some frames with bounding box predictions in the polyp video. As expected, the five models are able to detect the polyp correctly in almost all the frames (first row of Figure 7). However, in this specific polyp video, the new models even surpass the detection ability of the base model (slightly better recall, as shown in Table 6), where some of the false negatives (white box in the second row of Figure 7) in the original model were corrected by the not-polyp-trained models. Moreover, some false positive identifications (boxes not in the correct location of the polyp) emitted by the baseline model disappear in the not-polyp-trained models (second and third rows of Figure 7).
Regarding the not-polyp video, Figure 8 again shows three example frames. The base model produces false positive identifications in the three frames, while most of the not-polyp-trained models do not.
Focusing on the real out-of-distribution samples that we wanted to address in this study, we include two additional video clips found in two exploration videos. Figure 9 shows some frames of different situations found in two exploration videos included in the PIBAdb (provided as Supplementary Videos S3 and S4). These include, for instance, close to the wall of the colon images (first row), water jet (fourth row), or blurry images due to water (fifth row). The base model tends to produce false positives in such situations, as the first column shows, that the not-polyp-trained models tend to eliminate.

4. Conclusions

In this work, we retrained our polyp detection model with negative samples that were out-of-distribution images for our previous model in order to improve its positive predictive value (precision), while maintaining its sensitivity (recall), in the real setting. We have trained several models with different amounts of not-polyp images, incrementing our original training set from 2% up to 15% not-polyp images. We have evaluated them with both internal and public test partitions that include different amounts of such not-polyp images.
Regarding the performance of the models in our test partitions, or intra-dataset evaluation, we have observed that precision always increases as more not-polyp images are used for training the models, while recall oscillates slightly among them. The model trained with a 15% injection of not-polyp images was able to maintain the recall and reach the best precision of all the test partitions that include not-polyp images (average F1 of 0.893). Moreover, the original model was unable to maintain its original performance in the test partition that included the not-polyp images (descending from 0.881 to an average of 0.869), which illustrates the need for a test partition that includes this kind of image, which was kept out-of-distribution during the initial development.
On the other hand, we observed that the performance of the models in the public datasets containing not-polyp images also increases as more not-polyp images were used for training. Again, the model trained with a 15% injection of not-polyp images attained the best average F1 (0.722), and the biggest difference from the baseline model (average F1 of 0.695). In the public datasets with only polyp images, the results are more heterogeneous and differ depending on the dataset and model used. Altogether, the average F1 of each new model remained almost the same (~0.75), and, therefore, an unexpected worsening in the public datasets due to training with not-polyp images is discarded.
As noted before, the general trend is that the latest public datasets available include images without polyps, and our results suggest that it is worth using them (i) as negative samples for training, as it improves the performance in scenarios closer to real colonoscopies, and (ii) to correctly estimate the real performance of the detection models. The benefits are even observed when testing in datasets that only contain polyp images, although with higher variability.
Based on these observations, we think that there is room for improvement in polyp detection models using more data. Thus, future work to improve our model will be data-centric, as well. First, as the best performing model was the one trained with the highest amount of not-polyp images, also maintaining a good recall, it seems that the ceiling has not been reached in terms of the best proportion of not-polyp images to inject. More experiments in this direction could give us interesting insights. Our PIBAdb database still contains both annotated polyp images and not-polyp images that were not used to train the models, and it would be interesting to see if using more training data with both polyp and not-polyp images can increase the performance. Also, a lot of public datasets are available, and it would be interesting to see if training with some images from them (maybe the most challenging ones) can improve the performance in both our dataset and in the other public datasets.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/diagnostics13050966/s1, Supplementary Video S1 (S1_Video_polyp.mp4): an 18 s video of a polyp (450 frames). Supplementary Video S2 (S2_Video_Not_Polyp.mp4): (ii) a 14 s video of normal mucosa (not-polyp) (350 frames). Supplementary Video S3 (S3_Exploration_1.mp4): a 15 s video of an exploration from PIBAdb. Supplementary Video S4 (S4_Exploration_2.mp4): a 15 s video of an exploration from PIBAdb.

Author Contributions

Conceptualization, A.N.-R., D.G.-P. and H.L.-F.; data curation, A.N.-R.; funding acquisition, D.G.-P. and M.R.-J.; investigation, A.N.-R., D.G.-P. and H.L.-F.; methodology, A.N.-R., D.G.-P. and H.L.-F.; project administration, D.G.-P. and M.R.-J.; software, A.N.-R.; supervision, D.G.-P. and H.L.-F.; writing—original draft, A.N.-R., D.G.-P. and H.L.-F.; writing—review and editing, A.N.-R., D.G.-P., M.R.-J. and H.L.-F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by: (i) Ministerio de Ciencia y Competitividad and Ministerio de Ciencia e Innovación, Gobierno de España under the scope of the PolyDeep and PolyDeepAdvance projects (DPI2017-87494-R, PDC2021-121644-I00); (ii) by Consellería de Educación, Universidades e Formación Profesional (Xunta de Galicia) under the scope of the strategic funding ED431C2018/55-GRC and ED431C 2022/03-GRC Competitive Reference Group.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by Ethics Committee of Pontevedra-Vigo-Ourense research (code 2017/427 on 28 February 2018).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Publicly available datasets were analyzed in this study. Our not-polyp datasets are a subset of PIBAdb Colorectal Polyp Image Cohort, which can be requested here: https://www.iisgaliciasur.es/home/biobanco/cohorte-pibadb. Third-party datasets are available upon request to their owners.

Acknowledgments

We want to acknowledge Jorge Bernal for the support in accessing the CVC datasets and Heyato Itoh for giving access to the SUN dataset. The PICCOLO dataset included in this study was provided by the Basque Biobank (http://www.biobancovasco.org, accessed on January 2022). SING group thanks the CITI (Centro de Investigación, Transferencia e Innovación) from the University of Vigo for hosting its IT infrastructure.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, P.; Berzin, T.M.; Brown, J.R.G.; Bharadwaj, S.; Becq, A.; Xiao, X.; Liu, P.; Li, L.; Song, Y.; Zhang, D.; et al. Real-time automatic detection system increases colonoscopic polyp and adenoma detection rates: A prospective randomised controlled study. Gut 2019, 68, 1813–1819. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Gong, D.; Wu, L.; Zhang, J.; Mu, G.; Shen, L.; Liu, J.; Wang, Z.; Zhou, W.; An, P.; Huang, X.; et al. Detection of colorectal adenomas with a real-time computer-aided system (ENDOANGEL): A randomised controlled study. Lancet Gastroenterol. Hepatol. 2020, 5, 352–361. [Google Scholar] [CrossRef]
  3. Wang, P.; Liu, X.; Berzin, T.M.; Glissen Brown, J.R.; Liu, P.; Zhou, C.; Lei, L.; Li, L.; Guo, Z.; Lei, S.; et al. Effect of a deep-learning computer-aided detection system on adenoma detection during colonoscopy (CADe-DB trial): A double-blind randomised study. Lancet Gastroenterol. Hepatol. 2020, 5, 343–351. [Google Scholar] [CrossRef]
  4. Liu, W.-N.; Zhang, Y.-Y.; Bian, X.-Q.; Wang, L.-J.; Yang, Q.; Zhang, X.-D.; Huang, J. Study on detection rate of polyps and adenomas in artificial-intelligence-aided colonoscopy. Saudi J. Gastroenterol. 2020, 26, 13. [Google Scholar] [CrossRef] [PubMed]
  5. Su, J.-R.; Li, Z.; Shao, X.-J.; Ji, C.-R.; Ji, R.; Zhou, R.-C.; Li, G.-C.; Liu, G.-Q.; He, Y.-S.; Zuo, X.-L.; et al. Impact of a real-time automatic quality control system on colorectal polyp and adenoma detection: A prospective randomized controlled study (with videos). Gastrointest. Endosc. 2020, 91, 415–424.e4. [Google Scholar] [CrossRef] [PubMed]
  6. Repici, A.; Badalamenti, M.; Maselli, R.; Correale, L.; Radaelli, F.; Rondonotti, E.; Ferrara, E.; Spadaccini, M.; Alkandari, A.; Fugazza, A.; et al. Efficacy of Real-Time Computer-Aided Detection of Colorectal Neoplasia in a Randomized Trial. Gastroenterology 2020, 159, 512–520.e7. [Google Scholar] [CrossRef]
  7. Hann, A.; Troya, J.; Fitting, D. Current status and limitations of artificial intelligence in colonoscopy. United Eur. Gastroenterol. J. 2021, 9, 527–533. [Google Scholar] [CrossRef]
  8. Nogueira-Rodríguez, A.; Domínguez-Carbajales, R.; López-Fernández, H.; Iglesias, Á.; Cubiella, J.; Fdez-Riverola, F.; Reboiro-Jato, M.; Glez-Peña, D. Deep Neural Networks approaches for detecting and classifying colorectal polyps. Neurocomputing 2021, 423, 721–734. [Google Scholar] [CrossRef]
  9. Viscaino, M.; Torres Bustos, J.; Muñoz, P.; Auat Cheein, C.; Cheein, F.A. Artificial intelligence for the early detection of colorectal cancer: A comprehensive review of its advantages and misconceptions. World J. Gastroenterol. 2021, 27, 6399–6414. [Google Scholar] [CrossRef]
  10. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef] [Green Version]
  11. Nogueira-Rodríguez, A.; Domínguez-Carbajales, R.; Campos-Tato, F.; Herrero, J.; Puga, M.; Remedios, D.; Rivas, L.; Sánchez, E.; Iglesias, Á.; Cubiella, J.; et al. Real-time polyp detection model using convolutional neural networks. Neural Comput. Appl. 2021, 34, 10375–10396. [Google Scholar] [CrossRef]
  12. Nogueira-Rodríguez, A.; Reboiro-Jato, M.; Glez-Peña, D.; López-Fernández, H. Performance of Convolutional Neural Networks for Polyp Localization on Public Colonoscopy Image Datasets. Diagnostics 2022, 12, 898. [Google Scholar] [CrossRef] [PubMed]
  13. Misawa, M.; Kudo, S.; Mori, Y.; Hotta, K.; Ohtsuka, K.; Matsuda, T.; Saito, S.; Kudo, T.; Baba, T.; Ishida, F.; et al. Development of a computer-aided detection system for colonoscopy and a publicly accessible large colonoscopy video database (with video). Gastrointest. Endosc. 2021, 93, 960–967.e3. [Google Scholar] [CrossRef]
  14. Li, K.; Fathan, M.I.; Patel, K.; Zhang, T.; Zhong, C.; Bansal, A.; Rastogi, A.; Wang, J.S.; Wang, G. Colonoscopy polyp detection and classification: Dataset creation and comparative evaluations. PLoS ONE 2021, 16, e0255809. [Google Scholar] [CrossRef] [PubMed]
  15. Ma, Y.; Chen, X.; Cheng, K.; Li, Y.; Sun, B. LDPolypVideo Benchmark: A Large-Scale Colonoscopy Video Dataset of Diverse Polyps. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2021; de Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C., Eds.; Springer International Publishing: Cham, Germany, 2021; Volume 12905, pp. 387–396. [Google Scholar]
  16. Pogorelov, K.; Schmidt, P.T.; Riegler, M.; Halvorsen, P.; Randel, K.R.; Griwodz, C.; Eskeland, S.L.; de Lange, T.; Johansen, D.; Spampinato, C.; et al. KVASIR: A Multi-Class Image Dataset for Computer Aided Gastrointestinal Disease Detection. In Proceedings of the 8th ACM on Multimedia Systems Conference—MMSys’17, Taipei, Taiwan, 20–23 June 2017; ACM Press: Taipei, Taiwan, 2017; pp. 164–169. [Google Scholar]
  17. Bernal, J.; Sánchez, F.J.; Fernández-Esparrach, G.; Gil, D.; Rodríguez, C.; Vilariño, F. WM-DOVA maps for accurate polyp highlighting in colonoscopy: Validation vs. saliency maps from physicians. Comput. Med. Imaging Graph. 2015, 43, 99–111. [Google Scholar] [CrossRef] [PubMed]
  18. Bernal, J.; Sánchez, J.; Vilariño, F. Towards automatic polyp detection with a polyp appearance model. Pattern Recognit. 2012, 45, 3166–3182. [Google Scholar] [CrossRef]
  19. Vázquez, D.; Bernal, J.; Sánchez, F.J.; Fernández-Esparrach, G.; López, A.M.; Romero, A.; Drozdzal, M.; Courville, A. A Benchmark for Endoluminal Scene Segmentation of Colonoscopy Images. J. Healthc. Eng. 2017, 2017, 4037190. [Google Scholar] [CrossRef] [Green Version]
  20. Silva, J.; Histace, A.; Romain, O.; Dray, X.; Granado, B. Toward embedded detection of polyps in WCE images for early diagnosis of colorectal cancer. Int. J. Comput. Assist. Radiol. Surg. 2014, 9, 283–293. [Google Scholar] [CrossRef]
  21. Angermann, Q.; Bernal, J.; Sánchez-Montes, C.; Hammami, M.; Fernández-Esparrach, G.; Dray, X.; Romain, O.; Sánchez, F.J.; Histace, A. Towards Real-Time Polyp Detection in Colonoscopy Videos: Adapting Still Frame-Based Methodologies for Video Sequences Analysis. In Computer Assisted and Robotic Endoscopy and Clinical Image-Based Procedures; Cardoso, M.J., Arbel, T., Luo, X., Wesarg, S., Reichl, T., González Ballester, M.Á., McLeod, J., Drechsler, K., Peters, T., Erdt, M., et al., Eds.; Springer International Publishing: Cham, Germany, 2017; pp. 29–41. [Google Scholar]
  22. Bernal, J.J.; Histace, A.; Masana, M.; Angermann, Q.; Sánchez-Montes, C.; Rodriguez, C.; Hammami, M.; Garcia-Rodriguez, A.; Córdova, H.; Romain, O.; et al. Polyp Detection Benchmark in Colonoscopy Videos using GTCreator: A Novel Fully Configurable Tool for Easy and Fast Annotation of Image Databases. In Proceedings of the 32nd CARS Conference, Berlin, Germany, 20–23 June 2018. [Google Scholar]
  23. Sánchez-Peralta, L.F.; Pagador, J.B.; Picón, A.; Calderón, Á.J.; Polo, F.; Andraka, N.; Bilbao, R.; Glover, B.; Saratxaga, C.L.; Sánchez-Margallo, F.M. PICCOLO White-Light and Narrow-Band Imaging Colonoscopic Dataset: A Performance Comparative of Models and Datasets. Appl. Sci. 2020, 10, 8501. [Google Scholar] [CrossRef]
  24. Houwen, B.B.S.L.; Nass, K.J.; Vleugels, J.L.A.; Fockens, P.; Hazewinkel, Y.; Dekker, E. Comprehensive review of publicly available colonoscopic imaging databases for artificial intelligence research: Availability, accessibility, and usability. Gastrointest. Endosc. 2022, 97, 184–199.e6. [Google Scholar] [CrossRef]
  25. López-Fernández, H.; Graña-Castro, O.; Nogueira-Rodríguez, A.; Reboiro-Jato, M.; Glez-Peña, D. Compi: A framework for portable and reproducible pipelines. PeerJ Comput. Sci. 2021, 7, e593. [Google Scholar] [CrossRef] [PubMed]
  26. Nogueira-Rodríguez, A.; López-Fernández, H.; Graña-Castro, O.; Reboiro-Jato, M.; Glez-Peña, D. Compi Hub: A Public Repository for Sharing and Discovering Compi Pipelines. In Proceedings of the Practical Applications of Computational Biology & Bioinformatics, 14th International Conference (PACBB 2020), L’Aquila, Italy, 17–19 June 2020; Panuccio, G., Rocha, M., Fdez-Riverola, F., Mohamad, M.S., Casado-Vara, R., Eds.; Springer International Publishing: Cham, Germany, 2021; pp. 51–59. [Google Scholar]
Figure 1. Not-polyp datasets created from the original polyp images dataset [11].
Figure 1. Not-polyp datasets created from the original polyp images dataset [11].
Diagnostics 13 00966 g001
Figure 2. Random sample of the not-polyp images included in the injected datasets.
Figure 2. Random sample of the not-polyp images included in the injected datasets.
Diagnostics 13 00966 g002
Figure 3. Recall of the five trained models (blue) and their precision in both the original test partition (pink) and the same with the addition of 15% of not-polyp images (dark red) [11].
Figure 3. Recall of the five trained models (blue) and their precision in both the original test partition (pink) and the same with the addition of 15% of not-polyp images (dark red) [11].
Diagnostics 13 00966 g003
Figure 4. F1 of the five trained models in both the original test partition (light green) and the same with the addition of 15% of not-polyp images (dark green) [11].
Figure 4. F1 of the five trained models in both the original test partition (light green) and the same with the addition of 15% of not-polyp images (dark green) [11].
Diagnostics 13 00966 g004
Figure 5. Comparison between the F1 performances obtained by our baseline model in the ten public datasets (X axis) and the F1 performances obtained by the four new models trained with not-polyp images (Y axis). Public datasets containing not-polyp images are marked in light blue. Points above the diagonal line represent benefit.
Figure 5. Comparison between the F1 performances obtained by our baseline model in the ten public datasets (X axis) and the F1 performances obtained by the four new models trained with not-polyp images (Y axis). Public datasets containing not-polyp images are marked in light blue. Points above the diagonal line represent benefit.
Diagnostics 13 00966 g005
Figure 6. Relative F1 differences in the same test dataset with respect to our baseline model (trained without not-polyp images), and grouped test dataset type (i.e., containing or not containing not-polyp images).
Figure 6. Relative F1 differences in the same test dataset with respect to our baseline model (trained without not-polyp images), and grouped test dataset type (i.e., containing or not containing not-polyp images).
Diagnostics 13 00966 g006
Figure 7. The five models with three sample frames of the polyp video analyzed (Supplementary Video S1). The false positive identifications are marked with red cross marks. The false negatives are marked in white boxes (added manually for illustration purposes) [11].
Figure 7. The five models with three sample frames of the polyp video analyzed (Supplementary Video S1). The false positive identifications are marked with red cross marks. The false negatives are marked in white boxes (added manually for illustration purposes) [11].
Diagnostics 13 00966 g007
Figure 8. The five models with three sample frames of the not-polyp video segment analyzed (Supplementary Video S2). While the baseline model makes incorrect predictions (red cross marks) in the three cases, most of the not-polyp-trained models do not [11].
Figure 8. The five models with three sample frames of the not-polyp video segment analyzed (Supplementary Video S2). While the baseline model makes incorrect predictions (red cross marks) in the three cases, most of the not-polyp-trained models do not [11].
Diagnostics 13 00966 g008
Figure 9. The five models facing different situations found in two exploration videos included in PIBAdb (provided as Supplementary Videos S3 and S4), including close to the wall of the colon (first row), water jet (fourth row), or blurry images due to water jet (fifth row). False positives are marked as red cross marks [11].
Figure 9. The five models facing different situations found in two exploration videos included in PIBAdb (provided as Supplementary Videos S3 and S4), including close to the wall of the colon (first row), water jet (fourth row), or blurry images due to water jet (fifth row). False positives are marked as red cross marks [11].
Diagnostics 13 00966 g009
Table 1. Summary of the ten public colonoscopy image datasets for polyp location including: the name and publication/s related with the dataset (Dataset), the year when the dataset was published (Year), the image resolution (Resolution), the number of images in the dataset (No. of images), the method used to provide image location (Ground Truth), and whether the dataset includes images without polyps and the percentage of images with polyps in case they are included (Without polyps?).
Table 1. Summary of the ten public colonoscopy image datasets for polyp location including: the name and publication/s related with the dataset (Dataset), the year when the dataset was published (Year), the image resolution (Resolution), the number of images in the dataset (No. of images), the method used to provide image location (Ground Truth), and whether the dataset includes images without polyps and the percentage of images with polyps in case they are included (Without polyps?).
DatasetYearResolutionNo. of ImagesGround TruthWithout Polyps?
CVC-ColonDB [18,19]2012574 × 500300MaskNo
ETIS-Larib [20]20141225 × 966196MaskNo
CVC-ClinicDB [17]2015384 × 288612MaskNo
CVC-ClinicVideoDB [21,22]2017384 × 28811,954MaskYes (16.14%)
CVC-PolypHD [18,19]20181920 × 108056MaskNo
Kvasir-SEG [16]2020332 × 487 to 1920 × 10721000Bounding box and maskNo
PICCOLO [23]2020854 × 480,
1920 × 1080
3433MaskYes (1.25%)
KUMC [14]2021Various resolutions37,899Bounding boxYes (5.08%)
SUN [13]20211240 × 108049,136Bounding boxNo *
LDPolypVideo [15]2021560 × 48040,186Bounding boxYes (15.7%)
* The SUN dataset contains 109,554 frames without polyps that were not downloaded for our experiments.
Table 2. Performance metrics in each validation partition for the selected models.
Table 2. Performance metrics in each validation partition for the selected models.
Confidence Score ThresholdRecallPrecision F 1 AP
Nogueira et al. [11]0.190.9050.9120.9090.920
Model injection 2%0.240.8970.9170.9070.920
Model injection 5%0.180.9110.9150.9130.923
Model injection 10%0.220.8840.9180.9010.918
Model injection 15%0.150.9030.9200.9120.923
Table 3. Performance metrics of each model on the original test partition of Nogueira et al., which does not contain not-polyp images.
Table 3. Performance metrics of each model on the original test partition of Nogueira et al., which does not contain not-polyp images.
Training DatasetRecallPrecision F 1
Nogueira et al. [11]0.8720.8900.881
Model injection 2%0.8670.8940.880
Model injection 5%0.8750.9020.888
Model injection 10%0.8500.9040.876
Model injection 15%0.8800.9100.895
Table 4. Precision metrics obtained with different models against different not-polyp injections into the original test partition.
Table 4. Precision metrics obtained with different models against different not-polyp injections into the original test partition.
ModelTest Partitions (Percentage of Not-Polyp Augmentation)
2%5%10%15%
Nogueira et al. [11]0.8820.8710.8520.836
Model injection 2%0.8900.8860.8790.871
Model injection 5%0.8990.8950.8880.881
Model injection 10%0.9020.9010.8970.895
Model injection 15%0.9090.9070.9040.901
Table 5. F1 score of the five trained models in the 10 public datasets. Averages are provided for all datasets, datasets with both polyp and not-polyp images (4 datasets), and for datasets with only polyp images (6 datasets).
Table 5. F1 score of the five trained models in the 10 public datasets. Averages are provided for all datasets, datasets with both polyp and not-polyp images (4 datasets), and for datasets with only polyp images (6 datasets).
Training Dataset
Nogueira et al. [11]Not-Polyp Injection
2%5%10%15%
Datasets with both polyp and not-polyp images
LDPolypVideo0.5220.5630.5160.4910.564
ClinicVideoDB0.7740.8030.8130.8090.800
KUMC0.8180.8110.8190.7620.831
PICCOLO0.6670.6010.6910.7590.691
average0.6950.6940.7100.7050.722
Datasets only with polyp images
ClinicDB0.8450.8430.8670.7860.824
ColonDB0.8260.8480.8830.6890.797
SUN0.8050.7640.7380.7650.746
KVASIR0.8070.8000.7970.8400.830
Etis-Larib0.7180.7320.6790.5940.685
PolypHD0.8000.7290.8260.8200.820
average0.8000.7860.7990.7490.784
global average0.7580.7490.7630.7310.759
Table 6. Summary of the network predictions in both polyp and not-polyp videos. For the polyp video, the total number of distinct frames with at least one bounding box is reported, as a proxy of the recall (assuming that, at least one of the bounding boxes in the frame is placed over the polyp). For the not-polyp video, the total number of bounding boxes is reported, which are all false positives. Relative changes if these metrics are also reported among different models.
Table 6. Summary of the network predictions in both polyp and not-polyp videos. For the polyp video, the total number of distinct frames with at least one bounding box is reported, as a proxy of the recall (assuming that, at least one of the bounding boxes in the frame is placed over the polyp). For the not-polyp video, the total number of bounding boxes is reported, which are all false positives. Relative changes if these metrics are also reported among different models.
Polyp Video (450 Frames)Not-Polyp Video (350 Frames)
Frames with at Least One
Bounding Box
Relative ChangeTotal Bounding Boxes
(False Positives)
Relative Change
Nogueira et al. [11]404 (89.78%)-90-
Model injection 2%418 (92.89%)+3.47%43−52.22%
Model injection 5%419 (93.11%)+3.71%24−73.33%
Model injection 10%416 (92.44%)+2.97%16−82.22%
Model injection 15%407 (90.44%)+0.74%63−30.00%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nogueira-Rodríguez, A.; Glez-Peña, D.; Reboiro-Jato, M.; López-Fernández, H. Negative Samples for Improving Object Detection—A Case Study in AI-Assisted Colonoscopy for Polyp Detection. Diagnostics 2023, 13, 966. https://doi.org/10.3390/diagnostics13050966

AMA Style

Nogueira-Rodríguez A, Glez-Peña D, Reboiro-Jato M, López-Fernández H. Negative Samples for Improving Object Detection—A Case Study in AI-Assisted Colonoscopy for Polyp Detection. Diagnostics. 2023; 13(5):966. https://doi.org/10.3390/diagnostics13050966

Chicago/Turabian Style

Nogueira-Rodríguez, Alba, Daniel Glez-Peña, Miguel Reboiro-Jato, and Hugo López-Fernández. 2023. "Negative Samples for Improving Object Detection—A Case Study in AI-Assisted Colonoscopy for Polyp Detection" Diagnostics 13, no. 5: 966. https://doi.org/10.3390/diagnostics13050966

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop