skip to main content
10.1145/3651623.3651628acmotherconferencesArticle/Chapter ViewAbstractPublication PagesapitConference Proceedingsconference-collections
research-article
Open Access

A Self-Supervised Semantic Segmentation Method for Identifying Barrett' s Esophagus in Endoscopic Images

Published:07 May 2024Publication History

ABSTRACT

Barrett's esophagus is considered a precancerous condition that may lead to esophageal cancer. The condition is usually diagnosed through an endoscopy with biopsy. This careful imaging examination is quite labor-intensive and usually lacks of diagnostic consistency. In this paper, a computer-aided diagnosis (CAD) method was proposed to assist pathologist in Barrett's esophagus diagnosis from endoscopic images. The proposed semantic segmentation was built based on the U-Net architecture, which features capturing detailed spatial information and context to produce accurate pixel-wise predictions. An autoencoder, which is quite capable of learning a compact and meaningful representation of the input data, is incorporated to achieve self-supervised learning. A two-stage pretraining for the encoder and the decoder is adopted for better performance and reduced data requirements. The proposed method can extract features and patterns from unlabeled data without requiring human annotation or labels, which is very suitable in many biomedical image analyses. The experimental results show that the segmentation accuracy with mean pixel accuracy, mean dice coefficient, and mIoU reaches 98.20%, 87.96%, and 83.18% respectively, indicating that the proposed method has a good performance on the identification of Barrett's esophagus in endoscopic images.

References

  1. Fitzgerald RC, "Barrett's oesophagus and oesophageal adenocarcinoma: how does acid interfere with cell proliferation and differentiation,” Gut 2005; 54 (Suppl 1): i21-6.Google ScholarGoogle ScholarCross RefCross Ref
  2. Bus P, Siersema PD and van Baal JW, “Cell culture models for studying the development of Barrett's esophagus: a systematic review,” Cell Oncol (Dordr), 2012, 35: 149-61.Google ScholarGoogle ScholarCross RefCross Ref
  3. World Cancer Research Fund International, Oesophageal cancer statistics, retrieved September 9, 2023 from https://www.wcrf.org/cancer-trends/oesophageal-cancer-statistics/Google ScholarGoogle Scholar
  4. Y.-C. Lee , "Comparative analysis between psychological and endoscopic profiles in patients with gastroesophageal reflux disease: A prospective study based on screening endoscopy," J. Gastroenterol. Hepatol., vol. 21, no. 5, pp. 798–804, 2006.Google ScholarGoogle ScholarCross RefCross Ref
  5. H. Cheng , "Supplementation of Los Angeles classification with esophageal mucosa index of hemoglobin can predict the treatment response of erosive reflux esophagitis," Surg. Endosc., vol. 25, no. 8, pp. 2478–2486, 2011.Google ScholarGoogle ScholarCross RefCross Ref
  6. D. Armstrong , “The endoscopic assessment of esophagitis: A progress report on observer agreement,” Gastroenterology, vol. 111, no. 1, pp. 85–92, 1996Google ScholarGoogle ScholarCross RefCross Ref
  7. Ait Skourt B, El Hassani A, Majda A. Lung CT Image segmentation using deep neural networks. Procedia Comput Sci. 2018;127:109–13. https://doi. org/ 10. 1016/j. procs. 2018. 01. 104.Google ScholarGoogle Scholar
  8. Mittal M, Goyal LM, Kaur S, Kaur I, Verma A, Jude Hemanth D. Deep learning based enhanced tumor segmentation approach for MR brain images. Appl Soft Comput. 2019;78:346–54. https:// doi. org/ 10. 1016/j. asoc.2019.02.036.Google ScholarGoogle Scholar
  9. Pan W, Li X, Wang W, Zhou L, Wu J, Ren T, Liu C, Lv M, Su S, Tang Y. Identification of Barrett's esophagus in endoscopic images using deep learning. BMC Gastroenterol. 2021 Dec 17;21(1):479. doi: 10.1186/s12876-021-02055-2. PMID: 34920705; PMCID: PMC8684213Google ScholarGoogle ScholarCross RefCross Ref
  10. Liu G, Hua J, Wu Z, Meng T, Sun M, Huang P, He X, Sun W, Li X, Chen Y: Automatic classification of esophageal lesions in endoscopic images using a convolutional neural network. Ann Transl Med 2020,8(7):486.https:// doi. org/ 10. 21037/ atm.2020.03.24Google ScholarGoogle Scholar
  11. de Groof J, van der Sommen F, van der Putten J, Struyvenberg MR, Zinger S, Curvers WL, Pech O, Meining A, Neuhaus H, Bisschops R, The Argos project: The development of a computer-aided detection system to improve detection of Barrett's neoplasia on white light endoscopy. United Eur Gastroenterol J. 2019;7(4):538–47. https:// doi.org/10.1177/20506 40619 837443.Google ScholarGoogle ScholarCross RefCross Ref
  12. Martin Thoma, "A survey of semantic segmentation." arXiv preprint arXiv:1602.06541, 2016.Google ScholarGoogle Scholar
  13. J. Walsh, A. Othmani, M. Jain, and S. Dev, “Using U-Net network for efficient brain tumor segmentation in MRI images,” Healthcare Analytics, vol. 2, p. 100098, Nov. 2022.Google ScholarGoogle ScholarCross RefCross Ref
  14. Olaf Ronneberger, Philipp Fischer, and Thomas Brox, "U-Net: Convolutional networks for biomedical image segmentation," International Conference on Medical image computing and computer-assisted intervention. Springer, Cham, 2015.Google ScholarGoogle Scholar
  15. Long, Jonathan, Evan Shelhamer, and Trevor Darrell, "Fully convolutional networks for semantic segmentation," Proceedings of the IEEE conference on computer vision and pattern recognition. 2015.Google ScholarGoogle Scholar
  16. Liang-Chieh Chen , “Encoder-decoder with atrous separable convolution for semantic image segmentation,” Proceedings of the European conference on computer vision (ECCV), 2018.Google ScholarGoogle Scholar
  17. Chen, Xinlei, and Kaiming He. "Exploring simple siamese representation learning." Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2021.Google ScholarGoogle Scholar
  18. Chen, T., Kornblith, S., Norouzi, M., and Hinton, G. "A simple framework for contrastive learning of visual representations." International conference on machine learning. PMLR, 2020.Google ScholarGoogle Scholar
  19. Grill, Jean-Bastien, "Bootstrap your own latent-a new approach to self-supervised learning." Advances in neural information processing systems 33 (2020): 21271-21284.Google ScholarGoogle Scholar
  20. Bank, Dor, Noam Koenigstein, and Raja Giryes. "Autoencoders." Machine Learning for Data Science Handbook: Data Mining and Knowledge Discovery Handbook (2023): 353-374.Google ScholarGoogle Scholar
  21. He, Kaiming, "Deep residual learning for image recognition." Proceedings of the IEEE conference on computer vision and pattern recognition. 2016.Google ScholarGoogle Scholar
  22. S. Liu and W. Deng, "Very deep convolutional neural network based image classification using small training sample size," 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR), Kuala Lumpur, Malaysia, 2015, pp. 730-734, doi: 10.1109/ACPR.2015.7486599Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. A Self-Supervised Semantic Segmentation Method for Identifying Barrett' s Esophagus in Endoscopic Images

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      APIT '24: Proceedings of the 2024 6th Asia Pacific Information Technology Conference
      January 2024
      105 pages
      ISBN:9798400716218
      DOI:10.1145/3651623

      Copyright © 2024 Owner/Author

      This work is licensed under a Creative Commons Attribution International 4.0 License.

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 7 May 2024

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed limited
    • Article Metrics

      • Downloads (Last 12 months)10
      • Downloads (Last 6 weeks)10

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format .

    View HTML Format