skip to main content
10.1145/3584376.3584613acmotherconferencesArticle/Chapter ViewAbstractPublication PagesricaiConference Proceedingsconference-collections
research-article

A comparative study on fetal head circumference measurement from ultrasound images using deep learning models

Authors Info & Claims
Published:19 April 2023Publication History

ABSTRACT

Ultrasound imaging is the most commonly used imaging modality for the prenatal examination of pregnant women, with real-time imaging and no radiation characteristics. Through a ultrasound image of the fetal head, doctors can measure the fetal head circumference (HC) to evaluate fetal growth and potential delivery mode. In practice, fetal HC is usually measured manually by doctors based on ultrasound images. Manual measurement of fetal HC is subjective and time-consuming, which has a negative impact on measurement accuracy and efficiency. At present, deep learning is widely investigated in the medical field. Many researchers apply deep learning to measuring fetal HC to assist doctors to accurately and quickly completing the measurement of fetal HC. In this paper, we compare the performance of eight deep learning models (U-Net, Attention U-Net, GINet, global reasoning unit (GloRe), SegFormer, Segmenter, BiSeNet V2, and short-term dense concatenate network (STDC)) on two fetal HC measurement datasets. SegFormer achieves the best results in Dice similarity coefficient (DSC), Hausdorff distance (HD), and absolute Difference (ADF). The performance of Attention U-Net is slightly worse than that of SegFormer.

References

  1. M. Lipschuetz, S. M. Cohen, A. Israel, J. Baron, S. Porat, D. V. Valsky, O. Yagel, H. Amsalem, D. Kabiri, Y. Gilboa, , “Sono-graphic large fetal head circumference and risk of cesarean delivery,” American Journal of Obstetrics and Gynecology, vol. 218, no. 3, pp. 339–e1, 2018.Google ScholarGoogle ScholarCross RefCross Ref
  2. B. Deloison, G. E. Chalouhi, J.-P. Bernard, Y. Ville, and L. J. Salomon, “Outcomes of fetuses with small head circumference on second-trimester ultrasonography,” Prenatal Diagnosis, vol. 32, no. 9, pp. 869–874, 2012.Google ScholarGoogle Scholar
  3. W. Zeng, J. Luo, J. Cheng, and Y. Lu, “Efficient fetal ultrasound image segmentation for automatic head circumference mea-surement using a lightweight deep convolutional neural network,” Medical Physics, 2022.Google ScholarGoogle Scholar
  4. R. Napolitano, V. Donadono, E. Ohuma, C. Knight, S. Wanyonyi, B. Kemp, T. Norris, and A. Papageorghiou, “Scientific basis for standardization of fetal head measurements by ultrasound: A reproducibility study,” Ultrasound in Obstetrics & Gynecology, vol. 48, no. 1, pp. 80–85, 2016.Google ScholarGoogle ScholarCross RefCross Ref
  5. W. Lu, J. Tan, and R. Floyd, “Automated fetal head detection and measurement in ultrasound images by iterative randomized hough transform,” Ultrasound in Medicine & Biology, vol. 31, no. 7, pp. 929–936, 2005.Google ScholarGoogle ScholarCross RefCross Ref
  6. E. A. Anto, B. Amoah, and A. Crimi, “Segmentation of ultrasound images of fetal anatomic structures using random forest for low-cost settings,” in 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), pp. 793–796, IEEE, 2015.Google ScholarGoogle ScholarCross RefCross Ref
  7. J. Perez-Gonzalez, J. Mu ́noz, M. Porras, F. Ar ́ambula-Cos ́ıo, and V. Medina-Ba ́nuelos, “Automatic fetal head measurements from ultrasound images using optimal ellipse detection and texture maps,” in VI Latin American Congress on Biomedical Engineering CLAIB 2014, Paran ́a, Argentina 29, 30 & 31 October 2014, pp. 329–332, Springer, 2015.Google ScholarGoogle Scholar
  8. T. L. van den Heuvel, D. de Bruijn, C. L. de Korte, and B. v. Ginneken, “Automated measurement of fetal head circumference using 2D ultrasound images,” PloS One, vol. 13, no. 8, p. e0200412, 2018.Google ScholarGoogle ScholarCross RefCross Ref
  9. J. Li, Y. Wang, B. Lei, J.-Z. Cheng, J. Qin, T. Wang, S. Li, and D. Ni, “Automatic fetal head circumference measurement in ultrasound using random forest and fast ellipse fitting,” IEEE Journal of Biomedical and Health Informatics, vol. 22, no. 1, pp. 215–223, 2017.Google ScholarGoogle ScholarCross RefCross Ref
  10. D. Tian, B. Zhu, J. Wang, L. Kong, B. Gao, Y. Wang, D. Xu, R. Zhang, and Y. Yao, “Brachial plexus nerve trunk recognition from ultrasound images: A comparative study of deep learning models,” IEEE Access, vol. 10, pp. 82003–82014, 2022.Google ScholarGoogle ScholarCross RefCross Ref
  11. N. Wu, J. Phang, J. Park, Y. Shen, Z. Huang, M. Zorin, S. Jastrzebski, T. F ́evry, J. Katsnelson, E. Kim, , “Deep neural networks improve radiologists’ performance in breast cancer screening,” IEEE Transactions on Medical Imaging, vol. 39, no. 4, pp. 1184–1194, 2019.Google ScholarGoogle ScholarCross RefCross Ref
  12. L. Ma, H. Xue, G. Yang, Z. Zhang, C. Li, Y. Yao, and Y. Teng, “Scrdn: Residual dense network with self-calibrated convolutions for low dose CT image denoising,” Nuclear Instruments and Methods in Physics Research Section A: Accelerators, Spectrometers, Detectors and Associated Equipment, p. 167625, 2022.Google ScholarGoogle Scholar
  13. S. M. Amini, “Head circumference measurement with deep learning approach based on multi-scale ultrasound images,” Multimedia Tools and Applications, pp. 1–13, 2022.Google ScholarGoogle Scholar
  14. M. C. Fiorentino, S. Moccia, M. Capparuccini, S. Giamberini, and E. Frontoni, “A regression framework to head-circumference delineation from US fetal images,” Computer Methods and Programs in Biomedicine, vol. 198, p. 105771, 2021.Google ScholarGoogle ScholarCross RefCross Ref
  15. P. Li, H. Zhao, P. Liu, and F. Cao, “Automated measurement network for accurate segmentation and parameter modification in fetal head ultrasound images,” Medical & Biological Engineering & Computing, vol. 58, no. 11, pp. 2879–2892, 2020.Google ScholarGoogle ScholarCross RefCross Ref
  16. E. L. Skeika, M. R. Da Luz, B. J. T. Fernandes, H. V. Siqueira, and M. L. S. C. De Andrade, “Convolutional neural network to detect and measure fetal skull circumference in ultrasound imaging,” IEEE ACCESS, vol. 8, pp. 191519–191529, 2020.Google ScholarGoogle ScholarCross RefCross Ref
  17. Z. Sobhaninia, S. Rafiei, A. Emami, N. Karimi, K. Najarian, S. Samavi, and S. R. Soroushmehr, “Fetal ultrasound image segmentation for measuring biometric parameters using multi-task deep learning,” in 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), pp. 6545–6548, IEEE, 2019.Google ScholarGoogle ScholarCross RefCross Ref
  18. Z. Sobhaninia, S. Rafiei, A. Emami, N. Karimi, K. Najarian, S. Samavi, and S. R. Soroushmehr, “Fetal ultrasound image segmentation for measuring biometric parameters using multi-task deep learning,” in 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), pp. 6545–6548, IEEE, 2019.Google ScholarGoogle ScholarCross RefCross Ref
  19. S. Johnson, N. Sebire, R. Snijders, S. Tunkel, and K. Nicolaides, “Ultrasound screening for anencephaly at 10–14 weeks of gestation,” Ultrasound in Obstetrics and Gynecology: The Official Journal of the International Society of Ultrasound in Obstetrics and Gynecology, vol. 9, no. 1, pp. 14–16, 1997.Google ScholarGoogle ScholarCross RefCross Ref
  20. O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical Image Computing and Computer-assisted Intervention, pp. 234–241, Springer, 2015.Google ScholarGoogle ScholarCross RefCross Ref
  21. O. Oktay, J. Schlemper, L. L. Folgoc, M. Lee, M. Heinrich, K. Misawa, K. Mori, S. McDonagh, N. Y. Hammerla, B. Kainz, , “Attention u-net: Learning where to look for the pancreas,” arXiv preprint arXiv:1804.03999, 2018.Google ScholarGoogle Scholar
  22. T. Wu, Y. Lu, Y. Zhu, C. Zhang, M. Wu, Z. Ma, and G. Guo, “GINet: Graph interaction network for scene parsing,” in European Conference on Computer Vision, pp. 34–51, Springer, 2020.Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Y. Chen, M. Rohrbach, Z. Yan, Y. Shuicheng, J. Feng, and Y. Kalantidis, “Graph-based global reasoning networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 433–442, 2019.Google ScholarGoogle ScholarCross RefCross Ref
  24. R. Strudel, R. Garcia, I. Laptev, and C. Schmid, “Segmenter: Transformer for semantic segmentation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 7262–7272, 2021.Google ScholarGoogle ScholarCross RefCross Ref
  25. E. Xie, W. Wang, Z. Yu, A. Anandkumar, J. M. Alvarez, and P. Luo, “Segformer: Simple and efficient design for semantic segmentation with transformers,” Advances in Neural Information Processing Systems, vol. 34, pp. 12077–12090, 2021.Google ScholarGoogle Scholar
  26. C. Yu, C. Gao, J. Wang, G. Yu, C. Shen, and N. Sang, “BiSeNet V2: Bilateral network with guided aggregation for real-time semantic segmentation,” International Journal of Computer Vision, vol. 129, no. 11, pp. 3051–3068, 2021.Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. M. Fan, S. Lai, J. Huang, X. Wei, Z. Chai, J. Luo, and X. Wei, “Rethinking BiSeNet for real-time semantic segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9716–9725, 2021.Google ScholarGoogle ScholarCross RefCross Ref
  28. Y. Liu, L. Chu, G. Chen, Z. Wu, Z. Chen, B. Lai, and Y. Hao, “PaddleSeg: A High-Efficient Development Toolkit for Image Segmentation,” 2021.Google ScholarGoogle Scholar

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Other conferences
    RICAI '22: Proceedings of the 2022 4th International Conference on Robotics, Intelligent Control and Artificial Intelligence
    December 2022
    1396 pages
    ISBN:9781450398343
    DOI:10.1145/3584376

    Copyright © 2022 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 19 April 2023

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article
    • Research
    • Refereed limited

    Acceptance Rates

    Overall Acceptance Rate140of294submissions,48%
  • Article Metrics

    • Downloads (Last 12 months)26
    • Downloads (Last 6 weeks)1

    Other Metrics

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format .

View HTML Format