Skip to main content

A validation of an entropy-based artificial intelligence for ultrasound data in breast tumors

Abstract

Background

The application of artificial intelligence (AI) in the ultrasound (US) diagnosis of breast cancer (BCa) is increasingly prevalent. However, the impact of US-probe frequencies on the diagnostic efficacy of AI models has not been clearly established.

Objectives

To explore the impact of using US-video of variable frequencies on the diagnostic efficacy of AI in breast US screening.

Methods

This study utilized different frequency US-probes (L14: frequency range: 3.0-14.0 MHz, central frequency 9 MHz, L9: frequency range: 2.5-9.0 MHz, central frequency 6.5 MHz and L13: frequency range: 3.6-13.5 MHz, central frequency 8 MHz, L7: frequency range: 3-7 MHz, central frequency 4.0 MHz, linear arrays) to collect breast-video and applied an entropy-based deep learning approach for evaluation. We analyzed the average two-dimensional image entropy (2-DIE) of these videos and the performance of AI models in processing videos from these different frequencies to assess how probe frequency affects AI diagnostic performance.

Results

The study found that in testing set 1, L9 was higher than L14 in average 2-DIE; in testing set 2, L13 was higher in average 2-DIE than L7. The diagnostic efficacy of US-data, utilized in AI model analysis, varied across different frequencies (AUC: L9 > L14: 0.849 vs. 0.784; L13 > L7: 0.920 vs. 0.887).

Conclusion

This study indicate that US-data acquired using probes with varying frequencies exhibit diverse average 2-DIE values, and datasets characterized by higher average 2-DIE demonstrate enhanced diagnostic outcomes in AI-driven BCa diagnosis. Unlike other studies, our research emphasizes the importance of US-probe frequency selection on AI model diagnostic performance, rather than focusing solely on the AI algorithms themselves. These insights offer a new perspective for early BCa screening and diagnosis and are of significant for future choices of US equipment and optimization of AI algorithms.

Key points

1. The study explored the impact of ultrasound images with different frequencies on the diagnostic efficacy of artificial intelligence.

2. Ultrasound images obtained with different frequency probes exhibited variable levels of average two-dimensional image entropy, influencing the diagnostic performance of artificial intelligence models in nuanced ways.

3. Datasets with higher average two-dimensional image entropy were associated with superior artificial intelligence breast diagnostic efficacy.

Summary

The research on artificial intelligence-assisted breast diagnosis often relies on static images or dynamic videos obtained from ultrasound probes with different frequencies. However, the effect of frequency-induced image variations on the diagnostic performance of artificial intelligence models remains unclear. In this study, we aimed to explore the impact of using ultrasound images with variable frequencies on AI’s diagnostic efficacy in breast ultrasound screening. Our approach involved employing a video and entropy-based feature breast network to compare the diagnostic efficiency and average two-dimensional image entropy of the L14 (frequency range: 3.0-14.0 MHz, central frequency 9 MHz), L9 (frequency range: 2.5-9.0 MHz, central frequency 6.5 MHz) linear array probe and L13 (frequency range: 3.6-13.5 MHz, central frequency 8 MHz), and L7 (frequency range: 3-7 MHz, central frequency 4.0 MHz) linear array probes. The results revealed that the diagnostic efficiency of AI models differed based on the frequency of the ultrasound probe. It is noteworthy that ultrasound images acquired with different frequency probes exhibit different average two-dimensional image entropy, while higher average two-dimensional image entropy positively affect the diagnostic performance of the AI model. We concluded that a dataset with higher average two-dimensional image entropy is associated with superior diagnostic efficacy for AI-based breast diagnosis. These findings contribute to a better understanding of how ultrasound image variations impact AI-assisted breast diagnosis, potentially leading to improved breast cancer screening outcomes.

Peer Review reports

Introduction

Breast cancer (BCa) is the most prevalent cancer and the leading cause of cancer mortality in females worldwide [1, 2]. Early identification and intervention of BCa lead to significant improvements in 5-year relative survival rates [3, 4]. Ultrasound(US) is the imaging method of choice for the evaluation of breast disease because it is technically simple, cost-effective, and safe [5]. Also, the US is seen as a primary measure of BCa detection and mortality reduction [3]. However, US-image-based diagnosis of the breast greatly relies on the experiences of the sonographers [6], so it is significant to further explore the information carried by US images to enhance the detection rate and diagnostic accuracy of BCa in its early stages.

In recent years, artificial intelligence (AI) has brought opportunities for the advancement of medical imaging [7,8,9,10]. The algorithm is enabled to extract a large amount of information from medical images that cannot be observed by the naked eye for diagnosis and improve the computer detection rate of nodules [11,12,13]. US-based AI studies rely on the sonographers-selected images during the scanning process or partially on responsibility frames selected from video [14,15,16,17,18,19,20,21]. Therefore, the US image selection is particularly crucial in BCa AI diagnosis.

In information theory [22], entropy is the average amount of information contained in each received “message”. Image entropy (IE) is a statistical form of image features, reflecting the average amount of information in the image, which can reflect the distribution complexity of each pixel point of the image [23]. Most previous studies were image-based that required high-frequency probe acquisition as a dataset [24,25,26]. The principle of choosing a US probe is to ensure sufficient detection depth while maximizing the frequency to ensure the resolution of the US image [25]. Although the high-frequency probe images may aid the sonographer in making a diagnosis. Whether they are favorable for the training and diagnosis of AI models is not known yet. Earlier works [21] demonstrated that the richer the average information content of an image, the better its tumor classification. Thus, based on the principle that US low frequency corresponds to high penetration [27], US data obtained at different frequencies may carry different levels of information, thereby impacting the diagnosis of the AI model.

Therefore, this study introduces the feature entropy of breast US to calculate the magnitude of the average two-dimensional image entropy(2-DIE) at different frequencies. Further, to investigate whether the US images obtained at lower frequencies have higher average 2-DIE and are more beneficial to improve AI diagnosis.

Materials and methods

Participants

This retrospective research collected US videos examined at Shenzhen People’s Hospital from June 2021 to December 2021. As a retrospective study, informed consent was waived by the Medical Ethics Committee of Shenzhen People’s Hospital. All patient information was handled with strict confidentiality in compliance with ethical guidelines. The benignity and malignancy of the nodules obtained by the US were confirmed by pathology.

Inclusion criteria

(a) Simultaneous acquisition of US video images of tumors in the same patient with two different frequency probes. (b) US-detected nodules must be classified as 0, 2, 3, 4a, 4b, 4c, or 5 following the BI-RADS. (c) No biopsy or surgical treatment of the breast nodules is to be evaluated before the US scan. (d) Patients were biopsied or surgically treated within 1 week of US data acquisition, while pathological results were obtainable.

Exclusion criteria

(a) BI-RADS 6 in the US, DR (Digital mammography), or MRI (magnetic resonance imaging). (b) BI-RADS 1 in the US. (c) History of breast surgery. (d) Single-frequency US video data. (e) Missing pathological results. (f) Poor image quality.

In this study, a total of 668 breast tumors (260 malignancy and 408 benign) of US videos from 167 female patients were included and divided into 2 testing sets: (1) In testing set 1, breast US videos were obtained from Resona I9 (Mindray, China) with L14 (frequency range: 3.0-14.0 MHz, central frequency 9 MHz), L9 (frequency range: 2.5-9.0 MHz, central frequency 6.5 MHz) linear array probe. (2) Testing set 2, the data were obtained from DC-65 (Mindray, China) with L13 (frequency range: 3.6-13.5 MHz, central frequency 8 MHz), and L7 (frequency range: 3.0-7.0 MHz, central frequency 4.0 MHz) linear array probe, which was aimed to further evaluate our theory and discoveries.

Ultrasound examination and video acquisition

In this study, all US videos were acquired by the same radiologist with more than 10 years of experience. The researchers utilized 3 markers to localize the location of the mass during the collection process. Specifically, the largest section of the target tumor was first located using one of the frequency probes and marked on the body surface. Then a complete sweep was made horizontally along the largest section of the tumor to find two more markers ≥ 2 cm from the tumor margin, respectively. Finally, the whole tumor is swept along the marked direction, and the operation is repeated, keeping the direction and position of the probe consistent each time, until the US video acquisition of the four different frequencies is completed.

Processing of US-video and use of AI model

First, we conducted this study based on the constructed feature entropy breast network (FEBrNet), which inherits the pre-trained backbone of the fully connected layer and the weight-optimal model [21]. We use the AI model to select responsibility frames to reduce subjective dependence. Our method of selecting pivotal frames draws inspiration from established applications of entropy in information theory, such as in decision trees. Specifically, the Iterative Dichotomiser 3 (ID3) decision tree algorithm utilizes entropy to ascertain the most suitable parent node and its division. In our methodology, we aim to minimize the discrepancy between the FScore of the video and that of the chosen frame collection, where a smaller disparity indicates that the information content of the chosen frames closely mirrors that of the entire video. By incrementally adding frames to this collection, starting from one and increasing to n, and at each increment selecting the frame that least differs, we gradually form an optimally representative set of frames, each contributing unique features. Subsequently, for the final collection of these optimal frames, our study computes the two-dimensional image entropy for each frame using the FEBrNet model. We then determine the video’s image entropy by calculating the average two-dimensional image entropy (2-DIE) of all the chosen frames. Finally, pathological results were used as the gold standard to compare the ability of using image entropy of different frequencies in the differential diagnosis of benign and malignant breast tumors. The processing and validation of the data are based on the pre-trained entropy-based model (FEBrNet). For specific information about the model refer to this literature [21] and supplementary materials. We investigate and verify the effect of entropy on the diagnostic performance of AI models from the perspective of IE. US images obtained from US probes of different frequencies are various. The researchers compared the variations by collecting US data from the same patient at different frequencies simultaneously. This is used to research the difference in diagnostic efficacy of US images obtained at different frequencies for AI models. The flow chart is shown in Fig. 1.

Fig. 1
figure 1

The flow charts of this Study

Statistical analysis

Statistical analysis was performed using R 3.6.3 (Copyright (C) 2020 The R Foundation for Statistical Computing). The significance level was set at P < 0.05. A normality test was performed for each variable. T-test is used for the normally distributed numerical variables, the rank sum test is used for the non-normally distributed numerical variables, and the Chi-square test is used for the disordered classification variables. The paired sample t-test was used to compare the differences within the group. The specificity, sensitivity, accuracy, receiver operating characteristic curve (ROC), and area under the curve (AUC) were used to evaluate models.

Results

Participant characteristics

According to the inclusion and exclusion criteria, A total of 668 tumor videos from 167 patients were included in this study, including 260 videos of cancerous masses and 408 videos of benign tumors. There are 280 videos in the testing set 1 and 388 videos in testing set 2. Table 1 show the baseline distribution characteristics of the collected patients, respectively.

Table 1 The distribution of baseline characteristics based on testing sets

Distribution of 2-DIE in various frequencies

The values of the average 2-DIE obtained at different frequencies are variable. For testing set 1, the 2-DIE of the L9 linear probe was higher than that of L14 (Mean ± SD,11.49 ± 0.769 vs. 10.94 ± 0.835); For the testing set 2, the 2-DIE of the L13 linear probe was higher than that of L7 (Mean ± SD,11.82 ± 0.356 vs. 12.27 ± 0.476). This result is summarized in Fig. 2.

Fig. 2
figure 2

The results of the average 2-DIE of the two testing sets. Note: 2-DIE: two-dimensional image entropy (unless otherwise stated, the above measurements are average values); L9: L9 linear array probe, frequency range: 2.5-9.0 MHz, central frequency 6.5 MHz; L14: L14 linear array probe, frequency range: 3.0-14.0 MHz, central frequency 9 MHz; L7: L7 linear array probe, frequency range: 3-7 MHz, central frequency 4.0 MHz; L13: L13 linear array probe, frequency range: 3.6-13.5 MHz, central frequency 8 MHz; P: L9 vs. L14, L7 vs. L13

Diagnosis performance of AI models

The diagnostic efficacy of US data for AI models varies at distinct frequencies. For the testing set 1, L9 attained the best AUC (0.849), with a sensitivity of 76.9%, specificity of 93.2%, and accuracy of 87.1%. For the testing set 2, L13 reached the best AUC (0.920), sensitivity 89.7%, specificity 93.8%, as well as accuracy 91.0%. The detailed results are shown in Table 2; Fig. 3.

Fig. 3
figure 3

Comparison of diagnostic performance of the testing sets. Note: AUC: area under the curve; 95% CI: 95% confidence interval; L9: L9 linear array probe; L14: L14 linear array probe; L7: L7 linear array probe; L13: L13 linear array probe. (A): Testing 1; (B): Testing 2

Table 2 Comparison of the efficacy of AI models

Discussion

In this study, we used a video and entropy-based deep learning model [21] to compare the diagnosis performance of breast US. The assessment effect of variable frequencies on the AI model diagnosis validity was based on two retrospective data sets (Mindray L7/L13 and L9/L14). In testing set 1, compared to L14 (frequency range: 3.0-14.0 MHz, central frequency 9 MHz), the L9-had better diagnosis performance and 2-DIE. However, in testing set 2, compared to L7 (frequency range: 3-7 MHz, central frequency 4.0 MHz), the L13-had better diagnosis performance and 2-DIE. This observation suggests that US-data derived from probes operating at varying frequencies can significantly impact the diagnostic effectiveness of AI models. Another finding is that higher 2-DIE is accompanied by increased diagnostic efficacy.

In recent years, many AI-based studies have investigated the benign and malignant categorization of US breast nodules [18, 28,29,30,31,32,33,34,35,36,37]. The accuracy of their models fluctuates from 80 to 95%. While the literature recommends a frequency range of 5–17 [38] for breast US screening, it does not specify which one to use. Also, there is no literature examining the distinction in the diagnostic utility of AI for images acquired by various probes. The probe frequencies used in the studies in the literature mentioned above ranged from 1 to 42 MHz. Therefore, one reason for the variation in accuracy between these surveys may be the variance in the frequency of the probes utilized. So, we did this experience and discovered that the L9 (frequency range: 2.5-9.0 MHz, central frequency 6.5 MHz) had better diagnosis performance and higher 2-DIE. This may be contradicted by our clinical experience but offers another probability. While high-frequency US probes typically require more sophisticated technology and may be costlier, their application might not always correspond to improved diagnostic performance in AI models. Furthermore, primary hospitals may be unable to afford the purchase and maintenance of high frequency probe. The results of this experiment may now solve this challenge – using AI to aid diagnosis and compensate for the low accuracy of clinicians when using low frequency probe. On the other hand, excessively low-frequency probes do not enhance the diagnostic performance of AI models. In the testing set 2, we found that the L13 (frequency range: 3.6-13.5 MHz, central frequency 8 MHz) had better diagnosis performance. This is inconsistent with the results of our testing set 1. Possible reasons for this result include (a. The penetration is excessive, resulting in images that contain more confounding information unrelated to the lesion. b. The high frequency probe provides excellent spatial and soft-tissue resolution, greatly improving the differentiation of lesion saliency. However, the results for the 2-DIE of the L13 are higher than those of the low-frequency ones, which remains consistent with our previous findings.

Also, previous studies [21] by our team demonstrated that the richer the 2-DIE contained in US images, the more favorable the prediction of breast tumor benignity-malignancy. That is, a high 2-DIE in US images corresponds to rich image information. Meanwhile, US features varied depending on the pathological heterogeneity of the breast tumor [39]. The richer the information contained in the US image, the more comprehensive the information it may contain about the tumor characteristics. Moon et al. [34] also indicated that images with more information would help improve the diagnostic efficacy of the model. The results of this experiment are consistent with previous studies – both in the two sets, the higher the 2-DIE, the better the diagnostic performance. Because of the higher penetration of the low frequency probe, visualization of deep posterior tissues is made easy. More information related to the nodules may be captured. This information may not be recognized by the naked eye but facilitates machine learning. Accompanying the development of AI and the concept of medical-industrial integration, the application of AI-assisted diagnosis may become more extensive. However, previous research has focused more on the innovation and refinement of algorithms and hardware, ignoring the differences in images of different frequencies. Therefore, it is necessary to investigate the diagnostic efficacy of different frequency datasets on AI models. Images acquired at more appropriate frequencies will help improve diagnostic performance and provide a reference for future US image acquisition for AI models.

There are some limitations in this study. First, the study was a retrospective single-center study with smaller sample size and uneven image quality. Second, Lack of comparison of diagnostic efficacy of different frequencies for AI models. Therefore, we will further investigate the effect of other frequencies on AI diagnostic efficacy in the next research plan. Finally, Variations in sensitivity and inter-machine variability of various US devices were not considered.

Conclusion

This study indicate that US-data acquired using probes with varying frequencies exhibit diverse average 2-DIE values, and datasets characterized by higher average 2-DIE demonstrate enhanced diagnostic outcomes in AI-driven BCa diagnosis. Unlike other studies, our research emphasizes the importance of US-probe frequency selection on AI model diagnostic performance, rather than focusing solely on the AI algorithms themselves. These insights offer a new perspective for early BCa screening and diagnosis and are of significant guidance for future choices of US equipment and optimization of AI algorithms.

Data Availability

The datasets used and analyzed during the current study are available from the corresponding author on reasonable request.

Data Availability

The datasets used and analyzed during the current study are available from the corresponding author on reasonable request.

Abbreviations

BCa:

breast cancer

2-DIE:

two-dimensional image entropy

US:

ultrasound

AI:

artificial intelligence

IE:

Image entropy

DR:

Digital mammography

MRI:

magnetic resonance imaging

L9:

L9 linear array probe (frequency range:2.5-9.0 MHz, central frequency 6.5 MHz)

L14:

L14 linear array probe (frequency range:3.0-14.0 MHz, central frequency 9 MHz)

L7:

L7 linear array probe (frequency range:3-7 MHz, central frequency 4.0 MHz)

L13:

L13 linear array probe (frequency range:3.6-13.5 MHz, central frequency 8 MHz)

BI-RADS:

Breast Imaging-Reporting and Data System

ROC:

receiver operating characteristic curve

CI:

confidence interval

AUC:

area under the curve

IQR:

interquartile range

References

  1. Xia C, Dong X, Li H, Cao M, Sun D, He S, Yang F, Yan X, Zhang S, Li N, et al. Cancer statistics in China and United States, 2022: profiles, trends, and determinants. Chin Med J (Engl). 2022;135(5):584–90.

    Article  PubMed  Google Scholar 

  2. Cao W, Chen H-D, Yu Y-W, Li N, Chen W-Q. Changing profiles of cancer burden worldwide and in China: a secondary analysis of the global cancer statistics 2020. Chin Med J. 2021;134(7):783–91.

    Article  PubMed  PubMed Central  Google Scholar 

  3. Cedolini C, Bertozzi S, Londero AP, Bernardi S, Seriau L, Concina S, Cattin F, Risaliti A. Type of Breast cancer diagnosis, screening, and survival. Clin Breast Cancer. 2014;14(4):235–40.

    Article  PubMed  Google Scholar 

  4. Islami F, Ward EM, Sung H, Cronin KA, Tangka FKL, Sherman RL, Zhao J, Anderson RN, Henley SJ, Yabroff KR, et al. Annual Report to the Nation on the Status of Cancer, Part 1: National Cancer statistics. J Natl Cancer Inst. 2021;113(12):1648–69.

    Article  PubMed  PubMed Central  Google Scholar 

  5. Berg WA. Current status of Supplemental Screening in dense breasts. J Clin Oncology: Official J Am Soc Clin Oncol. 2016;34(16):1840–3.

    Article  Google Scholar 

  6. Sakamoto N, Tozaki M, Higa K, Abe S, Ozaki S, Fukuma E. False-negative ultrasound-guided vacuum-assisted biopsy of the breast: difference with US-detected and MRI-detected lesions. Breast cancer (Tokyo Japan). 2010;17(2):110–7.

    Article  PubMed  Google Scholar 

  7. Sechopoulos I, Teuwen J, Mann R: Artificial intelligence for breast cancer detection in mammography and digital breast tomosynthesis: State of the art. SEMINARS IN CANCER BIOLOGY 2021, 72:214-225.

  8. Yadav N, Dass R, Virmani J. Despeckling filters applied to thyroid ultrasound images: a comparative analysis. MULTIMEDIA TOOLS AND APPLICATIONS. 2022;81(6):8905–37.

    Article  Google Scholar 

  9. Dass R, Yadav N. Image quality assessment parameters for despeckling filters. Procedia Comput Sci. 2020;167:2382–92.

    Article  Google Scholar 

  10. Yadav N, Dass R, Virmani J. Deep learning-based CAD system design for thyroid Tumor characterization using ultrasound images. MULTIMEDIA TOOLS AND APPLICATIONS 2023:1–43.

  11. Hinton G. Deep Learning-A Technology with the potential to Transform Health Care. JAMA. 2018;320(11):1101–2.

    Article  PubMed  Google Scholar 

  12. Huang YQ, Liang CH, He L, Tian J, Liang CS, Chen X, Ma ZL, Liu ZY. Development and validation of a Radiomics Nomogram for Preoperative Prediction of Lymph Node Metastasis in Colorectal Cancer. J Clin Oncol. 2016;34(18):2157–64.

    Article  PubMed  Google Scholar 

  13. Ouyang D, He B, Ghorbani A, Yuan N, Ebinger J, Langlotz CP, Heidenreich PA, Harrington RA, Liang DH, Ashley EA, et al. Video-based AI for beat-to-beat assessment of cardiac function. Nature. 2020;580(7802):252–6.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  14. Aljuaid H, Alturki N, Alsubaie N, Cavallaro L, Liotta A. Computer-aided diagnosis for Breast cancer classification using deep neural networks and transfer learning. Comput Methods Programs Biomed. 2022;223:106951.

    Article  PubMed  Google Scholar 

  15. Bae MS, Kim HG. Breast Cancer Risk Prediction using deep learning. Radiology. 2021;301(3):559–60.

    Article  PubMed  Google Scholar 

  16. Hayashida T, Odani E, Kikuchi M, Nagayama A, Seki T, Takahashi M, Futatsugi N, Matsumoto A, Murata T, Watanuki R, et al. Establishment of a deep-learning system to diagnose BI-RADS4a or higher using breast ultrasound for clinical application. Cancer Sci. 2022;113(10):3528–34.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  17. Jiang M, Li CL, Luo XM, Chuan ZR, Lv WZ, Li X, Cui XW, Dietrich CF. Ultrasound-based deep learning radiomics in the assessment of pathological complete response to neoadjuvant chemotherapy in locally advanced Breast cancer. Eur J Cancer. 2021;147:95–105.

    Article  CAS  PubMed  Google Scholar 

  18. Qi X, Zhang L, Chen Y, Pi Y, Chen Y, Lv Q, Yi Z. Automated diagnosis of breast ultrasonography images using deep neural networks. Med Image Anal. 2019;52:185–98.

    Article  PubMed  Google Scholar 

  19. Qian X, Pei J, Zheng H, Xie X, Yan L, Zhang H, Han C, Gao X, Zhang H, Zheng W, et al. Prospective assessment of Breast cancer risk from multimodal multiview ultrasound images via clinically applicable deep learning. Nat Biomed Eng. 2021;5(6):522–32.

    Article  PubMed  Google Scholar 

  20. Zheng X, Yao Z, Huang Y, Yu Y, Wang Y, Liu Y, Mao R, Li F, Xiao Y, Wang Y, et al. Deep learning radiomics can predict axillary lymph node status in early-stage Breast cancer. Nat Commun. 2020;11(1):1236.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  21. Chen J, Jiang Y, Yang K, Ye X, Cui C, Shi S, Wu H, Tian H, Song D, Yao J, et al. Feasibility of using AI to auto-catch responsible frames in ultrasound screening for Breast cancer diagnosis. iScience. 2023;26(1):105692.

    Article  PubMed  Google Scholar 

  22. Shannon CE. A mathematical theory of communication. Bell Syst Tech J. 1948;27(3):379–423.

    Article  Google Scholar 

  23. Chen X, Zhang Q, Lin M, Yang G, He C. No-reference color image quality assessment: from entropy to perceptual quality. EURASIP J Image Video Process. 2019;2019(1):1–14.

    Article  Google Scholar 

  24. Berg WA, Mendelson EB. Technologist-performed handheld screening breast US imaging: how is it performed and what are the outcomes to date? Radiology. 2014;272(1):12–27.

    Article  PubMed  Google Scholar 

  25. Berg WA, Bandos AI, Mendelson EB, Lehrer D, Jong RA, Pisano ED. Ultrasound as the primary screening test for Breast Cancer: analysis from ACRIN 6666. J Natl Cancer Inst 2016, 108(4).

  26. Bevers TB, Helvie M, Bonaccio E, Calhoun KE, Daly MB, Farrar WB, Garber JE, Gray R, Greenberg CC, Greenup R, et al. Breast Cancer screening and diagnosis, Version 3.2018, NCCN Clinical Practice guidelines in Oncology. J Natl Compr Canc Netw. 2018;16(11):1362–89.

    Article  PubMed  Google Scholar 

  27. Popp RL, Macovski A. Ultrasonic diagnostic instruments. Sci (New York NY). 1980;210(4467):268–73.

    Article  CAS  Google Scholar 

  28. Cheng J-Z, Ni D, Chou Y-H, Qin J, Tiu C-M, Chang Y-C, Huang C-S, Shen D, Chen C-M. Computer-aided diagnosis with Deep Learning Architecture: applications to breast lesions in US images and pulmonary nodules in CT scans. Sci Rep. 2016;6:24454.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  29. Seung Yeon S, Soochahn L, Il Dong Y, Sun Mi K, Kyoung Mu L. Joint weakly and semi-supervised deep learning for localization and classification of masses in breast ultrasound images. IEEE Trans Med Imaging. 2019;38(3):762–74.

    Article  Google Scholar 

  30. Shan J, Alam SK, Garra B, Zhang Y, Ahmed T. Computer-aided diagnosis for breast Ultrasound using computerized BI-RADS features and machine learning methods. Ultrasound Med Biol. 2016;42(4):980–8.

    Article  PubMed  Google Scholar 

  31. Han S, Kang HK, Jeong JY, Park MH, Kim W, Bang WC, Seong YK. A deep learning framework for supporting the classification of breast lesions in ultrasound images. Phys Med Biol. 2017;62(19):7714–28.

    Article  PubMed  Google Scholar 

  32. Tanaka H, Chiu S-W, Watanabe T, Kaoku S, Yamaguchi T. Computer-aided diagnosis system for breast ultrasound images using deep learning. Phys Med Biol. 2019;64(23):235013.

    Article  PubMed  Google Scholar 

  33. Zhang E, Seiler S, Chen M, Lu W, Gu X. BIRADS features-oriented semi-supervised deep learning for breast ultrasound computer-aided diagnosis. Phys Med Biol. 2020;65(12):125005.

    Article  PubMed  Google Scholar 

  34. Moon WK, Lee Y-W, Ke H-H, Lee SH, Huang C-S, Chang R-F. Computer-aided diagnosis of breast ultrasound images using ensemble learning from convolutional neural networks. Comput Methods Programs Biomed. 2020;190:105361.

    Article  PubMed  Google Scholar 

  35. Zhang Q, Xiao Y, Dai W, Suo J, Wang C, Shi J, Zheng H. Deep learning based classification of breast tumors with shear-wave elastography. Ultrasonics. 2016;72:150–7.

    Article  PubMed  Google Scholar 

  36. Qian X, Zhang B, Liu S, Wang Y, Chen X, Liu J, Yang Y, Chen X, Wei Y, Xiao Q, et al. A combined ultrasonic B-mode and color doppler system for the classification of breast masses using neural network. Eur Radiol. 2020;30(5):3023–33.

    Article  PubMed  Google Scholar 

  37. Zhou Y, Xu J, Liu Q, Li C, Liu Z, Wang M, Zheng H, Wang S. A Radiomics Approach with CNN for Shear-Wave Elastography breast Tumor classification. IEEE Trans Bio Med Eng. 2018;65(9):1935–42.

    Article  Google Scholar 

  38. Hooley RJ, Scoutt LM, Philpotts LE. Breast ultrasonography: state of the art. Radiology. 2013;268(3):642–59.

    Article  PubMed  Google Scholar 

  39. Wang LC, Sullivan M, Du H, Feldman MI, Mendelson EB. US appearance of ductal carcinoma in situ. Radiographics: A Review Publication of the Radiological Society of North America Inc. 2013;33(1):213–28.

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

Not applicable.

Funding

This work is supported by the following grant: Clinical Scientist Training Program of Shenzhen People’s Hospital, SYWGSCGZH202202. The funding unit assumes the function of the research sponsor in this research.

Author information

Authors and Affiliations

Authors

Contributions

Study concept and design: ZH, KY. Data extraction: HW, JC. Data quality: ST, HT. Data analysis and interpretation: CC, SS, YJ. Manuscript preparation: ZH, KY. Manuscript review: JX, FD. Manuscript editing: ZH. All authors approved the final manuscript for submission.

Corresponding authors

Correspondence to Jinfeng Xu or Fajin Dong.

Ethics declarations

Ethics approval and consent to participate

This retrospective study was approved by the Institutional Review Board of the Shenzhen People’s Hospital, specifically the Medical Ethics Committee of Shenzhen People ‘s Hospital. Due to the retrospective nature of this study, the need for informed consent was waived by the Institutional Review Board of Shenzhen People’s Hospital, specifically by the Medical Ethics Committee of Shenzhen People’s Hospital. All methods were carried out in accordance with relevant guidelines and regulations.

Informed consent

This study was conducted in accordance with the Declaration of Helsinki and was approved by the institutional review board of the Shenzhen People’s Hospital. Informed consent was waived by the same ethics committee that approved the study.

Consent for publication

Not applicable.

Competing interests

The authors of this manuscript declare no relationships with any companies, whose products or services may be related to the subject matter of the article.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Electronic supplementary material

Below is the link to the electronic supplementary material.

Supplementary Material 1

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Huang, Z., Yang, K., Tian, H. et al. A validation of an entropy-based artificial intelligence for ultrasound data in breast tumors. BMC Med Inform Decis Mak 24, 1 (2024). https://doi.org/10.1186/s12911-023-02404-z

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12911-023-02404-z