Abstract
Epiretinal membrane (ERM) is a common ophthalmological disorder of high prevalence. Its symptoms include metamorphopsia, blurred vision, and decreased visual acuity. Early diagnosis and timely treatment of ERM is crucial to preventing vision loss. Although optical coherence tomography (OCT) is regarded as a de facto standard for ERM diagnosis due to its intuitiveness and high sensitivity, ophthalmoscopic examination or fundus photographs still have the advantages of price and accessibility. Artificial intelligence (AI) has been widely applied in the health care industry for its robust and significant performance in detecting various diseases. In this study, we validated the use of a previously trained deep neural network based-AI model in ERM detection based on color fundus photographs. An independent test set of fundus photographs was labeled by a group of ophthalmologists according to their corresponding OCT images as the gold standard. Then the test set was interpreted by other ophthalmologists and AI model without knowing their OCT results. Compared with manual diagnosis based on fundus photographs alone, the AI model had comparable accuracy (AI model 77.08% vs. integrated manual diagnosis 75.69%, χ2 = 0.038, P = 0.845, McNemar’s test), higher sensitivity (75.90% vs. 63.86%, χ2 = 4.500, P = 0.034, McNemar’s test), under the cost of lower but reasonable specificity (78.69% vs. 91.80%, χ2 = 6.125, P = 0.013, McNemar’s test). Thus our AI model can serve as a possible alternative for manual diagnosis in ERM screening.
Similar content being viewed by others
Introduction
Epiretinal membranes (ERMs) are the abnormal formation of non-vascular cellular or fibrous membrane at the inner layer of retina1. Typically, when the membrane locates around macular region, it is called epimacular membranes or premacular fibroplasis. Most ERMs are asymptomatic and remain stable or even regress automatically in absence of medical intervention2. However, there are cases that develop rapidly within months, causing blind spot or scotoma3. Further contraction of the membrane may tear the retina and lead to severe outcomes such as retinal hemorrhage, retinal edema, or even macular holes4,5. Reported prevalence of ERM ranges from 3.4% to 39% in different populations6,7,8,9,10,11,12,13,14,15. These published incidences can be under-estimated, due to difficulty in diagnosis of secondary ERM with the presence of primary ocular diseases.
Clinical management of ERM was observed at early asymptomatic stage or surgical intervention at late stage1. Surgical removal of ERM can restore retinal structure and improve visual acuity, but there is potentially residual impairment of visual function16. Better visual acuity improvements have been found with: (i) shorter latency between symptom onset and surgery17; (ii) better preoperative visual acuity17,18,19; (iii) less central foveal thickness measured by OCT20,21; (iv) maintained integrity of retinal structure18,19,20,22,23; (v) thinner ganglion cell inner plexiform layer24; (vi) absence of co-existing maculopathy25. Therefore, early detection of ERM and timely surgical intervention are crucial to achieving better visual outcomes.
With the high prevalence and possible severe consequence of the disease, fast and inexpensive approaches for ERM screening in large population are in need. Traditionally, ERMs were diagnosed and classified based on subjective symptoms and ophthalmoscopic examination26. Fundus photographs provide the possibility of review and compare ERM at different time points. Experienced ophthalmologists are required for correct interpretation of retinal images. Non-invasive modern examination, such as optical coherence tomography (OCT), provides high-resolution structural information of live retina and has been widely used as an essential diagnostic tool in multiple maculopathies. Cross-sectional macula-centered OCT images provide clear and intuitive structural relationship of epimacular membrane and the neuroretina, and have been proved to be more sensitive than other clinical examinations in ERM diagnosis27. Furthermore, linear scanning across the macular fovea provides multiple quantifiable structural parameters including the central foveal thickness, the volume of macular edema, and the ERM thickness18,21,28,29. Based on these quantifiable data, several OCT-based ERM classification systems have been established, contributing to a better understanding of ERM. However, the promotion of OCT in primary hospitals is difficult due to price issues.
Deep learning-based artificial intelligence (AI) has been widely applied in assisting detection and staging of common ocular disorders with distinctive features, such as diabetic retinopathy30,31,32,33, glaucoma34, cataract35, age-related macular degeneration36,37, and retinopathy of prematurity38,39,40. Unlike the traditional coding process where methodology has already been established by programmers before coding, decision-making strategy is figured out by deep-learning algorithms throughout the coding process, which means deep learning algorithms may detect novel features that highly related to disease diagnosis or staging41. AI models have shown reliable performance in detection of ERM based on OCT images42,43. However, currently there is no validated AI model for ERM detection based on fundus photographs, which is of easier access and lower cost. In this study, we compared manual or AI detection of ERM based on fundus photographs with the standard diagnosis of ERM based on OCT and other clinical information, validating the use of AI model in assisting detection of ERM using fundus photographs.
Methods
Case recruitment
Total 96 successive outpatients diagnosed with ERM from 2017 October to 2019 December at Beijing Tsinghua Changgung Hospital were recruited in this retrospective study (Fig. 1). Their basic information, brief medical history, fundus photographs of both eye (192 eyes) (macular-centered, by Canon CR-2 Digital Retinal Camera (Canon, Long Island, NY, US)), and ipsilateral OCT images (vertical and horizontal cross section at macular, by Heidelberg Engineering OCT Spectralis (Heidelberg Engineering, Heidelberg, Germany)) were anonymized and acquired from the hospital information system. This study was approved by the Ethical Committee of Beijing Tsinghua Changgung Hospital. The Ethical Committee of Beijing Tsinghua Changgung Hospital waived the need for informed consent from the participants. All methods were carried out in accordance with the Declaration of Helsinki and the International Ethical Guidelines for Biomedical Research Involving Human Subjects.
ERM diagnosis by OCT images
Corresponding OCT images of all recruited retinal images (including both ERM and non-ERM groups) were examined individually by two experienced ophthalmologists. Sex, age, and brief medical history of the patients were provided to the examiners in order to improve diagnostic accuracy. Both examiners gave binary diagnosis of ERM or non-ERM. When the two examiners reported contradicting diagnoses, a third specialist re-examined the corresponding OCT images and made the final OCT diagnosis. Figure 2D–E showed typical OCT images of ERM.
Manual and AI interpretation of fundus photographs
Fundus photographs of ERM and non-ERM eyes were mixed together and randomly labeled with successive numbers. Resulted retinal image set was interpreted independently by three different ophthalmologists without knowing the sex, age, medical history, and OCT result of the cases. The three ophthalmologists consist of one resident doctor, one attending, and one retina specialist, representing different experience levels in ERM diagnosis and management. Mixed fundus photographs set were analyzed by AI model simultaneously. Figure 2A–C were examples of visualized ERM detection of AI model in form of Grad-CAM44. The AI model outputs binary result (ERM or non-ERM).
Brief introduction of the AI model
The AI model used in our ERM recognition is a multiply retinal diseases classification model based on a deep neural network (DNN). Neural network inspired by animal and human visual system mimics how the human brain works in processing data and pattern abstraction. Due to its depth and massive layers, DNN has huge representation power to learn visual features of retinal diseases and discriminate them effectively. The disease classification model adapted a novel backbone neural network that combines Inception-Resnet-v245 and Xception46 architecture. An n-way binary classification layer is added to the top of the backbone network, which yields the probability of each disease. Tensorflow machine learning framework is used in the training model47. The AI model used in this study was previously trained with a different dataset consisting of 207,228 fundus photographs collected from 16 clinical settings with different disease distributions across China, with an average age of 40.83 ± 15.43 (registered with ClinicalTrials.gov, NCT04213430)48. It was annotated by an expert ophthalmologist team as 14 retinal abnormalities including ERM. Data augmentation methods during the training included Inception network family-related data sampling49 and random rotation that imitates intrinsic similarity of fundus image. Before image analysis, any given RGB fundus photograph is cropped and resized into 585 × 585 square with only the region around the posterior pole reserved. After analysis, the model returns a 1 × n vector whose every element represents the probability of a certain disease (ranging from 0 to 1). For each disease, a positive result is assigned to the image if the corresponding element exceeds the threshold of the disease. In this study, only the element linked to ERM was taken into consideration.
Statistical analysis
Performance of both manual and AI interpretations were calculated with respect to standard diagnosis based on OCT images. The accuracy of manual and AI detection for ERM were calculated using the equation accuracy = (TP + TN)/ALL (TP: true positive; TN: true negative; ALL: total sample size). Sensitivity and specificity of manual and AI detection were also calculated, using the equation sensitivity = TP/(TP + FN), specificity = TN/(TN + FP) (TP: true positive, TN: true negative, FP: false positive, FN: false negative). Kappa (κ) statistics were used to quantify the degree of agreement comparing different individual manual analyses, summarized manual results, AI interpretation with the standard OCT result. Statistical significance was considered at P < 0.05.
Results
Basic information of recruited cases
A total of 96 patients has been recruited in this study. Among them, 48 (50%) are female and 48 (50%) are male, with an average age of 66.75 ± 10.03 and 72.25 ± 12.09 respectively (Table 1). Our reported even distribution of ERM patients in different sex groups is in consensus with the previously reported consistent prevalence in both male and female groups. Among all these outpatients, 69 (71.87%) of them had symptoms of blurred vision, decreased visual acuity, or metamorphopsia. 27 (28.13%) of them had no obvious symptoms and the ERM was found accidentally through physical examination or during following up of other ocular diseases. The most frequent comorbidity in our recruited group was age-related cataract. Other comorbidities include macular edema, diabetic retinopathy, retinal detachment, and macular hole.
OCT diagnosis results
Totally 96 outpatients (192 eyes) were recruited in the beginning, among which 91 patients (182 eyes) have done OCT and were included in further analysis (Fig. 1A). All acquired OCT images were re-examined by retinal specialists in order to obtain standard ERM diagnosis. For all the OCT-examined patients, 24% of them have both eyes diagnosed with ERM. Left-eye only and right-eye only ERM took up 30% and 37% of the population respectively (Fig. 3).
After elimination of unclear fundus photographs due to cataracts or vitreous opacity, 153 fundus images from 91 patients were reserved. Among them, 144 eyes have corresponding OCT images and their diagnosis of ERM was re-confirmed by experienced ophthalmologists based on OCT images as well as basic information and brief medical history of the patients (Fig. 1B). Eventually, 83 affected eyes and 61 non-ERM eyes were figured out from the image set. There is no significant difference with right or left eye between the ERM and non-ERM groups (chi-square test, χ2 = 0.006, P = 0.9383) (Table 2). The mixture of both ERM and non-ERM eyes was reserved for subsequent analysis.
Manual and AI interpretation of fundus photograph set
Three ophthalmologists analyzed the fundus photograph set individually. The diagnostic accuracy of the three examiners were 74.31 ± 1.82%, 73.61 ± 1.84%, and 72.22 ± 1.87% (95% CI) respectively. Kappa tests between any two of them showed intermediate to high consistency among the three individual analyses. (ophthalmologist 1 & 2: kappa = 0.631, z value = 7.909, P = 0.000****; ophthalmologist 1 & 3: kappa = 0.529, z value = 7.081, P = 0.000****; ophthalmologist 2 & 3: kappa = 0.669, z value = 8.157, P = 0.000****). The AI model showed higher accuracy than all three manual interpretations, with an accuracy of 77.08 ± 1.75% (95% CI). However, the accuracy between any ophthalmologists and the AI model was not significantly different. (ophthalmologist 1 vs. AI: u = 0.5495, P > 0.05; ophthalmologist 2 vs. AI: u = 0.6836, P > 0.05; ophthalmologist 3 vs. AI: u = 0.9482, P > 0.05; u test). The highest sensitivity was achieved by the AI model (75.90%), whereas the highest specificity was found in manual interpretation (98.36%). Generally, diagnosis of ERM based on fundus photographs is an approach with high specificity (73.77–98.36%) but relatively low sensitivity (53.01–75.90%).
In order to investigate whether re-examination of the same fundus photograph would increase diagnostic accuracy, opinions from all three examiners were integrated. Diagnosis with most supporters was defined as the integrated manual diagnosis, representing the general performance of manual diagnosis. The integrated manual diagnosis resulted in an accuracy of 75.69%, which is slightly higher than the simple average of the manual accuracy (73.38%) and is higher than all of the examiners’ accuracy (Table 3). Although it is still lower than the accuracy of the AI model, no statistically significant difference was found between the accuracy of integrated manual diagnosis and the AI model. (u test, u = 0.2774, P > 0.05).
Further comparison between correct detection of ERM made by integrated manual analysis and AI model was shown in Fig. 4. Among all the correctly diagnosed ERM cases, 49 (73.13%) images were correctly detected by both manual approaches and AI model. 4 (5.97%) retinal images were diagnosed by manual approaches only, and another 14 (20.90%) ERM positive fundus photographs were detected by AI model only. The total number of the correctly categorized pictures will add up to 67 on either positive result. Therefore, taking both AI outputs and “manual intelligence” into consideration will obviously increase the sensitivity of diagnostic approaches. Comparison of correct elimination of ERM between integrated manual approaches and AI model were also conducted. Among all the correctly classified as non-ERM cases, 48 (85.71%) photographs were correctly eliminated both manually and with the AI model. 8 (14.29%) fundus images without ERM were detected by manual analysis only. All non-ERM cases recognized by the AI model were successfully figured out by manual diagnosis, which indicates better performance of ophthalmologists in negative prediction comparing with the AI model. Despite the differences between the two approaches, the AI model detection and manual diagnosis still greatly overlapped with each other (73.13% in correct ERM detection and 85.71% in non-ERM), which further proves the consistency between them.
To eliminate the bias caused by imbalanced case numbers between ERM and non-ERM, the performance of ERM detection of the AI model was presented as receiver operating characteristic (ROC) curve and areas under the curve (AUC) (Fig. 5). The performance of AI model was comparable with ophthalmologists, with one operating point above, one under, and one on the ROC curve of AI model. AUC of the AI model was 0.8566 ± 0.061 (95% CI) using OCT diagnosis as the standard diagnosis.
Evaluation of AI model in simulative clinical situation
In order to mimic the diagnosis of ERM in clinics, we next provided both fundus photographs and OCT results with other clinical information to 2 retinal specialists respectively, resulted in a comprehensive diagnosis. Diagnosis made by 2 specialists showed high consistency to each other (kappa = 0.795, z value = 9.603, P = 0.000****). ROC curves of the AI model using either specialists’ diagnosis as standard diagnosis was illustrated in Fig. 6. Comparable performance between AI and ophthalmologists was found in both situations, with AUC of 0.8879 ± 0.0542 (specialist 1, 95% CI) and 0.9065 ± 0.0473 (specialist 2, 95% CI).
Discussion
ERM is a retinal disease that can be commonly seen in ophthalmology clinics. Its diagnosis is mostly based on clinical findings. Variety of examinations have also been applied in ERM diagnosis: fluorescein angiography provides unique information for ERM diagnosis but is invasive and of more side effects; OCT, as a novel non-invasive imaging technique, provides an intuitive diagnosis of ERM. However, these examinations are less conducted in clinic than fundus photographs. Little studies have been reported comparing ERM diagnosis based on fundus photographs and OCT images. The only one we found in the literature reported a concordance rate of 89.13% between diagnosis based on non-mydriatic fundus images and OCT images in a group of 32 outpatients (46 eyes) with suspected idiopathic ERM50. Here we reported accuracy of ERM diagnosis based on fundus photographs between 72.22 and 74.31% (comparing with diagnosis based on OCT images), which is lower than the previously reported concordance rate. This may due to our inclusion of secondary ERM in the study, resulting in complicated fundus conditions. With a sensitivity of 63.86% by ophthalmologists, diagnosis of ERM based solely on fundus photographs take a huge risk of missed diagnosis. To our best knowledge, most prevalence and risk factor studies of ERM were based on diagnosis using fundus photographs13. The two studies that used OCT-based diagnosis were in urban and rural Chinese populations14,51, thus reported prevalence of ERM may be greatly underestimated in other populations.
Previously when manual labeling and AI testing were based on the same materials with manual diagnosis itself was set as standard diagnosis, it is impossible to prove AI is better than manual diagnosis since both sensitivity and specificity cannot exceed 1. Our study used a labeling method (diagnosis based on OCT images) with higher accuracy than the testing method (diagnosis based on retinal images). This kind of up-level labeling makes it possible to draw the conclusion that AI model have better performance than manual analysis and validate substitution of physicians with AI model in some conditions. Our reported AI model was based solely on retinal images. We expect higher sensitivity and specificity in ERM detection when further information, such as symptoms and medical history, are also included in AI model to improve its discrimination ability.
The application of our AI model in ERM detection, however, is not without limitation. ERM is an age-related disease. In other words, the prevalence of the ERM is higher in elder population. Prevalence of cataracts or opaque vitreous is also increased in this population, making it difficult to acquire clear retinal images. In our study, when unclear fundus images were eliminated, we reported an accuracy of 77.08% for AI detection. Most of the fundus images misclassified by AI model were of large area of darkness around macula. (Supplementary Figure 1). In the scenario of disease screening in population, however, when most people are healthy and have clear cornea, lens, and vitreous, it is easier to acquire fundus images of high quality and better performance of the AI model is expected. Moreover, the distribution of the disease and non-disease group of test image set in this study was slightly different from the dataset used in the AI model development. The average age of patients in this study is older, and the spectrum of the disease likely shifts to severe ERM, which makes the model tend to predict disease cases rather than exclude non-disease ones. This is in accord with our reported lower specificity of AI model (78.69%) compared with manual diagnosis (91.80%), in other words, more people without ERM are labeled as ERM by the AI model.
The choice of threshold of our AI model for positive prediction in this study addressed a balance between sensitivity and specificity. In Fig. 5, the operation points of ophthalmologist 1 and 2 are below or on the ROC curve of AI model, which means the model has a better or equivalent performance. For the diagnosis of ophthalmologist 3, although its operation point is above the curve due to higher specificity, its sensitivity is too low, indicating that its use is biased and limited. Furthermore, in a specific application, the model can be calibrated by choosing a threshold corresponding to a preferred sensitivity while achieving a comparable specificity with ophthalmologist’s diagnosis, and vice versa. Thus our AI model could be used as a possible alternative of acceptable accuracy at a relatively low cost when there is difficult access to ophthalmological specialists or advanced examination. Another limitation of our study is that we have not validated the use of this AI model in populational screening, which could be conducted in our sequential studies.
References
Johnson, T. M. & Johnson, M. W. Epiretinal membrane. In Ophthalmology 4th edn (eds Yankoff, M. & Duker, J.) 614–619 (Elsevier, 2013).
Fraser-Bell, S., Guzowski, M., Rochtchina, E., Wang, J. J. & Mitchell, P. Five-year cumulative incidence and progression of epiretinal membranes: The Blue Mountains Eye Study. Ophthalmology 110, 34–40. https://doi.org/10.1016/s0161-6420(02)01443-4 (2003).
Ghazi-Nouri, S. M., Tranos, P. G., Rubin, G. S., Adams, Z. C. & Charteris, D. G. Visual function and quality of life following vitrectomy and epiretinal membrane peel surgery. Br. J. Ophthalmol. 90, 559–562. https://doi.org/10.1136/bjo.2005.085142 (2006).
Inoue, M. & Kadonosono, K. Macular diseases: Epiretinal membrane. Dev. Ophthalmol. 54, 159–163. https://doi.org/10.1159/000360462 (2014).
Govetto, A. et al. Microcystoid macular changes in association with idiopathic epiretinal membranes in eyes with and without glaucoma: Clinical insights. Am. J. Ophthalmol. 181, 156–165. https://doi.org/10.1016/j.ajo.2017.06.023 (2017).
Cheung, N. et al. Prevalence and risk factors for epiretinal membrane: The Singapore Epidemiology of Eye Disease study. Br. J. Ophthalmol. 101, 371–376. https://doi.org/10.1136/bjophthalmol-2016-308563 (2017).
Fraser-Bell, S., Ying-Lai, M., Klein, R. & Varma, R. Prevalence and associations of epiretinal membranes in latinos: The Los Angeles Latino Eye Study. Investig. Ophthalmol. Vis. Sci. 45, 1732–1736. https://doi.org/10.1167/iovs.03-1295 (2004).
Kawasaki, R. et al. Prevalence and associations of epiretinal membranes in an adult Japanese population: The Funagata study. Eye (London) 23, 1045–1051. https://doi.org/10.1038/eye.2008.238 (2009).
Keel, S. et al. Prevalence and associations of epiretinal membranes in the Australian National Eye Health Survey. Acta Ophthalmol. 95, e796–e798. https://doi.org/10.1111/aos.13386 (2017).
Kim, J. M. et al. Epiretinal membrane: Prevalence and risk factors from the Korea National Health and Nutrition Examination Survey, 2008 through 2012. Korean J. Ophthalmol. 31, 514–523. https://doi.org/10.3341/kjo.2016.0098 (2017).
Ng, C. H. et al. Prevalence and risk factors for epiretinal membranes in a multi-ethnic United States population. Ophthalmology 118, 694–699. https://doi.org/10.1016/j.ophtha.2010.08.009 (2011).
Wang, S. B. et al. Prevalence and risk factors of epiretinal membrane in a cohort with cardiovascular disease risk, compared with the Blue Mountains Eye Study. Br. J. Ophthalmol. 99, 1601–1605. https://doi.org/10.1136/bjophthalmol-2015-306776 (2015).
Xiao, W., Chen, X., Yan, W., Zhu, Z. & He, M. Prevalence and risk factors of epiretinal membranes: A systematic review and meta-analysis of population-based studies. BMJ Open 7, e014644. https://doi.org/10.1136/bmjopen-2016-014644 (2017).
Ye, H. et al. Prevalence and associations of epiretinal membrane in an elderly urban Chinese population in China: The Jiangning Eye Study. Br. J. Ophthalmol. 99, 1594–1597. https://doi.org/10.1136/bjophthalmol-2015-307050 (2015).
You, Q., Xu, L. & Jonas, J. B. Prevalence and associations of epiretinal membranes in adult Chinese: The Beijing eye study. Eye (London) 22, 874–879. https://doi.org/10.1038/sj.eye.6702786 (2008).
Dawson, S. R., Shunmugam, M. & Williamson, T. H. Visual acuity outcomes following surgery for idiopathic epiretinal membrane: An analysis of data from 2001 to 2011. Eye (London) 28, 219–224. https://doi.org/10.1038/eye.2013.253 (2014).
Asaria, R., Garnham, L., Gregor, Z. J. & Sloper, J. J. A prospective study of binocular visual function before and after successful surgery to remove a unilateral epiretinal membrane. Ophthalmology 115, 1930–1937. https://doi.org/10.1016/j.ophtha.2008.05.020 (2008).
Falkner-Radler, C. I., Glittenberg, C., Hagen, S., Benesch, T. & Binder, S. Spectral-domain optical coherence tomography for monitoring epiretinal membrane surgery. Ophthalmology 117, 798–805. https://doi.org/10.1016/j.ophtha.2009.08.034 (2010).
Inoue, M. et al. Preoperative inner segment/outer segment junction in spectral-domain optical coherence tomography as a prognostic factor in epiretinal membrane surgery. Retina 31, 1366–1372. https://doi.org/10.1097/IAE.0b013e318203c156 (2011).
Kim, J. H., Kim, Y. M., Chung, E. J., Lee, S. Y. & Koh, H. J. Structural and functional predictors of visual outcome of epiretinal membrane surgery. Am. J. Ophthalmol. 153, 103-110.e101. https://doi.org/10.1016/j.ajo.2011.06.021 (2012).
Suh, M. H., Seo, J. M., Park, K. H. & Yu, H. G. Associations between macular findings by optical coherence tomography and visual outcomes after epiretinal membrane removal. Am. J. Ophthalmol. 147, 473-480.e473. https://doi.org/10.1016/j.ajo.2008.09.020 (2009).
Cobos, E. et al. Preoperative study of the inner segment/outer segment junction of photoreceptors by spectral-domain optical coherence tomography as a prognostic factor in patients with epiretinal membranes. Clin. Ophthalmol. 7, 1467–1470. https://doi.org/10.2147/opth.S44837 (2013).
Mitamura, Y., Hirano, K., Baba, T. & Yamamoto, S. Correlation of visual recovery with presence of photoreceptor inner/outer segment junction in optical coherence images after epiretinal membrane surgery. Br. J. Ophthalmol. 93, 171–175. https://doi.org/10.1136/bjo.2008.146381 (2009).
Kim, J. H., Kang, S. W., Kong, M. G. & Ha, H. S. Assessment of retinal layers and visual rehabilitation after epiretinal membrane removal. Graefe’s Arch. Clin. Exp. Ophthalmol. = Albrecht von Graefes Archiv fur klinische und experimentelle Ophthalmologie 251, 1055–1064. https://doi.org/10.1007/s00417-012-2120-7 (2013).
Miguel, A. I. & Legris, A. Prognostic factors of epiretinal membranes: A systematic review. J. Fr. Ophtalmol. 40, 61–79. https://doi.org/10.1016/j.jfo.2016.12.001 (2017).
McLeod, D., Hiscott, P. S. & Grierson, I. Age-related cellular proliferation at the vitreoretinal juncture. Eye (London) 1(Pt 2), 263–281. https://doi.org/10.1038/eye.1987.46 (1987).
Do, D. V. et al. The impact of optical coherence tomography on surgical decision making in epiretinal membrane and vitreomacular traction. Trans. Am. Ophthalmol. Soc. 104, 161–166 (2006).
Wilkins, J. R. et al. Characterization of epiretinal membranes using optical coherence tomography. Ophthalmology 103, 2142–2151. https://doi.org/10.1016/s0161-6420(96)30377-1 (1996).
Kinoshita, T. et al. Time course of changes in metamorphopsia, visual acuity, and OCT parameters after successful epiretinal membrane surgery. Investig. Ophthalmol. Vis. Sci. 53, 3592–3597. https://doi.org/10.1167/iovs.12-9493 (2012).
Ting, D. S. W. et al. Development and validation of a deep learning system for diabetic retinopathy and related eye diseases using retinal images from multiethnic populations with diabetes. JAMA 318, 2211–2223. https://doi.org/10.1001/jama.2017.18152 (2017).
He, J. et al. Artificial intelligence-based screening for diabetic retinopathy at community hospital. Eye (London) https://doi.org/10.1038/s41433-019-0562-4 (2019).
Gulshan, V. et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA 316, 2402–2410. https://doi.org/10.1001/jama.2016.17216 (2016).
Gargeya, R. & Leng, T. Automated identification of diabetic retinopathy using deep learning. Ophthalmology 124, 962–969. https://doi.org/10.1016/j.ophtha.2017.02.008 (2017).
Li, Z. et al. Efficacy of a deep learning system for detecting glaucomatous optic neuropathy based on color fundus photographs. Ophthalmology 125, 1199–1206. https://doi.org/10.1016/j.ophtha.2018.01.023 (2018).
Wu, X. et al. Universal artificial intelligence platform for collaborative management of cataracts. Br. J. Ophthalmol. 103, 1553–1560. https://doi.org/10.1136/bjophthalmol-2019-314729 (2019).
Burlina, P. M. et al. Automated grading of age-related macular degeneration from color fundus images using deep convolutional neural networks. JAMA Ophthalmol. 135, 1170–1176. https://doi.org/10.1001/jamaophthalmol.2017.3782 (2017).
Grassmann, F. et al. A deep learning algorithm for prediction of age-related eye disease study severity scale for age-related macular degeneration from color fundus photography. Ophthalmology 125, 1410–1420. https://doi.org/10.1016/j.ophtha.2018.02.037 (2018).
Brown, J. M. et al. Automated diagnosis of plus disease in retinopathy of prematurity using deep convolutional neural networks. JAMA Ophthalmol. 136, 803–810. https://doi.org/10.1001/jamaophthalmol.2018.1934 (2018).
Mao, J. et al. Automated diagnosis and quantitative analysis of plus disease in retinopathy of prematurity based on deep convolutional neural networks. Acta Ophthalmol. https://doi.org/10.1111/aos.14264 (2019).
Hu, J., Chen, Y., Zhong, J., Ju, R. & Yi, Z. Automated analysis for retinopathy of prematurity by deep neural networks. IEEE Trans. Med. Imaging 38, 269–279. https://doi.org/10.1109/tmi.2018.2863562 (2019).
LeCun, Y., Bengio, Y. & Hinton, G. Deep learning. Nature 521, 436–444. https://doi.org/10.1038/nature14539 (2015).
Lu, W. et al. Deep learning-based automated classification of multi-categorical abnormalities from optical coherence tomography images. Transl. Vis. Sci. Technol. 7, 41. https://doi.org/10.1167/tvst.7.6.41 (2018).
Sonobe, T. et al. Comparison between support vector machine and deep learning, machine-learning technologies for detecting epiretinal membrane using 3D-OCT. Int. Ophthalmol. 39, 1871–1877. https://doi.org/10.1007/s10792-018-1016-x (2019).
Salvaraju, R. et al. Grad-CAM: Visual explanations from deep networks via gradient-based localization. In The International Conference on Computer Vision, 618–626 (2017).
Szegedy, C. et al. Inception-v4, inception-resnet and the impact of residual connections on learning. Preprint arXiv:1602.07261 (2016).
Chollet, F. Xception: Deep learning with depthwise separable convolutions. In The IEEE Conference on Computer Vision and Pattern Recognition, 1251–1258 (2017).
Abadi, M. et al. Tensorflow: Large-scale machine learning on heterogeneous distributed systems. Preprint arXiv:1603.04467 (2016).
Lin, D. et al. Application of Comprehensive Artificial intelligence Retinal Expert (CARE) system: A national real-world evidence study. Lancet Digit. Health 3, e486–e495. https://doi.org/10.1016/S2589-7500(21)00086-8 (2021).
Szegedy, C. et al. Going deeper with convolutions. In IEEE Conference on Computer Vision and Pattern Recognition, 1–9 (2015).
Hong, W. Application of non-mydriatic fundus photography in the diagnosis of idiopathic macular epiretinal membranes. Anat. Res. 36, 457–459 (2014) (in Chinese).
Duan, X. R. et al. Prevalence and associations of epiretinal membranes in a rural Chinese adult population: The Handan Eye Study. Investig. Ophthalmol. Vis. Sci. 50, 2018–2023. https://doi.org/10.1167/iovs.08-2624 (2009).
Author information
Authors and Affiliations
Contributions
Idea came from Y.H. and B.W. Y.H. and E.S. designed the experiment. C.L. and X.Y. developed the AI model. E.S. and L.W. collected basic information, brief medical history, fundus photographs and OCT images. L.W., D.S, L.G. and Y.H. interpreted fundus photographs and OCT images. E.S. and C.L. analyzed data. E.S. wrote the manuscript. E.S. and C.L. prepared all figures. E.S., C.L., J.X., B.W. and Y.H. critically revised the manuscript. All authors reviewed the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Shao, E., Liu, C., Wang, L. et al. Artificial intelligence-based detection of epimacular membrane from color fundus photographs. Sci Rep 11, 19291 (2021). https://doi.org/10.1038/s41598-021-98510-x
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-021-98510-x
This article is cited by
-
Development of a generative deep learning model to improve epiretinal membrane detection in fundus photography
BMC Medical Informatics and Decision Making (2024)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.