Skip to main content

MINI REVIEW article

Front. Med., 23 February 2023
Sec. Obstetrics and Gynecology
Volume 10 - 2023 | https://doi.org/10.3389/fmed.2023.1098205

Current developments in artificial intelligence from obstetrics and gynecology to urogynecology

  • Department of Obstetrics and Gynecology, Ankara University School of Medicine, Ankara, Türkiye

In today’s medical practice clinicians need to struggle with a huge amount of data to improve the outcomes of the patients. Sometimes one clinician needs to deal with thousands of ultrasound images or hundred papers of laboratory results. To overcome this shortage, computers get in help of human beings and they are educated under the term “artificial intelligence.” We were using artificial intelligence in our daily lives (i.e., Google, Netflix, etc.), but applications in medicine are relatively new. In obstetrics and gynecology, artificial intelligence models mostly use ultrasound images for diagnostic purposes but nowadays researchers started to use other medical recordings like non-stress tests or urodynamics study results to develop artificial intelligence applications. Urogynecology is a developing subspecialty of obstetrics and gynecology, and articles about artificial intelligence in urogynecology are limited but in this review, we aimed to increase clinicians’ knowledge about this new approach.

Introduction

In contrast to natural intelligence produced by animals, including humans, artificial intelligence (AI) refers to the intelligence demonstrated by computer systems. We use AI in daily life routinely, with web search engines (like Google), recommendation systems on online entertainment platforms (YouTube and Netflix), speech recognition (Siri and Alexa), and self-driving cars (Tesla) are some examples. In 1956, John McCarthy organized the first academic meeting on the subject at Dartmouth College, coining the phrase “artificial intelligence.” Artificial intelligence can be categorized in three headings (Figure 1).

FIGURE 1
www.frontiersin.org

Figure 1. The three types of artificial intelligence.

Artificial narrow intelligence

It is possible to develop problem-solving skills using text, audio, or image data. A processor could easily complete a single task at this level of AI. In spite of AI’s supremacy in chess, it has an IQ of just 0 (zero).

Artificial general intelligence

At this degree of artificial intelligence, an intelligent creature can fully comprehend all complicated tasks in the same way that an individual can.

Artificial superintelligence

While humans will never attain this degree of intelligence, machines may be able to surpass humans in terms of intelligence and behavioral skills.

Working mechanism of artificial intelligence

Machine learning

Human brain-like capabilities for AI are mostly achieved via machine learning. Machine learning uses a big amount of data to figure out how factors interact with each other. Supervised, unsupervised, and reinforcement learning are the three main approaches of machine learning (Table 1). To begin the supervised learning method, the input data must first be labeled. This is a great method for classifying and calculating regressions. Algorithms are designed to use this method because we know exactly what they should learn, and it’s the most popular method of training by far.

TABLE 1
www.frontiersin.org

Table 1. The three main machine learning techniques.

For unsupervised learning, there is no requirement for labeled data, in contrast to supervised learning. In data analysis and the generation of novel hypotheses, it is routinely used to uncover patterns from the data that could otherwise go undetected (1).

It is possible to think of reinforcement learning as a hybrid between unsupervised and supervised. It allows the model to learn how to complete the tasks via a sequence of decisions without being explicitly instructed on how to do so.

Deep learning

Deep learning is a subset of machine learning and artificial neural networks mimic the human brain’s ability to make automatic predictions from training data sets using numerous layers of artificial neuronal networks. Deep learning refers to the use of numerous layers of filters, each of which provides an output score that serves as the input for the following layer. Table 2 describes the key differences between machine learning and deep learning.

TABLE 2
www.frontiersin.org

Table 2. Differences between machine learning and deep learning.

Artificial intelligence in obstetrics and gynecology

Artificial intelligence with its capabilities will enable a much more effective and efficient healthcare system. AI can aid in the determination of diagnosis and prognosis, therapy optimization, and drug discovery. Artificial intelligence is becoming popular in the field of obstetrics and gynecology rapidly. At the 29th World Congress of the International Society of Ultrasound in Obstetrics and Gynecology (ISUOG) in 2019, 14 abstracts explicitly addressing AI were presented, on the other hand, a total of 13 abstracts were presented at the previous six ISUOG World Congresses (2013–2018).

Ultrasound images for nearly all situations in obstetrics and gynecology are available. As a result of this large database, AI applications for perinatal medicine and gynecologic oncology were reported (212).

The great majority of the research relied on ultrasound images, but physicians also tried out AI-based cardiotocography analysis for intrapartum fetal monitoring (13). Idowu et al. (14) emphasized the importance of using artificial intelligence to detect preterm labor. Electrohysterography signals were used in this study, and they were identified using three different machine-learning algorithms to help detect true labor and reliably diagnose premature labor (14).

Couples are being referred for assisted reproductive therapies at an increasing rate, and AI solutions in this field are gaining ground daily. Guy and colleagues used data mining and artificial intelligence to create a computer model that can help clinicians in the prediction of pregnancy after in vitro fertilization (IVF). Data mining (DM) is a technique for uncovering patterns in large databases that combines artificial intelligence and advanced statistics. DM obtains the needed data and is also capable of finding other important factors that may influence the outcome, hence increasing the amount of data that can be used (15). Manna et al. (16) suggested combining AI and ANNs to extract texture descriptors from oocyte or embryo images, with this method AI can identify the most viable oocytes and embryos with a high chance of pregnancy.

Neural network models are being used to estimate prognoses for patients with ovarian cancer. In a report, Enshaei et al. revealed that ANNs could predict survival with a 97% accuracy (17). It can also predict which treatment will be most beneficial for each patient depending on their diagnosis. Researchers at Brigham and Women’s Hospital and Dana-Farber Cancer Institute used AI to manipulate vast volumes of microribonucleic acid (RNA) data to build models capable of diagnosing early ovarian cancer (17). In comparison to an ultrasound screening test, which detected abnormal results less than 5% of the time, the AI neural network could keep up with the intricate linkages between microRNA and accurately recognized nearly 100% of anomalies associated with ovarian cancer (18). Additionally, AI outperformed human experts when it came to analyze pre-cancer images of the cervical region. Deep learning algorithms based on artificial intelligence can collect a large number of photographs related to cervical cancer screening and reliably detect diseased tissue. Patients can be treated on the same visit because minimum training is required and the results are obtained quickly (19).

Physical AI has been employed in surgery more frequently than virtual AI. Virtual AI forecasts the outcome using established patient parameters, repeating patterns, and treatment algorithms, as opposed to the surgical field, which has a high number of independent variables. The consistency of different tissues, the surgeon’s skill level, the changes made to the surgical field while operating, and a unique variation between patients and their pathology are only a few of these variables; ultimately, these distinct characteristics make designing an algorithm challenging (20). In various ways, AI has helped in gynecological surgery, including imaging and spatial awareness. Artificial intelligence can aid surgeons by enhancing imaging both before and during surgery. Three-dimensional (3DP) printing that mimics the operation site is far superior to two-dimensional (2D) printing because it provides a more accurate picture of the actual mode (21). This enables more exact preoperative planning, realistic trainee practice, and previously impossible preoperative planning.

Additionally, AI has contributed to the reduction of operating time and precision, resulting in fewer operative issues. Such augmented reality is used to accomplish this. A computer reconstructs real-world items and digitally enhances them to generate a more informative visual representation in augmented reality. While the system has several drawbacks, it also has some advantages, such as increased precision, safety, and a reduction in the time it takes to execute tasks (22). AI has aided in the development of spatial awareness, notifying surgeons when vital vessels or structures are hidden, allowing them to be recognized quickly and vital structures to be protected. One example is isolating the ureters during gynecological surgery. An endoscopic system powered by artificial intelligence was used in a study to detect the depth and position of the ureters using algorithms, displaying greater accuracy and safety (23).

Artificial intelligence in urogynecology

As is the issue with many other fields of medicine, the use of AI in urogynecology is still experimental, and the optimal regions for AI in urogynecology are not clearly specified; but we will endeavor to summarize the current understanding regarding the use of AI in urogynecology.

To begin, advances in artificial intelligence allow telemedicine to advance. Patients’ conditions might be remotely monitored, tracked, and regulated with the help of wearable devices linked to AI systems. Electronic medical records software currently automates caregiver scheduling, organizes care plans, provides follow-up alerts, and automates payment, as well as offering patient and family portals. Virtual visits allow patients who live far away or have restricted mobility to receive follow-up care without having to undergo a physical examination, which can cut wait times. Virtual visits are becoming more common, particularly during the peak days of the COVID-19 pandemic. Wearable devices like urine incontinence monitors and post-void residual bladder volume scanners could help telemedicine in urogynecology. Wearable bladder volume monitors have recently been developed for use in children with nocturnal enuresis and other types of urinary incontinence (24). To monitor patients, treating teams may be able to access collected data remotely, or this may someday be done by an AI system.

Deep learning may be used to recognize medical pictures, as previously indicated, and Oral et al. (25) evaluated the role of dynamic MRI in the diagnosis and quantification of POP. From a midsagittal aspect, they looked at the 15 dynamic MR images. Although the writers did not appropriately identify the stage of prolapse in the publication, these pictures portray patients at varying levels of prolapse who had not previously received POP surgery. For all MRI reference sites, the semiautomated pelvic floor measurement method produced exceptionally consistent and exact placements. Furthermore, the model detects reference locations quicker than the old method. In the following years, Nekooeimehr et al. (26) proposed a method for automatically monitoring and segmenting pelvic organs on dynamic magnetic resonance imaging (MRI), followed by multiple-object trajectory classification, to aid in the understanding of pelvic organ prolapse (POP). According to their findings, the current method is capable of autonomously tracking and segmenting pelvic organs in 94 cases with a Dice similarity score of more than 78% and a Hausdorff distance of 5.2 mm. Statistically significant associations between various radiomic markers and urodynamic findings were discovered by Keene et al. The authors concluded that CT texture analysis of the bladder wall could be a useful tool for identifying patients with high-risk urodynamic features in spina bifida and that it could be employed instead of or in addition to urodynamics in neurogenic patient populations in the future (27). To reduce human interference and standardize urodynamic tract interpretation, Wang et al. (28) identified detrusor overactivty (DO) patterns linked with clinical findings. They used manifold learning (29, 30) and dynamic time-warping techniques. They reported an overall accuracy of 81.35%, a sensitivity of 76.92%, and a specificity of 81.41% for their AI-based model for identifying the DO. Cullingsworth et al. (31) focused examined the frequency and amplitudes of low-amplitude rhythmic detrusor contractions (LARC) in patients with DO and discovered a subset of them. The model was automatically constructed using a Fast Fourier Transform approach (32) that yielded 100% specificity. This level of precision was obtained while LARC was recognized independently of abdominal pressure traces, paving the way for future automated urodynamic trace interpretation. Niederhauser et al. (33) used ‘Wavelet’ time-frequency analysis in conjunction with AI-based algorithms to determine the existence of various subgroups of overactive bladder (OAB) patients based on the amplitude and frequency of non-voiding contractions and to investigate the possibility of automatically detecting DO in the third promising study on machine learning-based urodynamic automation. They were successful in identifying various OAB subgroups, and their machine learning-based algorithms for DO detection functioned well. There is a scarcity of information about AI’s therapeutic implications of in urogynecology. Whangbo and colleagues developed a device that can be used instead of a voiding diary. This is a smart bracelet that uses the patient’s particular postures to identify the time and the intervals between micturition. The voiding diary is the gold standard for examining patients with lower urinary tract symptoms, but it takes time and necessitates that patients stay home. This technology will provide data on actual VD and ease clinical evaluation (34).

In a cohort of 559 female patients treated with anticholinergic drugs for OAB, Sheyn et al. (35) built and externally validated a prediction model for anticholinergic response using an RF algorithm. The model performed brilliantly (sensitivity 80.4%; sensitivity 77.4%). This study relied on a subjective report of successful treatment, regardless of its good outcome.

Although it is generally intended by the surgeons on the field of urogynecology, the reconstruction of pelvic anatomy with surgery may not always improve health related quality-of-life (QoL) in all women with pelvic floor dysfunctions. Thus, predicting reduced health related QoL after a surgical procedure for a pelvic floor disorder provides the opportunity for individual targeted therapies of dysfunctions and disorders. Including symptom and quality-of-life scales in AI systems that evaluate the diagnosis and follow-up progress after treatments seem to be promising for the accurate and effective management of pelvic floor dysfunctions. However, these studies were limited in the literature and this issue should be investigated in future studies.

Challenges of the artificial intelligence

While there are many potential benefits to introducing AI into clinical practice, there are also many challenges and unknowns that may cause worry. The quantity and quality of the data used to generate the models is the fundamental constraint on the robustness of these methods. This is particularly true for rare disorders that have erratic or subjective symptoms. As a result, we have been asked to automate data collection and quality control to get more powerful tools. Medical software, electronic medical records, care platforms, ambulatory devices, patient parameters, surveys, and measures of patient-related outcomes should all be developed in collaboration. Finally, ethical means for archiving surgical and perioperative data must be devised. The collection of high-quality data that can ultimately help patients requires standardization of methods across broad territorial groups and coordination among institutions.

When applying AI in clinical practice, another factor to consider is its applicability. Non-mentioned clinical data are ignored by the AI model because AI models can only account for data that were ‘seen’ during training. As a result, a growing segment of healthcare for AI research is committed to constructing AI models that integrate imaging and electronic health record data to enable accurate diagnosis and treatment.

When using machine learning in a clinical setting, the most important issue to overcome is trust, which arises when clinicians and patients accept the system’s recommendations. The data is noisy, sophisticated, multidimensional, and skewed toward the catchment area of the founding hospital systems, where the model was created. Another issue is the possibility of professional liability for doctors who use AI. Should hospitals and doctors be held liable for judgments made by artificial intelligence software?

Conclusion

Throughout the last decade, excitement for artificial intelligence has increased massively. Numerous recent studies have focused on the potential of artificial intelligence in urogynecology not only to investigate the pathophysiology of lower urinary tract dysfunction but also to use it as a diagnostic tool by boosting the capabilities of existing techniques such as dynamic MRI, functional MRI, or urodynamics. By strengthening prediction models, AI may shortly have significant therapeutic benefits in the field of urogynecology. Clinical practice, on the other hand, has always required physicians to handle huge amounts of data, ranging from histories and physical examinations to laboratory and imaging investigations, and, most recently, genetic data. The capacity to handle this complexity will always distinguish good physicians.

Author contributions

All authors listed have made a substantial, direct, and intellectual contribution to the work, and approved it for publication.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1.Deo, RC. Machine learning in medicine. Circulation. (2015) 132:1920–30. doi: 10.1161/CIRCULATIONAHA.115.001593

PubMed Abstract | CrossRef Full Text | Google Scholar

2.Timmerman, D, Verrelst, H, Bourne, TH, de Moor, B, Collins, WP, Vergote, I, et al. Artificial neural network models for the preoperative discrimination between malignant and benign adnexal masses. Ultrasound Obstet Gynecol. (1999) 13:17–25. doi: 10.1046/j.1469-0705.1999.13010017.x

PubMed Abstract | CrossRef Full Text | Google Scholar

3.Eriksson, LSE, Epstein, E, Testa, AC, Fischerova, D, Valentin, L, Sladkevicius, P, et al. Ultrasound-based risk model for preoperative prediction of lymph-node metastases in women with endometrial cancer: model-development study. Ultrasound Obstet Gynecol. (2020) 56:443–52. doi: 10.1002/uog.21950

PubMed Abstract | CrossRef Full Text | Google Scholar

4.Huang, YL, and Chen, HY. Computer-aided diagnosis of urodynamic stress incontinence with vector-based perineal ultrasound using neural networks. Ultrasound Obstet Gynecol. (2007) 30:1002–6. doi: 10.1002/uog.4102

PubMed Abstract | CrossRef Full Text | Google Scholar

5.van den Noort, F, van der Vaart, CH, Grob, ATM, van de Waarsenburg, MK, Slump, CH, and van Stralen, M. Deep learning enables automatic quantitative assessment of puborectalis muscle and urogenital hiatus in plane of minimal hiatal dimensions. Ultrasound Obstet Gynecol. (2019) 54:270–5. doi: 10.1002/uog.20181

PubMed Abstract | CrossRef Full Text | Google Scholar

6.Neocleous, AC, Syngelaki, A, Nicolaides, KH, and Schizas, CN. Two-stage approach for risk estimation of fetal trisomy 21 and other aneuploidies using computational intelligence systems. Ultrasound Obstet Gynecol. (2018) 51:503–8. doi: 10.1002/uog.17558

PubMed Abstract | CrossRef Full Text | Google Scholar

7.Bonet-Carne, E, Palacio, M, Cobo, T, Perez-Moreno, A, Lopez, M, Piraquive, JP, et al. Quantitative ultrasound texture analysis of fetal lungs to predict neonatal respiratory morbidity. Ultrasound Obstet Gynecol. (2015) 45:427–33. doi: 10.1002/uog.13441

PubMed Abstract | CrossRef Full Text | Google Scholar

8.Bahado-Singh, RO, Sonek, J, McKenna, D, Cool, D, Aydas, B, Turkoglu, O, et al. Artificial intelligence and amniotic fluid multiomics: prediction of perinatal outcome in asymptomatic women with short cervix. Ultrasound Obstet Gynecol. (2019) 54:110–8. doi: 10.1002/uog.20168

PubMed Abstract | CrossRef Full Text | Google Scholar

9.Tsur, A, Batsry, L, Toussia-Cohen, S, Rosenstein, MG, Barak, O, Brezinov, Y, et al. Development and validation of a machine-learning model for prediction of shoulder dystocia. Ultrasound Obstet Gynecol. (2020) 56:588–96. doi: 10.1002/uog.21878

PubMed Abstract | CrossRef Full Text | Google Scholar

10.Jugović, D, Tumbri, J, Medić, M, Jukić, MK, Kurjak, A, Arbeille, P, et al. New Doppler index for prediction of perinatal brain damage in growth-restricted and hypoxic fetuses. Ultrasound Obstet Gynecol. (2007) 30:303–11. doi: 10.1002/uog.4094

PubMed Abstract | CrossRef Full Text | Google Scholar

11.Papageorghiou, AT, Kemp, B, Stones, W, Ohuma, EO, Kennedy, SH, Purwar, M, et al. Ultrasound-based gestational-age estimation in late pregnancy. Ultrasound Obstet Gynecol. (2016) 48:719–26. doi: 10.1002/uog.15894

PubMed Abstract | CrossRef Full Text | Google Scholar

12.Xie, HN, Wang, N, He, M, Zhang, LH, Cai, HM, Xian, JB, et al. Using deep-learning algorithms to classify fetal brain ultrasound images as normal or abnormal. Ultrasound Obstet Gynecol. (2020) 56:579–87. doi: 10.1002/uog.21967

CrossRef Full Text | Google Scholar

13.Brocklehurst, P, INFANT Collaborative Group. A study of an intelligent system to support decision making in the management of labour using the cardiotocograph – the INFANT study protocol. BMC Pregnancy Childbirth. (2016) 16:10. doi: 10.1186/s12884-015-0780-0

PubMed Abstract | CrossRef Full Text | Google Scholar

14.Ibrahim, OI, Paul, F, Abir, H, Chelsea, D, Mohammed, K, Raul, VCE, et al. Artificial intelligence for detecting preterm uterine activity in gynecology and obstetric care. In Proceeding of 2015 IEEE International Conference on Computer and Information Technology; Ubiquitous Computing and Communications; Dependable, Autonomic and Secure Computing; Pervasive Intelligence and Computing. IEEE. (2015).

Google Scholar

15.Guh, RS, Tsung-Chieh, JW, and Shao-Ping, W. Integrating genetic algorithm and decision tree learning for assistance in predicting in vitro fertilization outcomes. Expert Syst Appl. (2011) 38:4437–49. doi: 10.1016/j.eswa.2010.09.112

CrossRef Full Text | Google Scholar

16.Manna, C, Nanni, L, Lumini, A, and Pappalardo, S. Artificial intelligence techniques for embryo and oocyte classification. Reprod Biomed Online. (2013) 26:42–9. doi: 10.1016/j.rbmo.2012.09.015

PubMed Abstract | CrossRef Full Text | Google Scholar

17.Enshaei, A, Robson, CN, and Edmondson, RJ. Artificial intelligence systems as prognostic and predictive tools in ovarian cancer. Ann Surg Oncol. (2015) 22:3970–5. doi: 10.1245/s10434-015-4475-6

PubMed Abstract | CrossRef Full Text | Google Scholar

18.Kann, BH, Thompson, R, Thomas, CR Jr, Dicker, A, and Aneja, S. Artificial intelligence in oncology: current applications and future directions. Oncology (Williston Park). (2019) 33:46–53

PubMed Abstract | Google Scholar

19.Hu, L, Bell, D, Antani, S, Xue, Z, Yu, K, Horning, MP, et al. An observational study of deep learning and automated evaluation of cervical images for cancer screening. J Natl Cancer Inst. (2019) 111:923–32. doi: 10.1093/jnci/djy225

PubMed Abstract | CrossRef Full Text | Google Scholar

20.Moawad, G, Tyan, P, and Louie, M. Artificial intelligence and augmented reality in gynecology. Curr Opin Obstet Gynecol. (2019) 31:345–8. doi: 10.1097/GCO.0000000000000559

PubMed Abstract | CrossRef Full Text | Google Scholar

21.Ajao, MO, Clark, NV, Kelil, T, Cohen, SL, and Einarsson, JI. Case report: three-dimensional printed model for deep infiltrating endometriosis. J Minim Invasive Gynecol. (2017) 24:1239–42. doi: 10.1016/j.jmig.2017.06.006

PubMed Abstract | CrossRef Full Text | Google Scholar

22.Dirie, NI, Wang, Q, and Wang, S. Two-dimensional versus three-dimensional laparoscopic Systems in Urology: a systematic review and meta-analysis. J Endourol. (2018) 32:781–90. doi: 10.1089/end.2018.0411

PubMed Abstract | CrossRef Full Text | Google Scholar

23.Song, E, Yu, F, Liu, H, Cheng, N, Li, Y, Jin, L, et al. A novel endoscope system for position detection and depth estimation of the ureter. J Med Syst. (2016) 40:266. doi: 10.1007/s10916-016-0607-1

PubMed Abstract | CrossRef Full Text | Google Scholar

24.van Leuteren, PG, Nieuwhof-Leppink, AJ, and Dik, P. SENS-U: clinical evaluation of a full-bladder notification – a pilot study. J Pediatr Urol. (2019) 15:381.e1–5. doi: 10.1016/j.jpurol.2019.04.006

PubMed Abstract | CrossRef Full Text | Google Scholar

25.Onal, S, Lai-Yuen, S, Bao, P, Weitzenfeld, A, Greene, K, Kedar, R, et al. Assessment of a semiautomated pelvic floor measurement model for evaluating pelvic organ prolapse on MRI. Int Urogynecol J. (2014) 25:767–73. doi: 10.1007/s00192-013-2287-4

PubMed Abstract | CrossRef Full Text | Google Scholar

26.Nekooeimehr, I, Lai-Yuen, S, Bao, P, Weitzenfeld, A, and Hart, S. Automated contour tracking and trajectory classification of pelvic organs on dynamic MRI. J Med Imaging (Bellingham). (2018) 5:014008. doi: 10.1117/1.JMI.5.1.014008

PubMed Abstract | CrossRef Full Text | Google Scholar

27.Khene, ZE, Richard, C, Hascoet, J, Gasmi, A, Goujon, A, Brochard, C, et al. Contrast-enhanced CT texture parameters as predictive markers of high-risk urodynamic features in adult patients with spina bifida. Urology. (2019) 134:84–9. doi: 10.1016/j.urology.2019.09.023

PubMed Abstract | CrossRef Full Text | Google Scholar

28.Wang, HS, Cahill, D, Panagides, J, Nelson, CP, Wu, HT, and Estrada, C. Pattern recognition algorithm to identify detrusor overactivity on urodynamics. Neurourol Urodyn. (2021) 40:428–34. doi: 10.1002/nau.24578

PubMed Abstract | CrossRef Full Text | Google Scholar

29.Alagapan, S, Shin, HW, Fröhlich, F, and Wu, HT. Diffusion geometry approach to efficiently remove electrical stimulation artifacts in intracranial electroencephalography. J Neural Eng. (2019) 16:036010. doi: 10.1088/1741-2552/aaf2ba

PubMed Abstract | CrossRef Full Text | Google Scholar

30.Roweis, ST, and Saul, LK. Nonlinear dimensionality reduction by locally linear embedding. Science. (2000) 290:2323–6. doi: 10.1126/science.290.5500.2323

CrossRef Full Text | Google Scholar

31.Cullingsworth, ZE, Kelly, BB, Deebel, NA, Colhoun, AF, Nagle, AS, Klausner, AP, et al. Automated quantification of low amplitude rhythmic contractions (LARC) during real-world urodynamics identifies a potential detrusor overactivity subgroup. PLoS One. (2018) 13:e0201594. doi: 10.1371/journal.pone.0201594

PubMed Abstract | CrossRef Full Text | Google Scholar

32.Colhoun, AF, Speich, JE, Cooley, LF, Bell, ED III, Barbee, RW, Guruli, G, et al. Low amplitude rhythmic contraction frequency in human detrusor strips correlates with phasic intravesical pressure waves. World J Urol. (2017) 35:1255–60. doi: 10.1007/s00345-016-1994-0

PubMed Abstract | CrossRef Full Text | Google Scholar

33.Niederhauser, T, Gafner, ES, Cantieni, T, Grämiger, M, Haeberlin, A, Obrist, D, et al. Detection and quantification of overactive bladder activity in patients: can we make it better and automatic? Neurourol Urodyn. (2018) 37:823–31. doi: 10.1002/nau.23357

PubMed Abstract | CrossRef Full Text | Google Scholar

34.Whangbo, TK, Eun, SJ, Jung, EY, Park, DK, Kim, SJ, Kim, CH, et al. Personalized urination activity recognition based on a recurrent neural network using smart band. Int Neurourol J. (2018) 22:S91–S100. doi: 10.5213/inj.1836168.084

PubMed Abstract | CrossRef Full Text | Google Scholar

35.Sheyn, D, Ju, M, Zhang, S, Anyaeche, C, Hijaz, A, Mangel, J, et al. Development and validation of a machine learning algorithm for predicting response to anticholinergic medications for overactive bladder syndrome. Obstet Gynecol. (2019) 134:946–57. doi: 10.1097/AOG.0000000000003517

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: artificial intelligence, deep learning, machine learning, urogynecology, urinary incontinence, pelvic organ prolapse

Citation: Seval MM and Varlı B Current developments in artificial intelligence from obstetrics and gynecology to urogynecology. Front. Med. 10:1098205. doi: 10.3389/fmed.2023.1098205

Received: 14 November 2022; Accepted: 09 February 2023;
Published: 23 February 2023.

Edited by:

Omneya Attallah, Arab Academy for Science, Technology and Maritime Transport (AASTMT), Egypt

Reviewed by:

Milap Shah, Max Super Speciality Hospital, India
İsmail Güler, Gazi University, Türkiye

Copyright © 2023 Seval and Varlı This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Mehmet Murat Seval, ✉ seval@ankara.edu.tr

Download