Trustworthy and ethical AI-enabled cardiovascular care: a rapid review

Background Artificial intelligence (AI) is increasingly used for prevention, diagnosis, monitoring, and treatment of cardiovascular diseases. Despite the potential for AI to improve care, ethical concerns and mistrust in AI-enabled healthcare exist among the public and medical community. Given the rapid and transformative recent growth of AI in cardiovascular care, to inform practice guidelines and regulatory policies that facilitate ethical and trustworthy use of AI in medicine, we conducted a literature review to identify key ethical and trust barriers and facilitators from patients’ and healthcare providers’ perspectives when using AI in cardiovascular care. Methods In this rapid literature review, we searched six bibliographic databases to identify publications discussing transparency, trust, or ethical concerns (outcomes of interest) associated with AI-based medical devices (interventions of interest) in the context of cardiovascular care from patients’, caregivers’, or healthcare providers’ perspectives. The search was completed on May 24, 2022 and was not limited by date or study design. Results After reviewing 7,925 papers from six databases and 3,603 papers identified through citation chasing, 145 articles were included. Key ethical concerns included privacy, security, or confidentiality issues (n = 59, 40.7%); risk of healthcare inequity or disparity (n = 36, 24.8%); risk of patient harm (n = 24, 16.6%); accountability and responsibility concerns (n = 19, 13.1%); problematic informed consent and potential loss of patient autonomy (n = 17, 11.7%); and issues related to data ownership (n = 11, 7.6%). Major trust barriers included data privacy and security concerns, potential risk of patient harm, perceived lack of transparency about AI-enabled medical devices, concerns about AI replacing human aspects of care, concerns about prioritizing profits over patients’ interests, and lack of robust evidence related to the accuracy and limitations of AI-based medical devices. Ethical and trust facilitators included ensuring data privacy and data validation, conducting clinical trials in diverse cohorts, providing appropriate training and resources to patients and healthcare providers and improving their engagement in different phases of AI implementation, and establishing further regulatory oversights. Conclusion This review revealed key ethical concerns and barriers and facilitators of trust in AI-enabled medical devices from patients’ and healthcare providers’ perspectives. Successful integration of AI into cardiovascular care necessitates implementation of mitigation strategies. These strategies should focus on enhanced regulatory oversight on the use of patient data and promoting transparency around the use of AI in patient care. Supplementary Information The online version contains supplementary material available at 10.1186/s12911-024-02653-6.


Background
Artificial intelligence (AI) is increasingly used in healthcare to improve the prevention, diagnosis, treatment, and maintenance of health conditions [1].These interventions have enormous potential to assist in the management of cardiovascular diseases, the leading cause of death in the US, given the high number of AIbased devices authorized for use and under review by the FDA for cardiovascular diseases, the breadth of use cases spanning clinical practice to consumer-facing AIenabled solutions, and the potential for improving clinical outcomes [2][3][4][5].
Previous studies have shown that patients may be willing to accept the use of AI in healthcare and see its potential benefits if certain conditions are met, including transparency about the capture and use of their data by AI systems and the ability to opt out from data sharing at any time [6].Moreover, patients place a higher level of trust in a healthcare provider's assessment of their health compared to an AI and often want assurance that their physicians are involved in and ultimately are responsible for AI-enabled decisions due to the concerns about risks of AI failures during care [7,8].On a similar note, healthcare providers express specific needs for information transparency, such as explanations about known strengths and limitations of interventions when using AIbased software in clinical decision-making [9].Healthcare providers also recognize the potential impact of AI on patient-clinician trust and seek support for transparent and effective communication with patients about AI use in their care [10].Thus, to fully achieve the appropriate uptake of AI in medicine, patients' and healthcare providers' ethical and trust concerns must be addressed [11].
Although prior research has begun to explore patient and clinician perspectives on the use of AI in medicine, none have focused explicitly on stakeholders' transparency, trust, and ethical concerns; nor have studies focused explicitly on cardiovascular care, an area where there has been rapid and transformative recent growth [12].Accordingly, there remains a significant gap in understanding the specific barriers and facilitators to addressing these stakeholder concerns related to transparency, trust, and ethics when implementing AI in cardiovascular care.This gap could hinder the development of effective practice guidelines and regulatory policies necessary for ensuring the ethical and trustworthy use of AI in medicine.To bridge this gap and to provide actionable insights into the nuanced requirements for trusted use of these AI-based technologies, this study reviewed the literature to identify key ethical concerns, potential mitigation strategies, and barriers and facilitators to trustworthy AI-informed cardiovascular care.

Inclusion and exclusion criteria
We conducted a rapid review of the literature, a form of information synthesis aiming to generate evidence through a resource-efficient approach by simplifying or removing certain components of the traditional systematic review process [13].Eligible for inclusion were publications discussing transparency, trust, or ethical concerns (outcomes of interest) associated with AI-based medical devices (interventions of interest) in the context of cardiovascular care from patients' , caregivers' , or healthcare providers' perspectives.Our search was not limited by date or study design.All papers published as full manuscripts, including qualitative and quantitative analyses, commentaries, editorials, expert opinions, perspective pieces, and guidelines were included.Conference abstracts, book chapters, pre-prints, animal studies, and publications that were not in English were excluded.Prior to the formal article screening process, we conducted a calibration exercise by piloting the screening of 10% of the sample.This ensured that all authors involved in the screening process consistently applied the inclusion and exclusion criteria.

Search strategy and data sources
A medical librarian with literature review expertise (AAG) developed the search strategy with input from all authors.The search was developed as an Ovid Embase search strategy, which was subsequently reviewed by a second librarian not otherwise associated with the project using Peer Review of Electronic Search Strategies (PRESS) [14].After the strategy had been finalized and unanimously approved by all authors, it was adapted to the syntax and subject headings of other databases.Details on the search strategy can be found in Appendix 1.The search was conducted on the following six bibliographic databases: Cochrane Library, Embase, Google Scholar, Ovid Medline, Scopus, and Web of Science Core Collection, and was completed on May 24, 2022.

Study selection
Search results were downloaded to EndNote 20 (Clarivate, Philadelphia, PA), and duplicate citations were removed using the Yale Deduplicator Tool [15].Individual citations were ingested into Covidence, a software tool dedicated to literature review management that facilitates collaboration between independent reviewers in the article screening and review processes.The review process was divided into two major steps: title/abstract screening and full-text screening.Titles and abstracts of each paper identified by the search were independently screened by two authors [MM and AMS, AAG, or DWY] against the inclusion criteria.Next, full-text articles were obtained for all studies that had not been excluded at the first level of screening and were assessed by two independent reviewers [MM and AMS or DWY], with the reasoning for exclusions being recorded.Disagreements on eligibility were resolved by consensus or through the input of a third investigator.After screening, CitationChaser was used to perform citation chasing on all included studies to identify other potentially relevant studies [16].One reviewer [MM, AMS, or DWY] screened the identified papers to decide whether they met the eligibility criteria.Reviewers were not blinded to the journal titles, authors, or institutions.

Data extraction and synthesis
Using the Qualtrics software [17], data extraction was conducted by an author [MM, AMS, or DWY] for the following fields for each included paper: article type; article title; publication year; first author; purpose and indication(s) of AI-based medical device; and device users (patients, caregivers, and healthcare providers).Next, the conceptualization and characteristics used to describe barriers and facilitators of transparency and trust and ethical concerns from patients' , caregivers' , and healthcare providers' perspectives were recorded.For validation, a second reviewer independently performed data extraction on 20% of the final sample, selected at random [MM, AMS, or DWY].Disagreements were less than 5% and were resolved by discussion or through the input of a third investigator [JEM].Data generated from this project will be actively preserved for three years per Yale Research Data and Materials Policy-Retention 6001.2 unless otherwise required by the journal.Content analyses were performed by MM, using Qualtrics 2022 and Microsoft Excel 2018 (Microsoft Corp) to facilitate data management and organization.In keeping with content analyses methods, abstracted data were independently categorized by two researchers [JEM and MM] who then met to discuss and agree upon the final categorization of findings, through iterative discussion with 100% agreement.Categories where then summarized into key themes pertaining to concerns and mitigation strategies for ethics and barriers and facilitators for trust in AI-enabled care, with unanimous agreement among all researchers.

Search results
The search resulted in 10,171 papers, of which 7,925 were unique.After conducting the first level of screening, 7,799 titles and abstracts were excluded, leaving 126 full-text articles for review.Of those, 71 did not meet eligibility criteria due to ineligible area of care, i.e., non-cardiovascular (n = 10); ineligible intervention, i.e., non-AI tools (n = 26); ineligible outcome (n = 22); ineligible format, i.e., conference abstracts, book chapters, or preprints (n = 13), leaving a total of 55 eligible publications.Citation chasing of these articles resulted in 3,603 additional citations, 3,330 of which were eliminated upon title and abstract reviewing.Of the 273 reviewed full-texts, 90 articles were found to be eligible.The reasons for excluding the remaining papers included: ineligible area of care (n = 69), intervention (n = 14), outcome (n = 88), and format (n = 12).Overall, 145 papers were included in this review (Fig. 1).Since we reached information saturation upon reviewing the additional papers identified through citation chasing, we stopped subsequent rounds of citation chasing.
The AI-based interventions discussed in 43 (29.7%)papers were devices used for the diagnosis or monitoring of cardiovascular diseases (e.g., AI-enabled cardiac imaging), while 5 (3.4%) were therapeutic devices (e.g., clinical decision support tools for heart pump implants).The interventions discussed in the remaining papers (101 [69.7%]) included both diagnostic and therapeutic AI-based medical devices.The indications for use of the AI-based devices were not specified in most papers (122 [84.1%]).Among those that specified, arrhythmia was the highest reported indication (8 [5.5%]), followed by heart failure (7 [4.8%]).Although all papers discussed AI-based devices in the cardiovascular context, 88 (60.7%) were specific to the cardiovascular specialty, while the remaining articles also included other areas of medicine.
Among all the reviewed articles, 3 (2.1%)studied devices that were self-management software used directly by patients [19][20][21], whereas the main users of the other devices discussed by 48 (33.1%) papers were healthcare providers.The remaining 94 (64.8%) papers did not specify the users.Only 2 (1.4%) papers specified the device sponsor; both studied HeartMan, a personal decision support system for heart failure management, funded by the Horizon 2020 Framework Programme of the European Union [19,20].

Ethical concerns and mitigation strategies
There were six key ethical concerns discussed in the literature, which were privacy, security, or confidentiality issues; risk of healthcare inequity or disparity; risk of patient harm; accountability and responsibility concerns; problematic informed consent and potential loss of patient autonomy; and issues related to data ownership (Fig. 2).Three papers discussed the lack of human involvement in patient care and the altered relationship between patients and healthcare providers as an ethical concern associated with AI-enabled medical care [22][23][24].One paper debated the additional complexity that AI-based medical devices could add to endof-life care [25].

Privacy, security, and confidentiality concerns
Fifty-nine (40.7%) publications discussed ethical concerns related to privacy, security, or confidentiality.Specific concerns included potential inappropriate access to and misuse of personal information stored in medical devices and inadvertent release of private patient healthcare data [22,26].Protecting sensitive patient information from data leakage and cyberattacks, especially for data used by private for-profit organizations [27], and protecting the stored medical data, particularly by cloud-assisted AI medical devices or commercial smartphone-based applications with poorly secured servers, were other areas of concern [28,29].Moreover, transferring data between institutions for the reproducibility of results could cause additional security problems [30].Lastly, ensuring confidentiality could be difficult owing to the circulation of sensitive patient information among unregulated companies and a lack of de-identification of raw data input for AI algorithms [30,31].

Mitigation strategies
We identified mitigation strategies from the literature to address some of the aforementioned ethical concerns.Data de-identification or anonymization and using highly secure data platforms could protect patient data used for the development and training of AI-medical devices [31][32][33].Additionally, more secure health systems across different localities need to be built, and policymakers could help with constructing the adapted infrastructures and developing guidelines regarding patient privacy, data storage, and data sharing to ensure optimal implementation of AI tools in healthcare [34][35][36].Several papers emphasized the need for more regulation and legislation on patient data use, such as performing regular privacy audits, mandating security breach notifications, and setting greater penalties for data misuse [27,33,[37][38][39].

Risk of healthcare inequity or disparity
Thirty-six (24.8%) papers raised concerns that AI-based medical devices could create new or exacerbate healthcare inequities or disparities based on factors such as gender, race, ethnicity, or pathology-driven specificities.Potential unfairness in algorithmically automated decisions was described as the major cause of inequities and disparities.Papers discussed the risk of the AI intervention being less effective or providing inaccurate recommendations for under-represented patients if the training datasets for algorithms are based on unrepresentative patient samples [37,40].This in turn could lead to discrimination against certain patient populations and increase the gap in healthcare outcomes among different social groups.Furthermore, some were concerned that data could be used to improperly profile patients and differentially provide healthcare (e.g., avoidance of highest-cost or highest-risk patients) [26].There were also concerns regarding social justice and potential unfairness in the distribution of the benefits and burdens of AI applications [22].
Mitigation strategies Several papers described important considerations for the data sources used by AI tools to help healthcare providers recognize when it could be inappropriate to use a specific AI tool for certain patient groups and to ensure that access to AI-based tools is not affected by demographic, geographic, or temporal constraints [41][42][43].Strategies to mitigate concerns related to health inequity when using AI in medical care include using a balanced dataset through collecting sufficient data from under-represented populations, validating AI algorithms on different minority and low-income groups, and obtaining robust input from different stakeholders involved in the development, use, and regulation of AI tools [44][45][46].Moreover, creating a distinct algorithm in AI systems for each group of patients, rather than using a universal algorithm for all patients, could improve fairness in decision-making [47].Lastly, conducting evidence-based assessment and implementing further regulatory oversights could help to ensure the fairness of AI tools [28,45].

Risk of patient harm
Concerns about the risk of suboptimal care or patient harm associated with AI tools were raised by 24 (16.6%)papers.Inaccurate data used by AI-based decision tools, flawed AI algorithms, and deliberate hacking of algorithms were discussed as potentially leading to erroneous recommendations and patient harm on a massive scale [33,48].The risk of errors would be greater when the AI systems function independently with unchecked decision-making and actions [49], particularly in the setting where errors made by complex and untransparent AI systems are difficult to trace and debug [50].Moreover, the complexity of AI-based systems, potentially unpredictable system output, and the uncertainty of human-AI interactions could result in substantial variation in the performance of AI-based medical devices, causing further safety challenges [51].Lastly, there were concerns about AI-based devices programmed to function in unethical ways, for example by suggesting clinical actions that generate higher profits without patient care benefits [31].
Mitigation strategies Several papers described the importance of providing sufficient training to device users to reduce the risk of patient harm, with an emphasis on educating healthcare providers about the potential pitfalls and limitations of AI technologies [48,52].Additionally, rigorous validation and continuous assessment of the algorithms used in AI-based medical devices, including conducting clinical trials that compare AI-supported care with the standard of care, could identify potential bias in AI algorithms and minimize patient harm [50,[53][54][55].Establishing further regulatory and ethical guidelines in the postmarket stage and implementing standard frameworks for regular assessment of the safety of AI tools are also necessary [33,46].

Problematic informed consent and loss of patient autonomy
We found 17 papers (11.7%) discussing ethical concerns about obtaining informed consent for providing care with AI-enabled medical devices.The main reason leading to problematic informed consent is the lack of transparency and interpretability of AI tools and insufficient information about different aspects of care provided by AI-enabled medical devices [45,56,57].Moreover, informing patients about all aspects of health data collection and its use across different platforms and for training algorithms may not be always feasible [36,58].Withdrawing consent for the use of these data would cause further challenges [59].Eight papers (5.5%) argued that patient autonomy could be negatively affected when using AI-enabled care.This issue specifically is likely to happen if the devices function independently and have unchecked actions [49], which could damage patients' confidence in their ability to change their medical decisions, i.e. refuse care, if later desired [50].
Mitigation strategies To improve informed decisionmaking, several papers described the necessity of providing patients and healthcare providers with sufficient information and ensuring that patients are freely able to change their medical decisions if desired [50,60].Moreover, further regulations on obtaining valid unambiguous consent when using patient data should be established [27].

Accountability and responsibility concerns
Another key ethical concern raised by 19 (13.1%) papers was the issue related to accountability and responsibility.Since multiple groups of professionals are involved in the design, manufacture, and use of AI-based medical devices, accountability and liability of the decisions made by these devices could be difficult to determine.While some suggested that users of the devices should ultimately be responsible for the output of algorithms [25,61], there are considerable debates around the accountability of actions suggested or performed by AI-based technologies and the potential misuse of data [36,37].
The complexity, opaqueness, and lack of transparency of AI-based medical devices make the accountability and responsibility issues even more challenging [50,62].
Mitigation strategies To address questions of accountability, several papers described the importance of improving the engagement of all stakeholders, including physicians and developers.Papers also suggested improving the transparency of AI tools' function so that the reasons behind decisions and actions taken by the devices are clear [63,64].Moreover, there is a need for regulatory and legal systems to oversee the implementation of AIbased medical devices and determine the responsibilities of patients, healthcare providers, and others [65].

Data ownership issues
There were further ethical concerns discussed by 11 (7.6%) papers related to ownership of the patient data being used by AI-based technologies, particularly if the data is identifiable [66].The rules and regulations related to data ownership vary significantly across different regions and may be absent in some jurisdictions, which makes it unclear whether patients, hospitals, or private companies own the data analyzed by AI tools [67,68].This issue is directly associated with how AI and its data are monetized [68], as there are controversies about who should profit from the collected data and for how long these institutions or individuals can and should retain patient health information [69].
Mitigation strategies To address these concerns, several papers described the importance of clear regulations around data ownership and preparing models of health data ownership with rights to the individual ahead of using AI-based devices in healthcare [33,38].

Trust barriers and facilitators
We identified 53 (36.6%) and 58 (40.0%) papers discussing trust barriers and facilitators, respectively, from patients' and healthcare providers' perspectives when using AI-based medical devices in cardiovascular care (Fig. 3).

Shared (Patient and Healthcare Provider) Perspective Data privacy and security issues
Data privacy and security concerns were discussed as key trust barriers for patients and healthcare providers [17,62].In particular, patients were described as worried about the potential alteration of data, unauthorized use of data, information sharing with commercial partners, and data loss [59,70].These issues are specifically concerning in the absence of uniform federal privacy regulations regarding collecting, storing, and using patient health information in different settings [41].
Facilitators To address data privacy and security concerns, the literature discussed encrypting patient data according to the Health Insurance Portability and Accountability Act of 1996 (HIPAA), removing data identifiers, documenting the purpose of datasets, establishing ethical standards for data use and access, and securing communications between patients and healthcare providers [41,71,72].Regulatory bodies could ensure the competence of AI systems and their users and establish standardized codes of ethics and conduct for device developers [72].

Risk of suboptimal care or patient harm
Users have expressed concerns around the possibility of device malfunction and are hesitant about the trustworthiness of diagnostic decisions or automatically generated medical advice by AI tools, especially if the advice contradicts their previous experiences [50].Another important trust barrier is the uncertainty about the reliability and quality of the data used in the algorithms, which could be incomplete, unrepresentative, or outdated [73].This lack of generalizability could exacerbate health inequities, and further decrease trust in the populations who feel that AI would be inaccurate when applied to their cases [74].Certain populations may also feel that they may not equally benefit from AI technologies because of the deployment and marketing strategies that manufacturers might take [74].Healthcare providers are also concerned that AI-based medical devices could provide inaccurate or biased recommendations, especially if the systems are not regularly updated [75,76].Moreover, clinicians may not trust the generalizability of the outputs of AI systems for their own patients due to the lack of diversity in the clinical dataset [77][78][79].
Facilitators To address these trust barriers, the literature discussed the importance of keeping AI systems updated by introducing new rules and cases along with routine performance assessments to enhance the accuracy of decisions made by AI-based medical devices [75,80].Further regulations and legislation could also increase trust by ensuring the balance between innovation and patient safety and confirming that AI algorithms meet appropriate standards of clinical benefit [81,82].

Lack of transparency and insufficient knowledge
Substantial barriers to trust in AI-enabled medical devices are the lack of transparency, opaqueness (black box nature), and poor interpretability of the devices [76,83,84].Physicians tend to trust a device less if they do not fully understand how it functions or how its outputs are generated, even if the device performs well [37,40,54].Multiple barriers to transparent AI-based medical devices exist, including the lack of understanding of what information is being used by the AI tools, what the AI systems are learning, and how the AI algorithms reach conclusions based on the inputs [30,[85][86][87].Also, it could be difficult to achieve algorithmic transparency due to the complicated structure, dynamic learning, and constant evolution of AI algorithms [36,56].These factors make AI models difficult to explain and justify, and therefore, uninterpretable [88].Besides, inadequate education and experience with AI tools can cause additional barriers to trustworthy AI-enabled care [76,89].
Facilitators To improve explainability and physicians' understanding of AI-based medical devices, it is essential to clarify AI algorithm training data, explain the computational model and its output, and acknowledge the existing limitations of AI-based medical devices [76,78,87,90,91].Making the datasets, codes, and trained models publicly available and using interpretable models that will allow healthcare providers to review and provide feedback to the AI decision-making tools could further improve transparency [47,92].Some argued that healthcare providers may not need detailed explanations of the validated predictions and decisions made by AI-enabled medical devices but need to have sufficient information about the major components that affect the decisions [43].Additionally, a visual display of the consensus between decision support tools and clinicians' assessments could enhance clinicians' trust in AI systems [55].
Restricting the complexity of AI tools as well as providing clarity on how AI devices are regulated could facilitate patient trust [19,21,59,93].It is also essential to provide patients with appropriate education about how to use AI tools and enhance their engagement in different phases of the design and implementation of AI technologies [50,89,94,95].
Other important factors for facilitating transparency are to clarify all the interactions within and among different sectors that led to the development of AI systems and to maintain open and clear communication between healthcare providers and developers [88,96].Regulatory bodies could establish more rigorous regulations for the enforcement of transparency in datasets and algorithms used in AI-based medical devices [47,92].

Replacing human aspects of care
Patients and healthcare providers seem to trust AI tools less if the devices are meant to entirely replace the human aspect of care [53].
Facilitators Trust could improve if patients and healthcare providers are assured that AI-based devices are supplementary to care, rather than outright replacing clinicians or other human aspects of care [53,92].

Patient perspective Prioritizing profits over patients' interests
From the patient perspective, trust would be diminished when they feel AI devices are mainly used for economic efficiency at the cost of patient interests and benefits [72].
No facilitators were identified in the reviewed literature for this trust barrier.

Healthcare provider perspective Lack of robust evidence
A significant barrier to clinician trust is the lack of robust evidence for the accuracy and limitations of AI-based medical devices in addition to the inadequate education and training about the use of AI tools [76,97,98].
Facilitators Several papers argued that while it might not be feasible to explain all aspects of AI, generating more reliable evidence and standards through rigorous internal and external validations, prospective clinical trials in diverse cohorts which demonstrate safety, efficacy, and generalizability of AI devices, and peer-reviewed publications can improve trust [99][100][101][102][103].Therefore, collaborative practices with healthcare providers for the development and continuous assessment of AI devices are essential [75,98].Lastly, complying with the established legislations and regulations is essential when producing trustworthy AI research [88].

Discussion
In this rapid review of the literature on the use of AIbased interventions in cardiovascular care, which included more than 11,000 publications, we identified key stakeholder concerns among healthcare providers and patients that relate to transparency, trust, and ethical concerns around the use of AI in cardiovascular care.Concerns focused on data privacy and security, risk of patient harm, and the possibility that AI-based medical care could exacerbate healthcare inequities or advance unfair algorithmically automated decisions.Inadequately obtaining informed consent from patients regarding the use of AI and various forms of data collection while providing AI-enabled care was also described, as was determining who is ultimately responsible for regulating the development, performance, and use of AI in medicine and who owns the collected data.The absence of rigorous clinical trials to support the safety and efficacy of AI-enabled medical devices and the lack of transparency about the data used by AI devices and their subsequent recommendations remain other significant barriers to patients' and healthcare providers' trust.Given the rapid and transformative recent growth of AI in cardiovascular care [12], these challenges should be carefully identified and addressed to ensure that AI systems are developed and implemented in an ethical and trustworthy manner.
We identified mitigation strategies to address most key ethical and trust concerns about the use of AI in medicine, which requires a collaborative effort involving AI developers, regulators, hospital systems, healthcare providers, and patients.Regulatory agencies were identified as having multiple inroads to addressing patient and clinician concerns.Notably, we found that establishing further regulations and legislation around development, adoption, and use of AI in healthcare is a key facilitator for addressing almost all the identified ethics concerns and trust barriers.Certain proposed frameworks and guidance documents have carved out actions for oversight bodies to delineate the scope of liability, strengthen data privacy protections, and clarify data ownership regulations [104,105].Moreover, requiring postapproval studies could ensure continuous monitoring of AI devices' performance, potential biases, and unintended consequences.
AI developers similarly have a significant stake in addressing patient and clinician concerns and need to be attentive to data stewardship practices, safety, and transparency as models are researched, developed, and marketed.Moreover, current medical device labeling does not always address the unique challenges of the use of AI-based software, such as training data sources, model accuracy, potential biases, and opting out of use, which can hinder patient-shared decision-making and trust in AI-enabled care.Providing AI model facts labels will establish a clear and standardized communication of information with users and enhance transparency and trust [52].Furthermore, self-governance approaches may serve as a potential mechanism in tandem with regulatory intervention for implementing mitigation strategies.Submitting to a set of industry standards as well as certification processes may help to mitigate the risks of AI tools and help to facilitate trust in models [106].
Hospital systems and clinicians will also be faced with key decisions regarding AI tools adopted in their practices.As hospitals become a source of data for the development of numerous models, appropriate privacy protections and transparency about data use and model deployment would be relevant, especially as they act in coordination with third-party developers [107].As endusers of most healthcare AI tools, clinicians may become responsible for providing appropriate information about these systems to patients at the point of care and for appropriately integrating model insights into clinical decision-making.
While our findings are indicative of many strategies that would be taken up by clinical, technical, and regulatory stakeholders, there are also opportunities for including patients.Stakeholder engagement with patient populations and the public in the research and design of AI tools may be relevant to mitigating bias and developing trust, particularly by communicating the underlying design of AI tools in ways that are understandable to patients and leveraging advisory groups to inform the creation of such tools [108].Identifying opportunities for patient engagement will be incumbent upon all stakeholders with more formal decision-making authority.Thus, regulatory oversight on using and sharing patient information, safety and transparency of AI tools, and responsibilities of healthcare providers, device manufacturers, and patients would facilitate the application of AI in medical care.
Overall, we found that most papers briefly touched upon issues related to trust and ethics and potential mitigation strategies without providing in-depth information.Additional studies translating ethical principles into tangible tools and guidance for stakeholders will be an important next step in implementation of responsible and trustworthy AI-enabled healthcare [109].Moreover, we did not find any ethical concerns or trust barriers and facilitators from the caregivers' perspective, necessitating further research in this area.
Our study has limitations.First, similar to all reviews of published literature, publication and reporting biases may have affected our findings.Second, while we identified and reviewed a significant number of relevant papers, the vast majority were review articles and commentaries, editorials, or perspective pieces with fewer original research articles.While our search was very exhaustive, there was an inconsistency in the level of detail, which may have led to papers potentially being missed.However, citation chasing was undertaken to identify additional relevant articles that failed to include the three main concepts of our search.Lastly, this study focused on the use of AI in cardiovascular care and may not generalize to uses in other areas of medicine.

Conclusion
This rapid review of the literature on the use of AIbased interventions in cardiovascular care identified key ethical and trust concerns from patients' and healthcare providers' perspectives, including issues related to data privacy and security, potential inequity and bias, risk of patient harm, patient consent and autonomy, and a lack of transparency about the function of AI-based medical devices.Given the rapid and transformative recent growth of AI in cardiovascular care [12], certain mitigation strategies, particularly establishing further regulatory oversight on the use of patient data, and safety and transparency of AI tools seem necessary.

Fig. 1 Fig. 2
Fig. 1 Sample Construction Using the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) Diagram

Fig. 3
Fig. 3 Trust Barriers and Facilitators for the Use of Artificial intelligence-based Medical Devices in Cardiovascular Care