Exploring stakeholder attitudes towards AI in clinical practice
•,,.
...
Abstract
Objectives Different stakeholders may hold varying attitudes towards artificial intelligence (AI) applications in healthcare, which may constrain their acceptance if AI developers fail to take them into account. We set out to ascertain evidence of the attitudes of clinicians, consumers, managers, researchers, regulators and industry towards AI applications in healthcare.
Methods We undertook an exploratory analysis of articles whose titles or abstracts contained the terms ‘artificial intelligence’ or ‘AI’ and ‘medical’ or ‘healthcare’ and ‘attitudes’, ‘perceptions’, ‘opinions’, ‘views’, ‘expectations’. Using a snowballing strategy, we searched PubMed and Google Scholar for articles published 1 January 2010 through 31 May 2021. We selected articles relating to non-robotic clinician-facing AI applications used to support healthcare-related tasks or decision-making.
Results Across 27 studies, attitudes towards AI applications in healthcare, in general, were positive, more so for those with direct experience of AI, but provided certain safeguards were met. AI applications which automated data interpretation and synthesis were regarded more favourably by clinicians and consumers than those that directly influenced clinical decisions or potentially impacted clinician–patient relationships. Privacy breaches and personal liability for AI-related error worried clinicians, while loss of clinician oversight and inability to fully share in decision-making worried consumers. Both clinicians and consumers wanted AI-generated advice to be trustworthy, while industry groups emphasised AI benefits and wanted more data, funding and regulatory certainty.
Discussion Certain expectations of AI applications were common to many stakeholder groups from which a set of dependencies can be defined.
Conclusion Stakeholders differ in some but not all of their attitudes towards AI. Those developing and implementing applications should consider policies and processes that bridge attitudinal disconnects between different stakeholders.
Summary
What is already known?
Very little is known about the attitudes of different stakeholders towards artificial intelligence (AI) applications in healthcare.
While the AI industry see their applications as promising for improving healthcare, the views of clinicians, patients and other groups directly involved in delivering or receiving care may not be so favourable.
What does this paper add?
This paper provides an exploratory analysis of published reports of the attitudes and perceptions of different stakeholder groups towards AI applications in healthcare.
Stakeholder groups hold similar attitudes towards AI on some attributes but differ in their attitudes towards others.
In general, attitudes towards AI in healthcare were positive, more so for those with direct experience of AI in care delivery, but with the proviso that certain safeguards were met.
Those developing and implementing AI applications should consider policies and processes that bridge attitudinal disconnects between different stakeholders.
Introduction
Artificial intelligence (AI) refers to advanced computer programs that mimic intelligent human behaviours and assist humans with different tasks. Medical AI applications span a spectrum, from diagnosis and disease screening to treatment selection and prognostication,1 and aim to optimise care, improve efficiency and enhance clinician and consumer experience. Despite scores of AI applications having received regulatory approval for use in clinical settings in recent years, and many more having passed the proof-of-concept stage, relatively few that purport to directly assist decision-making have been adopted at scale into clinical practice.2 This limited uptake may be due, at least partly, to misperceptions of what the term AI actually means and negative attitudes towards AI held by key players in the healthcare ecosystem. Multiple stakeholders share interest in the performance and outcomes of AI applications, comprising clinicians, consumers, managers, researchers, regulators and industry. Their perceptions and expectations of AI may differ, and need to be understood and considered by AI developers and implementers if AI applications are to be designed and operationalised in ways acceptable to all parties.
Methods
We undertook an exploratory analysis of articles whose titles or abstracts contained the terms ‘artificial intelligence’ or ‘AI’ and ‘medical’ or ‘healthcare’ and ‘attitudes’, ‘perceptions’, ‘opinions’, ‘views’, ‘expectations’. Using a snowballing strategy, we searched PubMed and Google Scholar for articles published 1 January 2010 through 31 May 2021. Reference lists of retrieved articles were perused for additional studies. We excluded articles that did not employ a formal survey or interview tool and/or did not report quantified response measures for individual questions among respondents. We only selected articles dealing with non-robotic AI applications used to support clinician-mediated care-related tasks or decision-making, and excluded mobile or wearable applications that were exclusively consumer facing. Key findings were extracted and summarised in narrative form according to four categories of participants. We used these results to derive a thematic synthesis of stakeholder expectations and corresponding requirements (or dependencies) for developers of AI applications to consider.
Results
A total of 27 articles were included3–29 of which most (16, 59%) targeted clinicians,3–18 8 (30%) focused on consumers (including patients),19–26 1 (4%) on health executives27 and 2 (7%) on industry stakeholders comprising AI vendors, researchers and regulators.28 29 Detailed study descriptions are provided in the online supplemental appendix and summary results are listed in table 1. Most studies (23; 85%) used online surveys,3–20 22–24 27 28 of which only three (11%)15 17 24 were designed using the Checklist for Reporting Results of Internet E-Surveys.30 Three (11%) studies undertook face to face interviews,25,26,29 and one used a paper-based questionnaire.21 A specific definition or example of AI was provided to participants in only 10 (37%) studies,3 8 17 19 22–27 with generic descriptors (eg, ‘computers’ or ‘machines’) used in 6 (22%)5 13 14 16 28 29 and none in 11 (41%).4 6 7 9–12 15 18 20 21 Survey response rates were reported in 11 (41%) studies,5 6 9 12 13 15 17 18 21 23 28 ranging from <0.1% to 66%, with 6 (22%)7 8 10 11 14 16 reporting no response rates and the remainder using convenience samples19 20 22–27 29 of which one calculated a required sample size.19
Table 1
|
Stakeholder perceptions of clinical AI applications
Clinicians
Clinicians practising in imaging-based disciplines, where deep machine learning is most advanced, featured in several surveys. In an Australian survey of 632 specialists (ophthalmology (n=305), radiology/radiation oncology (n=230), dermatology (n=97)),3 most had never actually used any AI application in practice (81%), but predicted AI would improve their field (71%) and impact future workforce needs (86%). Most considered AI had to perform better than specialists for disease screening (64%) and diagnosis (80%). The top three perceived AI benefits were improved patient access to screening, greater diagnostic confidence and reduced specialist time spent on mundane tasks. The top three concerns were outsourcing application development to large commercial AI companies, clinician liability due to AI errors and decreased reliance on specialists (‘do-it-yourself’ medicine). Most respondents (86%) felt their professional colleges were ill prepared for introducing AI into practice, citing need for training curricula, guidelines and working groups with AI expertise.
Radiologist attitudes towards AI were mostly positive. Most surveyed Italian radiologists (n=1032) favoured adopting AI (77%), did not fear job loss due to AI (89%) and anticipated fewer diagnostic errors (73%) and optimised workflows (68%), although at the expense of some reputational loss and decreased demand for their services (60%).4 Among 270 French radiologists, most anticipated fewer errors (81%), reduced time spent on image interpretation (74%) and more time spent with patients (52%), with most wanting ongoing education in AI (69%).5
Trainees and medical students with an interest in radiology expressed more mixed views, with a third of 69 US radiology residents stating, with hindsight, they may have chosen a different career because of AI.6 Among 484 UK medical students, half (49%) were disinclined towards a radiology career, despite most (89%) seeing expertise in AI as benefitting them (89%) and wanting AI education included in medical degrees (78%).7 In Germany, 263 medical students thought AI will improve radiology (86%), not replace radiologists (83%), and desired further training in AI (71%).8 Canadian students (n=322) expressed similar views, but also voiced concerns about reduced radiologist demand (67%).9
Clinicians in pathology and dermatology also tended to view AI positively. Among 487 survey respondents in pathology from 59 countries, 73% expressed interest or excitement in AI as a diagnostic tool for improving workflow efficiency and quality assurance.10 Fewer than 20% feared displacement or negative career impacts, with most (73%) stating diagnostic decision-making should remain a predominantly human task or one shared equally with AI. While only 25% were concerned about AI errors, opinions about medico-legal responsibility were split, with 44% believing the AI vendor and pathologist should be held equally liable and 50% believing the pathologist should bear prime responsibility. Most (93%) pathologists supported AI if it resulted in more time being spent on academic or research efforts in answering questions previously not possible. Similarly, among 1271 dermatologists from 92 countries, 77% saw AI as improving diagnostic accuracy, particularly in regards to dermatoscopic images, and 80% thought AI should be part of medical training.11 Less than 6% saw dermatologists being replaced by AI, although 18% held non-specified fears of negative impacts. In contrast, being replaced by AI was of great concern to 27% of laboratory workers and non-clinical technicians in a survey of 1721 subjects, although most (64%) expressed support for AI projects within their organisation and 40% believed AI could reduce errors and save time in their routine work.12
Clinicians from non-imaging-based disciplines considered the potential of AI to be more limited. Among 720 UK general practitioners, most (>70%) thought human empathy and communication could not be emulated by AI, that value-based care required clinician judgement, and that benefits of AI would centre on reducing workflow inefficiencies, particularly administrative burdens.13 Similarly, most psychiatrist respondents (n=791) from 22 countries felt AI was best suited to documenting and updating medical records (75%) and synthesising information to reach a diagnosis (54%).14 Among 669 Korean doctors, most (83%) considered AI useful in analysing vast amounts of clinical data in real time, while more than a quarter (29%) thought AI would fail in dealing with uncommon scenarios owing to inadequate data.15 Respondents felt responsibility for AI-induced errors lay with doctors (49%), patients consenting to use of AI (31%) or AI companies that created the tools (19%). Most Chinese clinicians (82% of 191) were disinclined to use an AI diagnostic tool they did not trust or could not understand how it would improve care.16 Among 98 UK clinicians (including 34 doctors, 23 nurses, 30 allied health professionals), 80% expressed privacy concerns and 40% considered AI potentially dangerous (indeed as bad as nuclear weapons, although this response was primed by reference to a film in which Elon Musk expressed similar sentiments).17 However, 79% also believed AI could assist their field of work and 90% had no fear of job loss. In a survey of 250 hospital employees from four hospitals in Riyadh, Saudi Arabia (nurses=121; doctors=70; technicians=59), the majority stated AI could reduce errors (67%), speed up care processes (70%) and deliver large amounts of high-quality, clinically relevant data in real time (65%).18 However, most thought AI could replace them in their job (78%) despite AI limitations in being unable to provide opinions in every patient (66%) or in unexpected situations (64%), unable to sympathise with patients (67%) and developed by computer specialists with little clinical experience (68%).
Consumers
Consumer surveys of AI in healthcare are few and yield mixed views depending on who was surveyed and what AI functions were considered. Most clinical trials of AI tools also omit assessment of patient attitudes.31 In general, patients view AI more favourably than non-patients, but only if AI is highly trustworthy and associated with clinician oversight.
An online US survey of 50 individuals revealed dehumanisation of clinician–patient relations, low trustworthiness of AI advice and lack of regulatory oversight as significant risks which predominated over potential benefits, although privacy breaches or algorithm bias were not expressed as major concerns.19 In an online survey of 6000 adults from various countries, only 27% respondents expressed comfort with doctors using AI to influence clinical decisions.20
In a survey of 229 German patients, most (≥60%) favoured physicians over AI for history taking, diagnosis and treatment plans, but simultaneously acknowledged AI could help integrate the most recent scientific evidence into clinician decision-making.21 Most (>60%) preferred physician opinion to AI where the two disagreed, and were less accepting (≤45%) of AI use in cases of severe versus less severe disease. In a UK case-based questionnaire study involving 107 neurosurgery patients, most accepted using AI for image interpretation (66%), operative planning (76%) and real-time alert of potential complications (73%), provided the neurosurgeon was in control at all times.22 Among 1183 mostly female patients with various chronic conditions who were considering biometric monitoring devices and AI, only 20% considered benefits (such as improved access to care, better follow-up, reduced treatment burden) greatly outweighed risks and 35% would decline the use of AI-based tools in their care.23 The majority (>70%) of parents of paediatric patients (n=804) reported openness to AI-driven tools if accuracy was proven, privacy and shared decision-making were protected and care using AI was convenient, of low cost, and not in any way dehumanised.24 Among 48 US dermatology patients, most (60%) anticipated earlier diagnosis and better care access, while 94% saw the main function of AI as offering second opinions to physicians, and perceived AI as having both strengths (69% believed AI to be very accurate most of the time) and weaknesses (85% expected rare but serious misdiagnoses).25 A small study found 18 patients with meningioma wanted assurance that use of AI to allocate treatment was fair and equitable, that AI-mediated mistakes would be disclosed and reparations to patients forthcoming and that patient consent was obtained for any sharing of health data.26
Healthcare executives
In a global survey of 180 healthcare executives, 40% of respondents overall favoured increased use of AI applications, although this figure varied according to jurisdiction, with Australian executives (23%) being least in favour.27 Perceived AI benefits comprised improved cybersecurity (56%) operational efficiency (56%), analytics capacity (50%) and cost savings (43%). However, fewer respondents thought there would necessarily be improvements in patient satisfaction (13%), access to care (10%) or clinical outcomes (6%). Respondents cited success factors for AI implementation as comprising adequate staff training and expertise (73%), explicit regulator legislation (64%) and mature digital infrastructures (62%).
Industry professionals
Information technology (IT) specialists, technology and software vendors, researchers and regulators—the ‘insiders’ of AI—may harbour attitudes different to those of AI users such as clinicians, consumers and healthcare executives.
In one German survey (n=123; 42 radiologists, 55 IT specialists, 26 vendors), all three groups mostly agreed (>75%) that AI could improve efficiency of care, provided AI applications had been validated in clinical studies, were capable of being understood by clinicians and were referenced in medical education.28 However, only 25% of participants would advocate sole reliance on AI results, only 14% felt AI would render care more human and 93% required confirmation of high levels of accuracy. In interviews involving 40 French subjects (13 physicians, 7 industry representatives, 5 researchers, 7 regulators, 8 independent observers), all agreed reliable AI required access to large quantities of patient data, but such access had to be coupled with confidentiality safeguards and greater transparency in how data were gathered and processed to protect the integrity of physician–patient relationships.29 On other matters there were notable differences. Physicians highlighted many tools lacked proof of efficacy in clinical settings and they would not assume criminal liability if a tool they could not understand produced errors. Industry representatives wanted greater access to more high-quality data, while wanting to avoid injury liability as they believed this would hinder tool development. Regulators were urgently searching for robust procedures for assessing safety of constantly evolving AI tools, and resolving liability for AI error which would otherwise discourage clinicians and patients from using AI. Researchers with no commercial sponsors wanted more funding and more rapid translation of their findings into practice.
Expectations and dependencies
Our analysis identified certain stakeholder expectations of AI (table 2), with the most frequently cited being a need for accurate and trustworthy applications that improve clinical decision-making, workflow efficiencies and patient outcomes, but which do not diminish professional roles. These expectations, which varied in strength of expression across studies, reflect the dominance of clinician surveys in existing studies. The corresponding self-explanatory dependencies were extrapolated by the authors, and are aligned with those expressed in authoritative reports from the National Academy of Medicine32 and the WHO.33 According to these bodies, understanding stakeholder views is essential in formulating clinical AI policy and that AI designers should focus on education, communication and collaboration in bridging attitudinal disconnects between different stakeholders.
Table 2
|
Expectations and dependencies
Discussion
Overview of findings
The diversity in attitudes towards AI of different stakeholders and the cautionary sentiments expressed by many suggest AI applications should be seen as complex sociotechnical systems with many interacting components.34 However, stated positive or negative perceptions of AI may not consistently translate into adoption or resistance, or necessarily track what is possible or even probable in a still-developing technology. The failure of many survey studies to cite concrete examples of AI applications in the prelude to questionnaires (some justifying this as a way of avoiding the conjuring up of negative ‘Terminator’ or ‘cyborg’ images) may have caused confusion among respondents as to what they were being asked to conceptualise and respond to. Response rates were either low (<50%) or incalculable, with respondents more likely than non-respondents to hold strong attitudes. Priming effects in how AI was introduced and questions were worded may have biased some responses. Finally, responses in some studies appeared internally inconsistent in that, for example, radiology residents and students acknowledged AI would improve their discipline and wanted more AI training, but, at the same time, feared loss of professional status and held concerns about career choice.
Individuals without direct experience of AI who perceived it in the abstract tended to be more guarded in their views compared with the more optimistic views of direct users or recipients of AI. However, this optimism was more often grounded in views of workflow improvements and error minimisation, rather than perceptions of improved clinical outcomes, greater fairness of access or less risk to patient autonomy compared with current clinical practice. All stakeholders voiced concern about potential harm to patients from AI that lacks human oversight in its design, development and deployment, that the expected benefits of AI were by no means guaranteed, and that explicit regulatory standards must be formulated.
Applications which automate image interpretation and data synthesis were regarded more favourably by clinicians than those directly influencing clinical decisions or having potential to negatively impact clinician–patient relationships or clinician autonomy. Repetitive tasks using digitised data, such as radiological or dermatological diagnosis, are seen as more amenable to being performed by AI applications than interactive or procedural tasks such as consultations or surgical operations.35 Privacy breaches and inability to understand or control AI applications worried clinicians, while loss of clinician oversight and inability to properly share in decision-making worried consumers. There was a common desire to ensure humans remained at the centre of decision-making and preserve empathetic, contextualised communication in clinical encounters.36 Case studies have confirmed consumers prefer human advisers who can appreciate their unique circumstances, and see AI assisting, rather than replacing, clinician advice.37
All stakeholders wanted reassurance that AI-generated advice was trustworthy, and that this level of trust was context-dependent, with clinician opinion trumping AI advice where the two were discordant or where decisions relating to serious illness were being made. As others have also shown,38 stakeholders tend to be less forgiving towards error made by AI than error made by humans. Who should bear liability for error was much more contentious, both between and within stakeholder groups, and subject to considerable ongoing debate.39 In a very recent US survey study of 750 physicians and 1007 members of the public, the majority of both groups believed the physician should be held responsible for AI error, although more of the public held this view than did physicians (66% vs 57%; p=0.02).40 In contrast, more physicians believed the AI vendor (44% vs 33%; p=0.004) should share liability, while equal proportions of both groups conferred liability on regulatory authorities (23% vs 23%) or healthcare organisations purchasing the application (29% vs 23%).
Despite their reservations, clinicians overall were keen to receive further education in AI in recognition of its potential to increase diagnostic accuracy and workflow efficiencies, and this need is increasingly recognised.41 While some clinicians in imaging specialties were worried about potential negative impacts on job prospects and professional status, most clinicians felt AI could enhance professional satisfaction.
Perceptions and expectations
Understanding what drives stakeholder perceptions of AI is important as they critically influence predisposition towards accepting AI.42 Further in-depth research into why differing views of AI are held should assist in formulating operational solutions that accommodate such diversity of views. We note few studies considered the extent to which age, sex, clinical setting, level of expertise in computing or mathematics, personal beliefs and values, or other attributes of individuals impacted on their perceptions of AI in healthcare, which some investigators suggest as being important.43
Notwithstanding these considerations, certain expectations were inherent to many studies from which dependencies can be defined. While these dependencies are not necessarily unique to AI applications, being relevant to other computer-based technologies, the rapid evolution and potentially huge scope of AI magnifies the imperative for these dependencies to be enshrined in governance and ethics policies of government and industry.
Conclusion
A wide range of stakeholders have interest in how AI applications can be used in delivering better healthcare. In general, attitudes towards AI are positive, provided certain safeguards are met. While some concerns about AI are common to most groups, others are unique to a more select few. The challenge for AI developers and implementers is to understand these various concerns and respond appropriately if their applications are to be adopted at scale.
Contributors: IAS conceived the study design, undertook initial literature search and data analysis and wrote first draft of the manuscript. SC critically appraised the manuscript and contributed to additional text and table 2. EC critically appraised the manuscript and provided further text under Methods.
Funding: The authors have not declared a specific grant for this research from any funding agency in the public, commercial or not-for-profit sectors.
Competing interests: None declared.
Provenance and peer review: Not commissioned; internally peer reviewed.
Supplemental material: This content has been supplied by the author(s). It has not been vetted by BMJ Publishing Group Limited (BMJ) and may not have been peer-reviewed. Any opinions or recommendations discussed are solely those of the author(s) and are not endorsed by BMJ. BMJ disclaims all liability and responsibility arising from any reliance placed on the content. Where the content includes any translated material, BMJ does not warrant the accuracy and reliability of the translations (including but not limited to local regulations, clinical guidelines, terminology, drug names and drug dosages), and is not responsible for any error and/or omissions arising from translation and adaptation or otherwise.
Data availability statement
All data relevant to the study are included in the article or uploaded as supplementary information. Not applicable.
Shaw J, Rudzicz F, Jamieson T, et al. Artificial intelligence and the implementation challenge. J Med Internet Res2019; 21. doi:10.2196/13659•Google Scholar•PubMed
Scheetz J, Rothschild P, McGuinness M, et al. A survey of clinicians on the use of artificial intelligence in ophthalmology, dermatology, radiology and radiation oncology. Sci Rep2021; 11. doi:10.1038/s41598-021-84698-5•Google Scholar•PubMed
Coppola F, Faggioni L, Regge D, et al. Artificial intelligence: radiologists' expectations and opinions gleaned from a nationwide online survey. Radiol Med2021; 126:63–71. doi:10.1007/s11547-020-01205-y•Google Scholar•PubMed
Waymel Q, Badr S, Demondion X, et al. Impact of the rise of artificial intelligence in radiology: what do radiologists think? Diagn Interv Imaging2019; 100:327–36. doi:10.1016/j.diii.2019.03.015•Google Scholar•PubMed
Collado-Mesa F, Alvarez E, Arheart K, et al. The role of artificial intelligence in diagnostic radiology: a survey at a single radiology residency training program. J Am Coll Radiol2018; 15:1753–7. doi:10.1016/j.jacr.2017.12.021•Google Scholar•PubMed
Sit C, Srinivasan R, Amlani A, et al. Attitudes and perceptions of UK medical students towards artificial intelligence and radiology: a multicentre survey. Insights Imaging2020; 11. doi:10.1186/s13244-019-0830-7•Google Scholar•PubMed
Pinto Dos Santos D, Giese D, Brodehl S, et al. Medical students' attitude towards artificial intelligence: a multicentre survey. Eur Radiol2019; 29:1640–6. doi:10.1007/s00330-018-5601-1•Google Scholar•PubMed
Gong B, Nugent JP, Guest W, et al. Influence of artificial intelligence on Canadian medical students' preference for radiology specialty: ANational survey study. Acad Radiol2019; 26:566–77. doi:10.1016/j.acra.2018.10.007•Google Scholar•PubMed
Polesie S, Gillstedt M, Kittler H, et al. Attitudes towards artificial intelligence within dermatology: an international online survey. Br J Dermatol2020; 183:159–61. doi:10.1111/bjd.18875•Google Scholar•PubMed
Blease C, Kaptchuk TJ, Bernstein MH, et al. Artificial intelligence and the future of primary care: exploratory qualitative study of UK general practitioners' views. J Med Internet Res2019; 21. doi:10.2196/12802•Google Scholar•PubMed
Doraiswamy PM, Blease C, Bodner K, et al. Artificial intelligence and the future of psychiatry: insights from a global physician survey. Artif Intell Med2020; 102. doi:10.1016/j.artmed.2019.101753•Google Scholar•PubMed
Oh S, Kim JH, Choi S-W, et al. Physician confidence in artificial intelligence: an online mobile survey. J Med Internet Res2019; 21. doi:10.2196/12422•Google Scholar•PubMed
Fan W, Liu J, Zhu S, et al. Investigating the impacting factors for the healthcare professionals to adopt artificial intelligence-based medical diagnosis support system (AIMDSS). Ann Oper Res2020; 294:567–92. doi:10.1007/s10479-018-2818-y•Google Scholar
Abdullah R, Fakieh B. Health care employees' perceptions of the use of artificial intelligence applications: survey study. J Med Internet Res2020; 22. doi:10.2196/17620•Google Scholar•PubMed
Esmaeilzadeh P. Use of AI-based tools for healthcare purposes: a survey study from consumers’ perspectives. BMC Med Inform Decis Mak2020; 20. doi:10.1186/s12911-020-01191-1•Google Scholar
Pega Inc. What consumers really think about AI: a global study. Insights into the minds of consumers to help Businesses reshape their customer engagement strategies. Cambridge, MA: Pega. 2017;
Lennartz S, Dratsch T, Zopfs D, et al. Use and control of artificial intelligence in patients across the medical workflow: single-center questionnaire study of patient perspectives. J Med Internet Res2021; 23. doi:10.2196/24221•Google Scholar•PubMed
Palmisciano P, Jamjoom AAB, Taylor D, et al. Attitudes of patients and their relatives toward artificial intelligence in neurosurgery. World Neurosurg2020; 138:e627–33. doi:10.1016/j.wneu.2020.03.029•Google Scholar•PubMed
Tran V-T, Riveros C, Ravaud P, et al. Patients’ views of wearable devices and AI in healthcare: findings from the ComPaRe e-cohort. NPJ Digit Med2019; 2:1–8. doi:10.1038/s41746-019-0132-y•Google Scholar
Sisk BA, Antes AL, Burrous S, et al. Parental attitudes toward artificial intelligence-driven precision medicine technologies in pediatric healthcare. Children2020; 7. doi:10.3390/children7090145•Google Scholar•PubMed
Nelson CA, Pérez-Chada LM, Creadore A, et al. Patient perspectives on the use of artificial intelligence for skin cancer screening: a qualitative study. JAMA Dermatol2020; 156:501–12. doi:10.1001/jamadermatol.2019.5014•Google Scholar•PubMed
McCradden MD, Baba A, Saha A, et al. Ethical concerns around use of artificial intelligence in health care research from the perspective of patients with meningioma, caregivers and health care providers: a qualitative study. CMAJ Open2020; 8:E90–5. doi:10.9778/cmajo.20190151•Google Scholar•PubMed
Consulting A. Injecting intelligence into healthcare. Accenture executive survey on AI in healthcare Australia. Australia, Accenture2018; Google Scholar
Laï M-C, Brian M, Mamzer M-F, et al. Perceptions of artificial intelligence in healthcare: findings from a qualitative survey study among actors in France. J Transl Med2020; 18. doi:10.1186/s12967-019-02204-y•Google Scholar•PubMed
Eysenbach G. Improving the quality of web surveys: the checklist for reporting results of Internet E-Surveys (cherries). J Med Internet Res2004; 6. doi:10.2196/jmir.6.3.e34•Google Scholar•PubMed
Perlmutter A, Benchoufi M, Ravaud P, et al. Identification of patient perceptions that can affect the uptake of interventions using biometric monitoring devices: systematic review of randomized controlled trials. J Med Internet Res2020; 22. doi:10.2196/18986•Google Scholar•PubMed
Matheny M, Israni ST, Ahmed M, et al. Artificial intelligence in health care: the hope, the hype, the promise, the peril.
WHO. Ethics and governance of artificial intelligence for health: WHO guidance. Geneva, World Health Organisation2021; Google Scholar
Gerke S, Babic B, Evgeniou T, et al. The need for a system view to regulate artificial intelligence/machine learning-based software as medical device. NPJ Digit Med2020; 3:53. doi:10.1038/s41746-020-0262-2•Google Scholar•PubMed
Mesko B. The impact of digital health technologies on the future of medical specialties in one infographic. the medical futurist.
Mittelman M, Markham S, Taylor M, et al. Patient commentary: stop hyping artificial intelligence-patients will always need human doctors. BMJ2018; 363. doi:10.1136/bmj.k4669•Google Scholar•PubMed
Longoni C, Bonezzi A, Morewedge CK, et al. Resistance to medical artificial intelligence. J Consum Res2019; 46:629–50. Google Scholar
Dietvorst BJ, Simmons JP, Massey C, et al. Algorithm aversion: people erroneously avoid algorithms after seeing them err. J Exp Psychol Gen2015; 144:114–26. doi:10.1037/xge0000033•Google Scholar•PubMed
Khullar D, Casalino LP, Qian Y, et al. Public vs physician views of liability for artificial intelligence in health care. J Am Med Inform Assoc2021; 28:1574–7. doi:10.1093/jamia/ocab055•Google Scholar•PubMed
Lee MK. Understanding perception of algorithmic decisions: Fairness, trust, and emotion in response to algorithmic management. Big Data Soc2018; 5:205395171875668. doi:10.1177/2053951718756684•Google Scholar
Thurman N, Moeller J, Helberger N, et al. My friends, editors, algorithms, and I. Digit J2018; Google Scholar