Article Text

Download PDFPDF

Healthcare provider evaluation of machine learning-directed care: reactions to deployment on a randomised controlled study
  1. Julian C Hong1,2,3,
  2. Pranalee Patel4,
  3. Neville C W Eclov4,
  4. Sarah J Stephens4,
  5. Yvonne M Mowery4,5,
  6. Jessica D Tenenbaum6 and
  7. Manisha Palta4
  1. 1Department of Radiation Oncology, University of California San Francisco, San Francisco, California, USA
  2. 2Bakar Computational Health Sciences Institute, University of California San Francisco, San Francisco, California, USA
  3. 3Joint Program in Computational Precision Health, UCSF-UC Berkeley, San Francisco, California, USA
  4. 4Department of Radiation Oncology, Duke University, Durham, North Carolina, USA
  5. 5Department of Head and Neck Surgery & Communication Sciences, Duke University, Durham, North Carolina, USA
  6. 6Department of Biostatistics and Bioinformatics, Duke University, Durham, North Carolina, USA
  1. Correspondence to Dr Julian C Hong; julian.hong{at}ucsf.edu

Abstract

Objectives Clinical artificial intelligence and machine learning (ML) face barriers related to implementation and trust. There have been few prospective opportunities to evaluate these concerns. System for High Intensity EvaLuation During Radiotherapy (NCT03775265) was a randomised controlled study demonstrating that ML accurately directed clinical evaluations to reduce acute care during cancer radiotherapy. We characterised subsequent perceptions and barriers to implementation.

Methods An anonymous 7-question Likert-type scale survey with optional free text was administered to multidisciplinary staff focused on workflow, agreement with ML and patient experience.

Results 59/71 (83%) responded. 81% disagreed/strongly disagreed their workflow was disrupted. 67% agreed/strongly agreed patients undergoing intervention were high risk. 75% agreed/strongly agreed they would implement the ML approach routinely if the study was positive. Free-text feedback focused on patient education and ML predictions.

Conclusions Randomised data and firsthand experience support positive reception of clinical ML. Providers highlighted future priorities, including patient counselling and workflow optimisation.

  • Machine Learning
  • Delivery of Health Care

Data availability statement

Data are available on reasonable request. Research data are stored in an institutional repository, and will be shared on reasonable request to the corresponding author.

http://creativecommons.org/licenses/by-nc/4.0/

This is an open access article distributed in accordance with the Creative Commons Attribution Non Commercial (CC BY-NC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited, appropriate credit is given, any changes made indicated, and the use is non-commercial. See: http://creativecommons.org/licenses/by-nc/4.0/.

Statistics from Altmetric.com

Request Permissions

If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.

Introduction

Artificial intelligence (AI) and machine learning (ML) has the potential to transform medical practice. Despite many retrospective studies, randomised controlled trials (RCTs), particularly interventional trials, remain limited.1–3 Thus, there have been limited opportunities to formally characterise barriers to the implementation of healthcare AI and ML and identify solutions.4 5 There are minimal reports describing provider opinions following a prospective randomised interventional study of healthcare ML.

One application of healthcare ML is in the prediction and reduction of acute care (emergency visits and hospitalisations) during outpatient cancer therapy,1 6–9 prioritized by the Centers for Medicare and Medicaid Services.10 The System for High Intensity EvaLuation During Radiotherapy study (SHIELD-RT; NCT03775265) was a randomised controlled quality improvement study of an ML model predicting acute care visits (emergency department visits and/or hospitalisation) during radiotherapy (RT) or chemoradiotherapy (CRT).1 ,6 ML identified high-risk patients for supplemental clinical evaluations, which reduced acute care rates from 22.3% to 12.3%, with low-risk patients experiencing a 2.7% rate. Radiation oncology care uniquely requires a diverse clinical staff, including attending and resident physicians, advanced practice provider (APPs), nurses and radiation therapists (RTTs), each with different viewpoints on how ML can optimally play a role in delivering care. Following the completion but prior to final analysis of SHIELD-RT, we administered a survey to understand the perspectives of healthcare providers with regard to the acceptability and feasibility of ML-directed strategies, addressing key components of the implementation outcomes framework.11 The objective was to evaluate specific barriers to planned long-term implementation.

Methods

We conducted a single institution survey of perceptions of SHIELD-RT, during which all outpatient adult courses of RT and CRT initiated from 7 January 2019 to 30 June 2019 were evaluated during the first week of treatment by ML to identify high-risk patients with >10% risk of an acute care visit during RT.1 6 Patients were randomised to standard of care (mandatory weekly on-treatment and clinically indicated ad hoc visits) versus mandatory twice-weekly visits. Interventional second weekly visits were facilitated through an alert that notified RTTs to bring patients to an appropriate clinic room to then be seen by an APP, nurse clinician, resident physician or attending physician. The primary endpoint was rate of acute care visits during RT. Additional details of SHIELD-RT and its primary analysis and implementation workflow were previously reported.1 12

Involved attending and resident physicians, APPs, nurses and RTTs were invited to participate in an anonymous survey to characterise workflow satisfaction and evaluation of potential barriers to future adoption. This included eight questions on a Likert-type scale characterising respondents’ attitudes with an optional free-text comment field.

Results

A total of 59/71 (83%) of invited staff completed the survey, including 14/16 attending physicians (MD), 9/9 resident physicians, 3/5 APPs, 10/11 nurses, 23/30 RTTs (table 1). Eighty-one per cent of staff disagreed or strongly disagreed that the study disrupted their workflow. Only 51% of respondents agreed or strongly agreed that they were aware of their patients undergoing the intervention; 3% agreed that their clinical management beyond the study intervention was altered. Of those aware of patients seen twice weekly, 67% agreed or strongly agreed that patients undergoing intervention were high risk. Most staff (64%) neither agreed nor disagreed that patients understood the study. Willingness for future adoption was favourable, as 75% of respondents agreed or strongly agreed that they would implement the intervention routinely if the study was positive; 41% agreed or strongly agreed and none disagreed that their opinion of clinical ML improved following the study.

Table 1

Responses to survey questions

There were 8 (16%) free-text comments. Three (two RTT and one nurse) indicated confusion among staff and patients with the need and logistics of the supplemental visit. One nurse noted that they felt ML overestimated the risk of their patients (specifically in brain tumours). Two MD responses indicated that they had minimal contact with patients on study. Two (one MD and one RTT) responses expressed anticipation for the results of the study.

Discussion

Our study highlights an overall positive reception towards ML implementation in an academic radiation oncology clinic. Our survey supports that RCT results drive willingness to routinely adopt clinical ML. ML-guidance and supplemental visits were integrated successfully into our clinical workflow with minimal perceived disruption.

This analysis shows how some concerns regarding ML may be overcome. In addition to randomised evidence, direction observation of ML operating in a controlled setting may have improved subjective opinions of clinical ML prior to the study. This is instrumental given recent data demonstrating the limitations of commercial prediction models,13 and ultimately, subsequent to this survey, the SHIELD-RT analysis demonstrated a reduction in acute care events.1 While ML will continue to require complementary input from healthcare professionals, these survey results are promising for adoption.14 Our clinic is currently incorporating this ML-directed clinical strategy into routine practice.

Overall, ML implementation had limited provider-perceived impact on clinical workflows, to the point of reducing MD awareness as indicated by survey responses. This was intentional in the design to minimise extra cognitive and functional effort to improve the likelihood of MD adoption.1 12 One relative exception to this was surveyed APPs, the majority of whom participated in the interventional second mandatory clinical evaluation. This suggests that ML-guided interventions may place greater burden on specific staff. This cost must be considered in model and interventional design.

Among limited free-text comments, staff reservations focused on patient education and ML risk predictions. Patients were not surveyed, although staff both anecdotally and in the survey highlighted logistical challenges surrounding location and timing of supplemental visits. While patients were educated when undergoing the supplemental evaluation, the neutral evaluation of patient understanding and anecdotal responses highlight the reported challenges of explaining the algorithm and its clinical implications to patients. This emphasises the need for transparent and explainable approaches, especially given increasingly opaque AI methods. Despite the single comment noting concern for overestimation, calibration analyses previously reported in the primary study results demonstrated good model performance in comparison to clinicians who were more inconsistent, with wide CIs, and assigned a 0% risk to a patient who had an acute care event.1 It is possible that over time, both improved explainability and consistent observation of ML accuracy may demonstrate longitudinal improvements in clinician perception.

There are limitations to our study. We surveyed staff only following completion of the study, and direct comparisons pre-SHIELD-RT and post-SHIELD-RT were not possible. The results of this survey may be subject to bias, though we had a high rate of completion (83%) across a range of roles, with a high representation of non-academic staff (61% of respondents; APPs, nurses and RTTs).

The results of this study inform our future directions, primarily emphasising the importance of RCTs in demonstrated clinical ML benefit and highlighting the need for concerted efforts in patient and staff education. Other ongoing work focuses on optimising workflows, patient logistics, long-term ML surveillance and generalisability.

Data availability statement

Data are available on reasonable request. Research data are stored in an institutional repository, and will be shared on reasonable request to the corresponding author.

Ethics statements

Patient consent for publication

References

Footnotes

  • Twitter @julian_hong

  • Contributors JCH and MP: concept and design. JCH, PP, NCWE, SJS, YMM, JDT, MP: acquisition, analysis or interpretation of data. JCH, PP, NCWE, SJS, YMM, JDT, MP: drafting of the manuscript. JCH, PP, NCWE, SJS, YMM, JDT, MP: critical revision of the manuscript for important intellectual content.

  • Funding This study was supported in part by the Duke Endowment, the Radiation Oncology Institute and the Conquer Cancer Foundation which had no role in study design, data collection and analysis, decision to publish or preparation of the manuscript. The Duke Department of Radiation Oncology also provided funding. JCH is also supported by a career development grant from the American Society for Radiation Oncology and Prostate Cancer Foundation.

  • Competing interests JDT, MP and JH are coinventors on a pending patent, 'Systems and methods for predicting acute care visits during outpatient cancer therapy,' related to the current work.

  • Provenance and peer review Not commissioned; externally peer reviewed.