“Think Aloud” and “Near Live” usability testing generated unique and generalizable insights.
•
Feedback during “Think Aloud” testing primarily helped to improve the tools’ ease of use.
•
“Near Live” testing was helpful for eliciting key barriers to provider workflow and adoption.
•
During “Near Live” testing participants were more critical of the tools’ usefulness.
•
Usability testing helps decision support reach its potential to improve health outcomes.
Abstract
Objectives
Low provider adoption continues to be a significant barrier to realizing the potential of clinical decision support. “Think Aloud” and “Near Live” usability testing were conducted on two clinical decision support tools. Each was composed of an alert, a clinical prediction rule which estimated risk of either group A Streptococcus pharyngitis or pneumonia and an automatic order set based on risk. The objective of this study was to further understanding of the facilitators of usability and to evaluate the types of additional information gained from proceeding to “Near Live” testing after completing “Think Aloud”.
Methods
This was a qualitative observational study conducted at a large academic health care system with 12 primary care providers. During “Think Aloud” testing, participants were provided with written clinical scenarios and asked to verbalize their thought process while interacting with the tool. During “Near Live” testing participants interacted with a mock patient. Morae usability software was used to record full screen capture and audio during every session. Participant comments were placed into coding categories and analyzed for generalizable themes. Themes were compared across usability methods.
Results
“Think Aloud” and “Near Live” usability testing generated similar themes under the coding categories visibility, workflow, content, understand-ability and navigation. However, they generated significantly different themes under the coding categories usability, practical usefulness and medical usefulness. During both types of testing participants found the tool easier to use when important text was distinct in its appearance, alerts were passive and appropriately timed, content was up to date, language was clear and simple, and each component of the tool included obvious indicators of next steps. Participant comments reflected higher expectations for usability and usefulness during “Near Live” testing. For example, visit aids, such as automatically generated order sets, were felt to be less useful during “Near-Live” testing because they would not be all inclusive for the visit.
Conclusions
These complementary types of usability testing generated unique and generalizable insights. Feedback during “Think Aloud” testing primarily helped to improve the tools’ ease of use. The additional feedback from “Near Live” testing, which mimics a real clinical encounter, was helpful for eliciting key barriers and facilitators to provider workflow and adoption.
Section snippets
Background
Clinical decision support (CDS) has demonstrated the ability to shape health care provider behavior towards more evidence based clinical practice by improving diagnosis, treatment, and preventative care services [1], [2], [3], [4], [5], [6]. CDS is typically integrated into the electronic health record (EHR) and functions to bring key pieces of evidence or best practice guidelines to the point of care. These tools stand to improve the American healthcare system where on average it takes five
Methods
This was a qualitative observational study done at the University of Wisconsin, a large academic health care center. “Think Aloud” testing was completed with 4 participants. The tool was revised based on these results before “Near Live” testing was conducted with 8 participants. Different participants were recruited for each type of testing, as is typically the case, to minimize the time commitment required from each of these busy health care providers. Both “Think Aloud” and “Near Live”
Results
Participants were primarily medical doctors, along with one nurse practitioner and one physician assistant. (Table 1) Our sample was 42% female with an average age of 47.1, 18.7 years of post-graduate practice, and 7.7 years of experience using an EHR. Average SUS was 85.6 during “Think Aloud” testing and 81.3 during “Near Live” testing. SUS scores range from 0 to 100, with 100 being a perfect score [17]. An SUS score of 68 is considered average [18]. Those raw scores would correspond to the
Discussion
“Think Aloud” and “Near Live” usability testing of these two CDS tools generated unique insights and lessons generalizable to all forms of CDS. Participant commentary was consistent across participants and across the two tools. This is the first study to evaluate generalizable lessons learned from “Think Aloud” and “Near Live” usability testing of complex clinical decision support tools. Previous studies documented usability findings particular to the tools studied and the relative percentages
Conclusion
“Think Aloud” and “Near Live” usability testing provide CDS tool designers with complementary insights that when combined provide a more robust understanding of CDS usability. Participant comments made during “Think Aloud” testing about visibility, content, understand-ability and navigation primarily helped to improve the tools ease of use. Participant comments and observed behavior during “Near Live” testing, which mimics a real clinical encounter, were more helpful for eliciting key barriers
Conflict of interest
The authors declare that they have no competing interests.
Funding
This project was funded by the National Institutes of Health, National institute of Allergy and Infectious Diseases, under grant #5R01 AI108680-03. The funding body had no role in the design of the study or the collection, analysis, or interpretation of data.
Ethics approval and consent to participate
Written informed consent was obtained from all participants. The Institutional Review Boards at both institutions approved the research protocol.
Consent for publication
Not applicable.
AUTHORS’ CONTRIBUTIONS
Safiya Richardson, Rachel Mishuris and Alexander O’Connell analyzed and interpreted the qualitative data. David Feldstein, Rachel Hess, Paul Smith and Lauren McCullagh conducted usability testing with participants. Safiya Richardson, Thomas McGinn and Devin Mann were major contributors in writing the manuscript. All authors read and approved the final manuscript.
Availability of data and materials
Datasets analyzed during this study are available from the corresponding author on reasonable request.
SUMMARY POINTS
Already Known
•
Low provider adoption continues to be a significant barrier to realizing the potential of clinical decision support.
•
Efficiency, usefulness, information content, user interface, and workflow have been reported by clinicians to be the keys to effective decision support.
This Study Has Added:
•
“Think Aloud” and “Near Live” usability testing of these two CDS tools generated
Effect of clinical decision-support systems: a systematic review
Ann. Intern. Med.
(2012)
R. Kaushal et al.
Effects of computerized physician order entry and clinical decision support systems on medication safety: a systematic review
Arch. Intern. Med.
(2003)
P.S. Roshanov et al.
Features of effective computerised clinical decision support systems: meta-regression of 162 randomised trials
BMJ
(2013)
N.M. Souza et al.
Computerized clinical decision support systems for primary preventive care: a decision-maker-researcher partnership systematic review of effects on process of care and patient outcomes
Implement Sci.
(2011)
T.G. McGinn et al.
Efficacy of an evidence-based clinical decision support in primary care practices: a randomized clinical trial
JAMA Intern. Med.
(2013)
E.A. McGlynn et al.
The quality of health care delivered to adults in the United States
New Engl. J. Med.
(2003)
J. Lomas et al.
From evidence to practice in the United States, the United Kingdom, and Canada
Milbank Q.
(1993)
D.A. Feldstein et al.
Design and implementation of electronic health record integrated clinical prediction rules (iCPR): a randomized trial in diverse primary care settings
2023, Intelligence-Based Cardiology and Cardiac Surgery: Artificial Intelligence and Human Cognition in Cardiovascular Medicine
As artificial intelligence (AI) advances, and the domain of clinical medicine becomes more subspecialized, the knowledge gulf between the two grows. Cardiology is a prime example of a highly subspecialized medical field that stands to benefit substantially from AI. By updating cardiology curricula to include AI components, we can equip clinicians to direct the evolution of AI-augmented cardiology.
Several of our themes have been mentioned previously in the literature,24,49 the most recognizable being automation, which has long been cited as a key source of physician burnout related to EHRs but which remains a challenge as the EHR becomes increasingly capable yet complex. In our review of usability testing studies involving only end-users, in addition to automation, themes commonly related to workflow,18 level of interruption,37 content,41 functionality,51 and usefulness.39 This literature supports our findings that end-users are more likely to identify issues relating to workflow and automation.
Trauma clinical decision support systems improve adherence with evidence-based practice but suffer from poor usability and the lack of a user-centered design. The objective of this study was to compare the effectiveness of user and expert-driven usability testing methods to detect usability issues in a rib fracture clinical decision support system and identify guiding principles for trauma clinical decision support systems.
A user-driven and expert-driven usability investigation was conducted using a clinical decision support system developed for patients with rib fractures. The user-driven usability evaluation was as follows: 10 clinicians were selected for simulation-based usability testing using snowball sampling, and each clinician completed 3 simulations using a video-conferencing platform. End-users participated in a novel team-based approach that simulated realistic clinical workflows. The expert-driven heuristic evaluation was as follows: 2 usability experts conducted a heuristic evaluation of the clinical decision support system using 10 common usability heuristics. Usability issues were identified, cataloged, and ranked for severity using a 4-level ordinal scale. Thematic analysis was utilized to categorize the identified usability issues.
Seventy-nine usability issues were identified; 63% were identified by experts and 48% by end-users. Notably, 58% of severe usability issues were identified by experts alone. Only 11% of issues were identified by both methods. Five themes were identified that could guide the design of clinical decision support systems—transparency, functionality and integration into workflow, automated and noninterruptive, flexibility, and layout and appearance. Themes were preferentially identified by different methods.
We found that a dual-method usability evaluation involving usability experts and end-users drastically improved detection of usability issues over single-method alone. We identified 5 themes to guide trauma clinical decision support system design. Performing usability testing via a remote video-conferencing platform facilitated multi-site involvement despite a global pandemic.
Finally, we showed in a previously published qualitative study that efficiency and ease of use are two key facilitators for the adoption of CDSS.8 Even though the tools were developed iteratively with a testing process involving clinicians, we probably should have invested more time to involve end-users through an extensive user-experience process to optimise these aspects and improve user acceptance and adoption.30 The effectiveness of the COMPASS intervention to reduce overall antimicrobial use remains inconclusive.
Computerised decision-support systems (CDSSs) for antibiotic stewardship could help to assist physicians in the appropriate prescribing of antibiotics. However, high-quality evidence for their effect on the quantity and quality of antibiotic use remains scarce. The aim of our study was to assess whether a computerised decision support for antimicrobial stewardship combined with feedback on prescribing indicators can reduce antimicrobial prescriptions for adults admitted to hospital.
The Computerised Antibiotic Stewardship Study (COMPASS) was a multicentre, cluster-randomised, parallel-group, open-label superiority trial that aimed to assess whether a multimodal computerised antibiotic-stewardship intervention is effective in reducing antibiotic use for adults admitted to hospital. After pairwise matching, 24 wards in three Swiss tertiary-care and secondary-care hospitals were randomised (1:1) to the CDSS intervention or to standard antibiotic stewardship measures using an online random sequence generator. The multimodal intervention consisted of a CDSS providing support for choice, duration, and re-evaluation of antimicrobial therapy, and feedback on antimicrobial prescribing quality. The primary outcome was overall systemic antibiotic use measured in days of therapy per admission, using adjusted-hurdle negative-binomial mixed-effects models. The analysis was done by intention to treat and per protocol. The study was registered with ClinicalTrials.gov (identifier NCT03120975).
24 clusters (16 at Geneva University Hospitals and eight at Ticino Regional Hospitals) were eligible and randomly assigned to control or intervention between Oct 1, 2018, and Dec 31, 2019. Overall, 4578 (40·2%) of 11 384 admissions received antibiotic therapy in the intervention group and 4142 (42·8%) of 9673 in the control group. The unadjusted overall mean days of therapy per admission was slightly lower in the intervention group than in the control group (3·2 days of therapy per admission, SD 6·2, vs 3·5 days of therapy per admission, SD 6·8; p<0·0001), and was similar among patients receiving antibiotics (7·9 days of therapy per admission, SD 7·6, vs 8·1 days of therapy per admission, SD 8·4; p=0·50). After adjusting for confounders, there was no statistically significant difference between groups for the odds of an admission receiving antibiotics (odds ratio [OR] for intervention vs control 1·12, 95% CI 0·94–1·33). For admissions with antibiotic exposure, days of therapy per admission were also similar (incidence rate ratio 0·98, 95% CI 0·90–1·07). Overall, the CDSS was used at least once in 3466 (75·7%) of 4578 admissions with any antibiotic prescription, but from the first day of antibiotic treatment for only 1602 (58·9%) of 2721 admissions in Geneva. For those for whom the CDSS was not used from the first day, mean time to use of CDSS was 8·9 days. Based on the manual review of 1195 randomly selected charts, transition from intravenous to oral therapy was significantly more frequent in the intervention group after adjusting for confounders (154 [76·6%] of 201 vs 187 [87%] of 215, +10·4%; OR 1·9, 95% CI 1·1–3·3). Consultations by infectious disease specialists were less frequent in the intervention group (388 [13·4%] of 2889) versus the control group (405 [16·9%] of 2390; OR 0·84, 95% CI 0·59–1·25).
An integrated multimodal computerised antibiotic stewardship intervention did not significantly reduce overall antibiotic use, the primary outcome of the study. Contributing factors were probably insufficient uptake, a setting with relatively low antibiotic use at baseline, and delays between ward admission and first CDSS use.
Swiss National Science Foundation.
For the French and Italian translations of the abstract see Supplementary Materials section.