Research paper
“Think aloud” and “Near live” usability testing of two complex clinical decision support tools

https://doi.org/10.1016/j.ijmedinf.2017.06.003Get rights and content

Highlights

  • “Think Aloud” and “Near Live” usability testing generated unique and generalizable insights.

  • Feedback during “Think Aloud” testing primarily helped to improve the tools’ ease of use.

  • “Near Live” testing was helpful for eliciting key barriers to provider workflow and adoption.

  • During “Near Live” testing participants were more critical of the tools’ usefulness.

  • Usability testing helps decision support reach its potential to improve health outcomes.

Abstract

Objectives

Low provider adoption continues to be a significant barrier to realizing the potential of clinical decision support. “Think Aloud” and “Near Live” usability testing were conducted on two clinical decision support tools. Each was composed of an alert, a clinical prediction rule which estimated risk of either group A Streptococcus pharyngitis or pneumonia and an automatic order set based on risk. The objective of this study was to further understanding of the facilitators of usability and to evaluate the types of additional information gained from proceeding to “Near Live” testing after completing “Think Aloud”.

Methods

This was a qualitative observational study conducted at a large academic health care system with 12 primary care providers. During “Think Aloud” testing, participants were provided with written clinical scenarios and asked to verbalize their thought process while interacting with the tool. During “Near Live” testing participants interacted with a mock patient. Morae usability software was used to record full screen capture and audio during every session. Participant comments were placed into coding categories and analyzed for generalizable themes. Themes were compared across usability methods.

Results

“Think Aloud” and “Near Live” usability testing generated similar themes under the coding categories visibility, workflow, content, understand-ability and navigation. However, they generated significantly different themes under the coding categories usability, practical usefulness and medical usefulness. During both types of testing participants found the tool easier to use when important text was distinct in its appearance, alerts were passive and appropriately timed, content was up to date, language was clear and simple, and each component of the tool included obvious indicators of next steps. Participant comments reflected higher expectations for usability and usefulness during “Near Live” testing. For example, visit aids, such as automatically generated order sets, were felt to be less useful during “Near-Live” testing because they would not be all inclusive for the visit.

Conclusions

These complementary types of usability testing generated unique and generalizable insights. Feedback during “Think Aloud” testing primarily helped to improve the tools’ ease of use. The additional feedback from “Near Live” testing, which mimics a real clinical encounter, was helpful for eliciting key barriers and facilitators to provider workflow and adoption.

Section snippets

Background

Clinical decision support (CDS) has demonstrated the ability to shape health care provider behavior towards more evidence based clinical practice by improving diagnosis, treatment, and preventative care services [1], [2], [3], [4], [5], [6]. CDS is typically integrated into the electronic health record (EHR) and functions to bring key pieces of evidence or best practice guidelines to the point of care. These tools stand to improve the American healthcare system where on average it takes five

Methods

This was a qualitative observational study done at the University of Wisconsin, a large academic health care center. “Think Aloud” testing was completed with 4 participants. The tool was revised based on these results before “Near Live” testing was conducted with 8 participants. Different participants were recruited for each type of testing, as is typically the case, to minimize the time commitment required from each of these busy health care providers. Both “Think Aloud” and “Near Live”

Results

Participants were primarily medical doctors, along with one nurse practitioner and one physician assistant. (Table 1) Our sample was 42% female with an average age of 47.1, 18.7 years of post-graduate practice, and 7.7 years of experience using an EHR. Average SUS was 85.6 during “Think Aloud” testing and 81.3 during “Near Live” testing. SUS scores range from 0 to 100, with 100 being a perfect score [17]. An SUS score of 68 is considered average [18]. Those raw scores would correspond to the

Discussion

“Think Aloud” and “Near Live” usability testing of these two CDS tools generated unique insights and lessons generalizable to all forms of CDS. Participant commentary was consistent across participants and across the two tools. This is the first study to evaluate generalizable lessons learned from “Think Aloud” and “Near Live” usability testing of complex clinical decision support tools. Previous studies documented usability findings particular to the tools studied and the relative percentages

Conclusion

“Think Aloud” and “Near Live” usability testing provide CDS tool designers with complementary insights that when combined provide a more robust understanding of CDS usability. Participant comments made during “Think Aloud” testing about visibility, content, understand-ability and navigation primarily helped to improve the tools ease of use. Participant comments and observed behavior during “Near Live” testing, which mimics a real clinical encounter, were more helpful for eliciting key barriers

Conflict of interest

The authors declare that they have no competing interests.

Funding

This project was funded by the National Institutes of Health, National institute of Allergy and Infectious Diseases, under grant #5R01 AI108680-03. The funding body had no role in the design of the study or the collection, analysis, or interpretation of data.

Ethics approval and consent to participate

Written informed consent was obtained from all participants. The Institutional Review Boards at both institutions approved the research protocol.

Consent for publication

Not applicable.

AUTHORS’ CONTRIBUTIONS

Safiya Richardson, Rachel Mishuris and Alexander O’Connell analyzed and interpreted the qualitative data. David Feldstein, Rachel Hess, Paul Smith and Lauren McCullagh conducted usability testing with participants. Safiya Richardson, Thomas McGinn and Devin Mann were major contributors in writing the manuscript. All authors read and approved the final manuscript.

Availability of data and materials

Datasets analyzed during this study are available from the corresponding author on reasonable request.

SUMMARY POINTS

Already Known

  • Low provider adoption continues to be a significant barrier to realizing the potential of clinical decision support.

  • Efficiency, usefulness, information content, user interface, and workflow have been reported by clinicians to be the keys to effective decision support.

This Study Has Added:

  • “Think Aloud” and “Near Live” usability testing of these two CDS tools generated

Acknowledgements

Not applicable.

References (35)

  • R. Kaushal et al.

    Effects of computerized physician order entry and clinical decision support systems on medication safety: a systematic review

    Arch. Intern. Med.

    (2003)
  • P.S. Roshanov et al.

    Features of effective computerised clinical decision support systems: meta-regression of 162 randomised trials

    BMJ

    (2013)
  • N.M. Souza et al.

    Computerized clinical decision support systems for primary preventive care: a decision-maker-researcher partnership systematic review of effects on process of care and patient outcomes

    Implement Sci.

    (2011)
  • T.G. McGinn et al.

    Efficacy of an evidence-based clinical decision support in primary care practices: a randomized clinical trial

    JAMA Intern. Med.

    (2013)
  • E.A. McGlynn et al.

    The quality of health care delivered to adults in the United States

    New Engl. J. Med.

    (2003)
  • J. Lomas et al.

    From evidence to practice in the United States, the United Kingdom, and Canada

    Milbank Q.

    (1993)
  • D.A. Feldstein et al.

    Design and implementation of electronic health record integrated clinical prediction rules (iCPR): a randomized trial in diverse primary care settings

    Implement. Sci.

    (2017)
  • Cited by (66)

    • Education of artificial intelligence for cardiovascular clinicians

      2023, Intelligence-Based Cardiology and Cardiac Surgery: Artificial Intelligence and Human Cognition in Cardiovascular Medicine
    • Combined Expert and User-Driven Usability Assessment of Trauma Decision Support Systems Improves User-Centered Design

      2022, Surgery (United States)
      Citation Excerpt :

      Several of our themes have been mentioned previously in the literature,24,49 the most recognizable being automation, which has long been cited as a key source of physician burnout related to EHRs but which remains a challenge as the EHR becomes increasingly capable yet complex. In our review of usability testing studies involving only end-users, in addition to automation, themes commonly related to workflow,18 level of interruption,37 content,41 functionality,51 and usefulness.39 This literature supports our findings that end-users are more likely to identify issues relating to workflow and automation.

    • Impact of interactive computerised decision support for hospital antibiotic use (COMPASS): an open-label, cluster-randomised trial in three Swiss hospitals

      2022, The Lancet Infectious Diseases
      Citation Excerpt :

      Finally, we showed in a previously published qualitative study that efficiency and ease of use are two key facilitators for the adoption of CDSS.8 Even though the tools were developed iteratively with a testing process involving clinicians, we probably should have invested more time to involve end-users through an extensive user-experience process to optimise these aspects and improve user acceptance and adoption.30 The effectiveness of the COMPASS intervention to reduce overall antimicrobial use remains inconclusive.

    View all citing articles on Scopus
    View full text