Introduction
The knowledge crisis
Every week, hundreds of papers by expert medical researchers are published in high-quality journals and this research fuels continuous improvements in treatments and medical practice. New knowledge produced through research is often summarised and disseminated in the form of clinical practice guidelines (CPGs) which give short, evidence-based summaries of best practice for specific medical conditions. CPGs are generally considered to be a vital way of disseminating up-to-date recommendations for high quality and safe clinical practice.
However, there are important difficulties for achieving all the potential benefits of CPGs: they take time to read and absorb; they are difficult to keep up to date, and they only provide general guidance not patient-specific recommendations. Chidgey et al1 reviewed the successes and issues facing the UK National Institute for Health and Care Excellence (NICE) in improving the delivery of medical services in the National Health Service. The focus was the NICE programme of guideline development, ‘arguably the largest in the world’, whose goal is to carry out rigorous and up-to-date reviews of evidence for alternative treatments and develop recommendations for practice. Chidgey et al discussed successes in changing practice but concluded that the record of translating guidance into successful implementation was ‘mixed’.
The NICE guidance programme has grown to cover approaching 500 areas of medical practice and diversified into a wide range of products: CPGs, clinical pathways, quality standards, technology assessments, evidence summaries and more. NICE is presently in the process of reviewing its guidance development processes in a programme called NICE Connect (https://www.nice.org.uk/about/who-we-are/nice-connect) whose goal is to facilitate more effective use of NICE content in clinical practice.
In response to the emergence of ‘Rapid Learning Systems’ in healthcare and ‘Computable Biomedical Knowledge’, NICE is also investigating options for disseminating knowledge of best practice in a computable form that can support individualised care. There are many options to be considered, from restructuring and enriching guideline documents so they can be read by computers and applied in more context specific ways (eg, GEM Guideline Elements Model (https://www.astm.org/DATABASE.CART/HISTORICAL/E2210-02.htm), MAGIC enriched guideline content (http://magicproject.org/)) to more technical approaches like statistical and knowledge-based decision-support systems (https://en.wikipedia.org/wiki/Clinical_decision_support_system); rule-based systems for making patient-specific recommendations (eg, Arden syntax (https://en.wikipedia.org/wiki/Arden_syntax), Clinical Quality Language (https://ecqi.healthit.gov/cql)) to Computer Interpretable Clinical Guidelines or CIGs (Peleg et al2) which can also be integrated with the other methods if required.3
Computer interpretable clinical guidelines: publets
OpenClinical was conceived as a way of developing CIGs using a class of artificial intelligence techniques called knowledge engineering.3 It is a web-based knowledge-sharing platform which offers a radically different approach to summarising knowledge of best practice in the form of executable models of practice that we call publets. Publets include structured data models and executable logic (rules, decisions, pathways) as well as traditional CPG content. When a publet is consulted in a patient’s care, it can request and interpret information about the patient and offer personalised recommendations and explanations of the reasons for its recommendations and the supporting evidence if required. OpenClinical (https://www.openclinical.net/) was originally developed at Cancer Research UK and later at Oxford University and UCL/Royal Free Hospital in London. It was launched in its first ‘alpha’ version at the Royal Free in 2013. We plan to deploy a significantly improved technology and publishing platform in 2020.
Publets can support a clinician through just the parts of the CPG that are relevant for each patient, summarise the patient-specific pros and cons of each decision option and provide the supporting evidence. Users can either act on or critique suggestions as they see fit in light of the patient-specific rationale provided by the publet and their own professional judgement.
The OpenClinical concept is that clinicians, researchers and other authors develop models of best practice and submit them for review and publication. Authors will submit publets to OpenClinical in much the same way that researchers submit research papers to conventional research journals; each publet can be peer reviewed by independent experts and validated against cases before publication. Several publet demonstrations can be accessed at https://openclinical.net/index.php?id=68 and a repository of diverse examples is at https://openclinical.net/index.php?id=69.
Publets empower healthcare professionals and clinical researchers to develop and share models of practice in any area of medicine. Colleagues in other institutions, countries and specialties can download, assess and adapt the publet to meet local requirements and constraints.
Methods for creating publets: knowledge engineering
The tools for developing publets provided by OpenClinical exploit a range of artificial intelligence techniques known collectively as Knowledge Engineering. This discipline emerged from artificial intelligence and computer science but has also drawn on insights from cognitive science in that it exploits our understanding of human decision-making and the cognitive skills needed to carry out complex tasks. Medicine and clinical practice have been among the most important targets and challenges for this branch of artificial intelligence research.
An important use of knowledge engineering is to formalise human-readable CPGs (traditionally text, tables, flow charts, etc) in a logical or other symbolic form rather than as a mathematical function or algorithm. Like algorithms knowledge models can be executed on a computer and assist clinicians in routine tasks, but symbolic task models are easier for healthcare professionals to understand and critique than conventional algorithms.
In a classic paper, Peleg et al4 reviewed a number of approaches to modelling clinical guidelines that draw on concepts from knowledge engineering including EON (USA), ASBRU (Israel), GUIDE (Italy), GLIF (USA) and PROforma (UK5). Peleg et al systematically compared the abilities of these different methods in capturing practice guidelines. OpenClinical uses the PROforma language model and CIG authoring tools for modelling decision-making and pathways, but we wish to accommodate other approaches as they emerge.
There are many differences between publets and other CIG models that have been proposed. PROforma is based on a theory of human expertise and it is easily understood and written by clinicians so they can create, validate and share publets with their peers. Despite its naturalistic form, PROforma has sound foundations in logic and decision theory. In practical terms, it has proved to be a versatile, scalable and effective tool for improving decision-making in many clinical settings and specialties (see table). Tools for authoring and testing publets are available from OpenClinical.
The PROforma approach to designing and deploying artificial intelligence at the point of care is different from the currently popular machine learning and data science approaches, in that it is grounded in an understanding of human expertise and getting the best from a combination of natural and artificial intelligence. It has shown that knowledge engineering methods are effective and versatile and that publets can be deployed at scale in a way that healthcare professionals can understand, critically engage with and, if warranted, challenge.