Communication

Scaling equitable artificial intelligence in healthcare with machine learning operations

Machine learning operations (MLOps), a discipline concerned with the production, monitoring and maintenance of artificial intelligence (AI) and machine learning (ML) models at scale, applied in healthcare can facilitate the transition of AI/ML-enabled healthcare tools from research to sustainable deployment.1–3 Adherence to MLOps best practices can address persistent challenges with AI/ML tools deployed into clinical workflows where models often struggle with generalisability, integration and robustness. As AI regulations continue to evolve such as the Department of Health and Human Services Office of Civil Rights final ruling that requires healthcare providers to ensure their AI/ML tools do not discriminate,4 it becomes increasingly essential for MLOps in healthcare to prioritise health equity.

MLOps in healthcare contextualises core principles to serve the needs of AI/ML deployment at healthcare organisations.1–3 We highlight key principles and considerations to scale equitable health AI/ML model deployments and establish accountability measures5 that systematically quell health inequities and comply with AI/ML regulations (table 1).

Table 1
|
Considerations to prioritise health equity in MLOps healthcare infrastructure

One MLOps in healthcare principle is to optimise the clinical workflow integration of health AI/ML tools for care delivery. Additionally, ensuring models are accessible to all patient populations and adaptable to different clinical settings is crucial. A suitable functionality that prioritises health equity is to conduct clinical workflow analyses to identify key stakeholders, processes and resources required to execute AI/ML equitably for patient care. The paths identified inform the development of workflow orchestration components that serve as the foundation of MLOps pipelines at the healthcare organisation.

Another principle is that diverse stakeholders can collaborate to address the operational needs of health AI/ML tools. Effective operations require the coordination of knowledge and efforts across an organisation. Functionalities that prioritise health equity in collaboration include the documentation of health equity assessments6 and the creation of an equitable AI expert panel (eg, healthcare providers, researchers, patients, ethics and compliance officers, diversity, equity, and inclusion officers, privacy officers) to inform decision-making. These components will enhance transparency and help justify the performance, safety and equity thresholds set for deployed AI/ML tools in MLOps healthcare infrastructure.

A defining principle is the continuous maintenance, monitoring and support of health AI/ML tools over time to ensure the most optimal version is in production for clinical impact. Functionalities that promote health equity include reinforcing model monitoring components with the detection of fairness drifts in AI/ML tools. This will require setting triggers for model recalibration, updating or retraining based on key fairness metrics (eg, equalised odds, equal opportunity, demographic parity)7 prior to deployment. Continuously monitoring model performance, safety and equity over time and the ability to reconcile conflicting deviations ensures holistically calibrated AI/ML models reach patients. Furthermore, this component of MLOps in healthcare infrastructure prepares organisations to meet regulatory criteria (eg, predetermined change control plans) and can expedite the Food and Drug Administration-approval of home-grown AI/ML models for broader clinical use.8

MLOps in healthcare also follows the principle of automation regarding the continuous delivery (CD), continuous integration (CI) and continuous deployment of AI/ML tools. This principle requires the setup of CD/CI components that automate steps to enable rapid iteration and model optimisation on receiving feedback. The automation of manual processes can alleviate associated AI/ML operational burdens (eg, costs, time, quality) for an organisation and ensure consistency in bias and fairness checks. To promote health equity through this component, one functional requirement is that feature stores include features considered critical to advancing health equity such as social determinants of health. The ability to standardise health equity feature processing pipelines that feed into feature stores will enable engineers to scale and ensure models are ‘fit-for-equity’.

Finally, an underlying principle is that MLOps in healthcare require compliance with evolving regulations on health AI/ML and patient safety. Organisations must keep pace with changing standards and modify technical guardrails accordingly to provide proper oversight while ensuring AI/ML models do not inadvertently harm vulnerable populations. One functional requirement is to expand automated quality and assurance checks to screen AI/ML tools for bias and fairness before deployment and monitor for discrimination and jailbreaking after deployment. Those unable to meet minimum thresholds of quality in terms of performance, safety and equity are considered not compliant with regulation and pose an excessive risk to worsening health inequities or organisational reputation.

With the surge of health AI innovation, forethought on how to integrate functional health equity accountability measures into MLOps healthcare infrastructure can catalyse the adoption and scale of equitable AI/ML-enabled tools for broad clinical impact. The path forward in healthcare requires recognition, investment and adoption of robust MLOps pipelines to advance operational excellence for equitable AI.