Scaling equitable artificial intelligence in healthcare with machine learning operations
•,,,,.
...
Machine learning operations (MLOps), a discipline concerned with the production, monitoring and maintenance of artificial intelligence (AI) and machine learning (ML) models at scale, applied in healthcare can facilitate the transition of AI/ML-enabled healthcare tools from research to sustainable deployment.1–3 Adherence to MLOps best practices can address persistent challenges with AI/ML tools deployed into clinical workflows where models often struggle with generalisability, integration and robustness. As AI regulations continue to evolve such as the Department of Health and Human Services Office of Civil Rights final ruling that requires healthcare providers to ensure their AI/ML tools do not discriminate,4 it becomes increasingly essential for MLOps in healthcare to prioritise health equity.
MLOps in healthcare contextualises core principles to serve the needs of AI/ML deployment at healthcare organisations.1–3 We highlight key principles and considerations to scale equitable health AI/ML model deployments and establish accountability measures5 that systematically quell health inequities and comply with AI/ML regulations (table 1).
Table 1
|
Considerations to prioritise health equity in MLOps healthcare infrastructure
One MLOps in healthcare principle is to optimise the clinical workflow integration of health AI/ML tools for care delivery. Additionally, ensuring models are accessible to all patient populations and adaptable to different clinical settings is crucial. A suitable functionality that prioritises health equity is to conduct clinical workflow analyses to identify key stakeholders, processes and resources required to execute AI/ML equitably for patient care. The paths identified inform the development of workflow orchestration components that serve as the foundation of MLOps pipelines at the healthcare organisation.
Another principle is that diverse stakeholders can collaborate to address the operational needs of health AI/ML tools. Effective operations require the coordination of knowledge and efforts across an organisation. Functionalities that prioritise health equity in collaboration include the documentation of health equity assessments6 and the creation of an equitable AI expert panel (eg, healthcare providers, researchers, patients, ethics and compliance officers, diversity, equity, and inclusion officers, privacy officers) to inform decision-making. These components will enhance transparency and help justify the performance, safety and equity thresholds set for deployed AI/ML tools in MLOps healthcare infrastructure.
A defining principle is the continuous maintenance, monitoring and support of health AI/ML tools over time to ensure the most optimal version is in production for clinical impact. Functionalities that promote health equity include reinforcing model monitoring components with the detection of fairness drifts in AI/ML tools. This will require setting triggers for model recalibration, updating or retraining based on key fairness metrics (eg, equalised odds, equal opportunity, demographic parity)7 prior to deployment. Continuously monitoring model performance, safety and equity over time and the ability to reconcile conflicting deviations ensures holistically calibrated AI/ML models reach patients. Furthermore, this component of MLOps in healthcare infrastructure prepares organisations to meet regulatory criteria (eg, predetermined change control plans) and can expedite the Food and Drug Administration-approval of home-grown AI/ML models for broader clinical use.8
MLOps in healthcare also follows the principle of automation regarding the continuous delivery (CD), continuous integration (CI) and continuous deployment of AI/ML tools. This principle requires the setup of CD/CI components that automate steps to enable rapid iteration and model optimisation on receiving feedback. The automation of manual processes can alleviate associated AI/ML operational burdens (eg, costs, time, quality) for an organisation and ensure consistency in bias and fairness checks. To promote health equity through this component, one functional requirement is that feature stores include features considered critical to advancing health equity such as social determinants of health. The ability to standardise health equity feature processing pipelines that feed into feature stores will enable engineers to scale and ensure models are ‘fit-for-equity’.
Finally, an underlying principle is that MLOps in healthcare require compliance with evolving regulations on health AI/ML and patient safety. Organisations must keep pace with changing standards and modify technical guardrails accordingly to provide proper oversight while ensuring AI/ML models do not inadvertently harm vulnerable populations. One functional requirement is to expand automated quality and assurance checks to screen AI/ML tools for bias and fairness before deployment and monitor for discrimination and jailbreaking after deployment. Those unable to meet minimum thresholds of quality in terms of performance, safety and equity are considered not compliant with regulation and pose an excessive risk to worsening health inequities or organisational reputation.
With the surge of health AI innovation, forethought on how to integrate functional health equity accountability measures into MLOps healthcare infrastructure can catalyse the adoption and scale of equitable AI/ML-enabled tools for broad clinical impact. The path forward in healthcare requires recognition, investment and adoption of robust MLOps pipelines to advance operational excellence for equitable AI.
X: @MalvikaPillai
Contributors: MYN and AY contributed equally to this paper. Study concept and design: MYN, AY and TH-B. Analysis and interpretation of data: All authors (MYN, AY, MP, VS and TH-B). Writing—original draft preparation: MYN and AY. Writing—critical revision of the manuscript for important intellectual content: All authors (MYN, AY, MP, VS and TH-B). All authors read and approved the final manuscript. TH-B is the guarantor of this work.
Funding: Research reported in this publication was supported by the National Center for Advancing Translational Sciences under Award Number UL1TR003142, Agency for Healthcare Research and Quality (AHRQ) under Award Number R01HS027434 and the National Library of Medicine under Award Number T15LM007033. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health or AHRQ.
Competing interests: None declared.
Provenance and peer review: Not commissioned; externally peer reviewed.
Rajagopal A, Ayanian S, Ryu AJ, et al. Machine Learning Operations in Health Care: A Scoping Review. Mayo Clin Proc Dig Health2024; 2:421–37. doi:10.1016/j.mcpdig.2024.06.009•Google Scholar
Kreuzberger D, Kühl N, Hirschl S, et al. Machine Learning Operations (MLOps): Overview, Definition, and Architecture. IEEE Access2023; 11:31866–79. doi:10.1109/ACCESS.2023.3262138•Google Scholar
U.S. Department of Health and Human Services (HHS) Office for Civil Rights (OCR). HHS issues new rule to strengthen nondiscrimination protections and advance civil rights in health care. 2024;
Chin MH, Afsar-Manesh N, Bierman AS, et al. Guiding Principles to Address the Impact of Algorithm Bias on Racial and Ethnic Disparities in Health and Health Care. JAMA Netw Open2023; 6. doi:10.1001/jamanetworkopen.2023.45050•Google Scholar
Kim JY, Hasan A, Kellogg KC, et al. Development and preliminary testing of Health Equity Across the AI Lifecycle (HEAAL): A framework for healthcare delivery organizations to mitigate the risk of AI solutions worsening health inequities. PLOS Dig Health2024; 3. doi:10.1371/journal.pdig.0000390•Google Scholar