A distributed approach to the regulation of clinical AI

No organization by itself, including the FDA, has the expertise to oversee the use of clinical AI.

PLOS Digital Health

Given that clinical AI has to be recalibrated on the local population prior to deployment, continuously monitored and regularly updated to account for dataset shift and calibration drift,

the existing model where the FDA is solely responsible for approving an AI algorithm is no longer going to be adequate.

Abstract

Regulation is necessary to ensure the safety, efficacy and equitable impact of clinical artificial intelligence (AI). The number of applications of clinical AI is increasing, which, amplified by the need for adaptations to account for the heterogeneity of local health systems and inevitable data drift, creates a fundamental challenge for regulators. Our opinion is that, at scale, the incumbent model of centralized regulation of clinical AI will not ensure the safety, efficacy, and equity of implemented systems. We propose a hybrid model of regulation, where centralized regulation would only be required for applications of clinical AI where the inference is entirely automated without clinician review, have a high potential to negatively impact the health of patients and for algorithms that are to be applied at national scale by design. This amalgam of centralized and decentralized regulation we refer to as a distributed approach to the regulation of clinical AI and highlight the benefits as well as the pre-requisites and challenges involved.