As healthcare infrastructure develops, many are questioning whether clinical data should be treated as a public good to help improve services in the era of artificial intelligence (AI).
According to a special report published in the journal Radiology, clinical data should be treated as a public good when it is used for secondary purposes, such as research or the development of AI algorithms.
The rapid development of AI, coming on the heels of the widespread adoption of electronic medical records, has opened up exciting possibilities in medicine. AI can potentially streamline and improve the analysis of medical images, but first it must be trained on large troves of data from mammograms, CT scans and other imaging exams.
Ethical concerns around sharing clinical data
One of the current limitations of the advancement of AI-based tools is the lack of broad consensus on an ethical framework for sharing clinical data.
Study lead author Dr David B. Larson, M.D., from the Stanford University School of Medicine, California, said: “This means that, on one hand, clinical data should be made available to researchers and developers after it has been aggregated and all patient identifiers have been removed.
“On the other hand, all who interact with such data should be held to high ethical standards, including protecting patient privacy and not selling clinical data.
“Now that we have electronic access to clinical data and the data processing tools, we can dramatically accelerate our ability to gain understanding and develop new applications that can benefit patients and populations, but unsettled questions regarding the ethical use of the data often preclude the sharing of that information.”
To help answer those questions, Dr Larson and his colleagues at Stanford University developed a framework for using and sharing clinical data in the development of AI applications.
Arguments regarding the sharing of clinical data traditionally have fallen into one of two camps: either the patient owns the data or the institution does. Dr Larson and colleagues advocate for a third approach based on the idea that, when it comes to secondary use, nobody truly owns the data in the traditional sense.
He said that: “Medical data, which are simply recorded observations, are acquired for the purposes of providing patient care. When that care is provided, that purpose is fulfilled, so we need to find another way to think about how these recorded observations should be used for other purposes.
“We believe that patients, provider organisations, and algorithm developers all have ethical obligations to help ensure that these observations are used to benefit future patients, recognising that protecting patient privacy is paramount.”
Research and development
The authors’ framework supports the release of de-identified and aggregated clinical data for research and development, as long as those receiving the data identify themselves and act as ethical data stewards. Individual patient consent would not be required, and patients would not necessarily be able to opt out of allowing their clinical data to be used for research or AI algorithm development – so long as their privacy is protected.
The article states: ‘When used in this manner, clinical data are simply a conduit to viewing fundamental aspects of the human condition. It is not the data, but rather the underlying physical properties, phenomena and behaviours that they represent, that are of primary interest.’
The framework states that it is not ethical for clinical providers to sell clinical data for profit, especially under exclusive arrangements. Corporate entities could profit from AI algorithms developed from clinical data, provided they profit from the activities that they perform rather than from the data itself. In addition, provider organisations could share clinical data with industry partners who financially support their research, if the support is for research rather than for the data.
Safeguards to protect patient privacy include stripping the data of any identifying information.
Dr Larson said: “We strongly emphasise that protection of patient privacy is paramount. The data must be de-identified. In fact, those who receive the data must not make any attempts to re-identify patients through identifying technology.”
Dr Larson and his Stanford colleagues are putting the framework into the public domain for consideration by other individuals and parties, as they navigate the ethical questions surrounding AI and medical data-sharing.