Please cite as: Hannah van Kolfschooten, ‘Artificial Intelligence in Radiology: Safeguarding Patients’ Rights in the Digital Era’, European Radiology 2026.
Artificial intelligence (AI) is now firmly embedded in radiology practice. From automated abnormality detection on chest radiographs to workflow optimisation in triage, AI is increasingly shaping diagnostic processes. Its promise is substantial: improved efficiency, faster reporting, and better diagnostic accuracy. Yet these benefits come with risks that extend beyond technical performance. For radiologists, the critical challenge is to ensure that integration of AI into clinical practice does not compromise patients’ fundamental rights.
This is not only a professional duty, but also a legal obligation. Under the recently adopted EU Artificial Intelligence Act (AI Act), hospitals and clinicians deploying “high-risk” AI systems, including most radiology applications, are required to demonstrate adequate AI literacy.[1][2] Radiologists must be equipped not only to understand technical aspects but also to assess risks for patients.
Bias and beyond: multiple risks to patients
Recent studies in breast cancer screening illustrate one well-known problem: the performance of AI systems can vary according to patient characteristics such as age, race/ethnicity, and breast density.[3] Women from minority backgrounds and older women are more likely to receive false-positive results compared with white, middle-aged women.[4] These disparities reflect biases in training datasets and design choices, raising the possibility that AI amplifies existing health inequities.
But bias is only one part of the story. Even a technically “accurate” AI system may still create new risks for patients if its functioning undermines autonomy, obscures information, compromises privacy, or blurs accountability.[5] Radiologists need to think about AI not just in terms of performance metrics, but in terms of how it interacts with the rights and expectations of patients.[6]
Patients’ rights under pressure
AI integration in radiology places five established patients’ rights at risk:
- Right of access to care: If algorithms perform less reliably for certain groups, patients face unequal diagnostic quality. For example, studies of AI in mammography have shown higher false-positive rates in women with dense breast tissue or in Black women compared with white women of the same age.[3][4]
- Right to autonomy: Informed consent is weakened if AI-driven outputs are accepted without explanation or clinician oversight. A patient undergoing CT lung cancer screening may never know an AI algorithm flagged or dismissed a nodule, leaving no meaningful opportunity for informed decision-making.[7]
- Right to information: Black-box systems make it difficult for radiologists to explain to patients how a diagnosis was reached. For example, heatmaps may highlight suspicious areas without justifying why, leaving radiologists unable to fully explain results.[8]
- Right to privacy: The development of radiology AI depends on massive datasets of CT, MRI, and X-ray images. Hospitals face mounting pressure to pool or commercialize these datasets, sometimes through partnerships with technology companies. This raises concerns about genuine consent, reuse of data beyond care (e.g., insurance or marketing), and cross-border data safeguards.[9]
- Right to redress: When harm occurs, accountability is unclear. Imagine a case where an AI algorithm fails to detect a cerebral aneurysm on CTA, leading to delayed treatment. Was the error the responsibility of the radiologist for relying on the tool, the hospital for approving its use, or the developer for deficiencies in training data? Without transparency and clearly defined liability frameworks, patients may find it nearly impossible to pursue compensation.[10]
European regulation: progress but not yet enough
In Europe, several legal instruments govern AI in healthcare, but their protections remain fragmented. The General Data Protection Regulation (GDPR) provides safeguards against automated decision-making, yet its application to clinical diagnostic processes remains ambiguous.[10] The Medical Device Regulation (MDR) focuses on product safety and performance but does not directly address patients’ rights such as autonomy or redress.[11] The recently adopted AI Act introduces new requirements for “high-risk” AI systems, including medical imaging applications, but its provisions remain general. They create no new enforceable rights for patients affected by algorithmic decision-making. Current frameworks tend to view AI risks through a narrow lens of product conformity and safety, rather than the experiences and rights of patients.[1]
Towards stronger safeguards in radiology practice
Radiologists are uniquely positioned at the interface between patients, technology, and regulation. To ensure that AI strengthens rather than weakens patient trust, three complementary safeguards are essential.
First, a European Charter of Digital Patients’ Rights should clarify rights for the digital era, such as the right to explainability, human oversight, and protection against automated discrimination.[12] This also means that patients should be involved in the entire AI lifecycle, from design to post-market surveillance.[13] Radiologists, as the link between AI medical imaging systems and patients, play a key role in putting patients’ rights into practice.
Second, professional guidelines for AI in radiology are needed. Beyond law, professional societies should provide practical guidance. This could cover when and how AI outputs should be recorded in medical records, how radiologists should address false-positive results, and what information must be shared with referring physicians, radiographers, and patients. Radiologists’ direct involvement in guideline development, together with ethicists and legal experts, will help ensure that safeguards are tailored to clinical realities.[14][15]
Third, continuous education and AI literacy are crucial. The EU AI Act explicitly requires that deployers of high-risk AI systems, including hospitals and clinicians, ensure adequate AI literacy. For radiologists, this means structured training not only in technical dimensions but also in ethical and legal implications. Embedding AI literacy into training at all levels will ensure radiologists remain equipped to use AI responsibly.[16]
Together, these measures can help ensure that AI strengthens patient trust and safeguards radiology’s commitment to quality and safety.
Conclusion
The expansion of AI in radiology represents both an opportunity and a responsibility. Bias is an important challenge, but not the only one: opacity, threats to autonomy, data pressures, and unclear accountability all place patients’ rights under strain. By proactively addressing this broader spectrum of ethical and legal risks, radiologists can ensure that technological innovation translates into safer, more equitable patient care. Rather than viewing patients’ rights and AI as opposing forces, we should treat them as mutually reinforcing: only by safeguarding rights can AI deliver on its promise to improve radiology for all patients.
References
1. van Kolfschooten, H., & van Oirschot, J. (2024) The EU Artificial Intelligence Act (2024): Implications for healthcare. Health Policy 149, 105152. DOI: 10.1016/j.healthpol.2024.105152
2. Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 laying down harmonised rules on artificial intelligence and amending Regulations (EC) No 300/2008, (EU) No 167/2013, (EU) No 168/2013, (EU) 2018/858, (EU) 2018/1139 and (EU) 2019/2144 and Directives 2014/90/EU, (EU) 2016/797 and (EU) 2020/1828 (Artificial Intelligence Act) (Text with EEA relevance), Article 5. http://data.europa.eu/eli/reg/2024/1689/oj
3. Nguyen DL, Ren Y, Jones TM, Thomas SM, Lo JY, Grimm LJ (2024) Patient characteristics impact performance of AI algorithm in interpreting negative screening digital breast tomosynthesis studies. Radiology. DOI: 10.1148/radiol.232286
4. Net, JM & Collado-Mesa F (2025) Bias in Artificial Intelligence: Impact on Breast Imaging. Journal of Breast Imaging 7, 4. DOI: 10.1093/jbi/wbaf027
5. van Kolfschooten, H. (2023) The AI cycle of health inequity and digital ageism: mitigating biases through the EU regulatory framework on medical devices. Journal of Law and the Biosciences 10, 2. DOI: 10.1093/jlb/lsad031
6. Klontzas ME, Groot Lipman KBW, Akinci D’Antonoli T et al (2025) ESR Essentials: common performance metrics in AI – practice recommendations by the European Society of Medical Imaging Informatics. Eur Radiol. DOI: 10.1007/s00330-025-11890-w
7. Chau, M, Rahman MG, Debnath T (2025) From black box to clarity: Strategies for effective AI informed consent in healthcare. Artificial Intelligence in Medicine 167. DOI: 10.1016/j.artmed.2025.103169
8. De-Giorgio F, Benedetti B, Mancino M, Sala E, Pascali VL. (2025) The need for balancing ’black box’ systems and explainable artificial intelligence: A necessary implementation in radiology. Eur Radiol. DOI: 10.1016/j.ejrad.2025.112014
9. Murdoch, B. (2021) Privacy and artificial intelligence: challenges for protecting health information in a new era. BMC Med Ethics 22, 122. DOI: 10.1186/s12910-021-00687-3
10. van Kolfschooten, H. (2025) Towards an EU Charter of Digital Patients’ Rights in the Age of Artificial Intelligence. Digit. Soc. 4, 6. DOI: 10.1007/s44206-025-00159-w
11. Regulation (EU) 2016/679 of the European Parliament and of the Council of 27 April 2016 on the protection of natural persons with regard to the processing of personal data and on the free movement of such data, and repealing Directive 95/46/EC (General Data Protection Regulation) (Text with EEA relevance). http://data.europa.eu/eli/reg/2016/679/oj. NB: the UK has similar data protection laws, see: https://www.gov.uk/data-protection
12. Regulation (EU) 2017/745 of the European Parliament and of the Council of 5 April 2017 on medical devices, amending Directive 2001/83/EC, Regulation (EC) No 178/2002 and Regulation (EC) No 1223/2009 and repealing Council Directives 90/385/EEC and 93/42/EEC (Text with EEA relevance.) http://data.europa.eu/eli/reg/2017/745/oj. NB: the UK has similar medical devices laws, see: https://www.gov.uk/guidance/regulating-medical-devices-in-the-uk
13. Stogiannos N, Gillan C, Precht H et al (2024) A multidisciplinary team and multiagency approach for AI implementation: A commentary for medical imaging and radiotherapy key stakeholders. J Med Imaging Radiat Sci. DOI: 10.1016/j.jmir.2024.101717
14. de Almeida JG, Messiou C, Withey SJ et al (2025) Medical machine learning operations: a framework to facilitate clinical AI development and deployment in radiology. Eur Radiol. DOI: 10.1007/s00330-025-11654-6
15. Kotter E, Akinsi D’Antonoli T, Cuocolo R et al (2025) Guiding AI in radiology: ESR’s recommendations for effective implementation of the European AI Act. Insighs into Imaging. DOI: 10.1186/s13244-025-01905-x
16. Perchik JD, Smith AD, Elkassem AA, Park JM, Rothenberg SA, Tanwar M, Yi PH, Sturdivant A, Tridandapani S, Sotoudeh H (2023) Artificial intelligence literacy: developing a multi-institutional infrastructure for AI education. Acad Radiol 30. DOI: 10.1016/j.acra.2022.10.002

