Risks to Healthcare Data Integrity
Deepfake X-rays Fool Doctors and AI Models
Experts and advanced AI struggle to distinguish real X-ray images from deepfakes, exposing healthcare risks.

A chest X-ray on a computer screen showing subtle digital glitches and binary code, representing a deepfake medical image in a clinical setting.
Photo: Avantgarde News
A study published in the journal Radiology reveals that experts cannot reliably identify deepfake medical images [1]. Researchers tested experienced radiologists and cutting-edge large language models against AI-generated X-rays [2]. Neither the humans nor the digital models could consistently tell the difference between authentic scans and deepfakes [3]. These findings expose serious vulnerabilities in healthcare data security [1]. While AI technology offers clinical benefits, it also enables the creation of highly realistic fraudulent records [2]. Experts suggest that protecting the integrity of medical data is now a critical challenge for the industry [3].
Editorial notes
Transparency note
Drafted with LLM; human-edited
- AI assisted
- Yes
- Human review
- Yes
- Last updated
Risk assessment
Reviewed for sourcing quality and editorial consistency.
Sources
Related stories
View allTopics
About the author
Avantgarde News Desk covers risks to healthcare data integrity and editorial analysis for Avantgarde News.


