Deepfake X-rays fool radiologists 75% of the time, raising alarm over medical fraud and cybersecurity

Reviewed byNidhi Govil

4 Sources

Share

A new study published in Radiology reveals that radiologists can correctly distinguish real from AI-generated medical X-rays only 75% of the time, even when aware synthetic images are present. The research, involving 17 radiologists from six countries, found that 41% initially suspected deepfake infiltration. Large language models like ChatGPT and Gemini performed similarly poorly, achieving 57-85% accuracy, highlighting urgent concerns about fraudulent litigation, hospital cybersecurity, and research integrity.

News article

Deepfake X-rays Deceive Expert Radiologists Despite Awareness

Radiologists struggle to identify AI-generated medical X-rays, correctly distinguishing between real and AI-generated X-rays only 75% of the time on average, according to a study published March 24 in Radiology, the journal of the Radiological Society of North America

1

2

. The research involved 17 radiologists from 12 research centers across six countries—the United States, France, Germany, Turkey, the United Kingdom, and the United Arab Emirates—who reviewed 264 X-ray images split evenly between authentic scans and deepfake X-rays

3

. When participants were initially unaware of the study's purpose and asked about technical quality, only 41% spontaneously raised concerns that AI-generated medical images might have infiltrated the dataset

2

.

Lead study author Mickael Tordjman, a radiologist at the Icahn School of Medicine at Mount Sinai in New York, emphasized the severity of these findings. "Our study demonstrates that these deepfake X-rays are realistic enough to deceive radiologists, the most highly trained medical image specialists, even when they were aware that AI-generated images were present," he stated

3

. Individual performance varied widely, with radiologists correctly identifying ChatGPT-generated images between 58% and 92% of the time

2

. Importantly, professional experience ranging from zero to 40 years showed no correlation with detection accuracy, though musculoskeletal radiologists demonstrated significantly higher accuracy than other subspecialists

3

.

Large Language Models Struggle With Detection Too

The research team tested whether large language models (LLMs) might perform better than human experts at identifying synthetic medical images. Four multimodal LLMs—GPT-4o, GPT-5 from OpenAI, Gemini 2.5 Pro from Google, and Llama 4 Maverick from Meta—achieved accuracy rates ranging from 57% to 85% when evaluating ChatGPT-generated images

2

4

. Notably, even ChatGPT-4o, the model used to create the deepfake images, could not accurately detect all of them, though it performed better than competing models

3

. For RoentGen-generated chest X-rays—created using an open-source generative AI diffusion model developed by Stanford Medicine researchers—radiologists achieved accuracy rates between 62% and 78%, while AI models ranged from 52% to 89%

2

.

Risks of Deepfake Medical Images Span Multiple Domains

The implications extend far beyond diagnostic accuracy. Tordjman warned of high-stakes vulnerabilities for fraudulent litigation, where fabricated fractures could be indistinguishable from real ones

2

. Hospital cybersecurity threats represent another critical concern—if hackers gain access to a hospital's network, they could inject synthetic images to manipulate patient diagnoses or cause widespread clinical chaos by undermining the fundamental reliability of digital medical records

3

. Elisabeth Bik, a microbiologist and image-integrity specialist, told Nature that the results were "both disturbing and not very surprising," raising concerns for research integrity, clinical workflows, insurance claims, and legal contexts where imaging evidence is used

1

4

.

Another emerging threat involves training data contamination. Siwei Lyu, who researches media forensics at the University at Buffalo, warned that diagnostic AI models could become distorted if AI-generated data infiltrates their training datasets, causing models to "latch onto features that are not exactly relevant to real medical cases, but are purely artefacts of generative AI models"

1

.

Identifying Unnaturally Perfect Anatomical Features

Researchers identified several visual clues that can appear in synthetic images. "Deepfake medical images often look too perfect," Tordjman explained

2

. Common characteristics include bones that are overly smooth, spines unnaturally straight, lungs overly symmetrical, blood vessel patterns excessively uniform, and fractures that appear unusually clean and consistent, often limited to one side of the bone

3

. However, the study suggests that visual inspection alone is no longer a reliable safeguard

4

.

Digital Watermarking and Cryptographic Signatures as Countermeasures

To combat these threats, researchers recommend implementing advanced digital safeguards. These include invisible watermarks that embed ownership or identity data directly into images and cryptographic signatures automatically attached and linked to the technologist at the time of image capture

3

4

. The study authors have created an interactive quiz designed to teach researchers how to discern between AI-generated and real X-ray scans, along with a curated deepfake dataset for educational purposes

1

2

.

Tordjman cautioned that current challenges may represent only the beginning. "We are potentially only seeing the tip of the iceberg," he said, pointing to AI-generation of synthetic 3D medical images such as CT and MRI as the logical next step

2

4

. Establishing educational datasets and detection tools now is critical before these threats become even harder to manage, he emphasized.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo