Can you spot a ‘deepfake’ x-ray?

A study published March 24 in Radiology bears disturbing news: Radiologists are not able to easily distinguish AI-generated “deepfake” x-ray images from authentic ones. 

In an interview with AuntMinnie, lead author Mickael Tordjman, MD, of the Icahn School of Medicine at Mount Sinai in New York, explained that the study involved 17 radiologists from 12 different centers in six countries. They were tested on two distinct image sets: first on a dataset of real and ChatGPT-4o-generated images of multiple anatomical regions and secondly on a dataset of half authentic chest x-rays and half created by RoentGen, an open-source generative AI model developed by Stanford Medicine researchers.

Anatomy-matched real and GPT-4o-generated radiographs: (A) real and (B) GPT-4o-generated posteroanterior chest radiographs, (C) real and (D) GPT-4ogenerated lateral cervical spine radiographs, (E) real and (F) GPT-4o-generated posteroanterior hand radiographs, and (G) real and (H) GPT-4o-generated lateral lumbar spine radiographs. The pairs demonstrate that GPT-4o can produce radiographically plausible images across different anatomic regions.Anatomy-matched real and GPT-4o-generated radiographs: (A) real and (B) GPT-4o-generated posteroanterior chest radiographs, (C) real and (D) GPT-4ogenerated lateral cervical spine radiographs, (E) real and (F) GPT-4o-generated posteroanterior hand radiographs, and (G) real and (H) GPT-4o-generated lateral lumbar spine radiographs. The pairs demonstrate that GPT-4o can produce radiographically plausible images across different anatomic regions. RSNAThe findings expose a significant cybersecurity risk, Tordjman said. If hackers were to gain access to a hospital’s network, they could inject synthetic images to manipulate patient diagnoses or cause widespread clinical chaos by undermining the fundamental reliability of the digital medical record, he noted.

What’s needed are tools to potentially help clinicians identify such deepfakes, Tordjman said. To that end, the study also identified common features of synthetic x-rays, and the authors have published a curated deepfake dataset with interactive quizzes for educational purposes. 

"Deepfake medical images often look too perfect,” Tordjman said.

Potential solutions to clearly distinguish real and fake images and help prevent tampering include implementing advanced digital safeguards, such as invisible watermarks that embed ownership or identity data directly into the images and automatically attaching technologist-linked cryptographic signatures when the images are captured. 

“We are potentially only seeing the tip of the iceberg,” Tordjman noted. “The logical next step in this evolution is AI-generation of synthetic 3D images, such as CT and MRI. Establishing educational datasets and detection tools now is critical.” 

In an accompanying editorial, Rajesh Bhayana, MD, and Satheesh Krishna, MD, both of the University of Toronto, noted that the study demonstrates that reliance on visual inspection alone may no longer be sufficient for verifying image authenticity. The study is not an indictment of AI, however, but a call to augment imaging systems, they wrote. 

“As the boundary between the authentic and the synthetic blurs, we must ensure that trust is built into the system itself. In the age of generative AI, seeing is no longer believing,” Bhayana and Krishna concluded. 

The full study is available here

 

Page 1 of 385
Next Page