Abstract

Synthetic medical images have an important role to play in developing data-driven medical image processing systems. Using a relatively small amount of patient data to train generative models that can produce an abundance of additional samples could bridge the gap towards big-data in niche medical domains. These generative models are evaluated in terms of the synthetic data they generate using the Visual Turing Test (VTT), Fréchet Inception Distance (FID), and other metrics. However, these are generally interpreted at the group level, and do not measure the artificiality of individual synthetic images. The present study attempts to address the challenge of automatically identifying artificial images that are obviously-artificial-looking, which may be necessary for filtering out poorly constructed synthetic images that might otherwise deteriorate the performance of assimilating systems. Synthetic computed tomography (CT) images from a progressively-grown generative adversarial network (PGGAN) were evaluated with a VTT and their image embeddings were analyzed for correlation with artificiality. Images categorized as obviously-artificial (≥0. 7 probability of being rated as fake) were classified using a battery of algorithms. The top-performing classifier, a support vector machine, exhibited accuracy of 75.5%, sensitivity of 0.743, and specificity of 0.769. This is an encouraging result that suggests a potential approach for validating synthetic medical image datasets. Clinical Relevance - Next-generation medical AI systems for image processing will utilize synthetic images produced by generative models. This paper presents an approach towards verifying artificial image legibility for quality-control before being deployed for these purposes.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call