Abstract

The rise of neuroimaging in research and clinical practice, together with the development of new machine learning techniques has strongly encouraged the Computer Aided Diagnosis (CAD) of different diseases and disorders. However, these algorithms are often tested in proprietary datasets to which the access is limited and, therefore, a direct comparison between CAD procedures is not possible. Furthermore, the sample size is often small for developing accurate machine learning methods. Multi-center initiatives are currently a very useful, although limited, tool in the recruitment of large populations and standardization of CAD evaluation. Conversely, we propose a brain image synthesis procedure intended to generate a new image set that share characteristics with an original one. Our system focuses on nuclear imaging modalities such as PET or SPECT brain images. We analyze the dataset by applying PCA to the original dataset, and then model the distribution of samples in the projected eigenbrain space using a Probability Density Function (PDF) estimator. Once the model has been built, we can generate new coordinates on the eigenbrain space belonging to the same class, which can be then projected back to the image space. The system has been evaluated on different functional neuroimaging datasets assessing the: resemblance of the synthetic images with the original ones, the differences between them, their generalization ability and the independence of the synthetic dataset with respect to the original. The synthetic images maintain the differences between groups found at the original dataset, with no significant differences when comparing them to real-world samples. Furthermore, they featured a similar performance and generalization capability to that of the original dataset. These results prove that these images are suitable for standardizing the evaluation of CAD pipelines, and providing data augmentation in machine learning systems -e.g. in deep learning-, or even to train future professionals at medical school.

Highlights

  • With the rise of neuroimaging in research and practice and the development of the machine learning paradigm, there has been an exponential trend in computer-aided methodologies (Frisoni et al, 2010; Martinez-Murcia et al, 2016; Rathore et al, 2017)

  • We have proposed a novel brain synthesis algorithm that could be used, among other things, in standardizing evaluation of Computer Aided Diagnosis (CAD) systems or as a data augmentation procedure

  • The algorithm consists of an analysis of a existing dataset using Principal Component Analysis, building a space defined by these principal components, or eigenbrains

Read more

Summary

Introduction

With the rise of neuroimaging in research and practice and the development of the machine learning paradigm, there has been an exponential trend in computer-aided methodologies (Frisoni et al, 2010; Martinez-Murcia et al, 2016; Rathore et al, 2017). Multi-center initiatives, such as the Alzheimer’s Disease Neuroimaging Initiative (ADNI) (Weiner et al, 2012), the Autism Brain Imaging Data Exchange (ABIDE) (Di Martino et al, 2014) or the Parkinson’s Progressions Markers Initiative (PPMI) (Marek et al, 2011) provide large image samples that reduce the type I error and ease a direct comparison between systems. Inhomogeneities in scanner, population, or techniques might cause the apparition of false positives that are not related to the signal (Pearlson, 2009; Van Horn and Toga, 2009)

Objectives
Methods
Results
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call