Abstract

Deep neural networks are good at extracting low-dimensional subspaces (latent spaces) that represent the essential features inside a high-dimensional dataset. Deep generative models represented by variational autoencoders (VAEs) can generate and infer high-quality datasets, such as images. In particular, VAEs can eliminate the noise contained in an image by repeating the mapping between latent and data space. To clarify the mechanism of such denoising, we numerically analyzed how the activity pattern of trained networks changes in the latent space during inference. We considered the time development of the activity pattern for specific data as one trajectory in the latent space and investigated the collective behavior of these inference trajectories for many data. Our study revealed that when a cluster structure exists in the dataset, the trajectory rapidly approaches the center of the cluster. This behavior was qualitatively consistent with the concept retrieval reported in associative memory models. Additionally, the larger the noise contained in the data, the closer the trajectory was to a more global cluster. It was demonstrated that by increasing the number of the latent variables, the trend of the approach a cluster center can be enhanced, and the generalization ability of the VAE can be improved.

Highlights

  • Deep neural networks are good at extracting low-dimensional subspaces that represent the essential features inside a high-dimensional dataset

  • We numerically analyzed the dynamics of repeated infernce of variational autoencoders (VAEs)

  • We evaluated the denoising behavior of a VAE trained with Modified National Institute of Standards and Technology (MNIST) dataset for noisy inputs

Read more

Summary

Introduction

Deep neural networks are good at extracting low-dimensional subspaces (latent spaces) that represent the essential features inside a high-dimensional dataset. VAEs can eliminate the noise contained in an image by repeating the mapping between latent and data space. Compared to generating an output image from latent space to smoothly morph one image into another, repeating inferences several times improves the quality of the output i­mage[11] Most of these studies only qualitatively evaluate output data through a one-shot inference from the latent space to output data. The deep generative models extract a low-dimensional subspace in such a high-dimensional space by nonlinear mapping using neural networks. Because various factors, such as noise in real environments, cause original data points to deviate from this low-dimensional subspace, we are Scientific Reports | (2020) 10:16001.

Objectives
Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call