Abstract

Unsupervised learning is of growing interest because it unlocks the potential held in vast amounts of unlabeled data to learn useful representations for inference. Autoencoders, a form of generative model, may be trained by learning to reconstruct unlabeled input data from a latent representation space. More robust representations may be produced by an autoencoder if it learns to recover clean input samples from corrupted ones. Representations may be further improved by introducing regularization during training to shape the distribution of the encoded data in the latent space. We suggest denoising adversarial autoencoders (AAEs), which combine denoising and regularization, shaping the distribution of latent space using adversarial training. We introduce a novel analysis that shows how denoising may be incorporated into the training and sampling of AAEs. Experiments are performed to assess the contributions that denoising makes to the learning of representations for classification and sample synthesis. Our results suggest that autoencoders trained using a denoising criterion achieve higher classification performance and can synthesize samples that are more consistent with the input data than those trained without a corruption process.

Highlights

  • M ODELING and drawing data samples from complex, high-dimensional distributions are challenging

  • Results show that the DAAE and integrating denoising AAE (iDAAE) outperform the adversarial autoencoders (AAEs) on the classification task

  • The DAAE and iDAAE outperform a classifier trained on encodings obtained by applying principle component’s analysis (PCA) to the image samples, while the AAE does not, further showing the benefits of using denoising

Read more

Summary

INTRODUCTION

M ODELING and drawing data samples from complex, high-dimensional distributions are challenging. Two broad approaches to learning the state-of-the-art generative autoencoders that do not require labeled training data include: 1) introduction of a denoising criterion [5], [30], [31], where the model learns to reconstruct clean samples from corrupted ones and 2) regularization of the latent space to match a prior [14], [20]; for the latter, the priors take a simple form, such as multivariate normal distributions. When a denoising criterion is introduced to an adversarial autoencoder (AAE), we have a choice to either shape the conditional distribution of latent variables given corrupted samples to match the prior (as was done using a variational approach [12]) or to shape the full posterior conditional on the original data samples to match the prior.

Autoencoders
Denoising Autoencoders
Variational Autoencoders
Denoising Variational Autoencoders
Adversarial Training
Adversarial Autoencoders
DENOISING ADVERSARIAL AUTOENCODER
Construction
SYNTHESIZING NOVEL SAMPLES
Drawing Samples From Autoencoders
Proposed Method for Sampling DAAEs
IMPLEMENTATION
Autoencoder
Sampling
Data Sets
Architecture and Training
Sampling DAAEs and iDAAEs
Reconstruction
Classification
Tradeoffs in Performance
VIII. CONCLUSION
Omniglot
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call