Variational Autoencoder on the MNIST Dataset

I plot digit characterizations in the autoencoder's latent space illustrating that the trained model is able to group (cluster) MNIST images according to their class (label) In the image below, each digit image corresponds to one point. Digits of the same class have the same color.

By: Amr M. Saber

Expertise: Probabilistic learning, variational autoencoders with Bayes, deep learning.

Skills: Julia, Flux, Zygote, MNIST dataset.

This project reproduces and extends the results of Kingma and Welling (2013) implementing a Variational Autoencoder (Auto-Encoding Variational Bayes) to reconstruct MNIST digits. The MNIST dataset contains images of digits from 0 to 9.

Th code for this project is hosted on GitHub at github.com/amrmsab/variational_autoencoder_MNIST.

Benefit 1: Autoencoders can represent the figures in a reduced latent space that allows us to visualize the distinction between the digits' images.

Benefit 2: Autoencoders can help us generate more digit images.

I use this latent space representation to see what `interpolating’ between numbers looks like. For example, we generate numbers between 0 and 5.

Benefit 3: Autoencoders can help us reconstruct incomplete images.

I reconstruct a digit image given only its top half. Below is the whole image of which we give only the top half to the model. The model then reconstructs the bottom half..