See an extension of this work https://github.com/51616/split-vae
(https://arxiv.org/abs/1611.02648) (version 1) By Nat Dilokthanakul, Pedro A.M. Mediano, Marta Garnelo, Matthew C.H. Lee, Hugh Salimbeni, Kai Arulkumaran, Murray Shanahan
We study a variant of the variational autoencoder model with a Gaussian mixture as a prior distribution, with the goal of performing unsupervised clustering through deep generative models. We observe that the standard variational approach in these models is unsuited for unsupervised clustering, and mitigate this problem by leveraging a principled information-theoretic regularisation term known as consistency violation. Adding this term to the standard variational optimisation objective yields networks with both meaningful internal representations and well-defined clusters. We demonstrate the performance of this scheme on synthetic data, MNIST and SVHN, showing that the obtained clusters are distinct, interpretable and result in achieving higher performance on unsupervised clustering classification than previous approaches.
Luarocks packages:
- mnist ( torch-rocks install https://raw.github.com/andresy/mnist/master/rocks/mnist-scm-1.rockspec )
Python packages:
- torchfile (pip install torchfile)
To run a spiral experiment,
./run.sh spiral
To visualise spiral experiment (can be used while training)
cd plot
python plot_latent.py
python plot_recon.py
To turn off CV regularisation term (set alpha)
set flag
./run.sh spiral -cvWeight 0.0
To turn off z-prior term (set eta)
./run.sh spiral -zPriorWeight 0.0
To run MNIST
./run.sh mnist
To use GPU
set flag
./run.sh mnist -gpu 1
To run quick MNIST on fully-connected network
./run.sh mnist_fc
The neural network components are constructed with nngraph in the folder 'models'.
The network's components are put together in the main.lua file at line 126.
During optimisation, the GMVAE objectives (5 contributing terms) are calculated from line 202 until 250 in main.lua.
I would like to thanks the following, whose github's repos were used as inital templates for implementing this idea.
-
Rui Shu https://github.com/RuiShu/cvae
-
Joost van Amersfoort https://github.com/y0ast/VAE-Torch
-
Kai Arulkumaran https://github.com/Kaixhin/Autoencoders