Invertible Gaussian Reparameterization: Revisiting the Gumbel-Softmax
December 19, 2019 Β· Entered Twilight Β· π Neural Information Processing Systems
"Last commit was 5.0 years ago (β₯5 year threshold)"
Evidence collected by the PWNC Scanner
Repo contents: Models, README.md, Tests, Utils, approximations, igr_singularity.def, rebar_toy.py, requirements.txt, run_toy.py, structure_output_prediction, vae_experiments
Authors
Andres Potapczynski, Gabriel Loaiza-Ganem, John P. Cunningham
arXiv ID
1912.09588
Category
stat.ML: Machine Learning (Stat)
Cross-listed
cs.LG
Citations
40
Venue
Neural Information Processing Systems
Repository
https://github.com/cunningham-lab/igr
β 30
Last Checked
2 months ago
Abstract
The Gumbel-Softmax is a continuous distribution over the simplex that is often used as a relaxation of discrete distributions. Because it can be readily interpreted and easily reparameterized, it enjoys widespread use. We propose a modular and more flexible family of reparameterizable distributions where Gaussian noise is transformed into a one-hot approximation through an invertible function. This invertible function is composed of a modified softmax and can incorporate diverse transformations that serve different specific purposes. For example, the stick-breaking procedure allows us to extend the reparameterization trick to distributions with countably infinite support, thus enabling the use of our distribution along nonparametric models, or normalizing flows let us increase the flexibility of the distribution. Our construction enjoys theoretical advantages over the Gumbel-Softmax, such as closed form KL, and significantly outperforms it in a variety of experiments. Our code is available at https://github.com/cunningham-lab/igr.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β Machine Learning (Stat)
R.I.P.
π»
Ghosted
R.I.P.
π»
Ghosted
Distilling the Knowledge in a Neural Network
R.I.P.
π»
Ghosted
Layer Normalization
R.I.P.
π»
Ghosted
Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning
R.I.P.
π»
Ghosted
Domain-Adversarial Training of Neural Networks
R.I.P.
π»
Ghosted