MINE: Mutual Information Neural Estimation
release_ntbynn76avazhlmogbi7tsamp4
by
Mohamed Ishmael Belghazi, Aristide Baratin, Sai Rajeswar, Sherjil
Ozair, Yoshua Bengio, Aaron Courville, R Devon Hjelm
2018
Abstract
We argue that the estimation of mutual information between high dimensional
continuous random variables can be achieved by gradient descent over neural
networks. We present a Mutual Information Neural Estimator (MINE) that is
linearly scalable in dimensionality as well as in sample size, trainable
through back-prop, and strongly consistent. We present a handful of
applications on which MINE can be used to minimize or maximize mutual
information. We apply MINE to improve adversarially trained generative models.
We also use MINE to implement Information Bottleneck, applying it to supervised
classification; our results demonstrate substantial improvement in flexibility
and performance in these settings.
In text/plain
format
Archived Files and Locations
application/pdf 2.2 MB
file_byhj6c5t4bcgvdwegrv6fgml7q
|
arxiv.org (repository) web.archive.org (webarchive) |
1801.04062v3
access all versions, variants, and formats of this works (eg, pre-prints)