3260 papers • 126 benchmarks • 313 datasets
To estimate mutual information from samples, specially for high-dimensional variables.
(Image credit: Papersgraph)
These leaderboards are used to track progress in mutual-information-estimation-5
No benchmarks available.
Use these libraries to find mutual-information-estimation-5 models and implementations
No datasets available.
No subtasks available.
It is shown that structure matters: incorporating knowledge about locality in the input into the objective can significantly improve a representation’s suitability for downstream tasks and is an important step towards flexible formulations of representation learning objectives for specific end-goals.
This work highlights the potential of KR to advance the field of graph representation learning and enhance the performance of GNNs, using KR loss as the primary loss in self-supervised settings or as a regularization term in supervised settings.
This paper designs a novel estimator for mutual information of discrete-continuous mixtures and proves that the proposed estimator is consistent, and provides numerical experiments suggesting superiority of this estimator compared to other heuristics.
EDGE is the first non-parametric MI estimator that can achieve paramet- ric MSE rates with linear time complexity and the utility of EDGE is illustrated for the analysis of the information plane (IP) in deep learning.
This work proposes and study an estimator that can be easily implemented, works well in high dimensions, and enjoys faster rates of convergence, and discusses its direct implications for total correlation, entropy, and mutual information estimation.
This work uses a recently proposed neural estimator of mutual information to optimize the encoder for a maximized mutual information, only relying on channel samples, and shows that this approach achieves the same performance as state-of-the-art end-to-end learning with perfect channel model knowledge.
A novel regularizer is developed to improve the learning of long-range dependency of sequence data and shows how the `next sentence prediction (classification)' heuristic can be derived in a principled way from the authors' mutual information estimation framework, and be further extended to maximize the mutual information of sequence variables.
It is argued that the seemingly redundant intermediate step of entropy estimation allows one to improve the convergence by an appropriate reference distribution and faster convergence is possible by choosing the uniform distribution as the reference distribution instead.
This paper introduces an estimator for KL-Divergence based on the likelihood ratio by training a classifier to distinguish the observed joint distribution from the product distribution and shows how to construct several CMI estimators using this basic divergence estimator by drawing ideas from conditional generative models.
Adding a benchmark result helps the community track progress.