Abstract
Information theory is an outstanding framework for measuring uncertainty, dependence, and relevance in data and systems. It has several desirable properties for real-world applications: naturally deals with multivariate data, can handle heterogeneous data, and the measures can be interpreted. However, it has not been adopted by a wider audience because obtaining information from multidimensional data is a challenging problem due to the curse of dimensionality. We propose an indirect way of estimating information based on a multivariate iterative Gaussianization transform. The proposed method has a multivariate-to-univariate property: it reduces the challenging estimation of multivariate measures to a composition of marginal operations applied in each iteration of the Gaussianization. Therefore, the convergence of the resulting estimates depends on the convergence of well-understood univariate entropy estimates, and the global error linearly depends on the number of times the marginal estimator is invoked. We introduce Gaussianization-based estimates for Total Correlation, Entropy, Mutual Information, and Kullback-Leibler Divergence. Results on artificial data show that our approach is superior to previous estimators, particularly in high-dimensional scenarios. We also illustrate the method's performance in different fields to obtain interesting insights. We make the tools and datasets publicly available to provide a test bed for analyzing future methodologies.
| Original language | English |
|---|---|
| Pages (from-to) | 1293-1308 |
| Number of pages | 16 |
| Journal | IEEE Transactions on Pattern Analysis and Machine Intelligence |
| Volume | 47 |
| Issue number | 2 |
| Early online date | 11 Nov 2024 |
| DOIs | |
| Publication status | Published - 1 Feb 2025 |
Bibliographical note
Publisher Copyright:© 1979-2012 IEEE.
Keywords
- and learning in neural networks
- computer vision
- entropy
- Gaussianization
- geoscience
- information bottleneck
- Information theory
- Kullback-Leibler divergence
- multivariate data
- mutual information
- total correlation
- visual neuroscience
Fingerprint
Dive into the research topics of 'Estimating Information Theoretic Measures via Multidimensional Gaussianization'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver