Estimating Information Theoretic Measures via Multidimensional Gaussianization

Valero Laparra*, Juan Emmanuel Johnson, Gustau Camps-Valls, Raul Santos-Rodriguez, Jesus Malo

*Corresponding author for this work

Research output: Contribution to journalArticle (Academic Journal)peer-review

3 Citations (Scopus)

Abstract

Information theory is an outstanding framework for measuring uncertainty, dependence, and relevance in data and systems. It has several desirable properties for real-world applications: naturally deals with multivariate data, can handle heterogeneous data, and the measures can be interpreted. However, it has not been adopted by a wider audience because obtaining information from multidimensional data is a challenging problem due to the curse of dimensionality. We propose an indirect way of estimating information based on a multivariate iterative Gaussianization transform. The proposed method has a multivariate-to-univariate property: it reduces the challenging estimation of multivariate measures to a composition of marginal operations applied in each iteration of the Gaussianization. Therefore, the convergence of the resulting estimates depends on the convergence of well-understood univariate entropy estimates, and the global error linearly depends on the number of times the marginal estimator is invoked. We introduce Gaussianization-based estimates for Total Correlation, Entropy, Mutual Information, and Kullback-Leibler Divergence. Results on artificial data show that our approach is superior to previous estimators, particularly in high-dimensional scenarios. We also illustrate the method's performance in different fields to obtain interesting insights. We make the tools and datasets publicly available to provide a test bed for analyzing future methodologies.
Original languageEnglish
Pages (from-to)1293-1308
Number of pages16
JournalIEEE Transactions on Pattern Analysis and Machine Intelligence
Volume47
Issue number2
Early online date11 Nov 2024
DOIs
Publication statusPublished - 1 Feb 2025

Bibliographical note

Publisher Copyright:
© 1979-2012 IEEE.

Keywords

  • and learning in neural networks
  • computer vision
  • entropy
  • Gaussianization
  • geoscience
  • information bottleneck
  • Information theory
  • Kullback-Leibler divergence
  • multivariate data
  • mutual information
  • total correlation
  • visual neuroscience

Fingerprint

Dive into the research topics of 'Estimating Information Theoretic Measures via Multidimensional Gaussianization'. Together they form a unique fingerprint.

Cite this