Signal-to-noise ratio in reproducing kernel Hilbert spaces

Luis Gómez-Chova*, Raúl Santos-Rodríguez, Gustau Camps-Valls

*Corresponding author for this work

Research output: Contribution to journalArticle (Academic Journal)

1 Citation (Scopus)
127 Downloads (Pure)

Abstract

This paper introduces the kernel signal-to-noise ratio (kSNR) for different machine learning and signal processing applications. The kSNR seeks to maximize the signal variance while minimizing the estimated noise variance explicitly in a reproducing kernel Hilbert space (rkHs). The kSNR gives rise to considering complex signal-to-noise relations beyond additive noise models, and can be seen as a useful regularizer for feature extraction and dimensionality reduction. We show that the kSNR generalizes kernel PCA (and other spectral dimensionality reduction methods), least squares SVM, and kernel ridge regression to deal with cases where signal and noise cannot be assumed independent. We give computationally efficient alternatives based on reduced-rank Nyström and projection on random Fourier features approximations, and analyze the bounds of performance and its stability. We illustrate the method through different examples, including nonlinear regression, nonlinear classification in channel equalization, nonlinear feature extraction from high-dimensional spectral satellite images, and bivariate causal inference. Experimental results show that the proposed kSNR yields more accurate solutions and extracts more noise-free features when compared to standard approaches.

Original languageEnglish
Pages (from-to)75-82
Number of pages8
JournalPattern Recognition Letters
Volume112
Early online date5 Jun 2018
DOIs
Publication statusPublished - 1 Sep 2018

    Fingerprint

Keywords

  • Causal inference
  • Feature extraction
  • Heteroscedastic
  • Kernel methods
  • Noise model
  • Signal classification
  • Signal-to-noise ratio
  • SNR

Cite this