Generalization Bounds via Information Density and Conditional Information Density

Type: Article

Publication Date: 2020-11-01

Citations: 27

DOI: https://doi.org/10.1109/jsait.2020.3040992

Abstract

We present a general approach, based on an exponential inequality, to derive bounds on the generalization error of randomized learning algorithms. Using this approach, we provide bounds on the average generalization error as well as bounds on its tail probability, for both the PAC-Bayesian and single-draw scenarios. Specifically, for the case of sub-Gaussian loss functions, we obtain novel bounds that depend on the information density between the training data and the output hypothesis. When suitably weakened, these bounds recover many of the information-theoretic bounds available in the literature. We also extend the proposed exponential-inequality approach to the setting recently introduced by Steinke and Zakynthinou (2020), where the learning algorithm depends on a randomly selected subset of the available training data. For this setup, we present bounds for bounded loss functions in terms of the conditional information density between the output hypothesis and the random variable determining the subset choice, given all training data. Through our approach, we recover the average generalization bound presented by Steinke and Zakynthinou (2020) and extend it to the PAC-Bayesian and single-draw scenarios. For the single-draw scenario, we also obtain novel bounds in terms of the conditional α-mutual information and the conditional maximal leakage.

Locations

  • IEEE Journal on Selected Areas in Information Theory - View
  • arXiv (Cornell University) - View - PDF
  • Chalmers Research (Chalmers University of Technology) - View - PDF
  • DataCite API - View

Similar Works

Action Title Year Authors
+ Corrections to “Generalization Bounds via Information Density and Conditional Information Density” [Nov 20 824-839] 2021 Fredrik Hellström
Giuseppe Durisi
+ Fast-Rate Loss Bounds via Conditional Information Measures with Applications to Neural Networks 2020 Fredrik Hellström
Giuseppe Durisi
+ Fast-Rate Loss Bounds via Conditional Information Measures with Applications to Neural Networks 2020 Fredrik Hellström
Giuseppe Durisi
+ Tighter expected generalization error bounds via Wasserstein distance 2021 Borja Rodríguez-Gálvez
Germán Bassi
Ragnar Thobaben
Mikael Skoglund
+ PDF Chat Fast-Rate Loss Bounds via Conditional Information Measures with Applications to Neural Networks 2021 Fredrik Hellström
Giuseppe Durisi
+ Tighter Information-Theoretic Generalization Bounds from Supersamples 2023 Ziqiao Wang
Yongyi Mao
+ PDF Chat Information Complexity and Generalization Bounds 2021 Pradeep Banerjee
Guido Montúfar
+ Strengthened Information-theoretic Bounds on the Generalization Error 2019 Ibrahim Issa
Amedeo Roberto Esposito
Michael Gastpar
+ Generalization Bounds: Perspectives from Information Theory and PAC-Bayes 2023 Fredrik Hellström
Giuseppe Durisi
Benjamin Guedj
Maxim Raginsky
+ PDF Chat Assouad, Fano, and Le Cam with Interaction: A Unifying Lower Bound Framework and Characterization for Bandit Learnability 2024 Fan Chen
Dylan J. Foster
Yanjun Han
Jian Qian
Alexander Rakhlin
Yuqi Xu
+ Strengthened Information-theoretic Bounds on the Generalization Error. 2019 Ibrahim Issa
Amedeo Roberto Esposito
Michael Gastpar
+ PDF Chat Learning Algorithm Generalization Error Bounds via Auxiliary Distributions 2024 Gholamali Aminian
Saeed Masiha
Laura Toni
Miguel R. D. Rodrigues
+ Generalization Bounds via Convex Analysis 2022 Gábor Lugosi
Gergely Neu
+ Towards a Unified Information-Theoretic Framework for Generalization 2021 Mahdi Haghifam
Gintare Karolina Dziugaite
Shay Moran
Daniel M. Roy
+ PDF Chat Generalization Bounds via Conditional $f$-Information 2024 Ziqiao Wang
Yongyi Mao
+ Towards a Unified Information-Theoretic Framework for Generalization 2021 Mahdi Haghifam
Gintare Karolina Dziugaite
Shay Moran
Daniel M. Roy
+ PDF Chat Generalization Error Bounds via mth Central Moments of the Information Density 2020 Fredrik Hellström
Giuseppe Durisi
+ Robustness Implies Generalization via Data-Dependent Generalization Bounds 2022 Kenji Kawaguchi
Zhun Deng
Kyle Luh
Jiaoyang Huang
+ Generalization Error Bounds via $m$th Central Moments of the Information Density 2020 Fredrik Hellström
Giuseppe Durisi
+ Generalization Error Bounds via $m$th Central Moments of the Information Density 2020 Fredrik Hellström
Giuseppe Durisi