R.I.P.
๐ป
Ghosted
How Does Information Bottleneck Help Deep Learning?
May 30, 2023 ยท Entered Twilight ยท ๐ International Conference on Machine Learning
Repo contents: .gitignore, LICENSE, README.md, __init__.py, dnn, requirements.txt, toy
Authors
Kenji Kawaguchi, Zhun Deng, Xu Ji, Jiaoyang Huang
arXiv ID
2305.18887
Category
cs.LG: Machine Learning
Cross-listed
cs.AI,
cs.CL,
cs.CV,
cs.IT
Citations
114
Venue
International Conference on Machine Learning
Repository
https://github.com/xu-ji/information-bottleneck
โญ 63
Last Checked
2 months ago
Abstract
Numerous deep learning algorithms have been inspired by and understood via the notion of information bottleneck, where unnecessary information is (often implicitly) minimized while task-relevant information is maximized. However, a rigorous argument for justifying why it is desirable to control information bottlenecks has been elusive. In this paper, we provide the first rigorous learning theory for justifying the benefit of information bottleneck in deep learning by mathematically relating information bottleneck to generalization errors. Our theory proves that controlling information bottleneck is one way to control generalization errors in deep learning, although it is not the only or necessary way. We investigate the merit of our new mathematical findings with experiments across a range of architectures and learning settings. In many cases, generalization errors are shown to correlate with the degree of information bottleneck: i.e., the amount of the unnecessary information at hidden layers. This paper provides a theoretical foundation for current and future methods through the lens of information bottleneck. Our new generalization bounds scale with the degree of information bottleneck, unlike the previous bounds that scale with the number of parameters, VC dimension, Rademacher complexity, stability or robustness. Our code is publicly available at: https://github.com/xu-ji/information-bottleneck
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Machine Learning
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
R.I.P.
๐ป
Ghosted
Semi-Supervised Classification with Graph Convolutional Networks
R.I.P.
๐ป
Ghosted
Proximal Policy Optimization Algorithms
R.I.P.
๐ป
Ghosted