R.I.P.
๐ป
Ghosted
PAC-Bayes Compression Bounds So Tight That They Can Explain Generalization
November 24, 2022 ยท Entered Twilight ยท ๐ Neural Information Processing Systems
Repo contents: .gitignore, LICENSE, README.md, __init__.py, config, environment.yml, experiments, notebooks, pactl, requirements.txt, setup.cfg, setup.py, tests
Authors
Sanae Lotfi, Marc Finzi, Sanyam Kapoor, Andres Potapczynski, Micah Goldblum, Andrew Gordon Wilson
arXiv ID
2211.13609
Category
cs.LG: Machine Learning
Cross-listed
stat.ML
Citations
76
Venue
Neural Information Processing Systems
Repository
https://github.com/activatedgeek/tight-pac-bayes
โญ 19
Last Checked
1 month ago
Abstract
While there has been progress in developing non-vacuous generalization bounds for deep neural networks, these bounds tend to be uninformative about why deep learning works. In this paper, we develop a compression approach based on quantizing neural network parameters in a linear subspace, profoundly improving on previous results to provide state-of-the-art generalization bounds on a variety of tasks, including transfer learning. We use these tight bounds to better understand the role of model size, equivariance, and the implicit biases of optimization, for generalization in deep learning. Notably, we find large models can be compressed to a much greater extent than previously known, encapsulating Occam's razor. We also argue for data-independent bounds in explaining generalization.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Machine Learning
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
R.I.P.
๐ป
Ghosted
Semi-Supervised Classification with Graph Convolutional Networks
R.I.P.
๐ป
Ghosted
Proximal Policy Optimization Algorithms
R.I.P.
๐ป
Ghosted