Calibrated One Round Federated Learning with Bayesian Inference in the Predictive Space
December 15, 2023 ยท Entered Twilight ยท ๐ AAAI Conference on Artificial Intelligence
Repo contents: Dataset, README.md, datasets.py, fed_algos.py, kd.py, models.py, noniid_plot.py, process_results.py, process_results_regr.py, requirements.txt, run_exp.py, sample_scripts, swa.py, teacher_oneshot_fedkt_exp.py, train_nets.py, utils.py
Authors
Mohsin Hasan, Guojun Zhang, Kaiyang Guo, Xi Chen, Pascal Poupart
arXiv ID
2312.09817
Category
cs.LG: Machine Learning
Cross-listed
stat.ML
Citations
14
Venue
AAAI Conference on Artificial Intelligence
Repository
https://github.com/hasanmohsin/betaPredBayesFL
โญ 4
Last Checked
2 months ago
Abstract
Federated Learning (FL) involves training a model over a dataset distributed among clients, with the constraint that each client's dataset is localized and possibly heterogeneous. In FL, small and noisy datasets are common, highlighting the need for well-calibrated models that represent the uncertainty of predictions. The closest FL techniques to achieving such goals are the Bayesian FL methods which collect parameter samples from local posteriors, and aggregate them to approximate the global posterior. To improve scalability for larger models, one common Bayesian approach is to approximate the global predictive posterior by multiplying local predictive posteriors. In this work, we demonstrate that this method gives systematically overconfident predictions, and we remedy this by proposing $ฮฒ$-Predictive Bayes, a Bayesian FL algorithm that interpolates between a mixture and product of the predictive posteriors, using a tunable parameter $ฮฒ$. This parameter is tuned to improve the global ensemble's calibration, before it is distilled to a single model. Our method is evaluated on a variety of regression and classification datasets to demonstrate its superiority in calibration to other baselines, even as data heterogeneity increases. Code available at https://github.com/hasanmohsin/betaPredBayesFL
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Machine Learning
R.I.P.
๐ป
Ghosted
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
R.I.P.
๐ป
Ghosted
Semi-Supervised Classification with Graph Convolutional Networks
R.I.P.
๐ป
Ghosted
Proximal Policy Optimization Algorithms
R.I.P.
๐ป
Ghosted