R.I.P.
๐ป
Ghosted
AdaPI: Facilitating DNN Model Adaptivity for Efficient Private Inference in Edge Computing
July 08, 2024 ยท Entered Twilight ยท ๐ International Conference on Computer Aided Design
Repo contents: README.md, combine_eval.py, dataset_download, evaluate_wgt_relu.py, grad_mask.ipynb, models_cifar, models_snl, models_util, plot_feature, scripts, train_autoprune_wgt_relu.py, train_util, util_func
Authors
Tong Zhou, Jiahui Zhao, Yukui Luo, Xi Xie, Wujie Wen, Caiwen Ding, Xiaolin Xu
arXiv ID
2407.05633
Category
cs.LG: Machine Learning
Cross-listed
cs.CR
Citations
18
Venue
International Conference on Computer Aided Design
Repository
https://github.com/jiahuiiiiii/AdaPI
โญ 5
Last Checked
2 months ago
Abstract
Private inference (PI) has emerged as a promising solution to execute computations on encrypted data, safeguarding user privacy and model parameters in edge computing. However, existing PI methods are predominantly developed considering constant resource constraints, overlooking the varied and dynamic resource constraints in diverse edge devices, like energy budgets. Consequently, model providers have to design specialized models for different devices, where all of them have to be stored on the edge server, resulting in inefficient deployment. To fill this gap, this work presents AdaPI, a novel approach that achieves adaptive PI by allowing a model to perform well across edge devices with diverse energy budgets. AdaPI employs a PI-aware training strategy that optimizes the model weights alongside weight-level and feature-level soft masks. These soft masks are subsequently transformed into multiple binary masks to enable adjustments in communication and computation workloads. Through sequentially training the model with increasingly dense binary masks, AdaPI attains optimal accuracy for each energy budget, which outperforms the state-of-the-art PI methods by 7.3\% in terms of test accuracy on CIFAR-100. The code of AdaPI can be accessed via https://github.com/jiahuiiiiii/AdaPI.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Machine Learning
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
R.I.P.
๐ป
Ghosted
Semi-Supervised Classification with Graph Convolutional Networks
R.I.P.
๐ป
Ghosted
Proximal Policy Optimization Algorithms
R.I.P.
๐ป
Ghosted