Towards Certificated Model Robustness Against Weight Perturbations
Authors
Authors
- Luca Daniel
- Xue Lin
- Pin-Yu Chen
- Sijia Liu
- Pu Zhao
- Lily Weng
Authors
- Luca Daniel
- Xue Lin
- Pin-Yu Chen
- Sijia Liu
- Pu Zhao
- Lily Weng
Published on
02/12/2020
This work studies the sensitivity of neural networks to weight perturbations, firstly corresponding to a newly developed threat model that perturbs the neural network parameters. We propose an efficient approach to compute a certified robustness bound of weight perturbations, within which neural networks will not make erroneous outputs as desired by the adversary. In addition, we identify a useful connection between our developed certification method and the problem of weight quantization, a popular model compression technique in deep neural networks (DNNs) and a ‘must-try’ step in the design of DNN inference engines on resource constrained computing platforms, such as mobiles, FPGA, and ASIC. Specifically, we study the problem of weight quantization – weight perturbations in the non-adversarial setting – through the lens of certificated robustness, and we demonstrate significant improvements on the generalization ability of quantized networks through our robustness-aware quantization scheme.
Please cite our work using the BibTeX below.
@article{Weng_Zhao_Liu_Chen_Lin_Daniel_2020,
title={Towards Certificated Model Robustness Against Weight Perturbations},
volume={34},
url={https://ojs.aaai.org/index.php/AAAI/article/view/6105},
DOI={10.1609/aaai.v34i04.6105},
number={04},
journal={Proceedings of the AAAI Conference on Artificial Intelligence},
author={Weng, Tsui-Wei and Zhao, Pu and Liu, Sijia and Chen, Pin-Yu and Lin, Xue and Daniel, Luca},
year={2020},
month={Apr.},
pages={6356-6363}
}