DP-Mix: Mixup-based Data Augmentation for Differentially Private Learning

Wenxuan Bao1 , Francesco Pittaluga2 , Vijay Kumar B G2 , Vincent Bindschaedler1
1 University of Florida
2 NEC Labs America
Our method

Abstract

Data augmentation techniques, such as simple image transformations and combinations, are highly effective at improving the generalization of computer vision models, especially when training data is limited. However, such techniques are fundamentally incompatible with differentially private learning approaches, due to the latter's built-in assumption that each training image's contribution to the learned model is bounded. In this paper, we investigate why naive applications of multi-sample data augmentation techniques, such as mixup, fail to achieve good performance and propose two novel data augmentation techniques specifically designed for the constraints of differentially private learning. Our first technique, DP-Mix_Self, achieves SoTA classification performance across a range of datasets and settings by performing mixup on self-augmented data. Our second technique, DP-Mix_Diff, further improves performance by incorporating synthetic data from a pre-trained diffusion model into the mixup process.

Poster

Slides

BibTeX

@inproceedings{bao2023dpmix,
  title={DP-Mix: Mixup-based Data Augmentation for Differentially Private Learning},
  author={Bao, Wenxuan and Pittaluga, Francesco and Kumar B G, Vijay and Bindschaedler, Vincent},
  journal={Advances in Neural Information Processing Systems},
  year={2023}
}