InstaHide: Instance-hiding Schemes for Private Distributed Learning
Author(s): Huang, Yangsibo; Song, Zhao; Li, Kai; Arora, Sanjeev
DownloadTo refer to this page use:
http://arks.princeton.edu/ark:/88435/pr1vg2g
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Huang, Yangsibo | - |
dc.contributor.author | Song, Zhao | - |
dc.contributor.author | Li, Kai | - |
dc.contributor.author | Arora, Sanjeev | - |
dc.date.accessioned | 2021-10-08T19:50:54Z | - |
dc.date.available | 2021-10-08T19:50:54Z | - |
dc.date.issued | 2020 | en_US |
dc.identifier.citation | Huang, Yangsibo, Zhao Song, Kai Li, and Sanjeev Arora. "InstaHide: Instance-hiding Schemes for Private Distributed Learning." In Proceedings of the 37th International Conference on Machine Learning (2020): pp. 4507-4518. | en_US |
dc.identifier.uri | http://proceedings.mlr.press/v119/huang20i/huang20i.pdf | - |
dc.identifier.uri | http://arks.princeton.edu/ark:/88435/pr1vg2g | - |
dc.description.abstract | How can multiple distributed entities train a shared deep net on their private data while protecting data privacy? This paper introduces InstaHide, a simple encryption of training images. Encrypted images can be used in standard deep learning pipelines (PyTorch, Federated Learning etc.) with no additional setup or infrastructure. The encryption has a minor effect on test accuracy (unlike differential privacy). Encryption consists of mixing the image with a set of other images (in the sense of Mixup data augmentation technique (Zhang et al., 2018)) followed by applying a random pixel-wise mask on the mixed image. Other contributions of this paper are: (a) Use of large public dataset of images (e.g. ImageNet) for mixing during encryption; this improves security. (b) Experiments demonstrating effectiveness in protecting privacy against known attacks while preserving model accuracy. (c) Theoretical analysis showing that successfully attacking privacy requires attackers to solve a difficult computational problem. (d) Demonstration that Mixup alone is insecure as (contrary to recent proposals), by showing some efficient attacks. (e) Release of a challenge dataset to allow design of new attacks. | en_US |
dc.format.extent | 4507 - 4518 | en_US |
dc.language.iso | en_US | en_US |
dc.relation.ispartof | Proceedings of the 37th International Conference on Machine Learning | en_US |
dc.rights | Final published version. Article is made available in OAR by the publisher's permission or policy. | en_US |
dc.title | InstaHide: Instance-hiding Schemes for Private Distributed Learning | en_US |
dc.type | Conference Article | en_US |
pu.type.symplectic | http://www.symplectic.co.uk/publications/atom-terms/1.0/conference-proceeding | en_US |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Instahide.pdf | 1.87 MB | Adobe PDF | View/Download |
Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.