Opacus is a library that enables training PyTorch models with differential privacy. It supports training with minimal code changes required on the client, has little impact on training performance, and allows the client to online track the privacy budget expended at any given moment.
Target audience
This code release is aimed at two target audiences:
- ML practitioners will find this to be a gentle introduction to training a model with differential privacy as it requires minimal code changes.
- Differential Privacy researchers will find this easy to experiment and tinker with, allowing them to focus on what matters.
Installation
The latest release of Opacus can be installed via pip
:
pip install opacus
You can also install directly from the source for the latest features (along with its quirks and potentially ocassional bugs):
git clone https://github.com/pytorch/opacus.git
cd opacus
pip install -e .
Getting started
To train your model with differential privacy, all you need to do is to instantiate a PrivacyEngine
and pass your model, data_loader, and optimizer to the engine's make_private()
method to obtain their private counterparts.
# define your components as usual
model = Net()
optimizer = SGD(model.parameters(), lr=0.05)
data_loader = torch.utils.data.DataLoader(dataset, batch_size=1024)
# enter PrivacyEngine
privacy_engine = PrivacyEngine()
model, optimizer, data_loader = privacy_engine.make_private(
module=model,
optimizer=optimizer,
data_loader=data_loader,
noise_multiplier=1.1,
max_grad_norm=1.0,
)
# Now it's business as usual
The MNIST example shows an end-to-end run using opacus. The examples folder contains more such examples.
Migrating to 1.0
Opacus 1.0 introduced many improvements to the library, but also some breaking changes. If you've been using Opacus 0.x and want to update to the latest release, please use this Migration Guide
Learn more
Interactive tutorials
We've built a series of IPython-based tutorials as a gentle introduction to training models with privacy and using various Opacus features.
- Building an Image Classifier with Differential Privacy
- Training a differentially private LSTM model for name classification
- Building text classifier with Differential Privacy on BERT
- Opacus Guide: Introduction to advanced features
- Opacus Guide: Grad samplers
- Opacus Guide: Module Validator and Fixer
Blogposts and talks
If you want to learn more about DP-SGD and related topics, check our our series of blogposts and talks:
- Differential Privacy Series Part 1 | DP-SGD Algorithm Explained
- Differential Privacy Series Part 2 | Efficient Per-Sample Gradient Computation in Opacus
- PriCon 2020 Tutorial: Differentially Private Model Training with Opacus
- Differential Privacy on PyTorch | PyTorch Developer Day 2020
FAQ
Checkout the FAQ page for answers to some of the most frequently asked questions about Differential Privacy and Opacus.
Contributing
See the CONTRIBUTING file for how to help out. Do also check out the README files inside the repo to learn how the code is organized.
Citation
To cite Opacus in your papers (much appreciated!), please use the following:
@article{opacus,
title={Opacus: User-Friendly Differential Privacy Library in PyTorch},
author={A. Yousefpour and I. Shilov and A. Sablayrolles and D. Testuggine and K. Prasad and M. Malek and J. Nguyen and S. Ghosh and A. Bharadwaj and J. Zhao and G. Cormode and I. Mironov},
journal={arXiv preprint arXiv:2109.12298},
year={2021}
}
License
This code is released under Apache 2.0, as found in the LICENSE file.