README.md 2.67 KB
Newer Older
Xin Pan's avatar
Xin Pan committed
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
<font size=4><b>Deep Learning with Differential Privacy</b></font>

Authors:
Martín Abadi, Andy Chu, Ian Goodfellow, H. Brendan McMahan, Ilya Mironov, Kunal Talwar, Li Zhang

Open Sourced By: Xin Pan (xpan@google.com, github: panyx0718)


<Introduction>

Machine learning techniques based on neural networks are achieving remarkable 
results in a wide variety of domains. Often, the training of models requires 
large, representative datasets, which may be crowdsourced and contain sensitive 
information. The models should not expose private information in these datasets. 
Addressing this goal, we develop new algorithmic techniques for learning and a 
refined analysis of privacy costs within the framework of differential privacy. 
Our implementation and experiments demonstrate that we can train deep neural 
networks with non-convex objectives, under a modest privacy budget, and at a 
manageable cost in software complexity, training efficiency, and model quality.

paper: https://arxiv.org/abs/1607.00133


<b>Requirements:</b>

1. Tensorflow 0.10.0 (master branch)

Note: r0.11 might experience some problems

2. Bazel 0.3.1

3. Download MNIST data

TODO(xpan): Complete the link:
[train](http://download.tensorflow.org/models/)
[test](http://download.tensorflow.org/models/)

Alternatively, download the tfrecord format MNIST from:
https://github.com/panyx0718/models/tree/master/slim

<b>How to run:</b>

```shell
# Clone the codes under differential_privacy.
# Create an empty WORKSPACE file.
# Download the data to the data/ directory.
ls -R
.:
data  differential_privacy  WORKSPACE

./data:
mnist_test.tfrecord  mnist_train.tfrecord

./differential_privacy:
dp_mnist  dp_optimizer  __init__.py  per_example_gradients  README.md

./differential_privacy/dp_mnist:
BUILD  dp_mnist.py

./differential_privacy/dp_optimizer:
accountant.py  BUILD  dp_optimizer.py  dp_pca.py  sanitizer.py  utils.py

./differential_privacy/per_example_gradients:
BUILD  __init__.py  per_example_gradients.py

# Build the codes.
bazel build -c opt differential_privacy/...

# Run the mnist differntial privacy training codes.
bazel-bin/differential_privacy/dp_mnist/dp_mnist \
    --training_data_path=data/mnist_train.tfrecord \
    --eval_data_path=data/mnist_test.tfrecord \
    --save_path=/tmp/mnist_dir

...
step: 1
step: 2
...
step: 9
spent privacy: eps 0.1250 delta 0.72709
spent privacy: eps 0.2500 delta 0.24708
spent privacy: eps 0.5000 delta 0.0029139
spent privacy: eps 1.0000 delta 6.494e-10
spent privacy: eps 2.0000 delta 8.2242e-24
spent privacy: eps 4.0000 delta 1.319e-51
spent privacy: eps 8.0000 delta 3.3927e-107
train_accuracy: 0.53
eval_accuracy: 0.53
...

ls /tmp/mnist_dir/
checkpoint  ckpt  ckpt.meta  results-0.json
```