Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

added one liner documentation for examples in README.md #351

Merged
merged 1 commit into from
Apr 10, 2021
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
17 changes: 16 additions & 1 deletion examples/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,8 +6,23 @@

#### Carrots Demo

This is the first example using the PyDP, about animals using the library to aggregate their data (about how many carrots they have eaten) before reporting it to their owner. More about it [here](carrots_demo/README.md).
This is the first example using the PyDP, about animals using the library to aggregate their data (about how many carrots they have eaten) before reporting it to their owner. More about it [here](Tutorial_1-carrots_demo/README.md).

#### Restaurant Demo

A restaurant owner would share business statistics with her visitors or potential clients, uses the PyDP library to preserve privacy of visitors while keeping track of number of visitors entering the restaurant and how much time and money they spend there. More about it [here](Tutorial_2-restaurant_demo/README.md).

#### Titanic Demo

This example uses the infamous Titanic dataset and finds what sorts of people were more likely to survive by finding demographic of people on the ship using differentially private statistical methods. More about it [here](Tutorial_3-Titanic_demo/Titanic_notebook.ipynb).

#### Launch Demo

In this example, two copies of one database are created where they differ by one record. This is for demonstrating the general principle used by all differentially private algorithms to protect users from MIA(Membership Inference Attack). More about it [here](Tutorial_4-Launch_demo/DP proof.ipynb).

#### Laplace Demo

This demo shows it is required to add noise to make data private. Laplace distribution makes it easy to satisfy ε-differential privacy by setting the b parameter to 1/ε. Hence, Laplace noise is used for making the data differentially private. More about it [here](laplace_demo/laplace.ipynb).

### Example

Expand Down