https://github.com/strmprivacy/blogpost-dss
An example notebook to work with STRM Privacy batch jobs and the Data Subjects API to easily retrieve data for Data Subject Requests (like a DSAR)
https://github.com/strmprivacy/blogpost-dss
batch-processing gdpr gdpr-compliant privacy privacy-enhancing-technologies privacy-tools
Last synced: 8 months ago
JSON representation
An example notebook to work with STRM Privacy batch jobs and the Data Subjects API to easily retrieve data for Data Subject Requests (like a DSAR)
- Host: GitHub
- URL: https://github.com/strmprivacy/blogpost-dss
- Owner: strmprivacy
- License: mit
- Created: 2022-09-15T08:29:55.000Z (about 3 years ago)
- Default Branch: master
- Last Pushed: 2022-09-20T13:22:06.000Z (about 3 years ago)
- Last Synced: 2025-01-09T02:26:54.665Z (9 months ago)
- Topics: batch-processing, gdpr, gdpr-compliant, privacy, privacy-enhancing-technologies, privacy-tools
- Language: Jupyter Notebook
- Homepage: https://strmprivacy.io/posts/batchjobs-and-datasubjects/
- Size: 4.97 MB
- Stars: 1
- Watchers: 3
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: readme.md
Awesome Lists containing this project
README
# STRM Privacy Batch Jobs and Data Subjects API
STRM Privacy is a privacy-focused data platform. This notebook is part of the tutorial on working with STRM Privacy's Batch Jobs and Data Subjects API: https://strmprivacy.io/posts/batchjobs-and-datasubjects/
## What is what
- With *STRM Batch Jobs*, you can process data in batches according to a privacy-focused _data contract_, from and to batch destinations (mostly buckets/storage)
- With the *Data Subjects API* you can easily locate and operate on a data subject's data through its keylink (as long as that data is processed through STRM).## Content
In this notebook you can find an example of
- Processing the UCI online retail dataset as batch job
- Setting up a data connector to read from and write into storage
- Defining a data contract and sending data according to it
- Configuring the batch job via a config JSON
- Executing the job itself
- Parsing and processing the encrypted dataset and working with the keys
- Querying the Data Subjects API to find all data related to a Data Subject## Sounds great
It does! STRM's Batch Jobs and Data Subjects API means:
- You can process regular workloads and add privacy by design to them
- Re-process existing data (as you probably don't know how you obtained it - oopsie)
- Easily and cheaply fulfill Data Subject Request through the Data Subjects API (via an associated keylink)
- And for deletion requests (RTBF - Right to be Forgotten) you can even fulfill them by just throwing away the keys (and only keep the anonymous data)
-