Guides selection of privacy-enhancing technologies like differential privacy, homomorphic encryption, federated learning, and zero-knowledge proofs with use-case matrices, performance comparisons, and GDPR assessments.
npx claudepluginhub mukul975/privacy-data-protection-skills --plugin privacy-skills-completeThis skill uses the workspace's default tool permissions.
Privacy-Enhancing Technologies (PETs) are technical measures that protect personal data during collection, processing, storage, and sharing. The GDPR does not prescribe specific technologies but requires "appropriate technical measures" (Article 25(1), Article 32(1)) to implement data protection principles. The European Commission's February 2025 communication on PETs and the ENISA 2023 report ...
Conducts multi-round deep research on GitHub repos via API and web searches, generating markdown reports with executive summaries, timelines, metrics, and Mermaid diagrams.
Dynamically discovers and combines enabled skills into cohesive, unexpected delightful experiences like interactive HTML or themed artifacts. Activates on 'surprise me', inspiration, or boredom cues.
Generates images from structured JSON prompts via Python script execution. Supports reference images and aspect ratios for characters, scenes, products, visuals.
Privacy-Enhancing Technologies (PETs) are technical measures that protect personal data during collection, processing, storage, and sharing. The GDPR does not prescribe specific technologies but requires "appropriate technical measures" (Article 25(1), Article 32(1)) to implement data protection principles. The European Commission's February 2025 communication on PETs and the ENISA 2023 report on engineering privacy by design with PETs provide regulatory context for PET adoption.
Selecting the right PET depends on the processing scenario, data sensitivity, computational requirements, and the specific privacy property needed (confidentiality, anonymity, unlinkability, or transparency).
Privacy property: Statistical indistinguishability — the output of an analysis is approximately the same whether or not any individual's data is included.
| Characteristic | Detail |
|---|---|
| Approach | Add calibrated noise to query results or model gradients |
| Privacy guarantee | Mathematically provable (epsilon, delta) bounds |
| Data utility | Configurable via epsilon; lower epsilon = more privacy, less utility |
| Performance overhead | Minimal for query-time noise; moderate for DP-SGD training |
| Maturity | Production-ready (Apple, Google, US Census Bureau) |
| GDPR relevance | Supports Recital 26 anonymization when epsilon is sufficiently small; Art. 25(1) by design measure |
Best for: Statistical analytics, aggregate reporting, ML model training on sensitive data.
Libraries: Google DP Library, OpenDP, IBM diffprivlib, PyDP.
Privacy property: Computation on encrypted data — results are decrypted only by the data owner.
| Characteristic | Detail |
|---|---|
| Approach | Encrypt data such that mathematical operations on ciphertext produce valid results when decrypted |
| Privacy guarantee | Cryptographic (lattice-based hardness assumptions) |
| Data utility | Exact results (no noise added) |
| Performance overhead | High: 1,000x-1,000,000x slower than plaintext, depending on scheme and operation |
| Maturity | Emerging for production; active research on performance optimization |
| GDPR relevance | Art. 32(1)(a) encryption as security measure; enables processing without exposing plaintext to processor |
Schemes:
Libraries: Microsoft SEAL, IBM HELib, Google FHE (Fully Homomorphic Encryption transpiler), TFHE-rs, OpenFHE.
Privacy property: Joint computation without revealing individual inputs — each party learns only the final result.
| Characteristic | Detail |
|---|---|
| Approach | Secret sharing, garbled circuits, or oblivious transfer protocols |
| Privacy guarantee | Information-theoretic (for secret sharing) or computational (for garbled circuits) |
| Data utility | Exact results |
| Performance overhead | Moderate to high; depends on circuit complexity and number of parties |
| Maturity | Production deployments in financial services, healthcare consortia |
| GDPR relevance | Enables joint controllership (Art. 26) analytics without data sharing; supports data minimization (Art. 5(1)(c)) |
Frameworks: MP-SPDZ, CrypTen (Meta), MOTION, ABY/ABY3, Sharemind.
Privacy property: Model training without centralizing data — raw data never leaves the data owner's environment.
| Characteristic | Detail |
|---|---|
| Approach | Distribute model training to data owners; aggregate only model updates (gradients) |
| Privacy guarantee | Data locality (raw data stays on device); strengthened with secure aggregation and DP |
| Data utility | Depends on data distribution across participants (IID vs non-IID) |
| Performance overhead | Communication overhead for gradient exchange; multiple training rounds |
| Maturity | Production at Google (Gboard), Apple (Siri), hospitals (NVIDIA FLARE) |
| GDPR relevance | Supports Art. 5(1)(c) minimization; reduces cross-border transfer requirements (Chapter V) |
Frameworks: TensorFlow Federated, PySyft (OpenMined), NVIDIA FLARE, Flower, FedML.
Privacy property: Prove a statement is true without revealing the underlying data.
| Characteristic | Detail |
|---|---|
| Approach | Prover demonstrates knowledge of a secret (e.g., age ≥ 18) without revealing the secret (date of birth) |
| Privacy guarantee | Soundness (false statements cannot be proven) and zero-knowledge (verifier learns nothing beyond the statement's truth) |
| Data utility | Binary verification (proof valid/invalid); no data disclosed |
| Performance overhead | Proof generation: moderate to high; proof verification: fast |
| Maturity | Production in blockchain identity; emerging in enterprise identity verification |
| GDPR relevance | Ultimate data minimization — prove compliance without disclosing data; supports Art. 5(1)(c), Art. 25 |
Systems: zk-SNARKs (Groth16, PLONK), zk-STARKs, Bulletproofs.
Privacy property: Isolated computation — data is processed inside a hardware-protected enclave that even the system administrator cannot access.
| Characteristic | Detail |
|---|---|
| Approach | Hardware-enforced isolation using secure enclaves (Intel SGX, AMD SEV, ARM TrustZone, AWS Nitro) |
| Privacy guarantee | Hardware-based attestation and memory encryption |
| Data utility | Full computation capability inside enclave |
| Performance overhead | Low to moderate; limited enclave memory may require data streaming |
| Maturity | Production: Azure Confidential Computing, AWS Nitro Enclaves, Google Confidential VMs |
| GDPR relevance | Art. 32(1)(a) encryption in processing; Art. 28 processor guarantees; protects against insider threats |
| Use Case | DP | HE | SMPC | FL | ZKP | TEE | Recommended Primary |
|---|---|---|---|---|---|---|---|
| Aggregate analytics on customer data | 5 | 2 | 3 | 2 | 1 | 3 | Differential Privacy |
| ML model training on distributed hospital data | 3 | 1 | 2 | 5 | 1 | 3 | Federated Learning |
| Credit scoring without sharing financial records | 2 | 3 | 5 | 2 | 2 | 3 | Secure MPC |
| Age verification without revealing date of birth | 1 | 1 | 1 | 1 | 5 | 2 | Zero-Knowledge Proofs |
| Processing encrypted customer data in third-party cloud | 2 | 5 | 2 | 1 | 1 | 4 | Homomorphic Encryption |
| Anti-money laundering across banks | 2 | 2 | 5 | 3 | 2 | 3 | Secure MPC |
| Private genomic analysis | 3 | 4 | 3 | 3 | 1 | 4 | HE + TEE combination |
| Confidential inference on cloud-hosted ML models | 2 | 4 | 2 | 1 | 1 | 5 | Trusted Execution Environment |
| Privacy-preserving surveys and polls | 5 | 1 | 3 | 1 | 2 | 1 | Differential Privacy |
| Secure keyword search on encrypted database | 1 | 4 | 1 | 1 | 1 | 4 | Homomorphic Encryption |
Scores: 1 = poor fit, 5 = excellent fit
Many production deployments combine PETs for defense in depth:
| Combination | Benefit |
|---|---|
| FL + DP | Federated learning with differentially private gradient updates prevents gradient inversion attacks |
| FL + Secure Aggregation | Aggregation server never sees individual gradients |
| HE + TEE | Process encrypted data inside an enclave for double protection |
| SMPC + DP | Add noise to SMPC outputs for formal anonymization guarantee |
| ZKP + TEE | Prove computation was performed correctly inside an enclave without revealing inputs |