Zubnet AI学习Wiki › Differential Privacy
Safety

Differential Privacy

DP
一个在聚合数据分析和模型训练里保证个体隐私的数学框架。有了差分隐私,加入或移除任何单个个体的数据最多改变输出一个小的、有界的量。这意味着你能从数据集中学到有用的模式,不透露关于其中任何特定人的信息。

为什么重要

当 AI 在越来越个人化的数据上训练(健康记录、金融交易、消息),差分隐私提供了已知最强的保证 — 个体数据不能从模型中被提取。Apple(键盘预测)、Google(Chrome 使用分析)、美国人口普查局都在用。对 AI,它解决了 LLM 可能记忆并重现私人训练数据的顾虑。

Deep Dive

The formal guarantee: a mechanism M is ε-differentially private if for any two datasets D and D' that differ in one record, and any output S: P[M(D) ∈ S] ≤ e^ε · P[M(D') ∈ S]. Intuitively: the output looks essentially the same whether or not any specific individual's data is included. The privacy parameter ε controls the privacy-utility trade-off — smaller ε means stronger privacy but noisier (less useful) outputs.

DP in ML Training

DP-SGD (Differentially Private Stochastic Gradient Descent) adds calibrated noise to gradients during training, ensuring the trained model doesn't memorize individual examples. The trade-off: noise reduces model accuracy. For large models and datasets, the accuracy impact can be small. For small datasets, DP can significantly hurt performance. The practical challenge is choosing ε — too small and the model is useless, too large and privacy guarantees are meaningless.

The Memorization Problem

LLMs can memorize and reproduce training data verbatim — phone numbers, email addresses, proprietary code. This is a privacy violation even without intentional data extraction. Differential privacy during pre-training would prevent this memorization, but applying DP to models trained on trillions of tokens is computationally challenging and can degrade quality. Current practice uses a combination of: training data deduplication, output filtering, and careful data sourcing rather than formal DP guarantees. As regulation tightens, the pressure to adopt formal privacy guarantees will increase.

相关概念

← 所有术语
← Developer 工具 Diffusion Model →