Zubnet AI學習Wiki › Federated 學習ing
Training

Federated 學習ing

FL, Collaborative 學習ing
一種在多個裝置或組織間訓練模型而不共享原始資料的訓練方法。不是把資料發到中心伺服器,而是每個參與者在自己的資料上訓練本地模型副本,只把模型更新(梯度)發到中央協調者。協調者聚合所有參與者的更新來改進全域模型。

為什麼重要

聯邦學習啟用在因隱私、監管、或競爭顧慮而無法中心化的資料上做 AI 訓練。醫院可以協作訓練診斷模型而不共享病人記錄。公司可以改進共享模型而不暴露專有資料。它是規模化保隱私 AI 訓練最實用的方法。

Deep Dive

The standard federated learning algorithm (FedAvg): (1) the server sends the current model to selected participants, (2) each participant trains the model on their local data for several steps, (3) participants send their updated model weights (not data) to the server, (4) the server averages the updates and creates a new global model, (5) repeat. The key property: raw data never leaves the participant's device.

Challenges

Non-IID data: participants often have very different data distributions (a hospital in Tokyo has different patient demographics than one in São Paulo). This makes training unstable — updates from different participants may conflict. Communication cost: sending model updates (potentially billions of parameters) over the network is expensive, especially for mobile devices. Free-riders: participants who receive the improved model but contribute low-quality updates. These challenges make federated learning harder than centralized training, though each has active solutions.

Real-World Use

Apple uses federated learning for keyboard prediction (learning from what you type without sending your texts to Apple). Google uses it for search suggestion improvement. Healthcare consortiums use it for multi-hospital model training. The technique is most valuable when: the data is truly sensitive (medical, financial), regulation prevents data sharing (GDPR, HIPAA), or the data is too large to centralize (billions of mobile device interactions).

相關概念

← 所有術語
← Feature Feedforward Network →