← All projects

Personalized Privacy-Preserving Framework for Cross-Silo Federated Learning

IEEE Transactions on Emerging Topics in Computing · 2024
VinUni-Illinois Smart Health Center (VISHC)  ·  VinUniversity, Hanoi, Vietnam
Overall architecture of the PPPFL framework
Figure 1. Overall architecture of the PPPFL framework consisting of three consecutive stages. First, each participant locally synthesizes DP-guaranteed dataset D* by training their own DP-data-generator from secret dataset D. Next, in the federated training process, the synthetic data will be used to update the global model parameters θ. Lastly, the clients emerge from the federated training rounds and perform updates to their local model parameters θ* based on their secret training data.

Abstract

Federated learning (FL) is recently surging as a promising decentralized deep learning (DL) framework that enables DL-based approaches trained collaboratively across clients without sharing private data. However, in the context of the central party being active and dishonest, the data of individual clients might be perfectly reconstructed, leading to the high possibility of sensitive information being leaked. Moreover, FL also suffers from the nonindependent and identically distributed (non-IID) data among clients, resulting in the degradation in the inference performance on local clients' data.

In this paper, we propose a novel framework, namely Personalized Privacy-Preserving Federated Learning (PPPFL), with a concentration on cross-silo FL to overcome these challenges. Specifically, we introduce a stabilized variant of the Model-Agnostic Meta-Learning (MAML) algorithm to collaboratively train a global initialization from clients' synthetic data generated by Differential Private Generative Adversarial Networks (DP-GANs). After reaching convergence, the global initialization will be locally adapted by the clients to their private data.

Motivation

Cross-silo federated learning faces two critical, intertwined challenges that existing methods address only in isolation:

Challenge 1

Privacy Leakage from Dishonest Servers

In a semi-honest scenario, private information can be leaked from inversion attacks on updated gradients. In an active-and-dishonest scenario, the server can fully eavesdrop and modify the shared global model, setting up trap weights for recovering clients' data with zero reconstruction loss.

Challenge 2

Non-IID Data Distribution

In cross-silo FL settings, clients (companies, institutions, hospitals) possess vastly different sizes and partitions of private data. This statistical heterogeneity leads to a decline in the performance of the aggregated global model, especially when a single model must serve all clients.

Key Insight: PPPFL is the first framework to integrate two critical FL research areas — privacy preservation and handling non-IID data — into a single cohesive solution, providing differential privacy guarantees while achieving improved convergence and personalized performance.

PPPFL Framework

The proposed framework involves three consecutive stages:

Stage 1

Local Data Generation

Each client trains a differentially private data generator (based on DataLens) to synthesize artificial datasets. This ensures instance-level DP, providing stronger privacy guarantees than user-level DP. The original private data never leaves the client.

Stage 2

Federated Training

Clients use synthetic data for collaborative training via a stabilized MAML variant. The server meta-learns a generalized initialization θ from clients' feedback signals on synthetic data. Model EMA and Cosine Annealing Learning Rate prevent gradient instability.

Stage 3

Local Adaptation

After collaborative training, each client performs gradient descent updates on the well-initialized model using their secret private data, producing a personalized model θ* tailored to their local distribution.

Key Technical Innovations

Innovation 1

DataLens Integration for Instance-level DP

We consolidate DataLens — a state-of-the-art DP data generative model based on the PATE framework — into FL. DataLens uses multiple teacher discriminators with top-k gradient compression, stochastic sign quantization, and calibrated Gaussian noise to generate high-utility synthetic data while maintaining rigorous privacy guarantees.

Innovation 2

Stabilized MAML for Federated Training

The original MAML optimization is unstable due to multiple differentiations through the model. We propose a modified variant that employs Model Exponential Moving Average (EMA) in the inner loop and Cosine Annealing Learning Rate scheduling in the outer loop, enabling fast and stable convergence.

Innovation 3

Personalized Privacy Budgets

Clients can independently select their preferred privacy budget ε and incorporate it into the DP-GAN training. The server integrates clients' privacy budgets via a softmax-weighted objective, improving model performance while respecting individual privacy requirements.

Main Results

PPPFL is evaluated on four benchmarks: MNIST, Fashion-MNIST, CIFAR-10, and CIFAR-100, with 5 clients and 30 communication rounds under non-IID settings.

Performance on DP-Synthetic Data (Server-Client FL)

Table 1 · Average performance over clients, trained on synthetic, fine-tuned and tested on secret data
MethodMNISTFMNISTCIFAR-10CIFAR-100
BMT-F1BMTABMT-F1BMTABMT-F1BMTABMT-F1BMTA
FedAvg0.92230.90510.72670.79380.48980.65620.21420.3028
FedNova0.87730.89960.63030.76070.38470.66270.28550.3374
FedProx0.90060.90400.69940.77210.48560.67180.30140.3404
SCAFFOLD0.89060.90240.66980.80610.37320.66140.20960.3333
FedMeta0.93990.96160.83670.88000.68970.74660.27370.4133
PPPFL (ours)0.94200.96400.86220.91990.71630.80000.35260.4800
PPPFL outperforms all baselines across all four datasets in both BMT-F1 and BMTA metrics, while maintaining differential privacy guarantees.

Comparison to Decentralized FL Frameworks

Table 2 · BMTA comparison with decentralized FL frameworks
MethodMNISTFMNISTCIFAR-10CIFAR-100
AvgPush0.98720.89000.60710.2975
ProxyFL0.98700.89420.62920.3144
Dis-PFL0.90000.93330.83800.2612
PPPFL (ours)0.96400.91990.80000.4800
Despite using DP-synthetic data (not real private data), PPPFL achieves competitive performance and significantly outperforms on CIFAR-100, demonstrating strong utility-privacy balance.

Key Findings

Finding 1

Superior Accuracy with Privacy Guarantees

PPPFL outperforms all server-client FL baselines (FedAvg, FedNova, FedProx, SCAFFOLD, FedMeta) across all four benchmarks, achieving up to 8.0% improvement in BMT-F1 on CIFAR-100 over the second-best method.

Finding 2

Stabilized Convergence

The original MAML diverges after a few communication rounds. The stabilized MAML variant with Model EMA and Cosine Annealing LR achieves fast convergence within 10 rounds, validated on CIFAR-10.

Finding 3

Privacy-Utility Trade-off

Higher privacy budgets (ε) yield better model accuracy and synthetic data quality (lower FID). PPPFL demonstrates a strong correlation (Spearman's ρ = −0.90) between local model performance and collaboration gains, providing natural incentives for protocol participation.

Citation

If you find this work useful in your research, please consider citing:

@article{tran2024personalized,
  title     = {Personalized Privacy-Preserving Framework
               for Cross-Silo Federated Learning},
  author    = {Tran, Van-Tuan and Pham, Huy-Hieu
               and Wong, Kok-Seng},
  journal   = {IEEE Transactions on Emerging Topics
               in Computing},
  volume    = {12},
  number    = {4},
  pages     = {1014--1024},
  year      = {2024},
  publisher = {IEEE}
}

Acknowledgments

This work was supported by VinUni-Illinois Smart Health Center (VISHC), VinUniversity.