When Data is Scarce, Learn to Adapt: Robust Federated Learning via Adversarial Meta-Optimization
FAML achieves SOTA robustness with single adversarially trained client
Comparison of robust and clean accuracy on CIFAR10 and CIFAR100 with varying numbers of AT-performing clients in FAML. Notably, even with a single AT-performing client, FAML, even with a single AT-performing client, FAML attains robustness comparable to the full-AT setting while preserving higher clean accuracy.
Abstract
Contributions
- We propose FAML, a federated adversarial meta-learning framework designed for robust learning under data scarcity and client heterogeneity.
- We show that a single adversarially trained client is sufficient to propagate robustness across the federation, removing the need for adversarial training at every client.
- FAML achieves strong clean and robust performance using only 20% of the training data, enabling robust federated learning in highly data-scarce regimes.
Federated Adversarial Training & Client Drift
Federated meta-learning follows a bi-level optimization process (inner and outer updates). We first investigate where robustness regularization should be applied for maximum effectiveness under client heterogeneity. We compare two variants: applying robustness only during the outer update (FAMLquery) versus during both inner and outer updates (FAMLboth). As shown in the left figure, regularizing only the outer update is sufficient to achieve strong robust accuracy, while avoiding the extra computational cost of full bi-level adversarial training. We further analyze training stability. The right figure shows that FAML suffers from client drift across communication rounds with higher perturbation strength.
To address this issue, we propose a novel adaptive entropy-based mask and a KL-based drift regularizer to mitigate client drift and ensure stable convergence.
Robust Accuracy vs. Attack Strength. Introducing robustness regularization during the outer update is sufficient to achieve strong robust accuracy.
Model Drift over Communication Rounds. Without any regularization, FAML suffers from significant client drift across communication rounds with higher perturbation strength.
Experimental Results
Clean and Robust Performance
We evaluate FAML against state-of-the-art federated adversarial training baselines across CIFAR10, SVHN, FMNIST, and CIFAR100 under multiple attack settings. FAML consistently achieves stronger clean and robust performance, demonstrating that a single adversarially trained client is sufficient for SOTA performance across the federation while requiring only 20% of the training data.
Feature Representation Analysis
We visualize feature embeddings using t-SNE to examine class separability. Compared to prior FAT baselines, FAML produces more compact and well-separated clusters, indicating improved feature robustness and alignment across clients.
Performance under data scarce scenario
Clean and PGD-20 robust accuracy of FAT methods on CIFAR10 and SVHN with varying training data ratios. FAML achieves SOTA performance even with only 20% of the training data.
Scalability Performance
Clean and robust accuracy across different numbers of clients K = {10, 20, 50} on CIFAR10 and SVHN datasets under PGD-20 and AutoAttack (AA) at ϵ = 8/255. FAML consistently outperforms all baselines, demonstrating strong robustness and scalability
Non-IID vs IID Performance (β Analysis)
Under severe non-IID settings (β = 0.05) and IID conditions (β = 10), FAML maintains strong robustness and stability. The results show that our proposed approach achieves SOTA performance even in highly heterogeneous federated environments.
Out-of-Domain Generalization
To evaluate transfer robustness, we test models trained on one dataset and evaluate on out-of-domain datasets. FAML demonstrates significantly stronger cross-domain robustness compared to prior methods, highlighting the effectiveness of transferable robust priors learned via meta-optimization.
Funding
This material is partly based upon work supported by the U.S. National Science Foundation (NSF) under Grant No. CRII-IIS-RI-2553868. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the NSF.