Data Poisoning and Leakage Analysis in Federated Learning
Wenqi Wei (),
Tiansheng Huang,
Zachary Yahn,
Anoop Singhal,
Margaret Loper and
Ling Liu
Additional contact information
Wenqi Wei: School of Computer Science
Tiansheng Huang: School of Computer Science
Zachary Yahn: School of Computer Science
Anoop Singhal: National Institute of Standards and Technology
Margaret Loper: Georgia Tech Research Institute
Ling Liu: School of Computer Science
A chapter in Handbook of Trustworthy Federated Learning, 2025, pp 73-108 from Springer
Abstract:
Abstract Data poisoning and leakage risks impede the massive deployment of federated learning in the real world. This chapter reveals the truths and pitfalls of understanding two dominating threats: training data privacy intrusion and training data poisoning. We first investigate training data privacy threat and present our observations on when and how training data may be leaked during the course of federated training. One promising defense strategy is to perturb the raw gradient update by adding some controlled randomized noise prior to sharing during each round of federated learning. We discuss the importance of determining the proper amount of randomized noise and the proper location to add such noise for effective mitigation of gradient leakage threats against training data privacy. Then we will review and compare different training data poisoning threats and analyze why and when such data poisoning induced model Trojan attacks may lead to detrimental damage on the performance of the global model. We will categorize and compare representative poisoning attacks and the effectiveness of their mitigation techniques, delivering an in-depth understanding of the negative impact of data poisoning. Finally, we demonstrate the potential of dynamic model perturbation in simultaneously ensuring privacy protection, poisoning resilience, and model performance. The chapter concludes with a discussion on additional risk factors in federated learning, including the negative impact of skewness, data and algorithmic biases, as well as misinformation in training data. Powered by empirical evidence, our analytical study offers some transformative insights into effective privacy protection and security assurance strategies in attack-resilient federated learning.
Date: 2025
References: Add references at CitEc
Citations:
There are no downloads for this item, see the EconPapers FAQ for hints about obtaining it.
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:spr:spochp:978-3-031-58923-2_3
Ordering information: This item can be ordered from
http://www.springer.com/9783031589232
DOI: 10.1007/978-3-031-58923-2_3
Access Statistics for this chapter
More chapters in Springer Optimization and Its Applications from Springer
Bibliographic data for series maintained by Sonal Shukla () and Springer Nature Abstracting and Indexing ().