Privacy Protection Optimization in Federated Learning
Abstrak
Federated learning has emerged as a promising distributed machine learning paradigm that enables collaborative model training while preserving data privacy. However, the increasing sophistication of privacy attacks and evolving regulatory requirements have exposed critical vulnerabilities in current FL systems. This paper provides a comprehensive analysis of privacy threats in federated learning, identifying three primary attack surfaces: gradient-based reconstruction, aggregation-phase breaches, and membership leakage during participant selection. This paper examines how these vulnerabilities manifest differently across healthcare, financial, and industrial applications, with sector-specific risks ranging from medical image reconstruction to inference of sensitive financial attributes. The study systematically evaluates three categories of defense mechanisms: differential privacy techniques (including adaptive noise injection and hybrid approaches), cryptographic methods (homomorphic encryption and secure multi-party computation), and blockchain-based distributed architectures. This paper analyzes the inherent trade-offs between privacy protection and model performance, presenting optimization strategies such as adaptive privacy budgeting and lightweight encryption to mitigate accuracy degradation. The paper further discusses compliance challenges posed by emerging regulations like the EU AI Act and FDA guidelines, highlighting the need for verifiable privacy proofs in sensitive domains. Finally, this paper concludes with a summary and outlook.
Topik & Kata Kunci
Penulis (1)
Zhou Xinyi
Akses Cepat
- Tahun Terbit
- 2025
- Sumber Database
- DOAJ
- DOI
- 10.1051/itmconf/20257804003
- Akses
- Open Access ✓