Denial-of-Service or Fine-Grained Control: Towards Flexible Model Poisoning Attacks on Federated Learning
File version
Version of Record (VoR)
Author(s)
Yao, Z
Zhang, LY
Hu, S
Chen, C
Liew, A
Li, Z
Griffith University Author(s)
Primary Supervisor
Other Supervisors
Editor(s)
Date
Size
File type(s)
Location
Macao, China
License
Abstract
Federated learning (FL) is vulnerable to poisoning attacks, where adversaries corrupt the global aggregation results and cause denial-of-service (DoS). Unlike recent model poisoning attacks that optimize the amplitude of malicious perturbations along certain prescribed directions to cause DoS, we propose a flexible model poisoning attack (FMPA) that can achieve versatile attack goals. We consider a practical threat scenario where no extra knowledge about the FL system (e.g., aggregation rules or updates on benign devices) is available to adversaries. FMPA exploits the global historical information to construct an estimator that predicts the next round of the global model as a benign reference. It then fine-tunes the reference model to obtain the desired poisoned model with low accuracy and small perturbations. Besides the goal of causing DoS, FMPA can be naturally extended to launch a fine-grained controllable attack, making it possible to precisely reduce the global accuracy. Armed with precise control, malicious FL service providers can gain advantages over their competitors without getting noticed, hence opening a new attack surface in FL other than DoS. Even for the purpose of DoS, experiments show that FMPA significantly decreases the global accuracy, outperforming six state-of-the-art attacks.
Journal Title
Conference Title
Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence
Book Title
Edition
Volume
Issue
Thesis Type
Degree Program
School
Publisher link
Patent number
Funder(s)
Grant identifier(s)
Rights Statement
Rights Statement
© 2023 International Joint Conference on Artificial Intelligence. The attached file is reproduced here in accordance with the copyright policy of the publisher. Please refer to the Conference's website for access to the definitive, published version.
Item Access Status
Note
Access the data
Related item(s)
Subject
Artificial intelligence
Persistent link to this record
Citation
Zhang, H; Yao, Z; Zhang, LY; Hu, S; Chen, C; Liew, A; Li, Z, Denial-of-Service or Fine-Grained Control: Towards Flexible Model Poisoning Attacks on Federated Learning, Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, 2023, pp. 4567-4575