Towards Model Extraction Attacks in GAN-Based Image Translation via Domain Shift Mitigation

Loading...
Thumbnail Image
File version

Accepted Manuscript (AM)

Author(s)
Mi, D
Zhang, Y
Zhang, LY
Hu, S
Zhong, Q
Yuan, H
Pan, S
Griffith University Author(s)
Primary Supervisor
Other Supervisors
Editor(s)
Date
2024
Size
File type(s)
Location

Vancouver, Canada

License
Abstract

Model extraction attacks (MEAs) enable an attacker to replicate the functionality of a victim deep neural network (DNN) model by only querying its API service remotely, posing a severe threat to the security and integrity of pay-per-query DNN-based services. Although the majority of current research on MEAs has primarily concentrated on neural classifiers, there is a growing prevalence of image-to-image translation (I2IT) tasks in our everyday activities. However, techniques developed for MEA of DNN classifiers cannot be directly transferred to the case of I2IT, rendering the vulnerability of I2IT models to MEA attacks often underestimated. This paper unveils the threat of MEA in I2IT tasks from a new perspective. Diverging from the traditional approach of bridging the distribution gap between attacker queries and victim training samples, we opt to mitigate the effect caused by the different distributions, known as the domain shift. This is achieved by introducing a new regularization term that penalizes high-frequency noise, and seeking a flatter minimum to avoid overfitting to the shifted distribution. Extensive experiments on different image translation tasks, including image super-resolution and style transfer, are performed on different backbone victim models, and the new design consistently outperforms the baseline by a large margin across all metrics. A few real-life I2IT APIs are also verified to be extremely vulnerable to our attack, emphasizing the need for enhanced defenses and potentially revised API publishing policies.

Journal Title
Conference Title

Proceedings of the AAAI Conference on Artificial Intelligence

Book Title
Edition
Volume

38

Issue

18

Thesis Type
Degree Program
School
Publisher link
Patent number
Funder(s)
Grant identifier(s)
Rights Statement

This work is covered by copyright. You must assume that re-use is limited to personal use and that permission from the copyright owner must be obtained for all other uses. If the document is available under a specified licence, refer to the licence for details of permitted re-use. If you believe that this work infringes copyright please make a copyright takedown request using the form at https://www.griffith.edu.au/copyright-matters.

Rights Statement
Item Access Status
Note
Access the data
Related item(s)
Subject

Data security and protection

Artificial intelligence

Deep learning

Persistent link to this record
Citation

Mi, D; Zhang, Y; Zhang, LY; Hu, S; Zhong, Q; Yuan, H; Pan, S, Towards Model Extraction Attacks in GAN-Based Image Translation via Domain Shift Mitigation, Proceedings of the AAAI Conference on Artificial Intelligence, 2024, 38 (18), pp. 19902-19910