File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

박새롬

Park, Saerom
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Privacy-preserving inference resistant to model extraction attacks

Author(s)
Byun, JunyoungChoi, YujinLee, JaewookPark, Saerom
Issued Date
2024-12
DOI
10.1016/j.eswa.2024.124830
URI
https://scholarworks.unist.ac.kr/handle/201301/83420
Citation
EXPERT SYSTEMS WITH APPLICATIONS, v.256, pp.124830
Abstract
Privacy-Preserving Deep Learning (PPDL) has been successfully applied in the inference phase to preserve the privacy of input data. However, PPDL models are vulnerable to model extraction attacks, in which an adversary attempts to steal the trained model itself. In this paper, we propose a new defense method against model extraction attacks that is specifically designed for PPDL based on secure multi-party computations and homomorphic encryption. The proposed method confounds inference queries for out-of-distribution data by using a fake network with the target network while optimizing computational efficiency for PPDL environments. Furthermore, we introduce Wasserstein regularization to ensure that the fake network’s output distribution is indistinguishable from the target network, thwarting adversaries’ attempts to discern any discrepancies within the PPDL framework. The experimental results demonstrate that our defense method attains a good accuracy-security trade-off and is effective against a wide range of attacks, including adaptive attacks and transfer attacks. Our work contributes to the field of PPDL by providing an extended perspective to improve the algorithm’s security and reliability beyond privacy.
Publisher
PERGAMON-ELSEVIER SCIENCE LTD
ISSN
0957-4174

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.