Defending against model extraction attacks with physical unclonable function

  • Dawei Li
  • , Di Liu*
  • , Ying Guo
  • , Yangkun Ren
  • , Jieyu Su
  • , Jianwei Liu
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Machine learning models, especially deep neural network (DNN) models, have widespread and valuable applications in business activities. Training a deep learning model for commercial use requires plenty of private data, expert knowledge, and computing resources. The huge commercial value of such trained models has attracted the attention of attackers. Attackers can construct a dataset by repeatedly querying the target model for the output of the requested samples and then train a substitute model on this dataset that functions similarly to the target model. In this paper, we propose a defense scheme based on physical unclonable function (PUF) against such black-box model extraction attacks. We deploy a PUF on the user side and the corresponding PUF model on the service provider side to ensure that only legitimate users can obtain the correct model predictions. Our experimental results show that by choosing a suitable fuzzy extractor threshold d, legitimate users can recover more than 99.5% of the prediction results with a little additional computational overhead to the service provider. We perform a model extraction attack in the most favorable case for the attacker, and the prediction accuracy of the obtained substitute model is only about 10%, which demonstrates the effectiveness of our proposed scheme. Compared to existing defenses, our scheme not only effectively prevents black-box model extraction attacks but also ensures that the accuracy of the prediction service for legitimate users is not affected.

Original languageEnglish
Pages (from-to)196-207
Number of pages12
JournalInformation Sciences
Volume628
DOIs
StatePublished - May 2023

Keywords

  • Black-box model extraction
  • Fuzzy extractor
  • Machine learning
  • Physical unclonable function

Fingerprint

Dive into the research topics of 'Defending against model extraction attacks with physical unclonable function'. Together they form a unique fingerprint.

Cite this