Unknown-Aware Bilateral Dependency Optimization for Defending Against Model Inversion Attacks

Xiong Peng, Feng Liu, Nannan Wang, Long Lan, Tongliang Liu, Yiu-ming Cheung, Bo Han*

*Corresponding author for this work

Research output: Contribution to journalJournal articlepeer-review

Abstract

By abusing access to a well-trained classifier, model inversion (MI) attacks pose a significant threat as they can recover the original training data, leading to privacy leakage. Previous studies mitigated MI attacks by imposing regularization to reduce the dependency between input features and outputs during classifier training, a strategy known as unilateral dependency optimization. However, this strategy contradicts the objective of minimizing the supervised classification loss, which inherently seeks to maximize the dependency between input features and outputs. Consequently, there is a trade-off between improving the model's robustness against MI attacks and maintaining its classification performance. To address this issue, we propose the bilateral dependency optimization strategy (BiDO), a dual-objective approach that minimizes the dependency between input features and latent representations, while simultaneously maximizing the dependency between latent representations and labels. BiDO is remarkable for its privacy-preserving capabilities. However, models trained with BiDO exhibit diminished capabilities in out-of-distribution (OOD) detection compared to models trained with standard classification supervision. Given the open-world nature of deep learning systems, this limitation could lead to significant security risks, as encountering OOD inputs—whose label spaces do not overlap with the in-distribution (ID) data used during training-is inevitable. To address this, we leverage readily available auxiliary OOD data to enhance the OOD detection performance of models trained with BiDO. This leads to the introduction of an upgraded framework, unknown-aware BiDO (BiDO+), which mitigates both privacy and security concerns. As a highlight, with comparable model utility, BiDO-HSIC+ reduces the FPR95 by $55.02 and enhances the AUCROC by $9.52 compared to BiDO-HSIC, while also providing superior MI robustness.
Original languageEnglish
Pages (from-to)1-13
Number of pages13
JournalIEEE Transactions on Pattern Analysis and Machine Intelligence
DOIs
Publication statusE-pub ahead of print - 4 Apr 2025

User-Defined Keywords

  • Analytical models
  • Data models
  • Data privacy
  • Face recognition
  • Feature extraction
  • Model inversion attacks
  • Optimization
  • Privacy
  • Robustness
  • Security
  • Training
  • dependency optimization
  • out-of-distribution detection

Fingerprint

Dive into the research topics of 'Unknown-Aware Bilateral Dependency Optimization for Defending Against Model Inversion Attacks'. Together they form a unique fingerprint.

Cite this