TY - JOUR
T1 - Unknown-Aware Bilateral Dependency Optimization for Defending Against Model Inversion Attacks
AU - Peng, Xiong
AU - Liu, Feng
AU - Wang, Nannan
AU - Lan, Long
AU - Liu, Tongliang
AU - Cheung, Yiu-ming
AU - Han, Bo
N1 - Publisher copyright:
© 2025 IEEE.
PY - 2025/4/4
Y1 - 2025/4/4
N2 - By abusing access to a well-trained classifier, model inversion (MI) attacks pose a significant threat as they can recover the original training data, leading to privacy leakage. Previous studies mitigated MI attacks by imposing regularization to reduce the dependency between input features and outputs during classifier training, a strategy known as unilateral dependency optimization. However, this strategy contradicts the objective of minimizing the supervised classification loss, which inherently seeks to maximize the dependency between input features and outputs. Consequently, there is a trade-off between improving the model's robustness against MI attacks and maintaining its classification performance. To address this issue, we propose the bilateral dependency optimization strategy (BiDO), a dual-objective approach that minimizes the dependency between input features and latent representations, while simultaneously maximizing the dependency between latent representations and labels. BiDO is remarkable for its privacy-preserving capabilities. However, models trained with BiDO exhibit diminished capabilities in out-of-distribution (OOD) detection compared to models trained with standard classification supervision. Given the open-world nature of deep learning systems, this limitation could lead to significant security risks, as encountering OOD inputs—whose label spaces do not overlap with the in-distribution (ID) data used during training-is inevitable. To address this, we leverage readily available auxiliary OOD data to enhance the OOD detection performance of models trained with BiDO. This leads to the introduction of an upgraded framework, unknown-aware BiDO (BiDO+), which mitigates both privacy and security concerns. As a highlight, with comparable model utility, BiDO-HSIC+ reduces the FPR95 by $55.02 and enhances the AUCROC by $9.52 compared to BiDO-HSIC, while also providing superior MI robustness.
AB - By abusing access to a well-trained classifier, model inversion (MI) attacks pose a significant threat as they can recover the original training data, leading to privacy leakage. Previous studies mitigated MI attacks by imposing regularization to reduce the dependency between input features and outputs during classifier training, a strategy known as unilateral dependency optimization. However, this strategy contradicts the objective of minimizing the supervised classification loss, which inherently seeks to maximize the dependency between input features and outputs. Consequently, there is a trade-off between improving the model's robustness against MI attacks and maintaining its classification performance. To address this issue, we propose the bilateral dependency optimization strategy (BiDO), a dual-objective approach that minimizes the dependency between input features and latent representations, while simultaneously maximizing the dependency between latent representations and labels. BiDO is remarkable for its privacy-preserving capabilities. However, models trained with BiDO exhibit diminished capabilities in out-of-distribution (OOD) detection compared to models trained with standard classification supervision. Given the open-world nature of deep learning systems, this limitation could lead to significant security risks, as encountering OOD inputs—whose label spaces do not overlap with the in-distribution (ID) data used during training-is inevitable. To address this, we leverage readily available auxiliary OOD data to enhance the OOD detection performance of models trained with BiDO. This leads to the introduction of an upgraded framework, unknown-aware BiDO (BiDO+), which mitigates both privacy and security concerns. As a highlight, with comparable model utility, BiDO-HSIC+ reduces the FPR95 by $55.02 and enhances the AUCROC by $9.52 compared to BiDO-HSIC, while also providing superior MI robustness.
KW - Analytical models
KW - Data models
KW - Data privacy
KW - Face recognition
KW - Feature extraction
KW - Model inversion attacks
KW - Optimization
KW - Privacy
KW - Robustness
KW - Security
KW - Training
KW - dependency optimization
KW - out-of-distribution detection
UR - http://www.scopus.com/inward/record.url?scp=105002155575&partnerID=8YFLogxK
U2 - 10.1109/TPAMI.2025.3558267
DO - 10.1109/TPAMI.2025.3558267
M3 - Journal article
SN - 0162-8828
SP - 1
EP - 13
JO - IEEE Transactions on Pattern Analysis and Machine Intelligence
JF - IEEE Transactions on Pattern Analysis and Machine Intelligence
ER -