TY - JOUR
T1 - Knowledge Distillation-Based Anomaly Detection via Adaptive Discrepancy Optimization
AU - Li, Ning
AU - Liu, Ajian
AU - Zhu, Zhenwei
AU - Lin, Xuxin
AU - Ma, Hui
AU - Dai, Hong-Ning
AU - Liang, Yanyan
N1 - This work was supported in part by the Science and Technology Development Fund of Macao under Grant 0070/2020/AMJ and Grant 0096/2023/RIA2, in part by the National Natural Science Foundation of China under Grant 62406320, in part by the Macao Young Scholars ProgramAM2024016, and in part by Zhuhai City Polytechnic Research Project under Grant 2024KYBS02 and Grant KY2024Y03Z.
Publisher Copyright:
© 2025 IEEE
PY - 2025/6/11
Y1 - 2025/6/11
N2 - Knowledge distillation has emerged as a primary solution for anomaly detection, leveraging feature discrepancies between teacher–student (T–S) networks to locate anomalies. However, previous approaches suffer from ambiguous feature discrepancies, which hinder effective anomaly detection due to two main challenges: 1) overgeneralization, where the student network excessively mimics teacher features in anomalous regions, and 2) semantic bias between T–S networks in normal regions. To address these issues, we propose an Adaptive Discrepancy Optimization (Ado) block. The Ado block adaptively calibrates feature discrepancies by reducing overgeneralization in anomalous regions and selectively aligning semantic features in normal regions via learnable feature offsets. This versatile block can be seamlessly integrated into various distillation-based methods. Experimental results demonstrate that the Ado block significantly enhances performance across 11 different knowledge distillation frameworks on two widely used datasets. Notably, when integrated with the Ado block, RD4AD achieves a 22% relative improvement in pixel-level PRO on the VisA dataset. In addition, a real-world keyboard inspection application further validates the effectiveness of the Ado block.
AB - Knowledge distillation has emerged as a primary solution for anomaly detection, leveraging feature discrepancies between teacher–student (T–S) networks to locate anomalies. However, previous approaches suffer from ambiguous feature discrepancies, which hinder effective anomaly detection due to two main challenges: 1) overgeneralization, where the student network excessively mimics teacher features in anomalous regions, and 2) semantic bias between T–S networks in normal regions. To address these issues, we propose an Adaptive Discrepancy Optimization (Ado) block. The Ado block adaptively calibrates feature discrepancies by reducing overgeneralization in anomalous regions and selectively aligning semantic features in normal regions via learnable feature offsets. This versatile block can be seamlessly integrated into various distillation-based methods. Experimental results demonstrate that the Ado block significantly enhances performance across 11 different knowledge distillation frameworks on two widely used datasets. Notably, when integrated with the Ado block, RD4AD achieves a 22% relative improvement in pixel-level PRO on the VisA dataset. In addition, a real-world keyboard inspection application further validates the effectiveness of the Ado block.
KW - Anomaly detection
KW - defect detection
KW - knowledge distillation
KW - overgeneralization problem
KW - semantic bias
UR - http://www.scopus.com/inward/record.url?scp=105008145950&partnerID=8YFLogxK
U2 - 10.1109/TII.2025.3574540
DO - 10.1109/TII.2025.3574540
M3 - Journal article
SN - 1941-0050
JO - IEEE Transactions on Industrial Informatics
JF - IEEE Transactions on Industrial Informatics
ER -