TY - JOUR
T1 - REDAD
T2 - A Reliable Distillation for Image Anomaly Detection
AU - Hu, Wenrui
AU - Yu, Wei
AU - Tang, Yongqiang
AU - Xie, Yuan
AU - Zhang, Wensheng
N1 - Publisher Copyright:
© 1963-2012 IEEE.
PY - 2025
Y1 - 2025
N2 - The vanilla knowledge distillation (KD)-based approach for image anomaly detection (AD) encounters three main challenges: data shift, normality forgetting, and anomaly overgeneralization. To effectively address these challenges, we introduce REDAD, a reliable distillation strategy for AD, which avoids intricate network designs. Initially, we incorporate an adaptive teacher model (ATM) that dynamically adjusts the teacher’s batch normalization (BN) statistics during the distillation process, aligning the teacher with the target training data distribution. Next, we introduce a normality remembering enhancement (NRE) module, designed to compel the student to learn the most challenging normal feature with high distillation loss, thereby bolstering normality retention and reducing false positives. Finally, we present a novel direction-guided regularization (DGR) technique to robustly enlarge the divergence in abnormal feature pairs, preventing the oversight of abnormal regions, or false negatives. Comprehensive experiments on the MVTec, VisA, and MVTec3D datasets show that REDAD effectively resolves these three concerns, achieving superior AD performance compared with its baseline model (exceeding 3.5% (5.0%, 6.1%) in I-AUROC, 2.3% (1.6%, 1.5%) in P-AUROC, and 3.2% (5.4%, 3.8%) in PRO). In addition, two real-world industrial product inspection applications further underscore the efficacy and utility of the proposed REDAD method.
AB - The vanilla knowledge distillation (KD)-based approach for image anomaly detection (AD) encounters three main challenges: data shift, normality forgetting, and anomaly overgeneralization. To effectively address these challenges, we introduce REDAD, a reliable distillation strategy for AD, which avoids intricate network designs. Initially, we incorporate an adaptive teacher model (ATM) that dynamically adjusts the teacher’s batch normalization (BN) statistics during the distillation process, aligning the teacher with the target training data distribution. Next, we introduce a normality remembering enhancement (NRE) module, designed to compel the student to learn the most challenging normal feature with high distillation loss, thereby bolstering normality retention and reducing false positives. Finally, we present a novel direction-guided regularization (DGR) technique to robustly enlarge the divergence in abnormal feature pairs, preventing the oversight of abnormal regions, or false negatives. Comprehensive experiments on the MVTec, VisA, and MVTec3D datasets show that REDAD effectively resolves these three concerns, achieving superior AD performance compared with its baseline model (exceeding 3.5% (5.0%, 6.1%) in I-AUROC, 2.3% (1.6%, 1.5%) in P-AUROC, and 3.2% (5.4%, 3.8%) in PRO). In addition, two real-world industrial product inspection applications further underscore the efficacy and utility of the proposed REDAD method.
KW - Knowledge distillation (KD)
KW - self-supervised learning
KW - unsupervised anomaly detection (UAD)
UR - https://www.scopus.com/pages/publications/105001675365
U2 - 10.1109/TIM.2025.3551571
DO - 10.1109/TIM.2025.3551571
M3 - 文章
AN - SCOPUS:105001675365
SN - 0018-9456
VL - 74
JO - IEEE Transactions on Instrumentation and Measurement
JF - IEEE Transactions on Instrumentation and Measurement
M1 - 5018713
ER -