TY - JOUR
T1 - Leveraging multi-class background description and token dictionary representation for hyperspectral anomaly detection
AU - Wang, Zhiwei
AU - Tan, Kun
AU - Wang, Xue
AU - Zhang, Wen
N1 - Publisher Copyright:
© 2025
PY - 2026/1
Y1 - 2026/1
N2 - Hyperspectral anomaly detection is aimed at distinguishing between background and anomalous regions in hyperspectral images, and plays a crucial role iSn various applications. However, the existing deep learning methods face challenges when dealing with complex background distributions and insufficient training data. In this article, we propose a novel multi-class background description transformer network (MBDTNet) to address the problems of imprecise background distribution learning and poor anomaly detection. Firstly, we propose an image-level end-to-end data augmentation method based on self-supervised training, which enhances the diversity and quantity of the training samples through adaptive clustering and spatial masking strategies. Secondly, based on the principles of low-rank representation, a sparse self-attention mechanism based on token dictionary representation is designed to help the model focus on key background features and guide the model in recognizing anomalies. Finally, a token dictionary learning mechanism for multi-class background description is established by combining Gaussian discriminant analysis with a conditional distance function, and intra-class and inter-class losses are designed to enhance the model's ability to separate background and anomalies. Experiments on five benchmark datasets demonstrate the superiority and applicability of the proposed MBDTNet method, showing that it outperforms the current state-of-the-art hyperspectral anomaly detection methods.
AB - Hyperspectral anomaly detection is aimed at distinguishing between background and anomalous regions in hyperspectral images, and plays a crucial role iSn various applications. However, the existing deep learning methods face challenges when dealing with complex background distributions and insufficient training data. In this article, we propose a novel multi-class background description transformer network (MBDTNet) to address the problems of imprecise background distribution learning and poor anomaly detection. Firstly, we propose an image-level end-to-end data augmentation method based on self-supervised training, which enhances the diversity and quantity of the training samples through adaptive clustering and spatial masking strategies. Secondly, based on the principles of low-rank representation, a sparse self-attention mechanism based on token dictionary representation is designed to help the model focus on key background features and guide the model in recognizing anomalies. Finally, a token dictionary learning mechanism for multi-class background description is established by combining Gaussian discriminant analysis with a conditional distance function, and intra-class and inter-class losses are designed to enhance the model's ability to separate background and anomalies. Experiments on five benchmark datasets demonstrate the superiority and applicability of the proposed MBDTNet method, showing that it outperforms the current state-of-the-art hyperspectral anomaly detection methods.
KW - Data augmentation
KW - Hyperspectral anomaly detection
KW - Multi-class background description
KW - Token dictionary representation
UR - https://www.scopus.com/pages/publications/105007732213
U2 - 10.1016/j.patcog.2025.111945
DO - 10.1016/j.patcog.2025.111945
M3 - 文章
AN - SCOPUS:105007732213
SN - 0031-3203
VL - 169
JO - Pattern Recognition
JF - Pattern Recognition
M1 - 111945
ER -