TY - JOUR
T1 - CrossU-Net
T2 - Dual-modality cross-attention U-Net for segmentation of precancerous lesions in gastric cancer
AU - Wang, Jiansheng
AU - Zhang, Benyan
AU - Wang, Yan
AU - Zhou, Chunhua
AU - Vonsky, Maxim S.
AU - Mitrofanova, Lubov B.
AU - Zou, Duowu
AU - Li, Qingli
N1 - Publisher Copyright:
© 2024 Elsevier Ltd
PY - 2024/3
Y1 - 2024/3
N2 - Gastric precancerous lesions (GPL) significantly elevate the risk of gastric cancer, and precise diagnosis and timely intervention are critical for patient survival. Due to the elusive pathological features of precancerous lesions, the early detection rate is less than 10%, which hinders lesion localization and diagnosis. In this paper, we provide a GPL pathological dataset and propose a novel method for improving the segmentation accuracy on a limited-scale dataset, namely RGB and Hyperspectral dual-modal pathological image Cross-attention U-Net (CrossU-Net). Specifically, we present a self-supervised pre-training model for hyperspectral images to serve downstream segmentation tasks. Secondly, we design a dual-stream U-Net-based network to extract features from different modal images. To promote information exchange between spatial information in RGB images and spectral information in hyperspectral images, we customize the cross-attention mechanism between the two networks. Furthermore, we use an intermediate agent in this mechanism to improve computational efficiency. Finally, we add a distillation loss to align predicted results for both branches, improving network generalization. Experimental results show that our CrossU-Net achieves accuracy and Dice of 96.53% and 91.62%, respectively, for GPL lesion segmentation, providing a promising spectral research approach for the localization and subsequent quantitative analysis of pathological features in early diagnosis.
AB - Gastric precancerous lesions (GPL) significantly elevate the risk of gastric cancer, and precise diagnosis and timely intervention are critical for patient survival. Due to the elusive pathological features of precancerous lesions, the early detection rate is less than 10%, which hinders lesion localization and diagnosis. In this paper, we provide a GPL pathological dataset and propose a novel method for improving the segmentation accuracy on a limited-scale dataset, namely RGB and Hyperspectral dual-modal pathological image Cross-attention U-Net (CrossU-Net). Specifically, we present a self-supervised pre-training model for hyperspectral images to serve downstream segmentation tasks. Secondly, we design a dual-stream U-Net-based network to extract features from different modal images. To promote information exchange between spatial information in RGB images and spectral information in hyperspectral images, we customize the cross-attention mechanism between the two networks. Furthermore, we use an intermediate agent in this mechanism to improve computational efficiency. Finally, we add a distillation loss to align predicted results for both branches, improving network generalization. Experimental results show that our CrossU-Net achieves accuracy and Dice of 96.53% and 91.62%, respectively, for GPL lesion segmentation, providing a promising spectral research approach for the localization and subsequent quantitative analysis of pathological features in early diagnosis.
KW - Cross attention
KW - Dual-modality
KW - Hyperspectral imaging
KW - Image segmentation
UR - https://www.scopus.com/pages/publications/85183458176
U2 - 10.1016/j.compmedimag.2024.102339
DO - 10.1016/j.compmedimag.2024.102339
M3 - 文章
C2 - 38262134
AN - SCOPUS:85183458176
SN - 0895-6111
VL - 112
JO - Computerized Medical Imaging and Graphics
JF - Computerized Medical Imaging and Graphics
M1 - 102339
ER -