TY - JOUR
T1 - CCSD
T2 - cross-camera self-distillation for unsupervised person re-identification
AU - Chen, Jiyuan
AU - Gao, Changxin
AU - Sun, Li
AU - Sang, Nong
N1 - Publisher Copyright:
© The Author(s) 2023.
PY - 2023/12
Y1 - 2023/12
N2 - Existing unsupervised person re-identification (Re-ID) methods have achieved remarkable performance by adopting an alternate clustering-training manner. However, they still suffer from camera variation, which results in an inconsistent feature space and unreliable pseudo labels that severely degrade the performance. In this paper, we propose a cross-camera self-distillation (CCSD) framework for unsupervised person Re-ID to alleviate the effect of camera variation. Specifically, in the clustering phase, we propose a camera-aware cluster refinement mechanism, which first splits each cluster into multiple clusters according to the camera views, and then refines them into more compact clusters. In the training phase, we first obtain the similarity between the samples and the refined clusters from the same and different cameras, and then transfer the knowledge of similarity distribution from intra-camera to cross-camera. Since the intra-camera similarity is free from camera variation, our knowledge distillation approach is able to learn a more consistent feature space across cameras. Extensive experiments demonstrate the superiority of our proposed CCSD against the state-of-the-art approaches on unsupervised person Re-ID.
AB - Existing unsupervised person re-identification (Re-ID) methods have achieved remarkable performance by adopting an alternate clustering-training manner. However, they still suffer from camera variation, which results in an inconsistent feature space and unreliable pseudo labels that severely degrade the performance. In this paper, we propose a cross-camera self-distillation (CCSD) framework for unsupervised person Re-ID to alleviate the effect of camera variation. Specifically, in the clustering phase, we propose a camera-aware cluster refinement mechanism, which first splits each cluster into multiple clusters according to the camera views, and then refines them into more compact clusters. In the training phase, we first obtain the similarity between the samples and the refined clusters from the same and different cameras, and then transfer the knowledge of similarity distribution from intra-camera to cross-camera. Since the intra-camera similarity is free from camera variation, our knowledge distillation approach is able to learn a more consistent feature space across cameras. Extensive experiments demonstrate the superiority of our proposed CCSD against the state-of-the-art approaches on unsupervised person Re-ID.
KW - Deep learning
KW - Knowledge distillation
KW - Person re-identification
KW - Unsupervised learning
UR - https://www.scopus.com/pages/publications/85190432188
U2 - 10.1007/s44267-023-00029-4
DO - 10.1007/s44267-023-00029-4
M3 - 文章
AN - SCOPUS:85190432188
SN - 2097-3330
VL - 1
JO - Visual Intelligence
JF - Visual Intelligence
IS - 1
M1 - 27
ER -