TY - GEN
T1 - GSMC
T2 - 41st Computer Graphics International Conference, CGI 2024
AU - Huang, Yongqi
AU - Liu, Feng
AU - Zhou, Aimin
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2025.
PY - 2025
Y1 - 2025
N2 - Due to the ambiguity of facial expressions, the subjectivity of annotators, and the low quality of images, noisy labels pose a significant challenge in the field of facial expression recognition. We address the noise label problem based on the principle of contrastive learning. Through CAM visualization, we found that when noisy labels are present, the model can make correct predictions due to its inability to memorize noisy labels from different data augmentation perspectives. Inspired by this observation, we propose the GSMC method, which enforces model consistency between the predictions of the original image and its augmented versions. Specifically, we divide the consistency requirement into two different tasks. The first task uses attention map consistency, emphasizing preventing the model from memorizing noise labels. The second task employs self-distillation learning, where the consistency is achieved through the predictions of student and self-distillation networks, enabling the model to learn more robust label distributions. Our framework, based on ResNet18, achieves a 90.54% accuracy on the RAF-DB dataset, achieving State-Of-The-Art performance. The code will be available at https://github.com/ECNU-Cross-Innovation-Lab/GSMC.
AB - Due to the ambiguity of facial expressions, the subjectivity of annotators, and the low quality of images, noisy labels pose a significant challenge in the field of facial expression recognition. We address the noise label problem based on the principle of contrastive learning. Through CAM visualization, we found that when noisy labels are present, the model can make correct predictions due to its inability to memorize noisy labels from different data augmentation perspectives. Inspired by this observation, we propose the GSMC method, which enforces model consistency between the predictions of the original image and its augmented versions. Specifically, we divide the consistency requirement into two different tasks. The first task uses attention map consistency, emphasizing preventing the model from memorizing noise labels. The second task employs self-distillation learning, where the consistency is achieved through the predictions of student and self-distillation networks, enabling the model to learn more robust label distributions. Our framework, based on ResNet18, achieves a 90.54% accuracy on the RAF-DB dataset, achieving State-Of-The-Art performance. The code will be available at https://github.com/ECNU-Cross-Innovation-Lab/GSMC.
KW - attention consistency
KW - computational perception
KW - contrastive learning
KW - emotion recognition
KW - noisy labels
UR - https://www.scopus.com/pages/publications/86000439276
U2 - 10.1007/978-3-031-81806-6_22
DO - 10.1007/978-3-031-81806-6_22
M3 - 会议稿件
AN - SCOPUS:86000439276
SN - 9783031818059
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 297
EP - 308
BT - Advances in Computer Graphics - 41st Computer Graphics International Conference, CGI 2024, Proceedings
A2 - Magnenat-Thalmann, Nadia
A2 - Kim, Jinman
A2 - Sheng, Bin
A2 - Deng, Zhigang
A2 - Thalmann, Daniel
A2 - Li, Ping
PB - Springer Science and Business Media Deutschland GmbH
Y2 - 1 July 2024 through 5 July 2024
ER -