TY - GEN
T1 - Incremental Learning Based on Dual-Branch Network
AU - Dong, Mingda
AU - Zhang, Zhizhong
AU - Xie, Yuan
N1 - Publisher Copyright:
© 2024, The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
PY - 2024
Y1 - 2024
N2 - Incremental learning aims to overcome catastrophic forgetting. When the model learns multiple tasks sequentially, due to the imbalance of new and old classes numbers, the knowledge of old classes stored in the model is destroyed by large number of new classes. The existing single-backbone model is difficult to avoid catastrophic forgetting. In this paper, we proposes to use the dual-branch network model to learn new tasks to alleviate catastrophic forgetting. Different from previous dual-branch models that learn tasks in parallel, we propose to use dual-branch network to learn tasks serially. The model creates a new backbone for learning the remaining tasks, and freezes the previous backbone. In this way, the model can reduce damage to the previous backbone parameters used to learn old tasks. The model uses knowledge distillation to preserve the information of old tasks when the model learns new tasks. We also analyze different distillation methods for the dual-branch network model. In this paper we mainly focuses on the more challenging class incremental learning. We use common incremental learning setting on the ImageNet-100 dataset. The experimental results show that the accuracy can be improved by using the dual-branch network.
AB - Incremental learning aims to overcome catastrophic forgetting. When the model learns multiple tasks sequentially, due to the imbalance of new and old classes numbers, the knowledge of old classes stored in the model is destroyed by large number of new classes. The existing single-backbone model is difficult to avoid catastrophic forgetting. In this paper, we proposes to use the dual-branch network model to learn new tasks to alleviate catastrophic forgetting. Different from previous dual-branch models that learn tasks in parallel, we propose to use dual-branch network to learn tasks serially. The model creates a new backbone for learning the remaining tasks, and freezes the previous backbone. In this way, the model can reduce damage to the previous backbone parameters used to learn old tasks. The model uses knowledge distillation to preserve the information of old tasks when the model learns new tasks. We also analyze different distillation methods for the dual-branch network model. In this paper we mainly focuses on the more challenging class incremental learning. We use common incremental learning setting on the ImageNet-100 dataset. The experimental results show that the accuracy can be improved by using the dual-branch network.
KW - catastrophic forgetting
KW - incremental learning
KW - knowledge distillation
UR - https://www.scopus.com/pages/publications/85180767336
U2 - 10.1007/978-981-99-8435-0_21
DO - 10.1007/978-981-99-8435-0_21
M3 - 会议稿件
AN - SCOPUS:85180767336
SN - 9789819984343
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 263
EP - 272
BT - Pattern Recognition and Computer Vision - 6th Chinese Conference, PRCV 2023, Proceedings
A2 - Liu, Qingshan
A2 - Wang, Hanzi
A2 - Ji, Rongrong
A2 - Ma, Zhanyu
A2 - Zheng, Weishi
A2 - Zha, Hongbin
A2 - Chen, Xilin
A2 - Wang, Liang
PB - Springer Science and Business Media Deutschland GmbH
T2 - 6th Chinese Conference on Pattern Recognition and Computer Vision, PRCV 2023
Y2 - 13 October 2023 through 15 October 2023
ER -