跳到主要导航 跳到搜索 跳到主要内容

OD-DETR: Online Distillation for Stabilizing Training of Detection Transformer

科研成果: 书/报告/会议事项章节会议稿件同行评审

摘要

DEtection TRansformer (DETR) becomes a dominant paradigm, mainly due to its common architecture with high accuracy and no post-processing. However, DETR suffers from unstable training dynamics. It consumes more data and epochs to converge compared with CNN-based detectors. This paper aims to stabilize DETR training through the online distillation. It utilizes a teacher model, accumulated by Exponential Moving Average (EMA), and distills its knowledge into the online model in following three aspects. First, the matching relation between object queries and ground truth (GT) boxes in the teacher is employed to guide the student, so queries within the student are not only assigned labels based on their own predictions, but also refer to the matching results from the teacher. Second, the teacher's initial query is given to the online student, and its prediction is directly constrained by the corresponding output from the teacher. Finally, the object queries from teacher's different decoder stages are used to build the auxiliary group to accelerate the convergence. For each GT, two queries with the least matching costs are selected into this extra group, and they predict the GT box and participate the optimization. Extensive experiments show that the proposed OD-DETR successfully stabilizes the training, and significantly increases the performance without bringing in more parameters.

源语言英语
主期刊名Proceedings of the 33rd International Joint Conference on Artificial Intelligence, IJCAI 2024
编辑Kate Larson
出版商International Joint Conferences on Artificial Intelligence
1443-1451
页数9
ISBN(电子版)9781956792041
出版状态已出版 - 2024
活动33rd International Joint Conference on Artificial Intelligence, IJCAI 2024 - Jeju, 韩国
期限: 3 8月 20249 8月 2024

出版系列

姓名IJCAI International Joint Conference on Artificial Intelligence
ISSN(印刷版)1045-0823

会议

会议33rd International Joint Conference on Artificial Intelligence, IJCAI 2024
国家/地区韩国
Jeju
时期3/08/249/08/24

指纹

探究 'OD-DETR: Online Distillation for Stabilizing Training of Detection Transformer' 的科研主题。它们共同构成独一无二的指纹。

引用此