TY - JOUR
T1 - A closer look at branch classifiers of multi-exit architectures
AU - Lin, Shaohui
AU - Ji, Bo
AU - Ji, Rongrong
AU - Yao, Angela
N1 - Publisher Copyright:
© 2023 Elsevier Inc.
PY - 2024/2
Y1 - 2024/2
N2 - Multi-exit architectures consist of a backbone and branch classifiers that offer shortened inference pathways to reduce the run-time of deep neural networks. In this paper, we analyze different branching patterns that vary in their allocation of computational complexity for the branch classifiers. Constant-complexity branching keeps all branches the same, while complexity-increasing and complexity-decreasing branching place more complex branches later or earlier in the backbone respectively. Through extensive experimentation on multiple backbones and datasets, we find that complexity-decreasing branches are more effective than constant-complexity or complexity-increasing branches, which achieve the best accuracy-cost trade-off. We investigate a cause by using knowledge consistency to probe the effect of adding branches onto a backbone. Our findings show that complexity-decreasing branching yields the least disruption to the feature abstraction hierarchy of the backbone, which explains the effectiveness of the branching patterns.
AB - Multi-exit architectures consist of a backbone and branch classifiers that offer shortened inference pathways to reduce the run-time of deep neural networks. In this paper, we analyze different branching patterns that vary in their allocation of computational complexity for the branch classifiers. Constant-complexity branching keeps all branches the same, while complexity-increasing and complexity-decreasing branching place more complex branches later or earlier in the backbone respectively. Through extensive experimentation on multiple backbones and datasets, we find that complexity-decreasing branches are more effective than constant-complexity or complexity-increasing branches, which achieve the best accuracy-cost trade-off. We investigate a cause by using knowledge consistency to probe the effect of adding branches onto a backbone. Our findings show that complexity-decreasing branching yields the least disruption to the feature abstraction hierarchy of the backbone, which explains the effectiveness of the branching patterns.
KW - Branch classifiers
KW - Knowledge consistency
KW - Model compression and acceleration
KW - Multi-exit architectures
UR - https://www.scopus.com/pages/publications/85179838499
U2 - 10.1016/j.cviu.2023.103900
DO - 10.1016/j.cviu.2023.103900
M3 - 文章
AN - SCOPUS:85179838499
SN - 1077-3142
VL - 239
JO - Computer Vision and Image Understanding
JF - Computer Vision and Image Understanding
M1 - 103900
ER -