A closer look at branch classifiers of multi-exit architectures

  • Shaohui Lin
  • , Bo Ji
  • , Rongrong Ji
  • , Angela Yao*
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Multi-exit architectures consist of a backbone and branch classifiers that offer shortened inference pathways to reduce the run-time of deep neural networks. In this paper, we analyze different branching patterns that vary in their allocation of computational complexity for the branch classifiers. Constant-complexity branching keeps all branches the same, while complexity-increasing and complexity-decreasing branching place more complex branches later or earlier in the backbone respectively. Through extensive experimentation on multiple backbones and datasets, we find that complexity-decreasing branches are more effective than constant-complexity or complexity-increasing branches, which achieve the best accuracy-cost trade-off. We investigate a cause by using knowledge consistency to probe the effect of adding branches onto a backbone. Our findings show that complexity-decreasing branching yields the least disruption to the feature abstraction hierarchy of the backbone, which explains the effectiveness of the branching patterns.

Original languageEnglish
Article number103900
JournalComputer Vision and Image Understanding
Volume239
DOIs
StatePublished - Feb 2024

Keywords

  • Branch classifiers
  • Knowledge consistency
  • Model compression and acceleration
  • Multi-exit architectures

Fingerprint

Dive into the research topics of 'A closer look at branch classifiers of multi-exit architectures'. Together they form a unique fingerprint.

Cite this