TY - GEN
T1 - Data-Free Low-Bit Quantization via Dynamic Multi-teacher Knowledge Distillation
AU - Huang, Chong
AU - Lin, Shaohui
AU - Zhang, Yan
AU - Li, Ke
AU - Zhang, Baochang
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd 2024.
PY - 2024
Y1 - 2024
N2 - Data-free quantization is an effective way to compress deep neural networks under the situation where training data is unavailable, due to data privacy and security issues. Although Off-the-shelf data-free quantization methods achieve the relatively same accuracy as the fully-precision (FP) models for high-bit (e.g., 8-bit) quantization, low-bit quantization performance drops significantly to restrict their extensive applications. In this paper, we propose a novel data-free low-bit quantization method via Dynamic Multi-teacher Knowledge Distillation (DMKD) to improve the performance of low-bit quantization models. In particular, we first introduce a generator to synthesize the training data based on the input of random noise. The low-bit quantization models are then trained on these synthetic images by the dynamic knowledge from the FP model and the high-bit quantization models, which are balanced by learnable loss weight factors. The factors are controlled by a tiny learnable FP network to adaptively allocate the balanced weights for the knowledge from the FP model and the high-bit quantization models during training. For inference, we only kept the low-bit quantization model by safely removing other additional networks, such as the generator and the tiny model. Extensive experiments demonstrate the effectiveness of DMKD for low-bit quantization of widely-used convolutional neural networks (CNNs) on different benchmark datasets. Our DMKD ooon methods.
AB - Data-free quantization is an effective way to compress deep neural networks under the situation where training data is unavailable, due to data privacy and security issues. Although Off-the-shelf data-free quantization methods achieve the relatively same accuracy as the fully-precision (FP) models for high-bit (e.g., 8-bit) quantization, low-bit quantization performance drops significantly to restrict their extensive applications. In this paper, we propose a novel data-free low-bit quantization method via Dynamic Multi-teacher Knowledge Distillation (DMKD) to improve the performance of low-bit quantization models. In particular, we first introduce a generator to synthesize the training data based on the input of random noise. The low-bit quantization models are then trained on these synthetic images by the dynamic knowledge from the FP model and the high-bit quantization models, which are balanced by learnable loss weight factors. The factors are controlled by a tiny learnable FP network to adaptively allocate the balanced weights for the knowledge from the FP model and the high-bit quantization models during training. For inference, we only kept the low-bit quantization model by safely removing other additional networks, such as the generator and the tiny model. Extensive experiments demonstrate the effectiveness of DMKD for low-bit quantization of widely-used convolutional neural networks (CNNs) on different benchmark datasets. Our DMKD ooon methods.
KW - Data-Free Quantization
KW - Dynamic knowledge
KW - Knowledge Distillation
KW - Loss Weight Factors
UR - https://www.scopus.com/pages/publications/85181978030
U2 - 10.1007/978-981-99-8543-2_3
DO - 10.1007/978-981-99-8543-2_3
M3 - 会议稿件
AN - SCOPUS:85181978030
SN - 9789819985425
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 28
EP - 41
BT - Pattern Recognition and Computer Vision - 6th Chinese Conference, PRCV 2023, Proceedings
A2 - Sheng, Bin
A2 - Bi, Lei
A2 - Kim, Jinman
A2 - Magnenat-Thalmann, Nadia
A2 - Thalmann, Daniel
PB - Springer Science and Business Media Deutschland GmbH
T2 - 6th Chinese Conference on Pattern Recognition and Computer Vision, PRCV 2023
Y2 - 13 October 2023 through 15 October 2023
ER -