Knowledge Distillation Based on Transformed Teacher Matching

17 Feb 2024  ·  Kaixiang Zheng, En-hui Yang ·

As a technique to bridge logit matching and probability distribution matching, temperature scaling plays a pivotal role in knowledge distillation (KD). Conventionally, temperature scaling is applied to both teacher's logits and student's logits in KD. Motivated by some recent works, in this paper, we drop instead temperature scaling on the student side, and systematically study the resulting variant of KD, dubbed transformed teacher matching (TTM). By reinterpreting temperature scaling as a power transform of probability distribution, we show that in comparison with the original KD, TTM has an inherent R\'enyi entropy term in its objective function, which serves as an extra regularization term. Extensive experiment results demonstrate that thanks to this inherent regularization, TTM leads to trained students with better generalization than the original KD. To further enhance student's capability to match teacher's power transformed probability distribution, we introduce a sample-adaptive weighting coefficient into TTM, yielding a novel distillation approach dubbed weighted TTM (WTTM). It is shown, by comprehensive experiments, that although WTTM is simple, it is effective, improves upon TTM, and achieves state-of-the-art accuracy performance. Our source code is available at https://github.com/zkxufo/TTM.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Knowledge Distillation ImageNet WTTM (T:resnet50, S:mobilenet-v1) Top-1 accuracy % 73.09 # 15
Knowledge Distillation ImageNet WTTM (T: DeiT III-Small S:DeiT-Tiny) Top-1 accuracy % 77.03 # 12
CRD training setting # 1
Knowledge Distillation ImageNet WTTM (T: ResNet-34 S:ResNet-18) Top-1 accuracy % 72.19 # 21
CRD training setting # 1

Methods