Over-parameterized Student Model via Tensor Decomposition Boosted Knowledge Distillation
Increased training parameters have enabled large pre-trained models to excel in various downstream tasks. Nevertheless, the extensive computational requirements associated with these models hinder their widespread adoption within the community. We focus on Knowledge Distillation (KD), where a compac...
Saved in:
Main Authors | , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
10.11.2024
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Increased training parameters have enabled large pre-trained models to excel
in various downstream tasks. Nevertheless, the extensive computational
requirements associated with these models hinder their widespread adoption
within the community. We focus on Knowledge Distillation (KD), where a compact
student model is trained to mimic a larger teacher model, facilitating the
transfer of knowledge of large models. In contrast to much of the previous
work, we scale up the parameters of the student model during training, to
benefit from overparameterization without increasing the inference latency. In
particular, we propose a tensor decomposition strategy that effectively
over-parameterizes the relatively small student model through an efficient and
nearly lossless decomposition of its parameter matrices into higher-dimensional
tensors. To ensure efficiency, we further introduce a tensor constraint loss to
align the high-dimensional tensors between the student and teacher models.
Comprehensive experiments validate the significant performance enhancement by
our approach in various KD tasks, covering computer vision and natural language
processing areas. Our code is available at
https://github.com/intell-sci-comput/OPDF. |
---|---|
DOI: | 10.48550/arxiv.2411.06448 |