Knowledge Distillation with Multi-granularity Mixture of Priors for Image Super-Resolution

4
citations
#2351
in ICLR 2025
of 3827 papers
8
Top Authors
8
Data Points

Abstract

Knowledge distillation (KD) is a promising yet challenging model compression technique that transfers rich learning representations from a well-performing but cumbersome teacher model to a compact student model. Previous methods for image super-resolution (SR) mostly compare the feature maps directly or after standardizing the dimensions with basic algebraic operations (e.g. average, dot-product). However, the intrinsic semantic differences among feature maps are overlooked, which are caused by the disparate expressive capacity between the networks. This work presents MiPKD, a multi-granularity mixture of prior KD framework, to facilitate efficient SR model through the feature mixture in a unified latent space and stochastic network block mixture. Extensive experiments demonstrate the effectiveness of the proposed MiPKD method.

Citation History

Jan 26, 2026
0
Jan 27, 2026
0
Jan 27, 2026
0
Feb 1, 2026
4+4
Feb 6, 2026
4
Feb 13, 2026
4
Feb 13, 2026
4
Feb 13, 2026
4