TY - GEN
T1 - Compressing Deep Image Super-Resolution Models
AU - Jiang, Yuxuan
AU - Nawala, Jakub T
AU - Zhang, Fan
AU - Bull, David R
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024/6/26
Y1 - 2024/6/26
N2 - Deep learning techniques have been applied in the context of image super-resolution (SR), achieving remarkable advances in terms of reconstruction performance. Existing techniques typically employ highly complex model structures which result in large model sizes and slow inference speeds. This often leads to high energy consumption and restricts their adoption for practical applications. To address this issue, this work employs a three-stage workflow for compressing deep SR models which significantly reduces their memory requirement. Restoration performance has been maintained through teacher-student knowledge distillation using a newly designed distillation loss. We have applied this approach to two popular image super-resolution networks, SwinIR and EDSR, to demonstrate its effectiveness. The resulting compact models, SwinIRmini and EDSRmini, attain an 89% and 96% reduction in both model size and floating-point operations (FLOPs) respectively, compared to their original versions. They also retain competitive super-resolution performance compared to their original models and other commonly used SR approaches. The source code and pretrained models for these two lightweight SR approaches are released at https://pikapi22.github.io/CDISM/.
AB - Deep learning techniques have been applied in the context of image super-resolution (SR), achieving remarkable advances in terms of reconstruction performance. Existing techniques typically employ highly complex model structures which result in large model sizes and slow inference speeds. This often leads to high energy consumption and restricts their adoption for practical applications. To address this issue, this work employs a three-stage workflow for compressing deep SR models which significantly reduces their memory requirement. Restoration performance has been maintained through teacher-student knowledge distillation using a newly designed distillation loss. We have applied this approach to two popular image super-resolution networks, SwinIR and EDSR, to demonstrate its effectiveness. The resulting compact models, SwinIRmini and EDSRmini, attain an 89% and 96% reduction in both model size and floating-point operations (FLOPs) respectively, compared to their original versions. They also retain competitive super-resolution performance compared to their original models and other commonly used SR approaches. The source code and pretrained models for these two lightweight SR approaches are released at https://pikapi22.github.io/CDISM/.
KW - image super-resolution
KW - complexity reduction
KW - model compression
KW - knowledge distillation
UR - https://arxiv.org/abs/2401.00523
UR - https://2024.picturecodingsymposium.org/
U2 - 10.1109/PCS60826.2024.10566374
DO - 10.1109/PCS60826.2024.10566374
M3 - Conference Contribution (Conference Proceeding)
SN - 9798350358490
T3 - Picture Coding Symposium, PCS
BT - 2024 Picture Coding Symposium (PCS)
PB - Institute of Electrical and Electronics Engineers (IEEE)
T2 - Picture Coding Symposium 2024
Y2 - 12 June 2024 through 14 June 2024
ER -