Toward Effective Knowledge Distillation for Fine-Grained Object Recognition in Remote Sensing

Yangte Gao*, Chenwei Deng, Liang Chen

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

摘要

With advancements in on-board computing devices deployed on remote sensing platforms, the demand for efficiently processing remote sensing imagery has become increasingly prominent. Knowledge distillation, as an effective lightweight method, has been introduced into this domain. Intuitively, distillation from a larger teacher model is expected to yield better performance. However, in our investigation of fine-grained object recognition in remote sensing imagery, we observed a counter-intuitive phenomenon: as the size of the teacher model increases, the performance of the student model initially improves but then degrades. This capacity gap issue hinders effective utilization of stronger teacher models. To address this issue, we propose a novel distillation framework named BL-KD. It integrates two tailored components: the class-level learnable orthogonal projection (CLOP) module and the object rebalance (ORB) module, which are jointly optimized to mitigate the negative impact of the capacity gap while effectively adapting to the unique distributional patterns and challenges inherent in remote sensing imagery. Experiments conducted on multiple fine-grained object recognition tasks in remote sensing demonstrate that our method consistently improves student performance, particularly in scenarios involving large teacher–student gaps, and outperforms several widely used distillation baselines.

源语言英语
文章编号6011405
期刊IEEE Geoscience and Remote Sensing Letters
22
DOI
出版状态已出版 - 2025
已对外发布

指纹

探究 'Toward Effective Knowledge Distillation for Fine-Grained Object Recognition in Remote Sensing' 的科研主题。它们共同构成独一无二的指纹。

引用此