跳到主要导航 跳到搜索 跳到主要内容

Variational Self-Distillation for Remote Sensing Scene Classification

  • Yutao Hu
  • , Xin Huang
  • , Xiaoyan Luo*
  • , Jungong Han
  • , Xianbin Cao
  • , Jun Zhang
  • *此作品的通讯作者
  • Beihang University
  • Aberystwyth University
  • Ministry of Industry and Information Technology
  • Beijing Institute of Technology

科研成果: 期刊稿件文章同行评审

摘要

Supported by deep learning techniques, remote sensing scene classification, a fundamental task in remote image analysis, has recently obtained remarkable progress. However, due to the severe uncertainty and perturbation within an image, it is still a challenging task and remains many unsolved problems. In this article, we note that regular one-hot labels cannot precisely describe remote sensing images, and they fail to provide enough information for supervision and limiting the discriminative feature learning of the network. To solve this problem, we propose a variational self-distillation network (VSDNet), in which the class entanglement information from the prediction vector acts as the supplement to the category information. Then, the exploited information is hierarchically distilled from the deep layers into the shallow parts via a variational knowledge transfer (VKT) module. Notably, the VKT module performs knowledge distillation in a probabilistic way through variational estimation, which enables end-to-end optimization for mutual information and promotes robustness to uncertainty within the image. Extensive experiments on four challenging remote sensing datasets demonstrate that, with a negligible parameter increase, the proposed VSDNet brings a significant performance improvement over different backbone networks and delivers state-of-the-art results.

源语言英语
文章编号5627313
期刊IEEE Transactions on Geoscience and Remote Sensing
60
DOI
出版状态已出版 - 2022
已对外发布

指纹

探究 'Variational Self-Distillation for Remote Sensing Scene Classification' 的科研主题。它们共同构成独一无二的指纹。

引用此