Q-A2NN: Quantized All-Adder Neural Networks for Onboard Remote Sensing Scene Classification

Ning Zhang, He Chen*, Liang Chen, Jue Wang, Guoqing Wang, Wenchao Liu

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Plum Print visual indicator of research metrics
  • Captures
    • Readers: 3
  • Mentions
    • Blog Mentions: 1
    • News Mentions: 1
see details

Abstract

Performing remote sensing scene classification (RSSC) directly on satellites can alleviate data downlink burdens and reduce latency. Compared to convolutional neural networks (CNNs), the all-adder neural network (A2NN) is a novel basic neural network that is more suitable for onboard RSSC, enabling lower computational overhead by eliminating multiplication operations in convolutional layers. However, the extensive floating-point data and operations in A2NNs still lead to significant storage overhead and power consumption during hardware deployment. In this article, a shared scaling factor-based de-biasing quantization (SSDQ) method tailored for the quantization of A2NNs is proposed to address this issue, including a powers-of-two (POT)-based shared scaling factor quantization scheme and a multi-dimensional de-biasing (MDD) quantization strategy. Specifically, the POT-based shared scaling factor quantization scheme converts the adder filters in A2NNs to quantized adder filters with hardware-friendly integer input activations, weights, and operations. Thus, quantized A2NNs (Q-A2NNs) composed of quantized adder filters have lower computational and memory overheads than A2NNs, increasing their utility in hardware deployment. Although low-bit-width Q-A2NNs exhibit significantly reduced RSSC accuracy compared to A2NNs, this issue can be alleviated by employing the proposed MDD quantization strategy, which combines a weight-debiasing (WD) strategy, which reduces performance degradation due to deviations in the quantized weights, with a feature-debiasing (FD) strategy, which enhances the classification performance of Q-A2NNs through minimizing deviations among the output features of each layer. Extensive experiments and analyses demonstrate that the proposed SSDQ method can efficiently quantize A2NNs to obtain Q-A2NNs with low computational and memory overheads while maintaining comparable performance to A2NNs, thus having high potential for onboard RSSC.

Original languageEnglish
Article number2403
JournalRemote Sensing
Volume16
Issue number13
DOIs
Publication statusPublished - Jul 2024

Keywords

  • all-adder neural network
  • deep learning
  • onboard processing
  • quantization
  • scene classification

Fingerprint

Dive into the research topics of 'Q-A2NN: Quantized All-Adder Neural Networks for Onboard Remote Sensing Scene Classification'. Together they form a unique fingerprint.

Cite this

Zhang, N., Chen, H., Chen, L., Wang, J., Wang, G., & Liu, W. (2024). Q-A2NN: Quantized All-Adder Neural Networks for Onboard Remote Sensing Scene Classification. Remote Sensing, 16(13), Article 2403. https://doi.org/10.3390/rs16132403