Single image de-raining using spinning detail perceptual generative adversarial networks

Kaizheng Chen, Yaping Dai, Zhiyang Jia*, Kaoru Hirota

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)

Abstract

In this paper, Spinning Detail Perceptual Generative Adversarial Networks (SDP-GAN) is proposed for single image de-raining. The proposed method adopts the Generative Adversarial Network (GAN) framework and consists of two following networks: the rain streaks generative network G and the discriminative network D. To reduce the background interference, we propose a rain streaks generative network which not only focuses on the high frequency detail map of rainy image, but also directly reduces the mapping range from input to output. To further improve the perceptual quality of generated images, we modify the perceptual loss by extracting high-level features from discriminative network D, rather than pre-trained networks. Furthermore, we introduce a new training procedure based on the notion of self spinning to improve the final de-raining performance. Extensive experiments on the synthetic and real-world datasets demonstrate that the proposed method achieves significant improvements over the recent state-of-the-art methods.

Original languageEnglish
Pages (from-to)811-819
Number of pages9
JournalJournal of Advanced Computational Intelligence and Intelligent Informatics
Volume24
Issue number7
DOIs
Publication statusPublished - 20 Dec 2020

Keywords

  • Detail map
  • Generative adversarial networks
  • Image de-raining
  • Perceptual loss
  • Self spinning

Fingerprint

Dive into the research topics of 'Single image de-raining using spinning detail perceptual generative adversarial networks'. Together they form a unique fingerprint.

Cite this