Popeye: A Unified Visual-Language Model for Multisource Ship Detection From Remote Sensing Imagery

Wei Zhang, Miaoxin Cai, Tong Zhang, Guoqiang Lei, Yin Zhuang*, Xuerui Mao*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)

Abstract

Ship detection needs to identify ship locations from remote sensing scenes. Due to different imaging payloads, various appearances of ships, and complicated background interference from the bird's eye view, it is difficult to setup a unified paradigm for achieving multisource ship detection. To address this challenge, in this article, leveraging the large language models powerful generalization ability, a unified visual-language model called Popeye is proposed for multisource ship detection from RS imagery. Specifically, to bridge the interpretation gap across the multisource images for ship detection, a novel unified labeling paradigm is designed to integrate different visual modalities and the various ship detection ways, i.e., horizontal bounding box and oriented bounding box. Subsequently, the hybrid experts encoder is designed to refine multiscale visual features, thereby enhancing visual perception. Then, a visual-language alignment method is developed for Popeye to enhance interactive comprehension ability between visual and language content. Furthermore, an instruction adaption mechanism is proposed for transferring the pretrained visual-language knowledge from the nature scene into the RS domain for multisource ship detection. In addition, the segment anything model is also seamlessly integrated into the proposed Popeye to achieve pixel-level ship segmentation without additional training costs. Finally, extensive experiments are conducted on the newly constructed ship instruction dataset named MMShip, and the results indicate that the proposed Popeye outperforms current specialist, open-vocabulary, and other visual-language models in zero-shot multisource various ship detection tasks.

Original languageEnglish
Pages (from-to)20050-20063
Number of pages14
JournalIEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing
Volume17
DOIs
Publication statusPublished - 2024

Keywords

  • Multisource imagery
  • natural language interaction
  • ship detection
  • visual-language alignment

Fingerprint

Dive into the research topics of 'Popeye: A Unified Visual-Language Model for Multisource Ship Detection From Remote Sensing Imagery'. Together they form a unique fingerprint.

Cite this

Zhang, W., Cai, M., Zhang, T., Lei, G., Zhuang, Y., & Mao, X. (2024). Popeye: A Unified Visual-Language Model for Multisource Ship Detection From Remote Sensing Imagery. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 17, 20050-20063. https://doi.org/10.1109/JSTARS.2024.3488034