Semantic description of image based on I-NiC model

Chaoying Zhang, Yaping Dai, Hao Wang, Zhiyang Jia*, Kaoru Hirota

*Corresponding author for this work

Research output: Contribution to conferencePaperpeer-review

Abstract

In order to address the problems of misprediction and object missing in semantic description of image, an improved Neural Image Caption (I-NIC) model is proposed. It primarily consists of the Convolutional Neural Network (CNN) and Recurrent Neural Network (RNN). The model uses Inception-v4 model developed by Google to extract image features and iteratively optimizes the training process parameters through word-based loss function. Therefore, the I-NIC model can generate more relevant descriptions and improve the accuracy and efficiency of the system. Compared with NIC model, the experiment results show that the accuracy of I-NIC model is improved by 2.5% with BLEU-4 metrics, 1.2% with METEOR metrics and 7.5% with CIDEr metrics on the Microsoft COCO Caption dataset.

Conference

Conference8th International Symposium on Computational Intelligence and Industrial Applications and 12th China-Japan International Workshop on Information Technology and Control Applications, ISCIIA and ITCA 2018
Country/TerritoryChina
CityTengzhou, Shandong
Period2/11/186/11/18

Keywords

  • Convolutional Neural Network
  • Long Short-Term Memory
  • Neural Networks
  • Semantic Description of Image

Fingerprint

Dive into the research topics of 'Semantic description of image based on I-NiC model'. Together they form a unique fingerprint.

Cite this