Semantic description of image based on I-NiC model

Chaoying Zhang, Yaping Dai, Hao Wang, Zhiyang Jia*, Kaoru Hirota

*此作品的通讯作者

科研成果: 会议稿件论文同行评审

摘要

In order to address the problems of misprediction and object missing in semantic description of image, an improved Neural Image Caption (I-NIC) model is proposed. It primarily consists of the Convolutional Neural Network (CNN) and Recurrent Neural Network (RNN). The model uses Inception-v4 model developed by Google to extract image features and iteratively optimizes the training process parameters through word-based loss function. Therefore, the I-NIC model can generate more relevant descriptions and improve the accuracy and efficiency of the system. Compared with NIC model, the experiment results show that the accuracy of I-NIC model is improved by 2.5% with BLEU-4 metrics, 1.2% with METEOR metrics and 7.5% with CIDEr metrics on the Microsoft COCO Caption dataset.

会议

会议8th International Symposium on Computational Intelligence and Industrial Applications and 12th China-Japan International Workshop on Information Technology and Control Applications, ISCIIA and ITCA 2018
国家/地区中国
Tengzhou, Shandong
时期2/11/186/11/18

指纹

探究 'Semantic description of image based on I-NiC model' 的科研主题。它们共同构成独一无二的指纹。

引用此