Dynamically visual disambiguation of keyword-based image search

Yazhou Yao, Zeren Sun, Fumin Shen*, Li Liu, Limin Wang, Fan Zhu, Lizhong Ding, Gangshan Wu, Ling Shao

*此作品的通讯作者

科研成果: 书/报告/会议事项章节会议稿件同行评审

13 引用 (Scopus)

摘要

Due to the high cost of manual annotation, learning directly from the web has attracted broad attention. One issue that limits their performance is the problem of visual polysemy. To address this issue, we present an adaptive multi-model framework that resolves polysemy by visual disambiguation. Compared to existing methods, the primary advantage of our approach lies in that our approach can adapt to the dynamic changes in the search results. Our proposed framework consists of two major steps: we first discover and dynamically select the text queries according to the image search results, then we employ the proposed saliency-guided deep multi-instance learning network to remove outliers and learn classification models for visual disambiguation. Extensive experiments demonstrate the superiority of our proposed approach.

源语言英语
主期刊名Proceedings of the 28th International Joint Conference on Artificial Intelligence, IJCAI 2019
编辑Sarit Kraus
出版商International Joint Conferences on Artificial Intelligence
996-1002
页数7
ISBN(电子版)9780999241141
DOI
出版状态已出版 - 2019
已对外发布
活动28th International Joint Conference on Artificial Intelligence, IJCAI 2019 - Macao, 中国
期限: 10 8月 201916 8月 2019

出版系列

姓名IJCAI International Joint Conference on Artificial Intelligence
2019-August
ISSN(印刷版)1045-0823

会议

会议28th International Joint Conference on Artificial Intelligence, IJCAI 2019
国家/地区中国
Macao
时期10/08/1916/08/19

指纹

探究 'Dynamically visual disambiguation of keyword-based image search' 的科研主题。它们共同构成独一无二的指纹。

引用此