Improving the multimodal probabilistic semantic model by ELM classifiers

Yu Zhang*, Ye Yuan, Fangda Guo, Yishu Wang, Guoren Wang

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

4 Citations (Scopus)

Abstract

The multi-modal retrieval is considered as performing information retrieval among different modalities of multimedia information. Nowadays, it becomes increasingly important in the information science field. However, it is so difficult to bridge the meanings of different multimedia modalities that the performance of multimodal retrieval is deteriorated now. In this paper, we propose a new mechanism to build the relationship between visual and textual modalities and to verify the multimodal retrieval. Specifically, this mechanism depends on the multimodal binary classifiers based on the Extreme Learning Machine (ELM) to verify whether the answers are related to the query examples. Firstly, we propose the multimodal probabilistic semantic model to rank the answers according to their generative probabilities. Furthermore, we build the multimodal binary classifiers to filter out unrelated answers. The multimodal binary classifiers are called the word classifiers. It can improve the performance of the multimodal probabilistic semantic model. The experimental results show that the multimodal probabilistic semantic model and the word classifiers are effective and efficient. Also they demonstrate that the word classifiers based on ELM not only can improve the performance of the probabilistic semantic model but also can be easily applied to other probabilistic semantic models.

Original languageEnglish
Pages (from-to)1967-1990
Number of pages24
JournalJournal of the Franklin Institute
Volume355
Issue number4
DOIs
Publication statusPublished - Mar 2018
Externally publishedYes

Fingerprint

Dive into the research topics of 'Improving the multimodal probabilistic semantic model by ELM classifiers'. Together they form a unique fingerprint.

Cite this