Please use this identifier to cite or link to this item:
|Title:||Utilizing related samples to enhance interactive concept-based video search|
|Source:||Yuan, J., Zha, Z.-J., Zheng, Y.-T., Wang, M., Zhou, X., Chua, T.-S. (2011). Utilizing related samples to enhance interactive concept-based video search. IEEE Transactions on Multimedia 13 (6) : 1343-1355. ScholarBank@NUS Repository. https://doi.org/10.1109/TMM.2011.2168813|
|Abstract:||One of the main challenges in interactive concept-based video search is the problem of insufficient relevant samples, especially for queries with complex semantics. In this paper, related samples are exploited to enhance interactive video search. The related samples refer to those video segments that are relevant to part of the query rather than the entire query. Compared to the relevant samples which may be rare, the related samples are usually plentiful and easy to find in search results. Generally, the related samples are visually similar and temporally neighboring to the relevant samples. Based on these two characters, we develop a visual ranking model that simultaneously exploits the relevant, related, and irrelevant samples, as well as a temporal ranking model to leverage the temporal relationship between related and relevant samples. An adaptive fusion method is then proposed to optimally explore these two ranking models to generate search results. We conduct extensive experiments on two real-world video datasets: TRECVID 2008 and YouTube datasets. As the experimental results show, our approach achieves at least 96% and 167% performance improvements against the state-of-the-art approaches on the TRECVID 2008 and YouTube datasets, respectively. © 2006 IEEE.|
|Source Title:||IEEE Transactions on Multimedia|
|Appears in Collections:||Staff Publications|
Show full item record
Files in This Item:
There are no files associated with this item.
checked on Feb 22, 2018
WEB OF SCIENCETM
checked on Jan 24, 2018
checked on Feb 19, 2018
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.