Please use this identifier to cite or link to this item:
|Title:||Static saliency vs. Dynamic saliency: A comparative study|
|Citation:||Nguyen, T.V.,Xu, M.,Gao, G.,Kankanhalli, M.,Tian, Q.,Yan, S. (2013). Static saliency vs. Dynamic saliency: A comparative study. MM 2013 - Proceedings of the 2013 ACM Multimedia Conference : 987-996. ScholarBank@NUS Repository. https://doi.org/10.1145/2502081.2502128|
|Abstract:||Recently visual saliency has attracted wide attention of researchers in the computer vision and multimedia field. However, most of the visual saliency-related research was conducted on still images for studying static saliency. In this paper, we give a comprehensive comparative study for the first time of dynamic saliency (video shots) and static saliency (key frames of the corresponding video shots), and two key observations are obtained: 1) video saliency is often different from, yet quite related with, image saliency, and 2) camera motions, such as tilting, panning or zooming, affect dynamic saliency significantly. Motivated by these observations, we propose a novel camera motion and image saliency aware model for dynamic saliency prediction. The extensive experiments on two static-vs-dynamic saliency datasets collected by us show that our proposed method outperforms the state-of-the-art methods for dynamic saliency prediction. Finally, we also introduce the application of dynamic saliency prediction for dynamic video captioning, assisting people with hearing impairments to better entertain videos with only off-screen voices, e.g., documentary films, news videos and sports videos. Copyright © 2013 ACM.|
|Source Title:||MM 2013 - Proceedings of the 2013 ACM Multimedia Conference|
|Appears in Collections:||Staff Publications|
Show full item record
Files in This Item:
There are no files associated with this item.
checked on Aug 20, 2018
checked on Jul 6, 2018
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.