Please use this identifier to cite or link to this item:
https://doi.org/10.1109/TMM.2018.2879749
Title: | Discovering Latent Discriminative Patterns for Multi-Mode Event Representation | Authors: | Wenlong Xie Hongxun Yao Xiaoshuai Sun Tingting Han Sicheng Zhao Tat-Seng Chua |
Keywords: | Segment-level Visual topics Event representation Latent patterns Event epitomes |
Issue Date: | 5-Nov-2018 | Publisher: | Institute of Electrical and Electronics Engineers Inc. | Citation: | Wenlong Xie, Hongxun Yao, Xiaoshuai Sun, Tingting Han, Sicheng Zhao, Tat-Seng Chua (2018-11-05). Discovering Latent Discriminative Patterns for Multi-Mode Event Representation. IEEE Transactions on Multimedia 21 (6) : 1425 - 1436. ScholarBank@NUS Repository. https://doi.org/10.1109/TMM.2018.2879749 | Abstract: | Representation of videos is essential since it conveys an understanding of video content and enables many higher level tasks to be tackled efficiently. However, it is challenging to propose a rational representation for complex event videos, as most video information is either noisy or redundant. In this paper, we propose a compact event representation method that can concisely describe the inner modes of events. We deem that an optimal event representation scheme should reflect the long-term and high-level visual semantics (visual topics) of events, so different from previous frame-level video semantics representation methods and concept-based video representation methods, we investigate the problem from the perspective of segment-level video representations. We then present three appealing properties of segment-level visual semantics. Based on the observation, we propose different algorithms that rely on a novel deep-visual-word-based video encoding method to discover latent discriminative patterns of events. Finally, our multi-mode event representation is obtained by concatenating the discovered patterns as inner modes. We adopt our event representation for representative event parts mining, which can highlight the visual topics of events and remarkably prune the raw videos. We validate our event representation method based on complex event detection task. Experimental results on two standard benchmarking datasets, MED11 and CCV Dataset, show that the proposed method can significantly outperform the state-of-the-art approaches. © 1999-2012 IEEE. | Source Title: | IEEE Transactions on Multimedia | URI: | https://scholarbank.nus.edu.sg/handle/10635/168409 | ISSN: | 15209210 | DOI: | 10.1109/TMM.2018.2879749 |
Appears in Collections: | Staff Publications Elements |
Show full item record
Files in This Item:
File | Description | Size | Format | Access Settings | Version | |
---|---|---|---|---|---|---|
Discovering Latent Discriminative Patterns for Multi-Mode Event Representation.pdf | 3.09 MB | Adobe PDF | OPEN | Post-print | View/Download |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.