Please use this identifier to cite or link to this item: https://doi.org/10.1145/2964284.2964308
DC FieldValue
dc.titlePlay and Rewind: Optimizing Binary Representations of Videos by Self-Supervised Temporal Hashing
dc.contributor.authorHanwang Zhang
dc.contributor.authorMeng Wang
dc.contributor.authorRichang Hong
dc.contributor.authorTat-Seng Chua
dc.date.accessioned2020-04-28T02:30:00Z
dc.date.available2020-04-28T02:30:00Z
dc.date.issued2016-10-15
dc.identifier.citationHanwang Zhang, Meng Wang, Richang Hong, Tat-Seng Chua (2016-10-15). Play and Rewind: Optimizing Binary Representations of Videos by Self-Supervised Temporal Hashing. ACM Multimedia Conference 2016 : 781-790. ScholarBank@NUS Repository. https://doi.org/10.1145/2964284.2964308
dc.identifier.isbn9781450336031
dc.identifier.urihttps://scholarbank.nus.edu.sg/handle/10635/167291
dc.description.abstractWe focus on hashing videos into short binary codes for efficient Content-based Video Retrieval (CBVR), which is a fundamental technique that supports access to the evergrowing abundance of videos on the Web. Existing video hash functions are built on three isolated stages: frame pooling, relaxed learning, and binarization, which have not adequately explored the temporal order of video frames in a joint binary optimization model, resulting in severe information loss. In this paper, we propose a novel unsupervised video hashing framework called Self-Supervised Temporal Hashing (SSTH) that is able to capture the temporal nature of videos in an end-to-end learning-to-hash fashion. Specifically, the hash function of SSTH is an encoder RNN equipped with the proposed Binary LSTM (BLSTM) that generates binary codes for videos. The hash function is learned in a self-supervised fashion, where a decoder RNN is proposed to reconstruct the original video frames in both forward and reverse orders. For binary code optimization, we develop a backpropagation rule that tackles the non-differentiability of BLSTM. This rule allows efficient deep network training without suffering from the binarization loss. Through extensive CBVR experiments on two real-world consumer video datasets of Youtube and Flickr, we show that SSTH consistently outperforms state-of-theart video hashing methods, e.g., in terms of mAP@20, SSTH using only 128 bits can still outperform others using 256 bits by at least 9% to 15% on both datasets. © 2016 ACM.
dc.publisherAssociation for Computing Machinery, Inc
dc.subjectBinary LSTM
dc.subjectSequence learning
dc.subjectTemporal hashing
dc.subjectVideo retrieval
dc.typeConference Paper
dc.contributor.departmentDEPARTMENT OF COMPUTER SCIENCE
dc.description.doi10.1145/2964284.2964308
dc.description.sourcetitleACM Multimedia Conference 2016
dc.description.page781-790
dc.grant.idR-252-300-002-490
dc.grant.fundingagencyInfocomm Media Development Authority
dc.grant.fundingagencyNational Research Foundation
Appears in Collections:Staff Publications
Elements

Show simple item record
Files in This Item:
File Description SizeFormatAccess SettingsVersion 
Play and Rewind.pdf1.64 MBAdobe PDF

OPEN

NoneView/Download

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.