Please use this identifier to cite or link to this item: https://doi.org/10.1109/TCSVT.2016.2602832
DC FieldValue
dc.titleEncoded Semantic Tree for Automatic User Profiling Applied to Personalized Video Summarization
dc.contributor.authorYin, Yifang
dc.contributor.authorThapliya, Roshan
dc.contributor.authorZimmermann, Roger
dc.date.accessioned2021-09-20T07:48:34Z
dc.date.available2021-09-20T07:48:34Z
dc.date.issued2018-01-01
dc.identifier.citationYin, Yifang, Thapliya, Roshan, Zimmermann, Roger (2018-01-01). Encoded Semantic Tree for Automatic User Profiling Applied to Personalized Video Summarization. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY 28 (1) : 181-192. ScholarBank@NUS Repository. https://doi.org/10.1109/TCSVT.2016.2602832
dc.identifier.issn10518215
dc.identifier.issn15582205
dc.identifier.urihttps://scholarbank.nus.edu.sg/handle/10635/200728
dc.description.abstractWe propose an innovative method of automatic video summary generation with personal adaptations. User interests are mined from their personal image collections. To reduce the semantic gap, we propose to extract visual representations based on a novel semantic tree (SeTree). A SeTree is a hierarchy that captures the conceptual relationships between the visual scenes in a codebook. This idea builds upon the observation that such semantic connections among the elements have been overlooked in the previous work. To construct the SeTree, we adopt a normalized graph cut clustering algorithm by conjunctively exploiting visual features, textual information, and social user-image connections. Using this technique, we obtain an 8.1% improvement of normalized discounted cumulative gain in personalized video segments ranking compared with existing methods. Furthermore, to promote the interesting parts of a video, we extract a space-time saliency map and estimate the attractiveness of segments by kernel fitting and matching. A linear function is utilized to combine the two factors, based on which the playback rate of a video is adapted to generate the summary. We play the less important segments in a fast-forward mode to keep users updated with the context. Subjective experiments were conducted which showed that our proposed video summarization approach outperformed the state-of-the-art techniques by 6.2%.
dc.language.isoen
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
dc.sourceElements
dc.subjectScience & Technology
dc.subjectTechnology
dc.subjectEngineering, Electrical & Electronic
dc.subjectEngineering
dc.subjectSemantic modeling
dc.subjectuser profiling
dc.subjectvideo summarization
dc.subjectvisual attention
dc.subjectFRAMEWORK
dc.typeArticle
dc.date.updated2021-09-19T15:33:10Z
dc.contributor.departmentCHEMICAL & BIOMOLECULAR ENGINEERING
dc.description.doi10.1109/TCSVT.2016.2602832
dc.description.sourcetitleIEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY
dc.description.volume28
dc.description.issue1
dc.description.page181-192
dc.published.statePublished
Appears in Collections:Staff Publications
Elements

Show simple item record
Files in This Item:
File Description SizeFormatAccess SettingsVersion 
main.pdf2.1 MBAdobe PDF

CLOSED

None

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.