Please use this identifier to cite or link to this item: https://doi.org/10.1145/3343031.3351034
DC FieldValue
dc.titleMMGCN: Multimodal Graph Convolution Network for Personalized Recommendation of Micro-video
dc.contributor.authorYinwei Wei
dc.contributor.authorXiang Wang
dc.contributor.authorLiqiang Nie
dc.contributor.authorXiangnan He
dc.contributor.authorRichang Hong
dc.contributor.authorTat-Seng Chua
dc.date.accessioned2020-05-06T04:14:44Z
dc.date.available2020-05-06T04:14:44Z
dc.date.issued2019-10-21
dc.identifier.citationYinwei Wei, Xiang Wang, Liqiang Nie, Xiangnan He, Richang Hong, Tat-Seng Chua (2019-10-21). MMGCN: Multimodal Graph Convolution Network for Personalized Recommendation of Micro-video. ACM MM 2019 : 1437-1445. ScholarBank@NUS Repository. https://doi.org/10.1145/3343031.3351034
dc.identifier.isbn9781450368896
dc.identifier.urihttps://scholarbank.nus.edu.sg/handle/10635/167770
dc.description.abstractPersonalized recommendation plays a central role in many online content sharing platforms. To provide quality micro-video recommendation service, it is of crucial importance to consider the interactions between users and items (i.e., micro-videos) as well as the item contents from various modalities (e.g., visual, acoustic, and textual). Existing works on multimedia recommendation largely exploit multi-modal contents to enrich item representations, while less effort is made to leverage information interchange between users and items to enhance user representations and further capture user's fine-grained preferences on different modalities. In this paper, we propose to exploit user-item interactions to guide the representation learning in each modality, and further personalized micro-video recommendation. We design a Multimodal Graph Convolution Network (MMGCN) framework built upon the message-passing idea of graph neural networks, which can yield modal-specific representations of users and micro-videos to better capture user preferences. Specifically, we construct a user-item bipartite graph in each modality, and enrich the representation of each node with the topological structure and features of its neighbors. Through extensive experiments on three publicly available datasets, Tiktok, Kwai, and MovieLens, we demonstrate that our proposed model is able to significantly outperform state-of-the-art multi-modal recommendation methods. ? 2019 Association for Computing Machinery.
dc.subjectGraph Convolution Network
dc.subjectMicro-video Understanding
dc.subjectMulti-modal Recommendation
dc.typeConference Paper
dc.contributor.departmentDEPT OF COMPUTER SCIENCE
dc.description.doi10.1145/3343031.3351034
dc.description.sourcetitleACM MM 2019
dc.description.page1437-1445
dc.grant.idR-252-300-002-490
dc.grant.fundingagencyInfocomm Media Development Authority
dc.grant.fundingagencyNational Research Foundation
Appears in Collections:Staff Publications
Elements

Show simple item record
Files in This Item:
File Description SizeFormatAccess SettingsVersion 
3343031.3351034.pdf3.29 MBAdobe PDF

OPEN

NoneView/Download

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.