Please use this identifier to cite or link to this item: https://doi.org/10.1016/j.ipm.2020.102277
DC FieldValue
dc.titleMGAT: Multimodal Graph Attention Network for Recommendation
dc.contributor.authorZhulin Tao
dc.contributor.authorYinwei Wei
dc.contributor.authorXiang Wang
dc.contributor.authorXiangnan He
dc.contributor.authorXianglin Huang
dc.contributor.authorTat-Seng Chua
dc.date.accessioned2020-10-21T04:02:50Z
dc.date.available2020-10-21T04:02:50Z
dc.date.issued2020-05-12
dc.identifier.citationZhulin Tao, Yinwei Wei, Xiang Wang, Xiangnan He, Xianglin Huang, Tat-Seng Chua (2020-05-12). MGAT: Multimodal Graph Attention Network for Recommendation. Information Processing and Management 57 (5). ScholarBank@NUS Repository. https://doi.org/10.1016/j.ipm.2020.102277
dc.identifier.issn03064573
dc.identifier.urihttps://scholarbank.nus.edu.sg/handle/10635/178636
dc.description.abstractGraph neural networks (GNNs) have shown great potential for personalized recommendation. At the core is to reorganize interaction data as a user-item bipartite graph and exploit high-order connectivity among user and item nodes to enrich their representations. While achieving great success, most existing works consider interaction graph based only on ID information, foregoing item contents from multiple modalities (e.g., visual, acoustic, and textual features of micro-video items). Distinguishing personal interests on different modalities at a granular level was not explored until recently proposed MMGCN (Wei et al., 2019). However, it simply employs GNNs on parallel interaction graphs and treats information propagated from all neighbors equally, failing to capture user preference adaptively. Hence, the obtained representations might preserve redundant, even noisy information, leading to non-robustness and suboptimal performance. In this work, we aim to investigate how to adopt GNNs on multimodal interaction graphs, to adaptively capture user preference on different modalities and offer in-depth analysis on why an item is suitable to a user. Towards this end, we propose a new Multimodal Graph Attention Network, short for MGAT, which disentangles personal interests at the granularity of modality. In particular, built upon multimodal interaction graphs, MGAT conducts information propagation within individual graphs, while leveraging the gated attention mechanism to identify varying importance scores of different modalities to user preference. As such, it is able to capture more complex interaction patterns hidden in user behaviors and provide a more accurate recommendation. Empirical results on two micro-video recommendation datasets, Tiktok and MovieLens, show that MGAT exhibits substantial improvements over the state-of-the-art baselines like NGCF (Wang, He, et al., 2019) and MMGCN (Wei et al., 2019). Further analysis on a case study illustrates how MGAT generates attentive information flow over multimodal interaction graphs. © 2020 Elsevier Ltd
dc.publisherElsevier Ltd
dc.subjectPersonalized recommendation
dc.subjectGraph Gate mechanism
dc.subjectAttention mechanism
dc.subjectMicro-videos
dc.typeArticle
dc.contributor.departmentDEPARTMENT OF COMPUTER SCIENCE
dc.description.doi10.1016/j.ipm.2020.102277
dc.description.sourcetitleInformation Processing and Management
dc.description.volume57
dc.description.issue5
dc.description.codenIPMAD
dc.grant.idR-252-300-002-490
dc.grant.fundingagencyInfocomm Media Development Authority
dc.grant.fundingagencyNational Research Foundation
Appears in Collections:Elements
Staff Publications

Show simple item record
Files in This Item:
File Description SizeFormatAccess SettingsVersion 
MGAT-Multimodal Graph Attention Network for Recommendation.pdf1.69 MBAdobe PDF

OPEN

NoneView/Download

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.