Please use this identifier to cite or link to this item: https://doi.org/10.1007/s11263-023-01805-x
DC FieldValue
dc.titleMulti-view Consistent Generative Adversarial Networks for Compositional 3D-Aware Image Synthesis
dc.contributor.authorZhang, Xuanmeng
dc.contributor.authorZheng, Zhedong
dc.contributor.authorGao, Daiheng
dc.contributor.authorZhang, Bang
dc.contributor.authorYang, Yi
dc.contributor.authorChua, Tat-Seng
dc.date.accessioned2023-11-09T04:18:14Z
dc.date.available2023-11-09T04:18:14Z
dc.date.issued2023-08
dc.identifier.citationZhang, Xuanmeng, Zheng, Zhedong, Gao, Daiheng, Zhang, Bang, Yang, Yi, Chua, Tat-Seng (2023-08). Multi-view Consistent Generative Adversarial Networks for Compositional 3D-Aware Image Synthesis. INTERNATIONAL JOURNAL OF COMPUTER VISION 131 (8) : 2219-2242. ScholarBank@NUS Repository. https://doi.org/10.1007/s11263-023-01805-x
dc.identifier.issn0920-5691
dc.identifier.issn1573-1405
dc.identifier.urihttps://scholarbank.nus.edu.sg/handle/10635/245843
dc.description.abstractThis paper studies compositional 3D-aware image synthesis for both single-object and multi-object scenes. We observe that two challenges remain in this field: existing approaches (1) lack geometry constraints and thus compromise the multi-view consistency of the single object, and (2) can not scale to multi-object scenes with complex backgrounds. To address these challenges coherently, we propose multi-view consistent generative adversarial networks (MVCGAN) for compositional 3D-aware image synthesis. First, we build the geometry constraints on the single object by leveraging the underlying 3D information. Specifically, we enforce the photometric consistency between pairs of views, encouraging the model to learn the inherent 3D shape. Second, we adapt MVCGAN to multi-object scenarios. In particular, we formulate the multi-object scene generation as a “decompose and compose” process. During training, we adopt the top-down strategy to decompose training images into objects and backgrounds. When rendering, we deploy a reverse bottom-up manner by composing the generated objects and background into the holistic scene. Extensive experiments on both single-object and multi-object datasets show that the proposed method achieves competitive performance for 3D-aware image synthesis.
dc.language.isoen
dc.publisherSPRINGER
dc.sourceElements
dc.subjectScience & Technology
dc.subjectTechnology
dc.subjectComputer Science, Artificial Intelligence
dc.subjectComputer Science
dc.subjectGenerative adversarial networks
dc.subjectNeural radiance fields
dc.subject3D-aware image synthesis
dc.typeArticle
dc.date.updated2023-11-09T03:37:50Z
dc.contributor.departmentDEPARTMENT OF COMPUTER SCIENCE
dc.description.doi10.1007/s11263-023-01805-x
dc.description.sourcetitleINTERNATIONAL JOURNAL OF COMPUTER VISION
dc.description.volume131
dc.description.issue8
dc.description.page2219-2242
dc.published.statePublished
Appears in Collections:Staff Publications
Elements

Show simple item record
Files in This Item:
File Description SizeFormatAccess SettingsVersion 
Zhang_MVCGAN.pdfPublished version34.32 MBAdobe PDF

OPEN

PublishedView/Download

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.