Please use this identifier to cite or link to this item: https://doi.org/10.1109/TMM.2020.3014488
DC FieldValue
dc.titleVehicleNet: Learning Robust Visual Representation for Vehicle Re-Identification
dc.contributor.authorZheng, Zhedong
dc.contributor.authorRuan, Tao
dc.contributor.authorWei, Yunchao
dc.contributor.authorYang, Yi
dc.contributor.authorMei, Tao
dc.date.accessioned2023-11-14T03:00:43Z
dc.date.available2023-11-14T03:00:43Z
dc.date.issued2021
dc.identifier.citationZheng, Zhedong, Ruan, Tao, Wei, Yunchao, Yang, Yi, Mei, Tao (2021). VehicleNet: Learning Robust Visual Representation for Vehicle Re-Identification. IEEE TRANSACTIONS ON MULTIMEDIA 23 : 2683-2693. ScholarBank@NUS Repository. https://doi.org/10.1109/TMM.2020.3014488
dc.identifier.issn1520-9210
dc.identifier.issn1941-0077
dc.identifier.urihttps://scholarbank.nus.edu.sg/handle/10635/245916
dc.description.abstractOne fundamental challenge of vehicle re-identification (re-id) is to learn robust and discriminative visual representation, given the significant intra-class vehicle variations across different camera views. As the existing vehicle datasets are limited in terms of training images and viewpoints, we propose to build a unique large-scale vehicle dataset (called VehicleNet) by harnessing four public vehicle datasets, and design a simple yet effective two-stage progressive approach to learning more robust visual representation from VehicleNet. The first stage of our approach is to learn the generic representation for all domains (i.e., source vehicle datasets) by training with the conventional classification loss. This stage relaxes the full alignment between the training and testing domains, as it is agnostic to the target vehicle domain. The second stage is to fine-tune the trained model purely based on the target vehicle set, by minimizing the distribution discrepancy between our VehicleNet and any target domain. We discuss our proposed multi-source dataset VehicleNet and evaluate the effectiveness of the two-stage progressive representation learning through extensive experiments. We achieve the state-of-art accuracy of 86.07% mAP on the private test set of AICity Challenge, and competitive results on two other public vehicle re-id datasets, i.e., VeRi-776 and VehicleID. We hope this new VehicleNet dataset and the learned robust representations can pave the way for vehicle re-id in the real-world environments.
dc.language.isoen
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
dc.sourceElements
dc.subjectScience & Technology
dc.subjectTechnology
dc.subjectComputer Science, Information Systems
dc.subjectComputer Science, Software Engineering
dc.subjectTelecommunications
dc.subjectComputer Science
dc.subjectTraining
dc.subjectRobustness
dc.subjectAdaptation models
dc.subjectData models
dc.subjectAutomobiles
dc.subjectCameras
dc.subjectFeature extraction
dc.subjectVehicle re-identification
dc.subjectimage representation
dc.subjectconvolutional neural networks
dc.subjectPERSON REIDENTIFICATION
dc.typeArticle
dc.date.updated2023-11-10T11:43:23Z
dc.contributor.departmentDEPARTMENT OF COMPUTER SCIENCE
dc.contributor.departmentELECTRICAL AND COMPUTER ENGINEERING
dc.description.doi10.1109/TMM.2020.3014488
dc.description.sourcetitleIEEE TRANSACTIONS ON MULTIMEDIA
dc.description.volume23
dc.description.page2683-2693
dc.published.statePublished
Appears in Collections:Elements
Staff Publications

Show simple item record
Files in This Item:
File Description SizeFormatAccess SettingsVersion 
TMM20_compressed.pdfAccepted version684.16 kBAdobe PDF

CLOSED

None

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.