Please use this identifier to cite or link to this item: https://doi.org/10.1145/3394171.3413896
DC FieldValue
dc.titleUniversity-1652: A Multi-view Multi-source Benchmark for Drone-based Geo-localization
dc.contributor.authorZheng, Zhedong
dc.contributor.authorWei, Yunchao
dc.contributor.authorYang, Yi
dc.date.accessioned2023-11-14T07:44:38Z
dc.date.available2023-11-14T07:44:38Z
dc.date.issued2020
dc.identifier.citationZheng, Zhedong, Wei, Yunchao, Yang, Yi (2020). University-1652: A Multi-view Multi-source Benchmark for Drone-based Geo-localization. 28th ACM International Conference on Multimedia (MM) : 1395-1403. ScholarBank@NUS Repository. https://doi.org/10.1145/3394171.3413896
dc.identifier.isbn9781450379885
dc.identifier.urihttps://scholarbank.nus.edu.sg/handle/10635/245947
dc.description.abstractWe consider the problem of cross-view geo-localization. The primary challenge is to learn the robust feature against large viewpoint changes. Existing benchmarks can help, but are limited in the number of viewpoints. Image pairs, containing two viewpoints, e.g., satellite and ground, are usually provided, which may compromise the feature learning. Besides phone cameras and satellites, in this paper, we argue that drones could serve as the third platform to deal with the geo-localization problem. In contrast to traditional ground-view images, drone-view images meet fewer obstacles, e.g., trees, and provide a comprehensive view when flying around the target place. To verify the effectiveness of the drone platform, we introduce a new multi-view multi-source benchmark for drone-based geo-localization, named University-1652. University-1652 contains data from three platforms, i.e., synthetic drones, satellites and ground cameras of 1,652 university buildings around the world. To our knowledge, University-1652 is the first drone-based geo-localization dataset and enables two new tasks, i.e., drone-view target localization and drone navigation. As the name implies, drone-view target localization intends to predict the location of the target place via drone-view images. On the other hand, given a satellite-view query image, drone navigation is to drive the drone to the area of interest in the query. We use this dataset to analyze a variety of off-the-shelf CNN features and propose a strong CNN baseline on this challenging dataset. The experiments show that University-1652 helps the model to learn viewpoint-invariant features and also has good generalization ability in real-world scenarios.
dc.publisherASSOC COMPUTING MACHINERY
dc.sourceElements
dc.subjectScience & Technology
dc.subjectTechnology
dc.subjectComputer Science, Artificial Intelligence
dc.subjectComputer Science, Information Systems
dc.subjectComputer Science, Interdisciplinary Applications
dc.subjectComputer Science, Software Engineering
dc.subjectImaging Science & Photographic Technology
dc.subjectComputer Science
dc.subjectDrone
dc.subjectGeo-localization
dc.subjectBenchmark
dc.subjectImage Retrieval
dc.typeConference Paper
dc.date.updated2023-11-11T04:42:01Z
dc.contributor.departmentDEPARTMENT OF COMPUTER SCIENCE
dc.contributor.departmentELECTRICAL AND COMPUTER ENGINEERING
dc.description.doi10.1145/3394171.3413896
dc.description.sourcetitle28th ACM International Conference on Multimedia (MM)
dc.description.page1395-1403
dc.published.statePublished
Appears in Collections:Staff Publications
Elements

Show simple item record
Files in This Item:
File Description SizeFormatAccess SettingsVersion 
ACMMM20.pdfAccepted version9.82 MBAdobe PDF

OPEN

Post-printView/Download

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.