Please use this identifier to cite or link to this item: https://doi.org/10.1109/tpami.2019.2919301
Title: DART: Distribution Aware Retinal Transform for Event-based Cameras
Authors: Ramesh, Bharath 
Yang, Hong 
Orchard, Garrick Michael 
Le Thi, Ngoc Anh
Zhang, Shihao
Xiang, Cheng 
Keywords: cs.CV
cs.CV
Issue Date: 27-May-2019
Publisher: Institute of Electrical and Electronics Engineers (IEEE)
Citation: Ramesh, Bharath, Yang, Hong, Orchard, Garrick Michael, Le Thi, Ngoc Anh, Zhang, Shihao, Xiang, Cheng (2019-05-27). DART: Distribution Aware Retinal Transform for Event-based Cameras. IEEE Transactions on Pattern Analysis and Machine Intelligence abs/1710.10800 : 1-1. ScholarBank@NUS Repository. https://doi.org/10.1109/tpami.2019.2919301
Abstract: We introduce a generic visual descriptor, termed as distribution aware retinal transform (DART), that encodes the structural context using log-polar grids for event cameras. The DART descriptor is applied to four different problems, namely object classification, tracking, detection and feature matching: (1) The DART features are directly employed as local descriptors in a bag-of-features classification framework and testing is carried out on four standard event-based object datasets (N-MNIST, MNIST-DVS, CIFAR10-DVS, NCaltech-101). (2) Extending the classification system, tracking is demonstrated using two key novelties: (i) For overcoming the low-sample problem for the one-shot learning of a binary classifier, statistical bootstrapping is leveraged with online learning; (ii) To achieve tracker robustness, the scale and rotation equivariance property of the DART descriptors is exploited for the one-shot learning. (3) To solve the long-term object tracking problem, an object detector is designed using the principle of cluster majority voting. The detection scheme is then combined with the tracker to result in a high intersection-over-union score with augmented ground truth annotations on the publicly available event camera dataset. (4) Finally, the event context encoded by DART greatly simplifies the feature correspondence problem, especially for spatio-temporal slices far apart in time, which has not been explicitly tackled in the event-based vision domain.
Source Title: IEEE Transactions on Pattern Analysis and Machine Intelligence
URI: https://scholarbank.nus.edu.sg/handle/10635/156799
ISSN: 01628828
19393539
DOI: 10.1109/tpami.2019.2919301
Appears in Collections:Staff Publications
Elements

Show full item record
Files in This Item:
File Description SizeFormatAccess SettingsVersion 
PAMI-2019.pdfPublished version1.88 MBAdobe PDF

CLOSED

Published

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.