Please use this identifier to cite or link to this item: https://doi.org/10.3389/frai.2024.1377337
Title: Expandable-RCNN: toward high-efficiency incremental few-shot object detection
Authors: Li, Y 
Tian, S
Zhu, H 
Jin, Y
Wang, K
Ma, J
Xiang, C 
Vadakkepat, P 
Keywords: few-shot learning
incremental learning
long-tailed recognition
object detection
zero-shot learning
Issue Date: 1-Jan-2024
Publisher: Frontiers Media SA
Citation: Li, Y, Tian, S, Zhu, H, Jin, Y, Wang, K, Ma, J, Xiang, C, Vadakkepat, P (2024-01-01). Expandable-RCNN: toward high-efficiency incremental few-shot object detection. Frontiers in Artificial Intelligence 7 : 1377337-. ScholarBank@NUS Repository. https://doi.org/10.3389/frai.2024.1377337
Abstract: This study aims at addressing the challenging incremental few-shot object detection (iFSOD) problem toward online adaptive detection. iFSOD targets to learn novel categories in a sequential manner, and eventually, the detection is performed on all learned categories. Moreover, only a few training samples are available for all sequential novel classes in these situations. In this study, we propose an efficient yet suitably simple framework, Expandable-RCNN, as a solution for the iFSOD problem, which allows online sequentially adding new classes with zero retraining of the base network. We achieve this by adapting the Faster R-CNN to the few-shot learning scenario with two elegant components to effectively address the overfitting and category bias. First, an IOU-aware weight imprinting strategy is proposed to directly determine the classifier weights for incremental novel classes and the background class, which is with zero training to avoid the notorious overfitting issue in few-shot learning. Second, since the above zero-retraining imprinting approach may lead to undesired category bias in the classifier, we develop a bias correction module for iFSOD, named the group soft-max layer (GSL), that efficiently calibrates the biased prediction of the imprinted classifier to organically improve classification performance for the few-shot classes, preventing catastrophic forgetting. Extensive experiments on MS-COCO show that our method can significantly outperform the state-of-the-art method ONCE by 5.9 points in commonly encountered few-shot classes.
Source Title: Frontiers in Artificial Intelligence
URI: https://scholarbank.nus.edu.sg/handle/10635/248842
ISSN: 2624-8212
DOI: 10.3389/frai.2024.1377337
Appears in Collections:Elements
Staff Publications

Show full item record
Files in This Item:
File Description SizeFormatAccess SettingsVersion 
Expandable-RCNN toward high-efficiency incremental few-shot object detection.pdf890.25 kBAdobe PDF

OPEN

PublishedView/Download

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.