Please use this identifier to cite or link to this item:
https://scholarbank.nus.edu.sg/handle/10635/155291
DC Field | Value | |
---|---|---|
dc.title | Loss Guided Activation for Action Recognition in Still Images | |
dc.contributor.author | Liu, Lu | |
dc.contributor.author | Tan, Robby T | |
dc.contributor.author | You, Shaodi | |
dc.date.accessioned | 2019-06-07T01:39:17Z | |
dc.date.available | 2019-06-07T01:39:17Z | |
dc.date.issued | 2018 | |
dc.identifier.citation | Liu, Lu, Tan, Robby T, You, Shaodi (2018). Loss Guided Activation for Action Recognition in Still Images. ScholarBank@NUS Repository. | |
dc.identifier.uri | https://scholarbank.nus.edu.sg/handle/10635/155291 | |
dc.description.abstract | One significant problem of deep-learning based human action recognition is that it can be easily misled by the presence of irrelevant objects or backgrounds. Existing methods commonly address this problem by employing bounding boxes on the target humans as part of the input, in both training and testing stages. This requirement of bounding boxes as part of the input is needed to enable the methods to ignore irrelevant contexts and extract only human features. However, we consider this solution is inefficient, since the bounding boxes might not be available. Hence, instead of using a person bounding box as an input, we introduce a human-mask loss to automatically guide the activations of the feature maps to the target human who is performing the action, and hence suppress the activations of misleading contexts. We propose a multi-task deep learning method that jointly predicts the human action class and human location heatmap. Extensive experiments demonstrate our approach is more robust compared to the baseline methods under the presence of irrelevant misleading contexts. Our method achieves 94.06\% and 40.65\% (in terms of mAP) on Stanford40 and MPII dataset respectively, which are 3.14\% and 12.6\% relative improvements over the best results reported in the literature, and thus set new state-of-the-art results. Additionally, unlike some existing methods, we eliminate the requirement of using a person bounding box as an input during testing. | |
dc.source | Elements | |
dc.subject | cs.CV | |
dc.subject | cs.CV | |
dc.type | Article | |
dc.date.updated | 2019-06-03T11:38:29Z | |
dc.contributor.department | YALE-NUS COLLEGE | |
dc.published.state | Unpublished | |
Appears in Collections: | Staff Publications Elements |
Show simple item record
Files in This Item:
File | Description | Size | Format | Access Settings | Version | |
---|---|---|---|---|---|---|
1812.04194v1.pdf | 5.71 MB | Adobe PDF | OPEN | Post-print | View/Download |
Google ScholarTM
Check
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.