Please use this identifier to cite or link to this item: https://doi.org/10.1109/TPAMI.2011.232
DC FieldValue
dc.titleToward holistic scene understanding: Feedback enabled cascaded classification models
dc.contributor.authorLi C.
dc.contributor.authorKowdle A.
dc.contributor.authorSaxena A.
dc.contributor.authorChen T.
dc.date.accessioned2018-08-21T04:58:28Z
dc.date.available2018-08-21T04:58:28Z
dc.date.issued2012
dc.identifier.citationLi C., Kowdle A., Saxena A., Chen T. (2012). Toward holistic scene understanding: Feedback enabled cascaded classification models. IEEE Transactions on Pattern Analysis and Machine Intelligence 34 (7) : 1394-1408. ScholarBank@NUS Repository. https://doi.org/10.1109/TPAMI.2011.232
dc.identifier.issn01628828
dc.identifier.urihttp://scholarbank.nus.edu.sg/handle/10635/146139
dc.description.abstractScene understanding includes many related subtasks, such as scene categorization, depth estimation, object detection, etc. Each of these subtasks is often notoriously hard, and state-of-the-art classifiers already exist for many of them. These classifiers operate on the same raw image and provide correlated outputs. It is desirable to have an algorithm that can capture such correlation without requiring any changes to the inner workings of any classifier. We propose Feedback Enabled Cascaded Classification Models (FE-CCM), that jointly optimizes all the subtasks while requiring only a black box interface to the original classifier for each subtask. We use a two-layer cascade of classifiers, which are repeated instantiations of the original ones, with the output of the first layer fed into the second layer as input. Our training method involves a feedback step that allows later classifiers to provide earlier classifiers information about which error modes to focus on. We show that our method significantly improves performance in all the subtasks in the domain of scene understanding, where we consider depth estimation, scene categorization, event categorization, object detection, geometric labeling, and saliency detection. Our method also improves performance in two robotic applications: an object-grasping robot and an object-finding robot.
dc.sourceScopus
dc.subjectclassification
dc.subjectmachine learning
dc.subjectrobotics
dc.subjectScene understanding
dc.typeArticle
dc.contributor.departmentOFFICE OF THE PROVOST
dc.contributor.departmentDEPARTMENT OF COMPUTER SCIENCE
dc.description.doi10.1109/TPAMI.2011.232
dc.description.sourcetitleIEEE Transactions on Pattern Analysis and Machine Intelligence
dc.description.volume34
dc.description.issue7
dc.description.page1394-1408
dc.description.codenITPID
dc.published.statepublished
Appears in Collections:Staff Publications

Show simple item record
Files in This Item:
There are no files associated with this item.

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.