Please use this identifier to cite or link to this item: https://doi.org/10.1109/TIP.2002.804265
Title: Synergizing spatial and temporal texture
Authors: Peh, C.-H. 
Cheong, L.-F. 
Keywords: Normal flow
Spatiotemporal texture
Temporal texture
Issue Date: Oct-2002
Source: Peh, C.-H.,Cheong, L.-F. (2002-10). Synergizing spatial and temporal texture. IEEE Transactions on Image Processing 11 (10) : 1179-1191. ScholarBank@NUS Repository. https://doi.org/10.1109/TIP.2002.804265
Abstract: Temporal texture accounts for a large proportion of motion commonly experienced in the visual world. Current temporal texture techniques extract primarily motion-based features for recognition. We propose in this paper a representation where both the spatial and the temporal aspects of texture are coupled together. Such a representation has the advantages of improving efficiency as well as retaining both spatial and temporal semantics. Flow measurements form the basis of our representation. The magnitudes and directions of the normal flow are mapped as spatiotemporal textures. These textures are then aggregated over time and are subsequently analyzed by classical texture analysis tools. Such aggregation traces the history of a motion which can be useful in the understanding of motion types. By providing a spatiotemporal analysis, our approach gains several advantages over previous implementations. The strength of our approach was demonstrated in a series of experiments, including classification and comparisons with other algorithms.
Source Title: IEEE Transactions on Image Processing
URI: http://scholarbank.nus.edu.sg/handle/10635/57581
ISSN: 10577149
DOI: 10.1109/TIP.2002.804265
Appears in Collections:Staff Publications

Show full item record
Files in This Item:
There are no files associated with this item.

SCOPUSTM   
Citations

47
checked on Dec 5, 2017

WEB OF SCIENCETM
Citations

37
checked on Nov 7, 2017

Page view(s)

34
checked on Dec 9, 2017

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.