Please use this identifier to cite or link to this item: https://doi.org/10.1109/ICDAR.2011.235
DC FieldValue
dc.titleA new Fourier-moments based video word and character extraction method for recognition
dc.contributor.authorRajendran, D.
dc.contributor.authorShivakumara, P.
dc.contributor.authorSu, B.
dc.contributor.authorLu, S.
dc.contributor.authorTan, C.L.
dc.date.accessioned2013-07-04T08:36:59Z
dc.date.available2013-07-04T08:36:59Z
dc.date.issued2011
dc.identifier.citationRajendran, D., Shivakumara, P., Su, B., Lu, S., Tan, C.L. (2011). A new Fourier-moments based video word and character extraction method for recognition. Proceedings of the International Conference on Document Analysis and Recognition, ICDAR : 1165-1169. ScholarBank@NUS Repository. https://doi.org/10.1109/ICDAR.2011.235
dc.identifier.isbn9780769545202
dc.identifier.issn15205363
dc.identifier.urihttp://scholarbank.nus.edu.sg/handle/10635/41836
dc.description.abstractThis paper presents a new method based on Fourier and moments features to extract words and characters from a video text line in any direction for recognition. Unlike existing methods which output the entire text line to the ensuing recognition algorithm, the proposed method obtains each extracted character from the text line as input to the recognition algorithm because the background of a single character is relatively simple compared to the text line and words. Max-Min clustering criterion is introduced to obtain text cluster from the extracted Fourier and moments feature set. Union of the text cluster with Canny operation of the input video text line is proposed to obtain missing text candidates. Then a run length criterion is used for extraction of words. From the words, we propose a new idea for extracting characters from the text candidates of each word image based on the fact that the text height difference at the character boundary column is smaller than that at other columns of the word image. We evaluate the method on a large dataset at three levels namely text line, words and characters in terms of recall, precision and f-measure. In addition to this, we show that the recognition result for the extracted character is better than words and lines. Our experimental set up involves 3527 characters including Chinese. The dataset is selected from TRECVID database of 2005 and 2006. © 2011 IEEE.
dc.description.urihttp://libproxy1.nus.edu.sg/login?url=http://dx.doi.org/10.1109/ICDAR.2011.235
dc.sourceScopus
dc.subjectFourier-Moments
dc.subjectRun length
dc.subjectText height difference
dc.subjectVideo character extraction
dc.subjectVideo character recognition
dc.subjectVideo word segmentation
dc.typeConference Paper
dc.contributor.departmentCOMPUTER SCIENCE
dc.description.doi10.1109/ICDAR.2011.235
dc.description.sourcetitleProceedings of the International Conference on Document Analysis and Recognition, ICDAR
dc.description.page1165-1169
dc.identifier.isiut000343450700229
Appears in Collections:Staff Publications

Show simple item record
Files in This Item:
There are no files associated with this item.

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.