Please use this identifier to cite or link to this item:
https://doi.org/10.1109/ICDAR.2011.207
Title: | A gradient vector flow-based method for video character segmentation | Authors: | Phan, T.Q. Shivakumara, P. Su, B. Tan, C.L. |
Keywords: | Curved segmentation path Gradient vector flow Minimum cost path finding Video character segmentation |
Issue Date: | 2011 | Citation: | Phan, T.Q., Shivakumara, P., Su, B., Tan, C.L. (2011). A gradient vector flow-based method for video character segmentation. Proceedings of the International Conference on Document Analysis and Recognition, ICDAR : 1024-1028. ScholarBank@NUS Repository. https://doi.org/10.1109/ICDAR.2011.207 | Abstract: | In this paper, we propose a method based on gradient vector flow for video character segmentation. By formulating character segmentation as a minimum cost path finding problem, the proposed method allows curved segmentation paths and thus it is able to segment overlapping characters and touching characters due to low contrast and complex background. Gradient vector flow is used in a new way to identify candidate cut pixels. A two-pass path finding algorithm is then applied where the forward direction helps to locate potential cuts and the backward direction serves to remove the false cuts, i.e. those that go through the characters, while retaining the true cuts. Experimental results show that the proposed method outperforms an existing method on multi-oriented English and Chinese video text lines. The proposed method also helps to improve binarization results, which lead to a better character recognition rate. © 2011 IEEE. | Source Title: | Proceedings of the International Conference on Document Analysis and Recognition, ICDAR | URI: | http://scholarbank.nus.edu.sg/handle/10635/41837 | ISBN: | 9780769545202 | ISSN: | 15205363 | DOI: | 10.1109/ICDAR.2011.207 |
Appears in Collections: | Staff Publications |
Show full item record
Files in This Item:
There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.