Please use this identifier to cite or link to this item: https://scholarbank.nus.edu.sg/handle/10635/130419
DC FieldValue
dc.titleSplitting-Merging Model of Chinese Word Tokenization and Segmentation
dc.contributor.authorYao, Y.
dc.contributor.authorLua, K.T.
dc.date.accessioned2016-11-16T11:05:46Z
dc.date.available2016-11-16T11:05:46Z
dc.date.issued1998
dc.identifier.citationYao, Y., Lua, K.T. (1998). Splitting-Merging Model of Chinese Word Tokenization and Segmentation. Natural Language Engineering 4 (4) : 309-324. ScholarBank@NUS Repository.
dc.identifier.issn13513249
dc.identifier.urihttp://scholarbank.nus.edu.sg/handle/10635/130419
dc.description.abstractWord tokenization & segmentation in natural language processing of languages like Chinese, which have no blank space for word delimitation, are considered. Three major problems are faced: (1) tokenizing direction & efficiency, (2) insufficient tokenization dictionary & nonwords, & (3) ambiguity of tokenization & segmentation. Most existing tokenization & segmentation methods have not dealt with the above problems together. A novel dictionary-based method called the splitting-merging model for Chinese word tokenization & segmentation is presented. It uses the mutual information of Chinese characters to find the boundaries & the non-boundaries of Chinese words, & finally leads to word segmentation by resolving ambiguities & detecting new words.
dc.sourceScopus
dc.typeArticle
dc.contributor.departmentCOMPUTER SCIENCE
dc.description.sourcetitleNatural Language Engineering
dc.description.volume4
dc.description.issue4
dc.description.page309-324
dc.description.codenNLENF
dc.identifier.isiutNOT_IN_WOS
Appears in Collections:Staff Publications

Show simple item record
Files in This Item:
There are no files associated with this item.

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.