Please use this identifier to cite or link to this item:
|Title:||Learning model order from labeled and unlabeled data for partially supervised classification, with application to word sense disambiguation|
|Keywords:||Partially supervised classification|
Word sense disambiguation
|Citation:||Niu, Z.-Y., Ji, D.-H., Tan, C.L. (2007). Learning model order from labeled and unlabeled data for partially supervised classification, with application to word sense disambiguation. Computer Speech and Language 21 (4) : 609-619. ScholarBank@NUS Repository. https://doi.org/10.1016/j.csl.2007.02.001|
|Abstract:||Previous partially supervised classification methods can partition unlabeled data into positive examples and negative examples for a given class by learning from positive labeled examples and unlabeled examples, but they cannot further group the negative examples into meaningful clusters even if there are many different classes in the negative examples. Here we proposed an automatic method to obtain a natural partitioning of mixed data (labeled data + unlabeled data) by maximizing a stability criterion defined on classification results from an extended label propagation algorithm over all the possible values of model order (or the number of classes) in mixed data. Our experimental results on benchmark corpora for word sense disambiguation task indicate that this model order identification algorithm with the extended label propagation algorithm as the base classifier outperforms SVM, a one-class partially supervised classification algorithm, and the model order identification algorithm with semi-supervised k-means clustering as the base classifier when labeled data is incomplete. © 2007 Elsevier Ltd. All rights reserved.|
|Source Title:||Computer Speech and Language|
|Appears in Collections:||Staff Publications|
Show full item record
Files in This Item:
There are no files associated with this item.
checked on Jul 10, 2018
WEB OF SCIENCETM
checked on Jun 4, 2018
checked on Jul 7, 2018
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.