Please use this identifier to cite or link to this item:
|Title:||Non-negative semi-supervised learning||Authors:||Wang, C.
|Issue Date:||2009||Citation:||Wang, C.,Yan, S.,Zhang, L.,Zhang, H.-J. (2009). Non-negative semi-supervised learning. Journal of Machine Learning Research 5 : 575-582. ScholarBank@NUS Repository.||Abstract:||The contributions of this paper are three-fold. First, we present a general formulation for reaping the benefits from both non-negative data factorization and semi-supervised learning, and the solution naturally possesses the characteristics of sparsity, robustness to partial occlusions, and greater discriminating power via extra unlabeled data. Then, an efficient multiplicative updating procedure is proposed along with its theoretic justification of the algorithmic convergency. Finally, the tensorization of this general formulation for non-negative semi-supervised learning is also briefed for handling tensor data of arbitrary order. Extensive experiments compared with the state-of-the-art algorithms for non-negative data factorization and semi-supervised learning demonstrate the algorithmic properties in sparsity, classification power, and robustness to image occlusions. © 2009 by the authors.||Source Title:||Journal of Machine Learning Research||URI:||http://scholarbank.nus.edu.sg/handle/10635/71162||ISSN:||15324435|
|Appears in Collections:||Staff Publications|
Show full item record
Files in This Item:
There are no files associated with this item.
checked on Apr 19, 2019
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.