Please use this identifier to cite or link to this item:
Title: Non-negative semi-supervised learning
Authors: Wang, C.
Yan, S. 
Zhang, L.
Zhang, H.-J.
Issue Date: 2009
Citation: Wang, C.,Yan, S.,Zhang, L.,Zhang, H.-J. (2009). Non-negative semi-supervised learning. Journal of Machine Learning Research 5 : 575-582. ScholarBank@NUS Repository.
Abstract: The contributions of this paper are three-fold. First, we present a general formulation for reaping the benefits from both non-negative data factorization and semi-supervised learning, and the solution naturally possesses the characteristics of sparsity, robustness to partial occlusions, and greater discriminating power via extra unlabeled data. Then, an efficient multiplicative updating procedure is proposed along with its theoretic justification of the algorithmic convergency. Finally, the tensorization of this general formulation for non-negative semi-supervised learning is also briefed for handling tensor data of arbitrary order. Extensive experiments compared with the state-of-the-art algorithms for non-negative data factorization and semi-supervised learning demonstrate the algorithmic properties in sparsity, classification power, and robustness to image occlusions. © 2009 by the authors.
Source Title: Journal of Machine Learning Research
ISSN: 15324435
Appears in Collections:Staff Publications

Show full item record
Files in This Item:
There are no files associated with this item.

Google ScholarTM


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.