Please use this identifier to cite or link to this item:
|Title:||A novel Contrast Co-learning framework for generating high quality training data|
Noisy training data
Training data bias
|Source:||Zheng, Z.,Yan, J.,Yan, S.,Liu, N.,Chen, Z.,Zhang, M. (2010). A novel Contrast Co-learning framework for generating high quality training data. Proceedings - IEEE International Conference on Data Mining, ICDM : 649-658. ScholarBank@NUS Repository. https://doi.org/10.1109/ICDM.2010.23|
|Abstract:||The good performances of most classical learning algorithms are generally founded on high quality training data, which are clean and unbiased. The availability of such data is however becoming much harder than ever in many real world problems due to the difficulties in collecting large scale unbiased data and precisely labeling them for training. In this paper, we propose a general Contrast Co-learning (CCL) framework to refine the biased and noisy training data when an unbiased yet unlabeled data pool is available. CCL starts with multiple sets of probably biased and noisy training data and trains a set of classifiers individually. Then under the assumption that the confidently classified data samples may have higher probabilities to be correctly classified, CCL iteratively and automatically filtering out possible data noises as well as adding those confidently classified samples from the unlabeled data pool to correct the bias. Through this process, we can generate a cleaner and unbiased training dataset with theoretical guarantees. Extensive experiments on two public text datasets clearly show that CCL consistently improves the algorithmic classification performance on biased and noisy training data compared with several state-of-the-art classical algorithms. © 2010 IEEE.|
|Source Title:||Proceedings - IEEE International Conference on Data Mining, ICDM|
|Appears in Collections:||Staff Publications|
Show full item record
Files in This Item:
There are no files associated with this item.
checked on Dec 9, 2017
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.