Please use this identifier to cite or link to this item: https://doi.org/10.1109/TNN.2009.2032543
Title: Feature selection for MLP neural network: The use of random permutation of probabilistic outputs
Authors: Yang, J.-B.
Shen, K.-Q. 
Ong, C.-J. 
Li, X.-P. 
Keywords: Feature ranking
Feature selection
Multilayer perceptrons (MLPs)
Probabilistic outputs
Random permutation
Issue Date: Dec-2009
Source: Yang, J.-B., Shen, K.-Q., Ong, C.-J., Li, X.-P. (2009-12). Feature selection for MLP neural network: The use of random permutation of probabilistic outputs. IEEE Transactions on Neural Networks 20 (12) : 1911-1922. ScholarBank@NUS Repository. https://doi.org/10.1109/TNN.2009.2032543
Abstract: This paper presents a new wrapper-based feature selection method for multilayer perceptron (MLP) neural networks. It uses a feature ranking criterion to measure the importance of a feature by computing the aggregate difference, over the feature space, of the probabilistic outputs of the MLP with and without the feature. Thus, a score of importance with respect to every feature can be provided using this criterion. Based on the numerical experiments on several artificial and real-world data sets, the proposed method performs, in general, better than several selected feature selection methods for MLP, particularly when the data set is sparse or has many redundant features. In addition, as a wrapper-based approach, the computational cost for the proposed method is modest. © 2009 IEEE.
Source Title: IEEE Transactions on Neural Networks
URI: http://scholarbank.nus.edu.sg/handle/10635/51413
ISSN: 10459227
DOI: 10.1109/TNN.2009.2032543
Appears in Collections:Staff Publications

Show full item record
Files in This Item:
There are no files associated with this item.

SCOPUSTM   
Citations

27
checked on Dec 6, 2017

WEB OF SCIENCETM
Citations

17
checked on Nov 22, 2017

Page view(s)

18
checked on Dec 10, 2017

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.