Please use this identifier to cite or link to this item:
https://doi.org/10.1016/j.neucom.2007.02.009
DC Field | Value | |
---|---|---|
dc.title | Convex incremental extreme learning machine | |
dc.contributor.author | Huang, G.-B. | |
dc.contributor.author | Chen, L. | |
dc.date.accessioned | 2013-07-04T07:35:23Z | |
dc.date.available | 2013-07-04T07:35:23Z | |
dc.date.issued | 2007 | |
dc.identifier.citation | Huang, G.-B., Chen, L. (2007). Convex incremental extreme learning machine. Neurocomputing 70 (16-18) : 3056-3062. ScholarBank@NUS Repository. https://doi.org/10.1016/j.neucom.2007.02.009 | |
dc.identifier.issn | 09252312 | |
dc.identifier.uri | http://scholarbank.nus.edu.sg/handle/10635/39163 | |
dc.description.abstract | Unlike the conventional neural network theories and implementations, Huang et al. [Universal approximation using incremental constructive feedforward networks with random hidden nodes, IEEE Transactions on Neural Networks 17(4) (2006) 879-892] have recently proposed a new theory to show that single-hidden-layer feedforward networks (SLFNs) with randomly generated additive or radial basis function (RBF) hidden nodes (according to any continuous sampling distribution) can work as universal approximators and the resulting incremental extreme learning machine (I-ELM) outperforms many popular learning algorithms. I-ELM randomly generates the hidden nodes and analytically calculates the output weights of SLFNs, however, I-ELM does not recalculate the output weights of all the existing nodes when a new node is added. This paper shows that while retaining the same simplicity, the convergence rate of I-ELM can be further improved by recalculating the output weights of the existing nodes based on a convex optimization method when a new hidden node is randomly added. Furthermore, we show that given a type of piecewise continuous computational hidden nodes (possibly not neural alike nodes), if SLFNs fn (x) = underover(∑, i = 1, n) βi G (x, ai, bi) can work as universal approximators with adjustable hidden node parameters, from a function approximation point of view the hidden node parameters of such "generalized" SLFNs (including sigmoid networks, RBF networks, trigonometric networks, threshold networks, fuzzy inference systems, fully complex neural networks, high-order networks, ridge polynomial networks, wavelet networks, etc.) can actually be randomly generated according to any continuous sampling distribution. In theory, the parameters of these SLFNs can be analytically determined by ELM instead of being tuned. © 2007 Elsevier B.V. All rights reserved. | |
dc.description.uri | http://libproxy1.nus.edu.sg/login?url=http://dx.doi.org/10.1016/j.neucom.2007.02.009 | |
dc.source | Scopus | |
dc.subject | Convergence rate | |
dc.subject | Convex optimization | |
dc.subject | Generalized feedforward networks | |
dc.subject | Incremental extreme learning machine | |
dc.subject | Random hidden nodes | |
dc.type | Article | |
dc.contributor.department | COMPUTATIONAL SCIENCE | |
dc.description.doi | 10.1016/j.neucom.2007.02.009 | |
dc.description.sourcetitle | Neurocomputing | |
dc.description.volume | 70 | |
dc.description.issue | 16-18 | |
dc.description.page | 3056-3062 | |
dc.description.coden | NRCGE | |
dc.identifier.isiut | 000249908400047 | |
Appears in Collections: | Staff Publications |
Show simple item record
Files in This Item:
There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.