Please use this identifier to cite or link to this item: https://doi.org/10.1109/TNN.2005.852236
Title: Convergence analysis of a deterministic discrete time system of Oja's PCA learning algorithm
Authors: Yi, Z.
Ye, M.
Lv, J.C.
Tan, K.K. 
Keywords: Eigenvalue
Eigenvector
Neural network
Oja's learning algorithm
Principal component analysis (PCA)
Issue Date: Nov-2005
Source: Yi, Z., Ye, M., Lv, J.C., Tan, K.K. (2005-11). Convergence analysis of a deterministic discrete time system of Oja's PCA learning algorithm. IEEE Transactions on Neural Networks 16 (6) : 1318-1328. ScholarBank@NUS Repository. https://doi.org/10.1109/TNN.2005.852236
Abstract: The convergence of Oja's principal component analysis (PCA) learning algorithms is a difficult topic for direct study and analysis. Traditionally, the convergence of these algorithms is indirectly analyzed via certain deterministic continuous time (DCT) systems. Such a method will require the learning rate to converge to zero, which is not a reasonable requirement to impose in many practical applications. Recently, deterministic discrete time (DDT) systems have been proposed instead to indirectly interpret the dynamics of the learning algorithms. Unlike DCT systems, DDT systems allow learning rates to be constant (which can be a nonzero). This paper will provide some important results relating to the convergence of a DDT system of Oja's PCA learning algorithm. It has the following contributions: 1) A number of invariant sets are obtained, based on which we can show that any trajectory starting from a point in the invariant set will remain in the set forever. Thus, the nondivergence of the trajectories is guaranteed. 2) The convergence of the DDT system is analyzed rigorously. It is proven, in the paper, that almost all trajectories of the system starting from points in an invariant set will converge exponentially to the unit eigenvector associated with the largest eigenvalue of the correlation matrix. In addition, exponential convergence rate are obtained, providing useful guidelines for the selection of fast convergence learning rate. 3) Since the trajectories may diverge, the careful choice of initial vectors is an important issue. This paper suggests to use the domain of unit hyper sphere as initial vectors to guarantee convergence. 4) Simulation results will be furnished to illustrate the theoretical results achieved. © 2005 IEEE.
Source Title: IEEE Transactions on Neural Networks
URI: http://scholarbank.nus.edu.sg/handle/10635/55430
ISSN: 10459227
DOI: 10.1109/TNN.2005.852236
Appears in Collections:Staff Publications

Show full item record
Files in This Item:
There are no files associated with this item.

SCOPUSTM   
Citations

53
checked on Dec 7, 2017

WEB OF SCIENCETM
Citations

43
checked on Nov 23, 2017

Page view(s)

27
checked on Dec 11, 2017

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.