Please use this identifier to cite or link to this item:
Title: Convergence analysis of a deterministic discrete time system of Oja's PCA learning algorithm
Authors: Yi, Z.
Ye, M.
Lv, J.C.
Tan, K.K. 
Keywords: Eigenvalue
Neural network
Oja's learning algorithm
Principal component analysis (PCA)
Issue Date: Nov-2005
Citation: Yi, Z., Ye, M., Lv, J.C., Tan, K.K. (2005-11). Convergence analysis of a deterministic discrete time system of Oja's PCA learning algorithm. IEEE Transactions on Neural Networks 16 (6) : 1318-1328. ScholarBank@NUS Repository.
Abstract: The convergence of Oja's principal component analysis (PCA) learning algorithms is a difficult topic for direct study and analysis. Traditionally, the convergence of these algorithms is indirectly analyzed via certain deterministic continuous time (DCT) systems. Such a method will require the learning rate to converge to zero, which is not a reasonable requirement to impose in many practical applications. Recently, deterministic discrete time (DDT) systems have been proposed instead to indirectly interpret the dynamics of the learning algorithms. Unlike DCT systems, DDT systems allow learning rates to be constant (which can be a nonzero). This paper will provide some important results relating to the convergence of a DDT system of Oja's PCA learning algorithm. It has the following contributions: 1) A number of invariant sets are obtained, based on which we can show that any trajectory starting from a point in the invariant set will remain in the set forever. Thus, the nondivergence of the trajectories is guaranteed. 2) The convergence of the DDT system is analyzed rigorously. It is proven, in the paper, that almost all trajectories of the system starting from points in an invariant set will converge exponentially to the unit eigenvector associated with the largest eigenvalue of the correlation matrix. In addition, exponential convergence rate are obtained, providing useful guidelines for the selection of fast convergence learning rate. 3) Since the trajectories may diverge, the careful choice of initial vectors is an important issue. This paper suggests to use the domain of unit hyper sphere as initial vectors to guarantee convergence. 4) Simulation results will be furnished to illustrate the theoretical results achieved. © 2005 IEEE.
Source Title: IEEE Transactions on Neural Networks
ISSN: 10459227
DOI: 10.1109/TNN.2005.852236
Appears in Collections:Staff Publications

Show full item record
Files in This Item:
There are no files associated with this item.


checked on Jan 14, 2019


checked on Jan 14, 2019

Page view(s)

checked on Jan 12, 2019

Google ScholarTM



Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.