Global convergence of an adaptive minor component extraction algorithm
Dezhong Peng and
Zhang Yi
Chaos, Solitons & Fractals, 2008, vol. 35, issue 3, 550-561
Abstract:
The convergence of neural networks minor component analysis (MCA) learning algorithms is crucial for practical applications. In this paper, we will analyze the global convergence of an adaptive minor component extraction algorithm via a corresponding deterministic discrete time (DDT) system. It is shown that if the learning rate satisfies certain conditions, almost all the trajectories of the DDT system are bounded and converge to minor component of the autocorrelation matrix of input data. Simulations are carried out to illustrate the results achieved.
Date: 2008
References: View references in EconPapers View complete reference list from CitEc
Citations:
Downloads: (external link)
http://www.sciencedirect.com/science/article/pii/S0960077906005042
Full text for ScienceDirect subscribers only
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:eee:chsofr:v:35:y:2008:i:3:p:550-561
DOI: 10.1016/j.chaos.2006.05.051
Access Statistics for this article
Chaos, Solitons & Fractals is currently edited by Stefano Boccaletti and Stelios Bekiros
More articles in Chaos, Solitons & Fractals from Elsevier
Bibliographic data for series maintained by Thayer, Thomas R. ().