EconPapers    
Economics at your fingertips  
 

Global convergence of an adaptive minor component extraction algorithm

Dezhong Peng and Zhang Yi

Chaos, Solitons & Fractals, 2008, vol. 35, issue 3, 550-561

Abstract: The convergence of neural networks minor component analysis (MCA) learning algorithms is crucial for practical applications. In this paper, we will analyze the global convergence of an adaptive minor component extraction algorithm via a corresponding deterministic discrete time (DDT) system. It is shown that if the learning rate satisfies certain conditions, almost all the trajectories of the DDT system are bounded and converge to minor component of the autocorrelation matrix of input data. Simulations are carried out to illustrate the results achieved.

Date: 2008
References: View references in EconPapers View complete reference list from CitEc
Citations:

Downloads: (external link)
http://www.sciencedirect.com/science/article/pii/S0960077906005042
Full text for ScienceDirect subscribers only

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:eee:chsofr:v:35:y:2008:i:3:p:550-561

DOI: 10.1016/j.chaos.2006.05.051

Access Statistics for this article

Chaos, Solitons & Fractals is currently edited by Stefano Boccaletti and Stelios Bekiros

More articles in Chaos, Solitons & Fractals from Elsevier
Bibliographic data for series maintained by Thayer, Thomas R. ().

 
Page updated 2025-03-19
Handle: RePEc:eee:chsofr:v:35:y:2008:i:3:p:550-561