Convergence of Batch Split-Complex Backpropagation Algorithm for Complex-Valued Neural Networks
Huisheng Zhang,
Chao Zhang and
Wei Wu
Discrete Dynamics in Nature and Society, 2009, vol. 2009, 1-16
Abstract:
The batch split-complex backpropagation (BSCBP) algorithm for training complex-valued neural networks is considered. For constant learning rate, it is proved that the error function of BSCBP algorithm is monotone during the training iteration process, and the gradient of the error function tends to zero. By adding a moderate condition, the weights sequence itself is also proved to be convergent. A numerical example is given to support the theoretical analysis.
Date: 2009
References: Add references at CitEc
Citations:
Downloads: (external link)
http://downloads.hindawi.com/journals/DDNS/2009/329173.pdf (application/pdf)
http://downloads.hindawi.com/journals/DDNS/2009/329173.xml (text/xml)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:hin:jnddns:329173
DOI: 10.1155/2009/329173
Access Statistics for this article
More articles in Discrete Dynamics in Nature and Society from Hindawi
Bibliographic data for series maintained by Mohamed Abdelhakeem ().