Implementation of artificial neural networks into hardware: Concepts and limitations
Karl F. Goser
Mathematics and Computers in Simulation (MATCOM), 1996, vol. 41, issue 1, 161-171
Abstract:
The implementation of artificial neural networks into hardware will only show their true potential since the networks need full parallel processing for real-time applications. By this way neural networks are a genuine challenge to microelectronics: not only many synapses have to be integrated with a high density but also many interconnections should be provided on many layers. First this paper describes the state-of-the-art and the potential of silicon technologies for artificial neural networks. Secondly the potential of nanoelectronics is demonstrated by some suggestions. For nanoelectronics one needs a specific system technique which gains the most of the advantages from the ultra large scale integration (ULSI). The key issue of this paper shows exemplarily how the Schrödinger equation from quantum mechanics can control both the characteristics of the devices and the algorithm of self-organization on system level, and the self-structuring of a system during implementation. Such differential equations seem to be the key algorithms for the ultimate hardware concepts in electronics.
Date: 1996
References: View complete reference list from CitEc
Citations:
Downloads: (external link)
http://www.sciencedirect.com/science/article/pii/0378475495000682
Full text for ScienceDirect subscribers only
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:eee:matcom:v:41:y:1996:i:1:p:161-171
DOI: 10.1016/0378-4754(95)00068-2
Access Statistics for this article
Mathematics and Computers in Simulation (MATCOM) is currently edited by Robert Beauwens
More articles in Mathematics and Computers in Simulation (MATCOM) from Elsevier
Bibliographic data for series maintained by Catherine Liu ().