Neural Network Circuits and Parallel Implementations
Ke-Lin Du () and
M. N. S. Swamy
Additional contact information
Ke-Lin Du: Concordia University, Department of Electrical and Computer Engineering
M. N. S. Swamy: Concordia University, Department of Electrical and Computer Engineering
Chapter Chapter 28 in Neural Networks and Statistical Learning, 2019, pp 829-851 from Springer
Abstract:
Abstract Hardware and parallel implementations can substantially speed up machine learning algorithms to extend their widespread applications. In this chapter, we first introduce various circuit realizations for popular neural network learning methods. We then introduce their parallel implementations on graphic processing units (GPUs), systolic arrays of processors, and parallel computers.
Date: 2019
References: Add references at CitEc
Citations:
There are no downloads for this item, see the EconPapers FAQ for hints about obtaining it.
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:spr:sprchp:978-1-4471-7452-3_28
Ordering information: This item can be ordered from
http://www.springer.com/9781447174523
DOI: 10.1007/978-1-4471-7452-3_28
Access Statistics for this chapter
More chapters in Springer Books from Springer
Bibliographic data for series maintained by Sonal Shukla () and Springer Nature Abstracting and Indexing ().