EconPapers    
Economics at your fingertips  
 

Bayesian Lasso and multinomial logistic regression on GPU

Rok Češnovar and Erik Štrumbelj

PLOS ONE, 2017, vol. 12, issue 6, 1-17

Abstract: We describe an efficient Bayesian parallel GPU implementation of two classic statistical models—the Lasso and multinomial logistic regression. We focus on parallelizing the key components: matrix multiplication, matrix inversion, and sampling from the full conditionals. Our GPU implementations of Bayesian Lasso and multinomial logistic regression achieve 100-fold speedups on mid-level and high-end GPUs. Substantial speedups of 25 fold can also be achieved on older and lower end GPUs. Samplers are implemented in OpenCL and can be used on any type of GPU and other types of computational units, thereby being convenient and advantageous in practice compared to related work.

Date: 2017
References: View references in EconPapers View complete reference list from CitEc
Citations:

Downloads: (external link)
https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0180343 (text/html)
https://journals.plos.org/plosone/article/file?id= ... 80343&type=printable (application/pdf)

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:plo:pone00:0180343

DOI: 10.1371/journal.pone.0180343

Access Statistics for this article

More articles in PLOS ONE from Public Library of Science
Bibliographic data for series maintained by plosone (plosone@plos.org).

 
Page updated 2025-03-19
Handle: RePEc:plo:pone00:0180343