Latin hypercube sampling with dependence and applications in finance
Natalie Packham and
Wolfgang M. Schmidt
Journal of Computational Finance
Abstract:
ABSTRACT In Monte Carlo simulation, Latin hypercube sampling (LHS) (McKay et al (1979)) is a well-known variance reduction technique for vectors of independent random variables. The method presented here, Latin hypercube sampling with dependence (LHSD), extends LHS to vectors of dependent random variables. The resulting estimator is shown to be consistent and asymptotically unbiased. For the bivariate case and under some conditions on the joint distribution, a central limit theorem together with a closed formula for the limit variance are derived. It is shown that for a class of estimators satisfying some monotonicity condition, the LHSD limit variance is never greater than the corresponding Monte Carlo limit variance. In some valuation examples of financial payoffs, when compared to standard Monte Carlo simulation, a variance reduction of factors up to 200 is achieved. We illustrate that LHSD is suited for problems with rare events and for high-dimensional problems, and that it may be combined with quasi-Monte Carlo methods.
References: Add references at CitEc
Citations:
Downloads: (external link)
https://www.risk.net/journal-of-computational-fina ... lications-in-finance (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:rsk:journ0:2160356
Access Statistics for this article
More articles in Journal of Computational Finance from Journal of Computational Finance
Bibliographic data for series maintained by Thomas Paine ().