# Quasi-Monte Carlo methods in finance

High-dimensional integrals in hundreds or thousands of variables occur commonly in finance. These integrals have to be computed numerically to within a threshold $epsilon$. If the integral is of dimension $d$ then in the worst case, where one has a guarantee of error at most $epsilon$, the computational complexity is typically of order $epsilon^\{-d\}$. That is, the problem suffers the curse of dimensionality. In 1977 P. Boyle, University of Waterloo, proposed using Monte Carlo (MC) to evaluate optionsBoyle, P. (1977), Options: a Monte Carlo approach, J. Financial Economics, 4, 323-338.] . Starting in early 1992, J. F. Traub, Columbia University, and a graduate student at the time, S. Paskov, used quasi-Monte Carlo (QMC) to price a CMO with parameters specified by Goldman Sachs. Even though it was believed by the world's leading experts that QMC should not be used for high dimensional integration, Paskov and Traub found that QMC beat MC by one to three orders of magnitude and also enjoyed other desirable attributes. Their results were first publishedPaskov, S. H. and Traub, J. F. (1995), Faster evaluation of financial derivatives, J. Portfolio Management, 22(1), 113-120.] in 1995. Today QMC is widely used in the financial sector to value financial derivatives.

QMC is not a panacea for all high dimensional integrals. A number of explanations have been proposed for why QMC is so good for financial derivatives. This continues to be a very fruitful research area.

**Monte Carlo and quasi-Monte Carlo methods**Integrals in hundreds or thousands of variables are common in computational finance. These have to be approximated numerically to within an error threshold $epsilon$. It is well known that if a worst case guarantee of error at most $epsilon$ is required then the computational complexity of integration may be exponential in $d$, the dimension of the integrand; See Traub, J. F and Werschulz, A. G. (1998), Complexity and Information, Cambridge University Press, Cambridge, UK.] Ch. 3 for details. To break this curse of dimesionality one can use the Monte Carlo (MC) method defined by

:$varphi^\{mathop\{\; m\; MC(f)=frac\; 1n\; sum\_\{i=1\}^nf(x\_i),$

where the evaluation points $x\_i$ are randomly chosen. It is well know that the expected error of Monte Carlo is of order $n^\{-1/2\}$. Thus the cost of the algorithm that has error $epsilon$ is of order $epsilon^\{-2\}$ breaking the curse of dimensionality.

Of course in computational practice pseudo-random points are used. Figure 1 shows the distribution of 500 pseudo-random points on the unit square. Note there are regions where there are no points and other regions where there are clusters of points. It would be desirable to sample the integrand at uniformly distributed points. A rectangular grid would be uniform but even if there were only 2 grid points in each Cartesian direction there would be $2^d$ points. So the desiratum should be as few points as possible chosen as uniform as possible.

It turns out there is a well developed part of number theory which deals exactly with this desiratum. Discrepancy is a measure of deviation from uniformity so what one wants are low discrepancy sequences (LDS) Niederreiter, H. (1992), Random Number Generation and Quasi-Monte Carlo Methods, CBMS-NSF Regional Conference Series in Applied Mathematics, SIAM, Pliladelphia. ] . Numerous LDS have been created named after their inventors, e.g.

*Halton

*Hammersley

*Sobol

*Faure

*NiederreiterFigure 2. gives the distribution of 500 LDS points.The quasi-Monte Carlo (QMC) method is defined by

:$varphi^\{mathop\{\; m\; QMC(f)=frac\; 1n\; sum\_\{i=1\}^nf(x\_i),$

where the $x\_i$ belong to an LDS. The standard terminology quasi-Monte Carlo is somewhat unfortunate since MC is a randomized method whereas QMC is purely deterministic.

The uniform distribution of LDS is desirable. But the worst case error of QMC is of order

:$frac\{(log\; n)^d\}\{n\},$

where $n$ is the number of sample points. See for the theory of LDS and references to the literature. The rate of convergence of LDS may be contrasted with the expected rate of convergence of MC which is $n^\{-1/2\}$. For $d$ small the rate of convergence of QMC is faster than MC but for $d$ large the factor $(log\; n)^d$ is devastating. For example, if $d=360$, then even with $log\; n=2$ the QMC error is proportional to $2^\{360\}$. Thus it was widely believed by the world's leading experts that QMC should not be used for high dimensional integration. For example, in 1992 Bratley, Fox and Niederreiter Bratley, P., Fox, B. L. and Niederreiter, H. (1992), Implementation and tests of low-discrepancy sequences, ACM Transactions on Modelling and Computer Simulation, Vol. 2, No. 3, 195-213.] performed extrensive testing on certain mathematical problems. They conclude "in high-dimensional problems (say $d\; >\; 12$), QMC seems to offer no practical advantage over MC". In 1993, Rensburg and Torrievan Rensburg, E. J. J. and Torrie, G. M. (1993), Estimation of multidimensional integrals: is Monte Carlo the best method? J. Phys. A: Math. Gen., 26(4), 943-953.] compared QMC with MC for the numerical estimation of high dimensional integrals which occur in computing virial coefficients for the hard-sphere fluid. They conclude QMC is more effective than MC only if $d<10$. As we shall see, tests on 360-dimensional integrals arising from a collateralized mortgage obligation (CMO) lead to very different conclusions.

Woźniakowski's 1991 paper Woźniakowski, H. (1991), Average case complexity of multivariate integration, Bull. Amer. Math. Soc. (New Ser.), 24(1), 185-194.] showing the connection between average case complexity of integration and QMC led to new interest in QMC. Woźniakowski's result received considerable coverage in the scientific press Cipra, B. A. (1991), Multivariate Integration: It ain't so tough (on average), SIAM NEWS, 28 March.] Traub, J. F. and Woźniakowski, H. (1994), Breaking intractability, Scientific American, 270(1), January, 102-107.] . In early 1992, I. T. Vanderhoof, New York University, became aware of Woźniakowski's result and gave Woźniakowski's colleague J. F. Traub, Columbia University, a CMO with parameters set by Goldman Sachs. This CMO had 10 tranches each requiring the computation of a 360 dimensional integral. Traub asked a Ph.D. student, Spassimir Paskov, to compare QMC with MC for the CMO. In 1992 Paskov built a software system called FinDer and ran extensive tests. To the Columbia's research group's surprise and initial disbelief Paskov reported that QMC was always superior to MC in a number of ways. Details are given below. Preliminary results were presented by Paskov and Traub to a number of Wall Street firms in Fall 1993 and Spring 1994. The firms were initially skeptical of the claim that QMC was superior to MC for pricing financial derivatives. A January 1994 article in Scientific American by Traub and Woźniakowski discussed the theoretical issues and reported that "Preliminary results obtained by testing certain finance problems suggests the superiority of the deterministic methods in practice". In Fall 1994 Paskov wrote a Columbia University Computer Science Report which appeared in slightly modified form in 1997 Paskov, S. H., New methodologies for valuing derivatives, 545-582, in Mathematics of Derivative Securities, S. Pliska and M. Dempster eds., Cambridge University Press, Cambridge.] .

In Fall 1995Paskov and Traub published a paper in the "Journal of Portfolio Management". They compared MC and two QMC methods. The two deterministic methods used Sobol and Halton points. Since better LDS were created later we won't report on the comparison between Sobol and Halton points here. We do summarize the conclusions regarding MC and QMC on the 10 tranche CMO

*QMC methods converge significantly faster than MC

*MC is sensitive to the initial seed

*The convergence of QMC is smoother than the convergence of MC. This makes automatic termination easier for QMC.To summarize, QMC beats MC for the CMO on accuracy, confidence level, and speed.

This paper was followed by reports on tests by a number of researchers which also led to the conclusion the QMC is superior to MC for a variety of high-dimensional finance problems. This includes papers by Caflisch and Morokoff (1996) Caflisch, R. E. and Morokoff, W. (1996), Quasi-Monte Carlo computation of a finance problem, 15-30, in Proceedings Workshop on Quasi-Monte Carlo Methods and their Applications, 11 December, 1995, K.-T. Fang and F. Hickernell eds., Hong Kong Baptist University.] ,Joy, Boyle, Tan (1996)Joy, C., Boyle, P. P. and Tang, K. S. (1996), Quasi-Monte Carlo methods in numerical finance, Management Science, 42(6), 926-938.] , Ninomiya and Tezuka (1996)Ninomiya, S. and Tezuka, S. (1996), Toward real-time pricing of complex financial derivatives, Appl. Math. Finance, 3, 1-20.] , Papageorgiou and Traub (1996) Papageorgiou, A. and Traub, J. F. (1996), Beating Monte Carlo, Risk, 9(6), 63-65.] , Ackworth, Broadie and Glasserman (1997)Ackworth, P., Broadie, M. and Glasserman, P. (1997), A comparison of some Monte Carlo techniques for option pricing, 1-18, in Monte Carlo and Quasi-Monte Carlo Methods '96, H. Hellekalek, P. Larcher and G. Zinterhof eds., Springer Verlag, New York.]

Wewho confine ourselves here to reporting on further testing of the CMO carried out by Anargyros Papageorgiou who developed an improved version of the FinDer software system. The new results include the following:

***Small number of sample points:**For the hardest CMO tranche QMC using the generalized Faure LDS due to S. TezukaTezuka, S., Uniform Random Numbers:Theory and Practice, Kluwer, Netherlands.] achieves accuracy $10^\{-2\}$ with just 170 points. MC requires 2700 points for the same accuracy. The significance of this is that due to future interest rates and prepayment rates being unknown financial firms are content with accuracy of $10^\{-2\}$.

***Large number of sample points:**The advantage of QMC over MC is further amplified as the sample size and accuracy demands grow. In particular, QMC is 20 to 50 times faster that MC with moderate sample sizes, and can be up to 1000 times faster than MC when high accuracy is desired QMC.**Theoretical explanations**The results reported so far in this article are empirical. A number of possible theoretical explanations have been advanced. This has been a very research rich area leading to powerful new concepts but a definite answer has not been obtained.

A posible explanation of why QMC is good for finance is the following. Consider a tranche of the CMO mentioned earlier. The integral gives expected future cash flows from a basket of 30 year mortgages at 360 monthly intervals. Because of the discounted value of money variables representing future times are increasingly less important. In a seminal paper I. Sloan and H. WoźniakowskiSloan, I. and Woźniakowski, H. (1998), When are quasi-Monte Carlo algorithms efficient for high dimensional integrals?, J. Complexity, 14(1), 1-33.] introduced the idea of weighted spaces. In these spaces the dependence on the successive variables can be moderated by weights. If the weights decrease sufficiently rapidly the curse of dimensionality is broken even with a worst case guarantee. This paper led toa great amount of work on the tractability of integration and other problems. Novak, E. and Wo�zniakowski, H. (2008), Tractability of multivariate problems, European Mathematical Society, Zurich (forthcoming). ] A problem is tractable when its complexity is of order $epsilon^\{-p\}$ and $p$ is independent of the dimension.

On the other hand, "effective dimension" was proposed by Caflisch, Morokoff and Owen Caflisch, R. E., Morokoff, W. and Owen, A. B. (1997), Valuation of mortgage backedsecurities using Brownian bridges to reduce effective dimension, Journal ofComputational Finance, 1, 27-46.] as an indicatorof the difficulty of high dimensional integration. The purpose was to explainthe remarkable success of quasi-Monte Carlo (QMC) in approximating the veryhigh dimensional integrals in finance. They argued thatthe integrands are of low effective dimension and that is why QMC is muchfaster than Monte Carlo (MC).The impact of the arguments of Caflisch et al. was great. A number of papers deal with the relationship between the error of QMC and the effective dimension Hickernell, F. J. (1998), Lattice rules: how well do they measure up?, in P. Hellekalek and G. Larcher (Eds.), Random and Quasi-Random Point Sets, Springer, 109-166.] Wang, X. and Sloan, I. H. (2005), Why are high-dimensional finance problems often of low effective dimension?, SIAM Journal on Scientific Computing, 27(1),159-183. ] .

It is known that QMC fails for certain functions that have high effective dimension. Howerver, low effective dimension is not a necessary condition for QMC to beat MC and for high dimensional integration to be tractable. In 2005, Tezuka Tezuka, S. (2005), On the necessity of low-effective dimension, Journal of Complexity, 21, 710-721. ] exhibited a class of functions of$d$ variables, all with maximum effective dimension equal to $d$. For these functions QMC is very fast since its convergence rate is of order $n^\{-1\}$, where $n$ is the number of function evaluations.

**Isotropic integrals**QMC can also be superior to MC and to other methods for isotropic problems, that is, problems where all variables are equally important. For example, Papageorgiou and TraubPapageorgiou, A. and Traub, J. F. (1997), Faster evaluation of multidimensionalintegrals, Computers in Phyics, 11(6), 574-578.] reported test results on the model integration problems suggested by the physicist B. D. Keister Keister, B. D. (1996), Multidimensional quadrature algorithms, Computers in Physics, 10(20), 119-122.]

:$left(frac\; 1\{2pi\}\; ight)^\{d/2\}\; int\_\{mathbb\; R^d\}cos(|x|\; )e^\{-|\; x|^2\},dx,$

where $|cdot|$ denotes the Euclidean norm and $d=25$. Keister reports that using a standard numerical method some 220,000 points were needed to obtain a relative error on the order of $10^\{-2\}$. A QMC calculation using the generalized Faure low discrepancy sequence (QMC-GF) used only 500 points to obtain the same relative error. The same integral was tested for a range of values of $d$ up to $d=100$. Its error was

:$ccdot\; n^\{-1\},$

$c<110$, where $n$ is the number of evaluations of $f$. This may be compared with the MC method whose error was proportional to $n^\{-1/2\}$.

These are empirical results. In a theoretical investigation PapageorgiouPapageorgiou, A. (2001), Fast convergence of quasi-Monte Carlo for a class of isotropic integrals, Math. Comp., 70, 297-306.] proved that the convergence rate of QMC for a class of $d$-dimensional isotropic integrals which includes the integral defined above is of the order

:$sqrt\{log\; n\}/n.$

This is with a worst case guarantee compared to the expected convergence rate of $n^\{-1/2\}$ of Monte Carlo and shows the superiority of QMC for this type of integral.

In another theoretical investigation PapageorgiouPapageorgiou, A. (2003), Sufficient conditions for fast quasi-Monte Carlo convergence, J. Complexity, 19(3), 332-351.] presented sufficient conditions for fast QMC convergence. The conditions apply to isotropic and non-isotropic problems and, in particular, to a number of problems in computational finance. He presented classes of functions where even in the worst case the convergence rate of QMC is of order

::::$n^\{-1+p(log\; n)^\{-1/2,$

where $pge\; 0$ is a constant that depends on the class of functions.

But this is only a sufficient condition and leaves open the major question we pose in the next section.

**Open questions**# Characterize for which high-dimensional integration problems QMC is superior to MC.

# Characterize types of financial instruments for which QMC is superior to MC.**References**

*Wikimedia Foundation.
2010.*

### Look at other dictionaries:

**Monte Carlo methods in finance**— Monte Carlo methods are used in finance and mathematical finance to value and analyze (complex) instruments, portfolios and investments by simulating the various sources of uncertainty affecting their value, and then determining their average… … Wikipedia**Quasi-Monte Carlo method**— In numerical analysis, a quasi Monte Carlo method is a method for the computation of an integral (or some other problem) that is based on low discrepancy sequences. This is in contrast to a regular Monte Carlo method, which is based on sequences… … Wikipedia**Monte Carlo method**— Not to be confused with Monte Carlo algorithm. Computational physics … Wikipedia**Outline of finance**— The following outline is provided as an overview of and topical guide to finance: Finance – addresses the ways in which individuals, businesses and organizations raise, allocate and use monetary resources over time, taking into account the risks… … Wikipedia**List of numerical analysis topics**— This is a list of numerical analysis topics, by Wikipedia page. Contents 1 General 2 Error 3 Elementary and special functions 4 Numerical linear algebra … Wikipedia**List of mathematics articles (Q)**— NOTOC Q Q analog Q analysis Q derivative Q difference polynomial Q exponential Q factor Q Pochhammer symbol Q Q plot Q statistic Q systems Q test Q theta function Q Vandermonde identity Q.E.D. QED project QR algorithm QR decomposition Quadratic… … Wikipedia**Quantile function**— See also quantile. In probability theory, a quantile function of a probability distribution is the inverse F −1 of its cumulative distribution function (cdf) F . Assuming a continuous and strictly monotonic distribution function, scriptstyle… … Wikipedia**List of statistics topics**— Please add any Wikipedia articles related to statistics that are not already on this list.The Related changes link in the margin of this page (below search) leads to a list of the most recent changes to the articles listed below. To see the most… … Wikipedia**Information-based complexity**— (IBC) studies optimal algorithms and computational complexity for the continuous problems which arise in physical science, economics, engineering, and mathematical finance. IBC has studied such continuous problems as path integration, partial… … Wikipedia**Joseph F Traub**— Infobox Scientist name = Joseph Frederick Traub caption = birth date = Birth date|1932|06|24 birth place = death date = death place = residence = nationality = field = Computer Science work institution = Columbia University alma mater = doctoral… … Wikipedia