Hubbry Logo
Control variatesControl variatesMain
Open search
Control variates
Community hub
Control variates
logo
7 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Contribute something
Control variates
Control variates
from Wikipedia

The control variates method is a variance reduction technique used in Monte Carlo methods. It exploits information about the errors in estimates of known quantities to reduce the error of an estimate of an unknown quantity.[1] [2][3]

Underlying principle

[edit]

Let the unknown parameter of interest be , and assume we have a statistic such that the expected value of m is μ: , i.e. m is an unbiased estimator for μ. Suppose we calculate another statistic such that is a known value. Then

is also an unbiased estimator for for any choice of the coefficient . The variance of the resulting estimator is

By differentiating the above expression with respect to , it can be shown that choosing the optimal coefficient

minimizes the variance of . (Note that this coefficient is the same as the coefficient obtained from a linear regression.) With this choice,

where

is the correlation coefficient of and . The greater the value of , the greater the variance reduction achieved.

In the case that , , and/or are unknown, they can be estimated across the Monte Carlo replicates. This is equivalent to solving a certain least squares system; therefore this technique is also known as regression sampling.

When the expectation of the control variable, , is not known analytically, it is still possible to increase the precision in estimating (for a given fixed simulation budget), provided that the two conditions are met: 1) evaluating is significantly cheaper than computing ; 2) the magnitude of the correlation coefficient is close to unity. [3]

Example

[edit]

We would like to estimate

using Monte Carlo integration. This integral is the expected value of , where

and U follows a uniform distribution [0, 1]. Using a sample of size n denote the points in the sample as . Then the estimate is given by

Now we introduce as a control variate with a known expected value and combine the two into a new estimate

Using realizations and an estimated optimal coefficient we obtain the following results

Estimate Variance
Classical estimate 0.69475 0.01947
Control variates 0.69295 0.00060

The variance was significantly reduced after using the control variates technique. (The exact result is .)

See also

[edit]

Notes

[edit]

References

[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
Control variates is a variance reduction technique used in Monte Carlo simulation to improve the efficiency of estimators by adjusting them with auxiliary random variables, known as control variates, whose expectations are known exactly. This method exploits the correlation between the target variable and the control variate to reduce the overall variance of the simulation output, allowing for more accurate estimates with fewer computational resources. The core idea behind control variates involves rewriting the expectation of interest, E[f(X)]E[f(X)], as E[f(X)h(X)]+E[h(X)]E[f(X) - h(X)] + E[h(X)], where h(X)h(X) is the control variate with a computable expectation E[h(X)]E[h(X)], and the variance of f(X)h(X)f(X) - h(X) is smaller than that of f(X)f(X) due to correlation between f(X)f(X) and h(X)h(X). In practice, the estimator is formed as A^=1ni=1n[f(Xi)β(h(Xi)E[h(X)])]\hat{A} = \frac{1}{n} \sum_{i=1}^n [f(X_i) - \beta (h(X_i) - E[h(X)])], where β\beta is a coefficient chosen to minimize variance, optimally set to β=Cov(f(X),h(X))Var(h(X))\beta^* = \frac{\mathrm{Cov}(f(X), h(X))}{\mathrm{Var}(h(X))}. This optimal choice yields a reduced variance of Var(A^)=Var(f(X))(1ρ2)\mathrm{Var}(\hat{A}) = \mathrm{Var}(f(X)) (1 - \rho^2), where ρ\rho is the correlation between f(X)f(X) and h(X)h(X), potentially achieving substantial reductions if ρ|\rho| is close to 1. The method was first rigorously formalized in the 1980s by Lavenberg and Welch, building on earlier ideas in simulations from the . Control variates are particularly effective when a simpler or "crude" version of the problem has an explicit solution, such as in where Black-Scholes formulas serve as controls for . The technique extends to multiple control variates by solving a for the optimal coefficients, further enhancing . It also connects to other methods, including conditional Monte Carlo (where the control effectively sets β=1\beta = 1), antithetic variates, stratification, and even nonparametric in constrained settings. Applications span , terminating simulations, and optimization problems, with empirical examples showing variance drops of over 40% in cases like valuation.

Introduction

Definition and Motivation

Monte Carlo methods provide a sampling-based approach to estimate expectations of the form θ=E[h(X)]\theta = \mathbb{E}[h(X)], where XX is a , by generating independent samples and averaging the function values h(Xi)h(X_i). This crude estimator converges to the true value by the , but its variance, which determines the estimation error, scales as σ2/n\sigma^2 / n with sample size nn, often requiring large nn for precision. Control variates constitute a post-processing technique in simulation, wherein the crude is adjusted using an auxiliary —termed the control variate—that correlates with the target quantity and possesses a known expectation. Specifically, if YY approximates θ\theta and ZZ is the control variate with E[Z]=μ\mathbb{E}[Z] = \mu known, the adjusted takes the form θ^c=Y+c(Zμ)\hat{\theta}_c = Y + c(Z - \mu), where the cc is chosen to minimize variance while preserving unbiasedness. This method leverages the correlation to refine estimates without additional sampling overhead. The primary motivation for control variates arises from the limitations of crude , particularly its high in challenging scenarios such as rare event simulation or high-dimensional integration, where probabilities are small or dependencies amplify , necessitating impractically large sample sizes for reliable accuracy. By exploiting negative between the and control variate, the technique can achieve a factor of up to 1ρ21 - \rho^2, where ρ\rho is the , potentially decreasing the required samples by orders of magnitude when ρ|\rho| is close to 1. This efficiency gain is especially valuable in fields like and physics, where computational costs are high.

Historical Development

The , foundational to simulation techniques including variance reduction strategies like control variates, was introduced by and Stanislaw Ulam in their seminal 1949 paper, which outlined probabilistic approaches to solving complex physical problems such as neutron diffusion. Control variates emerged shortly thereafter in the early 1950s as part of broader efforts to enhance computational efficiency in simulations, particularly at where early applications demanded reduced variance in estimates. Herman Kahn and A. W. Marshall played pivotal roles in pioneering control variates during this period, applying them to simulations to minimize sample sizes while maintaining accuracy; their 1953 paper formalized key aspects of the technique, including derivations for optimal coefficients that remain influential today. By the , the method gained theoretical rigor through comprehensive treatments in statistical literature, notably in John Hammersley and David Handscomb's 1964 textbook Monte Carlo Methods, which systematically described control variates alongside other tools and emphasized their practical implementation in multidimensional integrals. Control variates saw widespread adoption in the 1970s and 1980s across physics and emerging financial modeling, where simulations of stochastic processes benefited from its ease of integration with existing Monte Carlo frameworks, as highlighted in Reuven Rubinstein's 1981 book Simulation and the Monte Carlo Method. In the 2000s, extensions to quasi-Monte Carlo integration revitalized the technique for high-dimensional problems, with statisticians like Christiane Lemieux contributing key advancements in combining control variates with low-discrepancy sequences to achieve superior variance reduction in computational finance and statistics.

Theoretical Foundations

Monte Carlo Methods Overview

Monte Carlo methods are computational algorithms that rely on repeated random sampling to obtain numerical results, particularly for approximating expectations in probability distributions. The core idea involves estimating the E[f(X)]E[f(X)], where XX is a with a known distribution and ff is a function, by generating nn independent and identically distributed (i.i.d.) samples X1,X2,,XnX_1, X_2, \dots, X_n from the distribution of XX and computing the sample μ^=1ni=1nf(Xi)\hat{\mu} = \frac{1}{n} \sum_{i=1}^n f(X_i). This estimator is unbiased, meaning E[μ^]=E[f(X)]E[\hat{\mu}] = E[f(X)], and its variance is Var(f(X))n\frac{\text{Var}(f(X))}{n}, which decreases as the number of samples increases. These methods are particularly valuable for evaluating high-dimensional integrals and performing simulations where analytical solutions are intractable, such as in for modeling neutron diffusion during the or in for estimating thermodynamic properties. In finance, simulations are widely applied to price complex derivatives, assess risk through value-at-risk calculations, and model stochastic processes like asset price paths under the Black-Scholes framework. By leveraging random sampling, these techniques handle the "curse of dimensionality" more effectively than deterministic quadrature methods, which suffer from exponential growth in computational cost with increasing dimensions. The reliability of Monte Carlo estimates stems from fundamental results in probability theory. The guarantees that, as nn \to \infty, the sample average μ^\hat{\mu} converges almost surely to the true expectation E[f(X)]E[f(X)], ensuring consistency of the method. Additionally, the implies that, for large nn, the distribution of n(μ^E[f(X)])\sqrt{n} (\hat{\mu} - E[f(X)])
Add your contribution
Related Hubs
Contribute something
User Avatar
No comments yet.