Recent from talks
Nothing was collected or created yet.
Quasi-maximum likelihood estimate
View on WikipediaIn statistics a quasi-maximum likelihood estimate (QMLE), also known as a pseudo-likelihood estimate or a composite likelihood estimate, is an estimate of a parameter θ in a statistical model that is formed by maximizing a function that is related to the logarithm of the likelihood function, but in discussing the consistency and (asymptotic) variance-covariance matrix, we assume some parts of the distribution may be mis-specified.[1][2] In contrast, the maximum likelihood estimate maximizes the actual log likelihood function for the data and model. The function that is maximized to form a QMLE is often a simplified form of the actual log likelihood function. A common way to form such a simplified function is to use the log-likelihood function of a misspecified model that treats certain data values as being independent, even when in actuality they may not be. This removes any parameters from the model that are used to characterize these dependencies. Doing this only makes sense if the dependency structure is a nuisance parameter with respect to the goals of the analysis. [3] As long as the quasi-likelihood function that is maximized is not oversimplified, the QMLE (or composite likelihood estimate) is consistent and asymptotically normal. It is less efficient than the maximum likelihood estimate, but may only be slightly less efficient if the quasi-likelihood is constructed so as to minimize the loss of information relative to the actual likelihood.[4] Standard approaches to statistical inference that are used with maximum likelihood estimates, such as the formation of confidence intervals, and statistics for model comparison,[5] can be generalized to the quasi-maximum likelihood setting.
See also
[edit]References
[edit]- ^ Lindsay, Bruce G. (1988). "Composite likelihood methods". Statistical inference from stochastic processes (Ithaca, NY, 1987). Contemporary Mathematics. Vol. 80. Providence, RI: American Mathematical Society. pp. 221–239. doi:10.1090/conm/080/999014. MR 0999014.
- ^ Davidson, Russel; MacKinnon, James (2004). Econometric Theory and Methods. New York, New York: Oxford University Press. ISBN 978-0-19-512372-2.
- ^ Gourieroux, Christian; Monfort, Alain; Trognon, Alain (1984). "Pseudo Maximum Likelihood Methods: Theory" (PDF). Econometrica. 52 (3): 681–700. doi:10.2307/1913471. JSTOR 1913471. S2CID 122981013.
- ^ Cox, D.R.; Reid, Nancy (2004). "A note on pseudo-likelihood constructed from marginal densities". Biometrika. 91 (3): 729–737. CiteSeerX 10.1.1.136.7476. doi:10.1093/biomet/91.3.729.
- ^ Varin, Cristiano; Vidoni, Paolo (2005). "A note on composite likelihood inference and model selection" (PDF). Biometrika. 92 (3): 519–528. doi:10.1093/biomet/92.3.519.
Quasi-maximum likelihood estimate
View on GrokipediaBackground Concepts
Maximum Likelihood Estimation
Maximum likelihood estimation (MLE) is a fundamental statistical method for estimating the parameters of a probabilistic model from observed data. Given a sample of independent and identically distributed (i.i.d.) observations drawn from a probability density function , where is the unknown parameter vector, the likelihood function is defined as . The MLE, denoted , is the value of that maximizes this likelihood, assuming the model is correctly specified.[4] To facilitate computation, the maximization is typically performed on the log-likelihood: . The first derivative of the log-likelihood with respect to , known as the score function , plays a central role; at the maximum, . Under standard regularity conditions, such as differentiability of the log-likelihood and the existence of finite moments, the expected value of the score is zero: , ensuring that the true parameter satisfies the first-order condition in expectation. The method was developed by Ronald A. Fisher in the early 1920s, with its formal introduction in his 1922 paper, where he presented MLE as an optimal estimation procedure under correct model specification, offering desirable properties like efficiency relative to other estimators. A key result in this framework is the information matrix equality, which states that the variance of the score function equals the negative expected value of the second derivative of the log-likelihood: , where is the Fisher information matrix measuring the amount of information the sample carries about . This equality underpins the asymptotic efficiency of the MLE. Under the same regularity conditions, the MLE exhibits asymptotic normality, converging in distribution to a normal random variable centered at the true parameter with covariance given by the inverse Fisher information.[4]Likelihood Misspecification
Likelihood misspecification arises when the parametric family of probability distributions specified for maximum likelihood estimation does not encompass the true data-generating process, such that the assumed density deviates from the true density .[1] In this scenario, the maximum likelihood estimator fails to converge to the true parameter values, resulting in inconsistency.[1] The consequences of such misspecification are profound, including biased parameter estimates that do not recover the underlying true values, invalid standard errors that undermine hypothesis testing and confidence intervals, and the violation of the information matrix equality, where the expected outer product of the score differs from the negative expected Hessian.[1] A prevalent form of misspecification in regression analysis involves assuming normally distributed residuals when the true error distribution exhibits heavier tails, such as the Student-t distribution; this can severely distort inferences about random effects and variance components. Despite these issues, misspecification does not necessarily destroy all structural properties of the likelihood; in particular, the score function may retain a zero expected value and finite variance at a pseudo-true parameter value, laying the groundwork for quasi-maximum likelihood approaches that exploit these moments for robust estimation.[1]Formal Definition
Quasi-Likelihood Function
The quasi-likelihood function in the context of quasi-maximum likelihood estimation is the log-likelihood constructed from an assumed parametric probability density , which may not match the true data-generating distribution . It is defined as where the average is over independent observations. This function is maximized to obtain parameter estimates that approximate the data under the assumed model, even under misspecification. The resulting estimator converges to the pseudo-true parameter that minimizes the expected Kullback-Leibler divergence .[1] A key property is that the expected score under the true distribution vanishes at : , provided suitable regularity conditions hold, such as compactness of the parameter space and identifiability. This ensures consistency of the estimator without requiring correct specification of . The score function is generally , and for specific assumed densities (e.g., Gaussian in linear regression), it simplifies to forms like weighted least squares, but the framework applies broadly to nonlinear models.[3] In contrast to full maximum likelihood, which requires the assumed density to be correctly specified for optimal efficiency, the quasi-likelihood approach yields consistent estimates as long as the pseudo-true parameter is well-defined, making it robust to distributional misspecification while retaining the computational advantages of likelihood optimization.Quasi-Maximum Likelihood Estimator
The quasi-maximum likelihood estimator (QMLE), denoted , is defined as , where is the quasi-likelihood function based on an assumed but potentially misspecified density . This estimator, introduced by White (1982) in the context of misspecified parametric models, seeks to find the parameter values that best approximate the data under the chosen quasi-likelihood, even when the true data-generating process differs from the assumed one.[1] To obtain the QMLE, the optimization problem is typically solved by setting the score equations to zero: , where represents the individual score contributions. When a closed-form solution is unavailable—which is common for nonlinear models—numerical methods such as the Newton-Raphson algorithm are applied iteratively to converge to the maximum, relying on approximations of the Hessian matrix for updates. These procedures ensure computational feasibility while maintaining the estimator's properties under misspecification.[3] For inference, standard errors are computed using the sandwich variance estimator, which accounts for potential misspecification in the assumed density: where is the average negative Hessian (approximating the expected information matrix), and is the score outer-product matrix. This robust form, derived from the asymptotic covariance structure under misspecification, provides consistent estimates of the variability without assuming the quasi-likelihood is correctly specified.[3] In practice, QMLE is frequently implemented in statistical software such as R and Stata, often assuming a Gaussian quasi-likelihood for its analytical simplicity and ease of optimization in models like conditional heteroskedasticity or panel data regressions. For instance, R'ssandwich package computes the corresponding robust covariance matrices, while Stata's ml command with the vce(robust) option facilitates quasi-maximum likelihood fitting.
