Recent from talks
Contribute something
Nothing was collected or created yet.
Probit
View on WikipediaThis article may be too technical for most readers to understand. (January 2013) |

In statistics, the probit function converts a probability (a number between 0 and 1) into a score. This score indicates how many standard deviations from the mean a value from a standard normal distribution (or "bell curve") is. For example, a probability of 0.5 (50%) represents the exact middle of the distribution, so its probit score is 0. A smaller probability like 0.025 (2.5%) is far to the left on the curve, corresponding to a probit score of approximately −1.96.
The function is widely used in probit models, a type of regression analysis for binary outcomes (e.g., success/failure or pass/fail). It was first developed in toxicology to analyze dose-response relationships, such as how the percentage of pests killed by a pesticide changes with its concentration.[1] The probit function is also used to create Q–Q plots, a graphical tool for assessing whether a dataset is normally distributed.
Mathematically, the probit function is the quantile function (the inverse of the cumulative distribution function (CDF)) associated with the standard normal distribution. If the CDF is denoted by , then the probit function is defined as:
- .
This means that for any probability , the probit function finds the value such that the area under the standard normal curve to the left of is equal to .
Conceptual development
[edit]The idea of the probit function was published by Chester Ittner Bliss in a 1934 article in Science on how to treat data such as the percentage of a pest killed by a pesticide.[1] Bliss proposed transforming the percentage killed into a "probability unit" (or "probit") which was linearly related to the modern definition (he defined it arbitrarily as equal to 0 for 0.0001 and 1 for 0.9999):[2]
These arbitrary probability units have been termed "probits" ...
He included a table to aid other researchers to convert their kill percentages to his probit, which they could then plot against the logarithm of the dose and thereby, it was hoped, obtain a more or less straight line. Such a so-called probit model is still important in toxicology, as well as other fields. The approach is justified in particular if response variation can be rationalized as a lognormal distribution of tolerances among subjects on test, where the tolerance of a particular subject is the dose just sufficient for the response of interest.
The method introduced by Bliss was carried forward in Probit Analysis, an important text on toxicological applications by D. J. Finney.[3][4] Values tabled by Finney can be derived from probits as defined here by adding a value of 5. This distinction is summarized by Collett (p. 55):[5] "The original definition of a probit [with 5 added] was primarily to avoid having to work with negative probits; ... This definition is still used in some quarters, but in the major statistical software packages for what is referred to as probit analysis, probits are defined without the addition of 5." Probit methodology, including numerical optimization for fitting of probit functions, was introduced before widespread availability of electronic computing. When using tables, it was convenient to have probits uniformly positive. Common areas of application do not require positive probits.
Symmetries
[edit]Largely because of the central limit theorem, the standard normal distribution plays a fundamental role in probability theory and statistics. If we consider the familiar fact that the standard normal distribution places 95% of probability between −1.96 and 1.96 and is symmetric around zero, it follows that
The probit function gives the 'inverse' computation, generating a value of a standard normal random variable, associated with specified cumulative probability. Continuing the example,
- .
In general,
and
Diagnosing deviation of a distribution from normality
[edit]In addition to providing a basis for important types of regression, the probit function is useful in statistical analysis for diagnosing deviation from normality, according to the method of Q–Q plotting. If a set of data is actually a sample of a normal distribution, a plot of the values against their probit scores will be approximately linear. Specific deviations from normality such as asymmetry, heavy tails, or bimodality can be diagnosed based on detection of specific deviations from linearity. While the Q–Q plot can be used for comparison to any distribution family (not only the normal), the normal Q–Q plot is a relatively standard exploratory data analysis procedure because the assumption of normality is often a starting point for analysis.
Computation
[edit]The normal distribution CDF and its inverse are not available in closed form, and computation requires careful use of numerical procedures. However, the functions are widely available in software for statistics and probability modeling, and in spreadsheets. In computing environments where numerical implementations of the inverse error function are available, the probit function may be obtained as
An example is MATLAB, where an 'erfinv' function is available. The language Mathematica implements 'InverseErf'. Other environments directly implement the probit function as is shown in the following R code.
> qnorm(0.025)
[1] -1.959964
> pnorm(-1.96)
[1] 0.02499790
Details for computing the inverse error function can be found at [1]. Wichura gives a fast algorithm for computing the probit function to 16 decimal places; this is used in R to generate random variates for the normal distribution.[6]
An ordinary differential equation for the probit function
[edit]Another means of computation is based on forming a non-linear ordinary differential equation (ODE) for probit, as per the Steinbrecher and Shaw method.[7] Abbreviating the probit function as , the ODE is
where is the probability density function of w.
In the case of the Gaussian:
Differentiating again:
with the centre (initial) conditions
This equation may be solved by several methods, including the classical power series approach. From this, solutions of arbitrarily high accuracy may be developed based on Steinbrecher's approach to the series for the inverse error function. The power series solution is given by
where the coefficients satisfy the non-linear recurrence
with . In this form the ratio as .
Logit
[edit]
Closely related to the probit function (and probit model) are the logit function and logit model. The inverse of the logistic function is given by
Analogously to the probit model, we may assume that such a quantity is related linearly to a set of predictors, resulting in the logit model, the basis in particular of logistic regression model, the most prevalent form of regression analysis for categorical response data. In current statistical practice, probit and logit regression models are often handled as cases of the generalized linear model.
See also
[edit]- Detection error tradeoff graphs (DET graphs, an alternative to the ROC)
- Logistic regression (a.k.a. logit model)
- Logit
- Probit model
- Multinomial probit
- Q–Q plot
- Continuous function
- Monotonic function
- Quantile function
- Sigmoid function
- Rankit analysis, also developed by Chester Bliss
- Ridit scoring
References
[edit]- ^ a b Bliss, C. I. (1934). "The method of probits". Science. 79 (2037): 38–39. Bibcode:1934Sci....79...38B. doi:10.1126/science.79.2037.38. JSTOR 1659792. PMID 17813446.
- ^ Bliss 1934, p. 39.
- ^ Finney, D.J. (1947). Probit Analysis (1st ed.). Cambridge, UK: Cambridge University Press.
- ^ Finney, D.J. (1971). Probit Analysis (3rd ed.). Cambridge University Press, Cambridge, UK. ISBN 0-521-08041-X. OCLC 174198382.
- ^ Collett, D. (1991). Modelling Binary Data. Chapman and Hall / CRC.
- ^ Wichura, M.J. (1988). "Algorithm AS241: The Percentage Points of the Normal Distribution" (PDF). Applied Statistics. 37 (3). Blackwell Publishing: 477–484. doi:10.2307/2347330. JSTOR 2347330.
- ^ Steinbrecher, G.; Shaw, W.T. (2008). "Quantile mechanics" (PDF). European Journal of Applied Mathematics. 19 (2): 87–112. doi:10.1017/S0956792508007341. S2CID 6899308.
External links
[edit]Probit
View on GrokipediaDefinition and History
Definition of the Probit Function
The probit function, denoted or , is defined as the inverse of the cumulative distribution function (CDF) of the standard normal distribution . It transforms a probability into the corresponding quantile on the real line such that . This mapping allows the probit to convert bounded probabilities into unbounded z-scores, which are useful for linearizing sigmoid response curves in statistical analysis. For instance, , reflecting the mean of the standard normal distribution, and , which marks the approximate upper limit of the central 95% probability interval. Equivalently, the probit function can be expressed in terms of the inverse error function: where denotes the inverse error function, leveraging the known relation between the normal CDF and the error function . In probabilistic interpretation, the probit function associates cumulative probabilities with normal quantiles, enabling standardization of data in models that assume underlying normal latent variables. The term "probit" originated from Chester Ittner Bliss, who defined it as 5 plus the normal deviate to ensure positive values for early computational tables, though the contemporary form omits this shift in favor of the direct inverse CDF.[7]Historical Development
The concept of the probit transformation emerged in the early 1930s as a method to linearize sigmoid dose-response curves in biological assays, particularly for quantal responses such as survival or mortality. In 1933, John H. Gaddum proposed using the inverse of the cumulative normal distribution function to model such relationships in his report on methods for biological assays depending on quantal responses, providing an early foundation for handling binary outcomes in toxicology and pharmacology.[8] This approach was formalized and popularized by Chester Ittner Bliss in 1934, who introduced the term "probit"—a contraction of "probability unit"—in his seminal paper analyzing pesticide effectiveness on insects. Bliss defined the probit as the normal equivalent deviate plus 5 to avoid negative values, applying it to transform empirical proportions of affected subjects into a scale suitable for linear regression in dose-response studies. His work, rooted in bioassays for agricultural and toxicological applications, marked the inception of probit analysis as a standard tool for estimating median effective doses.[9] David J. Finney significantly advanced the methodology in his 1947 book Probit Analysis: A Statistical Treatment of the Sigmoid Response Curve, which provided comprehensive tables, maximum likelihood estimation procedures, and extensions to multi-dose designs. Finney's contributions standardized probit methods, shifting emphasis toward rigorous statistical inference while retaining Bliss's framework; a second edition in 1962 and a third edition in 1971 incorporated computational refinements but preserved the core approach. Their collaboration, beginning in the late 1940s, further refined estimation techniques for quantal data.[10] Following World War II, probit analysis gained traction in econometrics and broader statistics during the 1950s, integrating into regression frameworks for modeling binary choices and probabilities beyond bioassays. This period saw its adaptation for economic applications, such as labor market decisions and consumer behavior, leveraging the normal distribution's interpretability. By the mid-20th century, probit had become a cornerstone of limited dependent variable modeling, with key milestones including its incorporation into statistical software and textbooks. Theoretical developments remained stable after Finney's 1971 edition, with no major shifts in the probit paradigm, though computational advances in the 1980s enhanced practicality. Notably, Michael J. Wichura's 1988 algorithm provided a high-precision method for evaluating the inverse normal cumulative distribution function, enabling efficient probit computations in numerical software and facilitating the transition to unshifted probits in modern implementations.Mathematical Properties
Functional Form and Symmetries
The probit function, denoted as , where is the cumulative distribution function (CDF) of the standard normal distribution , derives its functional form directly from the inversion of this CDF. The standard normal CDF exhibits even symmetry in its probability density function (PDF), , which implies the key symmetry property for all . Consequently, applying the inverse yields the probit symmetry: for .[11][12] This symmetry establishes the probit function as odd with respect to the midpoint , satisfying . At , , anchoring the function's antisymmetry around this point. In statistical modeling, this odd property facilitates balanced interpretation of binary outcomes, where deviations from 0.5 in probability correspond to symmetric positive and negative shifts in the underlying latent normal variable, promoting equitable treatment of complementary events such as success and failure.[12] A probit-specific relation arises in the ratio of the standard normal PDF to the CDF, where , which serves as the reverse hazard rate for the normal distribution and informs hazard-like interpretations in probit-based survival or selection analyses. The derivative of the probit function underscores this connection: . Near the boundaries, the function displays unbounded asymptotic behavior: as , , and as , , reflecting the infinite tails of the normal distribution without finite limits.[13][12]Relation to the Normal Distribution
The probit function serves as the quantile function of the standard normal distribution, defined such that where and . This formulation positions the probit as the inverse of the cumulative distribution function (CDF) , directly linking it to z-scores that quantify deviations from the mean in units of standard deviation. In this context, applying the probit transformation standardizes probabilities to a normal scale, facilitating comparisons across distributions and enabling the interpretation of as the area under the standard normal curve up to . This connection underpins the probit's utility in statistical modeling, where it maps bounded probabilities to the unbounded real line while preserving the properties of normality. A precise mathematical relation exists between the probit and the error function, a fundamental special function in probability theory. The error function is given by and its inverse allows expression of the probit as This derivation follows from the identity ; inverting both sides yields the probit in terms of the inverse error function, providing an analytical bridge to tabulated values and computational routines for the normal quantile. In generalized linear models for binary outcomes, the probit link function leverages this normal connection by assuming a latent continuous variable with , where the observed binary response is if and otherwise. The probability thus normalizes non-normal probabilities via the inverse CDF, effectively transforming the linear predictor to the probability scale under latent normal errors. This setup extends linear regression principles to dichotomous data while maintaining the normalizing properties of the standard normal. The selection of the normal CDF for the probit over alternatives like the logistic emphasizes theoretical foundations rooted in symmetry and the central limit theorem. The normal distribution's symmetry ensures balanced tail behavior, aligning with assumptions of equitable error distributions in latent models. Furthermore, when binary outcomes result from aggregating numerous independent small effects—such as in random utility maximization—the central limit theorem justifies approximating the latent error as normal, as the sum of such effects converges to normality under mild conditions. This rationale supports the probit's prominence in bioassay, econometrics, and choice modeling, where aggregation is common.Computation Methods
Numerical Algorithms
The probit function, defined as the inverse of the standard normal cumulative distribution function , has no closed-form expression and relies on numerical inversion techniques for evaluation.[14] These methods typically involve solving the root-finding problem iteratively, starting from an initial guess for based on the probability . Common approaches include the Newton-Raphson method, which updates the estimate as where is the standard normal density, and Halley's method, a higher-order variant that incorporates the second derivative for cubic convergence: . Both methods converge rapidly near but require careful initial approximations for tail probabilities to avoid slow convergence or overflow.[15] A widely adopted high-precision algorithm is AS 241 by Wichura (1988), which computes the inverse for using a minimax rational approximation in the central region combined with asymptotic expansions and continued fractions in the tails, achieving an absolute error bound of less than across the range.[14] For efficient approximations, rational function series derived from Chebyshev polynomials are employed; since , these provide relative errors below in double precision for most practical . Continued fractions offer another approximation strategy, particularly effective near , by expanding the inverse as a series that converges uniformly in the moderate tails. In modern software, these algorithms are implemented for seamless computation. The R functionqnorm(p) in the base stats package uses Wichura's AS 241 algorithm, returning with machine-precision accuracy and handling edge cases by yielding for and for .[16] Similarly, Python's SciPy library provides scipy.stats.norm.ppf(p) as the percent point function, relying on optimized C implementations of rational approximations and iterative refinement for the inverse CDF, with limits to at the boundaries. Accuracy is constrained by floating-point precision (typically 15-16 decimal digits in double precision), beyond which underflow or overflow issues arise in the tails. Historically, prior to electronic computers, probit values were derived from manually computed tables, such as the extensive working probits and weights tabulated by Finney and Stevens (1948) using mechanical calculators for bioassay applications.[17][18]
Differential Equation Formulation
The probit function , defined as the inverse of the standard normal cumulative distribution function , satisfies the first-order nonlinear ordinary differential equation (ODE) where is the probability density function of the standard normal distribution.[19] This ODE follows from the defining relation , upon differentiation with respect to , since . The initial condition is , reflecting the symmetry of the normal distribution where .[19] A power series solution to this ODE can be obtained via Taylor expansion centered at : with and subsequent coefficients determined recursively by substituting the series into the ODE and matching powers of . The recursion leverages the structure of to compute higher-order terms efficiently, enabling high-order accuracy; for example, up to 20 terms yield precision on the order of .[19] This ODE-based power series approach circumvents the need for root-finding iterations typical in numerical inversion of , making it particularly useful for symbolic computations or repeated evaluations where varies. If direct numerical integration of the ODE is required, standard solvers can be applied, though the series often provides superior efficiency for the probit due to the expansion point at the median. The formulation of such ODEs and their power series solutions for quantile functions, including the probit, was developed in the context of quantile mechanics by Steinbrecher and Shaw.[19]Applications
Probit Regression Models
Probit regression models are used to analyze binary outcome variables, where the probability of the outcome equals 1 given covariates is modeled using the cumulative distribution function of the standard normal distribution.[20] The model assumes an underlying latent variable , where , and the observed binary variable if and otherwise, leading to , with denoting the standard normal CDF.[20] Equivalently, the probit link function is , transforming the probability to the linear predictor.[20] Estimation of the probit model parameters is typically performed via maximum likelihood, maximizing the log-likelihood , or equivalently the likelihood .[20] Numerical optimization methods such as Newton-Raphson or BFGS are employed due to the absence of closed-form solutions, with the score function and Hessian facilitating iterative convergence.[20] The coefficients represent changes in the latent variable scale, but interpretation focuses on marginal effects, given by , where is the standard normal PDF; these effects vary across observations, unlike the constant marginal effects in linear probability models.[20] The model's normalization of the error variance to 1 introduces scale invariance issues, as estimates are identified only up to this fixed variance, precluding direct comparisons of effect magnitudes across models without rescaling.[20] Probit models offer advantages over ordinary least squares for binary outcomes by avoiding predicted probabilities outside [0,1] and heteroscedasticity inherent in linear approximations of nonlinear relationships.[20] Their adoption in econometrics surged post-1950s, building on early applications like probit analysis in bioassay, with influential work extending to economic choice modeling. In modern extensions, probit regression is widely applied in labor economics to model participation decisions, such as female labor force entry, where covariates like education and wages predict binary employment status.[21] Implementation is supported in software like Stata'sprobit command for maximum likelihood fitting and R's glm function with family=binomial(link="probit") for generalized linear modeling.[22]