In
statistics
Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
, a probit model is a type of
regression where the
dependent variable
A variable is considered dependent if it depends on (or is hypothesized to depend on) an independent variable. Dependent variables are studied under the supposition or demand that they depend, by some law or rule (e.g., by a mathematical functio ...
can take only two values, for example married or not married. The word is a
portmanteau
In linguistics, a blend—also known as a blend word, lexical blend, or portmanteau—is a word formed by combining the meanings, and parts of the sounds, of two or more words together. , coming from ''probability'' + ''unit''. The purpose of the model is to estimate the probability that an observation with particular characteristics will fall into a specific one of the categories; moreover, classifying observations based on their predicted probabilities is a type of
binary classification model.
A
probit model is a popular specification for a
binary response model. As such it treats the same set of problems as does
logistic regression
In statistics, a logistic model (or logit model) is a statistical model that models the logit, log-odds of an event as a linear function (calculus), linear combination of one or more independent variables. In regression analysis, logistic regres ...
using similar techniques. When viewed in the
generalized linear model
In statistics, a generalized linear model (GLM) is a flexible generalization of ordinary linear regression. The GLM generalizes linear regression by allowing the linear model to be related to the response variable via a ''link function'' and by ...
framework, the probit model employs a
probit link function. It is most often estimated using the
maximum likelihood procedure, such an estimation being called a probit regression.
Conceptual framework
Suppose a response variable ''Y'' is ''binary'', that is it can have only
two possible outcomes which we will denote as 1 and 0. For example, ''Y'' may represent presence/absence of a certain condition, success/failure of some device, answer yes/no on a survey, etc. We also have a vector of
regressors ''X'', which are assumed to influence the outcome ''Y''. Specifically, we assume that the model takes the form
:
where ''P'' is the
probability
Probability is a branch of mathematics and statistics concerning events and numerical descriptions of how likely they are to occur. The probability of an event is a number between 0 and 1; the larger the probability, the more likely an e ...
and
is the cumulative distribution function (
CDF) of the
standard normal distribution. The parameters ''β'' are typically estimated by
maximum likelihood.
It is possible to motivate the probit model as a
latent variable model. Suppose there exists an auxiliary random variable
:
where ''ε'' ~ ''N''(0, 1). Then ''Y'' can be viewed as an indicator for whether this latent variable is positive:
:
The use of the standard normal distribution causes no
loss of generality compared with the use of a normal distribution with an arbitrary mean and standard deviation, because adding a fixed amount to the mean can be compensated by subtracting the same amount from the intercept, and multiplying the standard deviation by a fixed amount can be compensated by multiplying the weights by the same amount.
To see that the two models are equivalent, note that
:
Model estimation
Maximum likelihood estimation
Suppose data set
contains ''n'' independent
statistical units corresponding to the model above.
For the single observation, conditional on the vector of inputs of that observation, we have:
:
:
where
is a vector of
inputs, and
is a
vector of coefficients.
The likelihood of a single observation
is then
:
In fact, if
, then
, and if
, then
.
Since the observations are independent and identically distributed, then the likelihood of the entire sample, or the
joint likelihood, will be equal to the product of the likelihoods of the single observations:
:
The joint log-likelihood function is thus
:
The estimator
which maximizes this function will be
consistent
In deductive logic, a consistent theory is one that does not lead to a logical contradiction. A theory T is consistent if there is no formula \varphi such that both \varphi and its negation \lnot\varphi are elements of the set of consequences ...
, asymptotically normal and
efficient provided that
exists and is not singular. It can be shown that this log-likelihood function is globally
concave in
, and therefore standard numerical algorithms for optimization will converge rapidly to the unique maximum.
Asymptotic distribution for
is given by
:
where
:
and
is the Probability Density Function (
PDF
Portable document format (PDF), standardized as ISO 32000, is a file format developed by Adobe Inc., Adobe in 1992 to present documents, including text formatting and images, in a manner independent of application software, computer hardware, ...
) of standard normal distribution.
Semi-parametric and non-parametric maximum likelihood methods for probit-type and other related models are also available.
Berkson's minimum chi-square method
This method can be applied only when there are many observations of response variable
having the same value of the vector of regressors
(such situation may be referred to as "many observations per cell"). More specifically, the model can be formulated as follows.
Suppose among ''n'' observations
there are only ''T'' distinct values of the regressors, which can be denoted as
. Let
be the number of observations with
and
the number of such observations with
. We assume that there are indeed "many" observations per each "cell": for each
.
Denote
:
:
Then Berkson's minimum chi-square estimator is a
generalized least squares
In statistics, generalized least squares (GLS) is a method used to estimate the unknown parameters in a Linear regression, linear regression model. It is used when there is a non-zero amount of correlation between the Residual (statistics), resi ...
estimator in a regression of
on
with weights
:
:
It can be shown that this estimator is consistent (as ''n''→∞ and ''T'' fixed), asymptotically normal and efficient. Its advantage is the presence of a closed-form formula for the estimator. However, it is only meaningful to carry out this analysis when individual observations are not available, only their aggregated counts
,
, and
(for example in the analysis of voting behavior).
Albert and Chib Gibbs sampling method
Gibbs sampling of a probit model is possible with the introduction of normally distributed latent variables ''z'', which are observed as 1 if positive and 0 otherwise. This approach was introduced in Albert and Chib (1993),
[Albert, J., & Chib, S. (1993). "Bayesian Analysis of Binary and Polychotomous Response Data." Journal of the American Statistical Association, 88(422), 669-679.] which demonstrated how Gibbs sampling could be applied to binary and polychotomous response models within a Bayesian framework. Under a multivariate normal
prior distribution
A prior probability distribution of an uncertain quantity, simply called the prior, is its assumed probability distribution before some evidence is taken into account. For example, the prior could be the probability distribution representing the ...
over the weights, the model can be described as
:
From this, Albert and Chib (1993)
[Albert, J., & Chib, S. (1993). "Bayesian Analysis of Binary and Polychotomous Response Data." Journal of the American Statistical Association, 88(422), 669-679.] derive the following full conditional distributions in the Gibbs sampling algorithm:
:
The result for
is given in the article on
Bayesian linear regression, although specified with different notation, while the conditional posterior distributions of the latent variables follow a
truncated normal distribution within the given ranges. The notation