In
statistics
Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
, Poisson regression is a
generalized linear model form of
regression analysis used to model
count data and
contingency tables. Poisson regression assumes the response variable ''Y'' has a
Poisson distribution, and assumes the
logarithm
In mathematics, the logarithm of a number is the exponent by which another fixed value, the base, must be raised to produce that number. For example, the logarithm of to base is , because is to the rd power: . More generally, if , the ...
of its
expected value
In probability theory, the expected value (also called expectation, expectancy, expectation operator, mathematical expectation, mean, expectation value, or first Moment (mathematics), moment) is a generalization of the weighted average. Informa ...
can be modeled by a linear combination of unknown
parameters. A Poisson regression model is sometimes known as a
log-linear model, especially when used to model contingency tables.
Negative binomial regression is a popular generalization of Poisson regression because it loosens the highly restrictive assumption that the variance is equal to the mean made by the Poisson model. The traditional negative binomial regression model is based on the Poisson-gamma mixture distribution. This model is popular because it models the Poisson heterogeneity with a gamma distribution.
Poisson regression models are
generalized linear models with the logarithm as the (canonical)
link function, and the
Poisson distribution function as the assumed probability distribution of the response.
Regression models
If
is a vector of
independent variables, then the model takes the form
:
where
and
. Sometimes this is written more compactly as
:
where
is now an (''n'' + 1)-dimensional vector consisting of ''n'' independent variables concatenated to the number one. Here
is simply
concatenated to
.
Thus, when given a Poisson regression model
and an input vector
, the predicted mean of the associated Poisson distribution is given by
:
If
are
independent observations with corresponding values
of the predictor variables, then
can be estimated by
maximum likelihood. The maximum-likelihood estimates lack a
closed-form expression and must be found by numerical methods. The probability surface for maximum-likelihood Poisson regression is always concave, making Newton–Raphson or other gradient-based methods appropriate estimation techniques.
Interpretation of coefficients
Suppose we have a model with a single predictor, that is,
:
:
Suppose we compute the predicted values at point
and
:
:
:
By subtracting the first from the second:
:
Suppose now that
. We obtain:
:
So the coefficient of the model is to be interpreted as the increase in the logarithm of the count of the outcome variable when the independent variable increases by 1.
By applying the rules of logarithms:
:
:
:
That is, when the independent variable increases by 1, the outcome variable is multiplied by the exponentiated coefficient.
The exponentiated coefficient is also called the ''incidence ratio''.
Average partial effect
Often, the object of interest is the average partial effect or average marginal effect
, which is interpreted as the change in the outcome
for a one unit change in the independent variable
. The average partial effect in the Poisson model for a continuous
can be shown to be:
:
This can be estimated using the coefficient estimates from the Poisson model
with the observed values of
.
Maximum likelihood-based parameter estimation
Given a set of parameters ''θ'' and an input vector ''x'', the mean of the predicted
Poisson distribution, as stated above, is given by
:
and thus, the Poisson distribution's
probability mass function is given by
:
Now suppose we are given a data set consisting of ''m'' vectors
, along with a set of ''m'' values
. Then, for a given set of parameters ''θ'', the probability of attaining this particular set of data is given by
:
By the method of
maximum likelihood, we wish to find the set of parameters ''θ'' that makes this probability as large as possible. To do this, the equation is first rewritten as a
likelihood function
A likelihood function (often simply called the likelihood) measures how well a statistical model explains observed data by calculating the probability of seeing that data under different parameter values of the model. It is constructed from the ...
in terms of ''θ'':
:
Note that the expression on the
right hand side has not actually changed. A formula in this form is typically difficult to work with; instead, one uses the ''log-likelihood'':
:
Notice that the parameters ''θ'' only appear in the first two terms of each term in the summation. Therefore, given that we are only interested in finding the best value for ''θ'' we may drop the ''y''
''i''! and simply write
:
To find a maximum, we need to solve an equation
which has no closed-form solution. However, the negative log-likelihood,
, is a convex function, and so standard
convex optimization techniques such as
gradient descent can be applied to find the optimal value of ''θ''.
Poisson regression in practice
Poisson regression may be appropriate when the dependent variable is a count, for instance of
events such as the arrival of a telephone call at a call centre. The events must be independent in the sense that the arrival of one call will not make another more or less likely, but the probability per unit time of events is understood to be related to covariates such as time of day.
"Exposure" and offset
Poisson regression may also be appropriate for rate data, where the rate is a count of events divided by some measure of that unit's ''exposure'' (a particular unit of observation). For example, biologists may count the number of tree species in a forest: events would be tree observations, exposure would be unit area, and rate would be the number of species per unit area. Demographers may model death rates in geographic areas as the count of deaths divided by person−years. More generally, event rates can be calculated as events per unit time, which allows the observation window to vary for each unit. In these examples, exposure is respectively unit area, person−years and unit time. In Poisson regression this is handled as an offset. If the rate is count/exposure, multiplying both sides of the equation by exposure moves it to the right side of the equation. When both sides of the equation are then logged, the final model contains log(exposure) as a term that is added to the regression coefficients. This logged variable, log(exposure), is called the offset variable and enters on the right-hand side of the equation with a parameter estimate (for log(exposure)) constrained to 1.
:
which implies
:
Offset in the case of a
GLM in
R can be achieved using the
offset()
function:
glm(y ~ offset(log(exposure)) + x, family=poisson(link=log) )
Overdispersion and zero inflation
A characteristic of the
Poisson distribution is that its mean is equal to its variance. In certain circumstances, it will be found that the observed
variance
In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
is greater than the mean; this is known as
overdispersion and indicates that the model is not appropriate. A common reason is the omission of relevant explanatory variables, or dependent observations. Under some circumstances, the problem of overdispersion can be solved by using
quasi-likelihood estimation or a
negative binomial distribution instead.
Ver Hoef and Boveng described the difference between quasi-Poisson (also called overdispersion with quasi-likelihood) and negative binomial (equivalent to gamma-Poisson) as follows: If ''E''(''Y'') = ''μ'', the quasi-Poisson model assumes var(''Y'') = ''θμ'' while the gamma-Poisson assumes var(''Y'') = ''μ''(1 + ''κμ''), where ''θ'' is the quasi-Poisson overdispersion parameter, and ''κ'' is the shape parameter of the
negative binomial distribution. For both models, parameters are estimated using
iteratively reweighted least squares. For quasi-Poisson, the weights are ''μ''/''θ''. For negative binomial, the weights are ''μ''/(1 + ''κμ''). With large ''μ'' and substantial extra-Poisson variation, the negative binomial weights are capped at 1/''κ''. Ver Hoef and Boveng discussed an example where they selected between the two by plotting mean squared residuals vs. the mean.
Another common problem with Poisson regression is excess zeros: if there are two processes at work, one determining whether there are zero events or any events, and a Poisson process determining how many events there are, there will be more zeros than a Poisson regression would predict. An example would be the distribution of cigarettes smoked in an hour by members of a group where some individuals are non-smokers.
Other
generalized linear models such as the
negative binomial model or
zero-inflated model may function better in these cases.
On the contrary, underdispersion may pose an issue for parameter estimation.
Use in survival analysis
Poisson regression creates proportional hazards models, one class of
survival analysis: see
proportional hazards models for descriptions of Cox models.
Extensions
Regularized Poisson regression
When estimating the parameters for Poisson regression, one typically tries to find values for ''θ'' that maximize the likelihood of an expression of the form
:
where ''m'' is the number of examples in the data set, and
is the
probability mass function of the
Poisson distribution with the mean set to
. Regularization can be added to this optimization problem by instead maximizing
:
for some positive constant
. This technique, similar to
ridge regression, can reduce
overfitting.
See also
*
Zero-inflated model
*
Poisson distribution
*
Fixed-effect Poisson model
*
*
References
Further reading
*
*
*
*
*
*
*
{{least squares and regression analysis
Generalized linear models
Categorical regression models
Poisson distribution
Mathematical and quantitative methods (economics)
Articles with example R code