GHK Algorithm
   HOME

TheInfoList



OR:

The GHK algorithm (Geweke, Hajivassiliou and Keane) is an
importance sampling Importance sampling is a Monte Carlo method for evaluating properties of a particular distribution, while only having samples generated from a different distribution than the distribution of interest. Its introduction in statistics is generally at ...
method for simulating choice probabilities in the
multivariate probit model In statistics and econometrics, the multivariate probit model is a generalization of the probit model used to estimate several correlated binary outcomes jointly. For example, if it is believed that the decisions of sending at least one child to ...
. These simulated probabilities can be used to recover parameter estimates from the maximized likelihood equation using any one of the usual well known maximization methods (
Newton's method In numerical analysis, Newton's method, also known as the Newton–Raphson method, named after Isaac Newton and Joseph Raphson, is a root-finding algorithm which produces successively better approximations to the roots (or zeroes) of a real-valu ...
, BFGS, etc.). Train has well documented steps for implementing this algorithm for a multinomial probit model. What follows here will applies to the binary multivariate probit model. Consider the case where one is attempting to evaluate the choice probability of \Pr(\mathbf , \mathbf, \Sigma) where \mathbf = (y_1, ..., y_J), \ (i = 1,...,N) and where we can take j as choices and i as individuals or observations, \mathbf is the mean and \Sigma is the covariance matrix of the model. The probability of observing choice \mathbf is : \begin \Pr(\mathbf, \mathbf, \Sigma) = & \int_\cdots\int_f_N(\mathbf^*_i, \mathbf, \Sigma) dy^*_1\dots dy^*_J \\ \Pr(\mathbf, \mathbf, \Sigma) = & \int \mathbb_ f_N(\mathbf^*_i, \mathbf, \Sigma) d\mathbf^*_i \end Where A = A_1 \times \cdots \times A_J and, : A_j = \begin (-\infty,0] & y_j = 0 \\ (0, \infty) & y_j = 1 \end Unless J is small (less than or equal to 2) there is no closed form solution for the integrals defined above (some work has been done with J = 3 ). The alternative to evaluating these integrals closed form or by quadrature methods is to use simulation. GHK is a simulation method to simulate the probability above using importance sampling methods. Evaluating \Pr(\mathbf, \mathbf, \Sigma) = \int \mathbb_ f_N(\mathbf^*_i, \mathbf, \Sigma) d\mathbf^*_i is simplified by recognizing that the latent data model \mathbf = \mathbf + \epsilon can be rewritten using a Cholesky factorization, \Sigma = CC' . This gives \mathbf = \mathbf + C\eta_i where the \eta_i terms are distributed N(0,\mathbf) . Using this factorization and the fact that the \eta_i are distributed independently one can simulate draws from a truncated multivariate normal distribution using draws from a univariate random normal. For example, if the region of truncation \mathbf has lower and upper limits equal to ,b (including a,b = \pm\infty ) then the task becomes : \begin a < & y^*_1 & < b \\ a < & y^*_2 & < b \\ \vdots & \vdots & \vdots \\ a < & y^*_J & < b \\ \end Note: \mathbf = \mathbf + C\eta_i , substituting: : \begin a < & x_1\beta_1 + c_\eta_1 & < b \\ a < & x_2\beta_2 + c_\eta_1 + c_\eta_2 & < b \\ \vdots & \vdots & \vdots \\ a < & x_J\beta_J + \sum_^J c_\eta_k & < b \\ \end Rearranging above, : \begin \frac & < \eta_1 < & \frac\\ \frac & < \eta_2 < & \frac \\ \vdots & \vdots & \vdots \\ \frac & < \eta_k < & \frac \\ \end Now all one needs to do is iteratively draw from the truncated univariate normal distribution with the given bounds above. This can be done by the inverse CDF method and noting the truncated normal distribution is given by, : u = \frac Where u will be a number between 0 and 1 because the above is a CDF. This suggests to generate random draws from the truncated distribution one has to solve for x giving, : x = \sigma F^(u*(F(\beta) - F(\alpha)) + F(\alpha)) + \mu where \alpha = \frac and \beta = \frac and F is the standard normal CDF. With such draws one can reconstruct the \mathbf by its simplified equation using the Cholesky factorization. These draws will be conditional on the draws coming before and using properties of normals the product of the conditional PDFs will be the joint distribution of the \mathbf , : q(\mathbf, \mathbf, \Sigma) = q(y^*_1, \mathbf, \Sigma)q(y^*_2, y^*_1, \mathbf, \Sigma)\dots q(y^*_J, y^*_1,\dots,y^*_,\mathbf, \Sigma) Where q(\cdot) is the multivariate normal distribution. Because y^*_j conditional on y_k, \ k is restricted to the set A by the setup using the Cholesky factorization then we know that q(\cdot) is a truncated multivariate normal. The distribution function of a
truncated normal In probability and statistics, the truncated normal distribution is the probability distribution derived from that of a normally distributed random variable by bounding the random variable from either below or above (or both). The truncated no ...
is, : \frac Therefore, y^*_j has distribution, : \begin q(\mathbf , \mathbf, \Sigma) & = \frac\times \dots \times \frac\\ & = \frac \end where \phi_j is the standard normal pdf for choice j . Because y^*_ \sim N(\mathbf + \sum_^ c_\eta_k, c_^2) the above standardization makes each term mean 0 variance 1. Let the denominator \prod_^J \Phi_j \Big( \frac \Big) - \Phi\Big( \frac\Big) = \prod_^J l_ and the numerator \prod_^\frac \phi_j\Big( \frac\Big) = f_(\mathbf , \mathbf, \Sigma) where f_N(\cdot) is the multivariate normal PDF. Going back to the original goal, to evaluate the : \begin \Pr(\mathbf, \mathbf, \Sigma) = & \int_f_N(\mathbf^*_i, \mathbf, \Sigma) dy^*_j \\ \end Using importance sampling we can evaluate this integral, : \begin \Pr(\mathbf, \mathbf, \Sigma) = & \int_f_N(\mathbf^*_i, \mathbf, \Sigma) dy^*_j \\ = &\int_ \frac q(\mathbf , \mathbf, \Sigma) dy^*_j \\ = & \int_ \frac q(\mathbf , \mathbf, \Sigma) dy^*_j \\ = & \mathbb_ \Big( \prod_^J l_ \Big) \\ \end This is well approximated by \frac \sum_^S \prod_^J l_ .


References

{{reflist * * * * Regression models