HOME

TheInfoList



OR:

Risk score (or risk scoring) is the name given to a general practice in applied statistics, bio-statistics,
econometrics Econometrics is the application of statistical methods to economic data in order to give empirical content to economic relationships. M. Hashem Pesaran (1987). "Econometrics," '' The New Palgrave: A Dictionary of Economics'', v. 2, p. 8 p. 8 ...
and other related disciplines, of creating an easily calculated number (the score) that reflects the level of
risk In simple terms, risk is the possibility of something bad happening. Risk involves uncertainty about the effects/implications of an activity with respect to something that humans value (such as health, well-being, wealth, property or the environm ...
in the presence of some
risk factor In epidemiology, a risk factor or determinant is a variable associated with an increased risk of disease or infection. Due to a lack of harmonization across disciplines, determinant, in its more widely accepted scientific meaning, is often ...
s (e.g. risk of mortality or disease in the presence of symptoms or genetic profile, risk financial loss considering credit and financial history, etc.). Risk scores are designed to be: * Simple to calculate: In many cases all you need to calculate a score is a pen and a piece of paper (although
some Some may refer to: *''some'', an English word used as a determiner and pronoun; see use of ''some'' *The term associated with the existential quantifier *"Some", a song by Built to Spill from their 1994 album ''There's Nothing Wrong with Love'' *S ...
scores use rely on more sophisticated or less transparent calculations that require a computer program). * Easily interpreted: The result of the calculation is a single number, and higher score usually means higher risk. Furthermore, many scoring methods enforce some form of
monotonicity In mathematics, a monotonic function (or monotone function) is a function between ordered sets that preserves or reverses the given order. This concept first arose in calculus, and was later generalized to the more abstract setting of orde ...
along the measured risk factors to allow a straight forward interpretation of the score (e.g. risk of mortality only increases with age, risk of payment default only increase with the amount of total debt the customer has, etc.). * Actionable: Scores are designed around a set of possible actions that should be taken as a result of the calculated score. Effective score-based policies can be designed and executed by setting thresholds on the value of the score and associating them with escalating actions.


Formal definition

A typical scoring method is composed of 3 components: # A set of consistent rules (or weights) that assign a numerical value ("points") to each risk factor that reflect our estimation of underlying risk. # A formula (typically a simple sum of all accumulated points) that calculates the score. # A set of thresholds that helps to translate the calculated score into a level of risk, or an equivalent formula or set of rules to translate the calculated score back into probabilities (leaving the nominal evaluation of severity to the practitioner). Items 1 & 2 can be achieved by using some form of regression, that will provide both the risk estimation and the formula to calculate the score. Item 3 requires setting an arbitrary set of thresholds and will usually involve expert opinion.


Estimating risk with GLM

Risk score are designed to represent an underlying probability of an adverse event denoted \lbrace Y = 1 \rbrace given a vector of P explaining variables \mathbf containing measurements of the relevant risk factors. In order to establish the connection between the risk factors and the probability we estimate a set of weights \beta is estimated using a generalized linear model: :\begin \operatorname(\mathbf , \mathbf) = \mathbf(\mathbf = 1 , \mathbf) = g^(\mathbf \beta) \end Where g^: \mathbb \rightarrow ,1/math> is a real-valued, monotonically increasing function that maps the values of the linear predictor \mathbf \beta to the interval ,1. GLM methods typically uses the
logit In statistics, the logit ( ) function is the quantile function associated with the standard logistic distribution. It has many uses in data analysis and machine learning, especially in data transformations. Mathematically, the logit is the ...
or
probit In probability theory and statistics, the probit function is the quantile function associated with the standard normal distribution. It has applications in data analysis and machine learning, in particular exploratory statistical graphics and s ...
as the
link function In statistics, a generalized linear model (GLM) is a flexible generalization of ordinary linear regression. The GLM generalizes linear regression by allowing the linear model to be related to the response variable via a ''link function'' and b ...
.


Estimating risk with other methods

While it's possible to estimate \mathbf(\mathbf = 1 , \mathbf) using other statistical or machine learning methods, the requirements of simplicity and easy interpretation (and monotonicity per risk factor) make most of these methods difficult to use for scoring in this context: * With more sophisticated methods it becomes difficult to attribute simple weights for each risk factor and to provide a simple formula for the calculation of the score. A notable exception are tree-based methods like CART, that can provide a simple set of decision rules and calculations, but cannot ensure the monotonicity of the scale across the different risk factors. * The fact that we are estimating underlying risk across the population, and therefore cannot tag people in advance on an ordinal scale (we can't know in advance if a person belongs to a "high risk" group, we only see observed incidences) classification methods are only relevant if we want to classify people into 2 groups or 2 possible actions.


Constructing the score

When using GLM, the set of estimated weights \beta can be used to assign different values (or "points") to different values of the risk factors in \mathbf (continuous or nominal as indicators). The score can then be expressed as a weighted sum: :\begin \text = \mathbf \beta = \sum_^ \mathbf_ \beta_ \end * Some scoring methods will translate the score into probabilities by using g^ (e.g. SAPS II score that gives an explicit function to calculate mortality from the score) or a look-up table (e.g. ABCD² score or the ISM7 (NI) Scorecard). This practice makes the process of obtaining the score more complicated computationally but has the advantage of translating an arbitrary number to a more familiar scale of 0 to 1. * The columns of \mathbf can represent complex transformations of the risk factors (including multiple
interactions Interaction is action that occurs between two or more objects, with broad use in philosophy and the sciences. It may refer to: Science * Interaction hypothesis, a theory of second language acquisition * Interaction (statistics) * Interactions o ...
) and not just the risk factors themselves. * The values of \beta are sometimes scaled or rounded to allow working with integers instead of very small fractions (making the calculation simpler). While scaling has no impact ability of the score to estimate risk, rounding has the potential of disrupting the "optimality" of the GLM estimation.


Making score-based decisions

Let \mathbf = \lbrace \mathbf_, ... ,\mathbf_ \rbrace denote a set of m \geq 2 "escalating" actions available for the decision maker (e.g. for credit risk decisions: \mathbf_ = "approve automatically", \mathbf_ = "require more documentation and check manually", \mathbf_ = "decline automatically"). In order to define a decision rule, we want to define a map between different values of the score and the possible decisions in \mathbf . Let \tau = \lbrace \tau_1, ... \tau_ \rbrace be a
partition Partition may refer to: Computing Hardware * Disk partitioning, the division of a hard disk drive * Memory partition, a subdivision of a computer's memory, usually for use by a single job Software * Partition (database), the division of a ...
of \mathbb into m consecutive, non-overlapping intervals, such that \tau_1 < \tau_2 < \ldots < \tau_ . The map is defined as follows: :\begin \text \in [\tau_,\tau_) \rightarrow \text \mathbf_ \end * The values of \tau are set based on expert opinion, the type and prevalence of the measured risk, consequences of miss-classification, etc. For example, a risk of 9 out of 10 will usually be considered as "high risk", but a risk of 7 out of 10 can be considered either "high risk" or "medium risk" depending on context. * The definition of the intervals is on right open-ended intervals but can be equivalently defined using left open ended intervals (\tau_,\tau_] . * For scoring methods that are already translated the score into probabilities we either define the partition \tau directly on the interval ,1 or translate the decision criteria into ^(\tau_),g^(\tau_)) , and the monotonicity of g ensures a 1-to-1 translation.


Examples


Biostatistics

* Framingham Risk Score * QRISK * TIMI * Rockall score * ABCD² score * CHA2DS2–VASc score * SAPS II (see more examples on the category page :Medical scoring system)


Financial industry

The primary use of scores in the financial sector is for Credit scorecards, or credit scores: * In many countries (such as the US) credit score are calculated by commercial entities and therefore the exact method is not public knowledge (for example the Bankruptcy risk score,
FICO score A credit score is a number that provides a comparative estimate of an individual's creditworthiness based on an analysis of their credit report. It is an inexpensive and main alternative to other forms of consumer loan underwriting. Lenders, s ...
and others). Credit scores in Australia and UK are often calculated by using
logistic regression In statistics, the logistic model (or logit model) is a statistical model that models the probability of an event taking place by having the log-odds for the event be a linear combination of one or more independent variables. In regression a ...
to estimate
probability of default Probability of default (PD) is a financial term describing the likelihood of a default over a particular time horizon. It provides an estimate of the likelihood that a borrower will be unable to meet its debt obligations. PD is used in a variet ...
, and are therefore a type of risk score. * Other financial industries, such as the
insurance Insurance is a means of protection from financial loss in which, in exchange for a fee, a party agrees to compensate another party in the event of a certain loss, damage, or injury. It is a form of risk management, primarily used to hedge ...
industry also use scoring methods, but the exact implementation remains a
trade secret Trade secrets are a type of intellectual property that includes formulas, practices, processes, designs, instruments, patterns, or compilations of information that have inherent economic value because they are not generally known or readily ...
, except for some rare cases


Social Sciences

* COMPAS score for recidivism, as reverse-engineered by ProPublica using logistic regression and Cox's proportional hazard model.


References

* {{Reflist Econometrics Applied statistics Medical scoring system Credit scoring