Lewis's Triviality Result
   HOME

TheInfoList



OR:

In the mathematical theory of probability, David Lewis's triviality result is a theorem about the impossibility of systematically equating the conditional probability P(B\mid A) with the probability of a so-called conditional event, A \rightarrow B.


Conditional probability and conditional events

The statement "The probability that if A, then B, is 20%" means (put intuitively) that event B may be expected to occur in 20% of the outcomes where event A occurs. The standard formal expression of this is P(B\mid A)=0.20, where the
conditional probability In probability theory, conditional probability is a measure of the probability of an Event (probability theory), event occurring, given that another event (by assumption, presumption, assertion or evidence) is already known to have occurred. This ...
P(B\mid A) equals, by definition, P(A \cap B)/P(A). Beginning in the 1960s, several philosophical logicians—most notably Ernest Adams and
Robert Stalnaker Robert Culp Stalnaker (born 1940) is an American philosopher who is Laurance S. Rockefeller Professor Emeritus of Philosophy at the Massachusetts Institute of Technology. He is a Fellow of the American Academy of Arts and Sciences and a Correspond ...
—floated the idea that one might also write P(A \rightarrow B) = 0.20, where A \rightarrow B is the conditional event "If A, then B". That is, given events A and B, one might suppose there is an event, A \rightarrow B, such that P(A \rightarrow B) could be counted on to equal P(B\mid A), so long as P(A) > 0. Part of the appeal of this move would be the possibility of embedding conditional expressions within more complex constructions. One could write, say, P(A \cup (B \rightarrow C)) = 0.75, to express someone's high subjective degree of confidence ("75% sure") that either A, or else if B, then C. Compound constructions containing conditional expressions might also be useful in the programming of automated decision-making systems. How might such a convention be combined with standard probability theory? The most direct extension of the standard theory would be to treat A \rightarrow B as an event like any other, i.e., as a set of outcomes. Adding A \rightarrow B to the familiar Venn- or
Euler diagram An Euler diagram (, ) is a diagrammatic means of representing Set (mathematics), sets and their relationships. They are particularly useful for explaining complex hierarchies and overlapping definitions. They are similar to another set diagrammi ...
of A and B would then result in something like Fig. 1, where s,t,\ldots, z are probabilities allocated to the eight respective regions, such that s + t + \cdots + z = 1. For P(A \rightarrow B) to equal P(B\mid A) requires that t + v + w + y = (s + t)/(s + t + x + y), i.e., that the probability inside the A \rightarrow B region equal the A \cap B region's proportional share of the probability inside the A region. In general the equality will of course not be true, so that making it reliably true requires a new constraint on probability functions: in addition to satisfying Kolmogorov's
probability axioms The standard probability axioms are the foundations of probability theory introduced by Russian mathematician Andrey Kolmogorov in 1933. These axioms remain central and have direct contributions to mathematics, the physical sciences, and real-worl ...
, they must also satisfy a new constraint, namely that P(A \rightarrow B) = P(B\mid A) for any events A and B such that P(A) > 0.


Lewis's result

pointed out a seemingly fatal problem with the above proposal: assuming a nontrivial set of events, the new, restricted class of P-functions will not be closed under
conditioning Conditioning may refer to: Science, computing, and technology * Air conditioning, the removal of heat from indoor air for thermal comfort ** Automobile air conditioning, air conditioning in a vehicle ** Ice storage air conditioning, air conditio ...
, the operation that turns probability function P into new function P_C (\cdot) = P(\cdot\mid C), predicated on event C's occurrence. That is, if P(A \rightarrow B) = P(B\mid A), it will not in general be true that P_C(A \rightarrow B) = P_C(B\mid A) as long as P(C)> 0. This implies that if rationality requires having a well-behaved probability function, then a fully rational person (or computing system) would become irrational simply in virtue of learning that arbitrary event C had occurred.
Bas van Fraassen Bastiaan Cornelis "Bas" van Fraassen (; ; born 5 April 1941) is a Dutch-American philosopher noted for his contributions to philosophy of science, epistemology and formal logic. He is a Distinguished Professor of Philosophy at San Francisco Stat ...
called this result "a veritable bombshell" (
1976 Events January * January 2 – The International Covenant on Economic, Social and Cultural Rights enters into force. * January 5 – The Pol Pot regime proclaims a new constitution for Democratic Kampuchea. * January 18 – Full diplomatic ...
, p. 273). Lewis's proof is as follows. Let a set of events be non-trivial if it contains two possible events, A and B, that are mutually exclusive but do not together exhaust all possibilities, so that P(A) > 0, P(B) > 0, P(A \cap B) = 0, and P(A \cup B) < 1. The existence of two such events implies the existence of the event A \cup B, as well, and, if conditional events are admitted, the event (A \cup B) \rightarrow A. The proof derives a contradiction from the assumption that such a minimally non-trivial set of events exists. # Consider the probability of (A \cup B) \rightarrow A after conditioning, first on A and then instead on its complement A'. #* Conditioning on A gives P_A((A \cup B) \rightarrow A) = P(((A \cup B) \rightarrow A) \cap A)/P(A). But also, by the new constraint on P-functions, P_A((A \cup B) \rightarrow A) = P_A((A \cup B) \cap A)/P_A(A \cup B) = P((A \cup B) \cap A\mid A)/P(A \cup B\mid A) = 1/1 = 1. Therefore, P(((A \cup B) \rightarrow A) \cap A) = P(A). #* Conditioning on A' gives P_((A \cup B) \rightarrow A) = P(((A \cup B) \rightarrow A) \cap A')/P(A'). But also, P_((A \cup B) \rightarrow A) = P_((A \cup B) \cap A)/P_(A \cup B) = P((A \cup B) \cap A\mid A')/P(A \cup B\mid A') = 0/P(A \cup B\mid A') = 0. (The mutual exclusivity of B and A ensures that P(A \cup B\mid A') \neq 0.) Therefore, P(((A \cup B) \rightarrow A) \cap A') = 0. # Instantiate the identity P(X \cap Y) + P(X \cap Y') = P(X) as P(((A \cup B) \rightarrow A) \cap A) + P(((A \cup B) \rightarrow A) \cap A') = P((A \cup B) \rightarrow A). By the results from Step 1, the left side reduces to P(A), while the right side, by the new constraint on P-functions, equals P((A \cup B) \cap A)/P(A \cup B) = P(A)/P(A \cup B). Therefore, P(A) = P(A)/P(A \cup B), which means that P(A \cup B) = 1, which contradicts the stipulation that P(A \cup B) < 1. This completes the proof.


Graphical version

A graphical version of the proof starts with Fig. 2, where the A and B from Fig. 1 are now disjoint and A \rightarrow B has been replaced by (A \cup B) \rightarrow A.A proof starting with overlapping A and B, as in Fig. 1, would use mutually exclusive events A \cap B' and A' \cap B in place of A and B. By the assumption that A and B are possible, x+y>0 and u+v>0. By the assumption that together A and B do not together exhaust all possibilities, u + v + x + y < 1. And by the new constraint on probability functions, P((A \cup B) \rightarrow A) = P(A\mid A \cup B) = P(A \cap (A \cup B))/P(A \cup B) = P(A)/P(A \cup B), which means that :(1) y + v + w =\frac, Conditioning on an event involves zeroing out the probabilities outside the event's region and increasing the probabilities inside the region by a common scale factor. Here, conditioning on A will zero out u, v and w and scale up x and y, to x_A and y_A, respectively, and so :(2) y_A + 0 + 0 = \frac, which simplifies to y_A = 1. Conditioning instead on A' will zero out x and y and scale up u, v and w, and so :(3) 0 + v_ + w_ = \frac, which simplifies to v_ + w_ = 0. From (2), it follows that x_A = 0, and since x_A is the scaled-up value of x, it must also be that x = 0. Similarly, from (3), v = w = 0. But then (1) reduces to y = y/(y + u), which implies that y + u = 1, which contradicts the stipulation that u + v + x + y < 1.


Later developments

In a follow-up article, noted that the triviality proof can proceed by conditioning not on A and A' but instead, by turns, on each of a finite set of mutually exclusive and jointly exhaustive events A, C, D, E, \ldots\,. He also gave a variant of the proof that involved not total conditioning, in which the probability of either A or A' is set to 1, but partial conditioning (i.e., Jeffrey conditioning), by which probability is incrementally shifted from A' to A. Separately, pointed out that even without conditioning, if the number of outcomes is large but finite, then in general P(B\mid A) = P(A \cap B)/P(A), being a ratio of two outputs of the P-function, will take on more values than any single output of the function can. So, for instance, if in Fig. 1 s, t, \ldots are all multiples of 0.01 (as would be the case if there were exactly 100 equiprobable outcomes), then P(A \rightarrow B) must be a multiple of 0.01, as well, but P(A \cap B)/P(A) need not be. That being the case, P(A \rightarrow B) cannot reliably be made to equal P(B \mid A). also argued that the condition P(A \rightarrow B) = P(B\mid A) caused acceptable P-functions to be implausibly sparse and isolated from one another. One way to put the point: standardly, any weighted average of two probability function is itself a probability function, so that between any two P-functions there will be a continuum of weighted-average P-functions along which one of the original P-functions gradually transforms into the other. But these continua disappear if the added P(A \rightarrow B) = P(B\mid A) condition is imposed. Now an average of two acceptable P-functions will in general not be an acceptable P-function.


Possible rejoinders

Assuming that P(A \rightarrow B) = P(B\mid A) holds for a minimally nontrivial set of events and for any P-function leads to a contradiction. Thus P(A \rightarrow B) = P(B\mid A) can hold for any P-function only for trivial sets of events—that is the triviality result. However, the proof relies on background assumptions that may be challenged. It may be proposed, for instance, that the referent event of an expression like “A \rightarrow B” is not fixed for a given A and B, but instead changes as the probability function changes. Or it may be proposed that conditioning on C should follow a rule other than P_C(\cdot) = P(\cdot\mid C). But the most common response, among proponents of the P(A \rightarrow B) = P(B\mid A) condition, has been to explore ways to model conditional events as something other than subsets of a universe set of outcomes. Even before Lewis published his result, had modeled conditional events as ordered ''pairs'' of sets of outcomes. With that approach and others in the same spirit, conditional events and their associated combination and complementation operations do not constitute the usual
algebra of sets In mathematics, the algebra of sets, not to be confused with the mathematical structure of ''an'' algebra of sets, defines the properties and laws of sets, the set-theoretic operations of union, intersection, and complementation and the re ...
of standard probability theory, but rather a more exotic type of structure, known as a conditional event algebra.


Notes


References

* * * * * * * {{refend Probability articles needing expert attention Logic Conditional probability