# Statistical paradoxes for sharping your Bayesian skills

Sometimes despite apparently valid intuitions, it is easy to convey to wrong conclusions when the solution of the problem runs contrary to our expectations. This is the case of paradoxes that are known to have valid arguments, and without a deep insight, their solutions sound counter-intuitive and illogical.

I collected a couple of problems accountable on probability theory on which most people’s impulsive intuition, including mine, falls short. I want to demonstrate you that those puzzles don’t require great talent if we apply some basic rules. Bayesian conditional probability is not difficult, it is just counting. But to apply it you need to sharpen your interpretation skills, and this is hard.

## Bertrand’s box paradox

A classic probability puzzle is the Bertrand’s box paradox and I prefer to narrate this slightly different version rather than the original one. Suppose I have 3 cards, the first card is white on both sides. The second card is white on one side and black on the other. The third has both sides black. Now I put one card on the table, and the visible side of this card is white. What is the probability that the other side is also white?

The tree cards have:

- White in both sides \((ww)\)
- White on one side and black on the other \((wb)\)
- Black on both sides \((bb)\)

The card on the table is white on the visible side.

This is a hard problem if we don’t solve those problems by applying conditional probability. The Bayesian rule tells us to leverage what we already know for refine your knowledge about we don’t know yet:

\[P(wish\; to\; know \mid known)\]In the case of the cards, we know how the different sides are distributed and we know the color of the proposed card. We wish to know the probability that the other side has the same color of the visible one. Bayes formulates as:

\[P(ww \mid w) = \frac{P(w \mid ww) \; P(ww)}{P(w)}\]This is the well known Bayes Rule applied into this problem. I introduce also the following formula - law of total probability - that helps us to find the probability of having \(w\) when we put the card on the table:

\[\begin{align} cards &= \{ww,\;wb,\;bb\} \\ P(w) &= \sum_i{P(w \mid cards_i)P(cards_i)} \end{align}\]First, let’s solve \(P(w)\):

\[\begin{align} P(w) & = P(w \mid ww)P(ww) + P(w \mid wb)P(wb) + P(w \mid bb)P(bb) \\ &= 1*\frac{1}{3} + \frac{1}{2} \frac{1}{3} + 0\frac{1}{3} \\ &= \frac{1}{2} \end{align}\]The chances to pick up a card is \(\frac{1}{3}\) for all cards:

\(P(ww)\) = \(P(wb)\) = \(P(bb)\) = \(\frac{1}{3}\)

Self explanatory:

\(P(w \mid ww)\) = 1

Now we can solve the equation:

\(\begin{align}
P(ww \mid w) &= \frac{P(w \mid ww) \; P(ww)}{P(w)} \\
&= \frac{1\frac{1}{3}}{\frac{1}{2}} \\
&= \frac{2}{3}
\end{align}\)

For having the right intuition about that, the trick is to count sides not the cards themselves. There are two cards with at least one white side and only one with both white. But it is the sides and not the cards that matter.

## The Monty Hall Problem

Suppose you are on a game show, and you are given the choice of three doors. Behind one door is a car, behind the others, goats. You peek a door, say \(\#1\), and the host, who knows what’s behind the doors, opens another door, say \(\#3\), which has a goat. Then he says you: *“Would you switch the door \(\#1\) with \(\#2\)?”*.

In your opinion, is it better to hold \(\#1\) or switch it to \(\#2\)?

The question is based on a popular televised game show called Let’s Make a Deal conducted by the host Monty Hall. In September 1990, a popular columnist in the US gave an answer to this tricky problem an it created great furor. The question was posed by a reader that asked what to choose. Hold on \(\#1\) or switch to \(\#2\)? The answer, defended by the “world’s smartest woman” - as billed by the magazine - argued that the contestant should switch doors.

Over the following months, she received thousands of letters even from PhDs in mathematics and statistics. Most of them fervently disagreed with her and one of them commented: *“May I suggest that you obtain and refer to a standard textbook on probability before you try to answer a question of this type again?”*.

Critics stated that switching or not, have the same probability, so the decision to change door should be totally random. Let’s understand the mechanics of this problem under the lens of Bayes rule.

When the contestant choose the door, the odds to find the car behind that door is uncontestedly \(\frac{1}{3}\):

\(P(c1)\)=\(P(c2)\)=\(P(c3)\)=\(\frac{1}{3}\)

If the conductor open the #3, what are the chances to have the prize behind door #1? In this game Bayes should be interpreted as:

\[P(c1 \mid \#3) = \frac{P(\#3 \mid c1) P(c1)}{P(\#3)}\]But what is \(P(\#3)\)? \

This is the point where misconceptions act cruelly on our judgment. According to the columnist solutions, might be that the producers of the game read the contestant’s mind, otherwise how else could they put the car so that it is more likely to find it behind the door has been not choose? \

The conductor is everything but fool on selecting \(\#3\) and we should take into account the rules of the game that tell us something that has not been observed. The conductor must choose the door that doesn’t cover the car and probabilities come conditional on this information:

and then:

\[P(c1 \mid \#3) = \frac{1}{3}\]It turns out that switching door, **double the chances** to win the car (\(\frac{1}{3}\) vs \(\frac{2}{3}\)).

Indeed, it is an eccentric case and it doesn’t involve telepathy nor any form of collusion with the game producers’ nor the car is running away from us. There is no causation between our first choice and the car, it sounds very unnatural that we should change door! The only one form of causation in this game is ascribable to the conductor, and his choice is determined by a couple of reasons: our initial door and the location of the car. Bayesian conditioning helps us to unfold it.

## Comments