Tree diagrams

Tree diagrams can represent conditional probabilities very well.

Tree diagrams can be oriented in many ways, here the focus is on left-to-right tree diagrams with more branches on the right side.

Contents

Construction

Events or decisions that occur earlier are further to the left on tree diagrams, meaning causation is left to right.

If-then statements are insufficient to deduce the order of causation. It all depends on the context and common sense.

  • Are you choosing a bag first, or picking a marble first?
  • Is it raining first, or is the event getting cancelled first?

It is possible that the occurrence order is different from the order that information become known. As an analogy, if you wrap a gift first in tissue paper then wrapping paper and gave it to your friend. Then when your friend unwraps it, they see the wrapping paper before the tissue paper.

Each branch records the conditional probability that the event to the right occurs, given that all events from the root up to that branch, have occurred. P(BA)\text{P}(B\,\vert\, A) is represented by a branch from event AA to BB.

Use the formula

P(BA)+P(BA)=1\text{P}(B\,\vert\, A) + \text{P}(B^\prime\,\vert\, A) = 1

to figure out some of the missing branches. It basically says BB either occurs or does not occur.

In HL there may be three branches to an event, all branches from an event must add up to 11.

Typically there are two levels of branches.

HL students may expect up to three levels of branches.

Interpretation

The chance of a particular sequence of events occurring can be found by multiplying the conditional probabilities along the corresponding path of branches.

Calculations

All formulas from Venn diagrams apply.

In addition, Bayes’ theorem, which comes from the definitions of intersection and conditional probability, says

P(AB)=P(BA)P(A)P(B)\text{P}(A\,\vert\, B) = \frac{\text{P}(B \,\vert\, A)\cdot\text{P}(A)}{\text{P}(B)}

Since our BB is fragmented into two branches, from AA to BB and from AA^\prime to BB, this means

P(B)=P(BA)+P(BA)=P(BA)P(A)+P(BA)P(A)\begin{align*} \text{P}(B) &= \text{P}(B \cap A) + \text{P}(B \cap A^\prime) \\ &= \text{P}(B \,\vert\, A)\cdot\text{P}(A) + \text{P}(B \,\vert\, A^\prime)\cdot\text{P}(A^\prime) \end{align*}

and in turn

P(AB)=P(BA)P(A)P(BA)P(A)+P(BA)P(A)\text{P}(A\,\vert\, B) = \frac{\text{P}(B \,\vert\, A)\cdot\text{P}(A)}{\text{P}(B \,\vert\, A)\cdot\text{P}(A) + \text{P}(B \,\vert\, A^\prime)\cdot\text{P}(A^\prime)}

and this is a way to find the “reverse” conditional probability.

HL students may be asked up to three events, which often means up to three terms in the denominator, as opposed to two in the current form. The formula is not worth memorizing, rather simply apply the conversion between intersection and conditional events in a systematic manner.

But Bayes’ Theorem is not SL?

It just means at SL, instead of asking you P(AB)\text{P}(A\,|\, B) in one step, they’ll first ask you for the intersections and P(B)\text{P}(B) separately. Knowing Bayes’ theorem helps you understand why certain questions are asked.

Applications

diagnosis

First the person either has or do not have the disease. Then a diagnosis (detection) is made, which can be right or wrong.

false positive:     P(tests positiveis negative)\;\;\text{P}(\text{tests positive} \,\vert\, \text{is negative})

false negative:     P(tests negativeis positive)\;\;\text{P}(\text{tests negative} \,\vert\, \text{is positive})