Basic Probability

Basic Probability #

Probability models uncertainty: what you don’t know yet, but want to reason about.

Key takeaway: Probability is a number between 0 and 1 that measures how likely an event is. The whole topic is about defining events clearly and applying a few core rules consistently.


Agenda #

You will learn:

  • Random experiments and assigning probabilities
  • Events and their probabilities
  • Basic relationships (complement, union, intersection)
  • Mutually exclusive vs independent events
  • Basic problem solving patterns

Core building blocks #

Random experiment #

A random experiment is an action whose outcome is not known in advance.

Examples:

  • Flip a coin (head/tail)
  • Roll a die (1 to 6)
  • Wait time at a bus stop
  • Number of customers arriving in an hour
  • Pulling a card from a deck

A random experiment is not “random because we don’t care”. It is random because the outcome is uncertain before performing it.


Sample space \( S \) #

Sample space: the set of all possible outcomes.

Examples:

  • Die roll: \( S=\{1,2,3,4,5,6\} \)
  • Quality test (accepted/rejected): \( S=\{a,r\} \)

Event \( A \) #

Event: a subset of the sample space.

It can be:

  • Empty event: \( \varnothing \)
  • Certain event: \( S \)
  • Any meaningful subset of outcomes

Example (die roll):

  • Event “even”: \( A=\{2,4,6\} \)

Operations on events (set operations) #

Complement #

Complement: event “not \( A \) ”.

\[ A^c = S \setminus A \] \[ P(A^c)=1-P(A) \]

Union #

Union: \( A\cup B \) means “ \( A \) OR \( B \) OR both”.

\[ A \cup B = \{ \omega : \omega \in A \text{ or } \omega \in B \} \]

Intersection #

Intersection: \( A\cap B \) means “ \( A \) AND \( B \) ”.

\[ A \cap B = \{ \omega : \omega \in A \text{ and } \omega \in B \} \]

Probability as a numerical measure #

Probability scale:

  • 0: impossible
  • 0.5: equally likely to happen or not happen
  • 1: certain

If you ever compute a probability less than 0 or greater than 1, something is wrong.


Three common “definitions” of probability #

1) Classical approach (equally likely outcomes) #

Used when outcomes are equally likely.

\[ P(A)=\frac{\text{number of favourable outcomes}}{\text{total number of possible outcomes}} \]

Example: rolling an even number on a fair die.

\[ P(\text{even})=\frac{3}{6}=\frac{1}{2} \]

2) Empirical approach (relative frequency) #

Used when probability is estimated from repeated observations.

\[ P(A)\approx\frac{\text{number of times event occurs}}{\text{total number of trials}} \]

Intuition: with many trials, empirical probability tends to stabilise.


3) Axiomatic approach (most general) #

Probability is assigned to events in a way that satisfies the axioms below. This is the approach used throughout probability theory.


Axioms of probability #

For any event \( A \) in sample space \( S \) :

1) Non-negativity #

\[ P(A)\ge 0 \]

2) Normalisation #

\[ P(S)=1 \]

3) Additivity (for mutually exclusive events) #

If \( A\cap B=\varnothing \) , then:

\[ P(A\cup B)=P(A)+P(B) \]

These axioms are “rules of the game”. Most probability formulas you use later are consequences of these.


The Addition Rule (general case) #

Even if \( A \) and \( B \) overlap, the union probability is:

\[ P(A\cup B)=P(A)+P(B)-P(A\cap B) \]

Why we subtract: the overlap \( A\cap B \) gets counted twice if we only add \( P(A) \) and \( P(B) \) .


Mutually exclusive vs independent (do not confuse) #

Mutually exclusive (disjoint) #

Cannot happen together.

\[ A\cap B=\varnothing \]

So:

\[ P(A\cap B)=0 \]

Example (one die roll):

  • \( A \) : roll a 2
  • \( B \) : roll a 5
    They cannot occur together.

Independent #

One event happening does not change the probability of the other.

\[ P(A\cap B)=P(A)\,P(B) \]

Example:

  • First coin toss is heads
  • Second coin toss is heads
    These are independent.

Mutually exclusive events are usually NOT independent (unless one event has probability 0). If ( A\cap B=\varnothing )

and both have positive probability, then ( P(A\cap B)=0 )

but ( P(A)P(B)>0 )

. So they cannot be equal.


Visual: deciding the relationship between events #

flowchart LR
  A["Two events: A and B"] --> B{"Can A and B happen together?"}
  B -->|No| C["Mutually exclusive\nA ∩ B = ∅"]
  B -->|Yes| D{"Does knowing A change B?"}
  D -->|No| E["Independent\nP(A ∩ B)=P(A)P(B)"]
  D -->|Yes| F["Dependent\nUse conditional probability next"]

Worked patterns you should recognise #

Pattern 1: Valid probability assignment #

If outcomes \( A,B,C,\dots \) are mutually exclusive and exhaustive, then a valid assignment must satisfy:

  • Each probability is between 0 and 1
  • Total must sum to 1

Checklist:

  1. Are all ( P(\cdot)\ge 0 )

? 2) Is ( \sum P(\text{outcomes})=1 )

? If both yes, it is permissible.


Pattern 2: “At least one” trick #

“At least one of X” is often easiest using complement.

Example idea: “At least one desktop”
= 1 − P(no desktop)
= 1 − P(both are laptops)

Whenever you see: “At least one…” try the complement first.


Mini-check (self-test) #

  1. If \( P(A)=0.4 \) and \( P(B)=0.3 \) and \( A,B \) are independent, find \( P(A\cap B) \) .
  2. If \( A,B \) are mutually exclusive, what is \( P(A\cap B) \) ?
  3. If \( P(A)=0.7 \) , what is \( P(A^c) \) ?
  4. If \( P(A)=0.6 \) , \( P(B)=0.5 \) , and \( P(A\cap B)=0.2 \) , find \( P(A\cup B) \) .

Answers: 1) ( 0.4\times 0.3=0.12 )

\( 0 \)
\( 1-0.7=0.3 \)
\( 0.6+0.5-0.2=0.9 \)

What’s next #

Next page: Conditional Probability & Bayes’ Theorem
This is where “given what I already know…” becomes mathematics.


Home | Statistics