Basic Probability #
Probability models uncertainty: what you don’t know yet, but want to reason about.
Key takeaway: Probability is a number between 0 and 1 that measures how likely an event is. The whole topic is about defining events clearly and applying a few core rules consistently.
Agenda #
You will learn:
- Random experiments and assigning probabilities
- Events and their probabilities
- Basic relationships (complement, union, intersection)
- Mutually exclusive vs independent events
- Basic problem solving patterns
Core building blocks #
Random experiment #
A random experiment is an action whose outcome is not known in advance.
Examples:
- Flip a coin (head/tail)
- Roll a die (1 to 6)
- Wait time at a bus stop
- Number of customers arriving in an hour
- Pulling a card from a deck
A random experiment is not “random because we don’t care”. It is random because the outcome is uncertain before performing it.
Sample space \( S \) #
Sample space: the set of all possible outcomes.
Examples:
- Die roll: \( S=\{1,2,3,4,5,6\} \)
- Quality test (accepted/rejected): \( S=\{a,r\} \)
Event \( A \) #
Event: a subset of the sample space.
It can be:
- Empty event: \( \varnothing \)
- Certain event: \( S \)
- Any meaningful subset of outcomes
Example (die roll):
- Event “even”: \( A=\{2,4,6\} \)
Operations on events (set operations) #
Complement #
Complement: event “not \( A \) ”.
\[ A^c = S \setminus A \] \[ P(A^c)=1-P(A) \]Union #
Union: \( A\cup B \) means “ \( A \) OR \( B \) OR both”.
\[ A \cup B = \{ \omega : \omega \in A \text{ or } \omega \in B \} \]Intersection #
Intersection: \( A\cap B \) means “ \( A \) AND \( B \) ”.
\[ A \cap B = \{ \omega : \omega \in A \text{ and } \omega \in B \} \]Probability as a numerical measure #
Probability scale:
- 0: impossible
- 0.5: equally likely to happen or not happen
- 1: certain
If you ever compute a probability less than 0 or greater than 1, something is wrong.
Three common “definitions” of probability #
1) Classical approach (equally likely outcomes) #
Used when outcomes are equally likely.
\[ P(A)=\frac{\text{number of favourable outcomes}}{\text{total number of possible outcomes}} \]Example: rolling an even number on a fair die.
\[ P(\text{even})=\frac{3}{6}=\frac{1}{2} \]2) Empirical approach (relative frequency) #
Used when probability is estimated from repeated observations.
\[ P(A)\approx\frac{\text{number of times event occurs}}{\text{total number of trials}} \]Intuition: with many trials, empirical probability tends to stabilise.
3) Axiomatic approach (most general) #
Probability is assigned to events in a way that satisfies the axioms below. This is the approach used throughout probability theory.
Axioms of probability #
For any event \( A \) in sample space \( S \) :
1) Non-negativity #
\[ P(A)\ge 0 \]2) Normalisation #
\[ P(S)=1 \]3) Additivity (for mutually exclusive events) #
If \( A\cap B=\varnothing \) , then:
\[ P(A\cup B)=P(A)+P(B) \]These axioms are “rules of the game”. Most probability formulas you use later are consequences of these.
The Addition Rule (general case) #
Even if \( A \) and \( B \) overlap, the union probability is:
\[ P(A\cup B)=P(A)+P(B)-P(A\cap B) \]Why we subtract: the overlap \( A\cap B \) gets counted twice if we only add \( P(A) \) and \( P(B) \) .
Mutually exclusive vs independent (do not confuse) #
Mutually exclusive (disjoint) #
Cannot happen together.
\[ A\cap B=\varnothing \]So:
\[ P(A\cap B)=0 \]Example (one die roll):
- \( A \) : roll a 2
- \( B \)
: roll a 5
They cannot occur together.
Independent #
One event happening does not change the probability of the other.
\[ P(A\cap B)=P(A)\,P(B) \]Example:
- First coin toss is heads
- Second coin toss is heads
These are independent.
Mutually exclusive events are usually NOT independent (unless one event has probability 0). If ( A\cap B=\varnothing )
and both have positive probability, then ( P(A\cap B)=0 )
but ( P(A)P(B)>0 )
. So they cannot be equal.
Visual: deciding the relationship between events #
flowchart LR
A["Two events: A and B"] --> B{"Can A and B happen together?"}
B -->|No| C["Mutually exclusive\nA ∩ B = ∅"]
B -->|Yes| D{"Does knowing A change B?"}
D -->|No| E["Independent\nP(A ∩ B)=P(A)P(B)"]
D -->|Yes| F["Dependent\nUse conditional probability next"]
Worked patterns you should recognise #
Pattern 1: Valid probability assignment #
If outcomes \( A,B,C,\dots \) are mutually exclusive and exhaustive, then a valid assignment must satisfy:
- Each probability is between 0 and 1
- Total must sum to 1
Checklist:
- Are all ( P(\cdot)\ge 0 )
? 2) Is ( \sum P(\text{outcomes})=1 )
? If both yes, it is permissible.
Pattern 2: “At least one” trick #
“At least one of X” is often easiest using complement.
Example idea:
“At least one desktop”
= 1 − P(no desktop)
= 1 − P(both are laptops)
Whenever you see: “At least one…” try the complement first.
Mini-check (self-test) #
- If \( P(A)=0.4 \) and \( P(B)=0.3 \) and \( A,B \) are independent, find \( P(A\cap B) \) .
- If \( A,B \) are mutually exclusive, what is \( P(A\cap B) \) ?
- If \( P(A)=0.7 \) , what is \( P(A^c) \) ?
- If \( P(A)=0.6 \) , \( P(B)=0.5 \) , and \( P(A\cap B)=0.2 \) , find \( P(A\cup B) \) .
Answers: 1) ( 0.4\times 0.3=0.12 )
\( 0 \)
\( 1-0.7=0.3 \)
\( 0.6+0.5-0.2=0.9 \)
What’s next #
Next page:
Conditional Probability & Bayes’ Theorem
This is where “given what I already know…” becomes mathematics.