<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Probability on Arshad Siddiqui</title><link>https://arshadhs.github.io/tags/probability/</link><description>Recent content in Probability on Arshad Siddiqui</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Wed, 18 Mar 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://arshadhs.github.io/tags/probability/index.xml" rel="self" type="application/rss+xml"/><item><title>Formula Sheet</title><link>https://arshadhs.github.io/docs/ai/statistics/00_formulas/</link><pubDate>Thu, 12 Mar 2026 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/statistics/00_formulas/</guid><description>&lt;h1 id="formula-sheet">
 Formula Sheet
 
 &lt;a class="anchor" href="#formula-sheet">#&lt;/a>
 
&lt;/h1>
&lt;p>This page is a quick reference of &lt;strong>definitions + formulas&lt;/strong>, grouped by the modules.&lt;/p>
&lt;hr>
&lt;h2 id="notation">
 Notation
 
 &lt;a class="anchor" href="#notation">#&lt;/a>
 
&lt;/h2>
&lt;ul>
&lt;li>Sample size: 
&lt;link rel="stylesheet" href="https://arshadhs.github.io/katex/katex.min.css" />
&lt;script defer src="https://arshadhs.github.io/katex/katex.min.js">&lt;/script>

 &lt;script defer src="https://arshadhs.github.io/katex/auto-render.min.js" onload="renderMathInElement(document.body, {
 &amp;#34;delimiters&amp;#34;: [
 {&amp;#34;left&amp;#34;: &amp;#34;$$&amp;#34;, &amp;#34;right&amp;#34;: &amp;#34;$$&amp;#34;, &amp;#34;display&amp;#34;: true},
 {&amp;#34;left&amp;#34;: &amp;#34;$&amp;#34;, &amp;#34;right&amp;#34;: &amp;#34;$&amp;#34;, &amp;#34;display&amp;#34;: false},
 {&amp;#34;left&amp;#34;: &amp;#34;\\(&amp;#34;, &amp;#34;right&amp;#34;: &amp;#34;\\)&amp;#34;, &amp;#34;display&amp;#34;: false},
 {&amp;#34;left&amp;#34;: &amp;#34;\\[&amp;#34;, &amp;#34;right&amp;#34;: &amp;#34;\\]&amp;#34;, &amp;#34;display&amp;#34;: true}
 ]
});">&lt;/script>

&lt;span>
 \( n \)
 &lt;/span>

 (sample), 
&lt;span>
 \( N \)
 &lt;/span>

 (population)&lt;/li>
&lt;li>Sample mean: 
&lt;span>
 \( \bar{x} \)
 &lt;/span>

, population mean: 
&lt;span>
 \( \mu \)
 &lt;/span>

&lt;/li>
&lt;li>Sample variance: 
&lt;span>
 \( s^2 \)
 &lt;/span>

, population variance: 
&lt;span>
 \( \sigma^2 \)
 &lt;/span>

&lt;/li>
&lt;li>Sample SD: 
&lt;span>
 \( s \)
 &lt;/span>

, population SD: 
&lt;span>
 \( \sigma \)
 &lt;/span>

&lt;/li>
&lt;li>Complement: 
&lt;span>
 \( A^c \)
 &lt;/span>

&lt;/li>
&lt;li>Intersection (“and”): 
&lt;span>
 \( A\cap B \)
 &lt;/span>

, union (“or”): 
&lt;span>
 \( A\cup B \)
 &lt;/span>

&lt;/li>
&lt;li>Conditional probability: 
&lt;span>
 \( P(A\mid B) \)
 &lt;/span>

&lt;/li>
&lt;/ul>
&lt;hr>
&lt;h1 id="1-basic-probability--statistics">
 1. Basic Probability &amp;amp; Statistics
 
 &lt;a class="anchor" href="#1-basic-probability--statistics">#&lt;/a>
 
&lt;/h1>
&lt;h2 id="11-measures-of-central-tendency">
 1.1 Measures of Central Tendency
 
 &lt;a class="anchor" href="#11-measures-of-central-tendency">#&lt;/a>
 
&lt;/h2>
&lt;h3 id="arithmetic-mean">
 Arithmetic mean
 
 &lt;a class="anchor" href="#arithmetic-mean">#&lt;/a>
 
&lt;/h3>
&lt;p>Sample mean (ungrouped):&lt;/p></description></item><item><title>Stats Formula Sheet</title><link>https://arshadhs.github.io/docs/ai/statistics/ism-formula-sheet/</link><pubDate>Wed, 25 Feb 2026 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/statistics/ism-formula-sheet/</guid><description>&lt;h1 id="stats-formula-sheet">
 Stats Formula Sheet
 
 &lt;a class="anchor" href="#stats-formula-sheet">#&lt;/a>
 
&lt;/h1>
&lt;p>Keep this page as a quick reference of &lt;strong>definitions + formulas&lt;/strong>.&lt;/p>
&lt;hr>
&lt;h2 id="notation">
 Notation
 
 &lt;a class="anchor" href="#notation">#&lt;/a>
 
&lt;/h2>
&lt;ul>
&lt;li>Sample size: 
&lt;span>
 \( n \)
 &lt;/span>

 (sample), 
&lt;span>
 \( N \)
 &lt;/span>

 (population)&lt;/li>
&lt;li>Mean: 
&lt;span>
 \( \bar{x} \)
 &lt;/span>

 (sample), 
&lt;span>
 \( \mu \)
 &lt;/span>

 (population)&lt;/li>
&lt;li>Variance: 
&lt;span>
 \( s^2 \)
 &lt;/span>

 (sample), 
&lt;span>
 \( \sigma^2 \)
 &lt;/span>

 (population)&lt;/li>
&lt;li>Standard deviation: 
&lt;span>
 \( s \)
 &lt;/span>

 (sample), 
&lt;span>
 \( \sigma \)
 &lt;/span>

 (population)&lt;/li>
&lt;/ul>
&lt;hr>
&lt;h2 id="module-1-basic-statistics">
 Module 1: Basic Statistics
 
 &lt;a class="anchor" href="#module-1-basic-statistics">#&lt;/a>
 
&lt;/h2>
&lt;h3 id="measures-of-central-tendency">
 Measures of Central Tendency
 
 &lt;a class="anchor" href="#measures-of-central-tendency">#&lt;/a>
 
&lt;/h3>
&lt;p>&lt;strong>Sample mean (ungrouped):&lt;/strong>&lt;/p></description></item><item><title>Conditional Probability &amp; Bayes’ Theorem</title><link>https://arshadhs.github.io/docs/ai/statistics/conditional-probability/</link><pubDate>Thu, 12 Mar 2026 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/statistics/conditional-probability/</guid><description>&lt;h1 id="conditional-probability--bayes-theorem">
 Conditional Probability &amp;amp; Bayes’ Theorem
 
 &lt;a class="anchor" href="#conditional-probability--bayes-theorem">#&lt;/a>
 
&lt;/h1>
&lt;p>Probability often changes when we &lt;strong>learn new information&lt;/strong>.&lt;/p>
&lt;p>Conditional probability and Bayes’ theorem give a structured way to &lt;strong>update beliefs&lt;/strong> using evidence.&lt;/p>
&lt;blockquote class="book-hint info">
&lt;p>Conditional probability updates probabilities after observing an event.&lt;/p>
&lt;p>Bayes’ theorem lets you estimate a hidden cause from observed evidence.&lt;/p>
&lt;p>Naïve Bayes turns Bayes’ theorem into a practical classifier by assuming conditional independence of features given the class.&lt;/p>
&lt;/blockquote>
&lt;hr>


&lt;script src="https://arshadhs.github.io/mermaid.min.js">&lt;/script>

 &lt;script>mermaid.initialize({
 "flowchart": {
 "useMaxWidth":true
 },
 "theme": "default"
}
)&lt;/script>




&lt;pre class="mermaid">
flowchart TD

A[Conditional&amp;lt;br/&amp;gt;probability] --&amp;gt;|foundation| B[Bayes&amp;lt;br/&amp;gt;theorem]
D[Independent&amp;lt;br/&amp;gt;events] --&amp;gt;|implies| C[Independence]
C --&amp;gt;|simplifies| A

E[Prior] --&amp;gt;|with likelihood| B
F[Likelihood] --&amp;gt;|updates| H[Posterior]
G[Evidence] --&amp;gt;|normalises| B
B --&amp;gt;|yields| H

I[Naïve&amp;lt;br/&amp;gt;Bayes] --&amp;gt;|uses| B
J[Naïve&amp;lt;br/&amp;gt;assumption] --&amp;gt;|assumes| C
K[Features] --&amp;gt;|given class| J
L[Class] --&amp;gt;|conditions| J
I --&amp;gt;|predicts| M[Classification]
M --&amp;gt;|selects| L

style A fill:#90CAF9,stroke:#1E88E5,color:#000
style B fill:#90CAF9,stroke:#1E88E5,color:#000
style C fill:#90CAF9,stroke:#1E88E5,color:#000

style D fill:#CE93D8,stroke:#8E24AA,color:#000
style E fill:#CE93D8,stroke:#8E24AA,color:#000
style F fill:#CE93D8,stroke:#8E24AA,color:#000
style G fill:#CE93D8,stroke:#8E24AA,color:#000
style J fill:#CE93D8,stroke:#8E24AA,color:#000
style K fill:#CE93D8,stroke:#8E24AA,color:#000
style L fill:#CE93D8,stroke:#8E24AA,color:#000

style H fill:#C8E6C9,stroke:#2E7D32,color:#000
style I fill:#C8E6C9,stroke:#2E7D32,color:#000
style M fill:#C8E6C9,stroke:#2E7D32,color:#000

&lt;/pre>

&lt;hr>
&lt;h2 id="quick-summary">
 Quick summary
 
 &lt;a class="anchor" href="#quick-summary">#&lt;/a>
 
&lt;/h2>
&lt;ul>
&lt;li>Conditional probability:
updates probability after an event is known.&lt;/li>
&lt;li>Multiplication rule:
computes joint probability from conditional parts.&lt;/li>
&lt;li>Independence:
tested using 
&lt;link rel="stylesheet" href="https://arshadhs.github.io/katex/katex.min.css" />
&lt;script defer src="https://arshadhs.github.io/katex/katex.min.js">&lt;/script>

 &lt;script defer src="https://arshadhs.github.io/katex/auto-render.min.js" onload="renderMathInElement(document.body, {
 &amp;#34;delimiters&amp;#34;: [
 {&amp;#34;left&amp;#34;: &amp;#34;$$&amp;#34;, &amp;#34;right&amp;#34;: &amp;#34;$$&amp;#34;, &amp;#34;display&amp;#34;: true},
 {&amp;#34;left&amp;#34;: &amp;#34;$&amp;#34;, &amp;#34;right&amp;#34;: &amp;#34;$&amp;#34;, &amp;#34;display&amp;#34;: false},
 {&amp;#34;left&amp;#34;: &amp;#34;\\(&amp;#34;, &amp;#34;right&amp;#34;: &amp;#34;\\)&amp;#34;, &amp;#34;display&amp;#34;: false},
 {&amp;#34;left&amp;#34;: &amp;#34;\\[&amp;#34;, &amp;#34;right&amp;#34;: &amp;#34;\\]&amp;#34;, &amp;#34;display&amp;#34;: true}
 ]
});">&lt;/script>

&lt;span>
 \( P(A\cap B)=P(A)P(B) \)
 &lt;/span>

.&lt;/li>
&lt;li>Total probability:
breaks a probability into weighted cases.&lt;/li>
&lt;li>Bayes’ theorem:
reverses conditioning to infer causes from evidence.&lt;/li>
&lt;/ul>
&lt;hr>
&lt;h2 id="whats-next">
 What’s next
 
 &lt;a class="anchor" href="#whats-next">#&lt;/a>
 
&lt;/h2>
&lt;p>Probability Distributions&lt;br>
Move from events to random variables and distributions.&lt;/p></description></item><item><title>Conditional Probability</title><link>https://arshadhs.github.io/docs/ai/statistics/conditional-probability/021_conditional_prob/</link><pubDate>Thu, 12 Mar 2026 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/statistics/conditional-probability/021_conditional_prob/</guid><description>&lt;h1 id="conditional-probability">
 Conditional Probability
 
 &lt;a class="anchor" href="#conditional-probability">#&lt;/a>
 
&lt;/h1>
&lt;p>Conditional probability updates the probability of an event when new information is available.&lt;/p>
&lt;p>It shows up whenever a question says:&lt;/p>
&lt;ul>
&lt;li>“given that…”&lt;/li>
&lt;li>“among those who…”&lt;/li>
&lt;li>“out of the items that…”&lt;/li>
&lt;li>“if it does not fail immediately…”&lt;/li>
&lt;/ul>
&lt;blockquote class="book-hint info">
&lt;p>Key takeaway:
Conditional probability is always:&lt;/p>
&lt;p>joint probability ÷ probability of the condition.&lt;/p>
&lt;p>The condition must not be an impossible event.&lt;/p>
&lt;/blockquote>
&lt;hr>
&lt;h2 id="prior-vs-posterior">
 Prior vs posterior
 
 &lt;a class="anchor" href="#prior-vs-posterior">#&lt;/a>
 
&lt;/h2>
&lt;ul>
&lt;li>
&lt;p>Prior probability:
probability with no condition (before new information)&lt;/p></description></item><item><title>Bayes’ Theorem</title><link>https://arshadhs.github.io/docs/ai/statistics/conditional-probability/022_bayes_theorem/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/statistics/conditional-probability/022_bayes_theorem/</guid><description>&lt;h1 id="bayes-theorem">
 Bayes’ Theorem
 
 &lt;a class="anchor" href="#bayes-theorem">#&lt;/a>
 
&lt;/h1>
&lt;h3 id="21-total-probability-needed-for-bayes">
 2.1 Total probability (needed for Bayes)
 
 &lt;a class="anchor" href="#21-total-probability-needed-for-bayes">#&lt;/a>
 
&lt;/h3>
&lt;p>Often we split the world into cases 
&lt;span>
 \( E_1,E_2,\dots,E_k \)
 &lt;/span>

 that:&lt;/p>
&lt;ul>
&lt;li>are mutually exclusive&lt;/li>
&lt;li>cover the whole sample space&lt;/li>
&lt;/ul>
&lt;p>Then for any event 
&lt;span>
 \( A \)
 &lt;/span>

:&lt;/p>
&lt;span style="color: red;">
 &lt;span>
 \[ 
P(A)=\sum_{i=1}^{k} P(A\mid E_i)\,P(E_i)
 \]
 &lt;/span>
&lt;/span>
&lt;p>Tree intuition:&lt;/p>


&lt;pre class="mermaid">
flowchart TD
 S[Start] --&amp;gt; E1[Case E1]
 S --&amp;gt; E2[Case E2]
 S --&amp;gt; E3[Case E3]
 E1 --&amp;gt; A1[&amp;#34;A happens&amp;#34;]
 E2 --&amp;gt; A2[&amp;#34;A happens&amp;#34;]
 E3 --&amp;gt; A3[&amp;#34;A happens&amp;#34;]
&lt;/pre>

&lt;hr>
&lt;h3 id="22-bayes-theorem-two-event-form">
 2.2 Bayes’ theorem (two-event form)
 
 &lt;a class="anchor" href="#22-bayes-theorem-two-event-form">#&lt;/a>
 
&lt;/h3>
&lt;p>Bayes&amp;rsquo; Theorem is a mathematical formula used to determine the &lt;strong>conditional probability of an event based on prior knowledge and new evidence&lt;/strong>.&lt;/p></description></item><item><title>Naïve Bayes</title><link>https://arshadhs.github.io/docs/ai/statistics/conditional-probability/023_naive_bayes/</link><pubDate>Thu, 12 Mar 2026 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/statistics/conditional-probability/023_naive_bayes/</guid><description>&lt;h1 id="naïve-bayes">
 Naïve Bayes
 
 &lt;a class="anchor" href="#na%c3%afve-bayes">#&lt;/a>
 
&lt;/h1>
&lt;p>Naïve Bayes is a &lt;strong>probabilistic classifier&lt;/strong>.&lt;/p>
&lt;ul>
&lt;li>Supervised Learning Problem&lt;/li>
&lt;li>Binary Classification - final target variable is considered in two classes&lt;/li>
&lt;li>Hypothesis is target which you want to classify&lt;/li>
&lt;li>Total Probability (Prior) of Yes and No is already calculated&lt;/li>
&lt;li>Post / Posterior is when you start studying data&lt;/li>
&lt;li>Based on max probability of hypotheses classify given instance into a class&lt;/li>
&lt;/ul>
&lt;p>It predicts a class label by computing:&lt;/p></description></item><item><title>Probability Distributions</title><link>https://arshadhs.github.io/docs/ai/statistics/probability_distributions/</link><pubDate>Sun, 22 Feb 2026 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/statistics/probability_distributions/</guid><description>&lt;h1 id="probability-distributions">
 Probability Distributions
 
 &lt;a class="anchor" href="#probability-distributions">#&lt;/a>
 
&lt;/h1>
&lt;p>Probability distributions are the bridge between:
real-world randomness and mathematical modelling.&lt;/p>
&lt;p>A random experiment produces outcomes.
A random variable turns those outcomes into numbers.
A probability distribution tells you how likely each number (or range of numbers) is.&lt;/p>
&lt;blockquote class="book-hint info">
&lt;p>Key takeaway:
A distribution is a complete “story” about uncertainty:
what values are possible, how likely they are, and how we summarise them (mean, variance).&lt;/p>
&lt;/blockquote>
&lt;hr>


&lt;script src="https://arshadhs.github.io/mermaid.min.js">&lt;/script>

 &lt;script>mermaid.initialize({
 "flowchart": {
 "useMaxWidth":true
 },
 "theme": "default"
}
)&lt;/script>




&lt;pre class="mermaid">
flowchart TD
	PD[&amp;#34;Probability&amp;lt;br/&amp;gt;distributions&amp;#34;] --&amp;gt; RV[&amp;#34;Random&amp;lt;br/&amp;gt;variables&amp;#34;]
	PD[&amp;#34;Probability&amp;lt;br/&amp;gt;distributions&amp;#34;] --&amp;gt; DS[&amp;#34;Common&amp;lt;br/&amp;gt;distributions&amp;#34;]

	style PD fill:#90CAF9,stroke:#1E88E5,color:#000
	style RV fill:#90CAF9,stroke:#1E88E5,color:#000
	style DS fill:#90CAF9,stroke:#1E88E5,color:#000
&lt;/pre>

&lt;hr>
&lt;h2 id="aiml-connection">
 AI/ML Connection
 
 &lt;a class="anchor" href="#aiml-connection">#&lt;/a>
 
&lt;/h2>
&lt;ul>
&lt;li>Many ML models are probabilistic:
they assume data (or errors) follow a distribution.&lt;/li>
&lt;li>Loss functions often come from distribution assumptions:
squared loss aligns with Gaussian noise.&lt;/li>
&lt;li>Naïve Bayes (from the previous module) becomes practical once you can model:

&lt;link rel="stylesheet" href="https://arshadhs.github.io/katex/katex.min.css" />
&lt;script defer src="https://arshadhs.github.io/katex/katex.min.js">&lt;/script>

 &lt;script defer src="https://arshadhs.github.io/katex/auto-render.min.js" onload="renderMathInElement(document.body, {
 &amp;#34;delimiters&amp;#34;: [
 {&amp;#34;left&amp;#34;: &amp;#34;$$&amp;#34;, &amp;#34;right&amp;#34;: &amp;#34;$$&amp;#34;, &amp;#34;display&amp;#34;: true},
 {&amp;#34;left&amp;#34;: &amp;#34;$&amp;#34;, &amp;#34;right&amp;#34;: &amp;#34;$&amp;#34;, &amp;#34;display&amp;#34;: false},
 {&amp;#34;left&amp;#34;: &amp;#34;\\(&amp;#34;, &amp;#34;right&amp;#34;: &amp;#34;\\)&amp;#34;, &amp;#34;display&amp;#34;: false},
 {&amp;#34;left&amp;#34;: &amp;#34;\\[&amp;#34;, &amp;#34;right&amp;#34;: &amp;#34;\\]&amp;#34;, &amp;#34;display&amp;#34;: true}
 ]
});">&lt;/script>

&lt;span>
 \( P(X\mid Y) \)
 &lt;/span>

 using suitable distributions.&lt;/li>
&lt;/ul>
&lt;blockquote class="book-hint warning">
&lt;p>In practice:
choosing a distribution is a modelling decision.
It affects:
prediction, uncertainty estimates, and what “rare” or “typical” means in your data.&lt;/p></description></item><item><title>Random Variables</title><link>https://arshadhs.github.io/docs/ai/statistics/probability_distributions/random-variables/</link><pubDate>Sun, 22 Feb 2026 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/statistics/probability_distributions/random-variables/</guid><description>&lt;h1 id="random-variables">
 Random Variables
 
 &lt;a class="anchor" href="#random-variables">#&lt;/a>
 
&lt;/h1>
&lt;p>A random variable is a way to attach numbers to outcomes of a random experiment.&lt;/p>
&lt;p>It lets us move from:
“what happened?”
to:
“what number should we analyse?”&lt;/p>
&lt;blockquote class="book-hint info">
&lt;p>Key takeaway:
A random variable is a &lt;em>function&lt;/em> from the sample space to real numbers.
Once you define the random variable clearly, the rest (pmf/pdf/cdf, mean, variance) becomes systematic.&lt;/p>
&lt;/blockquote>
&lt;hr>


&lt;script src="https://arshadhs.github.io/mermaid.min.js">&lt;/script>

 &lt;script>mermaid.initialize({
 "flowchart": {
 "useMaxWidth":true
 },
 "theme": "default"
}
)&lt;/script>




&lt;pre class="mermaid">
flowchart TD
PD[&amp;#34;Probability&amp;lt;br/&amp;gt;distributions&amp;#34;] --&amp;gt; RV[&amp;#34;Random&amp;lt;br/&amp;gt;variables&amp;#34;]

RV --&amp;gt; T[&amp;#34;Types&amp;#34;]
T --&amp;gt; RV1[&amp;#34;Discrete&amp;lt;br/&amp;gt;RVs&amp;#34;]
T --&amp;gt; RV2[&amp;#34;Continuous&amp;lt;br/&amp;gt;RVs&amp;#34;]

RV --&amp;gt; F[&amp;#34;PMF / PDF / CDF&amp;#34;]
RV --&amp;gt; S[&amp;#34;Mean / Variance&amp;lt;br/&amp;gt;Covariance&amp;#34;]
RV --&amp;gt; J[&amp;#34;Joint &amp;amp; Marginal&amp;lt;br/&amp;gt;distributions&amp;#34;]
RV --&amp;gt; X[&amp;#34;Transformations&amp;#34;]

style PD fill:#90CAF9,stroke:#1E88E5,color:#000
style RV fill:#90CAF9,stroke:#1E88E5,color:#000

style T fill:#CE93D8,stroke:#8E24AA,color:#000
style F fill:#CE93D8,stroke:#8E24AA,color:#000
style S fill:#CE93D8,stroke:#8E24AA,color:#000
style J fill:#CE93D8,stroke:#8E24AA,color:#000
style X fill:#CE93D8,stroke:#8E24AA,color:#000
style RV1 fill:#CE93D8,stroke:#8E24AA,color:#000
style RV2 fill:#CE93D8,stroke:#8E24AA,color:#000
&lt;/pre>

&lt;hr>
&lt;h2 id="1-definition">
 1) Definition
 
 &lt;a class="anchor" href="#1-definition">#&lt;/a>
 
&lt;/h2>
&lt;p>Random variable:
a rule that assigns a number to each outcome.&lt;/p></description></item><item><title>Common Probability Distributions</title><link>https://arshadhs.github.io/docs/ai/statistics/probability_distributions/common-distributions/</link><pubDate>Sun, 22 Feb 2026 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/statistics/probability_distributions/common-distributions/</guid><description>&lt;h1 id="common-probability-distributions">
 Common Probability Distributions
 
 &lt;a class="anchor" href="#common-probability-distributions">#&lt;/a>
 
&lt;/h1>
&lt;p>Once you can describe a random variable using a pmf or pdf, the next step is to use
named distributions that appear repeatedly in real data and in ML models.&lt;/p>
&lt;blockquote class="book-hint info">
&lt;p>Key takeaway:
Named distributions give you ready-made probability models for common patterns:
binary outcomes, counts, and measurement noise.&lt;/p>
&lt;/blockquote>
&lt;hr>


&lt;pre class="mermaid">
flowchart TD
PD[&amp;#34;Probability&amp;lt;br/&amp;gt;distributions&amp;#34;] --&amp;gt; DS[&amp;#34;Common&amp;lt;br/&amp;gt;distributions&amp;#34;]

DS --&amp;gt; DIS[&amp;#34;Discrete&amp;#34;]
DS --&amp;gt; CON[&amp;#34;Continuous&amp;#34;]

DIS --&amp;gt; D1[&amp;#34;Bernoulli&amp;#34;]
DIS --&amp;gt; D2[&amp;#34;Binomial&amp;#34;]
DIS --&amp;gt; D3[&amp;#34;Poisson&amp;#34;]

CON --&amp;gt; D4[&amp;#34;Normal&amp;lt;br/&amp;gt;(Gaussian)&amp;#34;]
CON --&amp;gt; D5[&amp;#34;t / Chi-square / F&amp;lt;br/&amp;gt;(intro)&amp;#34;]

style PD fill:#90CAF9,stroke:#1E88E5,color:#000
style DS fill:#90CAF9,stroke:#1E88E5,color:#000

style DIS fill:#CE93D8,stroke:#8E24AA,color:#000
style CON fill:#CE93D8,stroke:#8E24AA,color:#000

style D1 fill:#C8E6C9,stroke:#2E7D32,color:#000
style D2 fill:#C8E6C9,stroke:#2E7D32,color:#000
style D3 fill:#C8E6C9,stroke:#2E7D32,color:#000
style D4 fill:#C8E6C9,stroke:#2E7D32,color:#000
style D5 fill:#C8E6C9,stroke:#2E7D32,color:#000
&lt;/pre>

&lt;hr>
&lt;h2 id="1-bernoulli-distribution-binary">
 1) Bernoulli distribution (binary)
 
 &lt;a class="anchor" href="#1-bernoulli-distribution-binary">#&lt;/a>
 
&lt;/h2>
&lt;p>Use when:
one trial has two outcomes (success/failure).&lt;/p></description></item><item><title>Mathematical Foundation</title><link>https://arshadhs.github.io/docs/ai/maths/</link><pubDate>Wed, 18 Mar 2026 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/maths/</guid><description>&lt;h1 id="mathematical-foundations-for-machine-learning">
 Mathematical Foundations for Machine Learning
 
 &lt;a class="anchor" href="#mathematical-foundations-for-machine-learning">#&lt;/a>
 
&lt;/h1>
&lt;p>Machine Learning is built on &lt;strong>mathematical principles&lt;/strong> that allow models to:&lt;/p>
&lt;ul>
&lt;li>represent data&lt;/li>
&lt;li>learn patterns&lt;/li>
&lt;li>optimise performance&lt;/li>
&lt;/ul>


&lt;script src="https://arshadhs.github.io/mermaid.min.js">&lt;/script>

 &lt;script>mermaid.initialize({
 "flowchart": {
 "useMaxWidth":true
 },
 "theme": "default"
}
)&lt;/script>




&lt;pre class="mermaid">
flowchart LR
 DATA[Data]
 MATH[Math Models]
 OPT[Optimisation]
 MODEL[Trained Model]

 DATA --&amp;gt; MATH
 MATH --&amp;gt; OPT
 OPT --&amp;gt; MODEL
&lt;/pre>

&lt;p>ML requires &lt;strong>core mathematical tools&lt;/strong> to understand how ML algorithms work internally. Algebra deals with relationships between variables and quantities, while Calculus focuses on change and optimization.&lt;/p></description></item><item><title>Statistics</title><link>https://arshadhs.github.io/docs/ai/statistics/</link><pubDate>Thu, 12 Mar 2026 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/statistics/</guid><description>&lt;h1 id="statistics">
 Statistics
 
 &lt;a class="anchor" href="#statistics">#&lt;/a>
 
&lt;/h1>
&lt;p>&lt;strong>Statistical methods&lt;/strong> help you turn &lt;strong>raw data into reliable conclusions&lt;/strong>, while understanding &lt;strong>uncertainty, variability, and confidence&lt;/strong>.&lt;/p>
&lt;p>Statistics provides the &lt;strong>language and tools&lt;/strong> for reasoning about data, uncertainty, and inference.&lt;/p>
&lt;p>ML needs &lt;strong>understanding data behaviour&lt;/strong>, drawing conclusions, and validating machine learning models.&lt;/p>
&lt;ul>
&lt;li>Collect Data&lt;/li>
&lt;li>Present &amp;amp; Organise Data (in a systematic manner)&lt;/li>
&lt;li>Alalyse Data&lt;/li>
&lt;li>Infer about the Data&lt;/li>
&lt;li>Take Decision from the Data&lt;/li>
&lt;/ul>
&lt;hr>




&lt;ul>
 
 
 
 
 &lt;li>
 &lt;a href="https://arshadhs.github.io/docs/ai/statistics/00_formulas/">Formula Sheet&lt;/a>
 &lt;/li>
 
 
 
 
 &lt;li>
 &lt;a href="https://arshadhs.github.io/docs/ai/statistics/ism-formula-sheet/">Stats Formula Sheet&lt;/a>
 &lt;/li>
 
 
 
 
 &lt;li>
 &lt;a href="https://arshadhs.github.io/docs/ai/statistics/01_basic_statistics/">Basic Statistics&lt;/a>
 &lt;/li>
 
 
 
 
 &lt;li>
 &lt;a href="https://arshadhs.github.io/docs/ai/statistics/01_basic_probability/">Basic Probability&lt;/a>
 &lt;/li>
 
 
 
 
 
 
 
 
 
 
 &lt;li>
 &lt;a href="https://arshadhs.github.io/docs/ai/statistics/04_hypothesis_testing/">Hypothesis Testing&lt;/a>
 &lt;/li>
 
 
 
 
 &lt;li>
 &lt;a href="https://arshadhs.github.io/docs/ai/statistics/05_prediction_n_forecasting/">Prediction &amp;amp; Forecasting&lt;/a>
 &lt;/li>
 
 
 
 
 &lt;li>
 &lt;a href="https://arshadhs.github.io/docs/ai/statistics/06_prediction_n_forecasting/">Gaussian Mixture model &amp;amp; Expectation Maximization&lt;/a>
 &lt;/li>
 
 

 
 
 &lt;li>
 &lt;a href="https://arshadhs.github.io/docs/ai/statistics/conditional-probability/">Conditional Probability &amp;amp; Bayes’ Theorem&lt;/a>

 
 



&lt;ul>
 
 
 
 
 &lt;li>
 &lt;a href="https://arshadhs.github.io/docs/ai/statistics/conditional-probability/021_conditional_prob/">Conditional Probability&lt;/a>
 &lt;/li>
 
 
 
 
 &lt;li>
 &lt;a href="https://arshadhs.github.io/docs/ai/statistics/conditional-probability/022_bayes_theorem/">Bayes’ Theorem&lt;/a>
 &lt;/li>
 
 
 
 
 &lt;li>
 &lt;a href="https://arshadhs.github.io/docs/ai/statistics/conditional-probability/023_naive_bayes/">Naïve Bayes&lt;/a>
 &lt;/li>
 
 

 
 
&lt;/ul>

 &lt;/li>
 
 &lt;li>
 &lt;a href="https://arshadhs.github.io/docs/ai/statistics/probability_distributions/">Probability Distributions&lt;/a>

 
 



&lt;ul>
 
 
 
 
 &lt;li>
 &lt;a href="https://arshadhs.github.io/docs/ai/statistics/probability_distributions/random-variables/">Random Variables&lt;/a>
 &lt;/li>
 
 
 
 
 &lt;li>
 &lt;a href="https://arshadhs.github.io/docs/ai/statistics/probability_distributions/common-distributions/">Common Probability Distributions&lt;/a>
 &lt;/li>
 
 

 
 
&lt;/ul>

 &lt;/li>
 
&lt;/ul>


&lt;hr>
&lt;table>
 &lt;thead>
 &lt;tr>
 &lt;th>Statistics Topic&lt;/th>
 &lt;th>What you learn (plain English)&lt;/th>
 &lt;th>ML Connection&lt;/th>
 &lt;/tr>
 &lt;/thead>
 &lt;tbody>
 &lt;tr>
 &lt;td>1. Basic Probability &amp;amp; Statistics&lt;/td>
 &lt;td>Summarise data;&lt;br>understand spread;&lt;br>basic probability rules&lt;/td>
 &lt;td>Data understanding (EDA), feature sanity checks,&lt;br>detecting outliers, interpreting “average behaviour”&lt;/td>
 &lt;/tr>
 &lt;tr>
 &lt;td>2. Conditional Probability &amp;amp; Bayes&lt;/td>
 &lt;td>Update probability using new information;&lt;br>Bayes’ rule&lt;/td>
 &lt;td>Naïve Bayes, Bayesian thinking,&lt;br>posterior probabilities, probabilistic classification&lt;/td>
 &lt;/tr>
 &lt;tr>
 &lt;td>3. Probability Distributions&lt;/td>
 &lt;td>Model randomness with distributions;&lt;br>expectation/variance/covariance&lt;/td>
 &lt;td>Likelihood models, noise assumptions (Gaussian), sampling,&lt;br>probabilistic modelling foundations&lt;/td>
 &lt;/tr>
 &lt;tr>
 &lt;td>4. Hypothesis Testing&lt;/td>
 &lt;td>Sampling, CLT, confidence intervals,&lt;br>significance tests, ANOVA, MLE&lt;/td>
 &lt;td>A/B testing, evaluating model improvements,&lt;br>significance vs noise, parameter estimation (MLE)&lt;/td>
 &lt;/tr>
 &lt;tr>
 &lt;td>5. Prediction &amp;amp; Forecasting&lt;/td>
 &lt;td>Correlation, regression,&lt;br>time series (AR/MA/ARIMA/SARIMA etc.)&lt;/td>
 &lt;td>Linear regression, forecasting, sequential data modelling, baseline predictive modelling&lt;/td>
 &lt;/tr>
 &lt;tr>
 &lt;td>6. GMM &amp;amp; EM&lt;/td>
 &lt;td>Mixtures of Gaussians;&lt;br>iterative estimation with EM&lt;/td>
 &lt;td>Unsupervised learning (soft clustering),&lt;br>density estimation, latent-variable models&lt;/td>
 &lt;/tr>
 &lt;/tbody>
&lt;/table>
&lt;hr>


&lt;pre class="mermaid">
flowchart TD
 A[&amp;#34;Statistical Methods&amp;lt;br/&amp;gt;AIML ZC418&amp;#34;] --&amp;gt; B[&amp;#34;1. Basic Probability and Statistics&amp;#34;]
 A --&amp;gt; C[&amp;#34;2. Conditional Probability and Bayes&amp;#34;]
 A --&amp;gt; D[&amp;#34;3. Probability Distributions&amp;#34;]
 A --&amp;gt; E[&amp;#34;4. Hypothesis Testing&amp;#34;]
 A --&amp;gt; F[&amp;#34;5. Prediction and Forecasting&amp;#34;]
 A --&amp;gt; G[&amp;#34;6. Gaussian Mixture Model and EM&amp;#34;]

 B --&amp;gt; B1[&amp;#34;Central Tendency&amp;lt;br/&amp;gt;Mean - Median - Mode&amp;#34;]
 B --&amp;gt; B2[&amp;#34;Variability&amp;lt;br/&amp;gt;Range - Variance - SD - Quartiles&amp;#34;]
 B --&amp;gt; B3[&amp;#34;Basic Probability Concepts&amp;#34;]
 B3 --&amp;gt; B31[&amp;#34;Axioms of Probability&amp;#34;]
 B3 --&amp;gt; B32[&amp;#34;Definition of Probability&amp;#34;]
 B3 --&amp;gt; B33[&amp;#34;Mutually Exclusive vs Independent&amp;#34;]

 C --&amp;gt; C1[&amp;#34;Conditional Probability&amp;#34;]
 C --&amp;gt; C2[&amp;#34;Independence (conditional)&amp;#34;]
 C --&amp;gt; C3[&amp;#34;Bayes Theorem&amp;#34;]
 C --&amp;gt; C4[&amp;#34;Naive Bayes (intro)&amp;#34;]

 D --&amp;gt; D1[&amp;#34;Random Variables&amp;lt;br/&amp;gt;Discrete and Continuous&amp;#34;]
 D --&amp;gt; D2[&amp;#34;Expectation - Variance - Covariance&amp;#34;]
 D --&amp;gt; D3[&amp;#34;Transformations of RVs&amp;#34;]
 D --&amp;gt; D4[&amp;#34;Key Distributions&amp;#34;]
 D4 --&amp;gt; D41[&amp;#34;Bernoulli&amp;#34;]
 D4 --&amp;gt; D42[&amp;#34;Binomial&amp;#34;]
 D4 --&amp;gt; D43[&amp;#34;Poisson&amp;#34;]
 D4 --&amp;gt; D44[&amp;#34;Normal (Gaussian)&amp;#34;]
 D4 --&amp;gt; D45[&amp;#34;t - Chi-square - F (intro)&amp;#34;]

 E --&amp;gt; E1[&amp;#34;Sampling&amp;lt;br/&amp;gt;Random and Stratified&amp;#34;]
 E --&amp;gt; E2[&amp;#34;Sampling Distributions&amp;lt;br/&amp;gt;CLT&amp;#34;]
 E --&amp;gt; E3[&amp;#34;Estimation&amp;lt;br/&amp;gt;Confidence Intervals&amp;#34;]
 E --&amp;gt; E4[&amp;#34;Hypothesis Tests&amp;lt;br/&amp;gt;Means and Proportions&amp;#34;]
 E --&amp;gt; E5[&amp;#34;ANOVA&amp;lt;br/&amp;gt;Single and Dual factor&amp;#34;]
 E --&amp;gt; E6[&amp;#34;Maximum Likelihood&amp;#34;]

 F --&amp;gt; F1[&amp;#34;Correlation&amp;#34;]
 F --&amp;gt; F2[&amp;#34;Regression&amp;#34;]
 F --&amp;gt; F3[&amp;#34;Time Series Basics&amp;lt;br/&amp;gt;Components&amp;#34;]
 F --&amp;gt; F4[&amp;#34;Moving Averages&amp;lt;br/&amp;gt;Simple and Weighted&amp;#34;]
 F --&amp;gt; F5[&amp;#34;Time Series Models&amp;#34;]
 F5 --&amp;gt; F51[&amp;#34;AR&amp;#34;]
 F5 --&amp;gt; F52[&amp;#34;ARMA / ARIMA&amp;#34;]
 F5 --&amp;gt; F53[&amp;#34;SARIMA / SARIMAX&amp;#34;]
 F5 --&amp;gt; F54[&amp;#34;VAR / VARMAX&amp;#34;]
 F --&amp;gt; F6[&amp;#34;Exponential Smoothing&amp;#34;]

 G --&amp;gt; G1[&amp;#34;GMM&amp;lt;br/&amp;gt;Mixture of Gaussians&amp;#34;]
 G --&amp;gt; G2[&amp;#34;EM Algorithm&amp;lt;br/&amp;gt;E-step - M-step&amp;#34;]

 B -.-&amp;gt; C
 C -.-&amp;gt; D
 D -.-&amp;gt; E
 E -.-&amp;gt; F
 F -.-&amp;gt; G
&lt;/pre>

&lt;hr>
&lt;h2 id="data---types">
 Data - Types
 
 &lt;a class="anchor" href="#data---types">#&lt;/a>
 
&lt;/h2>


&lt;pre class="mermaid">
flowchart TD
	A[(Data)] --&amp;gt; B[&amp;#34;Categorical (Qualitative)&amp;#34;]
 A --&amp;gt; C[&amp;#34;Numerical (Quantitative)&amp;#34;]

 B --&amp;gt; B1[Nominal]
 B --&amp;gt; B2[Ordinal]

 C --&amp;gt; C1[Discrete]
 C --&amp;gt; C2[Continuous]

 C2 --&amp;gt; C21[Interval]
 C2 --&amp;gt; C22[Ratio]

 %% Styling
 style A fill:#E1F5FE,stroke:#333
 style B fill:#90CAF9,stroke:#333
 style B1 fill:#90CAF9,stroke:#333
 style B2 fill:#90CAF9,stroke:#333
 style C fill:#FFF9C4,stroke:#333
 style C1 fill:#FFF9C4,stroke:#333
 style C2 fill:#FFF9C4,stroke:#333
 style C21 fill:#FFF9C4,stroke:#333
 style C22 fill:#FFF9C4,stroke:#333
&lt;/pre>

&lt;div class="book-steps ">
&lt;ol>
&lt;li>
&lt;h2 id="categorical-qualitative">
 Categorical (Qualitative)
 
 &lt;a class="anchor" href="#categorical-qualitative">#&lt;/a>
 
&lt;/h2>
&lt;p>express a qualitative attribute
e.g. hair color, eye color&lt;/p></description></item><item><title>Principal Component Analysis (PCA)</title><link>https://arshadhs.github.io/docs/ai/maths/010-linear-algebra/07-dimensionality-reduction/pca/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/maths/010-linear-algebra/07-dimensionality-reduction/pca/</guid><description>&lt;h1 id="principal-component-analysis-pca">
 Principal Component Analysis (PCA)
 
 &lt;a class="anchor" href="#principal-component-analysis-pca">#&lt;/a>
 
&lt;/h1>
&lt;ul>
&lt;li>dimensionality reduction technique&lt;/li>
&lt;li>helps us to &lt;strong>reduce the number of features&lt;/strong> in a dataset while keeping the most important information.&lt;/li>
&lt;li>changes complex datasets by transforming correlated features into a smaller set of uncorrelated components.&lt;/li>
&lt;li>uses &lt;strong>linear algebra&lt;/strong> to transform data into &lt;strong>new features&lt;/strong> called principal components.&lt;/li>
&lt;li>finds these by calculating &lt;strong>eigenvectors (directions)&lt;/strong> and &lt;strong>eigenvalues (importance)&lt;/strong> from the &lt;strong>covariance matrix&lt;/strong>.&lt;/li>
&lt;li>PCA &lt;strong>selects the top components with the highest eigenvalues&lt;/strong> and &lt;strong>projects the data onto them simplify the dataset&lt;/strong>.&lt;/li>
&lt;/ul>
&lt;blockquote class="book-hint default">
&lt;p>PCA prioritizes the directions where the data varies the most because more variation = more useful information.&lt;/p></description></item></channel></rss>