Matrices are the core data structure of linear algebra and the workhorse of machine learning. Almost every ML model can be described as a sequence of matrix operations.
A square matrix is positive definite if pre-multiplying and post-multiplying it by the same vector always gives a positive number as a result, independently of how we choose the vector.
Positive definite symmetric matrices have the property that all their eigenvalues are positive.
The Cauchy–Schwarz Inequality is one of the most important results in linear algebra.
It places a fundamental bound on the inner product of two vectors.
If you see angle, cosine, similarity, or inner product bounds → think Cauchy–Schwarz Inequality
Key Idea:
The inner product (dot product) can never exceed the product of magnitudes.
This ensures all geometric interpretations (angles, cosine) are valid.
Decompositions reveal structure in matrices and power algorithms like PCA.
Matrix decompositions break complex matrices into simpler parts.
From the lecture introduction, matrices are used to describe mappings and transformations of vectors.
That is why decomposition is important:
it lets us understand a complicated transformation by rewriting it using simpler building blocks.
In the slides, the topic is introduced as part of three closely connected goals:
how to summarise matrices,
how matrices can be decomposed,
and how the decompositions can be used for matrix approximations.