๐ŸŽ“How I Study AIHISA
๐Ÿ“–Read
๐Ÿ“„Papers๐Ÿ“ฐBlogs๐ŸŽฌCourses
๐Ÿ’กLearn
๐Ÿ›ค๏ธPaths๐Ÿ“šTopics๐Ÿ’กConcepts๐ŸŽดShorts
๐ŸŽฏPractice
โฑ๏ธCoach๐ŸงฉProblems๐Ÿง Thinking๐ŸŽฏPrompts๐Ÿง Review
SearchSettings
How I Study AI - Learn AI Papers & Lectures the Easy Way

Concepts13

Groups

๐Ÿ“Linear Algebra15๐Ÿ“ˆCalculus & Differentiation10๐ŸŽฏOptimization14๐ŸŽฒProbability Theory12๐Ÿ“ŠStatistics for ML9๐Ÿ“กInformation Theory10๐Ÿ”บConvex Optimization7๐Ÿ”ขNumerical Methods6๐Ÿ•ธGraph Theory for Deep Learning6๐Ÿ”ตTopology for ML5๐ŸŒDifferential Geometry6โˆžMeasure Theory & Functional Analysis6๐ŸŽฐRandom Matrix Theory5๐ŸŒŠFourier Analysis & Signal Processing9๐ŸŽฐSampling & Monte Carlo Methods10๐Ÿง Deep Learning Theory12๐Ÿ›ก๏ธRegularization Theory11๐Ÿ‘๏ธAttention & Transformer Theory10๐ŸŽจGenerative Model Theory11๐Ÿ”ฎRepresentation Learning10๐ŸŽฎReinforcement Learning Mathematics9๐Ÿ”„Variational Methods8๐Ÿ“‰Loss Functions & Objectives10โฑ๏ธSequence & Temporal Models8๐Ÿ’ŽGeometric Deep Learning8

Category

๐Ÿ”ทAllโˆ‘Mathโš™๏ธAlgo๐Ÿ—‚๏ธDS๐Ÿ“šTheory

Level

AllBeginnerIntermediate
โˆ‘MathIntermediate

Pseudoinverse (Moore-Penrose)

The Mooreโ€“Penrose pseudoinverse generalizes matrix inversion to rectangular or singular matrices and is denoted Aโบ.

#pseudoinverse#moore-penrose#least squares+12
โˆ‘MathIntermediate

Kronecker Product & Vec Operator

The Kronecker product A โŠ— B expands a small matrix into a larger block matrix by multiplying every entry of A with the whole matrix B.

#kronecker product
12
Advanced
Group:
Linear Algebra
#vec operator
#block matrix
+12
โˆ‘MathIntermediate

Orthogonal & Unitary Matrices

Orthogonal (real) and unitary (complex) matrices are length- and angle-preserving transformations, like perfect rotations and reflections.

#orthogonal matrix#unitary matrix#conjugate transpose+12
โˆ‘MathIntermediate

Matrix Calculus Fundamentals

Matrix calculus extends single-variable derivatives to matrices so we can differentiate functions built from matrix multiplications, traces, and norms.

#matrix calculus#frobenius norm#trace trick+12
โˆ‘MathIntermediate

Low-Rank Approximation

Low-rank approximation replaces a big matrix with one that has far fewer degrees of freedom while preserving most of its action.

#low-rank approximation#eckart-young theorem#svd+12
โˆ‘MathIntermediate

Tensor Operations

A tensor is a multi-dimensional array that generalizes scalars (0-D), vectors (1-D), and matrices (2-D) to higher dimensions.

#tensor#multi-dimensional array#broadcasting+12
โˆ‘MathIntermediate

Matrix Norms & Condition Numbers

Matrix norms measure the size of a matrix in different but related ways, with Frobenius treating entries like a big vector, spectral measuring the strongest stretch, and nuclear summing all singular values.

#matrix norm#spectral norm#frobenius norm+12
โˆ‘MathIntermediate

Positive Definite Matrices

A real symmetric matrix A is positive definite if and only if x^T A x > 0 for every nonzero vector x, and positive semidefinite if x^T A x โ‰ฅ 0.

#positive definite#positive semidefinite#cholesky decomposition+11
โˆ‘MathIntermediate

Eigendecomposition

Eigendecomposition expresses a matrix as a change of basis times a diagonal scaling, revealing its natural stretching directions.

#eigendecomposition#eigenvalue#eigenvector+11
โˆ‘MathIntermediate

Inner Products & Norms

An inner product measures how much two vectors point in the same direction; in R^n it is the dot product.

#inner product#dot product#norm+12
โˆ‘MathIntermediate

Systems of Linear Equations

A system of linear equations asks for numbers that make several linear relationships true at the same time, which we compactly write as Ax = b.

#systems of linear equations#gaussian elimination#row echelon form+12
โˆ‘MathIntermediate

Matrix Operations & Properties

Matrix operations like multiplication and transpose combine or reorient data tables and linear transformations in predictable ways.

#matrix multiplication#transpose#trace+12