๐ŸŽ“How I Study AIHISA
๐Ÿ“–Read
๐Ÿ“„Papers๐Ÿ“ฐBlogs๐ŸŽฌCourses
๐Ÿ’กLearn
๐Ÿ›ค๏ธPaths๐Ÿ“šTopics๐Ÿ’กConcepts๐ŸŽดShorts
๐ŸŽฏPractice
๐Ÿ“Daily Log๐ŸŽฏPrompts๐Ÿง Review
SearchSettings
How I Study AI - Learn AI Papers & Lectures the Easy Way

Concepts16

Groups

๐Ÿ“Linear Algebra15๐Ÿ“ˆCalculus & Differentiation10๐ŸŽฏOptimization14๐ŸŽฒProbability Theory12๐Ÿ“ŠStatistics for ML9๐Ÿ“กInformation Theory10๐Ÿ”บConvex Optimization7๐Ÿ”ขNumerical Methods6๐Ÿ•ธGraph Theory for Deep Learning6๐Ÿ”ตTopology for ML5๐ŸŒDifferential Geometry6โˆžMeasure Theory & Functional Analysis6๐ŸŽฐRandom Matrix Theory5๐ŸŒŠFourier Analysis & Signal Processing9๐ŸŽฐSampling & Monte Carlo Methods10๐Ÿง Deep Learning Theory12๐Ÿ›ก๏ธRegularization Theory11๐Ÿ‘๏ธAttention & Transformer Theory10๐ŸŽจGenerative Model Theory11๐Ÿ”ฎRepresentation Learning10๐ŸŽฎReinforcement Learning Mathematics9๐Ÿ”„Variational Methods8๐Ÿ“‰Loss Functions & Objectives10โฑ๏ธSequence & Temporal Models8๐Ÿ’ŽGeometric Deep Learning8

Category

๐Ÿ”ทAllโˆ‘Mathโš™๏ธAlgo๐Ÿ—‚๏ธDS๐Ÿ“šTheory

Level

AllBeginnerIntermediate
โˆ‘MathIntermediate

Pseudoinverse (Moore-Penrose)

The Mooreโ€“Penrose pseudoinverse generalizes matrix inversion to rectangular or singular matrices and is denoted Aโบ.

#pseudoinverse#moore-penrose#least squares+12
๐Ÿ“šTheoryIntermediate

Spectral Regularization

Spectral regularization controls how much a weight matrix can stretch inputs by constraining its largest singular value (spectral norm).

#spectral regularization
12
Advanced
Filtering by:
#condition number
#spectral norm
#power iteration
+11
๐Ÿ“šTheoryIntermediate

Double Descent Phenomenon

Double descent describes how test error first follows the classic U-shape with increasing model complexity, spikes near the interpolation threshold, and then drops again in the highly overparameterized regime.

#double descent#interpolation threshold#overparameterization+12
โš™๏ธAlgorithmIntermediate

Matrix Factorizations (Numerical)

Matrix factorizations rewrite a matrix into simpler building blocks (triangular or orthogonal) that make solving and analyzing linear systems much easier.

#lu decomposition#qr factorization#householder reflections+12
โš™๏ธAlgorithmIntermediate

Iterative Methods for Linear Systems

The Conjugate Gradient (CG) method solves large, sparse, symmetric positive definite (SPD) linear systems Ax = b using only matrixโ€“vector products and dot products.

#conjugate gradient#iterative solver#krylov subspace+12
โˆ‘MathIntermediate

Numerical Stability

Numerical stability measures how much rounding and tiny input changes can distort an algorithmโ€™s output on real computers using floating-point arithmetic.

#numerical stability#forward error#backward error+12
โš™๏ธAlgorithmIntermediate

Gradient Descent

Gradient descent is a simple, repeatable way to move downhill on a loss surface by stepping in the opposite direction of the gradient.

#gradient descent#batch gradient descent#learning rate+12
โˆ‘MathIntermediate

Implicit Differentiation & Implicit Function Theorem

Implicit differentiation lets you find slopes and higher derivatives even when y is given indirectly by an equation F(x,y)=0.

#implicit differentiation#implicit function theorem#jacobian+12
โˆ‘MathIntermediate

Jacobian Matrix

The Jacobian matrix collects all first-order partial derivatives of a vector-valued function, describing how small input changes linearly affect each output component.

#jacobian matrix#partial derivatives#multivariable calculus+11
โˆ‘MathIntermediate

Matrix Norms & Condition Numbers

Matrix norms measure the size of a matrix in different but related ways, with Frobenius treating entries like a big vector, spectral measuring the strongest stretch, and nuclear summing all singular values.

#matrix norm#spectral norm#frobenius norm+12
โˆ‘MathIntermediate

Positive Definite Matrices

A real symmetric matrix A is positive definite if and only if x^T A x > 0 for every nonzero vector x, and positive semidefinite if x^T A x โ‰ฅ 0.

#positive definite#positive semidefinite#cholesky decomposition+11
โˆ‘MathIntermediate

Systems of Linear Equations

A system of linear equations asks for numbers that make several linear relationships true at the same time, which we compactly write as Ax = b.

#systems of linear equations#gaussian elimination#row echelon form+12