๐ŸŽ“How I Study AIHISA
๐Ÿ“–Read
๐Ÿ“„Papers๐Ÿ“ฐBlogs๐ŸŽฌCourses
๐Ÿ’กLearn
๐Ÿ›ค๏ธPaths๐Ÿ“šTopics๐Ÿ’กConcepts๐ŸŽดShorts
๐ŸŽฏPractice
๐Ÿ“Daily Log๐ŸŽฏPrompts๐Ÿง Review
SearchSettings
How I Study AI - Learn AI Papers & Lectures the Easy Way

Concepts187

Groups

๐Ÿ“Linear Algebra15๐Ÿ“ˆCalculus & Differentiation10๐ŸŽฏOptimization14๐ŸŽฒProbability Theory12๐Ÿ“ŠStatistics for ML9๐Ÿ“กInformation Theory10๐Ÿ”บConvex Optimization7๐Ÿ”ขNumerical Methods6๐Ÿ•ธGraph Theory for Deep Learning6๐Ÿ”ตTopology for ML5๐ŸŒDifferential Geometry6โˆžMeasure Theory & Functional Analysis6๐ŸŽฐRandom Matrix Theory5๐ŸŒŠFourier Analysis & Signal Processing9๐ŸŽฐSampling & Monte Carlo Methods10๐Ÿง Deep Learning Theory12๐Ÿ›ก๏ธRegularization Theory11๐Ÿ‘๏ธAttention & Transformer Theory10๐ŸŽจGenerative Model Theory11๐Ÿ”ฎRepresentation Learning10๐ŸŽฎReinforcement Learning Mathematics9๐Ÿ”„Variational Methods8๐Ÿ“‰Loss Functions & Objectives10โฑ๏ธSequence & Temporal Models8๐Ÿ’ŽGeometric Deep Learning8

Category

๐Ÿ”ทAllโˆ‘Mathโš™๏ธAlgo๐Ÿ—‚๏ธDS๐Ÿ“šTheory

Level

AllBeginnerIntermediateAdvanced
โš™๏ธAlgorithmIntermediate

Discrete Fourier Transform (DFT) & FFT

The Discrete Fourier Transform (DFT) converts a length-N sequence from the time (or spatial) domain into N complex frequency coefficients that describe how much of each sinusoid is present.

#dft#fft#cooley-tukey+12
โš™๏ธAlgorithmAdvanced

Natural Gradient Method

Natural gradient scales the ordinary gradient by the inverse Fisher information matrix to account for the geometry of probability distributions.

#natural gradient
12345
#fisher information
#empirical fisher
+12
โš™๏ธAlgorithmIntermediate

Numerical Differentiation & Finite Differences

Numerical differentiation uses finite differences to estimate derivatives when an analytical derivative is hard or impossible to obtain.

#numerical differentiation#finite differences#forward difference+12
โš™๏ธAlgorithmIntermediate

Numerical Integration & Monte Carlo

Numerical integration approximates the area under a curve when an exact antiderivative is unknown, using deterministic quadrature rules or random sampling (Monte Carlo).

#numerical integration#quadrature#trapezoidal rule+11
โš™๏ธAlgorithmIntermediate

Matrix Factorizations (Numerical)

Matrix factorizations rewrite a matrix into simpler building blocks (triangular or orthogonal) that make solving and analyzing linear systems much easier.

#lu decomposition#qr factorization#householder reflections+12
โš™๏ธAlgorithmIntermediate

Iterative Methods for Linear Systems

The Conjugate Gradient (CG) method solves large, sparse, symmetric positive definite (SPD) linear systems Ax = b using only matrixโ€“vector products and dot products.

#conjugate gradient#iterative solver#krylov subspace+12
โš™๏ธAlgorithmAdvanced

Interior Point Methods

Interior point methods solve constrained optimization by replacing hard constraints with a smooth barrier that becomes infinite at the boundary, keeping iterates strictly inside the feasible region.

#interior point method#logarithmic barrier#central path+12
โš™๏ธAlgorithmAdvanced

ADMM (Alternating Direction Method of Multipliers)

ADMM splits a hard optimization problem into two easier subproblems that communicate through simple averaging-like steps.

#admm#alternating direction method of multipliers#augmented lagrangian+11
โš™๏ธAlgorithmIntermediate

Proximal Operators & Methods

A proximal operator pulls a point x toward minimizing a function f while penalizing how far it moves, acting like a denoiser or projector depending on f.

#proximal operator#ista#fista+12
โš™๏ธAlgorithmIntermediate

Bootstrap & Resampling Methods

Bootstrap is a resampling method that estimates uncertainty by repeatedly sampling with replacement from the observed data.

#bootstrap#resampling#confidence intervals+12
โš™๏ธAlgorithmIntermediate

Gradient Clipping & Normalization

Gradient clipping limits how large gradient values or their overall magnitude can become during optimization to prevent exploding updates.

#gradient clipping#clipping by norm#clipping by value+12
โš™๏ธAlgorithmAdvanced

Newton's Method & Second-Order Optimization

Newton's method uses both the gradient and the Hessian to take steps that aim directly at the local optimum by fitting a quadratic model of the loss around the current point.

#newton's method#second-order optimization#hessian+12