๐ŸŽ“How I Study AIHISA
๐Ÿ“–Read
๐Ÿ“„Papers๐Ÿ“ฐBlogs๐ŸŽฌCourses
๐Ÿ’กLearn
๐Ÿ›ค๏ธPaths๐Ÿ“šTopics๐Ÿ’กConcepts๐ŸŽดShorts
๐ŸŽฏPractice
โฑ๏ธCoach๐ŸงฉProblems๐Ÿง Thinking๐ŸŽฏPrompts๐Ÿง Review
SearchSettings
How I Study AI - Learn AI Papers & Lectures the Easy Way

Concepts11

Groups

๐Ÿ“Linear Algebra15๐Ÿ“ˆCalculus & Differentiation10๐ŸŽฏOptimization14๐ŸŽฒProbability Theory12๐Ÿ“ŠStatistics for ML9๐Ÿ“กInformation Theory10๐Ÿ”บConvex Optimization7๐Ÿ”ขNumerical Methods6๐Ÿ•ธGraph Theory for Deep Learning6๐Ÿ”ตTopology for ML5๐ŸŒDifferential Geometry6โˆžMeasure Theory & Functional Analysis6๐ŸŽฐRandom Matrix Theory5๐ŸŒŠFourier Analysis & Signal Processing9๐ŸŽฐSampling & Monte Carlo Methods10๐Ÿง Deep Learning Theory12๐Ÿ›ก๏ธRegularization Theory11๐Ÿ‘๏ธAttention & Transformer Theory10๐ŸŽจGenerative Model Theory11๐Ÿ”ฎRepresentation Learning10๐ŸŽฎReinforcement Learning Mathematics9๐Ÿ”„Variational Methods8๐Ÿ“‰Loss Functions & Objectives10โฑ๏ธSequence & Temporal Models8๐Ÿ’ŽGeometric Deep Learning8

Category

๐Ÿ”ทAllโˆ‘Mathโš™๏ธAlgo๐Ÿ—‚๏ธDS๐Ÿ“šTheory

Level

AllBeginnerIntermediate
๐Ÿ“šTheoryIntermediate

Classifier-Free Guidance

Classifier-Free Guidance (CFG) steers diffusion sampling toward a condition (like a text prompt) without needing a separate classifier.

#classifier-free guidance#diffusion models#epsilon prediction+11
๐Ÿ“šTheoryAdvanced

Energy-Based Models (EBM)

Energy-Based Models (EBMs) define probabilities through an energy landscape: low energy means high probability, with p(x) = exp(-E(x)) / Z.

#energy-based models
Advanced
Group:
Generative Model Theory
#partition function
#langevin dynamics
+12
๐Ÿ“šTheoryIntermediate

Flow Matching

Flow matching learns a time-dependent vector field v_t(x, c) whose ODE transports simple noise to complex data, enabling fast, deterministic sampling.

#flow matching#conditional flow matching#rectified flow+11
๐Ÿ“šTheoryIntermediate

Autoregressive Models

Autoregressive (AR) models represent a joint distribution by multiplying conditional probabilities in a fixed order, using the chain rule of probability.

#autoregressive#ar model#n-gram+11
โˆ‘MathAdvanced

Stochastic Differential Equations for Generation

A forward stochastic differential equation (SDE) models a state that drifts deterministically and is shaken by random Brownian noise over time.

#stochastic differential equation#diffusion model#euler maruyama+12
๐Ÿ“šTheoryAdvanced

Diffusion Models (Score-Based)

Score-based diffusion models corrupt data by gradually adding Gaussian noise and then learn to reverse this process by estimating the score, the gradient of the log-density.

#diffusion models#score-based modeling#ddpm+7
๐Ÿ“šTheoryAdvanced

Normalizing Flows

Normalizing flows transform a simple base distribution (like a standard Gaussian) into a complex target distribution using a chain of invertible functions.

#normalizing flows#change of variables#jacobian determinant+12
โˆ‘MathIntermediate

Wasserstein Distance & Optimal Transport

Wasserstein distance (Earth Moverโ€™s Distance) measures how much โ€œworkโ€ is needed to transform one probability distribution into another by moving mass with minimal total cost.

#wasserstein distance#earth mover's distance#optimal transport+12
๐Ÿ“šTheoryAdvanced

GAN Theory & Training Dynamics

GANs frame learning as a two-player game where a generator tries to fool a discriminator, and the discriminator tries to detect fakes.

#gan#generator#discriminator+12
๐Ÿ“šTheoryAdvanced

Variational Autoencoders (VAE) Theory

A Variational Autoencoder (VAE) is a probabilistic autoencoder that learns to generate data by inferring hidden causes (latent variables) and decoding them back to observations.

#variational autoencoder#elbo#kl divergence+12
๐Ÿ“šTheoryIntermediate

Maximum Likelihood & Generative Models

Maximum Likelihood Estimation (MLE) picks parameters that make the observed data most probable under a chosen probabilistic model.

#maximum likelihood#generative models#naive bayes+12