Concepts27

๐Ÿ“šTheoryAdvanced

PAC-Bayes Theory

PAC-Bayes provides high-probability generalization bounds for randomized predictors by comparing a data-dependent posterior Q to a fixed, data-independent prior P through KL(Q||P).

#pac-bayes#generalization bound#kl divergence+12
๐Ÿ“šTheoryAdvanced

MCMC Theory

MCMC simulates a Markov chain whose long-run behavior matches a target distribution, letting us sample from complex posteriors without knowing the normalization constant.

#mcmc#metropolis-hastings#gibbs sampling+11
๐Ÿ“šTheoryAdvanced

Graph Neural Network Theory

Graph Neural Networks (GNNs) learn on graphs by repeatedly letting each node aggregate messages from its neighbors and update its representation.

#graph neural networks#message passing#weisfeiler-leman+12
๐Ÿ“šTheoryAdvanced

Differential Privacy Theory

Differential privacy (DP) guarantees that the output of a randomized algorithm does not change much when one personโ€™s data is added or removed.

#differential privacy#epsilon delta dp#laplace mechanism+12
๐Ÿ“šTheoryAdvanced

Information-Theoretic Lower Bounds

Information-theoretic lower bounds tell you the best possible performance any learning algorithm can achieve, regardless of cleverness or compute.

#information-theoretic lower bounds#fano inequality#le cam method+12
๐Ÿ“šTheoryAdvanced

Quantum Computing Theory

Quantum computing uses qubits that can be in superpositions, enabling interference-based computation beyond classical bits.

#quantum computing#qubit#superposition+12
๐Ÿ“šTheoryAdvanced

Streaming Algorithm Theory

Streaming algorithms process massive data one pass at a time using sublinearโ€”often polylogarithmicโ€”memory.

#streaming algorithms#count-min sketch#misra-gries+12
๐Ÿ“šTheoryAdvanced

Distributed Algorithm Theory

Distributed algorithm theory studies how many independent computers cooperate correctly and efficiently despite delays and failures.

#distributed algorithms#message passing#synchronous rounds+12
๐Ÿ“šTheoryAdvanced

Algorithmic Information Theory

Algorithmic Information Theory studies information content via the shortest programs that generate data, rather than via average-case probabilities.

#kolmogorov complexity#algorithmic probability#solomonoff induction+11
๐Ÿ“šTheoryAdvanced

Optimal Transport Theory

Optimal Transport (OT) formalizes the cheapest way to move one probability distribution into another given a cost to move mass.

#optimal transport#wasserstein distance#kantorovich+12
๐Ÿ“šTheoryAdvanced

Diffusion Models Theory

Diffusion models learn to reverse a simple noising process by estimating the score (the gradient of the log density) of data at different noise levels.

#diffusion models#ddpm#score matching+12
๐Ÿ“šTheoryAdvanced

Variational Inference Theory

Variational Inference (VI) replaces an intractable posterior with a simpler distribution and optimizes it by minimizing KL divergence, which is equivalent to maximizing the ELBO.

#variational inference#elbo#kl divergence+12