šŸŽ“How I Study AIHISA
šŸ“–Read
šŸ“„PapersšŸ“°BlogsšŸŽ¬Courses
šŸ’”Learn
šŸ›¤ļøPathsšŸ“šTopicsšŸ’”ConceptsšŸŽ“Shorts
šŸŽÆPractice
šŸ“Daily LogšŸŽÆPrompts🧠Review
SearchSettings
How I Study AI - Learn AI Papers & Lectures the Easy Way

Concepts16

Groups

šŸ“Linear Algebra15šŸ“ˆCalculus & Differentiation10šŸŽÆOptimization14šŸŽ²Probability Theory12šŸ“ŠStatistics for ML9šŸ“”Information Theory10šŸ”ŗConvex Optimization7šŸ”¢Numerical Methods6šŸ•øGraph Theory for Deep Learning6šŸ”µTopology for ML5🌐Differential Geometry6āˆžMeasure Theory & Functional Analysis6šŸŽ°Random Matrix Theory5🌊Fourier Analysis & Signal Processing9šŸŽ°Sampling & Monte Carlo Methods10🧠Deep Learning Theory12šŸ›”ļøRegularization Theory11šŸ‘ļøAttention & Transformer Theory10šŸŽØGenerative Model Theory11šŸ”®Representation Learning10šŸŽ®Reinforcement Learning Mathematics9šŸ”„Variational Methods8šŸ“‰Loss Functions & Objectives10ā±ļøSequence & Temporal Models8šŸ’ŽGeometric Deep Learning8

Category

šŸ”·Allāˆ‘Mathāš™ļøAlgošŸ—‚ļøDSšŸ“šTheory

Level

AllBeginnerIntermediate
āš™ļøAlgorithmIntermediate

Efficient Attention Mechanisms

Standard softmax attention costs O(n²) in sequence length because every token compares with every other token.

#linear attention#efficient attention#kernel trick+12
šŸ“šTheoryIntermediate

Scaled Dot-Product Attention

Scaled dot-product attention scores how much each value V should contribute to a query by taking dot products with keys K, scaling by \(\sqrt{d_k}\), applying softmax, and forming a weighted sum.

#scaled dot-product attention
12
Advanced
Filtering by:
#time complexity
#softmax
#transformer
+10
āˆ‘MathIntermediate

Convolution Theorem

The Convolution Theorem says that convolving two signals in time (or space) equals multiplying their spectra in the frequency domain.

#convolution theorem#fft#dft+12
āˆ‘MathIntermediate

Tensor Operations

A tensor is a multi-dimensional array that generalizes scalars (0-D), vectors (1-D), and matrices (2-D) to higher dimensions.

#tensor#multi-dimensional array#broadcasting+12
āš™ļøAlgorithmIntermediate

Complexity Analysis Quick Reference

Use an operation budget of about 10^8 simple operations per second on typical online judges; always multiply by the time limit and number of test files if known.

#time complexity#competitive programming#big-o+12
āš™ļøAlgorithmIntermediate

Debugging Strategies for CP

Systematic debugging beats guesswork: always re-read the statement, re-check constraints, and verify the output format before touching code.

#competitive programming#debugging#stress testing+12
āš™ļøAlgorithmIntermediate

Small-to-Large Principle

Small-to-large means always merge the smaller container into the larger one to keep total work low.

#small-to-large#sack technique#dsu on tree+11
āˆ‘MathIntermediate

Permutations and Combinations

Permutations count ordered selections, while combinations count unordered selections.

#permutations#combinations#binomial coefficient+12
āˆ‘MathIntermediate

Fast Exponentiation

Fast exponentiation (binary exponentiation) computes a^n using repeated squaring in O(log n) multiplications.

#binary exponentiation#fast power#modular exponentiation+11
āš™ļøAlgorithmAdvanced

DSU on Tree (Sack)

DSU on Tree (also called the Sack technique) answers many subtree queries in O(n \log n) by keeping data from the heavy child and temporarily re-adding light subtrees.

#dsu on tree#sack technique#subtree queries+12
āš™ļøAlgorithmAdvanced

Knuth Optimization

Knuth Optimization speeds up a class of interval dynamic programming (DP) from O(n^3) to O(n^2) by exploiting the monotonicity of optimal split points.

#knuth optimization#interval dp#quadrangle inequality+12
āš™ļøAlgorithmIntermediate

Breadth-First Search (BFS)

Breadth-First Search (BFS) explores a graph level by level, visiting all vertices at distance d from the source before any at distance d+1.

#bfs#breadth first search#graph traversal+12