Concepts280

πŸ“šTheoryAdvanced

P vs NP Problem

P vs NP asks whether every problem whose solutions can be verified quickly can also be solved quickly.

#p vs np#np-complete#np-hard+12
πŸ“šTheoryAdvanced

GAN Theory

Generative Adversarial Networks (GANs) set up a two-player game where a generator tries to make fake samples that look real while a discriminator tries to tell real from fake.

#gan minimax#wasserstein gan#js divergence+11
πŸ“šTheoryIntermediate

Complexity Theory

Complexity theory classifies problems by the resources required to solve or verify them, such as time and memory.

#complexity theory#p vs np#np-complete+12
πŸ“šTheoryAdvanced

Representation Learning Theory

Representation learning aims to automatically discover features that make downstream tasks easy, often without human-provided labels.

#representation learning#contrastive learning#infonce+12
πŸ“šTheoryAdvanced

Information Bottleneck Theory

Information Bottleneck (IB) studies how to compress an input X into a representation Z that still preserves what is needed to predict Y.

#information bottleneck#mutual information#variational information bottleneck+12
πŸ“šTheoryIntermediate

Contrastive Learning Theory

Contrastive learning learns representations by pulling together positive pairs and pushing apart negatives using a softmax-based objective.

#contrastive learning#infonce#nt-xent+12
πŸ“šTheoryAdvanced

Policy Gradient Theorem

The policy gradient theorem tells us how to push a stochastic policy’s parameters to increase expected return by following the gradient of expected rewards.

#policy gradient#reinforce#actor-critic+11
πŸ“šTheoryIntermediate

Bellman Equations

Bellman equations express how the value of a state or action equals immediate reward plus discounted value of what follows.

#bellman equation#value iteration#policy iteration+12
πŸ“šTheoryAdvanced

Transformer Theory

Transformers map sequences to sequences using layers of self-attention and feed-forward networks wrapped with residual connections and LayerNorm.

#transformer#self-attention#positional encoding+12
πŸ“šTheoryAdvanced

Reinforcement Learning Theory

Reinforcement Learning (RL) studies how an agent learns to act in an environment to maximize long-term cumulative reward.

#reinforcement learning#mdp#bellman equation+12
πŸ“šTheoryAdvanced

Neural Tangent Kernel (NTK) Theory

The Neural Tangent Kernel (NTK) connects very wide neural networks to classical kernel methods, letting us study training as if it were kernel regression.

#neural tangent kernel#ntk#infinite width+12
πŸ“šTheoryIntermediate

Attention Mechanism Theory

Attention computes a weighted sum of values V where the weights come from how similar queries Q are to keys K.

#attention#self-attention#multi-head attention+12