Groups
Category
Standard softmax attention costs O(n²) in sequence length because every token compares with every other token.
The kernel (lazy) regime keeps neural network parameters close to their initialization, making training equivalent to kernel regression with a fixed kernel such as the Neural Tangent Kernel (NTK).