Groups
Category
Sequence-to-sequence with attention lets a decoder focus on the most relevant parts of the input at each output step, rather than compressing everything into a single vector.
Transformer expressiveness studies what kinds of sequence-to-sequence mappings a Transformer can represent or approximate.