Transformer Architecture
A neural network architecture, introduced in the paper 'Attention Is All You Need,' that relies heavily on 'self-attention' mechanisms to process sequential data like text. It's the foundation for most modern Large Language Models (LLMs) due to its effectiveness in capturing long-range dependencies and contextual relationships.