The Transformer is the neural network architecture behind all modern LLMs. Introduced in the 2017 paper "Attention Is All You Need," it revolutionized AI by enabling models to process entire sequences in parallel using a mechanism called attention.
GPT, Claude, Llama, and virtually all current language models are built on transformer architecture.