Transformer neural network architecture powering large language models like ChatGPT, featuring stacked encoder-decoder layers, self-attention mechanisms, token embeddings, and positional encoding. Enables deep understanding of input context and generation of accurate, human-like responses through masked and multi-head attention strategies.