Translations:Attention Is All You Need/26/en
The paper's title — "attention Is All You Need" — became one of the most recognizable phrases in machine learning, and the architecture it introduced has been called one of the most influential contributions to artificial intelligence in the 2010s. As of 2026, the transformer remains the dominant architecture for large-scale neural network models across modalities.