Английская Википедия:Attention Is All You Need

Материал из Онлайн справочника
Перейти к навигацииПерейти к поиску

Шаблон:Short description Шаблон:Use dmy dates

Файл:The-Transformer-model-architecture.png
An illustration of main components of the transformer model from the paper

"Attention Is All You Need" is a 2017 research paper by Google.[1] Authored by eight scientists, it was responsible for expanding 2014 attention mechanisms proposed by Bahdanau et. al. into a new deep learning architecture known as the transformer. The paper is considered the founding document for modern artificial intelligence, as transformers became the main architecture of large language models.[2][3] At the time, the focus of the research was on improving Seq2seq techniques for machine translation, but even in their paper the authors saw the potential for other tasks like question answering and for what is now called multimodal Generative AI.

Шаблон:As of the paper was cited more than 80,000 times. All eight authors left Google.[4]

References

Шаблон:Reflist

Шаблон:Google LLC Шаблон:Authority control


Шаблон:Ai-stub

  1. Шаблон:Cite journal
  2. Ошибка цитирования Неверный тег <ref>; для сносок Forbes не указан текст
  3. Ошибка цитирования Неверный тег <ref>; для сносок Financial Times не указан текст
  4. Шаблон:Cite news