0 votes
in Deep Learning by
What is self-attention, and how does it work in transformers?

1 Answer

0 votes
by

Before the introduction of the Transformer model, the use of attention for neural machine translation was implemented by RNN-based encoder-decoder architectures. The Transformer model revolutionized the implementation of attention by dispensing with recurrence and convolutions and, alternatively, relying solely on a self-attention mechanism. 

We will first focus on the Transformer attention mechanism in this tutorial and subsequently review the Transformer model in a separate one. 

In this tutorial, you will discover the Transformer attention mechanism for neural machine translation. 

After completing this tutorial, you will know:

  1. How the Transformer attention differed from its predecessors
  2. How the Transformer computes a scaled-dot product attention
  3. How the Transformer computes multi-head attention
...