Tranformer

The transformer uses self-attention mechanisms to focus on different parts of the input text during each processing step. This allows the model to consider the context of each word in relation to other words in the input text.

image

Ref: https://www.youtube.com/watch?v=wjZofJX0v4M image

Refer: