Pytorch Transformers from Scratch (Attention is all you need)

In this video we read the original transformer paper “Attention is all you need” and implement it from scratch!

Github Repository:
https://github.com/aladdinpersson/Machine-Learning-Collection

Attention is all you need paper:
https://arxiv.org/abs/1706.03762

OUTLINE:

  • 0:00 - Introduction
  • 0:54 - Paper Review
  • 11:20 - Attention Mechanism
  • 27:00 - TransformerBlock
  • 32:18 - Encoder
  • 38:20 - DecoderBlock
  • 42:00 - Decoder
  • 46:55 - Putting it togethor to form The Transformer
  • 52:45 - A Small Example
  • 54:25 - Fixing Errors
  • 56:44 - Ending

#pytorch #deep-learning #machine-learning #artificial-intelligence #developer

Pytorch Transformers from Scratch (Attention is all you need)
34.65 GEEK