Pytorch Transformers from Scratch (Attention is all you need)
Aladdin Persson Aladdin Persson
73.9K subscribers
283,832 views
0

 Published On Jun 22, 2020

In this video we read the original transformer paper "Attention is all you need" and implement it from scratch!

Attention is all you need paper:
https://arxiv.org/abs/1706.03762

A good blogpost on Transformers:
http://www.peterbloem.nl/blog/transfo...

❤️ Support the channel ❤️
   / @aladdinpersson  

Paid Courses I recommend for learning (affiliate links, no extra cost for you):
⭐ Machine Learning Specialization https://bit.ly/3hjTBBt
⭐ Deep Learning Specialization https://bit.ly/3YcUkoI
📘 MLOps Specialization http://bit.ly/3wibaWy
📘 GAN Specialization https://bit.ly/3FmnZDl
📘 NLP Specialization http://bit.ly/3GXoQuP

✨ Free Resources that are great:
NLP: https://web.stanford.edu/class/cs224n/
CV: http://cs231n.stanford.edu/
Deployment: https://fullstackdeeplearning.com/
FastAI: https://www.fast.ai/

💻 My Deep Learning Setup and Recording Setup:
https://www.amazon.com/shop/aladdinpe...

GitHub Repository:
https://github.com/aladdinpersson/Mac...

✅ One-Time Donations:
Paypal: https://bit.ly/3buoRYH

▶️ You Can Connect with me on:
Twitter -   / aladdinpersson  
LinkedIn -   / aladdin-persson-a95384153  
Github - https://github.com/aladdinpersson

OUTLINE:
0:00 - Introduction
0:54 - Paper Review
11:20 - Attention Mechanism
27:00 - TransformerBlock
32:18 - Encoder
38:20 - DecoderBlock
42:00 - Decoder
46:55 - Putting it togethor to form The Transformer
52:45 - A Small Example
54:25 - Fixing Errors
56:44 - Ending

show more

Share/Embed