Published On Oct 16, 2020
This is the first part of the Transformer Series. Here, I present an intuitive understanding of the self-attention mechanism in transformer networks.
[Paper] Attention Is All You Need: https://papers.nips.cc/paper/7181-att...
Other Resources:
Video Lecture on Word2Vec: • Lecture 2 | Word Vector Representatio...
Great article on Word2Vec: https://jalammar.github.io/illustrate...
show more