In this video, I’ll try to present a comprehensive study on Ashish Vaswani and his coauthors’ renowned paper, “attention is all you need”
This paper is a major turning point in deep learning research. The transformer architecture, which was introduced in this paper, is now used in a variety of state-of-the-art models in natural language processing and beyond.
📑 Chapters:
0:00 Abstract
0:39 Introduction
2:44 Model Details
3:20 Encoder
3:30 Input Embedding
5:22 Positional Encoding
11:05 Self-Attention
15:38 Multi-Head Attention
17:31 Add and Layer Normalization
20:38 Feed Forward NN
23:40 Decoder
23:44 Decoder in Training and Testing Phase
27:31 Masked Multi-Head Attention
30:03 Encoder-decoder Self-Attention
33:19 Results
35:37 Conclusion
📝 Link to the paper:
👥 Authors:
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kais
1 view
59
13
2 weeks ago 00:18:42 1
Lovell (Risad Remix) Alsa Remix - Deep House Music Best Popular Trend 2024
3 weeks ago 00:00:23 1
Built Different: Turning Heads in Downtown Miami
1 month ago 00:03:12 1
Slay in Style: YMDUCH Ruched High Split Maxi Dress – Chic, Flirty, and Made to Turn Heads! - YouTube
1 month ago 00:28:16 1
🌈 Discoveries of Great Tailors. Let’s Expose their Forbidden Tricks! (Part #35)
1 month ago 00:32:55 1
💥✅ Mysterious Sewing Techniques. You’ve Been Sewing Wrong All this Time!
1 month ago 00:01:37 1
THE MECHANIC 3 - First Trailer | Jason Statham, Gal Gadot, Liam Neeson