YouTube doesn't want you know this subscribers secret
Get Free YouTube Subscribers, Views and Likes

Attention for Neural Networks Clearly Explained!!!

Follow
StatQuest with Josh Starmer

Attention is one of the most important concepts behind Transformers and Large Language Models, like ChatGPT. However, it's not that complicated. In this StatQuest, we add Attention to a basic SequencetoSequence (Seq2Seq or EncoderDecoder) model and walk through how it works and is calculated, one step at a time. BAM!!!

NOTE: This StatQuest is based on two manuscripts. 1) The manuscript that originally introduced Attention to EncoderDecoder Models: Neural Machine Translation by Jointly Learning to Align and Translate: https://arxiv.org/abs/1409.0473 and 2) The manuscript that first used the DotProduct similarity for Attention in a similar context: Effective Approaches to Attentionbased Neural Machine Translation https://arxiv.org/abs/1508.04025

NOTE: This StatQuest assumes that you are already familiar with basic EncoderDecoder neural networks. If not, check out the 'Quest:    • SequencetoSequence (seq2seq) Encode...  

If you'd like to support StatQuest, please consider...
Patreon:   / statquest  
...or...
YouTube Membership:    / @statquest  

...buying my book, a study guide, a tshirt or hoodie, or a song from the StatQuest store...
https://statquest.org/statqueststore/

...or just donating to StatQuest!
https://www.paypal.me/statquest

Lastly, if you want to keep up with me as I research and create new StatQuests, follow me on twitter:
  / joshuastarmer  

0:00 Awesome song and introduction
3:14 The Main Idea of Attention
5:34 A worked out example of Attention
10:18 The Dot Product Similarity
11:52 Using similarity scores to calculate Attention values
13:27 Using Attention values to predict an output word
14:22 Summary of Attention

#StatQuest #neuralnetwork #attention

posted by TubErrobeBory8y