YouTube doesn't want you know this subscribers secret
Get Free YouTube Subscribers, Views and Likes

MIT 6.S191: Recurrent Neural Networks Transformers and Attention

Follow
Alexander Amini

MIT Introduction to Deep Learning 6.S191: Lecture 2
Recurrent Neural Networks
Lecturer: Ava Amini
** New 2024 Edition **

For all lectures, slides, and lab materials: http://introtodeeplearning.com

Lecture Outline
0:00​ Introduction
3:42​ Sequence modeling
5:30​ Neurons with recurrence
12:20 Recurrent neural networks
14:08 RNN intuition
17:14​ Unfolding RNNs
19:54 RNNs from scratch
22:41 Design criteria for sequential modeling
24:24 Word prediction example
31:50​ Backpropagation through time
33:40 Gradient issues
37:15​ Long short term memory (LSTM)
40:00​ RNN applications
44:00 Attention fundamentals
46:46 Intuition of attention
49:13 Attention and search relationship
51:22 Learning attention with neural networks
57:45 Scaling attention and applications
1:00:08 Summary
Subscribe to stay up to date with new deep learning lectures at MIT, or follow us @MITDeepLearning on Twitter and Instagram to stay fullyconnected!!

posted by ei1na23m