Self Attention in Transformer Neural Networks (with Code!)

Let’s understand the intuition, math and code of Self Attention in Transformer Neural Networks ABOUT ME ⭕ Subscribe: 📚 Medium Blog: @dataemporium 💻 Github: 👔 LinkedIn: RESOURCES [ 1🔎] Code for video: [2 🔎] Transformer Main Paper: [3 🔎] Bidirectional RNN Paper: Recurrent Neural PLAYLISTS FROM MY CHANNEL ⭕ ChatGPT Playlist of all other videos: ⭕ Transformer Neural Networks: ⭕ Convolutional Neural Networks: ⭕ The Math You Should Know : ⭕ Probability Theory for Machine Learning: ⭕ Coding Machine Learning: MATH COURSES (7 day free trial) 📕 Mathematics for Machine Learning: 📕 Calculus: 📕 Statistics for Data Science: 📕 Bayesian Statistics: 📕 Linear Algebra: 📕 Probability: OTHER RELATED COURSES (7 day free trial) 📕 ⭐ Deep Learning Specialization: 📕 Python for Everybody: 📕 MLOps Course: 📕 Natural Language Processing (NLP): 📕 Machine Learning in Production: 📕 Data Science Specialization: 📕 Tensorflow: TIMSTAMPS 0:00 Introduction 0:34 Recurrent Neural Networks Disadvantages 2:12 Motivating Self Attention 3:34 Transformer Overview 7:03 Self Attention in Transformers 7:32 Coding Self Attetion
Back to Top