Live -Transformers Indepth Architecture Understanding- Attention Is All You Need
Вставка
- Опубліковано 2 вер 2020
- All Credits To Jay Alammar
Reference Link: jalammar.github.io/illustrated...
Research Paper: papers.nips.cc/paper/7181-att...
youtube channel : • Jay's Visual Intro to AI
Please donate if you want to support the channel through GPay UPID,
Gpay: krishnaik06@okicici
Discord Server Link: / discord
Telegram link: t.me/joinchat/N77M7xRvYUd403D...
Please join as a member in my channel to get additional benefits like materials in Data Science, live streaming for Members and many more
/ @krishnaik06
Please do subscribe my other channel too
/ @krishnaikhindi
Connect with me here:
Twitter: / krishnaik06
Facebook: / krishnaik06
instagram: / krishnaik06
Your content is amazing. But, you unnecessarily repeat the same thing again and again. I am watching your video at a playback speed of 2x.
Any one in 2024, watching and learning from Krish
You really skip the maths too much.. you got to luk into that..
It looks like you are too confused in this session, not confident giving each answer of comments. But hats off to your good effort.
truely bro learn how to teach,,,please,,,,
Why 64? The correct answer is not a hyperparameter! It's because the dimensionality of data divided by number of head ~> so 512/8 (heads) =64
Sir, Please release the video of Bert. Eagerly waiting for it.
Nice. Too many adds
This video not in Urdu?
For those getting confused with 8 heads, all the words would be going to all the heads. It's not one word per head. The X matrix remains the same only the W matrix would change in case of multi-head attention.
@
I cannot express the amount of appreciation enough of your videos, especially NLP deep learning related topics! They are extremely helpful and so easy to understand from scratch! Thank you very much!
Krish is a hard working person, not for himself but for our country in the best way he could...We need more persons like him in our country
I am very new to the world of AI. I was looking for easy videos to teach me about the different models. I cannot imagine that I was totally enthralled by this video as long as you taught. You are a very good teacher. Thank you for publishing this video free. Thanks to Jay as well for simplifying such complex topic.
You are a really good teacher that always check your audiences weather they get the concept or not. Also, I appreciate your patience and the way you try to rephrase to have a better explanations.
For anyone having a doubt at
This might help the guy who asked why we take the square root and also for other aspirants :
Thanks for explaining Jay's blog. To add to the explanation at
You can skim through all the youtube videos explaining transformers, but nobody comes close to this video.
Krish, I really see the honesty in you man, lot of humility, very humble person. In the beginning of this video, you gave credit to Jay several times who created amazing blog for Transformers. I really liked that. Be like that.