Live -Transformers Indepth Architecture Understanding- Attention Is All You Need

Поділитися
Вставка
  • Опубліковано 2 вер 2020
  • All Credits To Jay Alammar
    Reference Link: jalammar.github.io/illustrated...
    Research Paper: papers.nips.cc/paper/7181-att...
    youtube channel : • Jay's Visual Intro to AI
    Please donate if you want to support the channel through GPay UPID,
    Gpay: krishnaik06@okicici
    Discord Server Link: / discord
    Telegram link: t.me/joinchat/N77M7xRvYUd403D...
    Please join as a member in my channel to get additional benefits like materials in Data Science, live streaming for Members and many more
    / @krishnaik06
    Please do subscribe my other channel too
    / @krishnaikhindi
    Connect with me here:
    Twitter: / krishnaik06
    Facebook: / krishnaik06
    instagram: / krishnaik06

КОМЕНТАРІ • 228

  • @aditiseetha1
    @aditiseetha1 Рік тому +15

    Your content is amazing. But, you unnecessarily repeat the same thing again and again. I am watching your video at a playback speed of 2x.

  • @anoopitiss

    Any one in 2024, watching and learning from Krish

  • @paneercheeseparatha

    You really skip the maths too much.. you got to luk into that..

  • @vivekyadav-zl5dl

    It looks like you are too confused in this session, not confident giving each answer of comments. But hats off to your good effort.

  • @vamshi5745

    truely bro learn how to teach,,,please,,,,

  • @Mr.AIFella

    Why 64? The correct answer is not a hyperparameter! It's because the dimensionality of data divided by number of head ~> so 512/8 (heads) =64

  • @jeeveshkataria6439
    @jeeveshkataria6439 3 роки тому +22

    Sir, Please release the video of Bert. Eagerly waiting for it.

  • @pavangoyal6840

    Nice. Too many adds

  • @ibrahimnadeem1064

    This video not in Urdu?

  • @faezakamran3793
    @faezakamran3793 Рік тому +3

    For those getting confused with 8 heads, all the words would be going to all the heads. It's not one word per head. The X matrix remains the same only the W matrix would change in case of multi-head attention.

  • @mohammadmasum4483
    @mohammadmasum4483 Рік тому +12

    @

  • @dandyyu0220
    @dandyyu0220 2 роки тому +7

    I cannot express the amount of appreciation enough of your videos, especially NLP deep learning related topics! They are extremely helpful and so easy to understand from scratch! Thank you very much!

  • @story_teller_1987
    @story_teller_1987 3 роки тому +13

    Krish is a hard working person, not for himself but for our country in the best way he could...We need more persons like him in our country

  • @shrikanyaghatak

    I am very new to the world of AI. I was looking for easy videos to teach me about the different models. I cannot imagine that I was totally enthralled by this video as long as you taught. You are a very good teacher. Thank you for publishing this video free. Thanks to Jay as well for simplifying such complex topic.

  • @apppurchaser2268
    @apppurchaser2268 Рік тому

    You are a really good teacher that always check your audiences weather they get the concept or not. Also, I appreciate your patience and the way you try to rephrase to have a better explanations.

  • @suddhasatwaAtGoogle
    @suddhasatwaAtGoogle 2 роки тому +37

    For anyone having a doubt at

  • @roshankumargupta46
    @roshankumargupta46 3 роки тому +43

    This might help the guy who asked why we take the square root and also for other aspirants :

  • @harshitjain4923
    @harshitjain4923 3 роки тому +13

    Thanks for explaining Jay's blog. To add to the explanation at

  • @anusikhpanda9816
    @anusikhpanda9816 3 роки тому +27

    You can skim through all the youtube videos explaining transformers, but nobody comes close to this video.

  • @ss-dy1tw
    @ss-dy1tw 3 роки тому +1

    Krish, I really see the honesty in you man, lot of humility, very humble person. In the beginning of this video, you gave credit to Jay several times who created amazing blog for Transformers. I really liked that. Be like that.