transformer neural network simply explained

Поділитися
Вставка
  • Опубліковано 10 чер 2024
  • #transformer #neuralnetwork #nlp #machinelearning
    Hello, in this video I share a simple step by step explanation on how Transformer Neural Network work.
    Timestamps
    0:00 - Intro
    0:45 - Understanding attention technique
    1:40 - Problem with sequence networks
    1:57 - Motivation for Transformer networks
    3:22 - Positional Encoding
    4:32 - Vanilla Attention
    5:22 - Self Attention
    6:52 - Multi-Head Attention
    7:50 - Residual Connection & Normalization
    9:53 - Masked Multi-Head Attention

КОМЕНТАРІ • 31

  • @mahibbalde1433
    @mahibbalde1433 13 годин тому

    Amazing visualization on this topic

  • @wp1300
    @wp1300 24 дні тому

    3:00 1 Input Embedding
    3:40 2 Positional Encoding
    4:08 3 Encoder Layer
    4:30 3a. Multi-head Attention
    7:51 3.a.ii Residual Addition, Layer Normalization & Pointwise Feed-Forward
    9:28 4 Output Embedding & Positional Encoding
    9:33 5 Decoder Layer
    10:07 5.a. Masked Multi-head Attention

  • @williamkariampuzha7951
    @williamkariampuzha7951 2 роки тому +1

    It finally cleared up what attention is for me! Thank you so much!!

  • @akrsrivastava
    @akrsrivastava 3 роки тому

    Great explaination.

  • @sarahsalt3689
    @sarahsalt3689 9 місяців тому

    Thank you for sharing!

  • @Amapramaadhy
    @Amapramaadhy 2 роки тому +1

    Loved your explanation. What an amazing mix of clear presentation and insightful visualization! Please keep it up.

    • @eniolaa
      @eniolaa  2 роки тому

      Thank you kindly!

  • @d3v487
    @d3v487 3 роки тому +2

    You teach very intuitively ❤️... Please make more videos on Deep leaning concepts , people definitely like.. : )

    • @eniolaa
      @eniolaa  2 роки тому

      Thank you, I will

  • @lmnefg121
    @lmnefg121 2 роки тому

    this is soooo helpful~! Just love it

    • @eniolaa
      @eniolaa  2 роки тому

      Glad it was helpful!

  • @sreebvmcreation9388
    @sreebvmcreation9388 Місяць тому

    Nice explanation

  • @TosinAlabi
    @TosinAlabi 3 роки тому

    Well done eniola, totally enjoyed your video.!!

    • @eniolaa
      @eniolaa  2 роки тому

      Thank you so much!

  • @RicardoSilva-lm4hb
    @RicardoSilva-lm4hb 2 роки тому

    you won a subscribe!

  • @maqboolurrahimkhan
    @maqboolurrahimkhan 3 роки тому

    Perfect thank you so much

    • @eniolaa
      @eniolaa  2 роки тому +1

      You're welcome 😊

  • @cedricmanouan2333
    @cedricmanouan2333 3 роки тому

    😭 i’m subscribing directly 😌

  • @Dhirajkumar-ls1ws
    @Dhirajkumar-ls1ws Рік тому

    Thanks

  • @hodgeshaywood3764
    @hodgeshaywood3764 Рік тому

    Hi Great explanation! Which software did you use for your animation and presentation?

  • @anshulkumarprajapati8346
    @anshulkumarprajapati8346 2 роки тому

    Very nice explained ma'am can you please make more videos on ttn to get more clear

  • @satyamgaba
    @satyamgaba Рік тому

    So crispy and precise.
    When you're stacking multiple decoders, will we stack masked multi head attention too?

    • @eniolaa
      @eniolaa  Рік тому

      yes we do, the masked multi head attention unit is a part of the decoder.

  • @coreygates1162
    @coreygates1162 Місяць тому

    Hello! wonderful explanation and video! I had one question, when going from the encoder layer to the embedding layer, what is the input to the output embedding layer? is it still "I am a student"? How are the French words incorporated?

    • @coreygates1162
      @coreygates1162 Місяць тому

      I guess like for the output embedding, would the look up table be the learned french translations?

  • @marcosmartinez9241
    @marcosmartinez9241 2 роки тому

    what is recived in the output embeding?

  • @aguntuk10
    @aguntuk10 Місяць тому

    ok How does query ,key and values calculated?

    • @rahulsbytes
      @rahulsbytes Місяць тому

      By the Query ,Key and Values matrices which is randomly assigned in intial state.

  • @mehmetaliozer2403
    @mehmetaliozer2403 2 роки тому

    I couldn't get the idea of masked multi-heead attention, any comment i would be appreciated.