ATTENTION | An Image is Worth 16x16 Words | Vision Transformers (ViT) Explanation and Implementation

Поділитися
Вставка
  • Опубліковано 20 гру 2024

КОМЕНТАРІ • 12

  • @Explaining-AI
    @Explaining-AI  Рік тому

    *Github Code* - github.com/explainingai-code/VIT-Pytorch
    *Patch Embedding* - Vision Transformer (Part One) - ua-cam.com/video/lBicvB4iyYU/v-deo.html
    *Attention* in Vision Transformer (Part Two) - ua-cam.com/video/zT_el_cjiJw/v-deo.html
    *Implementing Vision Transformer* (Part Three) - ua-cam.com/video/G6_IA5vKXRI/v-deo.html

  • @sladewinter
    @sladewinter 9 місяців тому

    Best explanation of multi-head attention i have attended to! I already had a reasonable intuition but still gathered so much more, massive respect to your work 🙏

    • @Explaining-AI
      @Explaining-AI  9 місяців тому

      Thank you! Really glad that it was of any help.

  • @DrAIScience
    @DrAIScience 7 місяців тому

    Amazing explanation... i did not come accross the beautiful and easy explanation of transformers that seems extremely difficult... this channel deserves millions of subscribers 🎉

  • @shashankdevraj8633
    @shashankdevraj8633 3 місяці тому

    Sir Can you explain dual attention vision transformers (Davit)please

  • @sladewinter
    @sladewinter 5 місяців тому

    Would rearranging by heads before splitting into q, k , v cause any logical difference. Just means fewer lines of code, and operations, but mostly was just curious to verify as it felt same to me.

    • @Explaining-AI
      @Explaining-AI  5 місяців тому

      @sladewinter Yes, I agree with you, both seem the same to me as well.

  • @sebastiancavada
    @sebastiancavada 9 місяців тому

    Great content! This is helping a lot!! Keep it up :)

  • @muhammadawais2173
    @muhammadawais2173 9 місяців тому

    Helping, much appreciated. Sir how about self attention in image context

    • @Explaining-AI
      @Explaining-AI  9 місяців тому

      Thank you!. I didn't get what exactly you mean by self attention in image context. Could you clarify a bit.