Coding a Paper - Ep. 4: Adding in Position Embeddings

Поділитися
Вставка
  • Опубліковано 26 жов 2024

КОМЕНТАРІ • 6

  • @imotvoksim
    @imotvoksim 8 місяців тому +1

    Just commenting to say that this series is appreciated and I took the weekend to follow along! Time well spent. Hopefully continue next weekend

  • @wilfredomartel7781
    @wilfredomartel7781 Місяць тому

    Amazing work!

  • @imotvoksim
    @imotvoksim 8 місяців тому +1

    At the end when you are bringing out the heads dimension out of the resulting relative_position_values matrix shouldn't the operation be relative_position_values.transpose(1, -1).transpose(0, 1).unsqueeze(0) so we end up with (batch, heads, sequence, context) instead of (batch, heads, context, sequence)?

    • @ChrisMcCormickAI
      @ChrisMcCormickAI  8 місяців тому +2

      Good catch! Yes, the context and sequence are in the wrong order (and I've ignored the batch) - your solution puts things in the correct order. We switch to einops later as we put everything together so this will be corrected in later videos. Glad you're enjoying the series :)

  • @mm100latests5
    @mm100latests5 8 місяців тому +1

    awesome!

  • @Tripp111
    @Tripp111 8 місяців тому +1

    Thank you.