What Is Self-Supervised Learning and Why Care?

Поділитися
Вставка
  • Опубліковано 17 чер 2023
  • What is Self-Supervised Learning?
    “Self-Supervised methods […] are going to be the main method to train neural nets before we train them for difficult tasks”  -  Yann LeCun
    Let’s discover the fascinating machine learning domain of Self-Supervised Learning!
    It is one of the most important machine learning techniques in the current day and age.
    So, in this first video of the series we will motivate the idea of representation learning using Self-Supervised Learning!
    👾 Join the discord server!
    / discord
    Sources of clips used:
    • #86 - Prof. YANN LECUN...
    • Yann LeCun: Dark Matte...
    • Yann LeCun, Chief AI S...
    ⬇️ Follow me on my other socials and feel free to DM questions! ⬇️
    📲 Instagram: / m.m.boris
    ⚫⚪ Medium: / boris.meinardus
    🐦 Twitter: / borismeinardus
    #ai #education #SSL
  • Наука та технологія

КОМЕНТАРІ • 34

  • @borismeinardus
    @borismeinardus  11 місяців тому

    What Machine Learning Domain would you like to explore next?

    • @kiffbeatz
      @kiffbeatz 9 місяців тому

      Something that you find interesting.

  • @granttannert4457
    @granttannert4457 6 місяців тому

    Great Video! I'm taking a deep learning class right now, and this video really helped me understand the idea of self-supervised learning!

  • @maximinjoshua766
    @maximinjoshua766 5 місяців тому

    I was not able to get my head around pretext tasks for a long time and then I found this video. Thanks man.

  • @Opinionman2
    @Opinionman2 6 місяців тому +2

    Best explanation on topic i've seen or read. Thanks so much

    • @borismeinardus
      @borismeinardus  6 місяців тому

      💛 Really happy to hear that! ☺️

  • @engr_nadeemshah
    @engr_nadeemshah 6 місяців тому

    Well explained! without going into unnecessary complexities (details) like traditional professors in class or text books go.
    Thanks a lot. Continue making videos like these.

  • @abrarfahim2042
    @abrarfahim2042 7 місяців тому +3

    Thank you Boris. Such a clear Explanation!

    • @borismeinardus
      @borismeinardus  7 місяців тому +1

      I‘m really glad to hear it could help you! 🤩☺️

  • @InfraredScale
    @InfraredScale 9 місяців тому +7

    very good explanation, also love how Vsauce music pops in for a split second when you say "...or are they?" lol

    • @borismeinardus
      @borismeinardus  9 місяців тому +1

      thanks 🤩
      haha yeah, really appreciate people noticing the details! ☺️

  • @tfhfjhgkjhkjh
    @tfhfjhgkjhkjh 2 місяці тому

    You explain complex concepts really well! Thanks!

    • @borismeinardus
      @borismeinardus  2 місяці тому

      Thank you! Really glad you enjoyed it :)

  • @sergioorozco7331
    @sergioorozco7331 5 місяців тому

    Great video! Helps alot with my research

  • @r.s.e.9846
    @r.s.e.9846 7 місяців тому

    So if I stack an autoregressive LLM on top of a self-supervised model LLM (BERT), would it understand the context of the text better?

    • @borismeinardus
      @borismeinardus  7 місяців тому

      (Not thinking too long about the idea) I have two thoughts:
      - how do you want to use the BERT encoder? In the classical encoder-decoder transformer? Or do you want to use it in between the tokenizer and the actual input into the decoder?
      - If you pretrain the BERT model separately from the AR decoder and just use it as a frozen module, the decoder will not instantly understand the embeddings produced by the encoder. You will have to either fine-tune the encoder and decoder together, to align their representations, or freeze them and train some sort of connection module. This connection module can be a simple linear layer, an MLP, or something more complex.
      Of course, there are more things to consider, but those are the basic things you might need to think about.
      I hope this makes sense and helps!

  • @kiffbeatz
    @kiffbeatz 9 місяців тому +2

    Very good high-level explanation.

  • @athenaconstantinou5353
    @athenaconstantinou5353 5 місяців тому

    Amazing explaination, thank you

  • @anna040598
    @anna040598 10 місяців тому +1

    This helped me a lot 🤗

    • @borismeinardus
      @borismeinardus  10 місяців тому +1

      I‘m really happy to hear that ☺️

  • @binig.4591
    @binig.4591 10 місяців тому +1

    Nice. Can you do a video on how this might be related to reinforcement learning?

    • @borismeinardus
      @borismeinardus  10 місяців тому

      Do you have a specific paper in mind? :)

    • @binig.4591
      @binig.4591 10 місяців тому

      @@borismeinardus Not really. I was just wondering if ssl can be used to predict future rewards or sth...

    • @borismeinardus
      @borismeinardus  10 місяців тому

      @@binig.4591 I see. Well, in a sense, RL already is similar to SSL. The agent performs actions in it‘s environment and autonomously generates it’s own training data.
      I’m no expert in RL, but if I happen to learn more about it or find someone to talk to that is more educated in that direction, I will try to work on a video :)

  • @efesirin2870
    @efesirin2870 4 місяці тому

    Well explained thx

  • @user-us6sg8vn3o
    @user-us6sg8vn3o 2 місяці тому +4

    what is this background? headache

  • @pauceano945
    @pauceano945 3 місяці тому +1

    Please, remove the music, I had to stop 4 times to avoid stress... If your speech is interesting enough, I think it is, why confuse it and generate noise?