Reinforcement Learning through Human Feedback - EXPLAINED! | RLHF

Поділитися
Вставка
  • Опубліковано 10 гру 2023
  • We talk about reinforcement learning through human feedback. ChatGPT among other applications makes use of this.
    ABOUT ME
    ⭕ Subscribe: ua-cam.com/users/CodeEmporiu...
    📚 Medium Blog: / dataemporium
    💻 Github: github.com/ajhalthor
    👔 LinkedIn: / ajay-halthor-477974bb
    PLAYLISTS FROM MY CHANNEL
    ⭕ Reinforcement Learning: • Reinforcement Learning...
    Natural Language Processing: • Natural Language Proce...
    ⭕ Transformers from Scratch: • Natural Language Proce...
    ⭕ ChatGPT Playlist: • ChatGPT
    ⭕ Convolutional Neural Networks: • Convolution Neural Net...
    ⭕ The Math You Should Know : • The Math You Should Know
    ⭕ Probability Theory for Machine Learning: • Probability Theory for...
    ⭕ Coding Machine Learning: • Code Machine Learning
    MATH COURSES (7 day free trial)
    📕 Mathematics for Machine Learning: imp.i384100.net/MathML
    📕 Calculus: imp.i384100.net/Calculus
    📕 Statistics for Data Science: imp.i384100.net/AdvancedStati...
    📕 Bayesian Statistics: imp.i384100.net/BayesianStati...
    📕 Linear Algebra: imp.i384100.net/LinearAlgebra
    📕 Probability: imp.i384100.net/Probability
    OTHER RELATED COURSES (7 day free trial)
    📕 ⭐ Deep Learning Specialization: imp.i384100.net/Deep-Learning
    📕 Python for Everybody: imp.i384100.net/python
    📕 MLOps Course: imp.i384100.net/MLOps
    📕 Natural Language Processing (NLP): imp.i384100.net/NLP
    📕 Machine Learning in Production: imp.i384100.net/MLProduction
    📕 Data Science Specialization: imp.i384100.net/DataScience
    📕 Tensorflow: imp.i384100.net/Tensorflow

КОМЕНТАРІ • 11

  • @RameshKumar-ng3nf
    @RameshKumar-ng3nf 6 днів тому +1

    Brilliant Bro 👌. Excellent explanation. I never understand RLHF reading so many books and notes. Your examples are GREAT & simple to understand 👌
    I am new to your channel and subscribed.

  • @manigoyal4872
    @manigoyal4872 5 місяців тому

    what about the generation of rewards, will there be another model to check the relativity of the answer and the precision of the answer, cause we have a lot of data

  • @theartofwar1750
    @theartofwar1750 2 місяці тому +1

    At 6:58, you have an error: PPO is not used to build the reward model.

  • @neetpride5919
    @neetpride5919 5 місяців тому +4

    Great video! I have a few questions:
    1) Why do we need to manually train the reward model with human feedback if the point is to evaluate responses of another pretrained model? Can't we just cut out the reward model altogether, rate the responses directly using human feedback to generate a loss value for each response, then backpropagate on that? Does it require less human input to train the reward model than to train the GPT model directly?
    2) When backpropagating the loss, do you need to do recurrent backpropagation for a number of steps that is the same as the length of the token output?
    3) Does the loss value apply equally to every token that is output? Seems like this would overly punish some words e.g. if the question starts with "why" it's likely the response is going to start with "because" regardless of what comes after. Does RLHF only work with sentence embeddings rather than word embeddings?

    • @0xabaki
      @0xabaki 3 місяці тому

      1) I think the point is to minimize the human feed back volume so humans just give enough responses to train a model for all future feedback. this way humans are not going to always have to give feedback, but instead will lay the basis, and probably come back to re-evaluate what the reward model is doing so it is still acting human
      (2) and (3) seem more specific to the architecture of chatGPT and neither PPO nor RLHF. I would look into the other GPT specific videos he made

  • @manigoyal4872
    @manigoyal4872 5 місяців тому

    Acts as a randomizing factor depending on whom you are getting feedback from

  • @sangeethashowrya0318
    @sangeethashowrya0318 Місяць тому

    Sir ,please make a video on function approximation in RL

  • @ayeshariaz3382
    @ayeshariaz3382 26 днів тому

    where to det your slides?

  • @0xabaki
    @0xabaki 3 місяці тому

    haha quiz time again:
    0) when the person knows me well
    1)D
    2)B if proper human feedback
    3)C

  • @manigoyal4872
    @manigoyal4872 5 місяців тому +1

    Aren't we users are the humans in feedback loop for openai

    • @akzytr
      @akzytr 5 місяців тому +2

      Yeah, however openai has the final say on what feedback goes through