How to Use Learning Rate Scheduling for Neural Network Training

Поділитися
Вставка
  • Опубліковано 25 сер 2024
  • Neural Networks and neural network based architecturres are powerful models that can deal with abstract problems but they are known for taking a long time to train. In this video, we learn about learning rate scheduling that can really speed up the training process.
    Previous lesson: • Pruning a neural Netwo...
    Next lesson: • How to Make Neural Net...
    📙 Here is a lesson notes booklet that summarizes everything you learn in this course in diagrams and visualizations. You can get it here 👉 misraturp.gumr...
    👩‍💻 You can get access to all the code I develop in this course here: github.com/mis...
    ❓To get the most out of the course, don't forget to answer the end of module questions:
    fishy-dessert-...
    👉 You can find the answers here:
    fishy-dessert-...
    RESOURCES:
    🏃‍♀️ Data Science Kick-starter mini-course: www.misraturp....
    🐼 Pandas cheat sheet: misraturp.gumr...
    📥 Streamlit template (updated in 2023, now for $5): misraturp.gumr...
    📝 NNs hyperparameters cheat sheet: www.misraturp....
    📙 Fundamentals of Deep Learning in 25 pages: misraturp.gumr...
    COURSES:
    👩‍💻 Hands-on Data Science: Complete your first portfolio project: www.misraturp....
    🌎 Website - misraturp.com/
    🐥 Twitter - / misraturp

КОМЕНТАРІ • 7

  • @lakeguy65616
    @lakeguy65616 Рік тому +3

    A learning rate that is too small can get trapped in a local minima with no escape. A learning rate that is too large will never find the global minima because it will keep "stepping" over it.

  • @rahiiqbal1294
    @rahiiqbal1294 5 місяців тому

    Thank you, these videos are very helpful :)

  • @user-id6fq5ym1v
    @user-id6fq5ym1v 5 місяців тому

    what a helpful videos!Thank you! :)

  • @alexandergraf8855
    @alexandergraf8855 Рік тому +1

    Been tinkering for a day (and night lol)....
    Your training set has only 10 items and 10 classifiers assigned to it.
    You go through the list, make a prediction for every element. Everything is predicted wrong, because your network is fresh off the shelf and unknowing what you want from it.
    If an item is predicted wrong, backprop only once and then go to the next item. Skip items that have been predicted correctly.
    For every item in the dataset there is a count that's being incremented everytime you have to backprop and the learning rate will be count * 0.9 (really). For correct predictions count will reset to 1.
    So....items with low success rate will be trained everytime you loop though your data but at a monster rate, while successful items or items that have to be trained only occationally will have lower learning rate.
    The overall process is done, if no backprop happened inside the loop. 😘 And this is my criteria, it has to be incredible fast learning at high framerate while doing all the others stuff like extracting unknown objects from my desktop screen, all things that Johnny-Boy has never seen before.😁
    Training should not be the bottleneck of AGI, I mean....am I really so clever (lol)? Why not running me on a 500MHz Computer? It should be doable 😎

    • @alexandergraf8855
      @alexandergraf8855 Рік тому

      Another phenomenon about constant learning rate is that the last elements in the list finish training first because the network seems to forget about the beginning of the list. Learn something and forget about the others items! That's not what I want!
      It seems that "dynamic monster learning" where LR is constantly fluctuating doesn't have so much trouble about forgetting, so it's more independent of the ordering of the list!

    • @user-tz6rj8te3v
      @user-tz6rj8te3v 19 днів тому

      by if you have stick to learning you probably have discovered the answer huh 😁

  • @AguinaldoMulondeMulonde
    @AguinaldoMulondeMulonde 6 місяців тому

    you are so pretty..