Backpropagation Explained

Поділитися
Вставка
  • Опубліковано 3 лис 2024

КОМЕНТАРІ • 127

  • @RecursiveRuminations
    @RecursiveRuminations 6 років тому +45

    At 8:40 I think it is suppose to say something like
    Step 2: weights x inputs + bias
    step 3: calculate loss (mean squared error)
    step 4: find partial derivative for all weights
    step 5: calculate optimization direction on computational graph
    step 6: take step towards minima / optimized weights

  • @subhajitdas2784
    @subhajitdas2784 6 років тому +45

    Finally. It's 4 AM in India and I am out of bed waiting for this video.

    • @pyxelr
      @pyxelr 6 років тому +7

      2 AM in the UK, when I am writing this comment. Wanted to go to bed so much after work, but, when Siraj drops the video, I have to watch it in hope that the knowledge will optimize while sleeping.

    • @UnchartedWorlds
      @UnchartedWorlds 6 років тому +1

      Its 07:46 in Europe and I just woke up so I post this comment!

    • @JordanShackelford
      @JordanShackelford 6 років тому +1

      4:20 AM in Oklahoma, USA. 420 blaze it

    • @SirajRaval
      @SirajRaval  6 років тому +1

      bahaha

  • @MayankArora
    @MayankArora 6 років тому

    If you can’t explain it simply, you don’t understand it well enough. You, sir understand it way better than others 🔥🔥🔥

  • @muktasane6046
    @muktasane6046 3 роки тому

    This was SO helpful, concise video packed with so many concepts I needed to understand before I could understand the back propagation.

  • @dariocardajoli6831
    @dariocardajoli6831 6 років тому +13

    Its incredible to see how many effort you put in these videos, you deserves more subs, non just because your'e hella funny, but also because I can feel that your goal is not make views and money as the other youtubers, your'e teaching me machine learning but Im learning something more from you, respect from Italy 🇮🇹

    • @elgrimoriodelchamo1017
      @elgrimoriodelchamo1017 6 років тому +1

      I know, right? Siraj's passion for these incredible subjects is contagious! I love this guy. (Venezuela)

    • @TheTravisweb
      @TheTravisweb 6 років тому

      Dario Cardajoli he gets paid and he's professionally produced

  • @brendanhansknecht4650
    @brendanhansknecht4650 6 років тому +23

    8:35 I feel like this is a mistake in the video. Everything is Step 1 Random Initialization.

    • @SirajRaval
      @SirajRaval  6 років тому +2

      right it should say
      Step 2: weights x inputs + bias
      step 3: calculate loss (mean squared error)
      step 4: find partial derivative for all weights
      step 5: calculate optimization direction on computational graph
      step 6: take step towards minima / optimized weights

  • @FuZZbaLLbee
    @FuZZbaLLbee 6 років тому +2

    Maybe a quick mention of the chain rule would have also been nice, because that plays a big part in backpropigation.
    Cool video though, it is sort of a quick recap of the course I did last month.

  • @GuillaumeVerdonA
    @GuillaumeVerdonA 6 років тому

    Great video as always man! Been a fan since your early days, your videos got me into ML and now I have 2 papers on Quantum Machine Learning on the arXiv. Just wanted to let you know of your positive impact. Keep up the great work educating the world!

  • @dancingwithdestiny454
    @dancingwithdestiny454 6 років тому +2

    Rap album of ML in cheatsheet form would be a worthy theme track to overplay. As learning styles vary, sometimes going meta (which at the heart is to link use cases with concepts) can help the user differentiate the scope of tree and forest. Much love for the content and the wizard community!

    • @dancingwithdestiny454
      @dancingwithdestiny454 6 років тому

      I wrote that before seeing the intro, as my video was lagging. Trippy.

  • @akrylic_
    @akrylic_ 6 років тому +1

    Would love to see a video about using docker. Keep it up siraj!

  • @decode0126
    @decode0126 4 роки тому

    it really shaped my thoughts!!
    Thanks for this

  • @hugoropensourceai
    @hugoropensourceai 6 років тому

    Hi! Great video; but maybe you could have spent more time on the back propagation schema and explain the steps one by one, it seemed to me like the most interessant part, but I still haven't understand it even after the video...

  • @offchan
    @offchan 6 років тому +31

    7:54 I don't think that backpropagation is a rename of gradient descent.
    It's more like backprop followed by gradient descent.
    Backprop finds gradient and then gradient descent applies "weight = weight - gradient*learning_rate"
    So gradient descent only does the weight update. It doesn't need to find the gradient. The gradient is provided by backpropagation.

    • @SirajRaval
      @SirajRaval  6 років тому +4

      Gradient descent also includes derivation. See Andrew Ng's video on the topic, ua-cam.com/video/yFPLyDwVifc/v-deo.html The gradient is computed using calculus, specifically through partially deriving the weight variables of whatever function is being used to model some data. This process is part of the gradient descent process across all models, and it’s called back-propagation in the context of neural networks.

    • @offchan
      @offchan 6 років тому +7

      This is where it gets tricky. The definition of it is not so clear. I kind of see it like this:
      In order for gradient descent to work, it needs to call backpropagation to get the gradient.
      So you could say that the whole thing is gradient descent, that's right.
      But as far as I know, backpropagation does not include the weight update part. It only computes gradients.
      So gradient descent (in neural network context) is more like backprop plus weight update.
      Gradient descent needs learning rate as a hyperparameter but backprop doesn't.
      So I think they are not the same because of the weight update part.

    • @einemailadressenbesitzerei8816
      @einemailadressenbesitzerei8816 6 років тому +1

      please verify your knowledge before posting half-truth. MSE or Cross-Entropy is the Loss/Cost-function and not gradient descent, you use gradient descent on the cost function to update the weigths and biases. The chain rule is the chain rule and is not called backpropagation, it is used to calculate the gradient and is an essential part of backpropagation.

    • @einemailadressenbesitzerei8816
      @einemailadressenbesitzerei8816 6 років тому

      you said that the backpropagation is the chain rule derivative but that is not true. Backpropagation is an algorithm that is dependent on the chain-rule but it is not equivalent to the chain-rule. You said gradient descent/ascent is the cross entropy or MSE that is not true and misleading. The cross entropy or MSE are error functions.

    • @sidsr
      @sidsr 6 років тому

      Chanchana Sornsoontorn exactly, we can still use Advanced optimisers like lbfgs, and use backprop for getting the gradients alone, backprop and gradient descent are 2 different things

  • @amirabouamrane7151
    @amirabouamrane7151 4 роки тому

    so great explanation ,could give me a simple definition for the Back propagation

  • @bestintentions6089
    @bestintentions6089 6 років тому +1

    Finally your videos make sense, great work , your progress is my ease of learning . Looking forward to the next one. Aren't quantum computers of youre do optimization particularly well ?

  • @j1nchuika
    @j1nchuika 6 років тому

    I love your videos and have learnt a lot from them. How ever, I think you shouldn't use Python 2 examples anymore, since the industry is (finally) migrating from it to modern python. Great video!

  • @pallavirana3022
    @pallavirana3022 6 років тому +2

    Waking up to this

  • @aniketbanginwar
    @aniketbanginwar 6 років тому +2

    That Rap in the beginning though. Kind of summary of the whole video xD

  • @sethnuzum
    @sethnuzum 6 років тому

    That opening scene was beautiful❤️

  • @luis96xd
    @luis96xd 6 років тому +1

    This is an excelent video, well explained! Thanks!

  • @empiricistsacademy7181
    @empiricistsacademy7181 6 років тому

    Great video, I would of liked it more if there was actually an backprop example with a very simple neural net, but I guess there isn't enough space for that. Try to record the UCLA lecture if possible.

  •  6 років тому

    Iike the new setup of the video awesome!

  • @abhijeetghodgaonkar
    @abhijeetghodgaonkar 6 років тому

    Yay good video on Backpropagation!!!!

  • @ki-ka
    @ki-ka 6 років тому +1

    Excellent video on GD, but this video is being too general and covers the topic on the GD itself, rather than elaborating on application of the chain rule for derivation of the recursive formula for the gradient of a cost function wrt to the weigths of a feedforward NN.

  • @vikrambharadwaj6349
    @vikrambharadwaj6349 6 років тому

    Siraj a wavyy dude!
    Great video as always :)

  • @crisgonzales9441
    @crisgonzales9441 6 років тому

    It is a very interesting video congratulations, could you make an example of how to perform backpropagation with TensorFlow? of an example of mathematical calculation, as for example to predict the cost of a house with parameters of area location etc. Keep it up by sharing your knowledge. THANK YOU

  • @Nissearne12
    @Nissearne12 6 років тому

    I also think of use of gradient decent as find in the circuit (the mess of weights) where are the weight Knobs how make most ”bang for the buck” to tuning on to change the outcome situation

  • @aunabbas8582
    @aunabbas8582 6 років тому

    Hi... When do you come on live video? please mention the time.

  • @denisbaranoff
    @denisbaranoff 5 років тому +1

    Thank you for your excellent video and perfect english

  • @sm0na
    @sm0na 6 років тому

    Consistently impressed by your ability to collapse a half semesters worth of info into 10 minutes, as well as the great video production. Keep it up!!!

  • @aniketagarwal7821
    @aniketagarwal7821 6 років тому

    These videos are really great.

  • @vaibhavgeek
    @vaibhavgeek 6 років тому +1

    Great video. Love from India

  • @prasanthsrinivasan7709
    @prasanthsrinivasan7709 6 років тому

    Hi Siraj, Could you please tell how to do Data Augmentation for numerical data (I mean not for images)
    I created a ANN with backpropagation algorithm. Now I wish to do Data Augmentation.

  • @nicolasjomolca2205
    @nicolasjomolca2205 5 років тому

    ahhhhh dude! i knew you lived in LA! I had a feeling. You have the vibe!

  • @frankbraker
    @frankbraker 6 років тому

    Computing the error for the last layer of the neural network makes perfect sense, but calculating error on earlier layers... nope. I.e. it would be really nice to see the values of (for example) an XOR gate being trained from random weights on the minimal network that can implement an XOR function, using backpropagation.

  • @ahmedhussain1905
    @ahmedhussain1905 6 років тому

    You should make a video on how to convince VC and other investors to invest in our AI project.

  • @goldpuneetsingh2489
    @goldpuneetsingh2489 6 років тому +2

    Hey siraj!!.. you have a spelling mistake of usain (osain) in your UA-cam about section😊
    Btw thanks for your videos... #lovefromIndia

  • @AbhishekKumar-mq1tt
    @AbhishekKumar-mq1tt 6 років тому

    Thank you for this awesome video

  • @jtekmmx
    @jtekmmx 6 років тому +2

    Did you get a chance to record the UCLA lecture?

  • @RaviPrakash-dz9fm
    @RaviPrakash-dz9fm 6 років тому

    Could you make a playlist of ML models built from scratch without using pre-trained libs.
    That would be of great help.

  • @shivamrathore1887
    @shivamrathore1887 6 років тому +7

    4:00 AM in India and Siraj is uploading video..

    • @SirajRaval
      @SirajRaval  6 років тому +3

      i will do a better time next time

    • @mukeshbasira8198
      @mukeshbasira8198 6 років тому +1

      its okke :) we love that .. coders == owl :P

  • @mohdhammadsiddiqui7598
    @mohdhammadsiddiqui7598 6 років тому

    yr videos are awesome man keep uploading more,yr great man

  • @jalshah9108
    @jalshah9108 6 років тому +1

    A video of detailed explanation of Math behind populer ML Algorithms...

    • @youngtaiahn
      @youngtaiahn 6 років тому +1

      Look up "The Elements of Statistical Learning". PDF book is available for free online.

  • @jubertvera6496
    @jubertvera6496 6 років тому

    Can we make generative desing with it?

  • @Nissearne12
    @Nissearne12 6 років тому

    Ahh! so Momentum is used to come over a smaller hill to then go to global minimum instead of local minimum, am I correct. (Preventing overtranig ??)

  • @algaida94
    @algaida94 6 років тому

    One video of Gradient Descent Optimization Algorithms??

  • @robienob89
    @robienob89 6 років тому

    Can you explain back propogation through time please?

  • @OrcaChess
    @OrcaChess 6 років тому

    Fresh Explanation!

  • @bharath5673__
    @bharath5673__ 6 років тому

    Bob I'm ur fan 💕...I'm a front-end developer, but I love AI, n can u help Me Plzzz , What is the EXACT difference between machine learning and deep learning?.. (The cutting-edge tip like that seperates ML n DL.. How to make out).. And can v learn deep learning without learning ML??.. What are the functions for ML and other functions for DL?

    • @AkshayKumarTK
      @AkshayKumarTK 6 років тому +1

      Bharath kumar As per my knowledge, you only needs basics of machine learning and a bit of maths to start on deep learning. It is a subset and most popular part of Machine Learning

    • @bharath5673__
      @bharath5673__ 6 років тому

      hy,, thnx bro. but. im already halfway into neural-networks..
      like linear-algebra, logic gates,vector,matrixs, inputs,weights, tendons,networks nll but the problem is, im not getting how to fetch up n apply which functions for wat type of problems :(,,, like why this linear regressions, y sigmoid,y this tanH, y ReLu where to apply this nll // full on confusions bro :'( .

  • @sifiso5055
    @sifiso5055 6 років тому +1

    Siraj remember a long time ago: 'Backpropagate to update weights'

  • @emmanuelogbewe2334
    @emmanuelogbewe2334 6 років тому

    Awesome ! Appreciate it!

  • @deniscandido4116
    @deniscandido4116 6 років тому

    At 8:35 there's written 'step 1 random initialization' on every box. Was supposed to be like this??

  • @krishna7991
    @krishna7991 5 років тому +12

    You've discussed gradient descent more here than backpropagation. To be fair, there's almost nothing about backpropagation.

    • @ronmedina429
      @ronmedina429 5 років тому +1

      cause he's a fraud

    • @vcktorn7204
      @vcktorn7204 3 роки тому

      There’s a reason I’m telling all my friends to block him. He’s more interested in blocks than actually learning the concepts. He’s a fraud only interested in profits.

  • @Naufal331
    @Naufal331 6 років тому

    Yo siraj do you possess any information on the new AI algorithm "Augmented Random Search" (ARS)? If you do please make a video on it ;)

  • @scott7948
    @scott7948 6 років тому

    How is back propagation different to gradient descent?

  • @adenpower249
    @adenpower249 6 років тому

    Editing in this video was a little bit sloppy, a few grammatical and spelling mistakes. However it was quite a well made video overall.

  • @vedavivek1486
    @vedavivek1486 6 років тому

    Siraj, you are awesome :)

  • @infratechethiopia
    @infratechethiopia 6 років тому

    I'm going to watch this video back- propagating

  • @joesiu4972
    @joesiu4972 6 років тому

    Let's gooooo

  • @aguy3203
    @aguy3203 6 років тому

    Always good to review

  • @mikeg9b
    @mikeg9b 6 років тому

    6:48 "It's represented by this little squiggly character." Hahaha

  • @manuelgoncalves8206
    @manuelgoncalves8206 6 років тому

    F5 F5 F5 since the announcement on Twitter :-)

  • @FerroNeoBoron
    @FerroNeoBoron 6 років тому +1

    3:31 how many people caught that PPAP meme I wonder?

  • @finnwilliams1508
    @finnwilliams1508 6 років тому

    Great video. Cover backpropagation through time (context of RNN) in detail?

  • @kitgary
    @kitgary 6 років тому

    The maths is so difficult and I can't understand! Can anyone reference an article of step by step derivation of backpropagation?

  • @snowball7720
    @snowball7720 6 років тому

    Nice

  • @jukebox1209
    @jukebox1209 6 років тому

    Nice!

  • @aradarbel4579
    @aradarbel4579 6 років тому

    tho I know how back prop works- YAAAAY!!!!! love you channel! #BeingCoolWhileProgramming

  • @omarcusmafait7202
    @omarcusmafait7202 6 років тому

    well done ;)

  • @SouthernHerdsman
    @SouthernHerdsman 6 років тому

    Did I hear you say "derive the meaning of life?"...

  • @jinxblaze
    @jinxblaze 6 років тому +3

    4:11 AM squad

    • @MrYashpaunikar
      @MrYashpaunikar 6 років тому

      Ajinkya Jumbad itachi is alert all the time!!!!

  • @enriquebozzadutra9402
    @enriquebozzadutra9402 6 років тому

    I thought it was supposed to be a backpropagation explanation, not a full machine learning course.

  • @Murderface666
    @Murderface666 5 років тому

    Not to be a weenie, but this topic of NN on youtube is much like the plethora of many programming channels that talks like they are regurgitating from a book and less like "let me show you how you actually apply this in code so that you can you use it in your own projects." And most people don't read scientific notation

  • @CryptoPendu
    @CryptoPendu 6 років тому

    1st Bro Jai HInd #swachhcrypto

  • @faisal6491
    @faisal6491 6 років тому

    no programer to full ai in xx days

  • @simhadrimallikarjun1635
    @simhadrimallikarjun1635 6 років тому

    Upto 4 minutes i didn't found backpropagation

  • @amitbaderia4194
    @amitbaderia4194 6 років тому

    i clicked video to understand the concept, i am more confused now.

  • @sxtian3324
    @sxtian3324 6 років тому +1

    明明适合当歌手,非要做技术

  • @geekyprogrammer4831
    @geekyprogrammer4831 6 років тому

    You look like Mika Singh

  • @messina8933
    @messina8933 6 років тому

    Just thought I would leave this here, optimization doesn't always end well or as we might expect: ua-cam.com/video/tcdVC4e6EV4/v-deo.html

  • @volfegan
    @volfegan 6 років тому +1

    Machine Learning is so ripoff of Linear programming.

  • @krishnendusengupta6158
    @krishnendusengupta6158 6 років тому

    This video lacked Siraj-ness...

  • @codethings271
    @codethings271 6 років тому

    Why are u repeating it again and again?

    • @SirajRaval
      @SirajRaval  6 років тому +1

      just wait till next week, there is a bigger audience and im getting them up to speed

  • @kislaykunal8921
    @kislaykunal8921 6 років тому

    to those who think gradient descent and backprop r different things, guys don't embarrass yourselves by writing that. Also it's just a got damn partial derivatives for christ sake.

  • @swrhsn6300
    @swrhsn6300 3 роки тому

    Like the way he pretends

  • @shreyanmehta6999
    @shreyanmehta6999 6 років тому

    Haha pen apple pineapple pen😹😹 if you got it

  • @hhhgggds
    @hhhgggds 6 років тому

    Why do i think that he is actually clueless about all of this and is just there to present it?

    • @SirajRaval
      @SirajRaval  6 років тому

      Tell me why

    • @hhhgggds
      @hhhgggds 6 років тому

      I commented on the wrong video that poped up after the one i watched. In any case that was the impression i got after watching it. Equations just popping on the screen without in depth explanation. Is the purpose of videos to teach people or to just let them know that this stuff exists ?:)
      Edit : i was referring to machine learning video , your latest

    • @hhhgggds
      @hhhgggds 6 років тому

      You tell me if im wrong

  • @g_wzrd_9265
    @g_wzrd_9265 6 років тому

    First one here 😎

  • @Naufal331
    @Naufal331 6 років тому

    Yo siraj do you possess any information on the new AI algorithm "Augmented Random Search" (ARS)? If you do please make a video on it ;)