Residual Networks and Skip Connections (DL 15)

Поділитися
Вставка
  • Опубліковано 22 жов 2022
  • Davidson CSC 381: Deep Learning, Fall 2022

КОМЕНТАРІ • 76

  • @alexei.domorev
    @alexei.domorev Рік тому +17

    ResNets are tricky to conceptualise as there are many nuances to consider. Dr Bryce, you have done a great job here offering such a brilliant explanation that is both logical and easy to follow. You definitely have a gift of explaining complex ideas. Thank you!

  • @vernonmascarenhas1801
    @vernonmascarenhas1801 Місяць тому +1

    I am writing a thesis on content-based image retrieval and I had to understand the ResNet architecture in-depth and by far this is the most transparent explanation ever!!

  • @anirudhsarma937
    @anirudhsarma937 Рік тому +17

    very very very good explanation. almost all explanations on this forget about the influence of random weights on the forward propagation and focus solely on the backward gradient multiplication. which is why i never understood why you needed to feed forward the input. thanks a lot

  • @ashishbhong5901
    @ashishbhong5901 6 місяців тому +9

    i have seen a lot of online lectures but you are the best for two reasons, the way you speak is not monotonous which give time to comprehend and process what your are explaining, and the second is the effort put in video editing to speed up when writing things down on board which doesn't break the flow of the lecture. Liked your video. Thanks🙂!

  • @AdityaSingh-qk4qe
    @AdityaSingh-qk4qe 3 місяці тому +4

    This is the clearest video that I've ever seen which explains the resnet for a layman, while at the same time conveying all the very important and relevant information related to resnet - I couldn't understand the paper - but with this video finally understood it - thanks a lot Professor Bryce - hope you create more such videos on deep learning

  • @thelife5628
    @thelife5628 Місяць тому +1

    Another example of a random youtuber with very less subscriber explaining a complex topic so brilliantly...
    Thankyou so much sir

  • @zhen_zhong
    @zhen_zhong Місяць тому

    This tutorial is so clear that I can follow along as a non-native English speaker. Thanks a lot!

  • @Engrbrain
    @Engrbrain Рік тому +9

    I am going to complete the entire playlist. Thanks, Bryce, you are a life saver

  • @alissabrave424
    @alissabrave424 28 днів тому +1

    Brilliant explanation! Thank you so much, Professor Bryce!

  • @lallama202
    @lallama202 4 місяці тому +1

    Love your explanation, very easy to understand the concept and the flow of the ResNet in 17 mins! Really appreciate it

  • @garydalley2349
    @garydalley2349 2 місяці тому

    Awesome explanation. Got me through a learning hurdle that several others could not.

  • @giordano_vitale
    @giordano_vitale 5 місяців тому

    Every single second of this video conveys an invaluable amount of information to properly understand these topics. Thanks a lot!

  • @raulpena9865
    @raulpena9865 10 місяців тому +1

    Thank you professor Bryce, Resnets where brilliantly explained by you. I am looking forward for new videos on more recent deep learning architectures!

  • @sanjeevjangra84
    @sanjeevjangra84 Місяць тому

    So clear and well explained. Thank you!

  • @subramanianiyer3300
    @subramanianiyer3300 5 місяців тому

    Thank you Prof. Bruce for explaining this thing with minimal complicated technicality

  • @user-ux2gz7sm6z
    @user-ux2gz7sm6z 11 місяців тому

    your explanation is clear and concise! Thank you so much

  • @rishabhagarwal4702
    @rishabhagarwal4702 Місяць тому

    Brilliant explanation, the 3D diagrams were excellent and I could understand some tricky concepts, thank you so much!

  • @minkijung3
    @minkijung3 10 місяців тому

    Amazing. Thanks a lot. Your explanation is so clear. Please keep making videos professor!🙏

  • @nilishamp245
    @nilishamp245 Рік тому

    you are brilliant!! Thank you for explaining this so well!!!!❤❤❤

  • @rabindhakal
    @rabindhakal 3 місяці тому

    You have my respect, Professor.

  • @genericchannel8589
    @genericchannel8589 11 місяців тому

    Awesome explanation!! Thank you for your effort :)

  • @nguyentranconghuy6965
    @nguyentranconghuy6965 24 дні тому

    nice explanation, thank you very much Professor Bryce

  • @kindness_mushroom
    @kindness_mushroom 5 місяців тому

    Thank you for the clear, concise, yet comprehensive explanation!

  • @user-ol1dx3nb3d
    @user-ol1dx3nb3d 5 місяців тому

    Brilliant explanation. Thank you!

  • @jonathanzkoch
    @jonathanzkoch Рік тому +1

    Great video on this, super informative.

  • @abdulsaboorkhan8337
    @abdulsaboorkhan8337 4 місяці тому

    Thank you so much Mr Bryce.

  • @nikhilthapa9300
    @nikhilthapa9300 8 місяців тому

    Your explanations are very clear and well structured. Please never stop teaching.

  • @lalop4258
    @lalop4258 Рік тому

    Excellent class! I watched many videos before I came to this video and none explained the concept of residual networks as clearly as you did.
    Greetings from México!

  • @business_central
    @business_central Рік тому +2

    Omg this is so helpful! Thank you so much !!!

  • @rohithr2071
    @rohithr2071 2 місяці тому

    Best explanation i came across resnet so far.

  • @schmiede1998
    @schmiede1998 8 місяців тому

    Thank you so much for this video!

  • @vaibhavnakrani2983
    @vaibhavnakrani2983 6 місяців тому

    awesome.Loved it clear and concise!

  • @rhysm8167
    @rhysm8167 6 місяців тому

    this was fantastic - thank you

  • @strictly-ai
    @strictly-ai 3 місяці тому +1

    Best explanation of resnet on the internet

  • @ali57555
    @ali57555 3 місяці тому

    Thank you very much for putting the time and effort. This is one of the best explanations I've seen (including US uni. professors)

  • @beatbustersindia3641
    @beatbustersindia3641 6 місяців тому +1

    Brilliant explanation.

  • @ArtJug
    @ArtJug Рік тому

    Wow This explanation is amazing. So clear! I saw some videos about resNets but none of them describes what skip connections mean inside, what is their inside structure and working logic. But your explanation gives me much more. You explained the way of thinking and inside structure and advantages. Wow!

  • @1991liuyangyang
    @1991liuyangyang 2 місяці тому

    great explanation, simple and straightforward.

  • @user-rb7vn3lt8t
    @user-rb7vn3lt8t 10 місяців тому

    Really Great explanation. Thanks Prof. ♥

  • @user-uq7kc2eb1i
    @user-uq7kc2eb1i 6 місяців тому +1

    Very nice video!

  • @user-hd3uv9ym7f
    @user-hd3uv9ym7f 7 місяців тому

    Thanks so much! very informative brief explanation

  • @user-yv3ib9so5d
    @user-yv3ib9so5d Місяць тому

    What an explanation

  • @Bachelorarbeit-op4he
    @Bachelorarbeit-op4he 6 місяців тому

    great explanation, thank you!

  • @MrMiguelDonate
    @MrMiguelDonate 2 місяці тому

    Brilliant explanation!!!

  • @shobhitsrivastava9112
    @shobhitsrivastava9112 10 місяців тому +1

    Until now, this is the best Residual Network tutorial I have found. As constructive feedback, I would like you to dive more deeply into how shape mismatches are handled because that part is not at par with the rest of the highly intuitive explanations of various things happening in a ResNet.

  • @user-bg2vs5kh6n
    @user-bg2vs5kh6n 5 місяців тому

    Great explanation, congrats.

  • @sam-vv6gl
    @sam-vv6gl 2 місяці тому

    thank you for the great explanation

  • @sajedehtalebi902
    @sajedehtalebi902 Рік тому

    It was clear and useful. Tnx a lot

  • @jiaqint961
    @jiaqint961 Місяць тому

    Thanks for your video.

  • @puyushgupta1768
    @puyushgupta1768 4 місяці тому

    16 golden minutes.❤

  • @adityabhatt4173
    @adityabhatt4173 5 місяців тому

    Great Explanation !!!!

  • @AymanFakri-ou8ro
    @AymanFakri-ou8ro 5 місяців тому

    very nice! thank you!

  • @sharmashikhashikha3
    @sharmashikhashikha3 11 місяців тому

    You are a star!

  • @bakhoinguyen5156
    @bakhoinguyen5156 7 місяців тому

    Thank you!!!

  • @SatyamAnand-ow4ub
    @SatyamAnand-ow4ub 11 місяців тому

    Awesome explanation

  • @charlesd4572
    @charlesd4572 Рік тому

    Superb!

  • @swethanandyala
    @swethanandyala 27 днів тому

    Amazing expalinaton. Thank you sir

  • @amitabhachakraborty497
    @amitabhachakraborty497 Рік тому

    Best Explanation

  • @happyvioloniste08
    @happyvioloniste08 8 місяців тому

    Thank you 👏👏

  • @zanzmeraankit4820
    @zanzmeraankit4820 9 місяців тому

    got a meaningfull insights from this video

  • @axe863
    @axe863 6 місяців тому

    Loss landscape looking super smooth .....

  • @lovenyajain6026
    @lovenyajain6026 5 місяців тому

    Waow. Thankyou

  • @sashimiPv
    @sashimiPv 5 місяців тому

    Prof. Bryce is the GOAT!

  • @paulocezarcunha
    @paulocezarcunha Місяць тому

    great!

  • @wouladjecabrelwen1006
    @wouladjecabrelwen1006 8 місяців тому

    Who is this teacher? Damn he is good. Thank you

  • @kkjun7157
    @kkjun7157 Рік тому +2

    This is such a clean and helpful video! Thank you very much! The only thing I still don't know is during the propagation, we now have two sets of gradients for each block? One for going through the layers, one for going around the layers, then how do we know which one to use to update the weights and bias?

    • @csprof
      @csprof  Рік тому +1

      Good question. For any given weight (or bias), its partial derivative expresses how it affects the loss along *all* paths. That means we have to use both the around- and through-paths to calculate the gradient. Luckily, this is easy to compute because the way to combine those paths is just to add up their contributions!

  • @newbie8051
    @newbie8051 11 місяців тому

    Coudn't understand how we can treat the shape-mismatch 13:40
    Great lecture nonetheless, thank you sir !! Understood what Residual Networks are 🙏

  • @kranthikumar9998
    @kranthikumar9998 10 місяців тому

    @csprof, By consistently including the original information alongside the features obtained from each residual block, are we inadvertently constraining our ResNet model to closely adhere to the input data, possibly leading to a form of over-memorization?

  • @wege8409
    @wege8409 3 місяці тому

    10:10
    Concerns: shape mis-match
    nervous sweating

  • @anirudhsarma937
    @anirudhsarma937 Рік тому

    Can you please talk about GANs and if possible stable diffusion

  • @mohammadyahya78
    @mohammadyahya78 Рік тому

    Thank you very much. I am not sure yet how residual block lead to faster gradient passing when the gradient has to go through both paths please? It means as I understand that this adds more overhead to compute the gradient. Please correct me if I am wrong. Also can you please add more how 1x1 reduce the depth or make a video please if possible? For example, I am not sure how the entire depth say of size 255 gives output to one neuron.

    • @csprof
      @csprof  Рік тому +3

      You're right that the residual connections mean more-complicated gradient calculations, which are therefore slower to compute for one pass. The sense in which it's faster is that it takes fewer training iterations for the network to learn something useful, because each update is more informative. Another way to think about it is that the function you're trying to learn with a residual architecture is simpler, so your random starting point is a lot more likely to be in a place where gradient descent can make rapid downhill progress.
      For the second part of your question, whenever we have 2D convolutions applied to a 3D tensor (whether the third dimension is color channels in the initial image, or different outputs from a preceding convolutional layer) we generally have a connection from *every* input along that third dimension to each of the neurons. If you do 1x1 convolution, each neuron gets input from a 1x1 patch in the first two dimensions, so the *only* thing it's doing is computing some function over all the third-dimension inputs. And then by choosing how many output channels you want, you can change the size on that dimension. For example, say that you have a 20x20x3 image. If you use 1x1 convolution with 8 output channels, then each neuron will get input from a 1x1x3 sub-image, but you'll have 8 different functions computed on that same patch, resulting in a 20x20x8 output.

  • @user-bw3bv1nz9l
    @user-bw3bv1nz9l Рік тому

    👍

  • @rayananwar8106
    @rayananwar8106 10 днів тому

    Do you mean that RESNET is just a skip connection not an individual network ?????????

  • @davar5029
    @davar5029 5 місяців тому

    Brilliant explanation. Thank you!