The StatQuest Introduction to PyTorch

Поділитися
Вставка
  • Опубліковано 7 лют 2025

КОМЕНТАРІ • 369

  • @statquest
    @statquest  2 роки тому +27

    The code demonstrated this video can be downloaded here: lightning.ai/lightning-ai/studios/statquest-introduction-to-coding-neural-networks-with-pytorch?view=public§ion=all
    To learn more about Lightning: lightning.ai/
    This StatQuest assumes that you are already familiar with...
    Neural Networks: ua-cam.com/video/CqOfi41LfDw/v-deo.html
    Backpropagation: ua-cam.com/video/IN2XmBhILt4/v-deo.html
    The ReLU Activation Function: ua-cam.com/video/68BZ5f7P94E/v-deo.html
    Tensors: ua-cam.com/video/L35fFDpwIM4/v-deo.html
    To install PyTorch see: pytorch.org/get-started/locally/
    To install matplotlib, see: matplotlib.org/stable/users/getting_started/
    To install seaborn, see: seaborn.pydata.org/installing.html
    Support StatQuest by buying my books The StatQuest Illustrated Guide to Machine Learning, The StatQuest Illustrated Guide to Neural Networks and AI, or a Study Guide or Merch!!! statquest.org/statquest-store/

    • @yongjiewang9686
      @yongjiewang9686 2 роки тому +5

      REALLY Hope you can continue with this PyTorch tutorial.

    • @statquest
      @statquest  2 роки тому +2

      @@yongjiewang9686 Will do!

    • @shichengguo8064
      @shichengguo8064 2 роки тому

      Do we have video talking about transformer? Thanks.

    • @statquest
      @statquest  2 роки тому

      @@shichengguo8064 Not yet, but soon.

    • @Mayur7Garg
      @Mayur7Garg 2 роки тому

      Just a small comment. Any variable should not be named similar to any builtin in Python. The 'input' variable in forward should have been called something else since it is already a builtin function in Python. Otherwise, you end up overriding the builtin within that scope.

  • @santoshmohanram536
    @santoshmohanram536 2 роки тому +121

    Favorite teacher with my favorite Deep learning framework. Lucky to have you. Thanks brother🙏

  • @firesongs
    @firesongs 2 роки тому +71

    Please continue to go through every single line of code including the parameters with excruciating detail like you do.
    None of my professors went over each line like that cuz they always "assumed we already knew" and everyone in the class who didnt already know was afraid to ask to avoid looking stupid. Thank you.

  • @insushin6139
    @insushin6139 11 місяців тому +10

    StatQuest is the GOAT in statistics, machine learning, and deep learning! You're videos are really helping me understanding the concepts and outline of these fields! Love from Korea!

  • @youlahr7589
    @youlahr7589 2 роки тому +32

    Ive used PyTorch for projects before, but I can honestly say that I never fully understood the workings of building a model. I knew that i needed the peices you mentioned, but not why I needed them. You've just explained it incredibly. Please don't stop making this series!!

    • @statquest
      @statquest  2 роки тому +4

      Thank you very much! :)

  • @gummybear8883
    @gummybear8883 2 роки тому +12

    What a blessing this is. You are indeed the Richard Feynman of Data Science.

  • @karlnikolasalcala8208
    @karlnikolasalcala8208 Рік тому +3

    YOU ARE THE BEST TEACHER EVER JOSHH!! I wish you can feel the raw feeling we feel when we watch your videos

  • @jonnhw
    @jonnhw 2 роки тому +3

    Was looking for a pytorch resource and was disappointed when this channel didnt have one yet but then this got uploaded. Really a blessing to the people haha

  • @jamilahmed2926
    @jamilahmed2926 Рік тому +2

    I have lived long enough to watch videos and understand nothing about ML stuffs, until I saw your videos. I truly wish your well being

  • @andreblanco
    @andreblanco 2 роки тому +3

    Triple bam!

    • @statquest
      @statquest  2 роки тому

      BAM! Thank you very much for supporting StatQuest!!!!

  • @binhle9475
    @binhle9475 2 роки тому +5

    AMAZING video. This is exactly what beginners need to start the Pytorch journey with a semi solid footing instead of mindless copying.
    Yoy must have spent so much time for your AWESOME videos.
    GREATLY appreciate your effort. Keep up the good work.

    • @statquest
      @statquest  2 роки тому +2

      Thank you very much! :)

  • @footballistaedit25
    @footballistaedit25 2 роки тому +30

    Thanks for the best content you bring. I hope you continue to make a full pytorch playlist

  • @viveksundaram4420
    @viveksundaram4420 2 роки тому +3

    Man, you are love. I started my neural net journey from your videos and it's the best decision I made. Thank you

  • @kaanzt
    @kaanzt Рік тому +2

    That's really cool explanation! Please continue this PyTorch series, we really need it. BAM!

  • @ToyExamples
    @ToyExamples 4 місяці тому +2

    The style of storytelling is just so unique and friendly

  • @MugIce-lr6ui
    @MugIce-lr6ui 7 місяців тому +3

    Hello! Not sure if anyone's pointed this out yet, but the code on 10:14, 12:09, and 22:42 needs a small addition, `plt.show()`, or else it won't show the graph. Though, maybe 2 years ago when this video was made you didn't need that, I'm not sure, haha.
    Other than that, this is an awesome tutorial that quite literally takes anyone through the process step-by-step, and even tells you some neat fun facts (like the sns nickname) and explanations like how `loss.backward()` works.
    TRIPLE BAM indeed! Thanks for the awesome tutorials and videos you put out 👍

    • @statquest
      @statquest  7 місяців тому

      Thanks! Did you run my code or type it in yourself? I keep the jupyter notebook updated.

    • @ni3d4888
      @ni3d4888 4 місяці тому +1

      plt.show() helped me get the visualizations in Ubuntu under WSL on Windows 11. Thank you for the comment.

  • @AlbertsJohann
    @AlbertsJohann 2 роки тому +1

    What a great feeling when it all clicks after learning about all these concepts in isolation. All thanks to an incredibly brilliant teacher! Triple BAM!!!

  • @Nonexistent_007
    @Nonexistent_007 4 місяці тому +1

    Thank you sir. You have no idea how valuable and helpful your videos are. Keep this good work running

  • @nicolasreinaldet732
    @nicolasreinaldet732 2 роки тому +2

    Guess who was going to start programing a neural network in python today......
    God bless you Josh, becase He know how much you are blessing me with your work.
    And know that Jesus loves you and want to be part of your life.

  • @sapnasharma4476
    @sapnasharma4476 2 роки тому +2

    Thanks for the awesome tutorial! You make the most difficult things so easy to understand, specially with the visuals and the arrows and all! The comments written on the right hand side make it so more helpful to pause and absorb. I would never miss a video of your tutorials!

    • @statquest
      @statquest  2 роки тому

      Hooray! I'm glad you like my videos. :)

  • @Irrazzo
    @Irrazzo 2 роки тому +1

    Thank you, good explanation!
    16:00 Python prefers for-each-loops over index-based loops. See how this equivalent for-each loop looks much simpler.
    for input, label in zip(inputs, labels):
    output = model(input)
    loss = (output - label)**2
    loss.backward()
    total_loss += float(loss)

  • @michapiechowski1810
    @michapiechowski1810 3 дні тому +1

    I can be picky with what videos I learn from. Most of the time I get an ick i cannot get rid of and look for other resources.
    You on the other hand had me on board from the first fkn second, thank you

    • @statquest
      @statquest  2 дні тому

      Awesome! Thank you very much!

  • @aayushjariwala6256
    @aayushjariwala6256 2 роки тому +4

    It amazes me, when I see no NLP video on StatQuest! Josh your explanation are always higher than what one can expect and you have created so many series including maths and conceptual understanding. NLP has the same importance compared to computer vision and actually people are suffering to learn it by lack of content availability! I hope you would create a series or maybe a few videos on basic concepts which help people to get interested in NLP : ) Hope you are doing good in life Josh

    • @statquest
      @statquest  2 роки тому +5

      I'm working on NLP.

    • @vans4lyf2013
      @vans4lyf2013 2 роки тому +3

      @@statquest Yay so glad to hear this, we really need you because no one gives great explanations like you do. Also your youtube comments are the nicest I've ever seen which is a testament to how valued you are in this community.

    • @statquest
      @statquest  2 роки тому

      @@vans4lyf2013 Thank you very much!

  • @ais3153
    @ais3153 2 роки тому +2

    BIG LIKE before watching 👍🏻 please continue the pytorch series

  • @Luxcium
    @Luxcium Рік тому +1

    Wow 😮 I didn't knew I had to watch the *Neural Networks part 2* before I can watch the *The StatQuest Introduction To PyTorch* before I can watch the *Introduction to coding neural networks with PyTorch and Lightning* 🌩️ (it’s something related to the cloud I understand)
    I am genuinely so happy to learn about that stuff with you Josh❤ I will go watch the other videos first and then I will back propagate to this video...

  • @Luxcium
    @Luxcium 11 місяців тому +2

    I am someone who loves *SQ,* and *JS* style of teaching in byte 😅 pieces but I also hate _snakes…_ I love *JavaScript* and *TypeScript* but I’ve been learning *JavaScript* with the _strictest linting rules_ one would imagine… and given how *JavaScript* could be used without any sort of strict rules (and is very similar to *Python* in this context) it is frustrating that it makes *Python* very hard to understand despite being easier since it has not the same stricter rules I have imposed myself learning *JavaScript…* but I am also genuinely grateful that *JS* is the best instructor for this kind of topics because *JS* has a _Ukulele,_ *StatSquatch* and *Normalsaurus* which are all there to help *JS* make *SQ* awesome 🎉🎉🎉🎉 Thanks 😅😅😅❤

  • @_epe2590
    @_epe2590 2 роки тому +3

    finally! some simple to understand content on how to make an AI model using pytourch!!! TRIPLE BAM!!!!

  • @frederikschnebel2977
    @frederikschnebel2977 2 роки тому +4

    Thanks so much for this gem John! Literally got a PyTorch project coming up and your timing is just perfect. Greatly appreciate the content, keep up the good work :)

  • @kwang-jebaeg2460
    @kwang-jebaeg2460 2 роки тому +2

    Wonderful !!! Cant wait your pytorch lightning code for NN. Always thanks alot !!

  • @Ajeet-Yadav-IIITD
    @Ajeet-Yadav-IIITD 2 роки тому +2

    Thank you Josh, pls continue this series of pytorch!

  • @mahammadodj
    @mahammadodj 2 роки тому +1

    Thank you very much! I am new to Deep Learning. I can say that just in one week i learned a lot of things from your tutorials!

  • @anashaat95
    @anashaat95 Рік тому +1

    This series about neural networks and deep learning is very well explained. Thank you soooooooo much.

  • @justinhuang8034
    @justinhuang8034 2 роки тому +1

    Man the content keeps getting better

  • @veronikaberezhnaia248
    @veronikaberezhnaia248 2 роки тому +3

    Amazing content, as always. Before I was a bit afraid to start closing in torch, so thank you to encourage le to do that!

    • @statquest
      @statquest  2 роки тому

      bam! You can do it! :)

  • @saralagrawal7449
    @saralagrawal7449 7 місяців тому +1

    Just watched matrix multiplication of Transformers. My mind was blown away. Same things appear so complex but when this guy explains them, it's like peanuts.
    Triple BAM

  • @theblueplanet3576
    @theblueplanet3576 11 місяців тому +1

    Enjoying this series on machine learning. By the way there is no shame in self promotion, you deserve it 😁

  • @jessicas2978
    @jessicas2978 2 роки тому +2

    Thank you so much, Josh. I have been learning PyTorch and deep learning. This video helps me a lot!

  • @praptithapaliya6570
    @praptithapaliya6570 2 роки тому +1

    I love you Josh. God bless you. You're my favorite teacher.

  • @amirhosseinafkhami2606
    @amirhosseinafkhami2606 2 роки тому +4

    Great explanation as always! Thanks for making content like this, which complements the theoretical concepts.

  • @MariaHendrikx
    @MariaHendrikx Рік тому +1

    I love how you you visualize and synchronize the code with the maths behind it :) On top of that you are doing it step-wise which results in a really awesome and very eduSupercalifragilisticexpialidociouscational video! #ThankYou

    • @statquest
      @statquest  Рік тому

      I love it. Thank you very much! :)

  • @WeeeAffandi
    @WeeeAffandi 2 роки тому +1

    Josh explaining the code is far better than any programmer

  • @xedvap
    @xedvap 2 роки тому +1

    Looking forward to seeing your following videos! Excellent explanation!

  • @exxzxxe
    @exxzxxe Рік тому +1

    Another charming, fully informative masterpiece.

    • @statquest
      @statquest  Рік тому

      Thank you very much! BAM! :)

  • @the_real_cookiez
    @the_real_cookiez 2 роки тому +2

    Quality educational content! It's so cool to see your channel grow. Been here since ~90k subs! Very well earned.

    • @statquest
      @statquest  2 роки тому +2

      Wow! Thank you very much!!! BAM! :)

  • @someone5781
    @someone5781 2 роки тому +1

    Woo! Been waiting for this sort of a tutorial!!!

  • @StratosFair
    @StratosFair 2 роки тому +1

    Nice video, looking forward to the next ones on Pytorch Lightning !

  • @AHMAD9087
    @AHMAD9087 2 роки тому +4

    The tutorial we all needed 🙂

  • @ISK_VAGR
    @ISK_VAGR 2 роки тому

    That is a big leap. I need to check it several times to understand it since I am not a programmer. However, I really got a good feeling of what is happening inside the code. I actually use codeless systems such as KNIME. So if Mr. Sasquatch, get the idea of using KNIME to explain all this, It will be amazing. Thanks to be such a good teacher.

    • @statquest
      @statquest  2 роки тому

      I'll keep that in mind.

  • @arijitchaki1884
    @arijitchaki1884 Рік тому

    Hi Josh, sorry to be a spoil sport, but I used exact same code and my prediction is showing 0.5 for dosage of 0.5 and it is running for all 100 epoch and final b value comes out to be -16.51 😔. But yes the concept is clear!! Great work! I always ask people whoever are interested in learning about data science or machine learning to refer you channel. Seeing your channel grow from 10-20K to a Mn is pleasure to my eyes!! You are the "El Professor"!!

    • @statquest
      @statquest  Рік тому +1

      Thank you very much! If you look at my actual code (follow the link), you'll see that I actually pulled a trick with the data to get it to train faster.

  • @kleanthivoutsadaki5989
    @kleanthivoutsadaki5989 2 роки тому +1

    thanks Josh, you really make understanding Neural Networks concepts a great process!

  • @sceaserjulius9476
    @sceaserjulius9476 2 роки тому +1

    I am also learning Deep Learning, and want to apply it to make good projects,
    This is going to be great.

  • @emilyli6763
    @emilyli6763 2 роки тому +2

    honestly wish I had this a year ago when I was struggling, still watching now tho!

  • @stivraptor
    @stivraptor 2 роки тому +1

    Hey Josh!
    Guess what just arrived in the mail....
    My new statquest mug!!!!!
    Hooray!!!

    • @statquest
      @statquest  2 роки тому

      BAM!!! Thank you so much for supporting StatQuest!!!

  • @hitikagahalot4549
    @hitikagahalot4549 19 днів тому +1

    thank you very much please continue this series it is very useful

    • @statquest
      @statquest  18 днів тому

      I've got a lot more videos like this here: ua-cam.com/video/CqOfi41LfDw/v-deo.html

  • @darshuetube
    @darshuetube 2 роки тому +1

    it's great that you are making videos on coding as well.

  • @aabshaarahmad7853
    @aabshaarahmad7853 2 роки тому +1

    Hi! This is amazing. Are you gonna continue this series? Out of ten different rabbitholes I have been to, this video has been the most helpful for me with understanding PyTorch and starting off with my project. Please continue making more complicated models. Thank you :)

  • @Sandeepkumar-dm2bp
    @Sandeepkumar-dm2bp 2 роки тому +1

    very well explained, thank you for providing quality content, it's very helpful

  • @anggipermanaharianja6122
    @anggipermanaharianja6122 2 роки тому +3

    Awesome vid by the legend!

  • @jwilliams8210
    @jwilliams8210 Рік тому +2

    Absolutely brilliant!

  • @mohsenmoghimbegloo
    @mohsenmoghimbegloo 2 роки тому +1

    Thank you very much Mr Josh Starmer

  • @karrde666666
    @karrde666666 2 роки тому +2

    better than MIT or any university slides

  • @lloydchan9606
    @lloydchan9606 2 роки тому +2

    bless josh and this channel

  • @carlitosvh91
    @carlitosvh91 9 місяців тому +1

    Great explanation. Thank you very much

  • @isaacpeng3625
    @isaacpeng3625 2 роки тому +1

    great video and explanation! me have been struggling in pytorch coding

  • @Drstrange1337
    @Drstrange1337 2 роки тому +1

    Thanks for this amazing walk through.

  • @Hitesh10able
    @Hitesh10able Рік тому

    Another excellent video, one humble request please provide video on Stable Diffusion Models.

  • @ShawnShi-hy9ed
    @ShawnShi-hy9ed 8 місяців тому

    Hi Josh, thanks for your video. I am confused why it doesn't work when I am trying to optimize any other weights and bias.
    five minutes later, I think I have got the answer from the comments and your reply. Thanks again!

  • @jawadmansoor6064
    @jawadmansoor6064 2 роки тому +1

    Great series.

  • @arer90
    @arer90 2 роки тому +3

    Thank you for perfect lecture~!!!

  • @ぶらえんぴん
    @ぶらえんぴん 2 роки тому +1

    Your teaching video is awesome

    • @statquest
      @statquest  2 роки тому

      Thank you!

    • @ぶらえんぴん
      @ぶらえんぴん 2 роки тому

      @@statquest Do you have intro to lightning ? I kind of remember you mentioned in the video you seemed to have one?

    • @statquest
      @statquest  2 роки тому +2

      @@ぶらえんぴん That's going to be the next video in this series. It will come out in a few weeks.

  • @bjornnorenjobb
    @bjornnorenjobb 2 роки тому +1

    omg! I have really wanted this! awesome!!! :) :) :)

  • @joshstat8114
    @joshstat8114 11 місяців тому +1

    Nice video for the introduction of LSTM using PyTorch. There is also `torch` R package that doesn't need to install python and torch. It's so nice that R also has deep learning framework aside from `tensorflow` and I recommend you to maybe try it.

    • @statquest
      @statquest  11 місяців тому

      Thanks for the info!

    • @joshstat8114
      @joshstat8114 11 місяців тому +1

      @@statquest i strongly recommend it because it is so nice that R has own deep learning frameworks, besides h2o

  • @Luxcium
    @Luxcium 11 місяців тому +1

    So I paused at 18:04 because _it blew my mind_ that we were calling backwards() on the loss variable because I thought it was defined on the line above… 😅😅😅😅 but yeah I didn’t find anything so one hour later I was just watching the rest of the video and _to be honest_ in about 33 seconds it came out that it was normal for _my mind to be blown_ 😂😂😂😂 at 18:37

    • @statquest
      @statquest  11 місяців тому

      Totally! I was like, "what?!?!?" when I first saw that.

  • @peterkanini867
    @peterkanini867 Рік тому

    Please make an entire tutorial about the ins and outs of PyTorch!

    • @statquest
      @statquest  Рік тому

      I've made several PyTorch videos and will continue to make more. You can find the others here: statquest.org/video-index/

  • @petercourt
    @petercourt 2 роки тому +1

    Amazing work Josh!

  • @mohammadidreesbhat1109
    @mohammadidreesbhat1109 5 місяців тому +1

    That is how teaching should be.. Triple Bam

  • @aiforeveryone2941
    @aiforeveryone2941 2 роки тому +1

    Double bam new way to teach coding

  • @whispers191
    @whispers191 2 роки тому +1

    Thank you Josh!

  • @05747100
    @05747100 2 роки тому +1

    Thanks a lot, beg for Pytorch Series playlist.

  • @pfever
    @pfever 2 роки тому +1

    Best tutorial like usual! would be nice to see more advanced examples of in pytorch, like CNN for image classification :)

  • @나는강아지-w6x
    @나는강아지-w6x 2 роки тому +1

    KOREAN BAMMMM!!! TY StatQuest😁

  • @sanjaykrish8719
    @sanjaykrish8719 2 роки тому +1

    wow.. super excited

  • @carol8099
    @carol8099 2 роки тому +1

    this video is gold

  • @onkarpandhare
    @onkarpandhare 2 роки тому +1

    great video! very well explained!!!👍👍

  • @abbddos
    @abbddos 2 роки тому +1

    This was great... I hope you can simplify Tensorflow the same way... big big thank you.

  • @shamshersingh9680
    @shamshersingh9680 9 місяців тому

    Hi Josh, thanks again for allowing me to break the ice between me and Pytorch. Everytime I see your videos, I wonder if my instructor could have taught us like this probably our lives must have been much simpler and happier. I have a small doubt here. In the example you have shown gradient training of only final bias. But in reality, all the weights will have to be trained during backpropagation. So when I try to initialise the all weights with random values and then train the model, I do not get the final weights as shown in the video. The code is as follows :-
    class BasicNN(nn.Module):
    def __init__(self):
    super().__init__()
    self.w00 = nn.Parameter(torch.randn(1), requires_grad = True)
    self.b00 = nn.Parameter(torch.randn(1), requires_grad = True)
    self.w01 = nn.Parameter(torch.randn(1), requires_grad = True)
    self.w10 = nn.Parameter(torch.randn(1), requires_grad = True)
    self.b10 = nn.Parameter(torch.randn(1), requires_grad = True)
    self.w11 = nn.Parameter(torch.randn(1), requires_grad = True)
    self.b_final = nn.Parameter(torch.randn(1), requires_grad = True)
    def forward(self, input):
    input_top_relu = input * self.w00 + self.b00
    input_bottom_relu = input * self.w10 + self.b10
    output_top_relu = F.relu(input_top_relu) * self.w01
    output_bottom_relu = F.relu(input_bottom_relu) * self.w11
    input_final_relu = output_top_relu + output_bottom_relu + self.b_final
    output = F.relu(input_final_relu)
    return output
    # Create an instance of the neural network
    model = BasicNN()
    # Print parameters
    print('Parameters before training')
    for name, param in model.named_parameters():
    print(name, param.data)
    # Define inputs and corresponding labels
    inputs = torch.tensor([0., 0.5, 0.1])
    labels = torch.tensor([0., 1.0, 0.])
    # Define a loss function
    criterion = nn.MSELoss()
    # Define an optimizer
    optimizer = optim.SGD(model.parameters(), lr=0.01)
    # Number of epochs for training
    epochs = 1000
    # Training loop
    for epoch in range(epochs):
    total_loss = 0
    # Forward pass
    output = model(inputs)
    # Compute the loss
    loss = criterion(output, labels)
    total_loss += loss
    # Backward pass
    loss.backward() # Compute gradients
    optimizer.step() # Update weights
    optimizer.zero_grad() # Clear previous gradients
    # Print loss every 100 epochs
    if (epoch + 1) % 100 == 0:
    print(f"Epoch [{epoch+1}/{epochs}], Loss: {loss.item()}")
    if (total_loss < 0.00001):
    print(f'Epoch = {epoch}')
    break
    # Print final parameters
    print('Parameters after training')
    for name, param in model.named_parameters():
    print(name, param.data)
    # check the model performance
    input_doses = torch.linspace(start = 0, end = 1, steps = 11)
    output = model(input_doses)
    sns.set(style = 'whitegrid')
    sns.lineplot(x = input_doses, y = output.detach(), color = 'green', linewidth = 2)
    plt.xlabel("Input Doses")
    plt.ylabel("Effectiveness")
    plt.show()
    Request if you can help me with the code above.

    • @statquest
      @statquest  9 місяців тому +1

      This example only works to optimize the final bias term.

  • @neginamirabadi4595
    @neginamirabadi4595 2 роки тому +2

    Hello Josh! Thank you so much for your amazing videos! I have learned so much from your tutorials and would not have been able to advance without them!
    I wanted to ask whether it is possible for you to put some videos on times series analysis, including autoregression (AR), moving average (MA) and their combinations. I would be more than grateful if you can provide such a video. Thank you so much.

    • @statquest
      @statquest  2 роки тому +1

      I'll keep those topics in mind!

  • @sabaaslam781
    @sabaaslam781 Рік тому +1

    Hi Josh. I am a big fan of your videos. I have a question regarding this quest. In this video, we optimized only one parameter. How can we optimize all the parameters? Thanks in advance.

    • @statquest
      @statquest  Рік тому

      I show how to impute all of the parameters in this video on LSTMs in PyTorch: ua-cam.com/video/RHGiXPuo_pI/v-deo.html (if you want to learn about the theory of LSTMs, see: ua-cam.com/video/YCzL96nL7j0/v-deo.html

  • @arshadkazi4559
    @arshadkazi4559 2 роки тому +1

    Tensorflow developer is turning into PyTorch… bam! 💥

  • @mohammadalikhani7798
    @mohammadalikhani7798 Рік тому +1

    That Was Nice ! Thank You

  • @conlele350
    @conlele350 2 роки тому +1

    Thanks Josh, its Incredible video. Beside, recently the Bayes theorem application in fitting model (linear, logistic, random forest...) has became more and more popular in order to replace classic statistic method, could you pls take some time to explain to us some of its popular algorithm like BART, Linear regression via Bayesian Methods...

    • @statquest
      @statquest  2 роки тому +3

      I'm planning on doing a whole series on Bayesian stuff as soon as I finish this series on neural networks.

    • @conlele350
      @conlele350 2 роки тому +2

      @@statquest that's great news for today, thanks Josh, Im looking forward to see it soon

  • @scottabbas3335
    @scottabbas3335 2 місяці тому

    1. The optimizer = SGD (model.parameters (), lr = 0.1) at 14:22 should be changed to optimizer = SGD ([model.final_bias], lr = 0.1), otherwise the parameters other than final_bias will also be optimized. For example, w10 will become -8.6 at iteration 1 of round 1, causing the subsequent gradient descent to fail.
    2. Another thing is optimizer.zero_grad(). I changed it to be placed before the start of the iteration. Placing it as in the video will cause the derivatives to accumulate.
    3. After the optimization, you have to run the model again to output it to the drawing program. If you write it in the order of the video, the output obtained by the drawing program is the output before the model optimization.

    • @statquest
      @statquest  2 місяці тому

      Are you using my code or did you write your own? At 4:53 we set "requires_grad=False".

  • @pauledam2174
    @pauledam2174 Рік тому

    Amazing job! I plan to donate to your patreon page. You were confused because we could use .backward on loss (or at least I was confused by this). I guess one explanation is that loss is defined in terms of output_i and output_i is an instance of the model class. So it may make sense that we can access the backward attribute of loss. But I was, for the same reason a bit surprised that we can subtract a scalar from output_i. One other question. Wouldn't it be better to take the average of total loss? Otherwise the condition that uses 0.0001 is dependent on the the number of examples in the training set.

    • @statquest
      @statquest  Рік тому

      Taking the average is really common, but it doesn't change anything.

    • @pauledam2174
      @pauledam2174 Рік тому

      No, I know it doesn't change things but it means that your criteria has to be adjusted depending on the size of the dataset doesn't it?

    • @statquest
      @statquest  Рік тому +1

      @@pauledam2174 If we wanted to compare the loss among different datasets, then the average would be helpful.

  • @AIpodcast369
    @AIpodcast369 2 роки тому +1

    Sir, Please make videos on the time-series analysis, it's hard to find the videos with clear explaination.

    • @statquest
      @statquest  2 роки тому +1

      I'll keep that in mind.

  • @xray788
    @xray788 4 місяці тому

    Hi Josh, I've watched most of your playlist. It is amazing how you explain it. But can you please explain or point to some reference on where the values for weights come from? I see at start of video like w is 1.70 but confuses me where it came from and why those values are used. Thank you Josh and hopefully once i get that it will be a... Triple bam for me :)

    • @statquest
      @statquest  4 місяці тому

      To create this network, I gave each weight and bias a random initialization value and then tried to fit the neural network to the training data with backpropagation. I then repeated the process a ton of times until I discovered a set of initialization values that worked.

  • @aaronm6675
    @aaronm6675 2 роки тому +1

    Finally🔥

  • @kobic8
    @kobic8 2 роки тому

    great presentation!! thanks again for simplfying this topic! are you planning to post more on NN implementation? computer vision maybe or object detection?

    • @statquest
      @statquest  2 роки тому +1

      Yes, there will be many more videos on how to implement NNs.

  • @minglee5164
    @minglee5164 2 роки тому +1

    Wonderful

  • @junechu9701
    @junechu9701 2 роки тому +1

    Soooooooo thankful!

  • @rrrprogram8667
    @rrrprogram8667 2 роки тому +2

    MEGABAMMMMMM.....
    Hey josh... It's been a very long long time.... I am occupied with different subject right now..
    Hope you are doing good... Will catch you soon..