Ridge Regression Part 1 | Geometric Intuition and Code | Regularized Linear Models

Поділитися
Вставка
  • Опубліковано 1 чер 2021
  • Dive into the fundamentals of Ridge Regression with the first part of our series. We'll provide a clear geometric intuition, backed by practical code examples. Explore how Ridge Regression, a form of regularized linear models, can enhance your understanding of linear regression in the presence of multicollinearity.
    Code: github.com/campusx-official/1...
    ============================
    Do you want to learn from me?
    Check my affordable mentorship program at : learnwith.campusx.in/s/store
    ============================
    📱 Grow with us:
    CampusX' LinkedIn: / campusx-official
    CampusX on Instagram for daily tips: / campusx.official
    My LinkedIn: / nitish-singh-03412789
    Discord: / discord
    E-mail us at support@campusx.in

КОМЕНТАРІ • 50

  • @mdnoorsolaimansiam8821
    @mdnoorsolaimansiam8821 Рік тому +16

    For past two days, I was watching different videos and reading articles to understand the core of ridge regression. I got tired as I wasn't understanding. And here I am, after half past this video, I think I've got the grasp. It'd be biased to say previous contents didn't help me at all, but your lecture is so much insightful than those. Thank you very much for sharing your learnings with us.

  • @nitika9769
    @nitika9769 5 місяців тому +1

    Omg this is brilliant. Exactly what I've been looking for. Thanks for making our lives easier

  • @quicksciencecalculation5676
    @quicksciencecalculation5676 Рік тому +5

    Well explained sir please i request you never Stop

  • @23injeela79
    @23injeela79 Місяць тому

    very well explained . sir you have done a lot of hardwork on your lectures . keep going .

  • @mustafizurrahman5699
    @mustafizurrahman5699 Рік тому

    Mesmerising such lucid explanation

  • @sujithsaikalakonda4863
    @sujithsaikalakonda4863 Рік тому +2

    Great explanation sir.

  • @krishnakanthmacherla4431
    @krishnakanthmacherla4431 2 роки тому +1

    You are a game changer sir

  • @1234manasm
    @1234manasm Рік тому +1

    You are look like a genius and tech like a professor

  • @shashankbangera7753
    @shashankbangera7753 10 місяців тому

    Beautiful explaination!

  • @ParthivShah
    @ParthivShah 3 місяці тому +1

    Thank You Sir.

  • @harsh2014
    @harsh2014 Рік тому

    Thanks for this session.

  • @TomatoPotato874
    @TomatoPotato874 2 місяці тому +1

    Charansparsh aapko 🙏

  • @yuktashinde3636
    @yuktashinde3636 Рік тому +1

    at 10:51 why we considered training datapoints in testing but yes there are other training points who will create y-y^ value

  • @manasyadav3159
    @manasyadav3159 9 місяців тому

    Hi, how can we say that if the slope is very high then it will be the case of overfitting, it can be underfitting also. I think high slope doesn't mean it will perfectly fit on our training data. Please help me out.

  • @osho_magic
    @osho_magic Рік тому +1

    Sir what if the incorrect fit line is already on the right side and imaginary true fit is on left ,then ridge will shift it more right away from true fit,, ".? It become irregularisation. Isn’t it.?

  • @arman_shekh97
    @arman_shekh97 3 роки тому

    Sir this video has cames after 5 days , everything is fine now

  • @krishcp7718
    @krishcp7718 Рік тому

    Hi Nitish,
    Very nice video.
    Just one thing I noticed - around 04:00. For a given intercept b, when m changes, it is basically the orientation or the angle the line makes with the x-axis changing. So when m is either too small or too high, there is underfitting. As can be seen geometrically the line is quite away from the data points for high and low m. So the overfitting - meaning line is very close to the data points is for only certain values of m - particularly between high m and low m values. Please let me know your thoughts on this.
    Regards,
    Krish

  • @ATHARVA89
    @ATHARVA89 3 роки тому +5

    Sir will you including SVM t-sne and all ahead in the 100 days ML playlist?

  • @mohitkushwaha8974
    @mohitkushwaha8974 Рік тому

    Awesome

  • @arshad1781
    @arshad1781 3 роки тому

    THANK

  • @shaikhsaniya3585
    @shaikhsaniya3585 Рік тому

    Regularization is regression here ??

  • @d-pain4844
    @d-pain4844 2 роки тому +1

    Sir thoda dark marker use Karo

  • @pankajwalia155
    @pankajwalia155 Рік тому

    Koi please bata do training error generalized error testing error irreducible error kis section main hai mera exam hain 20 dec ko

  • @ghostofuchiha8124
    @ghostofuchiha8124 4 місяці тому

    Krish Naik hindi has explained this better ; Rest others till now ; CampusX seems good

    • @darshedits1732
      @darshedits1732 4 місяці тому

      For only this algorithm krish naik is explained better or for all algorithms krish naik are explained better?

  • @rohitdahiya6697
    @rohitdahiya6697 Рік тому

    why there is no learning rate hyperparameter in scikit-learn Ridge/lasso/Elasticnet . As it has a hyperparameter called max_iteration that means it uses gradient descent but still there is no learning rate present in hyperparameters . if anyone knows please help me out with it.

    • @near_.
      @near_. Рік тому

      Did u get the answer??

    • @rohitdahiya6697
      @rohitdahiya6697 Рік тому

      @@near_. no still waiting for some expert to reply

    • @YogaNarasimhaEpuri
      @YogaNarasimhaEpuri Рік тому +1

      I didn't thought about this...
      Just I seen from documentation, that All Solvers are not using Gradient Descent. (which i think)
      SAG - uses a Stochastic Average Gradient descent, The (step size/learning rate) is set to 1 / (alpha_scaled + L + fit_intercept) where L is the max sum of squares for over all samples.
      ‘svd’ uses a Singular Value Decomposition, (Matrices)
      cholesky, (Matrices)
      ...
      Otherwise, Like SAG, all solvers based upon the data and solver, automatically calculate the learning rate
      What's ur opinion?

  • @abhinavkale4632
    @abhinavkale4632 3 роки тому +1

    Just one issue.. why did you multiply 0.9*3 while calculating the loss at second point?

    • @sidindian1982
      @sidindian1982 Рік тому +1

      even i am confused on this .,..:-(

    • @balrajprajesh6473
      @balrajprajesh6473 Рік тому +1

      It is clearly mentioned by Sir in the video that it is just an assumption.

    • @somanshkumar1325
      @somanshkumar1325 Рік тому +2

      There are two points in our training dataset -> (1,2.3) and (3,5.3).
      For calculating the loss at the second point,
      Yi = 5.3, Xi = 3.
      Y_hat = m*Xi + b where m=0.9, Xi = 3, b=1.5.
      Y_hat = 0.9*3+1.5
      I hope it helps?

    • @abhinavkale4632
      @abhinavkale4632 Рік тому

      @@somanshkumar1325 yooo...

  • @We.shall.fly1
    @We.shall.fly1 5 місяців тому +1

    I appreciate your work and no one can teach you like you but there is just a thing, Overfitting doesn't mean high slope in simple linear regression. Overfitting means you have used very complex model which is not able to generalise for new data which is not in training. Simple linear regression is simplest model, so they can't be overfitting in it. There can be only underfitting.

    • @kindaeasy9797
      @kindaeasy9797 Місяць тому

      exactly , simple linear regression mai over fitting ho hi nahi skti , because the line will not bend to pass from each and every data point in the training data set , yes it can underfit and best fit

  • @lingasodanapalli615
    @lingasodanapalli615 3 місяці тому

    But sir why did you choose two training points above the actual dataset. If we chose those two training points below the actual dataset then the correct line's slope is higher than predicted lin's slope. So the loss of the predicted line's slow will be less

    • @ruchitkmeme4441
      @ruchitkmeme4441 2 місяці тому

      exactly for this problem i came into comment box!!
      i mean if we give all data and not only that two point normal linear regretion will also choose that line that we want after ridge regression

  • @ANUbhav918
    @ANUbhav918 2 роки тому

    I guess, Bias is more compared to variance in Overfitting. Vice versa in Underfitting. Please correct me

  • @asifkdas
    @asifkdas Рік тому

    the slope value in a linear regression model does not directly indicate overfitting

    • @Tusharchitrakar
      @Tusharchitrakar 6 місяців тому

      Yes ofcourse but I think what he's trying to suggest is that some suspiciously high values "might" be indicative of over fitting.

  • @kindaeasy9797
    @kindaeasy9797 Місяць тому

    but in the graph at 7:28 if jo 2 training points hai vo test points ke neeche hotai and best fit line ke slope ko increase krna padta na overfitting ko handle krne ke liye ??? I think is video ka logic flawed hai

  • @kindaeasy9797
    @kindaeasy9797 Місяць тому

    aapki jo overfitting wali line hai , vo toh low bias and high variance wali definition ko satisfy hi nahi kr rhi , i dont think that overfitting is possible in case of simple linear regg, because the line cant bend to pass from each and every data point of the training data set

  • @garimadhanania1853
    @garimadhanania1853 Місяць тому

    I really appreciate your effort and all your videos but I think the explanation is incorrect here.
    m being high is not the definition of overfitting
    in a typical linear regression with m + 1 weights, if we do not constrain the value of weights and let them be anything, then they can represent very complex functions and that causes overfitting
    we have to penalize large values of weights (by adding in the loss function) so that our function has lower capacity to represent complexity and hence it wont learn complex functions that just fit the training data well

  • @ashisharora9649
    @ashisharora9649 Рік тому

    Sorry to say but you share the bookish knowledge this time. Practical intuition is not there. Adding something parallelly shifts the line upward. How does it able to make change in the slope? You said, mostly ds keeps this model as default as it will only be active if there is a situation of overfitting, Kindly explain that how? How model is able to find the best fit line of test set is that you assumed it on your own. Does algorithm do the same?

    • @campusx-official
      @campusx-official  Рік тому

      Regularization: ua-cam.com/play/PLKnIA16_RmvZuSEZ24Wlm13QpsfLlJBM4.html
      Check out this playlist, maybe this will help