Lecture51 (Data2Decision) Addressing Multicollinearity

Поділитися
Вставка
  • Опубліковано 26 лип 2024
  • Methods for addressing multicollinearity: adding data, removing correlated variables, ridge regression.
    Course Website: www.lithoguru.com/scientist/st...

КОМЕНТАРІ • 9

  • @jiahengzhou9820
    @jiahengzhou9820 Рік тому

    thank you so much, it's quite helpful!

  • @rubansran422
    @rubansran422 6 років тому

    Thanks for the video, it's so helpful. One thing: Why in a prediction analysis we can let multicollinearity exist, won't it increase the dimension which is redundant, thus leading to poor performance on test set.

  • @lovemelodiesfromindiaandma7022

    Can you please explain in detail what you mean by, ‘restrict the scope of the model to coincide with the range of predictor variables that exhibit the same pattern of multicollinearity’ ?

  • @gjnet89
    @gjnet89 7 років тому

    Great talk. Quick question, when performing ridge regression. Are we changing one x value with sqrt of k or lots? You mentioned diagonally changing the x values...

    • @chrismack783
      @chrismack783 7 років тому +1

      If we have p regressor variables in the model, then we add p new "data points" to our data set. The first one will have a value of sqrt(k) for the first regressor, then zero for all other regressors. The second one will have a value of sqrt(k) for the second regressor, then zero for all other regressors. Etc. This looks like we are adding a diagonal matrix to our data set.

    • @gjnet89
      @gjnet89 7 років тому

      That makes sense. Cheers Chris!

  • @Dreaming-11
    @Dreaming-11 4 роки тому

    I have a doubt that I can't find the answer anywhere: If you have 2 predictor variables with perfect collinearity (one is a linear combination of the other), how can ridge regression decide which one to use and which one to get rid of, since both are in essence the same variable?

    • @ChrisMack
      @ChrisMack  3 роки тому

      With PERFECT collinearity, there is nothing you can do. But that doesn't happen in real life (except a possibly a coding error, which is fixed by fixing that error).

    • @dominicsavolainen24
      @dominicsavolainen24 3 роки тому

      Ridge regression will not perform subset selection and hence, ridge will not tell you which predictor to decide to use. Lasso regression on the other hand can shrink coefficients to 0 and can perform subset selection. But if you wanted to mix the benefits from Lasso and Ridge, you could look at elastic net. Elastic net performs subset selection but it includes grouping effects more than the Lasso which will often pick one variable in a group of highly correlated covariates.