Linear Regression Summary in R

Поділитися
Вставка
  • Опубліковано 13 чер 2020
  • Linear Regression Summary in R
    Linear regression is an essential tool in R, but the output can be a little difficult to interpret. In this video, I walk you through the basics of the output and let you know what type of values to look for when you're fitting a linear model to your data.

КОМЕНТАРІ • 82

  • @analyst42
    @analyst42 Місяць тому +1

    WOW! This is easily the best explanation of the summary statistics for a linear regression model that I've ever encountered. Thank you so much!

  • @nikolescobedo7024
    @nikolescobedo7024 3 роки тому +48

    learn more in 10min than in one semester of stat THANK U SOO MUCH

  • @Annielytics
    @Annielytics Рік тому +12

    This is the best explanation I've seen to date. Thanks for focusing more on what the numbers mean and the conclusions we can make about our models and less on the underpinning formulas.

  • @kennedygolfhead4356
    @kennedygolfhead4356 Рік тому +6

    MY GOD!! IT WAS SOOOOO HELPFUL!! Best explanation ever in just 10 minutes!!!!! How life saving this is for all academics and data scientists!! Would you please please consider going over the summary of Linear Mixed Modelling??? THANK YOU A MILLION!!

  • @georgie66549
    @georgie66549 2 роки тому +3

    I wish there was a love button for this video. Thank you so much!!! I read through my lecture and still had no real clue what I was looking at in R. My understanding is so much clearer now

  • @thesoaringdividend3830
    @thesoaringdividend3830 Місяць тому

    LOVED THIS video! Way better than how my Data Mining professor explained it 😅

  • @rohitekka2674
    @rohitekka2674 3 роки тому +9

    Thank you for the wonderful explanation. This video really felt like light at the end of the tunnel. An absolute enlightenment.

  • @edadila
    @edadila 2 роки тому +2

    this was all i was looking for. perfect explanation, thank you!

  • @CamViesky
    @CamViesky 2 роки тому +2

    Absolutely brilliant, clear, and concise explanation. Thank you.

  • @reubenbrown1995
    @reubenbrown1995 3 роки тому +1

    Brilliant, compressed a week of lectures into 10 minutes!

  • @luismi8936
    @luismi8936 2 роки тому +5

    Thank you man, so accurate. So much information well explained. You're amazing

  • @dodgecarlincila879
    @dodgecarlincila879 2 роки тому

    Thanks for the visual examples, very helpful. Can't wait for your Machine Learning videos!

  • @krisztinachristmon9978
    @krisztinachristmon9978 2 роки тому

    Thank you so much! Finally was able to make sense of the R output and interpret my data.

  • @annazheng3391
    @annazheng3391 Рік тому

    So clear and helpful!! Answered all my questions in 10 mins. Thank you!

  • @rasmusnordman6764
    @rasmusnordman6764 Рік тому +1

    Absolutely incredible video, thank you so much!

  • @cezreycor
    @cezreycor 3 роки тому

    pretty cool stuff man, many thanks for the clear explanation!

  • @ihsan3700
    @ihsan3700 3 роки тому +1

    THANK YOU. Such an amazing lecture .

  • @NamNguyen-kp1xu
    @NamNguyen-kp1xu Рік тому

    Such an underated channel, clean explanation and straight to the point !

  • @kelliegadeken8481
    @kelliegadeken8481 2 роки тому

    Wow, so clear and easy to understand. Thanks!

  • @marvinschumann6832
    @marvinschumann6832 3 роки тому

    Great video my man!!! Thank you so much!

  • @curasvitas96
    @curasvitas96 Рік тому

    oh my gosh this is one of the best explanations I have seen. thank you very much!! I love the part you touched on each of the indicators of the summary() code mean. thank you!!

  • @ianleboo
    @ianleboo 2 роки тому

    Dataslice is just wow... Precise, informative and accurate. Good work👏 Much love from 🇰🇪

  • @johners2006
    @johners2006 2 роки тому

    Very clear and helpful. Pace was good. Thanks for doing it.

  • @CanDoSo_org
    @CanDoSo_org 2 роки тому

    Thank you. Nice and clean.

  • @HuongGiangNguyen-qt3sm
    @HuongGiangNguyen-qt3sm Рік тому

    Thank you! Very well explained.

  • @jac6003
    @jac6003 3 роки тому

    Great channel! Thanks!!

  • @emmy8133
    @emmy8133 Рік тому

    This really solved my question, thanks a lot.

  • @chinthakajayarathne
    @chinthakajayarathne 2 роки тому

    Thank you for video. Much helpful !!

  • @marcodomenicoolivi352
    @marcodomenicoolivi352 2 роки тому

    in 10 minutes, you have explained the linear regression much better than my professor

  • @malinkata1984
    @malinkata1984 Рік тому

    Thank you. Great explanation. :)

  • @Kenkoopa44
    @Kenkoopa44 7 місяців тому

    Excellent!!!! Thanks so much!

  • @abdullahalayed5276
    @abdullahalayed5276 3 роки тому

    Very good. Appreciate the effort.

  •  2 роки тому

    Excellent! Thanks for sharing.

  • @danielsummers9345
    @danielsummers9345 3 роки тому

    Very helpful, quick and easy to understand!

  • @ahmed007Jaber
    @ahmed007Jaber 2 роки тому

    this is superb and clear. thank u for this

  • @asseflas
    @asseflas Рік тому

    This was great thanks :)

  • @amirmirou8037
    @amirmirou8037 Рік тому

    Dude thanks a lot!

  • @nextlevel6825
    @nextlevel6825 2 роки тому

    Incredible 🚀🚀

  • @amaebarnes
    @amaebarnes Рік тому +1

    I was just trying to brush up on my econ/statistics degree because it's been 5 years since I was in university. I just realized how much I've forgotten 😳 😬

  • @nooberinho
    @nooberinho 3 роки тому

    Very helpful, thanks

  • @tagamag
    @tagamag 2 роки тому

    Very well done. Please do a few such videos on Stepwise regression, Logistic regression etc.

  • @genevieveemefaasare8352
    @genevieveemefaasare8352 2 роки тому

    Thanks so much.

  • @brazilfootball
    @brazilfootball 2 роки тому +1

    Great video, thank you! Any chance you could make one for the summary output of other types of regression models (Logistic, neg. binomial, Poisson, etc.)?

  • @not_from_here4477
    @not_from_here4477 3 роки тому

    nailed it!

  • @Quetzal00358
    @Quetzal00358 Рік тому

    This was great! Have you made a video where you also include categorical variables?

  • @zakariya_am
    @zakariya_am 4 місяці тому

    Thank you

  • @darkrillen
    @darkrillen Рік тому

    well said

  • @shuvhamdigitalacademy3228
    @shuvhamdigitalacademy3228 3 роки тому +2

    Kindly make some videos on multiple regression analysis and interpretation.

  • @harentongaray3443
    @harentongaray3443 2 роки тому

    So, if the P value is less than. 0.5 the model is significant and therefore we void the null hypothesis?

  • @hashemfathi1646
    @hashemfathi1646 3 роки тому +2

    spot on.... just one question, what does the " on 2" mean in the F statistic part?

    • @dataslice
      @dataslice  3 роки тому +2

      Good question -- that's how many x variables we used for the regression

  • @CaseySchacher
    @CaseySchacher 2 роки тому

    How do I access the second video talking about diagnosing regression models? You mentioned that were going to make the new modeling/diagnostics video towards end of this one. Really appreciated this video!

    • @dataslice
      @dataslice  2 роки тому +1

      Yes, I'm hoping to release a video covering the linear regression plots soon and then potentially more regression videos down the line!

  • @aishwariyagupta619
    @aishwariyagupta619 2 роки тому

    Hi, can we determine the sample size from this output?

  • @dataslice
    @dataslice  2 роки тому

    Part 2 (Regression *Plots* Explained): ua-cam.com/video/rfH7pCFvFT0/v-deo.html

  • @jossri
    @jossri 3 роки тому

    Great video!. Thank you. I have a question more related to the type of object of the Lm output. If I’m doing several Lm, how can I extract values of the output and append in a df to compare the results?

    • @dataslice
      @dataslice  3 роки тому +1

      Great question! You can save the output of the model by assigning the summary of it to a variable, e.g. `x = summary(fish_model)`. Then, if you open up 'x', you'll notice it's a list object with different vectors and values -- you can do `names(x)` to see the different variables and access them accordingly. For instance, if you wanted to extract the r squared, you could call `x$r.squared`

    • @jossri
      @jossri 3 роки тому +1

      @@dataslice thank you so much for your answer.

  • @jives.
    @jives. 2 роки тому

    thank you dataslice gang

  • @aishwariyagupta619
    @aishwariyagupta619 2 роки тому

    How do we know residuals are normally distributed?

  • @nooberinho
    @nooberinho 3 роки тому

    Your video was helpful for some working I'm doing atm. Hopefully you could help me with this question I've asked elsewhere: Hello, I've seen many statistics courses note that for a single linear regression if you regress an outcome variable on a binary predictor variable the slope coefficient is the same as the difference in average outcomes between the two groups. Is this still accurate for a multiple linear regression for a binary predictor variable when you also have multiple other non-binary variables? Thanks!

    • @dataslice
      @dataslice  3 роки тому +1

      If you have a multiple linear regression with one binary predictor and multiple non-binary predictors, then the slope of the binary predictor is the same as the difference in average outcomes between the two groups *if you hold all other predictor variables constant*. An example of this would be if we were plotting weight as a function of height and gender (weight ~ height + isMale). If the coefficient for isMale is 20, then holding height constant, the difference between the avg male and female is 20. Note that this would be different if there were any interaction effects between isMale and height

    • @nooberinho
      @nooberinho 3 роки тому

      @@dataslice that's great, thanks very much

    • @noamills1130
      @noamills1130 3 роки тому +1

      @dataslice What if you have a categorical predictor that has more than two possible values? Would you have to use a different kind of regression model?

    • @dataslice
      @dataslice  3 роки тому +2

      @@noamills1130 Great question. So going back to the previous example with plotting weight as a function of height and gender (weight ~ height + isMale), let's say we had a 'race' variable that could either be white, hispanic, asian, or african american. We could then create 3 additional dummy variables (isHispanic, isAsian, isAfricanAmer) in our data for the regression. If isHispanic = 1, isAsian = 0, and isAfricanAmer = 0 then that represents the race as hispanic (and so on and so forth for asian and african american). If all three dummy variables are 0, then the person would be white (the baseline). When you make the regression, each dummy variable would be given a coefficient which could help you determine the prediction among the different races. Does that make sense?

    • @noamills1130
      @noamills1130 3 роки тому

      @@dataslice Great, thank you so much! I'll be using this for my research project analyzing wildfire trends in the US.

  • @apurbabiswas8072
    @apurbabiswas8072 2 роки тому

    Waiting for the 2nd video of this topic

  • @BarcenasJoel
    @BarcenasJoel 3 роки тому

    How would you interpret the estimate for intercept?

    • @dataslice
      @dataslice  3 роки тому +2

      Depending on what data you're trying to fit, the interpretation may vary. For instance, if we were fitting a line on a dataset of ages and weights (Y = weight, x = age), the interpretation of the intercept would be 'how much a person weighs when they are born (age 0)'. However, if you're fitting a line to data that's very far from the x-axis, your interpretation may be invalid

  • @BarcenasJoel
    @BarcenasJoel 3 роки тому

    What does it mean when you have a std. error that is higher than your coefficient?

    • @dataslice
      @dataslice  3 роки тому +1

      Again, it's hard to tell without seeing the data but typically this means that the variable may have no statistically significant effect. This could happen because there's actually no effect, or there are some outliers in the data that's affecting the fit

  • @maxh5522
    @maxh5522 3 роки тому

    Why did I watch an hour long lecture that you covered in ten minutes?!? Our teacher should’ve just sent us to your channel

  • @kennylee6768
    @kennylee6768 2 роки тому

    So what is the regression equation?

  • @justin2icy
    @justin2icy 3 роки тому

    is the "-433.576" known as the regression coefficient?

    • @ianleboo
      @ianleboo 2 роки тому

      It is considered the mean of the regression equation when predictor variables are all zero.

  • @mundane12321
    @mundane12321 2 роки тому

    where is part 2 available? (reading plots)

  • @gladiatorsfc9952
    @gladiatorsfc9952 3 роки тому

    7:35 R Squared Interpretation

  • @nol-sor1985
    @nol-sor1985 3 роки тому

    10 min >>>>>>>> 2 trimesters

  • @doumansarouei4523
    @doumansarouei4523 3 роки тому

    very helpful, thanks