Linear Regression Summary in R
Вставка
- Опубліковано 13 чер 2020
- Linear Regression Summary in R
Linear regression is an essential tool in R, but the output can be a little difficult to interpret. In this video, I walk you through the basics of the output and let you know what type of values to look for when you're fitting a linear model to your data.
WOW! This is easily the best explanation of the summary statistics for a linear regression model that I've ever encountered. Thank you so much!
learn more in 10min than in one semester of stat THANK U SOO MUCH
This is the best explanation I've seen to date. Thanks for focusing more on what the numbers mean and the conclusions we can make about our models and less on the underpinning formulas.
MY GOD!! IT WAS SOOOOO HELPFUL!! Best explanation ever in just 10 minutes!!!!! How life saving this is for all academics and data scientists!! Would you please please consider going over the summary of Linear Mixed Modelling??? THANK YOU A MILLION!!
I wish there was a love button for this video. Thank you so much!!! I read through my lecture and still had no real clue what I was looking at in R. My understanding is so much clearer now
LOVED THIS video! Way better than how my Data Mining professor explained it 😅
Thank you for the wonderful explanation. This video really felt like light at the end of the tunnel. An absolute enlightenment.
this was all i was looking for. perfect explanation, thank you!
Absolutely brilliant, clear, and concise explanation. Thank you.
Brilliant, compressed a week of lectures into 10 minutes!
Thank you man, so accurate. So much information well explained. You're amazing
Thanks for the visual examples, very helpful. Can't wait for your Machine Learning videos!
Thank you so much! Finally was able to make sense of the R output and interpret my data.
So clear and helpful!! Answered all my questions in 10 mins. Thank you!
Absolutely incredible video, thank you so much!
pretty cool stuff man, many thanks for the clear explanation!
THANK YOU. Such an amazing lecture .
Such an underated channel, clean explanation and straight to the point !
Wow, so clear and easy to understand. Thanks!
Great video my man!!! Thank you so much!
oh my gosh this is one of the best explanations I have seen. thank you very much!! I love the part you touched on each of the indicators of the summary() code mean. thank you!!
Dataslice is just wow... Precise, informative and accurate. Good work👏 Much love from 🇰🇪
Very clear and helpful. Pace was good. Thanks for doing it.
Thank you. Nice and clean.
Thank you! Very well explained.
Great channel! Thanks!!
This really solved my question, thanks a lot.
Thank you for video. Much helpful !!
in 10 minutes, you have explained the linear regression much better than my professor
Thank you. Great explanation. :)
Excellent!!!! Thanks so much!
Very good. Appreciate the effort.
Excellent! Thanks for sharing.
Very helpful, quick and easy to understand!
Thanks! :)
this is superb and clear. thank u for this
This was great thanks :)
Dude thanks a lot!
Incredible 🚀🚀
I was just trying to brush up on my econ/statistics degree because it's been 5 years since I was in university. I just realized how much I've forgotten 😳 😬
Very helpful, thanks
Very well done. Please do a few such videos on Stepwise regression, Logistic regression etc.
Thanks so much.
Great video, thank you! Any chance you could make one for the summary output of other types of regression models (Logistic, neg. binomial, Poisson, etc.)?
nailed it!
This was great! Have you made a video where you also include categorical variables?
Thank you
well said
Kindly make some videos on multiple regression analysis and interpretation.
So, if the P value is less than. 0.5 the model is significant and therefore we void the null hypothesis?
spot on.... just one question, what does the " on 2" mean in the F statistic part?
Good question -- that's how many x variables we used for the regression
How do I access the second video talking about diagnosing regression models? You mentioned that were going to make the new modeling/diagnostics video towards end of this one. Really appreciated this video!
Yes, I'm hoping to release a video covering the linear regression plots soon and then potentially more regression videos down the line!
Hi, can we determine the sample size from this output?
Part 2 (Regression *Plots* Explained): ua-cam.com/video/rfH7pCFvFT0/v-deo.html
Great video!. Thank you. I have a question more related to the type of object of the Lm output. If I’m doing several Lm, how can I extract values of the output and append in a df to compare the results?
Great question! You can save the output of the model by assigning the summary of it to a variable, e.g. `x = summary(fish_model)`. Then, if you open up 'x', you'll notice it's a list object with different vectors and values -- you can do `names(x)` to see the different variables and access them accordingly. For instance, if you wanted to extract the r squared, you could call `x$r.squared`
@@dataslice thank you so much for your answer.
thank you dataslice gang
How do we know residuals are normally distributed?
Your video was helpful for some working I'm doing atm. Hopefully you could help me with this question I've asked elsewhere: Hello, I've seen many statistics courses note that for a single linear regression if you regress an outcome variable on a binary predictor variable the slope coefficient is the same as the difference in average outcomes between the two groups. Is this still accurate for a multiple linear regression for a binary predictor variable when you also have multiple other non-binary variables? Thanks!
If you have a multiple linear regression with one binary predictor and multiple non-binary predictors, then the slope of the binary predictor is the same as the difference in average outcomes between the two groups *if you hold all other predictor variables constant*. An example of this would be if we were plotting weight as a function of height and gender (weight ~ height + isMale). If the coefficient for isMale is 20, then holding height constant, the difference between the avg male and female is 20. Note that this would be different if there were any interaction effects between isMale and height
@@dataslice that's great, thanks very much
@dataslice What if you have a categorical predictor that has more than two possible values? Would you have to use a different kind of regression model?
@@noamills1130 Great question. So going back to the previous example with plotting weight as a function of height and gender (weight ~ height + isMale), let's say we had a 'race' variable that could either be white, hispanic, asian, or african american. We could then create 3 additional dummy variables (isHispanic, isAsian, isAfricanAmer) in our data for the regression. If isHispanic = 1, isAsian = 0, and isAfricanAmer = 0 then that represents the race as hispanic (and so on and so forth for asian and african american). If all three dummy variables are 0, then the person would be white (the baseline). When you make the regression, each dummy variable would be given a coefficient which could help you determine the prediction among the different races. Does that make sense?
@@dataslice Great, thank you so much! I'll be using this for my research project analyzing wildfire trends in the US.
Waiting for the 2nd video of this topic
How would you interpret the estimate for intercept?
Depending on what data you're trying to fit, the interpretation may vary. For instance, if we were fitting a line on a dataset of ages and weights (Y = weight, x = age), the interpretation of the intercept would be 'how much a person weighs when they are born (age 0)'. However, if you're fitting a line to data that's very far from the x-axis, your interpretation may be invalid
What does it mean when you have a std. error that is higher than your coefficient?
Again, it's hard to tell without seeing the data but typically this means that the variable may have no statistically significant effect. This could happen because there's actually no effect, or there are some outliers in the data that's affecting the fit
Why did I watch an hour long lecture that you covered in ten minutes?!? Our teacher should’ve just sent us to your channel
So what is the regression equation?
is the "-433.576" known as the regression coefficient?
It is considered the mean of the regression equation when predictor variables are all zero.
where is part 2 available? (reading plots)
Working on it now! :)
7:35 R Squared Interpretation
10 min >>>>>>>> 2 trimesters
very helpful, thanks