NOTE: 25:39 I should have (Pfit - Pmean) instead of the other way around. Support StatQuest by buying my book The StatQuest Illustrated Guide to Machine Learning or a Study Guide or Merch!!! statquest.org/statquest-store/
Thank you for the great video! Please note that from the second 25:49 the degrees of freedom for the numerator should be (Pfit-Pmean), otherwise it is less than 0.
There are two different ways to fit the line to data. The one most commonly used is to simply do the math and solve for the optimal fit (take the derivative with respect to the squared residuals and solve for where it is equal to 0). However, that method only works in this specific situation. A more general method is based on the "rotate the line approach" that I illustrate in this video. To learn more about it (how to rotate the line), see my video on Gradient Descent: ua-cam.com/video/sDv4f4s2SB8/v-deo.html
Thank you for making this series of statistic videos. One question please: I want to calculate the least squares growth rate of sales for a company. Would I have "higher quality" growth rate by using quarterly sales (40 pieces of data) vs. annual sales (10 pieces of data). Would the seasonality (Christmas sales higher) affects of quarterly sales and distort the growth rate? Thanks,
It sort of depends on how exactly you want to model and what you want to get out of the model. If you want to take seasonality into account, then you need to fit a periodic function (like a sine function) to your quarterly data. That said, the easiest thing to do would be to start with annual sales and see how useful that is.
Great video overall! But I'm a little confused with your description of calculating a p-value for the R^2. Does this mean we are treating R^2 as a random variable itself and looking at its distribution? Because to me it seems like it is the f-statistic that follows an f-distribution, hence we are calculating a p-value for the f-stat, not the R^2 itself, which(correct me if I'm wrong) does not follow any specific distribution. So what exactly is the connection between the R^2 and the f-stat and its corresponding p-value?
I have a question, in 5:24 why the variance is calculated dividing by n instead of n-1, I thought all the observed data points are just a sample of a bigger population includes data points which we haven't observed yet. I'm sorry if my English confuse you because it isn't my mother tongue
is residual the difference between the observed value of the dependent variable and the predicted value or the difference between the overall mean of the dependent and the observed value
They are the same. However, I changed notation so that I could specify when which model we were using to make the predictions. SS(fit) is the RSS around the fitted line and the SS(mean) is the RSS around the mean.
I struggled understanding this topic through a textbook/ professor videos online, and this was just a great explanation. It was like watching this video, made all the pieces finally fit
I'm an electrical engineer who wanted to learn about machine learning, and your videos helped me understand all the fundamentals of this field. Thank you so much, sir
Wow...i was searching for this on your channel last week and I was so sad I didnt find it... luckily i still have time to study for the test. Thank you!
This is implementation of Linear regression from scratch in NumPy only. In-depth explanation of key concepts like Cost Function and Gradient Descent ua-cam.com/video/wxCQxZKo4hU/v-deo.html
Thank God i came accross your videos. Making my CFA journey towards statistics less overwhelming by explaining like you are explaining to a 5 year old...pheeewww.
Cool merch you could probably easily create would be a workbook to pair with your book where we could practice calculating R2 for exemple in different scenarios. That way, everytime you learn a new concept you can practice doing the formulas :) i'd totally buy that 😏 and maybe links to extra videos or explainations on the concepts that are a little harder to comprehend for people that are completely new to this field and a little slow lol(like linear regression 😅)
Hi josh, while getting to R^2, you give the formula y= (data-mean)^2. This contradicts your StatQuest "Fitting a line to the data", where your formula was "(b-y1)^2+(b-y2)^2+...", meaning "(intersect-data)^2. Now i already understood that by squaring the difference you get the same positive value, so the order doesn't matter for this purpose. Is there another reason why you put it in the order "(data-mean)^2" in this video? Thanks. Love the videos, just watching for fun
This is the best explanation I have ever come accross on Linear Regression. I have a much better intuitive understanding of what the mathematic formulas I was exposed to mean. I do have a question. At 21:42, should the numerator be interpreted as [SS(mean) -SS(fit)]/(Pfit - Pmean) or is it SS(mean) - [SS(fit)/(Pfit - Pmean)] ? The position of the square brackets is not clear to me. Kindly clarify.
I was always wondering why the model chooses to use R2 rather than absolute value of R, until you draw that polynomial out of all sum of squares. It makes sense now
Hello, I had one doubt. For calculating multiple F values, are we taking random samples from our original dataset itself? As in, if there are 100 data points in total, we will take 80, 70 and any random data points from 100 to plot F values on histogram? Could you please help me with this?
The example where we use random data is just an example of the concepts behind how the p-value is calculated. In practice, we use a curve generated by the F distribution (see 25:26) that represents what would happen if we had generated an infinite number of random datasets.
this is amazing, thanks a lot @statquest , please can you also do a video of linear mixed models and generalised linear mixed models, there a few videos about them on UA-cam, it would really helpful. Thank you for the good work
Your videos are awesome! Thanks a lot for making complex concepts simpler. It will be helpful if you clearly explained Discrete probability distributions
wow this make so much sense! I'm pissed why college professors don't teach like this, it was a waste of time to sit in their classes being so confused right from the start. I can't thank you enough for your videos!
If they were perpendicular, than we would lose the relationship between the variable on the x-axis and the variable on the y-axis, and the whole point is to use an x-axis value to predict a y-axis value. Thus, the residuals are parallel with the y-axis - this preserves the relationship that we want to use to make predictions.
@@statquest R^2 is just a metric right, and I can set the coefficients of independent variables in such a way that variance(error) exceeds variance(y),( as variance(error) = variance(y* - y), (where y* is the infered value, and y is the actual value) , I can always make y*-y infinitely high for one datapoint, by choosing appropriate coefficients ), or am I wrong? Please correct me.
@@mathematics6199 Yes, in theory, you can do that - but that's not linear regression. In linear regression we don't just set the coefficients to whatever we want. We set them so that they minimize the sum of the squared residuals. And this is why R^2 isn't negative in this context. However, in other contexts, where you can do whatever you want, yes, it can be negative.
This is an excellent video Josh, thank you! I understand all well until you explain about p-value 23:58. So we were using a dataset of mouse size/weight and weight/tail length/body length, but I'm confusing where the 'random dataset' comes from when you calculate p-value. Could you explain a bit further about this please?
The idea is to give you an intuitive sense of what the p-values associated with linear regression represent. So, to start with, we had 9 data points (9 pairs of weight/height measurements) and fitted a line to it and calculated the F value. That is the "observed" F value generated from the original, raw data. Now pair 9 random values for height (and these could be any reasonable values for height that you randomly select) with 9 random values for weight (and these could be any reasonable values for weight). Calculate the F for those pairs of random values and put that in a histogram. Then repeat until we've done that a lot of times and compare the observed F value from the original data to the histogram.
@@statquest Thanks for explaining all. Much appreciate it. So those 'random values' are completely random, just made up within the range of the normal dataset, right? Then when we are calculating F and p values in SPSS or R, do those softwares go through this process? It might be a bit silly questions, hopefully I'm not too far away!
@@gnosmik That's the idea. However, as mentioned at 25:26, in practice, people (and software) just use an F-distribution (which is an equation for a curved line) to calculate the p-value. The idea of using random data is just to give you an intuition of what the curved line created by the F-distribution represents.
Your explanations are wonderful. Please just recommend the book should be studied with your videos. Please make videos on chi-Squared distribution, Monte Carlo Simulations and Hypotheses testing. Thanks for your valuable help.
My favorite book to go along with my videos is The StatQuest Illustrated Guide To Machine Learning. You can get it here: statquest.org/statquest-store/
Question. Why are we calculating R2 value and the p value? Is it the industry standard? Or else What led to the decision that you included it with linear regression. Theoretically Lin reg is complete before that right?(Making concepts clear)
If you just want to fit a line to data, you can used the method of least squares. However, if you want to quantify how well that line fits your data, then you use Linear Regression. Linear Regression consists of using least squares to fit the line to the data and then calculating r^2 and its p-value to evaluate how well that line fits the data.
@@statquest still confused.. as you said 'how well it fits the data', so the r2 and p value are tests for evaluation right? dont they have alternatives? or is it necessary to do exactly these steps. I'll still get a logistic regression model but it may not be the best one without them? Or are you saying that these, or some other alternatives tests are necessary to do, to assess the model and this repeats iteratively until best fit?
@@Slayer1407-d9d They do have alternatives, so, as you say, you might think of r^2 and its corresponding p-values as the 'industry standards'. Pretty much every program that offers a linear regression function will give you those as outputs. However, there are alternatives, and you can read more about them here: developer.nvidia.com/blog/a-comprehensive-overview-of-regression-evaluation-metrics/ among other places.
Of course! I am the person who is embarrassed on the inside that I don't get the stats terms when thrown around at work, but know that I'm memorized them so know what they are, but really don't understand the "why" or how it all relates. Thank you so much for speaking slowly in your videos, reiterating concepts, sometimes with additional concepts in between, and your humor. It's fun. I'm grateful. @@statquest
I have contacted UA-cam about this problem, but, unfortunately, they are all on vacation until next week. :( The good news is that this video does a pretty good job summarizing the concepts in that other video.
Awesome, but can we do this without squaring? Why can't we just sum the residuals without any squaring, it looks like it should give us the sum of all distances and then we could plot it in the same way and pick the rotation that gives us the least sum of non-squared residuals and it should still work, curious why do we choose to square it, thank you so much for the video
If the "distances" below the line are negative, they will cancel out the ones above them, so that's a problem. However, we could then take the absolute value so that everything is positive. This could work if Linear Regression was actually solved the way I've presented it here. However, in practice, when you square the distances, you can solve for the optimal parameters directly by taking the derivative of the squared residuals with respect to each parameter, setting those derivatives equal to 0 and then solving for the parameter values.
I don't understand why least squares can cause any term that will make ss(fit) worse to be multiplied by 0. Is it because mean squares differential the equation? 15:20
Thank you for the nice video! I wonder for your explanation to the F curves around 25:53, shouldn't it be (p_{fit} - p_{mean})=1? In addition, would you please provide the link to your video about the degrees of freedom if that is already available?
NOTE: 25:39 I should have (Pfit - Pmean) instead of the other way around.
Support StatQuest by buying my book The StatQuest Illustrated Guide to Machine Learning or a Study Guide or Merch!!! statquest.org/statquest-store/
HONESTLY IF YOU STARTED A NEW RELIGION. I WOULD CONVERT
bam! :)
Thanks Josh, your channel is recommended from Murdoch University,Australia lecturers. Worth watching your channel
Thanks!
Great work! The graphics made it super easy to understand.
Glad it helped!
Thank you for the great video! Please note that from the second 25:49 the degrees of freedom for the numerator should be (Pfit-Pmean), otherwise it is less than 0.
Thanks! In theory UA-cam is supposed to alert people of that typo, but maybe it doesn't always work. (I just tried it and it worked for me).
absolute masterpiece
Thank you!
Around which point do we rotate the line ????????
Beautiful lecture..really easy to understand
There are two different ways to fit the line to data. The one most commonly used is to simply do the math and solve for the optimal fit (take the derivative with respect to the squared residuals and solve for where it is equal to 0). However, that method only works in this specific situation. A more general method is based on the "rotate the line approach" that I illustrate in this video. To learn more about it (how to rotate the line), see my video on Gradient Descent: ua-cam.com/video/sDv4f4s2SB8/v-deo.html
thank you. that was very clear
Thanks!
This is great
Thanks!
Thank you for making this series of statistic videos. One question please: I want to calculate the least squares growth rate of sales for a company. Would I have "higher quality" growth rate by using quarterly sales (40 pieces of data) vs. annual sales (10 pieces of data). Would the seasonality (Christmas sales higher) affects of quarterly sales and distort the growth rate? Thanks,
It sort of depends on how exactly you want to model and what you want to get out of the model. If you want to take seasonality into account, then you need to fit a periodic function (like a sine function) to your quarterly data. That said, the easiest thing to do would be to start with annual sales and see how useful that is.
@@statquest Thank you so much for taking the time to answer my question!
Thank you ever so much!
You're very welcome!
Great video overall! But I'm a little confused with your description of calculating a p-value for the R^2. Does this mean we are treating R^2 as a random variable itself and looking at its distribution? Because to me it seems like it is the f-statistic that follows an f-distribution, hence we are calculating a p-value for the f-stat, not the R^2 itself, which(correct me if I'm wrong) does not follow any specific distribution. So what exactly is the connection between the R^2 and the f-stat and its corresponding p-value?
The f-statistic is what we use to calculate the p-value for the r-squared.
Why was the original Linear Regression video removed for this one? Is the information of this more accurate or clearer?
Without telling me, UA-cam put the original video behind a paywall, so re-uploaded it so it would still be free
I have a question, in 5:24 why the variance is calculated dividing by n instead of n-1, I thought all the observed data points are just a sample of a bigger population includes data points which we haven't observed yet. I'm sorry if my English confuse you because it isn't my mother tongue
In this context, the way we use variation means that denominator will cancel out, so it really doesn't matter which one (n or n-1) we use.
is residual the difference between the observed value of the dependent variable and the predicted value or the difference between the overall mean of the dependent and the observed value
The residual is the difference between the observed and predicted values.
how do you come with the equation
What time point, minutes and seconds, are you asking about?
@26:21 Should the curves say ( P fit- P mean)=1 ?
Yes! That's funny that it's been like that forever, but you finally caught it. Thanks!
@@statquest Haha the credit goes to you for teaching the concepts so well to a newbie! BAM! 😁
what's the difference between RSS and SS(fit) ?
They are the same. However, I changed notation so that I could specify when which model we were using to make the predictions. SS(fit) is the RSS around the fitted line and the SS(mean) is the RSS around the mean.
yo how did u respond to me
I keep track of all of my videos. bam! :)
First :)
BAM! :)
I'm in my stats class but watching this instead of listening to my professor lol 💀
bam!
Double bam!
Ternary Bam!!!
Qudary bam!!!!
Penta bam!
I struggled understanding this topic through a textbook/ professor videos online, and this was just a great explanation. It was like watching this video, made all the pieces finally fit
Hooray! :)
Yo bruh seriously I don't understand anything 😭😞
This assisted me in delivering a presentation for a job interview -- landed the opportunity.
Thanks!
TRIPLE BAM!!! Congratulations!!! :)
I'm an electrical engineer who wanted to learn about machine learning, and your videos helped me understand all the fundamentals of this field. Thank you so much, sir
Happy to help! :)
Does n equals to the number of data points in F equation? For example, we should take 9 for n in 22:40 ?
Yes
I struggle understanding this topic but it is Great to learn from someone who can explain things in a simple manner with eloquence
Thanks!
@@statquest Agreed. You articulate well and make the subject simple and easy to understand.
Wow...i was searching for this on your channel last week and I was so sad I didnt find it... luckily i still have time to study for the test. Thank you!
Good luck! :)
u sound like technoblade
Thanks!
This is implementation of Linear regression from scratch in NumPy only. In-depth explanation of key concepts like Cost Function and Gradient Descent
ua-cam.com/video/wxCQxZKo4hU/v-deo.html
thank you friendly folks of the genetics departement of NC Chapel Hill , greetings from Paris France.
Thank you!
I wish I had your lecture 50 yrs ago.... never too late learning it again today. thank you
Thanks!
Thank God i came accross your videos. Making my CFA journey towards statistics less overwhelming by explaining like you are explaining to a 5 year old...pheeewww.
You got this!
Thank you Josh! You are truly helping me with the difficult reviewers' comments🤣.
Good luck!
Cool merch you could probably easily create would be a workbook to pair with your book where we could practice calculating R2 for exemple in different scenarios. That way, everytime you learn a new concept you can practice doing the formulas :) i'd totally buy that 😏 and maybe links to extra videos or explainations on the concepts that are a little harder to comprehend for people that are completely new to this field and a little slow lol(like linear regression 😅)
That's a great idea!
Hi josh, while getting to R^2, you give the formula y= (data-mean)^2. This contradicts your StatQuest "Fitting a line to the data", where your formula was "(b-y1)^2+(b-y2)^2+...", meaning "(intersect-data)^2. Now i already understood that by squaring the difference you get the same positive value, so the order doesn't matter for this purpose. Is there another reason why you put it in the order "(data-mean)^2" in this video?
Thanks. Love the videos, just watching for fun
Since order doesn't matter, it's hard for me to remember to be consistent.
Okay great, just was wondering if i was missing something here @@statquest
Thank you so much for this video
Happy to help!
This is the best explanation I have ever come accross on Linear Regression. I have a much better intuitive understanding of what the mathematic formulas I was exposed to mean. I do have a question. At 21:42, should the numerator be interpreted as [SS(mean) -SS(fit)]/(Pfit - Pmean) or is it SS(mean) - [SS(fit)/(Pfit - Pmean)] ? The position of the square brackets is not clear to me. Kindly clarify.
It's the former. It should be [SS(mean) -SS(fit)]/(Pfit - Pmean)
I was always wondering why the model chooses to use R2 rather than absolute value of R, until you draw that polynomial out of all sum of squares. It makes sense now
Hooray!
Hello, I had one doubt. For calculating multiple F values, are we taking random samples from our original dataset itself? As in, if there are 100 data points in total, we will take 80, 70 and any random data points from 100 to plot F values on histogram? Could you please help me with this?
The example where we use random data is just an example of the concepts behind how the p-value is calculated. In practice, we use a curve generated by the F distribution (see 25:26) that represents what would happen if we had generated an infinite number of random datasets.
this is amazing, thanks a lot @statquest , please can you also do a video of linear mixed models and generalised linear mixed models, there a few videos about them on UA-cam, it would really helpful. Thank you for the good work
I'll keep those topics in mind.
i had to like just because of the song
bam! :)
this is fucking fantastic
:)
You are indeed a God among mortals. And as such you shall be praised. Tons of gratitude for blessing us with your pristine insight Father Majesty.
Wow, thank you!
At 15:15, how does least squares cause any useless variable to be multiplied by 0? I thought Lasso regression excludes variables.
Least squares can do it in principle, but not very well. Lasso is much more effective, and lasso also works when there are more variables than data.
10QUVM for your valuable presentation!!! You made me feel proud in my STAT!!!
BAM! :)
Your videos are awesome! Thanks a lot for making complex concepts simpler. It will be helpful if you clearly explained Discrete probability distributions
I cover the binomial here: ua-cam.com/video/J8jNoF-K8E8/v-deo.html
wow this make so much sense! I'm pissed why college professors don't teach like this, it was a waste of time to sit in their classes being so confused right from the start. I can't thank you enough for your videos!
Thank you!
Hi Josh,
Very nice video!
Shouldn't the distances from the points to the line be a perpendicular?
If they were perpendicular, than we would lose the relationship between the variable on the x-axis and the variable on the y-axis, and the whole point is to use an x-axis value to predict a y-axis value. Thus, the residuals are parallel with the y-axis - this preserves the relationship that we want to use to make predictions.
Best explanation ❤
Thanks a lot 😊!
thanks for the video
You're welcome!
good video but god this is so cringe
:)
u just earned a subcriber
bam! :)
Hey hi, R squared can be negative as well right?
Not in the context of linear regression. In other contexts, though, it can be.
@@statquest R^2 is just a metric right, and I can set the coefficients of independent variables in such a way that variance(error) exceeds variance(y),( as variance(error) = variance(y* - y), (where y* is the infered value, and y is the actual value) , I can always make y*-y infinitely high for one datapoint, by choosing appropriate coefficients ), or am I wrong? Please correct me.
@@mathematics6199 Yes, in theory, you can do that - but that's not linear regression. In linear regression we don't just set the coefficients to whatever we want. We set them so that they minimize the sum of the squared residuals. And this is why R^2 isn't negative in this context. However, in other contexts, where you can do whatever you want, yes, it can be negative.
@@statquest Thank you so much.
This is an excellent video Josh, thank you! I understand all well until you explain about p-value 23:58. So we were using a dataset of mouse size/weight and weight/tail length/body length, but I'm confusing where the 'random dataset' comes from when you calculate p-value. Could you explain a bit further about this please?
The idea is to give you an intuitive sense of what the p-values associated with linear regression represent. So, to start with, we had 9 data points (9 pairs of weight/height measurements) and fitted a line to it and calculated the F value. That is the "observed" F value generated from the original, raw data. Now pair 9 random values for height (and these could be any reasonable values for height that you randomly select) with 9 random values for weight (and these could be any reasonable values for weight). Calculate the F for those pairs of random values and put that in a histogram. Then repeat until we've done that a lot of times and compare the observed F value from the original data to the histogram.
@@statquest Thanks for explaining all. Much appreciate it. So those 'random values' are completely random, just made up within the range of the normal dataset, right? Then when we are calculating F and p values in SPSS or R, do those softwares go through this process? It might be a bit silly questions, hopefully I'm not too far away!
@@gnosmik That's the idea. However, as mentioned at 25:26, in practice, people (and software) just use an F-distribution (which is an equation for a curved line) to calculate the p-value. The idea of using random data is just to give you an intuition of what the curved line created by the F-distribution represents.
@@statquest Excellent! Thanks Josh
What is the value n (that was mentioned while explaining the degrees of freedom)?
n = the number of data points in the graph.
What is the difference between this video and the previous one in 2017? ua-cam.com/video/nk2CQITm_eo/v-deo.html
I don't think there's a difference - I had to re-release this video (and my other linear models videos) because UA-cam made an error.
Not that it matters here but the shouldn't the sample variance formula have n-1 instead of n?
In this case it doesn't matter.
dude u are funny
Thanks!
Cheer~~~arranged in or extending along a straight or nearly straight line.😊
double :)
awesoommeeeeee!
Thanks!
Love the musical introduction. Such a nice touch to prime you beforehand :)
Thank you!
Awesome video, thank you Prof. Josh!!!!!
Thank you!
This was truly advanced concept for me !!! :)
You can do it! :)
wow
:)
Your explanations are wonderful. Please just recommend the book should be studied with your videos. Please make videos on chi-Squared distribution, Monte Carlo Simulations and Hypotheses testing.
Thanks for your valuable help.
My favorite book to go along with my videos is The StatQuest Illustrated Guide To Machine Learning. You can get it here: statquest.org/statquest-store/
Awesome channel! I just bought your book too!
TRIPLE BAM!!! Thank you very much for supporting StatQuest!!!
Question. Why are we calculating R2 value and the p value? Is it the industry standard? Or else What led to the decision that you included it with linear regression. Theoretically Lin reg is complete before that right?(Making concepts clear)
If you just want to fit a line to data, you can used the method of least squares. However, if you want to quantify how well that line fits your data, then you use Linear Regression. Linear Regression consists of using least squares to fit the line to the data and then calculating r^2 and its p-value to evaluate how well that line fits the data.
@@statquest still confused.. as you said 'how well it fits the data', so the r2 and p value are tests for evaluation right? dont they have alternatives? or is it necessary to do exactly these steps. I'll still get a logistic regression model but it may not be the best one without them?
Or are you saying that these, or some other alternatives tests are necessary to do, to assess the model and this repeats iteratively until best fit?
@@Slayer1407-d9d They do have alternatives, so, as you say, you might think of r^2 and its corresponding p-values as the 'industry standards'. Pretty much every program that offers a linear regression function will give you those as outputs. However, there are alternatives, and you can read more about them here: developer.nvidia.com/blog/a-comprehensive-overview-of-regression-evaluation-metrics/ among other places.
@@statquest Thanks a lot for clearing that
Thanks!
BAM!!! Thank you so much for supporting StatQuest!!!
Of course! I am the person who is embarrassed on the inside that I don't get the stats terms when thrown around at work, but know that I'm memorized them so know what they are, but really don't understand the "why" or how it all relates. Thank you so much for speaking slowly in your videos, reiterating concepts, sometimes with additional concepts in between, and your humor. It's fun. I'm grateful. @@statquest
I had to buy a study guide book after watching this video...! This is a great video!!
Thank you so much for your support!
I am not able to find the video 'Fitting a line to the data'
I have contacted UA-cam about this problem, but, unfortunately, they are all on vacation until next week. :( The good news is that this video does a pretty good job summarizing the concepts in that other video.
Thank you. Wonderfully explained!!
Glad it was helpful!
I got pregnant two times while learning SGD from you. This is the hundredth time i'm jumping from a video to another video.
ok
I would be lost without this channel
bam! :)
I always have a good time with Statquest :3
bam!
can you do Quantile Regression?
I'll keep that in mind.
I think at ua-cam.com/video/7ArmBVF2dCs/v-deo.html the slide meant to say (SS(mean) - SS(fit))/(p_fit - p_mean) for the numerator?
Yep
I am enjoying this teaching method 😍
Thank you!
its like years since u uploaded this
I know! This one is classic! It might even be "pre BAM!"
could i ask u my doubts in this video
Sure!
thank you for this
Thanks!
Awesome, but can we do this without squaring? Why can't we just sum the residuals without any squaring, it looks like it should give us the sum of all distances and then we could plot it in the same way and pick the rotation that gives us the least sum of non-squared residuals and it should still work, curious why do we choose to square it, thank you so much for the video
If the "distances" below the line are negative, they will cancel out the ones above them, so that's a problem. However, we could then take the absolute value so that everything is positive. This could work if Linear Regression was actually solved the way I've presented it here. However, in practice, when you square the distances, you can solve for the optimal parameters directly by taking the derivative of the squared residuals with respect to each parameter, setting those derivatives equal to 0 and then solving for the parameter values.
@@statquest Thank you so much , it makes sense now
you are a genius!
Thanks!
This video is BAMMMMMMMMMM
Thanks! :)
so is mouse size a confounder?
What time point, minutes and seconds, are you asking about?
I don't understand why least squares can cause any term that will make ss(fit) worse to be multiplied by 0. Is it because mean squares differential the equation?
15:20
or is it because things like ridge regression can shrink the coefficients to 0?
Least squares minimizes the sum of the squared residuals and if setting a parameter = 0 reduces the SSR, then that's what will happen.
I didn’t get what is actual R
It's the correlation coefficient. For details, see: ua-cam.com/video/xZ_z8KWkhXE/v-deo.html and ua-cam.com/video/2AQKmw14mHM/v-deo.html
Please add this video to the linear regression playlist and remove the existing video from there as it doesn't open
I'm working on getting the original video out from behind the paywall. I've contacted UA-cam but they're on holiday until next week.
insanely good video
Thank you! :)
Very helpful. Thank you
Thanks!
Very nice, thank you
:)
Thank you for the nice video! I wonder for your explanation to the F curves around 25:53, shouldn't it be (p_{fit} - p_{mean})=1? In addition, would you please provide the link to your video about the degrees of freedom if that is already available?
Yes, that is a typo. And, unfortunately, I haven't made the degrees of freedom video yet. However, it's still on the todo list.
@@statquest Thank you! I look foreward to your new ones
@@statquest looking forward to the degrees of freedom video too!
Thank you so much Josh !
Thanks!
how is math cool ??????
In so many ways.
Hats off to StatQuest!!!
Thank you!
Thanks for the video. Could you please explain more why SS(fit)/(n-pfit) instead of n here 22:48? Thanks a lot.
This has to do with "degrees of freedom" and one day I hope to cover that topic in full.
@@statquest Looking forward to the degrees of freedom video! Parameters have always been a confusing topic for me
Amazing explanation
Thanks!
Thank you :)
You're welcome!
Legend
Thanks!
Bam! Bam! Bam!
Triple bam!!! :)
so is R square , a correlation coefficient?
It is the square of the correlation coefficient.