This guy is super smart and he takes sophisticated concepts and explains it in a way where it's digestible without mocking the theory! What a great teacher!
I can't explain how grateful I am for your channel! I am doing an introductory machine learning course at Uni and it's extremely challenging as it's full of complex concepts and the basics aren't explored throughly. Many videos I came across on youtube were too overly simplified and only helped me very briefly to make sense of my course. However, your videos offer the perfect balance, you explore the complex maths and don't oversimplify it, but do so in a way that's easy to understand. I read through this concept several times before watching your video, but only now do I feel as if I TRULY understand it. I HIGHLY appreciate the work you do and look forward to supporting your channel.
This has been simultaneously the simplest, most detailed and yet most concise explanation of this topic I've come across so far. Much appreciated! I hope you keep making awesome content!
Question on the notation. The image shows that the vector between the central line and decision line is w. So, I think, that w is the length of the decision boundary. But then we go on to show that the length of the decision boundary is k=1/||w||. So I'm not clear on what w (or k, for that matter) are actually representing.
I'm a PhD student studying data mining and I just wanted commend you for this SUPERB explanation. I can't thank you enough for the explaining this so clearly. Keep up the excellent work!!
Just to add onto all the love, I'm a data scientist in marketing and you are my number one channel for reviewing concepts. You are a very talented individual!
This is the best and the most intuitive explanation for SVM. It is really hard for me to actually read research papers and understand what story each line of the equation is telling. But you made it soo intuitive. Thanks a ton! Please Please make more videos like this
Great video as usual! A possible side note - I find 3d picture even more intuitive. Adding z-direction which is basically can be shrunk to [-1;1] is our class prediction dimension and x1 x2 are feature dimensions. Hence, the margin hyperplane "sits" exactly on (x1; x1; 0) This is also helpful for further explanation of what SVM kernels are and why kernel alters the norms (e.g. distances) between data points, but not the data points themselves.
Dude thank you! now these equations don't feel like they were pulled out of thin air. and the best part is I can work them out too! I haven't done linear algebra in almost a decade so I got stuck on the ||w||/(w*w) part for a good bit but this pushed me to refresh some concepts and figure it out! Thank you
thanks! I did indeed kind of skip a step. The missing step is that the dot product of a vector with itself is the square of the magnitude of the vector. ie. w · w = ||w||^2
I love your channel. You explain difficult concepts that could be explained to my dear grandmother who never went to college. Excellent job sir! You should become a professor one day. You would be good.
In case you're also having trouble figuring out how we arrive at k=1/||w|| from k * (w*w/||w||) = 1: remember that the dot product of any vector with itself is equal to its squared magnitude. Then, w*w can also be expressed as ||w||^2. ||w||^2/||w|| simplifies to just ||w||. Finally bring ||w|| to the other side by dividing the whole equation by ||w||, and you're done :) if you also have trouble understanding why exactly the dot product of any vector with itself is equal to its squared magnitude it also helps to know that the magnitude of a vector is the square root of the sum of squares of its components and that sqrt(x) * sqrt(x) = x I hope that somehow makes sense if you're struggling, surely took me a while to get that lol
🌟Magnificient🌟I actually understood this loss function in by watching once. Very nice explanation of math. I saw lot of other lectures but you cant understand math without graphical visualization.
thank you for your genius explanation. At 5:11, before getting the value k, the equation k * ( w * w) / (magnitude of w) = 1 contains w * w, why the output k doesn't have w in the end.
Amazing explanation from the theoretical to the mathematical. Please tell me how you do it? So i can self-learn myself how you are able to understand and then explain these concepts or other concepts. what resources do you use ?
Hi, how exactly did you choose 1 and -1, the values for wx -b where x is a support vector? wx-b = 0 for x on the separating line makes sense however. Could it have other values?
I'm not sure but I think you forgot to say that in order to have margin = +-1 you should scale multiplying constants to w and b. Otherwise I don't explain how we could have distance of 1 from the middle The rest of the video is awesome, thank you very much :)
@ritvikmath - Thanks for this great explanation. I have noticed other material online advises the equation for the hyperplan is w.x+b=0 rather than w.x-b=0. Can you confirm which is accurate
Equation for points on margins are: w.x - b = 1 w.x - b = -1 That means we have fixed our margin to "2" (from -1 to +1). But our problem is to maximize the margin, so shouldn't we keep it a variable? like: w.x - b = +r w.x - b = -r where maximizing r is our goal?
This guy is underrated for real. UA-cam - throw him into recommendations.
I know... I recommend him all the time on Reddit.
True! He deserves way more subscription. He should prepare a booklet like statquest did but of his own. Would definitely buy it!
True!!
This guy is super smart and he takes sophisticated concepts and explains it in a way where it's digestible without mocking the theory! What a great teacher!
I can't explain how grateful I am for your channel! I am doing an introductory machine learning course at Uni and it's extremely challenging as it's full of complex concepts and the basics aren't explored throughly. Many videos I came across on youtube were too overly simplified and only helped me very briefly to make sense of my course. However, your videos offer the perfect balance, you explore the complex maths and don't oversimplify it, but do so in a way that's easy to understand. I read through this concept several times before watching your video, but only now do I feel as if I TRULY understand it. I HIGHLY appreciate the work you do and look forward to supporting your channel.
same
This has been simultaneously the simplest, most detailed and yet most concise explanation of this topic I've come across so far. Much appreciated! I hope you keep making awesome content!
Glad it was helpful!
You answered all the questions I had in mind without me even asking them to you. This was an amazing walkthrough. Thank you!
Another great video on SVM. As a mathematician I do appreciate your succinct yet accurate exposition not playing around with irrelevant details.
Question on the notation.
The image shows that the vector between the central line and decision line is w. So, I think, that w is the length of the decision boundary. But then we go on to show that the length of the decision boundary is k=1/||w||. So I'm not clear on what w (or k, for that matter) are actually representing.
I too expected k to equal the length of that vector w :-/
I'm a PhD student studying data mining and I just wanted commend you for this SUPERB explanation. I can't thank you enough for the explaining this so clearly. Keep up the excellent work!!
Just to add onto all the love, I'm a data scientist in marketing and you are my number one channel for reviewing concepts. You are a very talented individual!
This is the best and most comprehensible math video on hard margin SVM I have seen till date!
This is the best and the most intuitive explanation for SVM. It is really hard for me to actually read research papers and understand what story each line of the equation is telling. But you made it soo intuitive. Thanks a ton! Please Please make more videos like this
I think this might be top 5 explanations of SVM mathematics all-time. Very well done
Great video as usual!
A possible side note - I find 3d picture even more intuitive.
Adding z-direction which is basically can be shrunk to [-1;1] is our class prediction dimension and x1 x2 are feature dimensions.
Hence, the margin hyperplane "sits" exactly on (x1; x1; 0)
This is also helpful for further explanation of what SVM kernels are and why kernel alters the norms (e.g. distances) between data points, but not the data points themselves.
I finally get svm after watching a lot of tutorial on UA-cam. Clever explanation. Thank you
Very easy to follow the concept! Thanks for this wonderful video! Looking forward to seeing next video!
You and statquest are the perfect combination :) Thanks for all of your hardwork.
Dude thank you! now these equations don't feel like they were pulled out of thin air. and the best part is I can work them out too! I haven't done linear algebra in almost a decade so I got stuck on the ||w||/(w*w) part for a good bit but this pushed me to refresh some concepts and figure it out! Thank you
studying my masters in data science and this is a brilliant easy to understand explanation tying graphical and mathematical concepts - thank you!
So simple, so clear!!! Wish all the teachers are like this!
Thank you so much for this video! I am learning about SVM now and your tutorial perfectly breaks it down for me!
this guy explained what my professors couldn't explain in 2 hours 😂😂😂
your videos are what allowed me to take a spring break vacation bro, saved me so much time thank you
Great to hear!
This is giving "Jacked Kal Penn clearly explains spicy math" and | am HERE for it
Just Amazing Clarity of Topics!!
That's what i've been waiting for! Thanks a lot. Great video!
Glad it was helpful!
At 5:10, I don't get how you obtain K from the last simplification. Can you/someone please explain?
Btw beautiful video!
thanks! I did indeed kind of skip a step. The missing step is that the dot product of a vector with itself is the square of the magnitude of the vector. ie. w · w = ||w||^2
@@ritvikmath right, thank you!!
I love your channel. You explain difficult concepts that could be explained to my dear grandmother who never went to college. Excellent job sir! You should become a professor one day. You would be good.
Absolutely amazing channel! You're a great teacher
Great, thanks for this lucid explanation about the math behind SVM
It's so easy to understand thi s math stuff! Best explanation ever in such a short video.
Thank you for this video. Thanks for simplifying SVM.
Great video on SVM. Simple to understand.
This is a serious good stuff video. I have not seen a better svm explanation
Best high-level explanation of SVMs out there, huge thanks
Glad it was helpful!
Thank you so much. This is what i have been looking for so long time. would you please do the behind other ML and DL algorithms.
You explained this topic really well and helped me a lot! Great work!
Thanks for such brilliant explanation really appreciate your work!!
In case you're also having trouble figuring out how we arrive at k=1/||w|| from k * (w*w/||w||) = 1:
remember that the dot product of any vector with itself is equal to its squared magnitude. Then, w*w can also be expressed as ||w||^2.
||w||^2/||w|| simplifies to just ||w||. Finally bring ||w|| to the other side by dividing the whole equation by ||w||, and you're done :)
if you also have trouble understanding why exactly the dot product of any vector with itself is equal to its squared magnitude it also helps to know that the magnitude of a vector is the square root of the sum of squares of its components and that sqrt(x) * sqrt(x) = x
I hope that somehow makes sense if you're struggling, surely took me a while to get that lol
I almost forget this rule, thank you brother for saving my day
Such a clear explanation! Thank you!!!
very informative and helpful video to help understand the SVM! Thanks for such a great video! You deserve more subscribers
Thanks man great explaination , was trying to understand the math for 2 days , finally got it
Glad it helped!
very helpful! I always wanted to learn math behind the model! thanks!
Best video on large margin classifiers 👍
🌟Magnificient🌟I actually understood this loss function in by watching once. Very nice explanation of math. I saw lot of other lectures but you cant understand math without graphical visualization.
Excellent explanation Ritvik
Youre so unbelieveble good in explaining :)
This really helped me learn the math of svm thanks !!
I am very happy that I found Your YT Channel Awsome Videos I was unable to Understand SVM UntilNow !!!!
Amazing teaching skills - Thanks, a lot!
great, concise explanation !
What an amazing video bro. Keep going.
Once again, ritvikmath being a lifesaver for me. If I understand the underlying math behind this concepts, it is because of him
Amazing explanation!
Thank you Sir . You really simplified the concept. I have subscribed already waiting patiently for more videos 😊
Nice explanation and really easy to follow!
Incredible video
thank you for your genius explanation. At 5:11, before getting the value k, the equation k * ( w * w) / (magnitude of w) = 1 contains w * w, why the output k doesn't have w in the end.
Amazing explanation from the theoretical to the mathematical. Please tell me how you do it? So i can self-learn myself how you are able to understand and then explain these concepts or other concepts. what resources do you use ?
Pls also make one for svm regression.. you are amazing
brilliant explanation!
Great explanation!
Great video ! Why we can assume that right hand side of wx - b in those three lines is 1, 0, -1 ?
Very well explained, thank you !
Eagerly waiting for your video on SVM Soft margin :D
Loved it!
Hey Ritvik, Nice video, can you please cover the kernalization part too.
Holy shit what a banger of a video this is
This is very clearly defined. Thank you.
But could someone explain to me what w is? How can I visualize it and calculate it.
You explained this topic perfectly! Amazing!
Glad you think so!
very informative and intuitive
Hi ritvik! I wonder what is the geometric intuition of the vector w? We want to minimize ||w||, but what does w look like on the graph?
Easily Explained 👍,
Can you also explain how does SVM works with respect to regression problems?
Thank you so much!
Wow, that was so well explained.
You are an amazing elucidator👍
phenomenal
That was crystal clear !
what about the points within the margin? are they support vectors as well?
it was amazing thankyou so much
BRILLIANT!
Hi, how exactly did you choose 1 and -1, the values for wx -b where x is a support vector? wx-b = 0 for x on the separating line makes sense however. Could it have other values?
you are my savior
I'm not sure but I think you forgot to say that in order to have margin = +-1 you should scale multiplying constants to w and b. Otherwise I don't explain how we could have distance of 1 from the middle
The rest of the video is awesome, thank you very much :)
Smart! This is the easiest way to come up with the margin when given theta (or weight)... gosh..
Great video, great underappreciated channel! Thank you and keep up the good work!
Could you do the math behind each Machine learning algorithm, also would you be doing Neural Networks in the future?
along with the assumptions of supervised and un-supervised ML algorithms that deals specifically with structured data.
Yup neural nets are coming up
@@ritvikmath CNN's and Super Resolution PLEASE PLEASE PLEASE
You should mention that your W is an arbitrary direction vector of the hyperplane. (it is not the same size as the margin)
So good .. ThnQ
Bro, you're a superhero
@ritvikmath - Thanks for this great explanation. I have noticed other material online advises the equation for the hyperplan is w.x+b=0 rather than w.x-b=0. Can you confirm which is accurate
great video as always. thank you
Glad you enjoyed it!
Great explanation
explained it so well.
Gifted teacher!
Thank you! I am wodering why do we use "+1 and -1" instead of "+1 and 0" to classify these two areas?
thank you so much
Are SVMs only useful for binary classification, or can they be extended to multi-class predictions?
you are the smartest person I know
super explanation
Equation for points on margins are:
w.x - b = 1
w.x - b = -1
That means we have fixed our margin to "2" (from -1 to +1). But our problem is to maximize the margin, so shouldn't we keep it a variable? like:
w.x - b = +r
w.x - b = -r
where maximizing r is our goal?
Have you figured it out?