SVM Dual : Data Science Concepts

Поділитися
Вставка
  • Опубліковано 17 вер 2024

КОМЕНТАРІ • 93

  • @robertcanberkozturk7725
    @robertcanberkozturk7725 Рік тому +13

    The only man who actually explains the concepts on the internet. God bless your soul Ritvik

  • @mental4508
    @mental4508 5 місяців тому +5

    Man I swear you are one of the very few that actually understands what it takes to make these concepts understandable.

  • @jjabrahamzjjabrhamaz1568
    @jjabrahamzjjabrhamaz1568 3 роки тому +46

    Ritvik you my man are a godsend. Thank you for sharing your extremely
    technical expertise on youtube for free. You teach this material better
    than any platform (university or online).

  • @teegnas
    @teegnas 3 роки тому +28

    Important video for those who want to understand SVM properly ... thanks for uploading!

  • @SS-xh4wu
    @SS-xh4wu 2 роки тому +4

    Thank you so much for your videos. As a stats major person, I still learned so much from your channel. In many cases I learned the math in school, but no one ever talked about the intuition behind. Really appreciate your work here

  • @alaham2590
    @alaham2590 5 місяців тому +1

    You just give the exact amount of details needed to grasp these concept ! very nice !! thx

  • @SHREEANSHMOHANTY
    @SHREEANSHMOHANTY 9 місяців тому +2

    I have watched 6 times now, still can't wrap my head around it. Enjoy your views!!!

  • @pranjalpatil9659
    @pranjalpatil9659 2 роки тому +1

    althogh this dosen't include all the maths in depth but this is enough for most of us to understand SVM formulation very well

  • @atrus3823
    @atrus3823 Рік тому

    Along with StatQuest, these are by far the best ML videos on the UA-cam. Thank you!

    • @ritvikmath
      @ritvikmath  Рік тому

      Wow, thanks!

    • @atrus3823
      @atrus3823 Рік тому

      @@ritvikmath No problem! I took SVM's in school, and have read about them many times, and I have only ever seen 1 of 2 levels of explanation: 1) just stating the results without question, or 2) diving deeply into measure theory and other crazy math that though sounds interesting, but I don't really have time for. This is the first source I've found that explains the gist of how the kernel trick works without diving super deeply into the math.

  • @CodeEmporium
    @CodeEmporium 3 роки тому +4

    I love the way you throw the marker and hope to catch it. And if it's too far, you don't snap and point at the camera. Ah. And Nice videos. It comforts my soul

    • @ritvikmath
      @ritvikmath  3 роки тому +1

      hahaha, thanks for noticing something I wasn't even aware of :)

  • @adam_long_unique_name
    @adam_long_unique_name Рік тому +1

    Thanks for these great videos.
    I would suggest to reorder the playlists to single topics - for example SVM,
    and add in the description links to the playlist or all other videos on the topic

  • @mohamedgaal5340
    @mohamedgaal5340 Рік тому

    Thank you so much for explaining why we need the dual formation. This video made me understand some of the concepts I jotted down from other videos without understanding.

  • @Trubripes
    @Trubripes 4 місяці тому

    Amazing that SVM is derived from KKT conditions. Never noticed that until watching this.

  • @kally3432
    @kally3432 3 роки тому +2

    Great explanation, I enjoyed it a lot. Thanks

  • @yitang7577
    @yitang7577 2 роки тому

    Thanks for the wonderful video!!! It's really benefit for me who don't know the Lang and Dual problem.

  • @amrdel2730
    @amrdel2730 6 місяців тому

    Thanks for the explanation of crucial points in this topic , thank you for the effort

    • @ritvikmath
      @ritvikmath  6 місяців тому

      Glad it was helpful!

  • @user-xi5by4gr7k
    @user-xi5by4gr7k 3 роки тому +4

    Fantastic video. I have been binging your content. Any chance you will make a series on stochastic calculus?

    • @ritvikmath
      @ritvikmath  3 роки тому +2

      thanks for the suggestion!

  • @lumos309
    @lumos309 3 роки тому +2

    Hello, regarding the efficiency of the two forms of the problem: What about the inner product gives the dual form a complexity of N^2 rather than NP? It seems that the inner product operation would have complexity O(P) since it is dependent on the number of terms in the input vectors x(i) and x(j); similarly W(T)x(i) in the primal form also has complexity O(P). So this O(P) term shows up in the same place in both forms, which would mean they are both dependent on P. Is there some other place in which only the primal form complexity scales with P? Or is complexity not even the right way to analyse this?

  • @annakotova3508
    @annakotova3508 2 роки тому

    Very-very clear explanation! Thanks!

  • @sejmou
    @sejmou Рік тому +1

    Thanks for another wonderful video!
    However, there's one thing I really want to understand: in the formulation of the hard margin SVM problem (the 'SVM Math' video) you indeed stated that we want to minimize ||w||, not 1/2 ||w||^2. where does this difference come from, and why are both approaches equivalent? can anybody shed some light on this?

    • @k_1_1_2_3_5
      @k_1_1_2_3_5 8 місяців тому

      I am in the same boat.. really would like to understand that

  • @marisa4942
    @marisa4942 2 роки тому

    Thank you so much!! Very clear and intuitive explanation!!

  • @shashankbarole
    @shashankbarole 3 роки тому +2

    Please make more videos on SVM!!
    Thanks a lot

  • @bryankelly8820
    @bryankelly8820 2 роки тому +2

    Thank you for posting this informative video. You mention that alphas only need to be calculated for support vectors. That does simplify things considerably; however, how can one in practice determine which vectors are support vectors without doing the minimisation?

  • @muhammadaliyu3076
    @muhammadaliyu3076 3 роки тому +10

    I think you are one of the great teachers on youtube. But I think the only reason I understood this video is because I already have knowledge about Optimization, linear algebra and Multivariable Calculus. And also because I already understood dual SVM problem. To be honest, some of your videos are totally not for beginners. I think you should try to be proving the math behind the algorithms you are explaining from the first principal. Its better for beginners.

    • @ritvikmath
      @ritvikmath  3 роки тому +15

      thanks for the feedback! It's important for me to strike a balance between making the videos accessible to everyone and covering complex topics.

    • @jjabrahamzjjabrhamaz1568
      @jjabrahamzjjabrhamaz1568 3 роки тому +6

      @@ritvikmath I think CrimaCode might not be fair here, because if you even come to this video for SVM's you will have atleast some knowledge about the realm of linear algebra and multivariate calculus. Both are basic things taught at most highschool levels. I think ritvik is doing a great job and video is dope for beginners.

    • @kroth5810
      @kroth5810 3 роки тому +2

      I found the video very helpful despite not being well versed in svm before watching it. People learn in different ways :)

    • @quant-prep2843
      @quant-prep2843 3 роки тому

      @@kroth5810 okay dicc

    • @thirumurthym7980
      @thirumurthym7980 3 роки тому

      @@ritvikmath agree. May be you can try to cover basics of those basics separately for the pure beginners. But all your videos are awesome.

  • @liq3395
    @liq3395 3 роки тому +5

    Thanks for the great teaching!!! Just have 1 question that why it was Max{Σα_i - 1/2(Σ.....)} after substituting w, but not Max{Σα_i + 1/2(Σ.....)}? I tried for several times and still got "+" not "-", might I ask for illustration on this calculation?

    • @mateuszserocki4026
      @mateuszserocki4026 2 роки тому +3

      Hi man, I stuck with the same problem but we did it wrong! Resolved it finaly, first element w^T*w is reduced to 0! The secound element - sum(alpha_i * y_i * w^T*x_i +... you have to extend that w^T here then you have double sum that is not reduced to zero and here you go with this minus. Hope it helps you

    • @adefajemisin
      @adefajemisin Рік тому

      When you substitute, the first two terms are the same, except the first one in multiplied by 1/2. 0.5x - x = -0.5x.

  • @winstonong9593
    @winstonong9593 2 дні тому

    How are the support vectors determined in the first place, such that only pairs of support vectors need to be considered for the final minimisation problem?

  • @TheKingLikesArt
    @TheKingLikesArt 2 роки тому

    You are a legend my man

  • @johnspivack6520
    @johnspivack6520 8 місяців тому

    Wonderful video thank you

  • @lilianaaa98
    @lilianaaa98 4 місяці тому

    The math this video covers is kind of sophisticated :(

  • @e555t66
    @e555t66 Рік тому

    Really explained well. If you want to get the theoretical concepts one could try doing the MIT micromasters. It’s rigorous and demands 10 to 15 hours a week.

  • @anthonyburn6505
    @anthonyburn6505 6 місяців тому

    Excellent videos

  • @abhishekchandrashukla3814
    @abhishekchandrashukla3814 2 роки тому

    apostle of machine learning !!

  • @emptygirl296
    @emptygirl296 2 роки тому

    You are a legend really

  • @arminhejazian5306
    @arminhejazian5306 Рік тому

    great explanation

  • @SwapravaNath
    @SwapravaNath 6 місяців тому

    11:30 but while solving the optimization, we don't know what are the support vectors. so, we need to solve the convex program for all cross terms but we'll find \lambda to be zero for most of them.

  • @TheJuankpunk
    @TheJuankpunk 8 місяців тому

    Hi @ritvkikmath, thank you for these videos. In what type of degree do you usually cover these subjects? I would like to enroll in one.

  • @nimeesha0550
    @nimeesha0550 3 роки тому

    Keep going buddy!! Amazing work. Really helpful :) Thank You.

  • @EW-mb1ih
    @EW-mb1ih 2 роки тому

    At 6:30 Why do we take the alpha that MAXIMIZE the solution to the inner minimization?

  • @alexz3346
    @alexz3346 2 роки тому

    Cool jacket!

  • @johnstephen8041
    @johnstephen8041 5 місяців тому

    Bro - Please discuss on the VC dimension

  • @sagarpadhiyar3666
    @sagarpadhiyar3666 2 роки тому +1

    Hello ritvik,
    superb video of dual SVM. nicely explained. but I have one doubt in final equation of dual form of SVM. Where is that xj term coming from?? actually, what is that term?? if xi is our training data then what is xj??

    • @SwapravaNath
      @SwapravaNath 6 місяців тому

      a different training instance.

  • @nuamaaniqbal6373
    @nuamaaniqbal6373 2 роки тому

    You are the boss!

  • @LingkarPeduli
    @LingkarPeduli Рік тому

    can you make video explain about twin support vector machine. thank's in advances

  • @kanishkgarg423
    @kanishkgarg423 Рік тому

    thanks, this was very helpful 😀😀

  • @AmithAdiraju1994
    @AmithAdiraju1994 2 роки тому

    _/\_ for this.
    I do have on question though, regarding non-support vectors requiring alpha(i)'s to be zeros.
    Intuitively, that would mean only support vector data points ( which are very few ) would contribute to optimal weights of the model, wouldn't that be bad for model weights ? i.e., only few contributions ?
    I'm wondering if model would not generalize well with only few contributions from few data points ?

  • @nieeel9112
    @nieeel9112 2 роки тому

    super clear! really helpful:)

  • @TheInevitable360
    @TheInevitable360 Рік тому

    Thank you very much for this great video. I have a question: you said that for non-support vectors the alpha values would be zero due to the fact that they have no contribution to the final solution. But here we have considered the Hard Margin version of the SVM. how about the soft margin. because in the soft margin, all the points have a contribution to the solution and we can't ignore non-support vectors. are they still 0 or not?

  • @xinking2644
    @xinking2644 2 роки тому

    good video with clearly explain!!!!

  • @zhangeluo3947
    @zhangeluo3947 Рік тому

    Hello, sir I have 2 doubts: 1. Once you found the optimal values for alpha, how to determine the optimal value for b(bias)?
    2. You said that for non-SVs, the alpha value is just 0, but how to determine those non-SVs in the real data set? I know those non-SVs are outside the bottom lines for -1 and 1, but in practice, for given alpha values for each data point, you can get the optimal values for w or b , then you can get the decision boundary given w and b based on that set of alpha values, so to use that decision boundary to evaluate the support vectors and non-SVs? Is that true?
    Thank you!

  • @shaktijain8560
    @shaktijain8560 2 роки тому

    Thanks for the video 😄😄

  • @prajwalsatannavar4576
    @prajwalsatannavar4576 Рік тому

    Thank u...

  • @ziadtarek633
    @ziadtarek633 3 роки тому

    you're brilliant

  • @tryfonmichalopoulos5656
    @tryfonmichalopoulos5656 2 роки тому

    Using stationarity we can get rid of 'w' but how did we get rid of the 'b' that does not appear in the dual formulation of the problem on the top right?

    • @gautamchopra9939
      @gautamchopra9939 2 роки тому

      I think this is due to the dL/db = Σα_i*y_i. When we plug these value we have b*Σα_i*y_i such that the term containing b becomes 0. We don't have an explicit statement that tells us that b = 0.

  • @johnspivack6520
    @johnspivack6520 8 місяців тому

  • @RahulDable
    @RahulDable 3 роки тому

    thanks

  • @hari8568
    @hari8568 3 роки тому

    Hey when you code this out how exactly would he choose the alphas like what are the upper bounds of alpha?

  • @prysrek8858
    @prysrek8858 3 роки тому

    You're great

  • @GanadoFO
    @GanadoFO 2 роки тому

    Once we get the point where we are trying to minimize the equation on the right of the screen, I don't understand how you actually do the minimization.
    Let's say we have 3 support vectors, so we will have a function of three variables, alpha_1, alpha_2, alpha_3. How do you minimize a multi-variable function? What does that even mean? I've only ever done minimization of one variable.

  • @ayushtiwari6932
    @ayushtiwari6932 3 роки тому

    The term |w| is squared only for mathematical convenience ?

    • @SwapravaNath
      @SwapravaNath 6 місяців тому

      yes, and makes the objective function differentiable as well.

  • @PF-vn4qz
    @PF-vn4qz 3 роки тому

    how did you get rid of b?

  • @mwave3388
    @mwave3388 2 роки тому

    Thank you. But from scratch, it is too difficult.

  • @ryanleyland5565
    @ryanleyland5565 Рік тому

    Kumar?

  • @johnspivack6520
    @johnspivack6520 8 місяців тому

    Wonderful video thank you

  • @johnspivack6520
    @johnspivack6520 8 місяців тому

    Wonderful video thank you

  • @johnspivack6520
    @johnspivack6520 8 місяців тому

    Wonderful video thank you

  • @johnspivack6520
    @johnspivack6520 8 місяців тому

    Wonderful video thank you