Likelihood Estimation - THE MATH YOU SHOULD KNOW!

Поділитися
Вставка
  • Опубліковано 26 вер 2024

КОМЕНТАРІ • 98

  • @CodeEmporium
    @CodeEmporium  2 роки тому +13

    Thanks for watching! Checkout the description for the MEDIUM article (published in Towards Data Science) that accompanies this video. Hopefully that should answer questions. Also please follow here and on medium for fun updates like this!

    • @vsugar10101
      @vsugar10101 2 роки тому

      Tell me how do I use intuition vs probability to predict outcome of my 5 lottery deep training model? 😂

    • @undertaker7523
      @undertaker7523 Рік тому

      Could you please explain why we used mean and standard deviation when attempting to calculate the likelihood?

    • @CodeEmporium
      @CodeEmporium  Рік тому

      I can’t speak for every case. But in linear regression, we assume the distribution of the labels follows a normal distribution. And the normal distribution can be characterized by a mean and standard deviation. And if you substitute this is the “maximum likelihood estimation”, the math with simplify to optimizing the residual sum or squares ( which is proportional to the mean squared error ) to compute the coefficients in the linear regression hypothesis.
      I explain this in the entire probability and likelihood videos too if that helps

    • @sameepshah3835
      @sameepshah3835 4 місяці тому

      Thank you so much. Such a good and simple explanation sir.

  • @imamibo
    @imamibo Рік тому +8

    It was probably a subject that I had been trying to clarify in my head for a month and could not clarify it. Maybe because I'm a little detail-oriented. Thanks to you, brother, I understood the subject. Thanks to UA-cam, you have a brother from the other side of the world. Thank you very much.

  • @maulikshah9078
    @maulikshah9078 Місяць тому

    finally, my searching of 2-3 hours and many videos on the likelihood rests. thanks man...

  • @songweimai6411
    @songweimai6411 6 місяців тому +1

    Thank you. Studying mathematics and statistics in college. I really like this video. My professor told me that “ the most important thing for statistics is : you have to understand the basic logic first using a basic example or daily life example, know what u want and what you need to do“. The second important thing is to “remember the notation and to read the books and study by myself. I really like the first part of the video---that’s the key and core idea for most likely function. Why i watch this video! 😂, Because want to refresh the idea. Doing harder problems with only notations and symbols, get lost.

  • @srinathkumar1452
    @srinathkumar1452 Рік тому +4

    This is such a clear explanation. Great job my dude

  • @alexandrupapiu3310
    @alexandrupapiu3310 2 роки тому +7

    This is great! However it's really important to not confuse the probability density function (p(x)) with the probability of x. For one p(x) can be larger than 1!

  • @stochasticNerd
    @stochasticNerd Рік тому +4

    it's great that you thought of making a video on comparison between probability and likelihood. However, I think in the initial graphs, the y-axis do not represent probability values. They are probability-density values at various x.

  • @chyldstudios
    @chyldstudios 2 роки тому +6

    Very well done, clear and concise!

    • @CodeEmporium
      @CodeEmporium  2 роки тому

      Thank you! My first time trying this style out. So I’m glad it turned out well :)

  • @hannes7218
    @hannes7218 Рік тому +1

    one of the best explanations on youtube! well done sir!

  • @yvettewang7221
    @yvettewang7221 4 місяці тому

    This is the best explanation of likelihood function. thank you so much for the video.

  • @ShashankData
    @ShashankData 2 роки тому +2

    Another 🔥video! This man has an insane brain

    • @CodeEmporium
      @CodeEmporium  2 роки тому +1

      Thanks Shashank! I’m just happy it’s useful 🙂🙂

  • @corradoforza
    @corradoforza 2 роки тому +3

    This is very well explained, thank you!

  • @JingzhuChenClassicalPiano
    @JingzhuChenClassicalPiano Рік тому +1

    Thank you so much!! You made complicated concepts so easy to understand!!! Thanks again!

    • @CodeEmporium
      @CodeEmporium  Рік тому

      Super welcome and also very glad to hear :D

  • @lit-library
    @lit-library Рік тому

    Thank you! My confusion goes away after watching this. Thumb up.

    • @CodeEmporium
      @CodeEmporium  Рік тому

      You are very welcome. Thanks for watching !

  • @abhalla
    @abhalla 8 місяців тому

    You also take the logarithm of both sides because that leads to nice properties when differentiating (because log is strictly increasing, it maintains the property that if x1 < x2, then l(x1) < l(x2)). Addressing arithmetic underflow is definitely a useful added benefit too.

  • @shayanghafoori8222
    @shayanghafoori8222 7 місяців тому

    Seriously, one of the best explanations !

  • @prathameshdinkar2966
    @prathameshdinkar2966 4 місяці тому

    Nicely explained! I got better understanding of this, could you also include some examples which give some feel about the calculations...

  • @arezooalipanah
    @arezooalipanah 6 місяців тому

    Thank you so much. This video solved so many things for me.

  • @flakky626
    @flakky626 8 місяців тому

    Thank god, I clicked the videoooo
    Thanks man people out there really like to make easy things difficult ty og

  • @arastooajorian9069
    @arastooajorian9069 Рік тому +1

    The mean values are not well selected. Most of the samples are distributed around 200k. So the means have to be around 200k

  • @goelnikhils
    @goelnikhils 10 місяців тому

    Very good explanation of MLE. Amazing

  • @darshh.poetry2193
    @darshh.poetry2193 3 місяці тому

    Nice explanation

  • @harry8175ritchie
    @harry8175ritchie 2 роки тому +1

    Awesome stuff! Just to clarify: logistic regression uses the binomial distribution; let's not confuse viewers with link functions and sigmoids.

    • @alang.2054
      @alang.2054 Рік тому

      Aren't sigmoids a whole family of functions that have certain properties?

  • @FabioDBB
    @FabioDBB 5 місяців тому

    Great explanation sir! Thx a lot!

  • @nervechews6781
    @nervechews6781 10 місяців тому

    Great job man. Thanks so much!

  • @ranidu_lakshan
    @ranidu_lakshan 7 місяців тому

    Thank you so much that was really helpful

  • @simonlee1762
    @simonlee1762 Рік тому

    wonderful, thanks for your clear explaining, pretty good

  • @pauledam2174
    @pauledam2174 7 місяців тому

    Thanks! Great explanation at the beginning (up to about minute 8 which is how far I have gotten). Aren't your example choices of myu and sigma off by a factor of more than 1000 though? Just want to make sure I am clear about it.

  • @larshoernle5400
    @larshoernle5400 6 місяців тому

    awesome video. thank you!

  • @juandesalgado
    @juandesalgado Рік тому

    Nice introduction! Very clear and helpful, thanks. My only nitpick would be that, when you change to logarithms, maybe "L proportional to P" (i.e. "L = kP") should become "log L = log k + log P" - not a proportionality anymore, but a constant offset. The idea of monotonicity is still maintained.

    • @CodeEmporium
      @CodeEmporium  Рік тому +1

      Yep. Good catch. I think that's technically correct. Guess when making this type of video when teaching on the spot, sometimes details like this slip my mind

  • @anujlahoty8022
    @anujlahoty8022 Рік тому

    Simply amazing!

  • @stay-amazed1295
    @stay-amazed1295 2 роки тому

    Nice video! New topic...👍 Pl make video ML binary classification of time series forecasting using likelyhood equation. waiting for next video!

  • @naveenbabu24391
    @naveenbabu24391 8 місяців тому

    Hi sir can you do a video on why we use Basie n inferences and how to use them?

  • @abdulrahmankerim2377
    @abdulrahmankerim2377 4 місяці тому

    Excellent! Thanks :)

  • @KeringKirwa
    @KeringKirwa 6 місяців тому

    You were talking about sigma and mean , and everything was clear until when you started talking about theta , where did the sigma and mean go ? are we training the model to make predictions on the model parameters or the distribution parameters ?? Thanks tho.

  • @arpit743
    @arpit743 2 роки тому +1

    It should be probability Density on y axis. Not not probability since X a continious Random Variable

    • @CodeEmporium
      @CodeEmporium  2 роки тому +1

      Yep! Going to make some videos around probability theory soon to clear this up. Good catch!

    • @arpit743
      @arpit743 2 роки тому +1

      @@CodeEmporium yes please more probability theory videos is what we need

  • @rajanmoodley4822
    @rajanmoodley4822 Рік тому

    Super explanation. Thanks

  • @mellowandblue
    @mellowandblue Рік тому

    you are honestly #1

  • @RAHUDAS
    @RAHUDAS Рік тому

    Thanks for the video

  • @fatriantobong
    @fatriantobong 7 місяців тому

    iid? i tot it was independent but non identical distribution, the fact that our data may come from different parameter values

  • @vrajupadhyay4755
    @vrajupadhyay4755 2 місяці тому

    obeservations y1 , y2 ..... yn are joint probability ? i didn't get that part .

  • @davidthehudson
    @davidthehudson 11 місяців тому

    With X values in the six figures, how can mu be a double digit number?

  • @Burakyesilvlogs
    @Burakyesilvlogs 11 місяців тому

    Great video dude

  • @jiaqint961
    @jiaqint961 Рік тому

    Thanks a lot!
    I think you should include keyword: Maximum Likelihood, Log Likelihood Ratio, to your title to reach more audience.

    • @CodeEmporium
      @CodeEmporium  Рік тому

      Yea. I’ll keep this in mind. Thanks for the tip. Maybe I’ll change this title soon

  • @nickstaresinic9933
    @nickstaresinic9933 2 роки тому

    Very nice review. Thanks.

  • @nollix
    @nollix Рік тому

    Good stuff.

  • @Red_Coder
    @Red_Coder Рік тому

    Is it possible to find probability distribution? Looks like in real world we see only likelihood, couse can't obtain general observation (population), does it?

  • @harshitdtu7479
    @harshitdtu7479 5 місяців тому

    7.52

  • @user-wr4yl7tx3w
    @user-wr4yl7tx3w 2 роки тому

    Another great video

  • @thachnnguyen
    @thachnnguyen 7 місяців тому

    Should have clarified that housing prices in practice are not independent. Perhaps use a better example.

  • @CanDoSo_org
    @CanDoSo_org 2 роки тому +1

    Great explanation! Thanks, man. By the way, what Blackboard App are you using in this video?

    • @CodeEmporium
      @CodeEmporium  2 роки тому +1

      Thank you! The app is called “explain everything “

  • @josefernandezvargas9189
    @josefernandezvargas9189 Рік тому

    Bro I got 4 ads watching this video. I hope this guy is making bank off of these videos

  • @jijie133
    @jijie133 2 роки тому

    Great video.

  • @sukursukur3617
    @sukursukur3617 2 роки тому

    Why do we use pdf with well fitted parameter instead of histogram?

  • @alejorabirog1679
    @alejorabirog1679 2 роки тому +1

    Nice

  • @kiranthota5137
    @kiranthota5137 Рік тому

    I don't see any math explanation in this other than showing the equations. but good explanation theoretically. sorry to comment this but i would appreciate if i see actual math and its explinations. thanks

    • @CodeEmporium
      @CodeEmporium  Рік тому +1

      Thanks for commenting! This was my first time teaching in this way with a white boarding strategy. I have tried more for future videos (hopefully they have turned out better)

  • @undisclosedmusic4969
    @undisclosedmusic4969 2 роки тому +13

    Writing red and green on a black background is very hard to read for colourblind people

    • @CodeEmporium
      @CodeEmporium  Рік тому +2

      Yea. I didn’t think it would look this dark. In future videos , I try to correct this. :)

  • @nesconcen3427
    @nesconcen3427 10 місяців тому

    probabaly a stupid question but, P(y1,y2,y3...) is written as P(y1).P(y2).P(y3)...; P(y1,y2,y3...) isn't this a function, but taking the product P(y1).P(y2).P(y3)... gives me a number? and these two are the same thing?

    • @CodeEmporium
      @CodeEmporium  10 місяців тому

      P(y1,y2,y3) is the probability the first random variable (RV) has a value y1 AND the 2nd RV has a value y2 and the 3rd RV is y3. This is a number.
      Now, if each of these RVs are independent of each other, then yea you can write it out as a product of P(y1)P(y2)P(y3). This too is a product of 3 numbers which gives us a number. If they aren’t independent RVs, you are going to have to use the Bayes Rule to write it out in a compex equation. Ultimately, the outcome though is still some real number

  • @ajaybhadauriya4645
    @ajaybhadauriya4645 2 роки тому +1

    Waiting

  • @shahed2356
    @shahed2356 Рік тому

    Smarter version of Aziz Ansari!

  • @deveshnandan323
    @deveshnandan323 Рік тому

    Next Level Explanation , Subscriber+=1 :)

  • @bluejays440
    @bluejays440 10 місяців тому +1

    At no point in this video did you ever state what likelihood actually is, only what it is proportional to. I recognize you're trying to educate but this is a very poor job, similar to the article you wrote on this subject.

  • @genbid007
    @genbid007 Рік тому

    Fake accent nothing else

  • @akarshanmishra2351
    @akarshanmishra2351 Рік тому

    God damn you explain so much better than my college prof.

    • @CodeEmporium
      @CodeEmporium  Рік тому

      Thanks a ton ! Hope you enjoy the rest of these videos :)

  • @nashwahammoud4076
    @nashwahammoud4076 7 місяців тому

    I got the benefit and enjoyment thank you

  • @ZQutui
    @ZQutui Рік тому

    Thanks. It is such a nice explanation of the topic. Everything is explained well

    • @CodeEmporium
      @CodeEmporium  Рік тому

      Thanks so much for the compliment! And I am glad you liked it :)

  • @donaldduck3888
    @donaldduck3888 Рік тому

    Thx, life saver