What are Maximum Likelihood (ML) and Maximum a posteriori (MAP)? ("Best explanation on YouTube")

Поділитися
Вставка
  • Опубліковано 18 гру 2020
  • Explains Maximum Likelihood (ML) and Maximum a posteriori (MAP) estimation/detection using a Gaussian measurement/sampling example.
    * If you would like to support me to make these videos, you can join the Channel Membership, by hitting the "Join" button below the video, and making a contribution to support the cost of a coffee a month. It would be very much appreciated.
    Check out my search for signals in everyday life, by following my social media feeds:
    Facebook: profile.php?...
    Instagram: / iainexplains
    Related videos: (see: www.iaincollings.com)
    • What is a Probability Density Function (pdf)? • What is a Probability ...
    • What is a Multivariate Probability Density Function (PDF)? • What is a Multivariate...
    • What is a Random Variable? • What is a Random Varia...
    • What is a Gaussian Distribution? • What is a Gaussian Dis...
    • What is Gaussian Noise? • What is Gaussian Noise?
    For a full list of Videos and accompanying Summary Sheets, see the associated website: www.iaincollings.com
    .

КОМЕНТАРІ • 149

  • @ayushkumarrai1117
    @ayushkumarrai1117 3 роки тому +37

    You're the professor I wished I had in my college! Thankyou!!

  • @Peolorios
    @Peolorios 3 роки тому +11

    I found your videos at the right moment, they cover a lot of the basics of my 1st semester master courses. Thank you. A nice topic you could cover that comes up a lot in detection and estimation is the Cramer-Lao Lower Bound

    • @iain_explains
      @iain_explains  3 роки тому +2

      Thanks for your comment. Glad the videos are helpful. And thanks for the C-R suggestion, I'll add it to my "to do" list.

  • @junchiliu7350
    @junchiliu7350 2 роки тому +3

    The best explanation on ML and MAP! I finally understood them. Thank you!

    • @iain_explains
      @iain_explains  2 роки тому

      I'm so glad the video helped, and that you liked the explanation.

  • @maclee3841
    @maclee3841 2 роки тому +1

    Thanks a lot! One of the most simplest explanations on UA-cam

  • @aidankennedy6973
    @aidankennedy6973 3 роки тому +7

    This is a fantastic video that answered so many questions I had while working through my academic coursework. Thank you so much for uploading!

  • @SouthEsVideos
    @SouthEsVideos 3 роки тому +3

    I'm currently learning about autoencoders and it's based on this topic! very helpful and intuitive. Thank you!

  • @aqeelal-shakhouri7572
    @aqeelal-shakhouri7572 2 роки тому

    Thank you. you explained it clearly, just what I was looking for.

  • @ashokmahipathi8133
    @ashokmahipathi8133 Рік тому

    Thank you so much Sir Iain. You made my day. Great explanation regrading MAP and ML. Hats off Iain

  • @arvindp551
    @arvindp551 2 роки тому +1

    What you just fabulously explained in 15 lines, takes 4+ blackboards to many Indian teachers to explain even less than that. Thank you so much for sharing your knowledge here.

  • @burakkurt3027
    @burakkurt3027 2 роки тому +2

    You explain the concept not only very concise way but also in saving paper. I appreciate you for both the topic and the saved paper.

    • @iain_explains
      @iain_explains  2 роки тому +3

      Thanks. I hadn't realised how environmentally responsible I was being. 😀 I think it really helps to fit everything onto a single sheet of paper so that the whole explanation is visible all at once, so the viewer can easily refer back at any point.

    • @burakkurt3027
      @burakkurt3027 2 роки тому

      @@iain_explains This is very logical :D

  • @shahabansari5201
    @shahabansari5201 3 роки тому +1

    Wow! Amazing way of explaining these complex ideas.

  • @rahulrustagi7029
    @rahulrustagi7029 6 місяців тому +1

    Amazing. Make a series of Probabilistic ML Models!

  • @NeilDeshpande313
    @NeilDeshpande313 2 роки тому +2

    I finally get the difference between the two! Thank you!

  • @zhipenglin
    @zhipenglin День тому +1

    excellent video!!

  • @prietessutube
    @prietessutube Рік тому

    Thanks so much for this video, explained it much better with my lecturer!!!

  • @ashleyhart1907
    @ashleyhart1907 9 місяців тому

    Incredibly helpful. Thank you!

  • @AbCd-fo6ys
    @AbCd-fo6ys 2 роки тому

    What a clear explanation!
    Thank you so much.

  • @keremokyay6725
    @keremokyay6725 Рік тому

    Amazing, this was so clear to understand. Thank you very much!!!

  • @ivanrodionov9724
    @ivanrodionov9724 Рік тому +1

    Outstanding video! You sir have saved the day, again!

    • @iain_explains
      @iain_explains  Рік тому +1

      I'm so glad the video helped. It's great to read these comments, and know that my videos are making a difference for people. Thanks.

  • @jizong234
    @jizong234 Рік тому

    This is the best explanation in the world, thank you !

    • @iain_explains
      @iain_explains  Рік тому +1

      I'm so glad to hear that you liked the video.

  • @sayantankundu973
    @sayantankundu973 Рік тому

    Definitely "Best explanation on UA-cam" !! ❤ Thanks a lot Sir.

    • @iain_explains
      @iain_explains  Рік тому

      Thanks. I'm glad you think so. And I'm glad it was helpful.

  • @partheshvikassoni8601
    @partheshvikassoni8601 2 роки тому

    This was really informative! Thanks.

  • @gastonparra2971
    @gastonparra2971 Рік тому

    Gran explicación.. Gracias por subir el video.

  • @joaquinvanloon7341
    @joaquinvanloon7341 2 роки тому

    Amazing video, thanks!

  • @peterkimemiah9669
    @peterkimemiah9669 2 роки тому

    Thanks for your super simple explanation. I now understand how to apply it.

  • @ibrahimshikdaher7551
    @ibrahimshikdaher7551 3 роки тому +1

    Thank you

  • @tainacoimbra1940
    @tainacoimbra1940 2 роки тому

    Excelent explanation! Thank you very much :)

  • @nikkoNine
    @nikkoNine Рік тому

    Very intuitive explanation! 🙏

  • @saraho5338
    @saraho5338 Рік тому

    Thank you for the great video

  • @AK-yf4dp
    @AK-yf4dp 2 роки тому

    Thank you so much!!! very helpful video

  • @heroszeros
    @heroszeros Рік тому

    Very good explanation with right amount of details and relevant examples. Thanks a million.

  • @waynerooney4674
    @waynerooney4674 Рік тому +1

    ty

  • @abderrahmanetayebcherif7325
    @abderrahmanetayebcherif7325 2 роки тому

    best explanation of the ML and MAP on youtube
    thank you

  • @JoaoVitorBRgomes
    @JoaoVitorBRgomes 3 роки тому +1

    Beautiful explanation. Very helpful.

  • @Dutta1605
    @Dutta1605 7 місяців тому

    i beleive the title of the video is genuinely true.

    • @iain_explains
      @iain_explains  7 місяців тому

      Thanks. It was a comment someone else had made about the video, so it's good to know that you also agree.

  • @adonis7419
    @adonis7419 3 роки тому +1

    Very helpful

  • @gofaonef.mogobe1306
    @gofaonef.mogobe1306 3 роки тому +1

    Hi..very helpful video. Kindly assist me understand how I can factor in the concept of consistency of MLE with respect to the graph illustrations?
    Particularly, I've learned that as n gets large, mean turns to zero as MLE becomes an even more consistent estimate.

  • @alexanderpaskal1387
    @alexanderpaskal1387 2 роки тому

    Liked and subbed, very clear and accessible explanation of a concept that made no sense to me as it was presented in my class

  • @Balance-fl1zc
    @Balance-fl1zc 7 місяців тому

    Beautiful explanation sir, thank you!

  • @mehmetfatihayten8269
    @mehmetfatihayten8269 Рік тому

    Decent video! Thanks.

  • @Yohan87845
    @Yohan87845 2 роки тому

    Thank you very much.

  • @NoahElRhandour
    @NoahElRhandour 3 роки тому

    holy what a clear explanation. it ended my 2 day struggle of not getting it in 18 minutes!!!! thank you

  • @user-ns9ze8xf5z
    @user-ns9ze8xf5z 3 роки тому

    Really big thanks for your video!!
    May you take another video for explaining different pathloss models, such as Okumura-Hata or various COST model in wireless channel?

    • @iain_explains
      @iain_explains  3 роки тому +2

      Good suggestions thanks. I've added them to my "to do" list.

  • @MohammedFarag81
    @MohammedFarag81 9 місяців тому

    Great, thank you

  • @usmanshabbir4902
    @usmanshabbir4902 2 роки тому

    very precisely explained.

  • @passionatechristianmaker
    @passionatechristianmaker 4 місяці тому

    great explanation!

  • @Amine1z
    @Amine1z 3 роки тому +1

    Dear prof, you're the best

  • @skyful9
    @skyful9 Рік тому

    Can this be applied in marketing?

  • @AleeEnt863
    @AleeEnt863 Рік тому

    Great explanation...

  • @yuxinzhang4228
    @yuxinzhang4228 2 роки тому

    It's very helpful thanks sooooo much

  • @robsenponte3308
    @robsenponte3308 11 місяців тому +1

    Nice

  • @ronaldmtenga8093
    @ronaldmtenga8093 Рік тому

    are the differenct x values you are checking for maximum likelihood each a possible input signal or are we searching on a bit by bit basis?

    • @iain_explains
      @iain_explains  Рік тому

      They are a possible realisation of the random variable X. If X represents binary data, then it would be "searching on a bit by bit basis", but it X represented higher order modulation then it would be on a "symbol" basis.

  • @AlexColeAldems
    @AlexColeAldems Рік тому

    Does demodulating using ML require channel state information? (i.e. an estimation of the AWGN noise variance)

    • @iain_explains
      @iain_explains  Рік тому +1

      Yes. Good point. It's almost never mentioned. It's not too hard to get an estimate of the receiver noise - by taking measurements when nothing is being sent (of course you need to be able to work out when nothing is being sent!) It's harder to estimate other parameters, such as channel gain. And there's lots of things that are done to make that possible. See eg: "Channel Estimation for Mobile Communications" ua-cam.com/video/ZsLh01nlRzY/v-deo.html

  • @ks.4494
    @ks.4494 Рік тому

    Thanks for the Video, is there any reference ( book, ...) for that, particulary for numerical solution?

    • @iain_explains
      @iain_explains  Рік тому

      I'm not sure if this is what you're looking for exactly (eg. I'm not sure it has the numerical examples you might be looking for), but I like this book: H. Vincent Poor, “An Introduction to Signal Detection and Estimation”

  • @davebordenkircher8131
    @davebordenkircher8131 Рік тому

    I see now that the MAP estimator is like a weighted version of the ML estimator, where the weights come from prior knowledge of the measurement target. The different conditional distributions fy(y|xi) are “pushed up” or “pushed down” based on the value of the corresponding fx(xi). Of course, provided that all fx(xi) are equiprobable, the MAP estimator reduces to the ML estimator which we commonly see in optimal communications system analysis.
    I have a question for you, why is it that the equiprobable symbol scheme is considered most optimal? I am inclined to assume that it is because it yields the highest entropy. Also, I would like to know how it is that we ensure equiprobable signaling?
    Thank you

    • @iain_explains
      @iain_explains  Рік тому +1

      Excellent question. Yes, when a random source is compressed to its minimal representation (using an entropy achieving codebook) it results in a binary sequence that has equally likely ones and zeros. This video provides more insights: "What is Entropy? and its relation to Compression" ua-cam.com/video/FlaJPxP8sd8/v-deo.html

  • @loreychu7646
    @loreychu7646 3 роки тому

    Good explanation of a lot of concepts in wireless communication. I'm watching your video for the preparation of QE. Hope I can pass!

  • @javiergonzalezarmas8250
    @javiergonzalezarmas8250 Рік тому

    Amazing

  • @ravindratomar9916
    @ravindratomar9916 2 роки тому

    sir, how to estimate channel in case of correlated rayleigh fading channel. for example y1=hx_1 + hx_2 +n_1, y2=hx_1 + hx_2 +n_2.

    • @ravindratomar9916
      @ravindratomar9916 2 роки тому

      n_1 and n_2 are white gaussian noise with different variances.

  • @_Sam_-zh7sw
    @_Sam_-zh7sw 2 роки тому

    may be i am missing some pre-requisite knowledge because i am confused a little bit. have we inverted the graph of the function here? f(x) is plotted horizontally and x is plotted vertically. But how can there be a different distribution function of ax1,ax2...ax(n) if there is just one input and output?

    • @iain_explains
      @iain_explains  2 роки тому

      Um no. x is not plotted vertically. f_Y(y|x) is plotted vertically. This is the density of the random variable Y, given a specific value of the random variable X. This is a different function for each different realisation (value) of X (ie. x_1, x_2, ...).

  • @analysislearning9179
    @analysislearning9179 3 роки тому

    Great video, I have a question, if the variable is its self distributed with Nakagami distribution. Then how can we compute the MLE and MAP?

    • @iain_explains
      @iain_explains  3 роки тому +1

      The term f_X(x) is the density function for the random variable of interest. So, if it is Nakagami distributed, then f_X(x) equals the formula for the Nakagami p.d.f. which you can find in this video: ua-cam.com/video/ztpNbE-Vpaw/v-deo.html

    • @analysislearning9179
      @analysislearning9179 3 роки тому +1

      @@iain_explains Thankyou very much

  • @cartercheng5846
    @cartercheng5846 Рік тому

    Very helpful video. I have a question there. The MAP is explained as MLE weighed by the probability of the parameter x, and the parameter follows a certain distribution. If X is a continuous random variable, what is the mathematical meaning for f_y(y|x)f_x(x)?

  • @KCODivyanshuKatyan
    @KCODivyanshuKatyan Рік тому +1

    best in the game 🙌🙌

  • @kavineshs15
    @kavineshs15 12 днів тому

    I can't understand why the bell curve is shifting for every value of x.

  • @saipavan7703
    @saipavan7703 2 роки тому

    Hi Iain. Loved your explanation. I wanted to ask a question about MLE. In the plots of x1,x2,xn, When each x1/x2 give a single value for the function, Why does plot exist for x1 when the function takes a single value for x1. Thank You

    • @iain_explains
      @iain_explains  2 роки тому

      Sorry, I'm not sure what you're asking.

    • @rishabhsrivastava1241
      @rishabhsrivastava1241 2 роки тому

      I think I got what you are saying and there seems to be some gap in your understanding. Let me try to fill that although Ian mentioned it in this video.
      What you are saying is that for a given value of x, there is only one single value of y through it's distribution f(y/x) but that is not true. Actually, there are SEVERAL different distributions of y depending on the SEVERAL values of x's. So, when Ian says that for a given x, the distribution's center shifts, it is actually a new distribution centered around that given x value. Then comes the concept of a single value from these distributions, now that is y(bar), this is an observation of all the f(y/x) pdf value among all the distributions of y's for those SEVERAL x's. That is the single value that you are thinking of.
      Hope I was able to answer your question to some degree. :)

    • @mrr7115
      @mrr7115 2 роки тому

      The single value is a result of having measured/fixed y (the y bar hat equation). The plot is for all y (ie it’s a function of y not of x1). x1 is just a guess of the true parameter of the Gaussian distribution (proportional to mean). The horizontal axis (independent variable) is y. Also the function, which is a Gaussian, takes more than just a*x1, it also takes in the variance from the noise. To prove to yourself that the function takes y, look at the form of the Gaussian equation, see the y in there?

  • @kalyansunkara9464
    @kalyansunkara9464 2 роки тому

    Sir which text book should we follow for detection and estimation theory?

    • @iain_explains
      @iain_explains  2 роки тому +1

      I like the book: H. Vincent Poor, "An Introduction to Signal Detection and Estimation", Springer.

  • @mangomilkshakelol
    @mangomilkshakelol 6 місяців тому

    Thank you so much! That's clear. One question: for MAP, what's f_X(x)?

    • @iain_explains
      @iain_explains  5 місяців тому +1

      It's the probability density function for the variable X.

  • @NvkChaitanya777
    @NvkChaitanya777 2 роки тому

    Can you please make a video on softmax regression?

    • @iain_explains
      @iain_explains  2 роки тому

      Thanks for the suggestion, but I'm not familiar with it, sorry. I'll have to give it some thought.

  • @gamingandmusic9217
    @gamingandmusic9217 3 роки тому

    at a particular point in the density function, the probability is zero right? I'm a little confused.

    • @iain_explains
      @iain_explains  3 роки тому

      I'm not exactly sure what you're asking. The density function is a "density" (as the name indicates). This means you need to integrate it over some range of values, in order to find the probability. The probability of any _exact_ value is zero (since the base has zero width, for a single _exact_ value). See: "What is a Probability Density Function (pdf)?" ua-cam.com/video/jUFbY5u-DMs/v-deo.html

    • @gamingandmusic9217
      @gamingandmusic9217 3 роки тому

      @@iain_explains oh sorry, im wrong. Thank you so much sir.

  • @tuongnguyen9391
    @tuongnguyen9391 3 роки тому

    Do you ever have to do a rehearsal beforehand ? I see the explanation is quite smooth.

    • @iain_explains
      @iain_explains  3 роки тому +1

      Thanks, I put quite a bit of thought into how to explain things in the best way.

    • @tuongnguyen9391
      @tuongnguyen9391 3 роки тому

      Do you follow a systematic procedure to construct the explanation process. If so I really hope that you could share this procedure :). Although everything is short I find that the information is delivered clearly with many subtle points and detail carefully summarized. Thank you for your inspiring lecture.

  • @amitpalkar934
    @amitpalkar934 3 роки тому

    Can we have a video sometime on mmse and irc receivers ?
    Regards,
    Amit

    • @iain_explains
      @iain_explains  3 роки тому

      Thanks for the suggestion. I've added them to my "to do" list. I'll see what I can do (it's starting to become a long list).

  • @rezasoleimani6636
    @rezasoleimani6636 Рік тому

    The explanation is great. The only problem is using pen and paper instead of something more comfortable. The page is too small for this amount of writing.

    • @iain_explains
      @iain_explains  Рік тому +1

      On the other hand, having everything on the one page means you don't need to scroll back and forth through the video to see the links to earlier parts, and I can simply point to the earlier parts while explaining how they link to the later parts (as I'm doing in the thumbnail image). Perhaps it doesn't work so well on small screens ...

  • @robertpollock8617
    @robertpollock8617 8 місяців тому

    So it is L(x|y)? - we want to maximize the likelihood of x given the data values y? . So we are in a sense trying to say that we have high likelihood that this data observed could come from or be predicted by this model of x? Where the probability is P(y|x). Maybe you are saying that and I am not picking up on this. I think you might be but I might not be understanding your notation.

  • @donpkchannel7203
    @donpkchannel7203 2 роки тому

    When your typing the screens becomes blurry because paper is moving. Please stabilize the paper.

  •  2 роки тому +1

    MAP starts at 10:35

  • @human_agi
    @human_agi Рік тому

    if x is vector ?

  • @egorkrapovnitskiy997
    @egorkrapovnitskiy997 Рік тому

    didn't get the idea

  • @skatingscotland64
    @skatingscotland64 2 роки тому

    Iain you have offered me shelter in a howling wind, thank you - I can leave the library and go home now xo love from rory

    • @iain_explains
      @iain_explains  2 роки тому

      That's great. I'm so glad you found the video helpful. Hope you mange to stay out of the wind.

  • @tomaguci5102
    @tomaguci5102 Рік тому

    I think this video could be improved by providing a concrete example, also it's really mathy without much intuitive explanation