Introduction to Bayesian statistics, part 2: MCMC and the Metropolis-Hastings algorithm

Поділитися
Вставка
  • Опубліковано 21 лип 2024
  • An introduction to Markov chain Monte Carlo (MCMC) and the Metropolis-Hastings algorithm using Stata 14. We introduce the concepts and demonstrate the basic calculations using a coin toss experiment.
    www.stata.com
    Copyright 2011-2019 StataCorp LLC. All rights reserved.

КОМЕНТАРІ • 123

  • @AlessandroBottoni
    @AlessandroBottoni 6 років тому +149

    This is by far the best explanation of Metropolis-Hastings algorithm I was able to find on the web. Thanks a lot.

  • @vivekpetrolhead
    @vivekpetrolhead 6 місяців тому +3

    The rotated density plot clarified so many things. Thanks!

  • @dataminingincae
    @dataminingincae 8 років тому +65

    Brilliantly simple explanation of the Metropolis-Hasting Algorithm. Thanks.

  • @abdulelahaljeffery6234
    @abdulelahaljeffery6234 7 років тому +42

    best MCMC explanation on UA-cam.

  • @practicaldataengineering3955
    @practicaldataengineering3955 2 роки тому +3

    The best explanation of Metropolis-Hastings algorithm. Thank you for this, no university course nor research paper will provide this much of understandable clarification.

  • @14loosecannon
    @14loosecannon 2 роки тому +2

    As others have already said, this is hands down the best explanation of MCMC I've seen on UA-cam!

  • @flagrance2808
    @flagrance2808 5 років тому +2

    Simply the best illustration I have ever seen on the web - THANK YOU SO MUCH!

  • @area51xi
    @area51xi 6 років тому +2

    One of the best MCMC explanations on UA-cam and I'm just learning for fun with no formal math/stats background.

  • @keggluneq
    @keggluneq 4 роки тому +2

    Thank you. I've been looking for an intuitive explanation of MCMC for years!

  • @nathsujitkumar
    @nathsujitkumar 5 місяців тому

    The best explanation, available online, on this topic by far. The visualization of the algorithm is superb! Very much appreciated.

  • @Echodonut
    @Echodonut 6 років тому

    I have been trying to understand bayesian MCMC for the last couple of days, and this helped my understanding along greatly. Thanks!

  • @rajarajansb5690
    @rajarajansb5690 Рік тому

    The perspective of visualization & presenter's clarity made this the best explanation

  • @jimmylee6547
    @jimmylee6547 4 роки тому +19

    My son and I watched this entire video together. You did a great job teaching. My son understood this, I made sure to ask him questions to ensure he surly did. He's nearly 13 now and loves data-science.
    Great lecture and you have an excellent voice for teaching. Thanks again for this video.

  • @hfkssadfrew
    @hfkssadfrew 6 років тому

    Very clean explanation! Much better than other videos online

  • @naveenkumarparameswaran3862
    @naveenkumarparameswaran3862 5 років тому +2

    The best video on MCMC! Hands down.

  • @hongjingxia199
    @hongjingxia199 2 роки тому

    By far the most clear explanation I've seen. Thanks a lot!!!!!

  • @aflouh
    @aflouh 5 років тому +5

    The best explanation of MCMC ,,, thank you x 1000

  • @com0oan
    @com0oan 6 років тому

    Excelent explanation! Best I've found so far!

  • @JohnSegrave
    @JohnSegrave 8 років тому +6

    Really nicely done - a very clear and concise explanation. I got a better sense of MCMC and M-H from these 8 minutes than several hours of reading on the topic beforehand. Thank you.

  • @StephenRoseDuo
    @StephenRoseDuo 7 років тому +35

    easily the best mcmc explanation

  • @allwanamar1
    @allwanamar1 6 років тому

    wow. man ! i have no words ...just perfection.I appreciate the speed.

  • @yifansong547
    @yifansong547 4 роки тому

    Brilliantly clear explanation! Thanks a lot! Really Great Job!

  • @uniquenessexistence
    @uniquenessexistence 11 місяців тому +1

    Very clear explanation with plots. This is very helpful.

  • @anakwesleyan
    @anakwesleyan 6 років тому

    very practical and insightful. best explanation of MH / MCMC

  • @PedroRibeiro-zs5go
    @PedroRibeiro-zs5go 4 роки тому

    I agree with the previous comments! Best explanation on UA-cam by far

  • @fahdyazin82
    @fahdyazin82 4 роки тому

    Most intuitive explanation of the MCMC-MH

  • @lihuil3115
    @lihuil3115 2 роки тому +1

    very good. The best explanation so far.

  • @tripathi26
    @tripathi26 3 роки тому

    Crisp n Clear!
    Thanks for sharing your knowledge.

  • @nutellahasswag3394
    @nutellahasswag3394 4 роки тому

    Top comment says it all, really clicked for me when you broke down into three stages. Thank you

  • @DreamWorker-jm5xn
    @DreamWorker-jm5xn 5 років тому

    Simply the best MH video.

  • @holloloh
    @holloloh 5 років тому +1

    I think I finally understood what this algorithm actually does.
    Thank you.

  • @omidkeivani6028
    @omidkeivani6028 3 роки тому

    Great explanation. Awesome job.

  • @nontastsokanos1695
    @nontastsokanos1695 5 років тому

    Excellent explanation. Thank you!

  • @PedroRibeiro-zs5go
    @PedroRibeiro-zs5go 6 років тому

    That was the heck of a good video!! Thank you!!

  • @kalimismilequest
    @kalimismilequest Місяць тому

    Thank you so much. This was life changing

  • @eduardosr9859
    @eduardosr9859 6 років тому

    Thank you! Excellent video!

  • @hsl2916
    @hsl2916 4 роки тому

    Thank you so much. I finally understood MCMC!!!!

  • @porelort09
    @porelort09 8 місяців тому +1

    Great video!

  • @SalElder
    @SalElder 3 місяці тому

    Great overview, thanks.

  • @ohmyfly3501
    @ohmyfly3501 6 років тому +1

    If you want to know properly then watch at .75x speed
    Very nice explanation

  • @marseraser
    @marseraser Рік тому

    I love that visaualization !

  • @Veto2090
    @Veto2090 8 місяців тому

    Thank you so much for this. All the other explainations I found had me very confused

  • @rahulrustagi7029
    @rahulrustagi7029 4 місяці тому

    Hands Down the best

  • @florangelicapereda3130
    @florangelicapereda3130 7 років тому

    Very good tutorial useful for Stata's users

  • @edieespejo105
    @edieespejo105 6 років тому

    Really helpful! Thank you. :)

  • @320achacin
    @320achacin Рік тому +1

    Excellent.!!!

  • @RJone89
    @RJone89 2 роки тому

    This is a godsend. Thank you.

  • @AIVidya
    @AIVidya 3 роки тому

    Great Explanation

  • @s.z.4382
    @s.z.4382 7 років тому

    Very helpful. Thanks.

  • @karannchew2534
    @karannchew2534 2 роки тому +1

    Note for revision.
    Why "Markov Chain"?
    Draw a theta (which will be accepted or rejected based on Hastings algo), then drawn another theta based on the new theta. The new theta is dependent on the old theta, so will be each successive theta, hence Markov Chain.

  • @sabrihamad
    @sabrihamad 3 роки тому

    Thank you for this great tutorial! I have a question though: What happens if the mcmc sampler gives theta>1. You won't be able to calculate the likelihood or the prior. Do you discard this theta in this case? Shouldn't one use a uniform distribution u(0,1) to make sure we are getting a theta in [0, 1]?

    • @gordongoodwin6279
      @gordongoodwin6279 3 роки тому

      You’ll need to read up on the distinction between Metropolis and Metropolis Hastings. The MH was created for this very reason, to address proposal distributions for constrained parameters, like sigma. The regular Metropolis algorithm only works for non-constrained parameters where this isn’t an issue. Basically, MH adjusts for this by allowing non symmetric proposal distributions in that instance. Beyond the scope of this video, I’d recommend watching Ben Lamberts series on Bayesian Statisitcs. He has a video specific to MH and this question

  • @rizwanniaz9265
    @rizwanniaz9265 6 років тому +1

    how to calculate odd ratio in bayesian ordered logistic plz tell me

  • @Ivan-td7kb
    @Ivan-td7kb 5 років тому

    Wait so if we use MCMC to walk around the parameter space why do we even need a prior distribution? Is it only used to initialize the starting value of the MCMC?

  • @yoloswaggins2161
    @yoloswaggins2161 5 років тому

    concise and precise

  • @Ivkovic1971
    @Ivkovic1971 8 років тому +9

    Videoo that explains mcmc nice and clear. Thank you

  • @karannchew2534
    @karannchew2534 2 роки тому

    03:35 A new theta is drawn from the Proposal Distribution, which has a normal distribution in this example. How to decide/select the proposal distribution function?

  • @elsidiegbelhaj2016
    @elsidiegbelhaj2016 6 років тому

    Very good. Thanks

  • @weizhang7428
    @weizhang7428 5 років тому

    Very clear expression

  • @excel_wang
    @excel_wang 7 років тому +1

    Thanks! This explains MCMC perfectly. However it does not seem to explain why Markov chain is required? Why not just use Monte Carlo alone for drawing from the distribution?

    • @rockstarchileno2
      @rockstarchileno2 7 років тому

      The Random Walk is a Markov Chain, problem they using to model the problem not to solve the problem

  • @tpof314
    @tpof314 7 років тому

    Best MCMC explanation on UA-cam. Thank you !

  • @ege5785
    @ege5785 3 роки тому

    Excellent!

  • @alexbode6894
    @alexbode6894 6 років тому +3

    Can someone please explain to me how step one is calculated (posterior theta new / posterior theta old) because if you are sampling from a binomial distribution the output would be 0 or 1. and sampling from beta (1,1) is just sampling from a uniform distribution. for example what does beta(1,1,0.088) and binomial(10,4,0.88) each equal individially? Thanks in advance

    • @RYBAAC28
      @RYBAAC28 6 років тому +3

      With the first example in the video:
      R code:
      prior1=dbeta(shape1=1,shape2=1,x=0.517)
      likelihood1=dbinom(x=4,size=10,prob=0.517)
      prior2=dbeta(shape1=1,shape2=1,x=0.380)
      likelihood2=dbinom(x=4,size=10,prob=0.380)
      r=(prior2*likelihood2)/(prior1*likelihood1)
      r
      1.31
      Each part equals:
      prior1=1
      prior2=1
      likelihood1=0.19
      likelihood2=0.25

  • @zhen3356
    @zhen3356 2 роки тому

    Amazing

  • @malharjajoo7393
    @malharjajoo7393 4 роки тому +7

    How is the proposal distribution chosen ? And Why isn't that a limitation of the MH algorithm ?

    • @gordongoodwin6279
      @gordongoodwin6279 3 роки тому

      It depends on the parameter type. If it’s an unconstrained parameter, then a symmetrical is fine, usually the normal curve is used. If it’s a constrained parameter (e.g. sigma), then the Hastings variation of the Metropolis algorithm is used because it allows for non symmetric proposal distributions. The key to the algorithm and how it works to map out the posterior is the acceptance ratio, which can be affected by the size of the steps proposed (sigma in the normal curve proposal distribution). Auto-tuning allows the proposal distribution to be adjusted periodically to make sure the acceptance ratio is within the target threshold.

  • @moslemasgarihassanluie
    @moslemasgarihassanluie 6 місяців тому

    Finally I understand MCMC WOooOw😍😍😍😍😍

  • @reabo22
    @reabo22 5 років тому

    Awesome!

  • @ahmedatta6508
    @ahmedatta6508 3 роки тому

    thanks for great video
    why you using uniform distribution to generate random variable and compare the result with it ?

    • @gordongoodwin6279
      @gordongoodwin6279 3 роки тому

      This is how MH accepts candidate/proposed values where the posterior ratio of proposed/current is less than 1. When it’s less than 1, a uniform distribution is used to draw a random number. If that random number is below the acceptance ratio, we accept the candidate value. It’s just a random number generator to make sure we accept our values at that probability

  • @archanamaurya89
    @archanamaurya89 4 роки тому

    Can you PLEASEEE add subtitles in your video? Somehow for these particular video series, automatic closed caption is not available either :(

  • @PiercingSight
    @PiercingSight 4 роки тому +3

    My question is why?
    Why do we generate a new distribution from an already existing distribution?
    Why is each new generated value based on the previous value? Coin tosses are independent, are they not? Why would we do this to distributions of independent probabilities?

    • @gordongoodwin6279
      @gordongoodwin6279 3 роки тому +3

      You wouldn’t do this for simple situations where it’s possible to sample independently. This is just a simplified example. MCMC algorithms are designed for use in multidimensional models where it isn’t feasible or efficient to calculate p(y), and thus a proper absolute posterior probability of theta. In such situations, when p(y) is intractable, dependent MCMC sampling is needed because it allows for the exploration of posterior parameter space through the use of relative frequencies of posterior samples rather than absolute probabilities

  • @user-ye2ni2hi9v
    @user-ye2ni2hi9v 5 років тому

    Please explain this for me...why do you use Beta(1,1) as prior distribution which is flat, why not use Beta(30, 30) you used in previous video?

  • @nihaarrshah
    @nihaarrshah 6 років тому

    I am not able to link the binomial likelihood and beta prior with the proposal and the target distributions. Do they correspond to each other?

    • @gordongoodwin6279
      @gordongoodwin6279 3 роки тому

      The target distribution is just the posterior, which is proportional to the prior * likelihood. A beta prior is a conjugate prior for the beta binomial posterior (target) distribution. Not all priors are conjugate priors. Regarding the proposal distribution, a normal distribution is often used b/c it is a symmetric distribution, but it’s not specifically mandated to be a normal

  • @pazenriqueguillermo
    @pazenriqueguillermo 2 місяці тому

    I would like to know which programa did tou use to make the trace anda histograma plot simulation. I found It Very useful for classes

  • @zachkim1624
    @zachkim1624 5 років тому

    what's the program tht he's using? looks so much simpler than rjags that i'm using right now

  • @malharjajoo7393
    @malharjajoo7393 4 роки тому +2

    1:59 - it's very important to mention Law of Large numbers (LLN) at this point !!

  • @lxl274
    @lxl274 5 років тому +2

    what is the proposal distribution, which generating theta for each time?

    • @gordongoodwin6279
      @gordongoodwin6279 3 роки тому

      It’s a normal curve in this example, centered around the theta value sampled in the previous iteration

  • @meshackamimo1945
    @meshackamimo1945 8 років тому +4

    thanx for a beautiful video that has made me form an image of what mcmc is all about.
    God bless you.
    do kalman filters demo ,eg, for time dpseries predictions,using stata.

  • @rongarza9488
    @rongarza9488 4 роки тому

    All this is excellent for physical, electrical / electronic systems, and the like. However, I have seen people try to use it for human based systems (like predicting the stock market, yeah, right). For that, it would serve best to know if a person likes Shakespeare, Frost, or limericks.

  • @bhaveshsolanki8765
    @bhaveshsolanki8765 7 років тому

    excellent

  • @Thelarryyy1
    @Thelarryyy1 5 років тому

    beautiful

  • @karannchew2534
    @karannchew2534 2 роки тому

    How would it work if theta consist of multiple variables?

  • @danielnakamura6430
    @danielnakamura6430 3 роки тому

    congratulations

  • @princejohn4681
    @princejohn4681 8 років тому +1

    Simply the best!

  • @pranavkhanna9459
    @pranavkhanna9459 6 років тому

    Thanks

  • @user-ex1fj5rn2o
    @user-ex1fj5rn2o 2 роки тому

    Best ever

  • @subhramaity
    @subhramaity 5 років тому

    thanks

  • @ahmedhassanGent
    @ahmedhassanGent 3 роки тому

    why do we do step 3 at 4:00?

  • @yangchen9983
    @yangchen9983 6 років тому

    The best

  • @SahibYar
    @SahibYar 8 років тому +2

    Well explained

  • @aref_m2024
    @aref_m2024 7 років тому +1

    Thanks for the nice lecture.

  • @mtheory85
    @mtheory85 8 місяців тому

    I just love how unapologetically Windows this is.

  • @nyanity
    @nyanity 3 роки тому

    Under every video I've watched, someone comments that "this is the best video on MH algorithm!!" but I still don't understand it 😭

  • @wp1300
    @wp1300 Рік тому

    1:13 Monte Carlo
    2:04 Markov Chain

  • @franciscogallegos7152
    @franciscogallegos7152 4 роки тому

    There is a mistake at 4:38. I.e., .247 > 0.039, not less.

  • @victorvolkov4169
    @victorvolkov4169 2 роки тому

    It seems, usage of Beta(1,1,0.286), Beta(1,1,0.380), Binomial (10,4,0.286) and Binomial(10,4,0.380) at the time frame 4:16 has no meaning: the first two are constants and equal to each other, and the latter two describe PDF to express outside [0,1] region, as shown in the plot.

  • @jennyapl1791
    @jennyapl1791 4 роки тому

    This is great but the typo = should say "proportional" on your definition of posterior

  • @juggernautuci8253
    @juggernautuci8253 3 роки тому

    how do you make animation

  • @ASHISHDHIMAN1610
    @ASHISHDHIMAN1610 2 роки тому

    3:14 I think we accept theta new when u > acceptance probability

  • @andreneves6064
    @andreneves6064 6 років тому

    English subtitles, please

  • @ajaydhungana1921
    @ajaydhungana1921 2 роки тому

    Sir..so MCMC is not seprarete algorithm but uses Meterpolis hasting algorithm.. i thought MCMC was a seperate procedure that helps us to simulate data to help us identify posterior distribution.
    Also i want to learn more.. how can i do so?

    • @statacorp
      @statacorp  2 роки тому

      Contact us at tech-support@stata.com for assistance.

  • @FinallyAFreeUsername
    @FinallyAFreeUsername 5 років тому

    Is that Barry Greenstein?