Gaussian Naive Bayes, Clearly Explained!!!

Поділитися
Вставка
  • Опубліковано 25 лис 2024

КОМЕНТАРІ • 483

  • @statquest
    @statquest  4 роки тому +21

    NOTE: This StatQuest is sponsored by JADBIO. Just Add Data, and their automatic machine learning algorithms will do all of the work for you. For more details, see: bit.ly/3bxtheb BAM!
    Corrections:
    3:42 I said 10 grams of popcorn, but I should have said 20 grams of popcorn given that they love Troll 2.
    Support StatQuest by buying my book The StatQuest Illustrated Guide to Machine Learning or a Study Guide or Merch!!! statquest.org/statquest-store/

    • @phildegreat
      @phildegreat 3 роки тому

      website not working?

    • @statquest
      @statquest  3 роки тому

      @@phildegreat Thanks! The site is back up.

    • @anirbanpatra3017
      @anirbanpatra3017 Рік тому

      8:15 There's a minor error in the slide 'help use decide' .
      You really are a great teacher.Wish I could Meet you in person some day.

  • @rohan2609
    @rohan2609 3 роки тому +136

    4 weeks back I had no idea what is machine learning, but your videos have really made a difference in my life, they are all so clearly explained and fun to watch, I just got a job and I mentioned some of the learnings I had from your channel, I am grateful for your contribution in my life.

  • @raa__va4814
    @raa__va4814 2 роки тому +26

    Im at the point where my syllabus does not require me to look into all of this but im just having too much fun learning with you. Im glad i took this course up to find your videos

  • @mildlyinteresting1925
    @mildlyinteresting1925 4 роки тому +70

    Following your channel for over 6 months now sir, your explanations are truly amazing..

    • @statquest
      @statquest  4 роки тому +2

      Thank you very much! :)

  • @tassoskat8623
    @tassoskat8623 4 роки тому +58

    This is by far my favorite educational UA-cam channel.
    Everything is explained in a simple, practical and fun way.
    The videos are full of positive vibes just from the beginning with the silly song entry. I love the catch phrases.
    Statquest is addictive!

    • @statquest
      @statquest  4 роки тому +2

      Thank you very much! :)

  • @amirrezamousavi5139
    @amirrezamousavi5139 3 роки тому +2

    My little knowledge about machine learning could not be derived without your tutorials. Thank you very much

  • @sakhawath19
    @sakhawath19 4 роки тому +7

    If I remember all the best educator's name on UA-cam, you always come at the beginning! You are a flawless genius!

  • @TheVijaySaravana
    @TheVijaySaravana 3 роки тому +2

    I have watched over 2-3 hours of lecture about Gaussian Naive Bayes. Now is when I feel my understanding is complete.

  • @minweideng4595
    @minweideng4595 4 роки тому +8

    Thank you Josh. You deserve all the praises. I have been struggling with a lot of the concepts on traditional classic text books as they tend to "jump" quite a lot. You channel brings all of them to life vividly. This is my go to reference source now.

    • @statquest
      @statquest  4 роки тому +2

      Awesome! I'm glad my videos are helpful.

  • @mohit10singh
    @mohit10singh 4 роки тому +3

    I am a beginner in Machine Learning field, and your channel helped me alot, almost went through all the videos, very nice way of explaining. Really appreciate you for making these videos and helping everyone. You just saved me ... Thank you very much...

    • @statquest
      @statquest  4 роки тому

      Thank you very much! :)

  • @argonaise_jay
    @argonaise_jay 2 роки тому +1

    One of the best channel for learners that the world can offer..

  • @leowei2575
    @leowei2575 Рік тому +2

    WOOOOOOW. I watched every video of yours, recommended in the description of this video, and now this video. Everything makes much more sense now. It helped me a lot to undersand the Gaussian Naive Bayes algorithm implemented and available from scikit-learn for applications in machine learning. Just awesome. Thank you!!!

  • @pinesasyg9894
    @pinesasyg9894 2 роки тому +2

    amazing kowledge with incredible communication skills..world will change if every student has such great teacher

  • @samuelbmartins
    @samuelbmartins 3 роки тому +3

    Hi, Josh.
    Thank you so much for all the exceptional content from your channel.
    Your work is amazing.
    I'm a professor in Brazil of Computer Science and ML and your videos have been supporting me a lot.
    You're an inspiration for me.
    Best.

  • @yuxinzhang4228
    @yuxinzhang4228 4 роки тому +5

    It's amazing! Thank you so much !
    Our professor let us self-teach the Gaussian naive bayes and I absolutely don't understand her slides with many many math equations. Thanks again for your vivid videos !!

  • @hli2147
    @hli2147 3 роки тому +2

    This is the only lecture that makes me feel not stupid...

  • @Godofwarares1
    @Godofwarares1 Рік тому +11

    This is crazy I went to school for Applied Mathematics and it never crossed my mind that what I learned was machine learning as chatgpt came into the lime light I started looking into it and almost everything I've learned so far is basically everything I've learned before but in a different context. My mind is just blown that I was assuming ML was something unattainable for me and it turns out I've been doing it for years

    • @statquest
      @statquest  Рік тому +5

      bam!

    • @yx1474
      @yx1474 3 місяці тому

      same applied math undergraduate student who switched to AI field as a postgraduate student now🙂

  • @WorthyVII
    @WorthyVII 2 роки тому +2

    Literally the best video ever on this.

  • @zitravelszikazii894
    @zitravelszikazii894 5 місяців тому +1

    Thank you for the prompt response. I’m fairly new to Stats. But this video prompted me to do a lot more research and I’m finally confident on how you got to the result. Thank you for your videos. They are so helpful

    • @statquest
      @statquest  5 місяців тому

      Glad it was helpful!

  • @sairamsubramaniam8316
    @sairamsubramaniam8316 3 роки тому +1

    Sir, this playlist is a one-stop solution for quick interview preparations. Thanks a lot sir.

    • @statquest
      @statquest  3 роки тому

      Good luck with your interviews! :)

  • @joganice2197
    @joganice2197 5 місяців тому +1

    this was the best explanation i've ever seen in my life, (i'm not even a english native speaker, i'm brazilian lol)

    • @statquest
      @statquest  5 місяців тому +1

      Muito obrigado! :)

  • @haofu1673
    @haofu1673 4 роки тому +2

    Great video! If people are willing to spend time on videos like this rather than Tiktok, the wold would be a much better place.

    • @statquest
      @statquest  4 роки тому

      Thank you very much! :)

  • @har_marachi
    @har_marachi 2 роки тому +1

    😅😅😅😅It's the "Shameless Self Promotion" for me... Thank you very much for this channel. Your videos are gold. The way you just know how to explain these hard concepts in a way that 5-year-olds can understand... To think that I just discovered this goldmine this week.
    God bless you😇

    • @statquest
      @statquest  2 роки тому

      Thank you very much! :)

  • @qbaliu6462
    @qbaliu6462 7 місяців тому +1

    This channel has helped me so much during my studies 🎉

    • @statquest
      @statquest  7 місяців тому

      Happy to hear that!

  • @chenzhiyao834
    @chenzhiyao834 3 роки тому +1

    you explained much clearer than my lecturer in ML lecture.

  • @tianhuicao3297
    @tianhuicao3297 4 роки тому +9

    These videos are amazing !!! Truly a survival pack for my DS class👍

  • @Theviswanath57
    @Theviswanath57 3 роки тому +3

    In Stats Playlist, we used following notation for P( Data | Model ) for probability & L(Model | Data) for likelihood;
    Here we are writing likelihood as L(popcorn=20 | Loves) which I guess L( Data | Model );

    • @statquest
      @statquest  3 роки тому +2

      Unfortunately the notation is somewhat flexible and inconsistent - not just in my videos, but in the the field in general. The important thing is to know that likelihoods are always the y-axis values, and probabilities are the areas.

    • @Theviswanath57
      @Theviswanath57 3 роки тому +1

      @@statquest understood; somewhere in the playlist you mentioned that likelihood is relative probability; and I guess this neatly summaries how likelihood and probability

    • @radicalpotato666
      @radicalpotato666 Рік тому

      I just had the exact same question when I started writing the expression in my notebook. I am more acquainted with the L(Model | Data) notation.

  • @WillChannelUS
    @WillChannelUS 4 роки тому +2

    This channel should have 2.74M subscribers instead of 274K.

    • @statquest
      @statquest  4 роки тому

      One day I hope that happens! :)

  • @sampyism
    @sampyism 5 місяців тому +1

    Your videos and voice make ML and statistics fun to learn. :)

    • @statquest
      @statquest  5 місяців тому

      Glad you like them!

  • @sudhashankar1040
    @sudhashankar1040 4 роки тому +1

    This video on Gaussian Naive Bayes has been very well explained. Thanks a lot.😊

  • @samuelschonenberger
    @samuelschonenberger 2 роки тому +2

    These gloriously wierd examples really are needed to understand a concept

  • @MrRynRules
    @MrRynRules 3 роки тому +1

    Daym, your videos are so good at explaining complicated ideas!! Like holy shoot, I am going to use this, multiple predictors ideas to figure out the ending of inception, Was it dream, or was it not a dream!

  • @maruthiprasad8184
    @maruthiprasad8184 2 місяці тому +1

    superb cool explanation. I am big fan of your explanation. Once I went through your explanation, I don't want any further reference for that topic.

  • @georgeruellan
    @georgeruellan 4 роки тому +3

    This series is helping me so much with my dissertation, thank you!!

    • @statquest
      @statquest  4 роки тому +1

      Awesome and good luck with your disertation!

  • @Adam_0464
    @Adam_0464 4 роки тому +1

    Thank you, You have made the theory concrete and visible!

  • @CyberGimen
    @CyberGimen 3 місяці тому +1

    Bam! I love your teaching style!!!

    • @statquest
      @statquest  3 місяці тому

      Thanks!

    • @CyberGimen
      @CyberGimen 3 місяці тому

      @@statquest I think you should explain some formula briefly. Like in Naive Bayes algorithm, you'd better explain why P(N)*P(Dear|N)*P(Friend|N)=P(N|Dear,Friend). I use GPT to finally understand it.

    • @statquest
      @statquest  3 місяці тому

      @@CyberGimen I've got a whole video about that here: ua-cam.com/video/9wCnvr7Xw4E/v-deo.html However, the reason I don't mention it in this video is that it's actually not critical to using the method.

  • @liranzaidman1610
    @liranzaidman1610 4 роки тому +5

    How do people come up with these crazy ideas? it's amazing, thanks a lot for another fantastic video

  • @anje889
    @anje889 Рік тому +1

    contents are excellent and also i love your intro quite a lot (its super impressive for me) btw. thanking for doing this at the fisrt place as a beginner some concepts are literally hard to understand but after watching your videos things are a lot better than before. Thanks :)

    • @statquest
      @statquest  Рік тому

      I'm glad my videos are helpful! :)

  • @auzaluis
    @auzaluis 4 роки тому +1

    The world needs more Joshuas!

  • @akashchakraborty6431
    @akashchakraborty6431 4 роки тому +1

    You have really helped me a lot. Thanks Sir. May you prosper more and keep helping students who cant afford paid content :)

  • @Vivaswaan.
    @Vivaswaan. 4 роки тому +1

    The demarcation of topics in the seek bar is useful and helpful. Nice addition.

    • @statquest
      @statquest  4 роки тому +1

      Glad you liked it. It's a new feature that UA-cam just rolled out so I've spent the past day (and will spend the next few days) adding it to my videos.

    • @anitapallenberg690
      @anitapallenberg690 4 роки тому +2

      @@statquest We really appreciate all your dedication into the channel!
      It's 100% awesomeness :)

    • @statquest
      @statquest  4 роки тому

      @@anitapallenberg690 Hooray! Thank you! :)

  • @ADESHKUMAR-yz2el
    @ADESHKUMAR-yz2el 4 роки тому +1

    i promise i will join the membership and buy your products when i get a job... BAM!!!

    • @statquest
      @statquest  4 роки тому

      Hooray! Thank you very much for your support!

  • @Steve-3P0
    @Steve-3P0 4 роки тому +1

    +5000 for using an example as obscure and as obscene as Troll 2.

  • @Mustafa-099
    @Mustafa-099 2 роки тому +1

    Hey Josh I hope you are having a wonderful day, I was searching for a video on " Gaussian mixture model " on your channel but couldn't find one, I have a request for that video since the concept is a bit complicated elsewhere
    Also btw your videos enabled to get one of the highest scores in the test conducted recently in my college, all thanks to you Josh, you are awesome

    • @statquest
      @statquest  2 роки тому +1

      Thanks! I'll keep that topic in mind.

  • @prashuk-ducs
    @prashuk-ducs 5 місяців тому

    Why the fuck does this video make it look so easy and makes 100 percent sense?

  • @meysamamini9473
    @meysamamini9473 3 роки тому +1

    I'm Having great time watching Ur videos ❤️

  • @yuniprastika7022
    @yuniprastika7022 4 роки тому +1

    can't wait for your channel to BAAM! going worldwide!!

  • @jiheonlee4065
    @jiheonlee4065 4 роки тому +2

    Thank you for another excellent Statquest !~

  • @camilamiraglia8077
    @camilamiraglia8077 4 роки тому

    Thanks for the great video!
    I would just like to point out that in my opinion if you are talking about log() when the base is e, it is easier (and more correct) to write ln().

    • @statquest
      @statquest  4 роки тому +1

      In statistics, programming and machine learning, "ln()" is written "log()", so I'm just following the conventions used in the field.

  • @sayanbhowmick9203
    @sayanbhowmick9203 8 місяців тому +1

    Great style of teaching & also thank you so much for such a great video (Note : I have bought your book "The StatQuest illustrated guide to machine learning") 😃

    • @statquest
      @statquest  8 місяців тому +1

      Thank you so much for supporting StatQuest!

  • @konstantinlevin8651
    @konstantinlevin8651 Рік тому +1

    I'm a simple man, I watch statquests in the nights, leave a like and go chat about it with chatgpt.That's it.

  • @tcidude
    @tcidude 4 роки тому

    Josh. I love you your videos. I've been following your channel for a while. Your videos are absolutely great!
    Would you consider covering more of Bayesian statistics in the future?

  • @heteromodal
    @heteromodal 3 роки тому +1

    Thank you Josh for another great video! Also, this (and other vids) makes think I should watch Troll 2, just to tick that box.

    • @statquest
      @statquest  3 роки тому

      Ha! Let me know what you think!

  • @therealbatman664
    @therealbatman664 2 роки тому +1

    Your videos are really great !! my prof made it way harder!!

  • @ahhhwhysocute
    @ahhhwhysocute 4 роки тому +1

    Thanks for the video !! it was very helpful and easy to understand

  • @r0cketRacoon
    @r0cketRacoon 4 місяці тому

    A really comprehensive video. Thank you!
    Sir, I have some questions about the conditions when applying this algo:
    1. Is it compulsory that all features contain continuous value?
    2. What happens if a feature doesn't have gaussian distribution? Is it worth to apply this algo?
    3. If that, I will find a function that makes that feature have gaussian distribution. Can it work?
    And also, Do u plan to do a video about Bernoulli Naive Bayes?

    • @statquest
      @statquest  4 місяці тому +1

      1. No - you can mix things up. I illustrate this in my book.
      2. You can use other distributions
      3. No need, just use the other distribution.
      4. Not in the short term.

  • @mukulsaluja6109
    @mukulsaluja6109 3 роки тому +1

    Best video i have ever seen

  • @rogertea1857
    @rogertea1857 3 роки тому +1

    Another great tutorial, thank you!

  • @ahmedshifa
    @ahmedshifa 8 місяців тому

    These videos are extremely valuable, thank you for sharing them. I feel that they really help to illuminate the material.
    Quick question though: where do you get the different probabilities, like for popcorn, soda pop, and candy? How do we calculate those in this context? Do you use the soda a person drinks and divide it by the total soda, and same with popcorn, and candy?

    • @statquest
      @statquest  8 місяців тому

      What time point are you asking about (in minutes and seconds). The only probabilities we use in this video are if someone loves or doesn't love troll 2. Everything else is a likelihood, which is just a y-axis coordinate.

  • @sejongchun8350
    @sejongchun8350 4 роки тому +2

    Troll 2 is an awesome classic, and should not be up for debate. =)

  • @konmemes329
    @konmemes329 3 роки тому +1

    Your video just helped me a lot !

  • @tagoreji2143
    @tagoreji2143 2 роки тому +1

    Tqsm Sir for the Very Valuable Information

  • @Geza_Molnar_
    @Geza_Molnar_ 4 роки тому

    Hi - another great explanation!
    I wonder what would be the result if you normalise the probabilies of the 3 values.
    - Would it affect the outcome of the example in this video?
    - Which areas of values are affected: different outcomes with non-normalised and normalised distributions (=probability or likelihood here)?

    • @statquest
      @statquest  4 роки тому

      Interesting questions! You should try it out and see what you get.

    • @Geza_Molnar_
      @Geza_Molnar_ 4 роки тому

      @@statquest Hi, that only make sense with real data. Without that, only juggling with equations and abstract parameters, the thing is not enough 'visual', IMO. Though, could run through the calculations with e.g. 2x scale, 10x scale and 100x scale... Maybe, when I have free few hours.

  • @alanamerkhanov6040
    @alanamerkhanov6040 11 місяців тому +1

    Hi, Josh. Troll 2 is a good movie... Thanks

  • @worksmarter6418
    @worksmarter6418 3 роки тому +1

    Super awesome, thank you. Useful for my Intro to Artificial Intelligence course.

  • @AmanKumar-oq8sm
    @AmanKumar-oq8sm 4 роки тому

    Hey Josh, Thank you for making these amazing videos. Please make a video on the "Bayesian Networks" too.

  • @dipinpaul5894
    @dipinpaul5894 4 роки тому +1

    Excellent explanation. Any NLP series coming up ? Struggling to find good resources.

    • @statquest
      @statquest  4 роки тому +4

      I'm working on Neural Networks right now.

    • @ragulshan6490
      @ragulshan6490 4 роки тому +1

      @@statquest it's going to be BAM!!

  • @콘충이
    @콘충이 4 роки тому +3

    Can you talk about Kernel estimation in the future?? Bam!

  • @johnel4005
    @johnel4005 3 роки тому +1

    BAM! Someone is going to pass the exam this semester .

  • @initdialog
    @initdialog 4 роки тому

    Finally worked up to the Gaussian Naive Bayes. BAM! "If you are not familiar with
    ..." :(

    • @anitapallenberg690
      @anitapallenberg690 4 роки тому +1

      You can do it! :)
      StatQuest made me lose my anxiety for statistics. It's truly brilliant, just start with the next video!

    • @statquest
      @statquest  4 роки тому

      BAM! :)

  • @samuelbmartins
    @samuelbmartins 3 роки тому

    Josh, a question about the formulation of Bayes' Theorem, especially considering the likelihood.
    For Naive Bayes, the formula is:
    P(class | X) = P(class) * P(X | class), in which the last term. is the likelihood
    In your video, you represented the likelihood as L, so that, apparently, the formula would be:
    P(No Love | X) = P(No Love) * L(X | No Love)
    (1) Is my assumption correct? Is it just a change of letters to mean the same thing?
    (2) Or is there any other math under the hoods?
    For example, something like: P(X | class) = L(No Love | X)
    Thanks in advance.

    • @statquest
      @statquest  3 роки тому

      When I use the notation "L(something)" for "likelihood", I mean that we want the corresponding y-axis coordinate for that something. However, not everyone uses that notation. Some put p(something) and you have to figure out from the context whether or not they are talking about a likelihood (y-axis coordinate) or, potentially, a probability (since "p" often refers to "probability"). So, if you use my notation, then you are correct, you get: P(No Love | X) = P(No Love) * L(X | No Love)

  • @MinhPham-jq9wu
    @MinhPham-jq9wu 3 роки тому +1

    So great, this video so helpful

    • @statquest
      @statquest  3 роки тому +1

      Glad it was helpful!

  • @aicancode5676
    @aicancode5676 4 роки тому

    I dont even know why there is people disliking this video!!

    • @statquest
      @statquest  4 роки тому

      It's always a mystery. :)

  • @MrElliptific
    @MrElliptific 4 роки тому

    Thanks for this super clear explanation. Why would we prefer this method for classification over a gradient boosting algorithm? When we have too few samples?

    • @statquest
      @statquest  4 роки тому

      With relatively small datasets it's simple and fast and super lightweight.

  • @RFS_1
    @RFS_1 3 роки тому +1

    Love the explaination BAM!

  • @nzsvus
    @nzsvus 4 роки тому

    BAM! thanks, Josh! It would be amazing if you can make a StatQuest concerning A/B testing :)

    • @statquest
      @statquest  4 роки тому +1

      It's on the to-do list. :)

  • @sheebanwasi2925
    @sheebanwasi2925 3 роки тому +1

    Hey JOSH Thanks for making such amazing video. Keep up the work. I just have a quick question if you don't mind.
    I can't understand how you got the likelihood eg: L(soda = 500 | LOVES) how you calculating that value.

    • @statquest
      @statquest  3 роки тому

      We plugged the mean and standard deviation of soda pot for people that loved Troll2 into the equation for a normal curve and then determined the y-axis coordinate when the x-axis value = 500.

  • @patrycjakasperska7272
    @patrycjakasperska7272 Рік тому +1

    Love your channel

  • @hawkiyc
    @hawkiyc 4 роки тому

    Dear Mr. Josh,
    I have taken another course have the following equation for the probability of Naive Bayes,
    P( Loves Troll 2 | new data ) = [ P( new data | Loves Troll 2 ) * P( Loves Troll 2 ) ] / P( new data)
    P( new data ) called marginal likelihood,
    and P( new data | Loves Troll 2 ) called likelihood
    And then, the way to calculate marginal likelihood and likelihood is to calculate the probability nearby the data at a certain distance, and the distance is adjustable while you are building the algorithm. For instance, there is a circle in which the center is new data and you can adjust the radius if your data is 2-D data.
    After watching both courses, I am wondering how can these two equations be equivalent?
    I deeply appreciate your time for answering my question.
    Sincerely,
    Gavin

    • @statquest
      @statquest  4 роки тому

      The marginal likelihood is often omitted because both p(Loves Troll 2 | data) and p(Does not love Troll 2 | data) are divided by it. In other words, the only thing that makes p(Loves Troll 2 | data) different from p(Does not love Troll 2 | data) is what is in the numerator. And because it is usually really hard to calculate the marginal likelihood, we just omit it because it will not change the results.

  • @vinaykumardaivajna5260
    @vinaykumardaivajna5260 Рік тому +1

    Awesome as always

  • @jonathanjacob5453
    @jonathanjacob5453 10 місяців тому +1

    Looks like I have to check out the quests before getting to this one😂

  • @sumanbindu2678
    @sumanbindu2678 2 роки тому +1

    Amazing videos. The beep boop sound reminds me of squid games

    • @statquest
      @statquest  2 роки тому +1

      Maybe they got the sound from my video! :)

  • @danielasanabria3242
    @danielasanabria3242 4 роки тому +2

    So we use Gaussian when ALL our features are continuous and multinomial when ALL our features are categorical?

    • @statquest
      @statquest  4 роки тому

      You can mix them by multiplying the different likelihoods. For more details, see: sebastianraschka.com/faq/docs/naive-bayes-vartypes.html

  • @diraczhu9347
    @diraczhu9347 3 роки тому +1

    Great video!

  • @kirilblazevski8329
    @kirilblazevski8329 Рік тому +2

    Since the likelihood can be greater than 1, doesn't that mean that we could get probability that is greater than 1?

    • @statquest
      @statquest  Рік тому +1

      No, probability is the area under the curve and those are defined such that the total area under the curve is always 1. For details, see: ua-cam.com/video/pYxNSUDSFH4/v-deo.html

    • @kirilblazevski8329
      @kirilblazevski8329 Рік тому

      @@statquest
      Dear Dr. Starmer,
      Thank you for your reply. I have another follow-up question regarding the calculation of probabilities for continuous random variables (i.e. what this video is about).
      From my understanding, when we have discrete random variables, the probability of a given outcome P(Y=y|X1,X2,..Xn) is proportional to the product of the probabilities of the individual variables given the outcome, times the prior probability (assuming conditional independence).
      i.e. P(Y=y) * the product of P(Xi=xi | Y=y)
      This makes sense to me, because the result is a probability value between 0 and 1.
      However, in the case of continuous random variables, the probability of a given outcome is zero, so we instead calculate the likelihood of the outcome. This means that the product of the individual likelihoods is no longer a probability value between 0 and 1. Is this correct?
      What I mean is: P(Y=y) * the product of L(Xi=xi | Y=y) is not guaranteed to be a value between 0 and 1.
      Thank you for your expertise and for being such a valuable educator. 💖

    • @statquest
      @statquest  Рік тому +1

      @@kirilblazevski8329 That's correct, with the continuous version, we do not end up with probabilities. However, if you saw my video on the discrete version of Naive Bayes ( ua-cam.com/video/O2L2Uv9pdDA/v-deo.html ) you'll notice that I call the results "scores" instead of probabilities. The reason for this is that in both cases (discrete and continuous), to get the correct probabilities for the results, you need to divide the results (what I call "scores") by the sum of the scores for the two possibilities. By doing this, you normalize the scores for the two possibilities so that they will add up to 1.

    • @kirilblazevski8329
      @kirilblazevski8329 Рік тому +1

      @@statquest Now I understand what I was missing. Thank you for clarifying, I really appreciate it!!

  • @kicauburungmania2430
    @kicauburungmania2430 4 роки тому

    Thanks for the awesome explanation. But I've a question. Is GNB can be used for sentiment analysis?

    • @statquest
      @statquest  4 роки тому

      Presumably you could use GNB, but I also know that normal NB (aka multinomial naive bayes) is used for sentiment analysis.

  • @roymillsdixton7941
    @roymillsdixton7941 Рік тому

    A nice video on Gaussian Naive Bayes Classification model. Well done! But I have a quick question for you, Josh. I only understand that Lim ln(x) as x approaches o is negative infinity. How is the Natural log of a really small unknown number very close to zero assumed to be equal to -115 and -33.6 as in the case of L(candy=25|Love Troll 2) and L(popcorn=20|does not Love Troll 2) respectively? What measure was used to determine these values?

    • @statquest
      @statquest  Рік тому

      log(1.1*10^-50) = -115 and log(2.5*10^-15) = -33.6

  • @taotaotan5671
    @taotaotan5671 4 роки тому

    Hi, Josh. Thanks for this clear explanation. Since this Naive Bayes could be applied to Gaussian distribution, I guess it could also be applied to other distributions like Poisson distribution, right? Then a question is: how to determine the distribution of a feature? I believe this will be quite important to build a reasonable model.
    Thanks again for the nice video.

    • @statquest
      @statquest  4 роки тому

      One day (hopefully not too long from now), I'm going to cover the different distributions, and that should help people decide which distributions to use with their data.

  • @mohammadelghandour1614
    @mohammadelghandour1614 2 роки тому

    Great work ! In 8:11 How can we use cross validation with Gaussian Naive Bayes? I have watched the Cross validation video but I still can't figure out how to employ cross validation to know that candy can make the best classification.

    • @statquest
      @statquest  2 роки тому

      to apply cross validation, we divide the training data into different groups - then we use all of the groups, minus 1, to create a gaussian naive bayes model. Then we use that model to make predictions based on the last group. Then we repeat, each time using a different group to test the model.

  • @franssjostrom719
    @franssjostrom719 3 роки тому +1

    Tough being a ML teacher these days with you around

  • @BrianRisk
    @BrianRisk 4 роки тому +3

    People who don't like this video are obviously trolls

  • @jacobwinters6648
    @jacobwinters6648 5 місяців тому

    Hello! Does it matter if the data in one of the columns (say popcorn) is not normally distributed? Or should the assumption be that we will have a large enough sample size to use the central limit theorem?
    Thanks for all of your videos! I love them and can’t wait for your book to be delivered (just ordered it yesterday).

    • @statquest
      @statquest  5 місяців тому +1

      It doesn't matter how the data are distributed. As long as we can calculate the likelihoods, we are good to go. BAM! :) And thank you so much for supporting StatQuest!!! TRIPLE BAM!!! :)

  • @shailukonda
    @shailukonda 4 роки тому +1

    Could you please make a video on Time Series Analysis (Arima model)?

    • @statquest
      @statquest  4 роки тому +1

      One day I'll do that.

  • @imbored8699
    @imbored8699 4 роки тому +2

    Can we also say that this person can be an outlier? Because of having very high likelihood of popcorn and soda pop scores given that he likes troll 2 and only but high variance according to 3rd category we can also say consider him under the outlier category, can't we? Can you clear this doubt for me, please! And also thanks a lot for your effort and work..

    • @statquest
      @statquest  4 роки тому

      Maybe. It depends on how much data we have in the training dataset - because that will define how confident we are that we have correctly modeled the two categories.

    • @imbored8699
      @imbored8699 4 роки тому +1

      @@statquest
      Yes!
      If the training dataset contains a good enough number of data then we can calculate the margin of error too at various confidence levels with the given sample size and present our output.
      Thank you!

  • @Darkev77
    @Darkev77 2 роки тому +1

    3:38, shouldn’t the notation be L(Loves | popcorn=20), since we’re given that he eats 20g of popcorn, how likely is that sample generated from the Loves distribution. Isn’t that right?

    • @statquest
      @statquest  2 роки тому

      The notation in the video is most common, however, the notation doesn't really matter as long as it is clear that we want the y-axis coordinate.

  • @deepshikhaagarwal4125
    @deepshikhaagarwal4125 Рік тому +1

    Thank you josh your videos are amazing! HoW to buy study guides from statquest

    • @statquest
      @statquest  Рік тому

      See: statquest.gumroad.com/

  • @ArinzeDavid
    @ArinzeDavid 2 роки тому +1

    awesome stuff for real

  • @adamkinsey3139
    @adamkinsey3139 3 роки тому

    Error in Video ("Probably")
    Hey Josh, I think you should be saying "Probability" when you are saying Likelihood.
    I know the distinction can be tough, but they are VERY different, so it is important to get right. And I believe you explain it correctly in your other video, so I'm surprised to see it used incorrectly here.
    When you already have a known distribution (not changing), and you calculate a value for an event based on that (known) distribution, that is a Probability.
    "Likelihood" occurs when you have some real data, but you're not sure what the "true" distribution is that the data came from. So you take a stab and pick one possible distribution, and then calculate the "likelihood" that the data came from that distribution that you guessed.
    And seriously, I think that the difference between "probability" and "likelihood" is as big as consequential as the difference between differentiation and integration, so it's really important to clearly distinguish the two concepts.
    Or am I wrong?

    • @statquest
      @statquest  3 роки тому

      With continuous distributions, the probability of an individual event is always 0, this is because probability is the area under the curve between two points. If we only have 1 point, then the width of the area is 0 and thus, the probability is 0. So, in Naive Bayes we do not use probability, instead we use the y-axis coordinate, which I call the likelihood because that is what the y-axis coordinates are called when we do "maximum likelihood". I know that statisticians only call it a likelihood when we are specifically changing parameters and holding the data as fixed, but I believe we can also call it a likelihood since the values are equal to each other and they are calculated the same way. In other words, if x = a number that is calculated the same way as a likelihood and is the same value as a likelihood, then I'll say x is equal to a likelihood.

    • @adamkinsey3139
      @adamkinsey3139 2 роки тому

      @@statquest Hmmm. Well, in the beginning, you take training data, and just ASSUME that the distributions of the data in each variable is normal, when you actually have no idea whether the distributions of each variable are actually "normal"/gaussian. All of the values that I suggested should be called "probabilities" are calculated based on that assumption that the distributions are normal. But since we don't know that they are normal, I suppose they truly are "likelihoods". They are the likelihoods of the normal distribution assumption being correct given the actual data. So maybe that is why they are called "Likelihoods" in Naive Bayes.

    • @statquest
      @statquest  2 роки тому

      @@adamkinsey3139 In statistics, when we have a continuous distribution like the normal distribution, likelihoods are distinct from probabilities (for details, see: ua-cam.com/video/pYxNSUDSFH4/v-deo.html ). That said, we do assume that the distribution is normal, however, that is not a requirement. We can assume that the data come from different distributions and the method will work (in a technical sense) just as well. Whether or not the results will be improved, however, depends on the actual underlying distributions.

  • @pinksincerely
    @pinksincerely 3 роки тому

    Hi Josh, as always thanks so much for the very informative video!!! Quick question, how did you calculate for the likelihoods? :D

    • @statquest
      @statquest  3 роки тому

      I plugged the x-axis coordinate, the mean and the standard deviation into the dnorm() function in R.

    • @atravellingstudent
      @atravellingstudent 2 роки тому

      @@statquest I am confused, I thought pdf != likelihood, but videos suggest otherwise.

    • @statquest
      @statquest  2 роки тому

      @@atravellingstudent PDF is just the curve for a continuous distribution if we integrate that curve between two points, we get the probability of something happening between those two points. If we look at the y-axis value for a specific point, we get the likelihood. For details, see: ua-cam.com/video/pYxNSUDSFH4/v-deo.html

  • @mahadmohamed2748
    @mahadmohamed2748 3 роки тому

    Thanks for these great videos! Quick question: In other resources the likelihood is actually the probability of the data given the hypothesis rather than the likelihood of the data given the hypothesis. Which one would be correct, or is it fine to use either?

    • @statquest
      @statquest  3 роки тому +1

      Generally speaking, you use the likelihoods of the data. However, we can normalize them to be probabilities. This does not offer an advantages and takes longer to do, so people usually omit that step and just use the likelihoods.