Training Latent Dirichlet Allocation: Gibbs Sampling (Part 2 of 2)

Поділитися
Вставка
  • Опубліковано 27 чер 2024
  • This is the second of a series of two videos on Latent Dirichlet Allocation (LDA), a powerful technique to sort documents into topics. In this video, we learn to train an LDA model using Gibbs sampling.
    The first video is here: • Latent Dirichlet Alloc...
  • Наука та технологія

КОМЕНТАРІ • 118

  • @amirrahimi6979
    @amirrahimi6979 3 роки тому +34

    This is the best tutorial that I could find around the web. I'm going to code and put it in my dissertation references. Also, I've watched it several times 😁 thank you

  • @Ke_Mis
    @Ke_Mis 3 роки тому +4

    Exceptional skills in breaking down complex topics into simple and relatable words! Well done!

  • @ravenfortyfive
    @ravenfortyfive 2 роки тому +4

    Words fail to thank you, your explanation is very, very wonderful. I hope you will continue explaining all the models of machine learning in this way and add subtitles to your videos.

  • @marijanmanoilov9407
    @marijanmanoilov9407 2 роки тому +11

    Luis, you are very talented as tutor/teacher. I'am impressed! Glad to found you and I am your new subscriber :) Keep the good work and wish you all the best!

  • @wasiimo
    @wasiimo 4 роки тому +7

    Fantastic content my man. Keep doing what you're doing.

  • @shivadhanush3131
    @shivadhanush3131 4 роки тому +3

    God... Just loved your videos on LDA... I had been running all over the sources to understand LDA. You dropped in like a GOD.. Sigmoid(Thanks)

  • @apah
    @apah 2 роки тому +5

    Man oh man Luis, you videos are a treat. You really deserve to know how much you're helping me and my fellow learners out, and how much it means to us. First saw you in the deep learning introductory course in udacity, and loved the content. Can't express how happy I was to discover you had a youtube channel ! Thank you !
    PS : it was also fun rewatching the RNN/LSTM part of that course, and noticing that the person teaching us there was none other than Alexis Cook, who I had some occasions of briefly discussing with in Kaggle learn forums. Realizing that gave me a Machine Learning Cinematic Universe vibes hahah

    • @SerranoAcademy
      @SerranoAcademy  2 роки тому +1

      Thank you for your kind message, it's so nice to hear the videos are helpful!
      Yes, I enjoyed working from Alexis and I learned a lot from her. Check out her blog, she has some nice DL posts: alexisbcook.github.io/

  • @coffeezealot8535
    @coffeezealot8535 4 роки тому

    best material on LDA online! this should have far more views.

  • @pulkitsapra1115
    @pulkitsapra1115 4 роки тому

    A very simple and easy to understand explanation for a complex model. Amazing Video!

  • @crna4
    @crna4 3 роки тому

    Muchas gracias Luis. You are very good at explaining complex things in a comprehensive way. Your videos help me a lot. Keep going!

  • @jmrludan
    @jmrludan 3 роки тому

    Amazingly intuitive once again! Great series.

  • @glowish1993
    @glowish1993 4 роки тому +3

    the best youtube videos are always the ones kept secret... thank you so so much the explanation is superb. i would not have gotten a role as a data scientist if not for amazing teachers like yo

  • @dragolov
    @dragolov 4 роки тому +2

    Thank you very much, Luis Serrano! You are so talented teacher!

  • @yiyichanmyaewinshein8248
    @yiyichanmyaewinshein8248 Рік тому

    this is the right place I have been looking around for LDA. Thank you very much. It will be better when you also provide the lectures with the code.

  • @JK-hi2nx
    @JK-hi2nx 3 роки тому

    Your explanation is so easy to understand and very clear. Thank you so much.

  • @akshatbhandari9798
    @akshatbhandari9798 2 роки тому

    This was an amazing series. Thanks a ton!!

  • @danielibanez7952
    @danielibanez7952 3 роки тому

    I found it clear and engaging. Thanks for your work!

  • @aleschudarek4672
    @aleschudarek4672 Рік тому

    Man, I think that our teacher bases his lessons and our projects themes of of your vids :D. Thanks. Really helped.

  • @lettry5297
    @lettry5297 2 роки тому +1

    You are a true gem in teaching 😍. Please make more video as possible

  • @astaconteazamaiputin
    @astaconteazamaiputin 3 роки тому

    Best intuitive explanation of LDA I have seen. Thank you for sharing ! Looking forward to Grokking NLP ;)

  • @amoldumrewal
    @amoldumrewal 3 роки тому

    Wow...what an explanation! Great work man. Next time I sit for an interview, this is the best video to help me revise one of my internships :)

  • @haxujun8007
    @haxujun8007 4 роки тому

    Awesome video for LDA! Thank you very much!

  • @mj3096
    @mj3096 4 роки тому

    Amazing explanation. Very easy to understand. Thanks a lot!

  • @petercourt
    @petercourt 4 роки тому +1

    Wonderful, and so clearly explained. Many thanks :)

  • @muhammadsarimmehdi
    @muhammadsarimmehdi 3 роки тому

    Very nice video. The part where you explained where alpha and beta came from felt like a plot twist (because I had seen the prequel to this video). Please continue making videos like these!

  • @madhavimourya1157
    @madhavimourya1157 4 роки тому

    Found very very helpful in understanding complex things in very easy and simple manner. YOU made my day :)

  • @thomasmthoppil9486
    @thomasmthoppil9486 2 роки тому

    This is amazing content. Really enjoyed the work

  • @jonasveland2655
    @jonasveland2655 3 роки тому +2

    Amazing, thank you so much. Would love to see a follow up explaining Collapsed Gibbs sampling.

  • @swagatmishra9350
    @swagatmishra9350 4 роки тому +1

    What an awsum explanation... Thank you very much!!

  • @radiaabdi6285
    @radiaabdi6285 3 роки тому

    Wonderful, and so clearly explained. Many thanks

  • @user-ad7rt9cv8h
    @user-ad7rt9cv8h Рік тому

    You made me finally understand LDA after so many times of reading many materials 😂 Thank you.

  • @lakshikaswarnamali5328
    @lakshikaswarnamali5328 3 роки тому

    Thank you very much for the clear and simple explanations !!

  • @aamirahmadi1584
    @aamirahmadi1584 3 роки тому

    excellent explanation of LDA and Dirichlet distribution. Thank you very much

  • @marcospaulorodriguescorrei8839
    @marcospaulorodriguescorrei8839 3 роки тому

    Man, for real, you explain better than 99,8% of the entire world (including the Indian hacks that help us w/ python)

  • @user-vh6cs1fg8g
    @user-vh6cs1fg8g 11 місяців тому

    Thank you for this video. I watched the first video on LDA, well presented. very talented with your pictorial presentation.

  • @777bloomingdale
    @777bloomingdale 4 роки тому +2

    Pls do a video on Metropolis Hastings and Gibbs Gibbs sampling. I am yet to find a best video on the MH with Gibbs sampling. Your intuitive approach was marvelous.

  • @suryamgupta4467
    @suryamgupta4467 Рік тому

    This is a brilliant video, thank you so much!

  • @souravdey1227
    @souravdey1227 3 роки тому

    Absolute genius. You can explain string theory to a baby. I wish I met you earlier in my life. Waiting more videos, especially maths based

  • @amardeepsingh9001
    @amardeepsingh9001 Рік тому

    Amazing! very well explained. Just outstanding. Thanks a lot!

  • @SonLe-mk4sq
    @SonLe-mk4sq Рік тому

    So good. Much appreciated.

  • @Elephant11ish
    @Elephant11ish 4 роки тому +1

    Thanks very much!! This is very helpful!!

  • @drunknoodle3188
    @drunknoodle3188 3 роки тому

    Amazing explanation!

  • @daleanfer7449
    @daleanfer7449 6 місяців тому

    you videos is remarkable!

  • @alessandrarizzi
    @alessandrarizzi 3 роки тому

    Fantastic explanation. Thank you

  • @bigchallenge7987
    @bigchallenge7987 2 роки тому

    Appreciate a great video. Thanks a lot!

  • @mariagaetanaagnesitois4844
    @mariagaetanaagnesitois4844 3 роки тому

    Thank you so much. You explain very well!

  • @HariomGautam
    @HariomGautam 4 роки тому +1

    Thanks Serrano, You made LDA simple. Can you please make next part for Variational Inference and also the different variation/modification of LDA.

  • @zhexu1727
    @zhexu1727 4 роки тому

    Super clear explanation, thx so much

  • @lokeshjindal87
    @lokeshjindal87 3 роки тому

    Thanks for the information. It helped me a lot

  • @Anshuman063
    @Anshuman063 3 роки тому

    Great content, thanks!

  • @rashmisharma4070
    @rashmisharma4070 3 роки тому

    Too good, You are superb!

  • @aravindramanan3679
    @aravindramanan3679 3 роки тому

    Thank you! this is a great video !!

  • @iconstleo
    @iconstleo Рік тому

    Great video! thanks!

  • @divyaank
    @divyaank 3 роки тому

    Great explanation

  • @monicabandaru3548
    @monicabandaru3548 4 роки тому

    Very clear explaination thank you so much

  • @liebesaktsamy
    @liebesaktsamy 4 роки тому

    Awesome teaching!!! Made LDA look very simple. Could you please make a video on Gaussian Mixture model

  • @HassanAli-os3py
    @HassanAli-os3py 3 роки тому

    Very well explained

  • @droneacharjyo7364
    @droneacharjyo7364 2 роки тому

    Amazing explaination

  • @ricky3dec1
    @ricky3dec1 2 роки тому

    Brilliant!!

  • @charliel7041
    @charliel7041 4 роки тому

    That is sick!

  • @danielxing1034
    @danielxing1034 4 роки тому

    amazingly clear.......

  • @tomasnobrega8087
    @tomasnobrega8087 2 роки тому

    amazing content thank you!

  • @InAweofhisglory
    @InAweofhisglory Рік тому

    GOAT, a colombian legend

  • @shadhopson5737
    @shadhopson5737 3 роки тому

    Thanks so much!

  • @generationgap416
    @generationgap416 3 роки тому

    I thought that I understood LDA, until now. Luis Serrano is the Richard Feynman of Machine Learning. Thanks,

  • @nidcxl4223
    @nidcxl4223 3 роки тому +1

    I understand the analogy of the room where you iteratively bring objects together assuming the others are already correctly positioned. But how did you know in the first place that the coat hanger and the pants had the same color? How do you know what the color of the words is?

  • @user-qe9gi7ok6r
    @user-qe9gi7ok6r Рік тому

    Thank you so much

  • @katateo328
    @katateo328 Рік тому

    hahha, exellent illustration the language of machine! human cannot understand what it means but we all reach the same goal in the end. Stunning!

  • @josephpareti9156
    @josephpareti9156 2 роки тому

    awesome

  • @johnbornmann4592
    @johnbornmann4592 2 роки тому

    thank you for this, I have been looking for a simple example of LDA with a visual element to show how the process works. However, in your example the "random" colors are actually very close to an accurate model. I would have loved to see it with a start point that was not pre-biased toward the correct answer.

    • @mitra.ashutosh
      @mitra.ashutosh Рік тому

      My doubt is at what point in this algorithm do we compare the generated docs and the actual docs.

  • @truongphan5275
    @truongphan5275 3 роки тому

    It is awesome! Could you explain more about dividers in Gibbs sampling (in demo)?

  • @includestdio.h8492
    @includestdio.h8492 7 місяців тому

    I didn't expected quechua, hahaha.. que grande q eres!

  • @abyssus9934
    @abyssus9934 2 роки тому

    Thank you so much for this explanation. Very intuitive. I have one more question about this tiny alpha and beta add to avoid zero-out. Since we have no idea what these two 'real dirichlet' look, how could we estimate or determine the value for that? Appreciate it!

  • @aladdinpersson1915
    @aladdinpersson1915 3 роки тому

    Thanks for the video! Do you have any resources to read about LDA and Gibbs sampling? Would specifically like to learn more about the math behind it

  • @MrStudent1978
    @MrStudent1978 4 роки тому +5

    Very interesting!
    How is this different from Naive Bayes based document classification?

    • @atifadib
      @atifadib 3 роки тому +1

      you need labels for naive bayes, its supervised

  • @AI_Financier
    @AI_Financier 2 роки тому

    thank you sir, a great video, how do you come up with best values for alpha and beta?

  • @gautamcheek1
    @gautamcheek1 2 роки тому

    Very nicely explained. How about cosine similarity vs Gibbs sampling for topic modelling ?

  • @akshaybhandari8951
    @akshaybhandari8951 3 місяці тому

    This has to be the best video.
    Do you think i can use this to organize files using the metadata and name?

  • @Justin-General
    @Justin-General 2 роки тому

    Wonderful explanation, still can't understand the equations though. Having great difficulty with that leap.

  • @arwan07
    @arwan07 Рік тому

    Very sharp explanation. How to build search a documents with lda? How to get best query which could locate cluster of documents?

  • @AshutoshRaj
    @AshutoshRaj 2 роки тому

    This video explains I need to know about LDA but one small doubt @4:08min of the video. How did you assign color to each word? Is it random or some rules we need to follow.

  • @trek8388
    @trek8388 Рік тому

    I wish you were my teacher!

  • @swagatmishra9350
    @swagatmishra9350 3 роки тому

    Thanks Luis.. Awsum work.
    Just had one confusion. at @4:13 you assigned the topics to the words. Is it assigned randomly or is there any logic behind this?
    Thanks again for the amazing work.

    • @mariagaetanaagnesitois4844
      @mariagaetanaagnesitois4844 3 роки тому

      i think it was assigned randomly, and then will be update with the gibbs sampling process until get the right one

  • @NawshadFarruque
    @NawshadFarruque 4 роки тому +1

    How do you assign colors to each words at the very beginning? Is it random or follows some distribution?

  • @vicvic553
    @vicvic553 4 роки тому

    Could say more about the situation when articles have different number of words? In the previous video you mentioned Poisson distribution, but could you develop this topic, please?

  • @kennguyen9092
    @kennguyen9092 3 роки тому

    Hi, one question I have is how does Gibbs sampling eventually correctly label every word after many iteration? In the tidying up a room example, we were able to put things together because we know the association between shirt and pants, but for Gibbs sampling, we randomly assign every word a label. In other words, a more accurate example would be we have all these objects in the room and we randomly assign them labels, how do we sort them now? Thanks for your help!

  • @kirindagegayashan5801
    @kirindagegayashan5801 4 роки тому +1

    Can you do a video for LDA and Gibbs Sampling implementation using python to identify topics like in this video?

  • @alexbadin
    @alexbadin Рік тому

    at 12-48 when we try to allocate a color to the word 'ball', we then check what colours are there in this document and what are coolers of word 'ball' in other documents. But where do we take these colours? Considering we just started with all black words in all documents. Do we initialise them randomly and then they converge to more definitive colours?

  • @learn5081
    @learn5081 4 роки тому

    are those four documents in the example the documents generated by the "LDA machine" in the first video?

  • @awaisnawaz2791
    @awaisnawaz2791 Рік тому

    I have a question, while documents can be monochromatic, but same cannot be said for words as they can have different contexts. For example, 'war' can refer to either world war I or world war II... what can be done if we run into this situation?

  • @anondoggo
    @anondoggo 2 роки тому

    Will the algorithm converge to the true posterior when frequency stop changing significantly if we assign the colour with highest probability instead of sampling a colour from all the topic probabilities computed?

    • @anondoggo
      @anondoggo 2 роки тому

      Would this correspond to MAP vs Bayesian approaches to LDA training?

  • @compassionkicksass
    @compassionkicksass 2 роки тому

    Jeeeez this is an awesome tutorial! Amazing work, thank you very much.
    faith_in_humanity += 1

  • @peasant12345
    @peasant12345 7 місяців тому

    What is the state of art of choosing number of topics K? Cross validation AIC/BIC? or use non parametric method?

  • @improvementbeyond2974
    @improvementbeyond2974 3 роки тому

    Now with all these advanced neural architectures (transformers etc) does this technique is considered depreciated?🤔

  • @peasant12345
    @peasant12345 7 місяців тому

    It look like the more prevailing a word is, the more monochromatic that word is. Frequent words tend to reinforce themselves. However, each of them might have more different meanings than obsolete words do. How to get rid of such bias?

  • @emmarocheteau5788
    @emmarocheteau5788 3 роки тому

    How do you stop the words from all being the same colour? That would solve both learning objectives

  • @katateo328
    @katateo328 Рік тому

    just look into the dirichlet distribution, we can color right away all words in a document. Why we need to reference other documents to color each word in one document??

  • @harshtrivedi700
    @harshtrivedi700 3 роки тому

    what do you mean by parameters @18:02, its really confusing how you connect the math with the layman explaination, would be great if you can elaborate in the comments

  • @sharangkulkarni5718
    @sharangkulkarni5718 3 роки тому

    u r ~ human ur god!

  • @vicvic553
    @vicvic553 4 роки тому +1

    Actually, something seems, for me, not right here, but probably (or definitely) it is right, but I need some explanation. The very first coloring of the words is completely random (4:35) and based on this, the rest coloring happens.. does it makes sense? Since the very first coloring was random, it could have been completely wrong, so why is it a benchmark for the further coloring? Please, could you or anyone else explain it to me?

    • @jdp11
      @jdp11 3 роки тому

      I have this same question.