Diffusion Models Beat GANs on Image Synthesis | ML Coding Series | Part 2

Поділитися
Вставка
  • Опубліковано 17 гру 2024

КОМЕНТАРІ • 26

  • @TheAIEpiphany
    @TheAIEpiphany  2 роки тому +5

    Continuing on with diffusion models! ;) Let me know how you like this format - is it better or worse compared to classic paper overviews? What are the pros and cons?

    • @prabhavkaula9697
      @prabhavkaula9697 2 роки тому +1

      Better but your coding sessions would add to it

    • @TheAIEpiphany
      @TheAIEpiphany  2 роки тому

      @@prabhavkaula9697 meaning coding from scratch?

    • @varunsai9736
      @varunsai9736 2 роки тому +3

      these are amazing they serve as a guide for some one who is looking to build from scratch and also to get better technical understanding and insights which we don't get from the classical paper some times... there aren't many con's though as your explanation is good... coding from scratch might bring good feel as we are walking along with you and views but doing it for these big projects can be a lot sometimes, .. just a view from a lazy person , we don't know until we try and get that videos feed back .. anyways thanks for the effort man.

    • @prabhavkaula9697
      @prabhavkaula9697 2 роки тому

      @@TheAIEpiphany yes please
      this would help in catching up the bottom up approach and maybe finding some new novel solutions :)

    • @masternobody1896
      @masternobody1896 2 роки тому

      @@TheAIEpiphany you are a legend

  • @InturnetHaetMachine
    @InturnetHaetMachine 2 роки тому +6

    Oh man please continue the coding series. Like I've said before, digging through codes is often where I've found out intricacies of the paper that I hadn't realized while just reading. For a non-CS background amateur person like me, going through codes isn't the easiest thing, so a guide like this is so helpful. Thanks for producing these, its really appreciated!

    • @TheAIEpiphany
      @TheAIEpiphany  2 роки тому +2

      Happy to hear that, that's the idea, thanks!

  • @vikramsandu6054
    @vikramsandu6054 7 місяців тому

    Legend Right here, Ladies and gentlemen. Explained Beautifully. Thanks a lot!!

  • @billyli7053
    @billyli7053 2 роки тому

    Oh, man. You are so much ahead of 80% of the ML researchers. Watching your stuff saves me so much time to get a hang of the codebase!

  • @Trumanshoww
    @Trumanshoww Рік тому

    I once couldn't understand how the sampling actually works, but now I know thanks to your video.
    Always thank U -!!

  • @seerproductions9141
    @seerproductions9141 2 роки тому +2

    Really enjoyed this video. Keep up the great work.

  • @prabhavkaula9697
    @prabhavkaula9697 2 роки тому +2

    Thank you for the videos on diffusion models. It would be great if you could also do a coding session for single and multiple gpus and known and custom datasets

  • @피카라이언
    @피카라이언 2 роки тому

    coding from scratch would really help as well :) !! Thank you for the awesome videos always

  • @tonytu3995
    @tonytu3995 Рік тому

    I wonder if there's any specific requirement for the classifier architecture. In the appendix of the paper it is mentioned that “Our classifier architecture is simply the downsampling trunk of the UNet model with an attention pool [49] at the 8x8 layer to produce the final output.” I wonder if there is a reason that the classifier is just a simplified version of the denoising network - UNet? If you change the classifier with another architecture for image classification, would you be able to guide the generation properly? In other words would the gradients from a different classifier architecture be meaningful in guiding the sampling process?

  • @97mayo97
    @97mayo97 2 роки тому

    Great Video! so nicely explained, thank you!
    Do you know what kind of accuracy values to expect when training the classifier? and for how long to train?
    I am trying to learn a custom dataset and as soon as I add the "--noised True"-Flag, my model does not learn at all (without noise it achieves 95 % accuracy easily). Loss is also not decreasing...😕

  • @likeapple1929
    @likeapple1929 Рік тому

    Great video! This really helps me a lot! Just one question about how they embed the class embedding into the unet model, it seems that you did not mention it in the video? I look throguth the code and I think the class is embedded after time is embedded, such that: emb = emb + self.label_emb(y) and the class informatin is so called lebel_emb. Do I get it correctly?

  • @junghunkim8467
    @junghunkim8467 Рік тому

    Hi, this is awesome video. Thank you.
    I have a question,
    the OS you tested is linux..? or window?

  • @ChowzorLoves2Tap
    @ChowzorLoves2Tap 2 роки тому

    You def read clean code.

  • @convolutionalnn2582
    @convolutionalnn2582 2 роки тому +2

    Do Computer vision require 3D geometry? Any recommendations on roadmap of Computer Vision?

    • @TheAIEpiphany
      @TheAIEpiphany  2 роки тому

      Depends which subfield of CV you care about. Definitely not a prerequisite for many topics. E2e learning is taking care of the geometry for you :))

    • @convolutionalnn2582
      @convolutionalnn2582 2 роки тому

      @@TheAIEpiphany Do you think I need geometry to become Research scientist in computer vision ? Any resources?

  • @rewixx69420
    @rewixx69420 2 роки тому

    plese cover k-diffusion