MIT 6.S191 (2019): Deep Generative Modeling

Поділитися
Вставка
  • Опубліковано 2 січ 2025

КОМЕНТАРІ •

  • @kaankorkmaz8180
    @kaankorkmaz8180 5 років тому +16

    Awesome content Alexander and Ava, your efforts are much appreciated!

  • @kalebbruwer
    @kalebbruwer 5 років тому

    It's quite amazing what is possible in this field and just how fast it is developing. I mean, I try to stay on top of these things and even I get blown away from time to time. For example just how different, yet accurate the faces at 40:00 are

  • @arc6fire
    @arc6fire 3 роки тому

    5.56
    4th row, 4th column: shahid afridi

  • @dccn
    @dccn 5 років тому +2

    Thank you for sharing the class. These courses are great, including a lot of aspects in deep learning.

  • @bokkieyeung504
    @bokkieyeung504 5 років тому +1

    one question for slide - VAEs: Latent perturbation (around 29:00):
    if "smile" and "head pose" are independent to each other, when look at faces by rows, why the "extent of smiling" are different? it's supposed that only "head pose" varies by rows but "smiling" remains to be constant?

    • @bokkieyeung504
      @bokkieyeung504 5 років тому

      problems seem only to appear at 4th&5th rows

    • @AAmini
      @AAmini  5 років тому +4

      Great question, while an ideal encoding would promote independence among latent variables, this is not always the case with the variables that are actually learned via gradient descent. This leads to the problem you describe, where the variables are mostly "smile" and "head pose" but still share some overlap (or entanglement). In fact, a huge field of research goes into "disentangling" latent variables so they are independent of each other. Check out work on "Beta-VAEs" which try to impose constraints during learning to promote this type of behavior.

  • @linzihong9520
    @linzihong9520 5 років тому

    I am a bit confused about the continuity of the latent variable at around 28:00 (VAEs: Latent perturbation)... What exactly does this continuity have to do with the normal distribution? Won't latent perturbation still work if we use the traditional AEs? -- If the case is that it won't work, I could surmise it is due to the network not being penalized for "cheating" in traditional AEs? But then it would be the regularization rather than the random factor \epsilon that is responsible for this property of continuity? Also, this leads to another confusion: Previously, I intuitively understood the advantage of VAEs to be that the stochastic approach introduces randomness to "loosen" the network a bit to prevent it from cheating/doing something similar to overfitting, but then it seems the regularization is doing this job. So what exactly is the role of that randomness in the advantage of VAEs? Is it simply because to have such regularization as this we must introduce some stochastic elements into our model? Sorry for bombarding you with questions! And many thanks in advance!!

  • @parthvasoya3562
    @parthvasoya3562 4 роки тому +1

    superb content looking forward for more of these kind....

  • @arhamjain6686
    @arhamjain6686 4 роки тому

    in autoencoders and virtual encoders our loss function was just a "squared error" , so why are we using discriminator in GANs, instead of just a "squared error".
    Is it because in AEs we were going into compressed latent space from an original image to the predicted so it needed a little bit of tuning, while in GANs we are going from noise to predicted image so it need higher tuning ,please clarify this doubt>>>

  • @T4l0nITA
    @T4l0nITA 4 роки тому

    27:05 shouldn't the backdrop at phi on the reparametrized form be (partial z / partial phi) ?

    • @danielperaza1514
      @danielperaza1514 4 роки тому

      In backprop you are always looking for the gradient of the last function (in this case f) with respect to the weights (phi), i.e you want (partial f / partial phi). Using the chain rule: (partial f / partial phi) = (partial f / partial z) * (partial z / partial phi). In computational diagrams it is usually shown as in the slides.

  • @r00t67
    @r00t67 5 років тому +1

    Александр, Ава. Курс лекций очень хорош! Если бы использовали в лабах tf2.0 то цены бы ему не было. Но в любом случае отличная работа. From Russia with love ;)

  • @intuitivej9327
    @intuitivej9327 4 роки тому

    Learnt a lot!!
    Thank you from korea♡

  • @randalllionelkharkrang4047
    @randalllionelkharkrang4047 4 роки тому

    Since you've already given a prior, how are means and variances stochastic in nature? i dont get it. In the original VAEs

  • @hhumar987
    @hhumar987 4 роки тому

    can you also teach how to write code for it?

  • @DiegoSantos-yj3cg
    @DiegoSantos-yj3cg 5 років тому +4

    thanks a lot for sharing! great content

  • @mateusdeassissilva8009
    @mateusdeassissilva8009 5 років тому

    The only thing that i'd like to ask is: what's the difference between encoding and embedding?

    • @ogsconnect1312
      @ogsconnect1312 5 років тому

      To me it's synonymous. Only that embedding is mostly used in Natural Language context e.g. text, word, sentence, document

  • @ashishvishwakarma5362
    @ashishvishwakarma5362 5 років тому +4

    Thank you very much for sharing.

  • @zt0t0s
    @zt0t0s 4 роки тому

    Do you know some resources to do some deep fakes like the Obama introduction of the course?

  • @prasannautube1
    @prasannautube1 5 років тому +3

    Thank you guys for sharing this valuable content.

  • @ogsconnect1312
    @ogsconnect1312 5 років тому +1

    Excellent!

  • @johnarmstrong6714
    @johnarmstrong6714 5 років тому +1

    Awesome Stuff

  • @sanjaykrish8719
    @sanjaykrish8719 4 роки тому

    Brilliant

  • @jiansong7460
    @jiansong7460 5 років тому +1

    i love this vedio!!!!

  • @sapito169
    @sapito169 5 років тому +8

    i feel small
    i feel stupid

  • @maplex2656
    @maplex2656 4 роки тому

    Both faces are fake.The man has an asymmetric mustache and the womon's shirt is wired (right shoulder is not fully covered but the left shoulder is fully covered).

  • @leonhardeuler9839
    @leonhardeuler9839 5 років тому +2

    There is an audio problem with this series of lectures, do you all agree?

    • @parthvasoya3562
      @parthvasoya3562 5 років тому

      Sony is reasonably cheaper and better than the rest...(if you're planning...)