Stanford CS236: Deep Generative Models I 2023 I Lecture 10 - GANs

Поділитися
Вставка
  • Опубліковано 5 тра 2024
  • For more information about Stanford's Artificial Intelligence programs visit: stanford.io/ai
    To follow along with the course, visit the course website:
    deepgenerativemodels.github.io/
    Stefano Ermon
    Associate Professor of Computer Science, Stanford University
    cs.stanford.edu/~ermon/
    Learn more about the online course and how to enroll: online.stanford.edu/courses/c...
    To view all online courses and programs offered by Stanford, visit: online.stanford.edu/

КОМЕНТАРІ • 1

  • @CPTSMONSTER
    @CPTSMONSTER 7 днів тому

    22:15? Reverse KL
    35:40 Instead of optimizing over all possible functions T, optimize over an arbitrary set of neural network architectures. Similar flavor to importance sampling in VAEs.
    36:45? Essentially the same as minimax training objective of GANs, p and q are expectations, T is the discriminator and optimization does not depend on likelihoods (p and q)
    40:00? Supporting hyperplanes, convex hull, tangent?
    41:35 KL divergences vs f divergences, doesn't depend on likelihoods, doesn't measure compression of data, flexible loss functions
    48:40 Summary slide on divergences and training objectives
    51:00? Example
    1:04:00? Lipschitz constant
    1:08:00? Earth mover distance, how is this different to the minimax training objective of GAN
    1:09:30 Unlike f divergence, Earth mover distance doesn't give bounds
    1:22:20 BiGAN, encoder is similar to VAE except deterministic, not trained by minimizing KL divergences like in the ELBO, trained by minimizing two sample test optimized by the discriminator