Stanford CS236: Deep Generative Models I 2023 I Lecture 10 - GANs

For more information about Stanford's Artificial Intelligence programs visit: stanford.io/ai
To follow along with the course, visit the course website:
deepgenerativemodels.github.io/
Stefano Ermon
Associate Professor of Computer Science, Stanford University
cs.stanford.edu/~ermon/
Learn more about the online course and how to enroll: online.stanford.edu/courses/c...
To view all online courses and programs offered by Stanford, visit: online.stanford.edu/

Пікірлер: 1

  • @CPTSMONSTER
    @CPTSMONSTER11 күн бұрын

    22:15? Reverse KL 35:40 Instead of optimizing over all possible functions T, optimize over an arbitrary set of neural network architectures. Similar flavor to importance sampling in VAEs. 36:45? Essentially the same as minimax training objective of GANs, p and q are expectations, T is the discriminator and optimization does not depend on likelihoods (p and q) 40:00? Supporting hyperplanes, convex hull, tangent? 41:35 KL divergences vs f divergences, doesn't depend on likelihoods, doesn't measure compression of data, flexible loss functions 48:40 Summary slide on divergences and training objectives 51:00? Example 1:04:00? Lipschitz constant 1:08:00? Earth mover distance, how is this different to the minimax training objective of GAN 1:09:30 Unlike f divergence, Earth mover distance doesn't give bounds 1:22:20 BiGAN, encoder is similar to VAE except deterministic, not trained by minimizing KL divergences like in the ELBO, trained by minimizing two sample test optimized by the discriminator