DeepBean
DeepBean
  • 21
  • 228 335
Understanding GANs (Generative Adversarial Networks) | Deep Learning
GANs use an elegant adversarial learning framework to generate high quality samples of everything from images to audio. Here, we explore the theoretical underpinnings, as well as some practical problems that can plague training, such as non-convergence and mode collapse.
Timestamps
--------------------
00:00 Introduction
01:28 Generative modelling
04:46 The GAN approach
07:37 Loss function
12:14 Game theory perspective
13:18 Optimal discriminator
15:33 Optimal generator
17:26 Training dynamics
19:45 Optimal discriminator problem
21:39 Training steps
22:13 Non-convergence
23:39 Mode collapse
Links
--------
- Original GAN paper (arxiv.org/abs/1406.2661)
- Analysis of vanishing/unstable gradients (arxiv.org/abs/1701.04862)
- Analysis of mode collapse (arxiv.org/abs/1606.03498)
- Wasserstein GAN paper (arxiv.org/abs/1701.07875)
- Keras CGAN tutorial (keras.io/examples/generative/conditional_gan/)
- PyTorch DCGAN tutorial (pytorch.org/tutorials/beginner/dcgan_faces_tutorial.html)
Переглядів: 1 043

Відео

Vanishing Gradients: Why Training RNNs is Hard
Переглядів 5893 місяці тому
Here, we run down how RNNs are trained via backpropagation through time, and see how this algorithm is plagued by the problems of vanishing and exploding gradients. We present an intuitive and mathematical picture by flying through the relevant calculus and linear algebra (so feel free to pause at certain bits!) Timestamps 00:00 Introduction 00:46 RNN refresher 03:42 Gradient calculation of W 0...
Vector-Quantized Variational Autoencoders (VQ-VAEs) | Deep Learning
Переглядів 3,5 тис.3 місяці тому
The Vector-Quantized Variational Autoencoder (VQ-VAE) forms discrete latent representations, by mapping encoding vectors to a limited size codebook. But, how does it do this, and why would we want to do it anyway? Link to my video on VAEs: ua-cam.com/video/HBYQvKlaE0A/v-deo.html Timestamps 00:00 Introduction 01:09 VAE refresher 02:42 Quantization 04:46 Posterior 06:09 Prior 07:06 Learned prior ...
Disentanglement with beta-VAEs | Deep Learning
Переглядів 8733 місяці тому
Link to my VAE video for a refresher: ua-cam.com/video/HBYQvKlaE0A/v-deo.html In this video, we explore how and why modifying the VAE loss function enables us to achieve disentanglement in the latent space, with different latent variables corresponding to different semantic features of the data. We take a look at the original beta-VAE formulation, as well as controlled capacity increase, and th...
Convolutional Neural Networks (CNNs) | Deep Learning
Переглядів 2,3 тис.6 місяців тому
CNNs are a go-to deep learning architecture for many computer vision tasks, from image classification to object detection and more. Here, we take a look at the basics, and see how they use biologically-inspired hierarchical feature extraction to do what they do. Timestamps Introduction 00:00 Kernel convolutions 00:41 Common kernels 02:30 Why flipping? 03:30 Convolution as feature extraction 04:...
Understanding Variational Autoencoders (VAEs) | Deep Learning
Переглядів 14 тис.7 місяців тому
Here we delve into the core concepts behind the Variational Autoencoder (VAE), a widely used representation learning technique that uncovers the hidden factors of variation throughout a dataset. Timestamps Introduction 0:00 Latent variables 01:53 Intractability of the marginal likelihood 05:08 Bayes' rule 06:35 Variational inference 09:01 KL divergence and ELBO 10:14 ELBO via Jensen's inequalit...
The Geiger-Marsden Experiments | Nuclear Physics
Переглядів 1,5 тис.Рік тому
In 1908-13, nuclear physics was born as Hans Geiger and Ernest Marsden embarked on the experiments that would discover the atomic nucleus and revolutionise our understanding of atomic structure. Here we explore why and how they carried out the famous gold-leaf experiment, as well as how Ernest Rutherford arrived at his startling conclusions. CHAPTERS Introduction 00:00 Alpha Particles 00:20 The...
Dijkstra's Algorithm: Finding the Shortest Path
Переглядів 928Рік тому
Dijkstra's algorithm is a neat way of finding the minimum-cost path between any two nodes in a graph. Here we see briefly how can use it to optimize our path through a graph, and also explore why it performs as well as it does. Feel free to like, comment and subscribe if you appreciate what I do!
Einstein's Ladder Paradox; Simply Explained
Переглядів 13 тис.Рік тому
In special relativity, the ladder paradox (or, "barn-pole" paradox) occurs due to they symmetry of length contraction. Here we explore how this apparent paradox can be solved using the relativity of simultaneity. If you're interested in more special relativity content, check out the series below! SPECIAL RELATIVITY SERIES I. The Michelson-Morley Experiment (ua-cam.com/video/DFQtVFEp_3E/v-deo.ht...
Solving Einstein's Twin Paradox
Переглядів 3,4 тис.Рік тому
Many solutions have been proposed to Einstein's twin paradox, but many of them miss the vital reasons why the Earth twin is correct and the Spaceship twin is wrong. Here we condense the solution of the twin paradox to its essentials, and also discuss why applying general relativity to the problem is unnecessary. CHAPTERS What is the Twin Paradox? 00:00 Time Dilation 00:43 The "Paradox" 01:17 Th...
Relativistic Velocity Addition | Special Relativity
Переглядів 6 тис.Рік тому
Here, we briefly derive the equation for relativistic velocity addition, using only the Lorentz transformation equations we derived back in Part 3. Please like, subscribe and leave a comment if you appreciate what I do! SPECIAL RELATIVITY SERIES I. The Michelson-Morley Experiment (ua-cam.com/video/DFQtVFEp_3E/v-deo.html) II. Time Dilation and Length Contraction (ua-cam.com/video/bArTzG3Mkmk/v-d...
What is Spacetime? | Special Relativity
Переглядів 2,7 тис.Рік тому
In this fifth video, we explore how space and time can be treated as one interrelated entity, and how the Lorentz transformations can be given a geometric interpretation. We also explore how paradoxes are avoided by the preservation of causality. Please like, comment and subscribe if you appreciate what I do! SPECIAL RELATIVITY SERIES I. The Michelson-Morley Experiment (ua-cam.com/video/DFQtVFE...
Deriving the General Lorentz Transformation | Special Relativity
Переглядів 8 тис.Рік тому
In this fourth video of the Special Relativity series, we derive the general (matrix) form of the Lorentz transformations for an arbitrary boost velocity in 3D space. Please like, comment and subscribe if you appreciate what I do! SPECIAL RELATIVITY SERIES I. The Michelson-Morley Experiment (ua-cam.com/video/DFQtVFEp_3E/v-deo.html) II. Time Dilation and Length Contraction (ua-cam.com/video/bArT...
Deriving the Lorentz Transformations | Special Relativity
Переглядів 23 тис.Рік тому
In this third video of the Special Relativity series, we derive the Lorentz transformations, which map events in one reference frame to another reference frame that moves at a constant relative velocity. We also demonstrate how these transformations can be used to derive the phenomena of time dilation and length contraction that we explored more informally in the last video. Please like, commen...
Time Dilation and Length Contraction | Special Relativity
Переглядів 6 тис.Рік тому
Here we explore how Einstein's postulates imply that moving clocks tick slower and moving trains become shorter. Please like, comment and subscribe if you appreciate what I do! SPECIAL RELATIVITY SERIES I. The Michelson-Morley Experiment (ua-cam.com/video/DFQtVFEp_3E/v-deo.html) II. Time Dilation and Length Contraction III. Deriving the Lorentz Transformations (ua-cam.com/video/FvqutkaPmas/v-de...
The Michelson-Morley Experiment | Special Relativity
Переглядів 32 тис.Рік тому
The Michelson-Morley Experiment | Special Relativity
The Physics of Nuclear Weapons
Переглядів 1,9 тис.Рік тому
The Physics of Nuclear Weapons
How YOLO Object Detection Works
Переглядів 46 тис.Рік тому
How YOLO Object Detection Works
Optimization for Deep Learning (Momentum, RMSprop, AdaGrad, Adam)
Переглядів 55 тис.Рік тому
Optimization for Deep Learning (Momentum, RMSprop, AdaGrad, Adam)
Backpropagation: How Neural Networks Learn
Переглядів 3,8 тис.Рік тому
Backpropagation: How Neural Networks Learn
Transformers, Simply Explained | Deep Learning
Переглядів 4,6 тис.Рік тому
Transformers, Simply Explained | Deep Learning

КОМЕНТАРІ

  • @arashakbari6986
    @arashakbari6986 3 дні тому

    perfect

  • @na50r24
    @na50r24 4 дні тому

    Can w be considered as vector that represents all adjustable parameters? I.e., not just weights of one linear transformation matrix from input to hidden layer but all of them + bias values. So when you compute gradient of L with respect to w, you compute a vector for which each entry is a the partial derivative of L with respect to w_i?

    • @deepbean
      @deepbean 3 дні тому

      Yup, that's correct!

  • @adamchelchowski
    @adamchelchowski 6 днів тому

    Pleasure to watch

  • @rishidixit7939
    @rishidixit7939 7 днів тому

    Nice, will revisit this because the maths is overwhelming right now

  • @rishidixit7939
    @rishidixit7939 7 днів тому

    From where to study Probability for Deep Learning ? Specifically the one used in here? I have studied probability but not this much. If anyone can give resources it will be very helpful

    • @deepbean
      @deepbean 2 дні тому

      I'm not sure of the best resources for deep learning in general, but for theory related to VAEs I found this helpful ermongroup.github.io/cs228-notes/

  • @TheKenigham
    @TheKenigham 10 днів тому

    Your videos are amazing to understand the subject of relativity! Unfortunately, I have no background in physics and mathematics so I often struggle to understand a few things. One question I have from the video is: I can grasp that the Michelson e Morley’s experiment proved that there is no special medium in relation to which light would be propagating. But, how does it also prove that light’s speed is constant in relation to all other things including that which emitted it? Did the experiment also move the light source in relation to the detector? If the frequency of the interference remained the same even when the emitter and the detector were in movement with relation to each other, I think it would be a more straightforward way to guarantee the the fact that light’s speed remain constant regardless of relative velocity. Sorry if that’s a stupid question, but I’d really like to understand the subject.

  • @AbhayShuklaSilpara
    @AbhayShuklaSilpara 12 днів тому

    At 8:46 why is joint probability tractable? Why are others not tractable?

    • @deepbean
      @deepbean 11 днів тому

      The joint probability is traceable under our model because it's easy to estimate the probability of say, a particular z given a particular x; however the marginal probability of x requires integration over all z, which makes it intractable.

  • @ThomasHaron
    @ThomasHaron 15 днів тому

    The video is woow

  • @klausziegler60
    @klausziegler60 24 дні тому

    Excellent analysis

  • @3B1bIQ
    @3B1bIQ Місяць тому

    My brother, please, can you create a course to learn the manim library for beginners? ☺️🤍🤍

    • @deepbean
      @deepbean 2 дні тому

      I'm not too familiar with it, but 3Blue1Brown has a tutorial!

  • @dmitriizhilenkov2673
    @dmitriizhilenkov2673 Місяць тому

    One question. Why max{ \int_x {f(x) dx} } is the same thing as max{ f(x) }?

    • @deepbean
      @deepbean Місяць тому

      Good question! This is legitimate because by finding our optimal D(x), we are directly maximizing the integrand at every point x. Since our integrand can never be higher for any other D(x) at a specific point x; then if we find this maximum for every x, then we are guaranteed to maximize the integral as a whole.

  • @nguyenvothanh9230
    @nguyenvothanh9230 Місяць тому

    Always expect the earliest video from your channel

  • @nguyenvothanh9230
    @nguyenvothanh9230 Місяць тому

    Well, i can’t miss your video

  • @rafa_br34
    @rafa_br34 Місяць тому

    Well done! Very thoroughly explained.

  • @RuizSerra
    @RuizSerra Місяць тому

    Fantastic explanation. Small erratum: the variable 𝜖 is epsilon, not eta (η) ua-cam.com/video/HBYQvKlaE0A/v-deo.htmlsi=k6EBUeCbMUl4JYWw&t=970

    • @deepbean
      @deepbean Місяць тому

      Ah, that's right!

  • @GreenyX1
    @GreenyX1 Місяць тому

    You know neutrinos are 65 billion solar neutrinos, per second per square centimeter and they didn't detect that interaction either. The Ether is likely even smaller. Discounting the existence of an Aether using stone knives and bear skins. What a joke. This is why you don't discount anything, and try to evolve your experiments.

  • @BenjaminEvans316
    @BenjaminEvans316 Місяць тому

    Great video. A good balance of professional, modern and unique style with in-depth technical description.

  • @HaiderAli-l5z1c
    @HaiderAli-l5z1c Місяць тому

    confusing there must be a simpler derivation

  • @nabinbk1065
    @nabinbk1065 Місяць тому

    thanks

  • @gilrosario7224
    @gilrosario7224 Місяць тому

    I’m here because of Lord Jamar. His interview on the Godfrey Comedy channel was very interesting….

  • @PepysFlora-t8p
    @PepysFlora-t8p Місяць тому

    Williams John Young Gary Clark Jessica

  • @Kir-f4j
    @Kir-f4j Місяць тому

    Очень классные ролики, очень интересно и понятно смотреть Только переводить не легко бывает, привет из России ❤

  • @fzigunov
    @fzigunov Місяць тому

    You're the best explanation out there in my opinion. I appreciate you!!

  • @lambda4931
    @lambda4931 Місяць тому

    Why wouldn’t going against the aether be the opposite of going with it. They should cancel out.

  • @raihanpahlevi6870
    @raihanpahlevi6870 Місяць тому

    predicted Ci is calculated with IoU if the cell have object, then how to calculate predicted Ci if the cell doesnt have object?

  • @BradleyJohnson-t2e
    @BradleyJohnson-t2e 2 місяці тому

    Ortiz Mountains

  • @oinotnarasec
    @oinotnarasec 2 місяці тому

    Beautiful video. Thank you

  • @everythingisalllies2141
    @everythingisalllies2141 2 місяці тому

    Your error is at ua-cam.com/video/FvqutkaPmas/v-deo.html If the spherical wave is centred and expanding from K origin, it cant also be expanding from a different center at location K prime's origin which is also in motion. Your whole explanation has failed at this point.

  • @bradleymorris161
    @bradleymorris161 2 місяці тому

    Thank you so much for this, really cleared up how VAEs work

  • @nielsniels5008
    @nielsniels5008 2 місяці тому

    Thank you so much for these videos

  • @everythingisalllies2141
    @everythingisalllies2141 2 місяці тому

    the is all BS. For jack, his ladder doesn't shrink, because he can say the barn is doing the moving, so the barn is not as big as it was before. So the ladder certainly cant fit. Now that we know it doesn't fit for two reasons, its not going to fit if you invent a third option. Your error is in your stupid simultaneity example. make up your mind, it cant be two things that light does. The centre of an expanding sphere of light cant have two different origins, one not moving and the other moving. That is where you make the error of simple logic and simple physics.

  • @fullerholiday2872
    @fullerholiday2872 2 місяці тому

    Martin Jessica Moore Carol Taylor Dorothy

  • @KwangrokRyoo
    @KwangrokRyoo 2 місяці тому

    this is amazing 🤩

  • @Chachaboyz
    @Chachaboyz 2 місяці тому

    By far the best resource I've found on VAEs, after _lots_ of reading and video watching. This puts it all together intelligently and clearly. Thank you!!

  • @chhotiverma5019
    @chhotiverma5019 2 місяці тому

    Wow wonderful explanation ❤️ ⭐⭐⭐⭐⭐

  • @jeffburton1326
    @jeffburton1326 2 місяці тому

    There is a difference between a paradox and BS. This is BS ........ not a paradox.

  • @truthbetold818
    @truthbetold818 2 місяці тому

    I think the Aether does exist

  • @qualquan
    @qualquan 2 місяці тому

    confusing

  • @romansate2854
    @romansate2854 2 місяці тому

    Zero point energy (bare vacuum) needs to be included in this.

  • @HojjatMonzavi
    @HojjatMonzavi 3 місяці тому

    As a junior AI developer, this was the best toturial of Adam and Other optimizers I've ever seen. Simply explained but not too simply to be a useless overview Thanks

  • @ligezhang4735
    @ligezhang4735 3 місяці тому

    This video is amazing! I like how you get "reparametrization trick" into the picture, that you first calculate the gradient seperately to show the potential issue. Super clear!

  • @ayahouassim4095
    @ayahouassim4095 3 місяці тому

    Earth does not move!

  • @sokrozayeng7691
    @sokrozayeng7691 3 місяці тому

    Great Explaination! Thank you.

  • @mehdizahedi2810
    @mehdizahedi2810 3 місяці тому

    awesome explanation, contains a lot of information which are missing in the paper. Thank you

  • @ancientseed2607
    @ancientseed2607 3 місяці тому

    Yeah, I don’t know about that.

  • @rafa_br34
    @rafa_br34 3 місяці тому

    Very interesting indeed, I feel kinda stupid cuz I barely understand the math tho lol.

    • @deepbean
      @deepbean 3 місяці тому

      I appreciate the comment! Yeah, the theory of VAEs can get a bit heavy at times... though I hope some points are conveyed well enough without the need for equations

  • @tahmidislamtasen1602
    @tahmidislamtasen1602 3 місяці тому

    Finally thats the counterargument that came to my mind

  • @deepbean
    @deepbean 3 місяці тому

    Note on 16:38. The classifier doesn't directly classify the ground-truth factor corresponding to each latent variable; it classifies the factor that was kept constant in each input data pair. However, the structure of the problem, and the limitation of a linear classifier, ensures that it can only do this by mapping latent variables to ground-truth factors.

    • @mohamedfarag4882
      @mohamedfarag4882 27 днів тому

      But this means we need to know the concepts that creates the chair before hand

    • @mohamedfarag4882
      @mohamedfarag4882 27 днів тому

      Which in return means we may overlook some important concepts

  • @isiisorisiaint
    @isiisorisiaint 3 місяці тому

    why don't i see such detailed explanation videos as PDFs? do you guys think you'll get rich from a couple hundred views on youtube??? make a PDF (including ALL your explanations during the video!). having only an online version, plus having to click stop and juggle with the video position slider and never having a continuous presentation in front of my eyes is just plain nonsense for this kind of material. basically, it seems (from what i've been able to follow) to be a very well done presentation in terms of content , but totally useless in the form the content is delivered PS: never heard this so clearly stated before: "a VAE maps a point in X to a distribution in Z by pushing the distributions towards the prior of Z which is a unit guasian, which encourages the distributions to overlap and fill all the space of the prior of Z". just brilliant.

  • @stephen7774
    @stephen7774 3 місяці тому

    The sheer arrogance of Michelson and Morley in ignoring the obvious fact that gravity is ether flow totally bewilders me to this day.

    • @jojojo9240
      @jojojo9240 Місяць тому

      I think you are a little confused. These experiments should be sufficient.