Adian Liusie
Adian Liusie
  • 6
  • 255 089
Intuitively Understanding the Shannon Entropy
Intuitively Understanding the Shannon Entropy
Переглядів: 92 783

Відео

Intuitively Understanding the Cross Entropy Loss
Переглядів 78 тис.3 роки тому
This video discusses the Cross Entropy Loss and provides an intuitive interpretation of the loss function through a simple classification set up. The video will draw the connections between the KL divergence and the cross entropy loss, and touch on some practical considerations. Twitter: AdianLiusie
Beginners Overview of Machine Learning and Artificial Intelligence
Переглядів 8303 роки тому
This is a recorded talk which I created for my old school, Dubai College. This video is an introduction for complete beginners into artificial intelligence and machine learning, gives the general idea of the field and an overview of the machine learning approach.
Understanding Neural Networks Training
Переглядів 7723 роки тому
This video gives an overview of general optimisation of neural networks and explains how training is done through loss minimisation. This video will provide a foundation to then take a deeper look at more complex practical algorithms like stochastic gradient descent and backwards propagation which will be covered in future videos.
Intuitively Understanding the KL Divergence
Переглядів 82 тис.3 роки тому
This video discusses the Kullback Leibler divergence and explains how it's a natural measure of distance between distributions. The video goes through a simple proof, which shows how with some basic maths, we can get under the KL divergence and intuitively understand what it's about.
Understanding deep neural networks
Переглядів 7843 роки тому
This video gives a basic introduction to neural networks and discusses what they are, how they work and ways to see the system in a matrix framework. This is the first video in a series which will ultimately build up to programming neural networks and running the back propagation algorithm from scratch in python.

КОМЕНТАРІ

  • @MissPiggyM976
    @MissPiggyM976 2 дні тому

    Very good!

  • @yingjiawan2514
    @yingjiawan2514 18 днів тому

    This is so well explained. thank you so much!!! Now I know how to understand KL divergence, cross entropy, logits, normalization, and softmax.

  • @chunheichau7947
    @chunheichau7947 19 днів тому

    I wish more professors can hit all the insights that you mentioned in the video.

  • @Sars78
    @Sars78 25 днів тому

    Well done, Adian. I just found out-though I'm not surprised at all, in the Shannon sense 🤓 -that you're doing a PhD at Cambridge. Congratulations! Best wishes for everything 🙂

  • @user-bi2jm1cn1h
    @user-bi2jm1cn1h Місяць тому

    How does the use of soft label distributions, instead of one-hot encoding hard labels, impact the choice of loss function in training models? Specifically, can cross-entropy loss still be effectively utilized, or should Kullback-Leibler (KL) divergence be preferred?

  • @sanjayadhith
    @sanjayadhith Місяць тому

    Useful video.

  • @HaykTarkhanyan
    @HaykTarkhanyan Місяць тому

    great video, thank you!

  • @AdeshBenipal
    @AdeshBenipal 2 місяці тому

    Nice video

  • @Micha-ku2hu
    @Micha-ku2hu 2 місяці тому

    What a great and simple explanation of the topic! Great work 👏

  • @xyzct
    @xyzct 2 місяці тому

    Excellent. Short and sweet.

  • @genkidama7385
    @genkidama7385 2 місяці тому

    distirbution

  • @avatar00001
    @avatar00001 3 місяці тому

    thank you codexchan

  • @mathy642
    @mathy642 3 місяці тому

    Thank you for the best explanation

  • @franklyvulgar1
    @franklyvulgar1 3 місяці тому

    this is a great explanation thank you !

  • @debasishraychawdhuri
    @debasishraychawdhuri 4 місяці тому

    It does not explain the most important part - how the formula for non-uniform distribution came about

  • @shahriarrahman8425
    @shahriarrahman8425 4 місяці тому

    Great explanation. Thank you so much!

  • @ian-haggerty
    @ian-haggerty 4 місяці тому

    Adian. I know you're probably super busy doing PhD things, but come back & make some more videos! You're a gifted orator.

  • @ian-haggerty
    @ian-haggerty 4 місяці тому

    Best explanation on the interwebs!

  • @madarahuchiha1133
    @madarahuchiha1133 4 місяці тому

    what is true class distribution?

    • @elenagolovach384
      @elenagolovach384 2 місяці тому

      the frequency of occurrence of a particular class depends on the characteristics of the objects

  • @ian-haggerty
    @ian-haggerty 4 місяці тому

    So a Kale Divergence of zero means identical distributions? What do the || lines mean?

  • @ian-haggerty
    @ian-haggerty 4 місяці тому

    <3 this. 👌

  • @user-ut4zh3pw7l
    @user-ut4zh3pw7l 4 місяці тому

    wowowowo

  • @adityakulkarni5577
    @adityakulkarni5577 4 місяці тому

    Perfectly explained in 5 minutes. Wow.

  • @LiHongxuan-ee7qs
    @LiHongxuan-ee7qs 5 місяців тому

    So clear explanation! Thanks!

  • @charleswilliams8368
    @charleswilliams8368 5 місяців тому

    Three bits to tell the guy on the other side of the wall what happened, and it suddenly made sense. Thanks.

  • @EricPham-gr8pg
    @EricPham-gr8pg 5 місяців тому

    This is an incomplete hypothesis because on p(x) and ln(p(x)) come without proof is grossly insulting and it ignored the basic principle of conservation of energy meaning if pressure goes up then volume also goes up to release presured to go back to normal and that is the frustrating truth of life. The beast never dies it only change its phases or faces

  • @SunilKumarSamji
    @SunilKumarSamji 5 місяців тому

    Excellent video. Can someone help me understand why is it called Divergence in the first place? Why are we taking 1/N power to normalise it to sample space, I did not understand the logic behind this.

  • @ajitzote6103
    @ajitzote6103 5 місяців тому

    not really a great explaination, so many terms were thrown in. that's not a good way to explain something.

  • @user-cf2yo5qf3h
    @user-cf2yo5qf3h 6 місяців тому

    Thankkkk youuuuu.

  • @brianlee4966
    @brianlee4966 6 місяців тому

    Thank you so much for this video and clear explanation!

  • @jiwoni523
    @jiwoni523 6 місяців тому

    make more videos please , you are awesome

  • @ananthakrishnank3208
    @ananthakrishnank3208 6 місяців тому

    Excellent expositions on KL divergence and Cross Entropy loss within 15 mins! Really intuitive. Thanks for sharing.

  • @cuongnguyenuc1776
    @cuongnguyenuc1776 7 місяців тому

    Great video! Can you make a video about soft actor critic?

  • @sirelegant2002
    @sirelegant2002 7 місяців тому

    Thank you!

  • @zizo-ve8ib
    @zizo-ve8ib 7 місяців тому

    Bro really explained it in less than 10 mins when my professors don't bother even if it could be done in 5 secs, true master piece thus video keep it up man 🔥🔥🔥

  • @wynandwinterbach455
    @wynandwinterbach455 7 місяців тому

    I'm just rewatching this video to freshen up my deep learning fundamentals. Super clear video, thank you so much!

  • @bodwiser100
    @bodwiser100 7 місяців тому

    I appreciate your effort, but the video is quite confusing. For example, in the example about 8 football teams, you explain why 3 bits are required by flat out stating as a starting premise that 3 bits are required! It's a circular argument.

  • @nicholaselliott2484
    @nicholaselliott2484 8 місяців тому

    Dude, took information theory from a rigorously academic and formal professor. I'm a little slow and under the pressure of getting assignments done, couldn't always see the forest for the trees. Just the sentence "how much information, on average, would we need to encode an outcome from a distribution" just summed up the whole motivation and intuition. Thanks!

  • @nyx8017
    @nyx8017 8 місяців тому

    god this is an incredible video thank you so much

  • @aj7_gauss
    @aj7_gauss 9 місяців тому

    can someone explain the triplets part

  • @kukuster
    @kukuster 9 місяців тому

    Thanks for the explanation!! One thing is, formulas were confusing with how you denoted *q1* & *q2* for probabilities for coin 2, instead of *p2* & *q2=1-p2*

  • @mormonteg4073
    @mormonteg4073 9 місяців тому

    Thank you a lot

  • @akidnag
    @akidnag 9 місяців тому

    Only that is not a distance ('cause is not symmetric), but a pseudo distance. Great video!

  • @diy_mo
    @diy_mo 9 місяців тому

    I expected something else, but it's also ok.

  • @karthikeyans3
    @karthikeyans3 9 місяців тому

    Great video. Thanks for sharing. Really intuitive.

  • @maxlehtinen4189
    @maxlehtinen4189 9 місяців тому

    for everyone trying to understand this concept even more thoroughly, towardsdatascience's article "The intuition behind Shannon’s Entropy" is amazing. it gives added insight on why information is the reciprocal of probability

  • @bluejays440
    @bluejays440 10 місяців тому

    Please make more videos this is literally the only time I've ever seen entropy be explained in a way that makes sense

  • @derickd6150
    @derickd6150 10 місяців тому

    Great video!

  • @adamtaylor2142
    @adamtaylor2142 10 місяців тому

    Great content! Thank you.

  • @kvnptl4400
    @kvnptl4400 10 місяців тому

    This one I would say is a very nice explanation of Cross Entropy Loss.