Reinforcement Learning 10: Classic Games Case Study

Поділитися
Вставка
  • Опубліковано 15 чер 2024
  • David Silver, Research Scientist, discusses classic games as part of the Advanced Deep Learning & Reinforcement Learning Lectures.
  • Наука та технологія

КОМЕНТАРІ • 33

  • @LuisYu
    @LuisYu 5 років тому +10

    amazing high quality lectures. especially enjoyed attention, memory, alpha zero talks.

  • @stevecarson7031
    @stevecarson7031 2 роки тому

    Thankyou so much for this series of lectures!

  • @samagrasharma7755
    @samagrasharma7755 5 років тому +1

    Two lectures (CNN and RNN) are missing from this series. Can anyone tell if they are available online?

  • @Kingstanding23
    @Kingstanding23 4 роки тому +3

    A Nash equilibrium sounds like what happens on roads where traffic evens itself out amongst all the roads towards some destination. When a new road is built, nothing really changes because the traffic just redistributes itself to an new equilibrium.

  • @TheGreatBlackBird
    @TheGreatBlackBird 3 роки тому +1

    Shouldn't there also be a reward present in TD error at 42:30 and 50:25 ?
    edit: ok, it's explained a bit more in the 2015 lecure that this version assumes no intermediate reward

  • @helinw
    @helinw 3 роки тому +1

    Did David do another RL course in 2018? Or just one lecture?

    • @ShortVine
      @ShortVine 3 роки тому +1

      i was thinking the same & searched a lot, but i think he did just one lecture in 2018

  • @alexanderyau6347
    @alexanderyau6347 5 років тому +4

    I can comment now. See you again David.

  • @johangras3522
    @johangras3522 5 років тому +6

    It is possible to access to the course slides ?

    • @TuhinChattopadhyay
      @TuhinChattopadhyay 3 роки тому

      @@Sigmav0 Link not working

    • @Sigmav0
      @Sigmav0 3 роки тому +2

      @@TuhinChattopadhyay The slide has been moved to www.davidsilver.uk/wp-content/uploads/2020/03/games.pdf
      Hope this helps !

    • @TuhinChattopadhyay
      @TuhinChattopadhyay 3 роки тому

      @@Sigmav0 Got it... many thanks

    • @Sigmav0
      @Sigmav0 3 роки тому +1

      @@TuhinChattopadhyay No problem ! 👍

    • @dojutsu6861
      @dojutsu6861 3 роки тому

      @@Sigmav0 these slides are from an older UCLxDeepMind lecture series lead primarily by David Silver. They do not include content on the newer AlphaZero models. Do you by any chance know if these updated slides are available online

  • @jakubbielan4784
    @jakubbielan4784 5 років тому +3

    Anyone know what was the exact hardware used to train Alpha Go Zero?

  • @domino14
    @domino14 2 роки тому +2

    The level of computer play in Scrabble is not superhuman. Quackle beats Maven, and the best humans can 50-50 Quackle in a long series.

  • @Dina_tankar_mina_ord
    @Dina_tankar_mina_ord 5 років тому

    I would love to see how deepmind would build a city on its own in Cityskyline. See how its optimization would create the best and most efficient layout in real time. Maybe we could learn alot from that.

  • @mohammadkhan5430
    @mohammadkhan5430 3 роки тому +2

    I love him, how sad the room is empty

  • @KayzeeFPS
    @KayzeeFPS 4 роки тому +1

    Here's a link to the same video but with slides visible ua-cam.com/video/N1LKLc6ufGY/v-deo.html

  • @yidingyu2739
    @yidingyu2739 5 років тому +1

    Why so many empty seats?

    • @yoloswaggins2161
      @yoloswaggins2161 5 років тому +5

      This stuff not on the exam

    • @matveyshishov
      @matveyshishov 5 років тому +2

      The number of people is lower with later lectures for some reason.

    • @markdonald4538
      @markdonald4538 5 років тому

      @@matveyshishov stupid ppl

  • @julioandresgomez3201
    @julioandresgomez3201 5 років тому

    Despite the success of A 0 nets in several games, I feel that is better starting point playing (random number) games with humans. Only then, when it has grasped some basic basics (by itself, not forcibly inserted by hand), let it play against itself. This way it could accomplish in thousands of self-play games what from scratch it´d take millions of self-play games, due to the total randomness and clueless of the first games. It´s not the absolute zero approach, but it has no "artificial" parameters handcrafted either. It learns from its own games all the way.

    • @Avandale0
      @Avandale0 4 роки тому +1

      Playing with humans takes considerably more time than running simulations - so actually, playing millions of games by itself is still faster than playing 100 games from playing humans. Knowing that a game of go takes around 1h, you'd have finished 3 games with a human in the time that it took AlphaZero to reach human level play.
      Same for chess, when you realise it took Alpha Zero 4 hours reach a level higher than Stockfish...
      It should be clear from these examples that one of the particularities of AlphaZero is the speed at which it learns. Playing humans here both defeats the purpose of self-learning and actually wastes time.

  • @omarcusmafait7202
    @omarcusmafait7202 5 років тому

    why does nobody take notes?

    • @yoloswaggins2161
      @yoloswaggins2161 5 років тому +1

      Not on the exam

    • @Sigmav0
      @Sigmav0 5 років тому +4

      @William Davis Sure... In primary school...

    • @vijayabhaskarj3095
      @vijayabhaskarj3095 5 років тому +4

      because slides are available online and lectures are available online, I would listen carefully first in the class