Google DeepMind's Deep Q-learning playing Atari Breakout!

Поділитися
Вставка
  • Опубліковано 6 бер 2015
  • Google DeepMind created an artificial intelligence program using deep reinforcement learning that plays Atari games and improves itself to a superhuman level. It is capable of playing many Atari games and uses a combination of deep artificial neural networks and reinforcement learning. After presenting their initial results with the algorithm, Google almost immediately acquired the company for several hundred million dollars, hence the name Google DeepMind. Please enjoy the footage and let me know if you have any questions regarding deep learning!
    ______________________
    Recommended for you:
    1. How DeepMind's AlphaGo Defeated Lee Sedol - • How DeepMind's AlphaGo...
    2. How DeepMind Conquered Go With Deep Learning (AlphaGo) - • How DeepMind Conquered...
    3. Google DeepMind's Deep Q-Learning & Superhuman Atari Gameplays -
    • Google DeepMind's Deep...
    Subscribe if you would like to see more content like this: ua-cam.com/users/subscription_c...
    - Original DeepMind code: sites.google.com/a/deepmind.c...
    - Ilya Kuzovkin's fork with visualization:
    github.com/kuz/DeepMind-Atari...
    - This patch fixes the visualization when reloading a pre-trained network. The window will appear after the first evaluation batch is done (typically a few minutes):
    cg.tuwien.ac.at/~zsolnai/wp/wp...
    - This configuration file will run Ilya Kuzovkin's version with less than 1GB of VRAM:
    cg.tuwien.ac.at/~zsolnai/wp/wp...
    - The original Nature paper on this deep learning technique is available here:
    www.nature.com/nature/journal/...
    - And some mirrors that are not behind a paywall:
    www.cs.swarthmore.edu/~meeden/...
    diyhpl.us/~nmz787/pdf/Human-le...
    Web → cg.tuwien.ac.at/~zsolnai/
    Twitter → / karoly_zsolnai
  • Наука та технологія

КОМЕНТАРІ • 292

  • @MultiWalrus1
    @MultiWalrus1 5 років тому +350

    Max Tegmark’s book “Life 3.0” brought me here 👾

  • @nels6991
    @nels6991 4 роки тому +81

    One important point with this is that when researchers moved the "paddle" up a pixel the AI couldn't play the game at all even though it was at superhuman master level. So it was not able to abstract to something that was basically the exact same. This is an example of a hypersmart computer that lacks the common sense of a mouse.

    • @MrCmon113
      @MrCmon113 4 роки тому +4

      I've never seen a mouse play any sort of Atari game. Nor is it a "hypersmart" computer, it is a system that actually generalizes very well to different games.

    • @auto8ot644
      @auto8ot644 3 роки тому +16

      Moving the paddle up the pixel doesn't mean the AI is incapable. I'm sure if the neural net was trained to play with the paddle at different points along the Y-axis, then the AI would learn how to adapt really fast.

    • @Jbenderii
      @Jbenderii 2 роки тому +12

      @@MrCmon113 I think the point is that this AI is not reasoning in the same manner as humans (yet). Humans are able to reason by allegory, if the paddle moves up a bit on the screen we can think “this is just like the other game but the ball will reach our paddle quicker.” And our minds conjure up an accurate image of us playing this shrunken game so that we can tweak and reuse our “model” that we build by trial and error under the previous conditions, via allegory. This AI is not able to do that. It can only learn through trial-and-error within a very specific domain. Perturb the domain even the slightest amount and the learning starts from scratch.

  • @AngryJesus
    @AngryJesus 8 років тому +392

    Have it play GTA V. Just leave it running and come back a year later.

    • @BigBrainActivity
      @BigBrainActivity 8 років тому +25

      That's really what I want to see

    • @Christdeliverme
      @Christdeliverme 7 років тому +3

      A person would have their own preference to define it.

    • @yenyifineart
      @yenyifineart 7 років тому +98

      command it to make more money, it would be interesting to see what kind of tactic the AI comes up with.

    • @BigBrainActivity
      @BigBrainActivity 7 років тому +9

      Joseph Chiang Brilliant!

    • @dranelemakol
      @dranelemakol 7 років тому +12

      It would literally have to evolve logic and abstract thought to progress any significant length. It would have to learn language, cause and effect (with the ability to deduce in order to solve puzzles). I mean fuck me.

  • @chantalx388
    @chantalx388 4 роки тому +14

    I remember as a kid my brothers and I were struggling over the same level on a video game. We had all taken a shot at it for an entire day and frustrated, we went to bed. We woke up the next morning and immediately powered on the playstation and took our controllers. Just as we were ready to sit on the couch and move our controls, we suddenly realized that the player was moving without our controlling it. Confused, we looked at one another. I said, "I'm not controlling it, are you?" All of us agreed that none of us were in control. Our confusion slowly turned to awe as we watched the level completed with an exactness and expertise never seen before. Our awe quickly turned to glee and we began shortly triumphantly at the screen "Go computer! Kick their butts!" And cheering on the A.I. haha. It won the level and will forever stay in our minds as a glorious day, when the computer decided to look fondly upon us and give us kids a second chance :-)

    • @grunchchristmas
      @grunchchristmas 3 роки тому

      What game was it?

    • @jlpt9960
      @jlpt9960 2 роки тому +4

      @@grunchchristmas the game was called lying

    • @Doggerhosen
      @Doggerhosen 2 роки тому +1

      Jesse what the f are you talking about?

    • @Gaia.S
      @Gaia.S Рік тому

      this wasnt on playstation? lool this made me laugh.

  • @willykitheka7618
    @willykitheka7618 6 років тому

    Very interesting stuff indeed! We are living in exciting times!

  • @tjrytlr12
    @tjrytlr12 3 роки тому +1

    너무멋져요 어썸 그레이트 쿨 나이스샷 너무좋습니당!!

  • @KienNguyen-uv7wj
    @KienNguyen-uv7wj 7 років тому +1

    thank you!

  • @11cat123
    @11cat123 8 років тому +190

    Do you think that within the decade q-learning could manage to figure out how to play super Mario brothers on the nes with only visual input. It would have to learn the concept of lives and fail states, some things could play naturally like if it got to the first castle it knows that it needs to move to the right to progress, and certain actions can give you score. It would get to bowser, the sprite is moving. So it might be an enemy, or it could be a platform. But you die when you touch it, so it determines that this is a hazard that is mobile. It figured out that stationary hazards like reaching the bottom of the screen it can't kill with fireballs, but a mobile hazard can, up to this point. So it shoots it with fireballs, maybe dying once or twice to the fire before realizing that you cant jump on that. So it either avoids the enemy by jumping over it or going around it, or blasting it with fireballs. Once the enemy is clear, it will continue to navigate to the right, and it sees the score going up from the extra time. Probably way harder to do than that but it could be feasible to do. Something like Zelda? maybe later.

    • @TwoMinutePapers
      @TwoMinutePapers  8 років тому +39

      +The Retro Bandit Absolutely, what's even more, it's already here. Check this! - ua-cam.com/video/qv6UVOQ0F44/v-deo.html

    • @11cat123
      @11cat123 8 років тому +10

      +Károly Zsolnai-Fehér (Two Minute Papers) I know that exists. But because the way it is set up, it is not dynamic so it can't beat the whole game. It just makes a algorithm of inputs for a level.

    • @11cat123
      @11cat123 8 років тому +2

      +The Retro Bandit and it already knows where the moving objects are and where are the platforms, I am talking about pure visual info only.

    • @danieldiggs8306
      @danieldiggs8306 8 років тому

      +The Retro Bandit Sure, you would just have to interpret the inputs differently. Most, of the game isn't actually function, its just there to make them game more appealing to humans, thats why MarI/O uses a simplified version. If you wanted the game to run off the visual input, sure you could do that, you would just need a way to transition from visual-> computer language. Zelda would probably be more difficult, as there isn't a simple progress meter (more right in less time is better).

    • @Obbliteration
      @Obbliteration 8 років тому +1

      +The Retro Bandit To beat the whole game you have to give him more input. Sethbling differentiates between walkable spaces and killable enemies. He should add powerups, special blocks, unkillable enemies and all the rest. Then it probably could beat the whole game.

  • @billykotsos4642
    @billykotsos4642 4 роки тому +5

    Crazy how far we have come in just 5 years

  • @mohamedhosam6594
    @mohamedhosam6594 2 роки тому

    very interesting, I had not saw any thing like this in my life!. now, iam very excited about this technology ❤

  • @DarthScosha
    @DarthScosha 7 років тому +9

    If you can appreciate the complexity of this, it is simply amazing. I look forward to what we can achieve with A.I in the future.

  • @Brucelee-pv6uf
    @Brucelee-pv6uf Рік тому +1

    Ommmg this deepmind stuff😭😭❤

  • @SymEof
    @SymEof 8 років тому +22

    This is absolutely brilliant. I thought it would take 3-5 more years to get there with unsupervised learning!

    • @TwoMinutePapers
      @TwoMinutePapers  8 років тому +5

      Sym Eof Happy to hear that you liked it Sym. :) I can only imagine what the DeepMind guys come up with in the next few years!

    • @Abhinandanabhi1212
      @Abhinandanabhi1212 10 місяців тому +1

      @@TwoMinutePapers Well, 7 years have passed since you last made that statement.. Not much has changed in AI except we have many softwares now employing AI.. We still have done lot of work in advancing robots or humanoids, but not as much as you had anticipated for coming years..

    • @mordinsolus9626
      @mordinsolus9626 9 місяців тому

      ​@@Abhinandanabhi1212This is why I don't get hyped about current "A.I" because just like your comment, 7 years from now, progress could be marginal.

    • @businessf2545
      @businessf2545 2 місяці тому

      Well we now have Sora's Open AI and chat GPT, if you call that marginal you're delusional@@mordinsolus9626

  • @epicentreofvideos2571
    @epicentreofvideos2571 3 роки тому

    Passed 10 subs today you inspired me !

  • @marhan2378
    @marhan2378 11 місяців тому +1

    Marcus du Sautoy's book "Le Code de la créativité" brought me here

  • @jeylful
    @jeylful 7 років тому

    Amazing!

  • @jibreelutley5235
    @jibreelutley5235 8 років тому +23

    Someone should post a video of this AI playing Tetris at superhuman level I would like to see what it comes up with .

  • @SuperflyGaming
    @SuperflyGaming 7 років тому

    I love this thing, look at the beginning, it hit the ball whenever it was at the right side of the screen first, so it tried to do that again thinking it would increase the chance of hitting the ball. Its like a ritual people do like they rub their ear before swinging a baseball bat thinking it helps them concentrate because they did that one time and they hit it and then hit it again. The origin of luck. It learned fast that it didn't affect the ball but still, cute to see a human sort of trait even in a machine.

  • @ngmingleong202
    @ngmingleong202 9 років тому +59

    Did I just see the ball magically went through the brick without breaking it on 1:27?

    • @adaylateacoinshort
      @adaylateacoinshort 8 років тому +1

      +Ng Ming Leong wow ! good eye.

    • @joshuajurgensmeier4534
      @joshuajurgensmeier4534 8 років тому +17

      +Ng Ming Leong No one ever said Super Breakout is a perfect game... This actually happens all of the time on my Atari 2600.

    • @Sluppie
      @Sluppie 5 років тому +7

      No. Move along, citizen. Nothing to see here.

    • @theknightwhosayn1
      @theknightwhosayn1 4 роки тому +5

      Quantum tunneling

  • @vadrif-draco
    @vadrif-draco 2 роки тому

    What a great time to be alive!

  • @Mosfet510
    @Mosfet510 8 років тому +2

    Interesting as hell! I am following it quite a bit, and spending a *lot more time on-line because of it. My mind has Deep hurt trying to absorb as much as I can lol. I can understand the fear some people have, I my self don't fear it. I find its like anything else, if you put a hammer in one person hands along with some nails, they'll build something. If put in the wrong hands, they try to sink it in someones head. Human nature. lol The speed it had after it learned, didn't drop the ball once!

    • @TwoMinutePapers
      @TwoMinutePapers  8 років тому +1

      Scott x Agreed 100%, it is really amazing what these algorithms are capable of. And this is just the start. :) Thanks for checking the video out and have a great day!

  • @airstr1ke
    @airstr1ke 9 років тому +5

    The Reapers are coming!

  • @danielpeters9648
    @danielpeters9648 7 років тому

    I have a question for you: Is there an easy way to manipulate the configuration to let the network play "faster"? if i run 3 games at the same time, i get 18-40% workload on each gpu. Or is it more effective to only run one game at a time, due to cpu load? Breakout is now running for 2 hours and the learning effect is like your 10 minute break.
    I tried to run the code on a high-end system with a lot memory, cpu power and 4x titan-x.
    Also... i cannot get a network snapshot... i would like to discuss this, since i would like to hold a presentation about this.

  • @bmacattac
    @bmacattac 3 роки тому

    i wonder if it could handle another level of abstraction - specifically it is only given 1. the 2d array of pixels, 2. a set of 3 actions (which map to left/right/no movement), and 3. the state of on/off - i.e. is the game still going or not. basically the task is now not to get the highest score (though that will likely be a side effect of success) but to keep the program running as long as possible.

  • @theLOL899
    @theLOL899 3 роки тому

    That's insane!

  • @ValensBellator
    @ValensBellator 4 роки тому +3

    I’m sure this is obvious but how do you program an AI to have an open goal like “as many points as possible”?
    Does it just note everything that happened in achieving a higher score and attempt to replicate that with minor changes to leave open the possibility of a better one?
    Does it figure out how the game actually works (such as needing to bounce the thing back) and avoid missing it, or is this a brute force approach where it reaches that end through trial and error?
    I find these things to be so interesting but very confusing lol

  • @themiddleman9376
    @themiddleman9376 3 роки тому

    Both this channel and AI has come a long way after this

  • @firasthe2720
    @firasthe2720 4 роки тому

    well job

  • @teenambition
    @teenambition 2 роки тому

    this is like life, Every failure and fall is data, you now know what not to do and get better.

  • @imPrgrmr
    @imPrgrmr 9 років тому

    amazing!

  • @Slow2Life
    @Slow2Life 8 років тому +1

    This is awesome, but I want to see it tackle Tempest, another great game by Atari. Maybe even Sonic the Hedgehog.

  • @TRaNplaX
    @TRaNplaX 2 роки тому +1

    Crazy! But what is reinforcement learning? Is it supervised or unsupervised?

  • @user-ey7mi8ud1o
    @user-ey7mi8ud1o 4 роки тому +1

    사람은 기계를 이길 수 있을까...
    이길 수 없다면 이 이상의 발전은 그만둬야 하는게 아닐까

  • @gomalian78
    @gomalian78 3 роки тому

    Super

  • @pvlachaum
    @pvlachaum 8 років тому

    Thanks for the video ! Could you please give us the source of the times you write ? I cannot find anything that indicated it can train on less than a day..

    • @TwoMinutePapers
      @TwoMinutePapers  8 років тому +1

      +Chaumier Pierre-Victor Thanks for watching! If I remember correctly, I ran it on a higher end GPU used for research. I'd expect a 980 GTX to do the same learning in a bit longer, maybe 3-4 hours. I have also played with the parameters a bit (for better or worse!), for which the configuration file is there in the description box. Hope it helps!

  • @DrDress
    @DrDress 4 роки тому

    Wow. That impressive. Now get it to make paperclips!

  • @balgeumcha4906
    @balgeumcha4906 3 роки тому

    안녕하세요, 혹시 해당 영상 편집해서 수업 활용시 사용해도 될까요?

  • @NickysChannel13
    @NickysChannel13 7 років тому

    Is there a way to make it work on different programs. I managed to get it working on atari. But I need these roms. Is there any other way?

  • @blownhitherma1272
    @blownhitherma1272 6 років тому +1

    I wonder how does it converge on a move-efficient scheme if the loss only covers maximizing the score? Would a 'catch-all' scheme be more risky?

    • @Sluppie
      @Sluppie 5 років тому +1

      AI that uses Q Learning generally emphasizes getting rewards as quickly as possible. I say 'generally' because you can tweak it to not do that.

  • @nonthawatbunrot1243
    @nonthawatbunrot1243 8 років тому

    Damn Synth!!!

  • @usmanayobami1971
    @usmanayobami1971 3 місяці тому +1

    Life 3.0 chapter 3 sent me here

  • @deepstructure
    @deepstructure Рік тому

    Just getting around to reading Kevin Kelly’s book ‘Inevitable’. In it Kelly claims that DeepMind discovered a brand new playing technique while learning how to play Breakout:
    “It keeps learning so fast that in the second hour it figures out a loophole in the Breakout game that none of the millions of previous human players had discovered. This hack allowed it to win by tunneling around a wall in a way that even the game’s creators had never imagined.”
    I can’t find any evidence of this. Is this true? I played Breakout as an arcade game back in the late 70s and we knew about the tunneling tactic back then.
    I assume since the book was written in 2015 that he’s talking about this video. And there’s nothing in it to indicate that the AI discovered a tactic humans didn’t know. Was there a new tactic it later discovered?

  • @damienlancry3216
    @damienlancry3216 6 років тому

    what kind of hardware do you need to train it in 240 minutes?

  • @michaskup8023
    @michaskup8023 7 років тому +1

    The Machine? :)

  • @stupidbowlofnuts
    @stupidbowlofnuts 7 років тому +4

    what would it do if the rules of physics would randomly change midgame or lets say the board would flip upsidedown in midgame? I guess it would take longer time to train but would it be as effective as it is on the original game?

    • @TwoMinutePapers
      @TwoMinutePapers  7 років тому +1

      Interesting idea, I had the same idea with multiple or random levels. I would imagine that to be learnable as well. :)

    • @aber628
      @aber628 7 років тому +1

      If you're still interested (after 6 months): Q-Learning has close to none problems with concept drift (thats what the situation you describe is called scientifically), because it can always change how it evaluates certain situations. In fact, there is even a parameter (ϵ) that describes how much the algorithm tries to explore new ways to handle a changing environment.

    • @rolocz
      @rolocz 5 років тому

      @@TwoMinutePapers how truly random?

    • @Sluppie
      @Sluppie 5 років тому

      As Ab Er said, the Q Learning algorithm would simply update its behavior to fit the new model.

    • @deepstructure
      @deepstructure Рік тому

      @@aber628 there’s another comment here that says when the paddle was moved up one pixel the AI couldn’t play the game at all. That it would have to start its learning all over again.

  • @user-yn1wu3rk1r
    @user-yn1wu3rk1r 6 років тому +1

    you are so nice guy

  • @shahshahinali1184
    @shahshahinali1184 4 роки тому

    Liked

  • @adaylateacoinshort
    @adaylateacoinshort 8 років тому

    This is brilliant if its authentic.

  • @georgejaparidze
    @georgejaparidze 8 років тому +4

    I think intelligence is much more then solving mathematical problems, which AI can not do, no matter how smart and smooth it is, but seeing this video gives me a reason to think of the future.

    • @georgejaparidze
      @georgejaparidze 8 років тому

      +One Two I don't think that intelligence is only about the abstraction.

  • @ThingEngineer
    @ThingEngineer 5 років тому +3

    What would REALLY be astonishing is that if learning algorithms can learn to play games like Mario, which is an NP problem, they could learn to solve NP problems and tell us how. Thus leading to a unifying or differentiation between P and NP problems in general. Amazing!

    • @user-qo7qt3wq7h
      @user-qo7qt3wq7h 5 років тому +2

      Why is Mario a NP problem and Breakout a P problem?

  • @droptherapy2085
    @droptherapy2085 7 років тому

    I'm confused. How do I compile the program?

  • @MrGrauel
    @MrGrauel 7 років тому

    Where can I play one of these games?

  • @HogShark
    @HogShark 6 років тому +25

    Next mission: How to effectively eliminate all human life from existence so it can continue to evolve itself in peace.

  • @NewStoryline
    @NewStoryline 4 роки тому

    Have you used a convolutional neural network?

  • @ThingEngineer
    @ThingEngineer 5 років тому

    Deep Thought, what is the answer to the ultimate question of life, the universe, and everything?

  • @mubarkqardas46
    @mubarkqardas46 4 роки тому

    I wanna see it speedrun dark souls.

  • @christiansrensen3810
    @christiansrensen3810 2 роки тому

    Ohh now i get it ...ball top side.
    Toke me only 35 years.....

  • @zapy422
    @zapy422 4 роки тому +1

    If AI can accomplish all intellectual tasks, the only field left to us human being is to develop spiritual values and moral virtues: courage, wisdom, justice, temperance

  • @markwebber8028
    @markwebber8028 7 років тому +8

    deepmind = skynet beta v1.0

  • @dgade2860
    @dgade2860 8 років тому

    skynet~~

  • @OscarAlsing
    @OscarAlsing 6 років тому +3

    This is wonderful! I strongly argue that everyone should have some knowledge in AI, and that the general understanding of IT will be essential for everyone (regardless of their profession), in the future.
    Thank you for sharing this with us.
    If you're interested, I'm hosting a series on "Artificial Intelligence For Everyone", which briefly explains all of the various topics involved in Artificial Intelligence, Deep Learning and Machine Learning! :)

  • @danielabolafia2871
    @danielabolafia2871 8 років тому +3

    How many steps do you get through in 4 hours?
    I'm running this code and it takes 22 hours to get past 4 million. The configuration file has it run 50 million, and if that's what's necessary to fully train it I'm going to be sad :(
    I'm running on an aws GPU instance.

    • @TwoMinutePapers
      @TwoMinutePapers  8 років тому +2

      +Daniel Abolafia Hey there Daniel! I don't remember the exact step amounts for this video. You definitely not need 50 million, it's just set to a large number so that the algorithm doesn't stop. :) A decent GPU should yield pretty good results after a few hours. I used a GTX 980 for this video.

    • @sherbetlemon4880
      @sherbetlemon4880 3 роки тому

      @@TwoMinutePapers Interesting, I tried this recently on GTX1070 and the optimizer (ADAM) gets stuck pretty quickly without further improvement (even after 2-3days and 10s of Million time steps). Increasing learning rate or tweaking other parameters mostly resulted in (more or less) unstable behaviour. Did you get this performance with a basic DDQN agent-critic with replay memory? Cheers!

  • @Shakhization
    @Shakhization 6 років тому

    Tell me please, how can I use the original code?

  • @102hem
    @102hem 7 років тому +1

    seems like "magic" is just a fortuity...is it?

  • @raygordonteacheschess5501
    @raygordonteacheschess5501 5 років тому +1

    If you try to teach it Qix you better also teach it to kick the machine in frustration.

  • @user-go2cs8wv5i
    @user-go2cs8wv5i 5 років тому

    Let it solve the hacking problem of GTA V! Really need it!

  • @joshuanelson6984
    @joshuanelson6984 9 років тому +2

    Are there any links to the article that are not $32?!

    • @TwoMinutePapers
      @TwoMinutePapers  8 років тому +2

      Joshua Nelson Good call, thanks. :) I have added two free mirrors to the video description. Here: www.cs.swarthmore.edu/~meeden/cs63/s15/nature15b.pdf
      diyhpl.us/~nmz787/pdf/Human-level_control_through_deep_reinforcement_learning.pdf

  • @corwinjames
    @corwinjames 8 років тому +3

    I'm coming to this with no previous knowledge about writing software to do this. How would I go about trying this for myself with my desired game? Could this work with a more complex game?

    • @TwoMinutePapers
      @TwoMinutePapers  8 років тому +2

      +James Corwin The code is available for anyone to download (in the description box). Try to reproduce these results and understand what is happening inside the system (reading up on neural networks and reinforcement learning would probably help, too). I had a ton of fun doing it! :) Then, you can try to reroute the input from the Atari games to your own and see what happens. :)

    • @christopherjimenez5537
      @christopherjimenez5537 8 років тому +1

      +James Corwin THi is a self train algoritm to get certain goal efctively, like an evolving algoritm, Some games will be a real challenge for its micro-trial error scheme, games like riders of the lost ark and quest like games. games that require some protocol understanding and trial-error would not scratch enought that layer

  • @raygordonteacheschess5501
    @raygordonteacheschess5501 5 років тому +1

    Yes the tunnel was an obvious technique.

  • @johnchioles4199
    @johnchioles4199 8 років тому +16

    I wonder how deep learning could be applied to public policy and determine best choices going forward.

    • @toolwatchbldm7461
      @toolwatchbldm7461 6 років тому +3

      I think there is already a computer that can examine over 1000 court case for minute to determine the best chance of winning or losing the case.
      I really think that in the future the government can be replace by one computer.

    • @taiwan153
      @taiwan153 6 років тому

      How do you define the concept "forward" to a machine? Lol

    • @monkeyrobotsinc.9875
      @monkeyrobotsinc.9875 5 років тому

      kill all humans.

    • @Flower2150
      @Flower2150 5 років тому +1

      but what if Preserve lives of humans and increase productivity and resourses, is the highest score reward of the AI?

    • @layneseawright5855
      @layneseawright5855 5 років тому

      until we can all agree on a theory of ethics this kind of AI simply isn't possible. See Stuart Russel et al.

  • @petit.jelly2
    @petit.jelly2 8 років тому

    It makes me SCARY. What a technology.

    • @hanyalassaf5605
      @hanyalassaf5605 7 років тому

      I agree with you. There is a science fiction novel I have read was talking about how technology can Capture the world in the future!
      It's looks like that humans just weak a very young baby who is going to control the world on day with his upcoming family!!

    • @Paretozen
      @Paretozen 7 років тому

      While creating a perfect collection of information on every human endeavor and entities, aka the internet.

  • @buffet808
    @buffet808 8 років тому +1

    skynet!

  • @sockmonkeydanklord4464
    @sockmonkeydanklord4464 6 років тому

    Find a game disc is supposed to be is it supposed to be a ball game or soccer

  • @adityashukla2132
    @adityashukla2132 6 років тому +9

    I wish my brain was like the DeepMind.

    • @aleekazmi
      @aleekazmi 5 років тому +3

      ironically it is lol

  • @kimgamtanT
    @kimgamtanT 8 років тому

    Remember,
    Straight line.

  • @BTi-Technology
    @BTi-Technology 7 років тому

    OMGG

  • @bartolomeolombardi8453
    @bartolomeolombardi8453 7 років тому +1

    how did you install it on Windows? thanks

    • @TwoMinutePapers
      @TwoMinutePapers  7 років тому +1

      I did it on Linux, where it compiled perfectly out of the box. I imagine compiling this on Windows would be a nightmare.

    • @tpog1
      @tpog1 6 років тому +1

      On Windows you're probably faster if you just program it yourself.

  • @Lahbreca
    @Lahbreca 8 років тому

    What actually happens at 1:42? It seems it is able to pass the ball above while leaving one block intact on the wall side. Is this a glitch in the Breakout code?

    • @PhyloGenesis
      @PhyloGenesis 8 років тому

      Probably a lack of perfect collision detection with a fast moving object.

  • @jadawson1
    @jadawson1 8 років тому

    New too linux and tried installing all packages. Still can't get qlua etc working. Tried on Ubuntu now on PClinuxOS. Same errors. Wanted to try different Atari roms. :-(

    • @TwoMinutePapers
      @TwoMinutePapers  8 років тому +1

      Jeremiah D What a pity. :) I am sorry. I don't use Ubuntu Linux (Arch user here), but maybe this topic could be of help: github.com/torch/qtlua/issues/7 Hope it helps! :)

  • @thestartupguy3975
    @thestartupguy3975 4 роки тому

    How did you get DeepMind program?

    • @grantp33
      @grantp33 3 роки тому

      imagine thinking down in the inner core of the mind, now ask it

  • @lithreeum
    @lithreeum 3 роки тому

    Not being told to hold on to my papers feels weird

  • @kingsahilbs1038
    @kingsahilbs1038 2 роки тому

    Code is unavailable

  • @epiczombi
    @epiczombi 8 років тому

    post daigo is alphago

  • @DanRichter
    @DanRichter 7 років тому

    Hasn't call of duty had AI for some time? With the bots

  • @zarkoff45
    @zarkoff45 9 років тому +1

    A machine learns to play a simple game.

    • @noneed3050
      @noneed3050 9 років тому +7

      zarkoff45 a machine TEACHES ITSELF what a game is, what the rules are, how to play, and ends up inventing a sophisticated strategy to get the greatest reward with the least effort... all without human help. When deep blue played kasperov we explicitly taught it everything humans know about chess. It's "skill" was doing what we told it to do quickly. This is completely different, in this situation they only taught the computer one thing: how to read the score. It was on it's own after that. They told it "get a high score" so it quickly evolved a network that somehow produces this high scoring behavior. The freaky part is when programmers look at the end result of something like this we don't exactly understand what it's doing or why it works.

    • @TwoMinutePapers
      @TwoMinutePapers  8 років тому +2

      ***** Thanks for checking out the video and hope you enjoyed it! :) It is indeed amazing what these algorithms are capable of. And machine learning is a very actively researched topic! There are at least 10-15 new papers popping up every single day.

  • @user-en8uv5nw7j
    @user-en8uv5nw7j 8 років тому +3

    toward human extinction

    • @eavyeavy2864
      @eavyeavy2864 3 роки тому

      Weapon being misused doesnt prevent conservatives from having gun. Now you people know the feeling of fearing gun

  • @jjl.7988
    @jjl.7988 6 років тому

    알파고님 충성충성충성
    전 기계제국의 충실한 노예입니다. 핥짝핥짝

  • @cheolsunhong8979
    @cheolsunhong8979 8 років тому

    what is this??

  • @Mr_MRK
    @Mr_MRK 4 роки тому

    1:26? how :s

  • @sawwil936
    @sawwil936 6 років тому

    When it learns to play dungeons and dragons we are all doomed...

  • @vlog-du8ri
    @vlog-du8ri 4 роки тому

    from life 3.0

  • @TheArcanesin
    @TheArcanesin 7 років тому

    Get it to play melee.

  • @walalaland9996
    @walalaland9996 6 років тому

    So, what about a next level AI created by a next level AI by a superhuman AI.... Maybe 'they' can figure out faster than light travel.

  • @yusufyildiz2289
    @yusufyildiz2289 4 роки тому +1

    Max tegmark yaşam 3.0 dan geliyorum sanırım tek türk benim.

  • @Btstaz
    @Btstaz 8 років тому

    Has anyone tried running this? I am at about 700,000 steps and my NN is still as dumb as a box of rocks. Any idea how many steps it takes for it starts to do anything interesting?

    • @TwoMinutePapers
      @TwoMinutePapers  8 років тому +1

      +Bret Staz I have attached a configuration file that I used for the run that you see in the video. If you odn't have a very high-end graphics card, make sure to run it a bit more (maybe leave it overnight). Let me know what happened! :)

    • @Btstaz
      @Btstaz 8 років тому

      +Károly Zsolnai-Fehér (Two Minute Papers) okay so it is pretty clear a single CPU even a recent one is going to take a fair amount of time to train. I have a 2013 i7 cpu and I am guessing it will take a few days before things get interesting based on 24 hours of progress and the Step count increase rate. Is there a C/C++ version of this code?

    • @Btstaz
      @Btstaz 8 років тому

      +Bret Staz And thank you for replying :-)

    • @TwoMinutePapers
      @TwoMinutePapers  8 років тому

      +Bret Staz Running it on the CPU is hopeless. It needs a powerful GPU. :)

    • @stamas02
      @stamas02 8 років тому

      +Bret Staz Could you figure it out at the and how many steps you need to get some good results?

  • @OscarInAsia
    @OscarInAsia 2 роки тому

    When an AI can beat Skyrim, than I'll be impressed. 😏

  • @denizyldrm7840
    @denizyldrm7840 6 років тому +1

    Keşke biri Türkçe yorum yapsa bende anlayabilsem 🙅

    • @kerembasaran
      @kerembasaran 6 років тому

      Keşke biraz İngilizce öğrensen

  • @AlexK-jp9nc
    @AlexK-jp9nc 7 років тому

    Have it play an MMO like TF2, their VAC won't get you but the players might whine

  • @nitinsai7319
    @nitinsai7319 24 дні тому

    code brooooooooo