This Image Is Fine. Completely Fine. 🤖

Поділитися
Вставка
  • Опубліковано 31 лип 2024
  • ❤️ Check out Lambda here and sign up for their GPU Cloud: lambdalabs.com/papers
    📝 The paper "The Sensory Neuron as a Transformer: Permutation-Invariant Neural Networks for Reinforcement Learning" is available here:
    attentionneuron.github.io/
    🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:
    Aleksandr Mashrabov, Alex Haro, Andrew Melnychuk, Angelos Evripiotis, Benji Rabhan, Bryan Learn, Christian Ahlin, Eric Haddad, Eric Martel, Gordon Child, Ivo Galic, Jace O'Brien, Javier Bustamante, John Le, Jonas, Kenneth Davis, Klaus Busse, Lorin Atzberger, Lukas Biewald, Matthew Allen Fisher, Mark Oates, Michael Albrecht, Nikhil Velpanur, Owen Campbell-Moore, Owen Skarpness, Rajarshi Nigam, Ramsey Elbasheer, Steef, Taras Bobrovytsky, Thomas Krcmar, Timothy Sum Hon Mun, Torsten Reil, Tybie Fitzhugh, Ueli Gallizzi.
    If you wish to appear here or pick up other perks, click here: / twominutepapers
    Thumbnail background design: Felícia Zsolnai-Fehér - felicia.hu
    Meet and discuss your ideas with other Fellow Scholars on the Two Minute Papers Discord: discordapp.com/invite/hbcTJu2
    Károly Zsolnai-Fehér's links:
    Instagram: / twominutepapers
    Twitter: / twominutepapers
    Web: cg.tuwien.ac.at/~zsolnai/
  • Наука та технологія

КОМЕНТАРІ • 352

  • @csoest24
    @csoest24 2 роки тому +497

    I love the reference to the reverse steering bike example. That is a very clear comparison to shuffles being thrown at the neural networks.

    • @sebastianjost
      @sebastianjost 2 роки тому +12

      But the network was trained with shuffled inputs right?
      If humans practice like that they might also be able to cope with the shuffled inputs.

    • @TheMarcusrobbins
      @TheMarcusrobbins 2 роки тому +2

      @@sebastianjost I agree a human could learn to play the game with these inputs. They may never realise they are playing pong - but they will be able to work out the rules. You won't get the benefit of using the priors that the visual cortex has, but so what.

    • @MrFEARFLASH
      @MrFEARFLASH 2 роки тому +6

      In fact, the neural network handles such problems poorly. Pong always arrives at one point, and the car is not driving along the road, albeit in the right direction. It seems to me that this is not a complete victory!

    • @sunboy4224
      @sunboy4224 2 роки тому +3

      I would argue that the reverse steering bike is more complicated. Turning the handlebars involves moving your center of mass, and on typical bikes it moves in a way that is contusive to steering in that direction. To ride the reverse bike, it's not just relearning that clockwise = left and counter-clockwise = right, it's entirely relearning how to posture yourself on the bike.
      Humans are actually pretty good (albeit slow) at purely switching inputs (see, well, many examples of video games in which this happens as a "status effect").

    • @nadiaplaysgames2550
      @nadiaplaysgames2550 2 роки тому +2

      In the case of backwards bike our internal brain networks was only trained on one way turning bikes to make if some rode both backward and forwards bikes there internal model accommodate that. for it 100% fair the neural net work should be trained based off same away if also for a human to learn a scrambled version of the game they could. but it would take longer because our brains already levvage older models to build new ones

  • @antiHUMANDesigns
    @antiHUMANDesigns 2 роки тому +393

    One very important difference is that humans are always trying to save energy. Do as little work as possible.
    For a human to play a game the way that AI does, it would have to be pounding the keys constantly, but humans instead try to move in straight lines and only adjust their heading when needed.
    And to re-learn things take energy, which humans need to care about but AIs don't.
    So, make a lazier AI and see if it behaves more like a human.

    • @ViRiXDreamcore
      @ViRiXDreamcore 2 роки тому +35

      That’s a really good point. Also there are no physical limitations on an AI other than the circuitry it runs on.

    • @toby3927
      @toby3927 2 роки тому +39

      Yeah, that makes sense. Also human brain is entirely optimized for the earth environment where these strange things never happen, but AIs are trained from the ground up to handle these things.

    • @anywallsocket
      @anywallsocket 2 роки тому +33

      Yes we are hunter gatherers not kaleidoscopic pong players 😂

    • @SemperFine
      @SemperFine 2 роки тому +3

      But sometimes we gets bursts of energy too

    • @pvic6959
      @pvic6959 2 роки тому +9

      this is really funny but imagine an ai that is efficient.. it would become so much better! we have evolved over millions of years to be efficient meat machines lol

  • @saudfata6236
    @saudfata6236 2 роки тому +219

    (right at the end) - car drives completely off the road - "no issues whatsoever!"

    • @yash1152
      @yash1152 2 роки тому +5

      7:05

    • @whiterottenrabbit
      @whiterottenrabbit 2 роки тому +1

      Reminds me of Boosting Stop-Motion to 60 fps using AI (ua-cam.com/video/sFN9dzw0qH8/v-deo.html)

    • @Ajay_Pathak_
      @Ajay_Pathak_ 2 роки тому

      Lol

    • @clray123
      @clray123 2 роки тому +3

      Now imagine this is a robotic dog shooting guns at people - vendor's ad: "no issues whatsoever!"

  • @WilliamDye-willdye
    @WilliamDye-willdye 2 роки тому +31

    I'm not convinced that shuffling the image reveals differences in how we think so much as differences in how we see. Still, it is a clever and worthwhile idea. Well done.

    • @warrenarnold
      @warrenarnold 2 роки тому

      Wow Just like me with my broken screen

  • @spencereaston8292
    @spencereaston8292 2 роки тому +146

    My question would be how many failure-resets happen before a success. Sure it took Dustin a week before he figured the bike out, but how many failure-resets was that? Human vs AI in this realm seems to be scale of time not capacity. Which in itself makes it a wonderful time to be alive!

    • @realmetatron
      @realmetatron 2 роки тому +15

      It took Destin 8 months or so, practicing 5 minutes every day. His son did it much faster because a child's brain learns easier.

    • @getsideways7257
      @getsideways7257 2 роки тому +6

      @@realmetatron And that's what we should be comparing it too. An adult's brain is basically solidified and almost unable to progress with these kinds of things (not to mention being riddled with all the useless data and conventions not helpful for the task at hand), but a clean slate kiddo brain is much closer to an artificial neural network in its learning ways.

    • @baumkuchen6543
      @baumkuchen6543 2 роки тому +2

      The question is as well what is a failure-reset. A normal human will usually 'hit a reset' before failure, which in case of a bike ride is falling off and getting injured.

    • @tim40gabby25
      @tim40gabby25 2 роки тому +1

      I noticed learning how to tightrope walk - a foot off the ground, it's ok - aged (55) was made easier by simply ignoring what I was trying to do, chatting with friends and family etc. Concentrating on the task was worse than useless. Curious, that.

    • @rayujohnson1302
      @rayujohnson1302 2 роки тому +4

      ​@@tim40gabby25 The unconscious mind is processing 500,000 x more information per second then your puny frontal cortex can muster. By not directly focusing on a task you are no longer slowing down the unconscious mind. This is also why taking a break from solving a hard problem actually helps you solve said problem.

  • @franticsledder
    @franticsledder 2 роки тому +98

    Car drifting off road and in the ditch.
    Tesla AI: "No issues whatsoever!"

    • @khiemgom
      @khiemgom 2 роки тому +5

      Well, in racing that grass patch so i mean a little off road is not rlly a big deal

    • @Andytlp
      @Andytlp 2 роки тому

      A.i would never do that in all a.i road traffic. The conclusion you came up with is a human idea, human error. A human would drive into a ditch and complain about something irrelevant to avoid responsibility of facing reality.

    • @brightblackhole2442
      @brightblackhole2442 2 роки тому

      @@Andytlp 7:04 is an example of what they were actually trying to say

  • @LanceThumping
    @LanceThumping 2 роки тому +106

    I mean doesn't the bike example and such not include time for training that the AI received?
    I've had various games that reshuffle controls as a penalty or a challenge and after doing it long enough you can adapt rather quickly.
    The fact that the brain can adapt to it's visual input completely inverting at all in adults means that testing could be done to see how far it can be pushed.
    Maybe a child that grows up with glasses that randomly shuffle their vision every second would adapt.
    I feel we don't know enough about the human brain at these extremes to say for sure that the AI is acting in an entirely non-human manner outside of the fact that it has computer speed to allow it to switch up faster than a physical network of neurons can.

    • @swe223
      @swe223 2 роки тому +16

      I remember that video on YT where a guy puts on glasses with a mirror to see the world upside down. The beginning was awful, but after a few days he adapted and was able to do things like pouring milk into his cup just fine. Then the inverse happened when he took them off^^ (although adaptation was much faster)

    • @mattp1337
      @mattp1337 2 роки тому +5

      "A child that grows up with glasses that randomly shuffle their vision every second". Good writing prompt for a sci-fi story, given basic understanding of the topic. Would a human brain that adapted to these conditions perceive more reliably than everyday humans? Conversely, can any common optical illusions in human vision be traced to our vision NOT being trained to deal with perturbation? I'm a migraine sufferer, and that messes with my vision quite often: can it explain my moderately-above-average artistic and mathematical aptitudes? What a time to be alive.

    • @ToyKeeper
      @ToyKeeper 2 роки тому +10

      Yeah... Humans can definitely adapt to stuff like this. It just doesn't happen as fast. I don't think the shuffling bit shows a qualitative difference between human and AI processing methods... it just shows that AI can adapt to changes faster.

    • @user-sl6gn1ss8p
      @user-sl6gn1ss8p 2 роки тому

      I think just maybe your experiment ins't going to pass the ethics committee : /

    • @user255
      @user255 2 роки тому +3

      If human neurons were that quick to adapt, they might adapt too much into some random non-sense and be in troubles 99% of the cases.

  • @Logqnty
    @Logqnty 2 роки тому +37

    For the pong example, the ai was just moving up slightly, and every time the ball would spawn in the same direction. because of how it spawned, the ai could just move up slightly and still win every time, regardless of the shuffled input.

    • @anywallsocket
      @anywallsocket 2 роки тому +1

      Watch more closely, that’s only somewhat true.

    • @Nulley0
      @Nulley0 2 роки тому +1

      5:08 I agree, the ball starts at center, and it is the same everytime. Other simulations are fine, pong is broken

  • @perschistence2651
    @perschistence2651 2 роки тому +8

    I think this simply means that it does not matter how WE see the input as humans. The problem why we need to adapt, for example with the bike is probably because we already created a model in our head, an abstraction layer, you could say and if you change how the bike works, our whole abstraction-layer needs to be rewritten. The AI is simply more primitive, it has no abstraction layer but processes the information directly. Our thinking/abstraction makes us slow, but it also enables us to solve way more general problems.

    • @AirNeat
      @AirNeat 2 роки тому

      We're also slow because of biology and chemistry. Silicon is much much faster at calculating things.

  • @MrrVlad
    @MrrVlad 2 роки тому +27

    for driving, it may use wiggling to map the squares. the constant left-right motion and corresponding translation-rotation within squares tell exactly how far a certain square is.

    • @anywallsocket
      @anywallsocket 2 роки тому +5

      That is true, and actually a general trend for how neural networks fit a function: they exploit differential motion (tweaking nobs slightly) to see how they affect one another, gradually mapping the function space.

  • @Paulo_Dirac
    @Paulo_Dirac 2 роки тому +31

    Doesn't the fact that it was looking to the side of the "road" for curvature indications a human like feature?

    • @1xdreykex1
      @1xdreykex1 2 роки тому +2

      You can compare it or even mimic human behavior but it can never be identical because of how deep rooted our behavior is to biology we’d have to understand psychology and neurology a lot more as a species

    • @sychuan3729
      @sychuan3729 2 роки тому +2

      Yeah I also think about it. If you provided with this task you'll try to find meaningfull parts which were several blocks, while others were distraction

    • @quantumblauthor7300
      @quantumblauthor7300 2 роки тому

      @@1xdreykex1 I mean, I don't think our biology naturally accounts for driving behaviors

    • @anywallsocket
      @anywallsocket 2 роки тому

      Sure, but a human could not navigate the shuffled road: Our understanding of the game of ‘steering an object’ demands a continuous environment, simply because we have no experience with counter examples. Things tend not to teleport spatially in our daily lives! But to this AI it is business as usual.

    • @sychuan3729
      @sychuan3729 2 роки тому

      @@anywallsocket I think human could adapt to this after training. Blind people or people with visual diparities could navigate in environment , also some people have hallucination but still could. AI more adaptable but it isn't so much different

  • @ReynaSingh
    @ReynaSingh 2 роки тому +52

    I don’t know if we can ever mirror human thinking perfectly but this is impressive

    • @tiefensucht
      @tiefensucht 2 роки тому +5

      but we could already simulate insects to 100%. human brain for sure some day. the greater problem is that if you want a human like beeing, you would have to raise it like a human.

    • @TheMostGreedyAlgorithm
      @TheMostGreedyAlgorithm 2 роки тому +15

      I don't know if we will ever need to do this. The goal is not to create a virtual human. The goal is to solve a task in a best way. I think that the idea of that video: "Machines doesn't think like us and they shouldn't". The ability to solve a task in cases, where humans can not is not human-like, but this is what we want from machine.

    • @juraganposter
      @juraganposter 2 роки тому

      Yes, but still long way to go. I predict in the 250-500 years.

    • @dgollas
      @dgollas 2 роки тому

      You don’t FEEL we can mirror human thinking.

    • @NanoMan737400
      @NanoMan737400 2 роки тому +3

      Why mirror a human brain when you can make algorithms like these, that do things humans never could? I think that's actually much more helpful to us than just digitized versions of ourselves.

  • @DouglasThompson
    @DouglasThompson 2 роки тому +1

    Your mic setup sounds excellent! Quality overall is excellent, great job with another one!

  • @noutram1000
    @noutram1000 2 роки тому +54

    This is a good example of how AI is already 'beyond Human' in its ability to adapt. The fact it doesn't attribute too much to the stream of information coming in (like a lifetime of relying on your eyes to consider structures, lines, angles, etc.) is actually a plus -to the AI it is just a huge sequence of 1's and 0s that the nueral network provides the optimal answer for. It would be almost impossible for a Human to take the raw 'computer stream' and interpret it. We cannot just see constructs in the binary stream, we have to apply years of Human learning to make any sense of it...

    • @abdullahalqwain3490
      @abdullahalqwain3490 2 роки тому

      If the quantum computer is integrated, it is an answer to the problems facing artificial intelligence. It is very fast, super fast, a penalty of a second, very complex information, and very huge information in seconds. It is a future with artificial intelligence.
      The solution to many human problems is very complex

    • @manzell
      @manzell 2 роки тому +9

      I think what this reveals is that humans have a built in reward function for throwing information down strongly-established neural pathways, even if the data doesn't fit. Since the learning algo doesn't have this - its reward function is tied strictly to whatever it's programmed for - it can adapt much more rapidly. This is what's behind anthropomorphization, CBT-style "thought distortions" and so on.

    • @rocksfire4390
      @rocksfire4390 2 роки тому

      it's great to see the leaps humans have made in AI but AI itself isn't magic. it's all predetermined, based on the inputs and limitations given to the program. it's only as good as we make it. a purpose built AI is going to iterate much faster on it's sole task then a human could ever simply because a AI doesn't need to think about anything else but what's it's programed for.
      however because of this fact it cannot actually ever compete with a human brain.
      computers are recreations (attempt) of our brains. however they too are far more focused with a single task and that task is computation. however computation can be used to simulate other areas of our brains but they are still no where near our level as these simulations are not purpose built like a computer is.
      you would be quite surprised by how much information our brains process on a daily basis.
      the level of data our brain's sort through every second is quite hard to grasp. most of that data you don't even notice because it's just not something people think about but is required in order to stay alive. like take all of your organs for instance, the brain keeps all of that stuff running but you never have any sort of idea of the data involved in doing such a thing. walking, talking, hearing, seeing, eating, sleeping. they all sound simple but they are not. ask any robot AI expect how hard it is to get something to walk around without falling over.
      computation is great but it's only a part of what the human brain can do. computation is pointless if you can't form a thought about why you would even want/need such a thing.
      lastly our brains are very power effective compared to a computer. it's kinda amazing.

    • @HoloDaWisewolf
      @HoloDaWisewolf 2 роки тому +2

      Is a bat's brain beyond a human brain just because it can process sound through echolocation? Just like it's be impossible for a human to make sense of trillions of 0s and 1s, our brain is simply not wired for being a sonar. Just because computers are better than humans for some specific tasks, I wouldn't say they're "beyond Human" in their ability to adapt. Not to mention that the definition of "interpreting a raw computer stream" is somewhat loose.

    • @anywallsocket
      @anywallsocket 2 роки тому

      @@manzell literally every cognitive bias

  • @ConceptsMadeEasyByAli
    @ConceptsMadeEasyByAli 2 роки тому +14

    *Sweats nervously* This thing can is amazing. It means if we can replicate the functionality in robotics, it will make the hardware work with faulty camera, rotor, angles etc.

    • @SwervingLemon
      @SwervingLemon 2 роки тому +8

      If we're smart enough to use some sort of conditional warnings, so they don't just continue performing until they're in complete ruins, it could allow some flexibility in maintenance.
      AI in production line bots... I just have a scene in my head of somebody trying to take one offline and it throwing a tantrum. "But I wanna solder! Nooooooooo!"

    • @firecatflameking
      @firecatflameking 2 роки тому +1

      @@SwervingLemon 😂

  • @annaclarafenyo8185
    @annaclarafenyo8185 2 роки тому +22

    This is not an appropriate comparison, as human visual networks have continuity, location markers, straight line detection, and 3-d geometry inference built in, which means they aren't going to be any good in a permutation invariant challenge. If you trained an AI to do the human things, lines, geometry inference, and so on, it would be equally bad at permutation challenges.

  • @catcatcatcatcatcatcatcatcatca
    @catcatcatcatcatcatcatcatcatca 2 роки тому +4

    I notice that in both pong and the racing game much of the important data is about angles relative to the player view, which is unaffected by the shuffle resistant to randomly blocking some data.
    I wonder if altering the orientation would cause more problems for the AI?

  • @Veptis
    @Veptis 2 роки тому +1

    We usually flatten images into a single vector. Meaning the neural network has to do learn the weights correctly so it can easily understand that locality makes a difference. And even in convolutional models the flattening is usually rows by rows just stacked ontop of each other. Maybe a spiral or Hamiltonian path could do it better.
    Drop out is used for regularization, so of course only giving half the input will easily work.

  • @AbhishekThakur-wl1pl
    @AbhishekThakur-wl1pl 2 роки тому +21

    Now do it with multiple artifacts flying around.

  • @blinded6502
    @blinded6502 2 роки тому +18

    Well, it's not like spinal cord thinks like human brain either. It figures out the map of what our limbs feel, and then this preprocessed info is fed into the brain for easier comprehension.
    So this this human-likeliness test is just silly.

  • @mattp1337
    @mattp1337 2 роки тому +97

    Sort of reinforces Donald Hoffman's thesis that our adaptive understanding of reality--the way we think the world is--likely bears no resemblance to reality. All evolution cared about is whether our cognition kept us alive and playing the game reasonably successfully, which we do.

    • @getsideways7257
      @getsideways7257 2 роки тому +5

      Not to mention that our optical sensor suite is way far from ideal.

    • @anywallsocket
      @anywallsocket 2 роки тому +5

      You should be clear about what you’re suggesting here. Minimizing the difference between our model of the world and the world is baked into natural selection. If you’re suggesting however that our model for an apple doesn’t itself resemble an apple, then quite clearly yes, of course, this isomorphism is unnecessary.
      On the other hand, these AIs are deep learning, meaning they are playing with hyper parameters within their response to given information, which is about relations between components, regardless of overall structure.
      I’m sure there are examples where human minds can do similar things, e.g. abstracting the notion of ‘something coming at you’ no matter the angle, environment, or thing, but it is in no way obvious that’s generally the case.

    • @mattp1337
      @mattp1337 2 роки тому +1

      @@anywallsocket Your assumptions seem reasonable, obvious even...right up until Hoffman demolishes them ¯\_(ツ)_/¯

    • @laykefindley6604
      @laykefindley6604 2 роки тому +5

      If our sense of reality was widely different from the way that it really is, we wouldn't be able to predict it and thus likely die off as other species have. I would say we are actually the best so far at seeing reality for the way it truly is, at least at the scale of humans.

    • @mattp1337
      @mattp1337 2 роки тому +1

      @@laykefindley6604 By all means, go tell Hoffman he's wrong without hearing his argument.

  • @jacquesbroquard
    @jacquesbroquard 2 роки тому +7

    Every single time. My mind is blown. Thanks for putting these together and sharing with the world.

  • @ds920
    @ds920 2 роки тому +1

    Thank you, sir for all your amazing videos!

  • @Abyss-Will
    @Abyss-Will 2 роки тому

    the shuffle thing reminded me of that scene in the matrix where they look at the green letters in the screen and can make sense of it and see the world they represent

  • @neelmehta9092
    @neelmehta9092 2 роки тому +1

    i think this also has to do with how images are being perceived, us humans see them as a summation of all the pixels simulating a moving object but for a machine its only matrices, so just shuffling the inner data values wont make much difference to a machine

  • @jl6723
    @jl6723 2 роки тому

    I think this kind of thing is impressive in terms that shuffling up the data and having it missing can still allow for an AI to act within a system. I could imagine that sort of decision making and algorithm design being useful for say dealing with dirty lens for self-driving cars or analyzing large groups of interconnected statistics on a subject to generate some hypothesis on a related topic.

  • @falnesioghander6929
    @falnesioghander6929 2 роки тому +1

    So this means that the AI treats all small sections of input homogeneously to later piece them together while we are more trained or overfit in relation to cohesion in a bigger picture based on past experience interacting with the world (aside from the task at hand)?

  • @TheMazyProduction
    @TheMazyProduction 2 роки тому +1

    Finally some David Ha appreciation. ❤️

  • @maxwibert
    @maxwibert 2 роки тому +4

    The input shuffling recovery reminds me of a scene from Naruto: Tsunade permutes the terminals of Kabuto's nervous system and he has to relearn how to walk mid-battle

  • @piotrarturklos
    @piotrarturklos 2 роки тому

    This is quite eye opening, because it gives some more intuition about the nature of things that should be possible using a neural network, even though they are impossible for a human.

  • @cherubin7th
    @cherubin7th 2 роки тому +2

    Example 1 shows that it is very different. A human looks at the big pictures and doesn't just analyse tile by tile. So in the later stages, because AI thinks very simple and doesn't care about the relations between tiles, it isn't confused. But humans see the picture holistically and uses prior knowledge to make sense of it, so humans get confused because we are not used to use tiles independently of the context.

  • @bzikarius
    @bzikarius 2 роки тому

    Incredible things are going on!

  • @gorkemvids4839
    @gorkemvids4839 2 роки тому

    So there is a layer which inverts reshufling so top of the network always gets the unchanged correct information right?

  • @OlivioSarikas
    @OlivioSarikas 2 роки тому +2

    1:42 - how did it get behind the blocks, before it has opened the tunnel? Did the AI find a dirty little cheat? ;)

    • @SwervingLemon
      @SwervingLemon 2 роки тому +1

      You can actually do that with a bit of practice, depending on the game.
      Some of them allow the ball to squirt through if you hit the exact corner between a block and the wall.

    • @OlivioSarikas
      @OlivioSarikas 2 роки тому

      @@SwervingLemon Sure, but that would be a glitch, not how it actually works. ;)

  • @evilkidm93b
    @evilkidm93b 2 роки тому +1

    Would have been interesting to see a control experiment, where the background has the same color as the racing track.

    • @TKZprod
      @TKZprod 2 роки тому

      Yes. For pong also, I'd like to see a control experiment where the model has no input at all. If the model fails, it would confirm that it's using "vision" to play, and not only blindly moves to the same spot each game.

  • @dieselguitar1440
    @dieselguitar1440 2 роки тому +1

    You can still tell what angle the ball is moving at and what angle the track is at (and when there's a turn) when it's reshuffled. I'd bet there are some humans out there who could manage. A platformer game, on the other hand, would probably be a lot harder to manage while it is reshuffled. I wonder how an AI would do playing a reshuffled platformer.

  • @wrOngplan3t
    @wrOngplan3t 2 роки тому

    I suppose every pixel (not just blocks) in view could be at a randomized position and it would still adapt with sufficient training. It's just another layer of translating to the real pixel x,y position. Or maybe that wouldn't even be necessary. But my guess is it would be more efficient for generalizing (I'm no expert in this field at all, just speculating).
    Afaik this was sliglty different with on-the-fly adjustment though. Looks impressive from a human standpoint!

  • @RandomGuy-hi2jm
    @RandomGuy-hi2jm 2 роки тому +2

    4:43 seems impossible,
    But for those who know how Deep Learning works, knows how its was done and a simple task for neural networks to do that.

  • @pw1169
    @pw1169 2 роки тому +12

    "No issues whatsoever" - except for the fact the car isn't being driven on the road :D

    • @MiguelAngel-fw4sk
      @MiguelAngel-fw4sk 2 роки тому

      At least it can mantain the car near the road, thats more what a human can do

    • @codetech5598
      @codetech5598 2 роки тому

      But is there a penalty for driving on the grass?

  • @dva_kompota
    @dva_kompota 2 роки тому +1

    +1 for SmarterEveryDay's reversed bicycle :)

  • @tristenrouse8596
    @tristenrouse8596 2 роки тому

    Great thought experiments here. I'm curious what this means for failsafe features on lets say a self-driving car. If for example a video feed is corrupted, would the AI still be able to process it the same?

  • @trainjumper
    @trainjumper 2 роки тому

    I'd argue the road example is still human-playable since angles are well-preserved - it seems that you can quite easily see the angle of the upcoming road by looking at changes in slopes of the road edges in the jumbled version. Still a very impressive performance from an AI

  • @cliffthecrafter
    @cliffthecrafter 2 роки тому

    I don't think you can really compare the time it takes the AI to adapt to reshuffled inputs to the time it took Dustin to learn how to ride the bike. The AI was trained with the inputs constantly reshuffling. Dustin trained with the inputs one way his whole life and had to re-train himself when the inputs switched. When he switched back to a normal bike it took him much less time, and if he trained himself to constantly switch between the two he could probably learn to switch as fast as the AI.

  • @RicardoNapoli
    @RicardoNapoli 2 роки тому

    Woooooooow !!!!! That's insane !!!!!

  • @thomasr1051
    @thomasr1051 2 роки тому

    Amazing as always. This is my source of staying up to date on ever evolving AI technology

  • @AirNeat
    @AirNeat 2 роки тому

    The reason the AI can deal with the shuffling, is because it can read the entire screen at once and easily identify which tiles are the edges. A human has to read each tile and process it slowly. It's simply a processing speed issue. It does think like a human, only faster.

  • @Lttlemoi
    @Lttlemoi 2 роки тому +7

    I disagree with how you compare AI training time with human training time. When you initially learn to ride a bike, it takes some time until you're able to steer precisely and perfectly as well. It seems to me that the slowness with which the human brain adapts to changing situations is only slow compared to the computer, but not that slow when compared to the human brain itself.

    • @Rybz
      @Rybz 2 роки тому +1

      did you just say a snail is fast compared to snails. what does that mean 🤣

    • @Lttlemoi
      @Lttlemoi 2 роки тому +1

      @@Rybz I meant that the human brain adapting is slow compared to the AI adapting, but not necessarily slow when comparing to the initial training time of each. It's like comparing acceleration and top speed of a human with that of a car. The car can accelerate faster and has a higher top speed than a human, but that doesn't mean a human has to take more time before he can accelerate to his top speed than a car requires to accelerate to its top speed.

    • @Rybz
      @Rybz 2 роки тому +1

      @@Lttlemoi But what does that matter if the AI is better than the human brain in this task?

  • @DanFrederiksen
    @DanFrederiksen 2 роки тому +1

    I assume training starts over, otherwise it could of course not handle a reshuffling.

    • @codetech5598
      @codetech5598 2 роки тому +1

      No the training does not start over.

  • @mogarbobac1472
    @mogarbobac1472 2 роки тому +1

    Can someone help me? I dont understand how this would even work.
    The first part where its switching up the streams of data, my guess is it trys a movement and then *snaps that control to the correct data stream (from experience). If you did that of course it would be able to quickly fix itself. This is unlike a human where our structures are permanent and we are literally forcefully training to work against a remembered system.
    But the 2nd permutations thing makes absolutely no sense to me. Even if it did the same thing as before after finding similar locations, it would be extrememly difficult to determine not only where the ball is but also the paddle unless you straight up remembered where it was heading and you could literally predict the entirety of the game ahead of time.
    ANY help with an explanation would be helpful

  • @1.4142
    @1.4142 2 роки тому

    Some people even temporarily forget how to ride a normal bike after learning the backwards steering bike.

  • @jawadmansoor6064
    @jawadmansoor6064 2 роки тому

    computation is fundamentally different from thinking. And COMPUTEr can COMPUTE faster than we can so ... no competition there.

  • @weaseloption
    @weaseloption 2 роки тому

    Wow, what a time to be alive

  • @xxiemeciel
    @xxiemeciel 2 роки тому

    very interesting, did the algorithm had to redo its training each time the complexity was increased (shuffling and removing parts) ?

  • @HD-Grand-Scheme-Unfolds
    @HD-Grand-Scheme-Unfolds 2 роки тому

    Do they use Convolution Nets or Capsules? Or if I should instead ask how do they achieve pixle permutation invariance. This is blowing my mind to know that they retrain the networks.

  • @yimingqu2403
    @yimingqu2403 2 роки тому

    A small question, in the ping-pong game, why does the agent always succeeds by aiming at the corner?

  • @MrTomyCJ
    @MrTomyCJ 2 роки тому +3

    I watched these videos to get a glimpse on how these amazing results are achieved, not only to see the incredible results. I think part of it is being lost lately...

  • @FungIsSquish
    @FungIsSquish 2 роки тому

    Can’t wait for these things to be able to play Mario kaizo

  • @DimiShimi
    @DimiShimi 2 роки тому +1

    I think this rather shows that the neural network is functioning like a brain, but unlike a human brain it's a purpose-trained brain that solves one particular kind of problem, not many. We know that in some rare cases people are capable of superhuman mental feats (sometimes while being deficient in some other common human ability). My theory is that in these cases brain resources are allocated in an unusual way, so we get results more comparable to a purpose-built neural network. - I might be wildly off. This is pure speculation.

  • @MrProfizmus
    @MrProfizmus 2 роки тому

    It doesn't think like a human does if you remove it far enough from what you'd consider a human-accessible example. Makes sense if you ask me. This is literally demonstrating what happens if you approach UX with the opposing intentions. Another commenter also addressed the cost of re-arranging your thoughts, which is a great point. For the AI, it comes free, as it's part of it's training lifecycle.
    I guess it is a good reinforcement of the idea of how unspecific it really is when we say "human thinking". I remember seeing a small Q&A where Feynman also addressed this, using planes vs birds as an example. It's only really mindblowing if you look at it from a humane perspective, less so from a theoretical one.

  • @justinwhite2725
    @justinwhite2725 2 роки тому

    Reminds me of the experiment where people wore goggles that made the world upside down. Wore them for a week.
    Confused them at first then they adjusted. When they were taken off they were confused again but then readjusted.

  • @Paruthi.618
    @Paruthi.618 2 роки тому

    Yea .. what a time to be alive

  • @JMPDev
    @JMPDev 2 роки тому

    @Károly Zsolnai-Fehér: The graphics from 3:27 to 6:26 in this video are surprisingly poor quality. It looks like it was bilinearly upscaled from a really low resolution source, and seems to have additional compression artifacts too. If the source data was really that low res, nearest neighbor or integer upscaling would really have been preferable. Was this a compositing/editing error? It looks awful, even at at the max 2160p60 :/

  • @tctrainconstruct2592
    @tctrainconstruct2592 2 роки тому

    instead of just shuffling the blocks, it should rotate them too
    because a human player trained for this could technically keep in mind the ball's position and velocity, as well as the paddle, really easily
    btw for the racing game example, humans can play games BLINDFOLDED: no sight whatsoever! So it is normal that the bot realigns itself with the information it sees

  • @chounoki
    @chounoki 2 роки тому +2

    This means the AI has actually been trained to work on a higher level of abstraction, dumping everything that is not directly related to the judgement of action, which is basically a trait of born geniuses if it were a person.

  • @SudiptoChandraDipu
    @SudiptoChandraDipu 2 роки тому +1

    We humans see images differently. We focus on speicific point more, our focus/attention gradually decrease surrounding that point. But AI gives entire image 100% focus. What would happen if we introduce this characterics to the AI? Will it behave the same like humans then?

  • @dietrevich
    @dietrevich 2 роки тому

    I think people are missing the point that the computer abstracts the data differently than we do. It can do away with lots of things that to us are necessary and that without it wouldnt make sense, but that to the program it mathematically pans out to a solution.

  • @CandidDate
    @CandidDate 2 роки тому

    I held my papers so tight, they turned back into a tree.

  • @venjsystems
    @venjsystems 2 роки тому

    amazing paper

  • @benjabkn12
    @benjabkn12 2 роки тому

    6:18 the minus ones just keep on coming and they don't stop coming

  • @1xdreykex1
    @1xdreykex1 2 роки тому +1

    The way humans think is more or less flawed so maybe we can design ai’s that think deeper than we can

    • @geli95us
      @geli95us 2 роки тому

      I wouldn't call it flawed, more like "greedy" (in the computer sense of the word), humans make a lot of assumptions and optimizations to quickly be able to process the world around them, and that's usually helpful, computers are, in a sense, more polyvalent in that they can be completely reprogrammed very easily, and that makes me hopeful of what we'll be able to achieve with them once they become more powerful, but right now, and until computers have as much computing power as a human brain and more (we're a ways off that), AI will have to make assumptions and optimizations too, just not the same ones as we do

  • @Craxin01
    @Craxin01 2 роки тому

    I learned about the laxative effect of sorbitol a long time ago. I don't chew gum anymore.

  • @mello7992
    @mello7992 2 роки тому

    what a time to be alive

  • @phpn99
    @phpn99 2 роки тому +1

    Let's be clear : There is no "thinking" in machines. There is gradient descent. Period.

  • @erlendpowell7446
    @erlendpowell7446 2 роки тому

    I think this nicely illustrates why artifical networks such as this tend to generalise poorly. The human brain tends to preserve the spatial locality of information, while the typically fully connected ANN throw most of this locality away. The problem with throwing this locality away, though, is that there's a ton of information that is lost as a result. This also illustrates why ANN's are prone to adversarial attacks - without this spatial information, features can easily be hidden in what looks like noise to a brain.

  • @deadpianist7494
    @deadpianist7494 2 роки тому +2

    Hi, i hope yall doing good, have a good day.

  • @edoardoschnell
    @edoardoschnell 2 роки тому

    All this has utility, indeed.

  • @Antonio_Vizcarra
    @Antonio_Vizcarra 2 роки тому

    Ok now i'm scared i've seen sci-fi movies and i know how this ends

  • @amortalbeing
    @amortalbeing 2 роки тому

    Gorgeous Papers, thanks doc
    by the way why dont you always talk the same way you talk in the advertisements?, sometimes it seems you are whispering !

  • @wibiyoutube6173
    @wibiyoutube6173 2 роки тому

    "What a scary time to be alive."

  • @DonVitoCS2workshop
    @DonVitoCS2workshop 2 роки тому

    Would it be possible for this AI to play(learn) pong at the same time as driving(learning) the racecar on the new background? That would be amazing and clear evidence it's better/different than the human brain

  • @juhotuho10
    @juhotuho10 2 роки тому

    Machine learning algorithms are amazing for specialized environments and tasks but lack the things like able to do game theory, people are very good at handling generalized work and people can be amazing at game theory and theorizing solutions to problems

  • @languagew9577
    @languagew9577 2 роки тому

    can you tell me any AI that can duplicate cloths of character from image ? and cam make in 3D ?

  • @perhlom
    @perhlom 2 роки тому

    I don't know about the -1's for the two last examples. To a neural network, the pixels are arbitrary. If you shuffle them, it just trains on that. The strict X-Y coordinate space is just our preferred way to see the pixels, but there's no privileged coordinate space to understand the pixels in. If all of us had grown up wearing cell-swapping glasses that remapped parts of our vision to other areas of our retinas, our brains would just have trained on that and formed a cohesive image anyway. Then we'd just not be able to see without those glasses. The neural network doesn't care about the X-Y layout, and I don't think we strictly do either. Inside the neural network, and in our minds, the image doesn't have a coordinate space at all.

  • @zaneg
    @zaneg 2 роки тому

    I am not sure this means they don’t think like us. We hen we look at a screen we see it completely differently then a neural network does. We can only see one part of the screen at a time while the neural network is focused on the entire screen all at once.

  • @antivanti
    @antivanti 2 роки тому +1

    The pong game seemed to play out exactly the same every point so it's possible the AI didn't even need to see the game to play just repeat the exact same input every time

  • @donaldduck830
    @donaldduck830 2 роки тому +3

    And I recently saw a report in how machine learning failed: They tried to identify wolves. The pictures of wolves in the data for learning were all in snow. The machine learning algrorithm looked only at a small patch of pixels in the corner, if snow, then wolf, else dog. Ended in fail with different pictures. Be careful with your enthusiasm. In the end a computer algorithm will "see" your body after a crash on the road and say to itself "trash".

    • @geli95us
      @geli95us 2 роки тому +2

      In the end, it's all about complexity, the more information, the more processing power, the more memory it has, it will be able to look at more complex patterns, it's not like humans are any different in that sense, in fact, if you grab a human that doesn't know what snow/dogs/wolfs are and do this to them, I'm pretty much sure they will reach the same conclusion as the computer.
      the way I see it, getting computers to think like humans once we have computers that are as powerful as a human brain, will be pretty much trivial, seeing the things we are doing with them having fractions of that

    • @donaldduck830
      @donaldduck830 2 роки тому

      @@geli95us In the end I had a perfectly functional table calculations program (better than almost all versions of excel) and a word processor (better than all versions of word) on the very first pc I had access to (it was my Dad's around 1990, late '80s).
      And even then the adage "garbage in, garbage out" held true. And if you program a woke AI, it will put garbage out no matter what the input.
      Like Michael Mann's hockey stick graph: No matter the input, his algo always spit out the same graph.
      Thus my "less enthusiasm, more care" to guard against the Terminator-Matrix events timelines. Oh, and even if the machines do not become self-aware, somebody who controls the machines might be evil. Or the person listening to the machine might be stupid: Like the countless people who got stranded in the wilderness after listening to their Nav.com. that said "turn right on the next" and misunderstood what the next right tuen was.
      Btw, saw it happen, heard about it happening and happened to me myself, so not dissing anybody, just saying to be careful and think for yourself.

    • @SwervingLemon
      @SwervingLemon 2 роки тому

      @@geli95us I think their greatest value is in how they DON'T think like us. They've arrived at novel solutions to challenges that I don't think would have occurred to a human.

    • @MushookieMan
      @MushookieMan 2 роки тому +1

      This channel is always overhyping "AI", or more accurately, unintelligent machine learning. It is very useful, but not 'artificial intelligence'.

    • @ddjoray1042
      @ddjoray1042 2 роки тому +1

      If a person with no general knowledge was only given training photos of wolves in snow, they would also think that the only difference between dogs and wolves was that wolves are dogs in snow. Your example seems to be a failure in giving the algorithm adequate training data.

  • @DJZofPCB
    @DJZofPCB 2 роки тому

    Does anyone remember Eliza? Guess the animal? In DOS? 1960's same concept but the AI asked questions to figure out your animal. If it got it wrong, it asked What question should I ask next time;) ps Eliza was a chatbot therapist

  • @MatthiasPitscher
    @MatthiasPitscher 2 роки тому

    Are they actually using the same model? Or just the same architecture trained on the randomized input?
    Is there some kind of transfer learning happening?

  • @mattomanx77
    @mattomanx77 2 роки тому

    I'd actually argue the last few examples are still in favor of human-like learning. The difference is key, humans are limited by their physical brain, and not having to do just one task but dozens upon dozens of tasks and constantly learn new ones. The AI only needs to run the one. One thing that's also noticeable is examples of people who's senses become 'shuffled' (I.e. mirror-goggles) adapt faster the second time (Taking the goggles off after adapting)
    It's likely that with enough practice, someone's vision could be cubed and jumbled every day and they'd adapt quickly after the first year or two.
    We don't do that as quickly as AI because that's prohibitively energy expensive, and we are limited by the speed at which our brain can physically build and adjust new and old pathways, which were made without accounting for complete visual input shuffling.
    Even when it comes to missing information, our brain finds ways to fill in the gaps. You ever heard of the blind spots in your eyes? All that was done in the second '-1' was give the AI blind spots... And it adjusted. Just like humans.
    People can lose an eye, people's vision can get blurry, some people can just be born colorblind, but in every case even though they are inhibited most can still do many of their everyday tasks.
    As for the third example of introducing unnecessary information, again, there are human-like analogs. In fact this is probably the most human-like of the three -1s. Think of how we listen to each other when speaking. Our environments are often noisy, and contain an incredible amount of varied backgrounds. Even still, we can hear and identify many noises and even speech with great accuracy.
    This leads back to what I started with about AIs only having to work one job here when we work many. Like the AI, we don't have many varied tasks involving our ears, most boil down to identifying and putting together noises.
    The same could be said for the background in visual information as well, come to think of it! If I changed your desktop background to a house, you wouldn't be suddenly unable to use your computer. If I put you in the woods instead of a town, you wouldn't forget how to walk forward or follow a path. We were built to hunt the relevant information and discard the unnecessary.
    All in all, we can be missing information, have it jumbled up, and various background oddities thrown in at almost random, but as humans we adapt almost instantly because it was just part of our lives, as important a role language has played in our evolution.
    You may say that the last 3 were proof against AI thinking like humans, but I think the last 3 were perhaps the strongest evidence of all in favor of AI thinking like humans.
    (Sorry for rambling, I tend to be bad with putting my thoughts into words, especially in long form)

  • @mishafinadorin8049
    @mishafinadorin8049 2 роки тому

    If I remember correctly, infants can recognize the face of their mother even if the picture is shuffled, but this ability disappears early during development. In this sense these neural networks are similar to neural networks of newborn children.

  • @BHFJohnny
    @BHFJohnny 2 роки тому

    wow. that't amazing. And also scary

  • @SuperXzm
    @SuperXzm 2 роки тому +1

    Does machine think the way we think? Or do we think the way machine works?

  • @packediceisthebestminecraf9007
    @packediceisthebestminecraf9007 2 роки тому

    I think it's better to compare the AI adapting to new input to the way humans can adapt to a different keyboard layout, because most humans have just never learned to get used to constantly changing input from their eyes.

  • @andytroo
    @andytroo 2 роки тому

    is it possible that the steering AI is simply going of 'if i see a left curb, turn left; if i see a right curb turn right" 6:54 shows the car driving of the road without an attempt to return.

  • @LKDesign
    @LKDesign 2 роки тому

    This is fine.🔥

  • @yugecheng8941
    @yugecheng8941 2 роки тому +2

    Why that sumarai cutting fruit video was deleted?

  • @harleygarrett8063
    @harleygarrett8063 2 роки тому

    I feel like splitting and shuffling an image is the equivalent torture for neural networks that Boston Dynamics inflicts upon Atlas and co

    • @SineN0mine3
      @SineN0mine3 2 роки тому

      Why else would we program them to feel pain?

  • @JohnyCilohokla
    @JohnyCilohokla 2 роки тому +4

    it appears that all of those "shuffled" image are effectively unused, the pong example just repeats the same thing over and over again (which you could compare to muscle memory, the image is irrelevant, a human can easily do that too), as for the racing example (if it's even random enough to not just be memorized) it can be easily explained by simple pattern recognition on the red-white part of the road, it doesn't really matter at what part of the screen it appears at, also helps that it's so zoomed in as that pattern will appear only when it's time to act on it, something a human could easily do as well even with a "shuffled" image sure it would take few days to learn it but it's definitely not impossible

    • @Monoffel
      @Monoffel 2 роки тому +1

      I'm aware of the reverse bike and upside down glasses for humans, and that it takes a while to adjust. But if a human shuffles often enough they may also learn to adapt to shuffling very quickly. On a similar note, a human may learn to adapt to shuffling without retraining if the shuffling is done from the start. It would definitely be interesting to really put humans to the test with some of these examples. Just make them try to learn these shuffled situations for 120 hours or more and maybe we're not that different to the AI

    • @HoD999x
      @HoD999x 2 роки тому +1

      Yes, I wanted to point out the same thing. The pong AI lost 0 to 14 repeating the exact same moves.

  • @coolbanda5446
    @coolbanda5446 2 роки тому

    It's easier for machines to do this because this was what machines were made for xD
    Not surprised. I'm more interested in its applications, what else can it optimize? That would be beneficial to us

  • @DustinRodriguez1_0
    @DustinRodriguez1_0 2 роки тому

    The fact you can remove large amounts of the input and still have the system know what to produce as output makes me wonder whether concepts close to this could be used to accomplish lossy data compression. Like could you have a player application that is loaded with such a network, pre-trained on general video/audio, which would only require a very small amount of data to be able to reconstitute a good enough approximation of the source material? I imagine the size of the network and its weights would need to be known, I don't have any familiarity with just how big these networks are... are we talking megabytes of weight data, or gigabytes or more?