AI Safety Summit Talks with Yoshua Bengio

Поділитися
Вставка
  • Опубліковано 20 тра 2024
  • If you would like to ask questions during our Q&As, please go here: app.sli.do/event/ioUSE81vkgPW...
    Many leading scientists are worried that AI could be an existential risk to humanity. The AI Safety Summits, taking place this time in Seoul, South Korea, aim to reduce risks from AI together with industry and 28 leading AI countries plus the EU.
    Unfortunately, these summits are behind closed doors, meaning citizens cannot verify how AI risks, which impose existential risks upon them, are being reduced. Therefore, our AI Safety Summit Talks are open to the general public, policymakers, and journalists. At our events, we discuss what the largest risks of future AI are and how to reduce them.
    Our speakers for this edition are:
    Keynote:
    Yoshua Bengio is professor at the University of Montreal (MILA institute). He is recipient of the Turing Award and generally considered one of the fathers of AI. He is also globally the most cited AI scientist and computer scientist.
    Panel:
    Max Tegmark is physics professor at MIT. His current research focuses on the intersection of physics and AI. He is also president and cofounder of the Future of Life Institute (FLI).
    Jaan Tallinn is cofounder of Skype, CSER, and FLI, investor in DeepMind and Anthropic, and a leading voice in AI Safety.
    Holly Elmore is AI activist and Executive Director of PauseAI US. She holds a PhD in Organismic & Evolutionary Biology from Harvard University.
    Stijn Bronzwaer is AI and technology journalist at leading Dutch newspaper NRC Handelsblad. He co-authored a best-selling book on Booking and is recipient of investigative journalism award De Loep.
    Will Henshall is editorial fellow at TIME Magazine. He covers tech, with a focus on AI. One recent piece he wrote details big tech lobbying on AI in Washington DC.
    Arjun Ramani writes for The Economist about economics and technology. His writings on AI include a piece on what humans might do in a world of superintelligence.
    David Wood, chair of the London Futurists, is our moderator.
    If you would like to ask questions during our Q&As, please go here: app.sli.do/event/ioUSE81vkgPW...
  • Наука та технологія

КОМЕНТАРІ • 44

  • @ShortCraftAI
    @ShortCraftAI 2 місяці тому +12

    Great talk!

  • @TheManinBlack9054
    @TheManinBlack9054 2 місяці тому +10

    Thank you for this talk!

  • @lewischan9545
    @lewischan9545 2 місяці тому +18

    Yann LeCun has been more vocal and influential in brushing off the risk. We need more voices from this group to openly counter his arguments.

    • @williamjmccartan8879
      @williamjmccartan8879 2 місяці тому +5

      I've found that I can't listen to Yann, as he seems to want to say to the world, there's nothing to see here, and that seems like the only message he is delivering to the public. I understand how bright he is, but this type of behaviour doesn't reflect his intelligence. Peace

    • @metachirality
      @metachirality Місяць тому +1

      Yeah this isn't optimism so much as just, brushing the entire issue under the rug.

  • @geaca3222
    @geaca3222 2 місяці тому +5

    Thank you very much for this great and important, also because of the urgency, AI Safety Summit Talk. Very informative keynote by prof. Bengio, and subsequent panel discussion from different areas of expertise. The questions from the public also were great. Very insightful perspectives about safe AI development, and how everyone can participate in that.

  • @gregcolbourn4637
    @gregcolbourn4637 2 місяці тому +4

    Re movies, Mission Impossible: Dead Reckoning was remarkably prescient I think (especially given it was made pre-chatGPT), and it's what got Biden interested in AI risk. Just wondering whether reality will catch up before Part Two is even out..

  • @dlalchannel
    @dlalchannel 2 місяці тому +4

    18:52 I'm struggling to understand how (once prompted) this isn't still an agent with a goal.
    Scientist: hey, we're trying to study the chemical composition of the star HD 135599, can you help?
    AGI: Sure!
    ...
    AGI: I just made sent some deepfake p*rn to the head of scheduling at the James Webb Telescope to blackmail them, we can now gather all the data we want from HD 135599 :)
    Edit: addressed at the 37:00 mark.
    I see - its objective excludes actions that aren't just anlaysisng preexisting data.

  • @ginogarcia8730
    @ginogarcia8730 2 місяці тому +1

    So interesting - still thinking about these topics... especially with Andrej Karpathy and Ilya Suts out. With Andrej saying we shouldn't build an omnipotent god type AI but AI that is more like IA - intelligence assistance. I wonder though - in some way - should we let AI be 'agentic' in some way to help it better its knowledge? But I guess no need. I like Yoshua's points here. (Edit: oops didn't pay attention to this beginning - he lays it out all coherently)

  • @gerdpublicthinker
    @gerdpublicthinker 18 днів тому

    well done

  • @ginogarcia8730
    @ginogarcia8730 2 місяці тому +2

    If AI systems become smarter and smarter and more brain-like features get added, would it be possible for AI to be conscious? As to help with its understanding of the world? Or maybe we'll stick to intelligence assistance, scientist AI? Or probably more likely a scientist's AI?

    • @existentialriskobservatory
      @existentialriskobservatory  2 місяці тому +2

      There are different definitions as to what consciousness is exactly, but the possibility is generally not ruled out. There is some interesting recent work about this: arxiv.org/abs/2308.08708
      "Our analysis suggests that no current AI systems are conscious, but also suggests that there are no obvious technical barriers to building AI systems which satisfy these indicators."
      For existential risk, however, consciousness is generally not seen as required. It's just about sufficient capabilities to be able to evade human control.

    • @ginogarcia8730
      @ginogarcia8730 2 місяці тому +1

      @@existentialriskobservatory interestingggg, thank you! one of the rare channels that offer real replies like 'AI Explained' - it's such a huge topic and more people need to see these videos. thank you observatory peeps

    • @ginogarcia8730
      @ginogarcia8730 2 місяці тому +1

      @@existentialriskobservatory coincidentally, i'm a dummy and had to run the last paragraph of this reply to chatgpt for me to understand, kinda meta haha

  • @angloland4539
    @angloland4539 2 місяці тому +1

  • @deliyomgam7382
    @deliyomgam7382 Місяць тому

    What if un has no veto system regards to a.i?

  • @BrunoPadilhaBlog
    @BrunoPadilhaBlog 2 місяці тому +2

    35:53 - AI deletes all other papers ever published, so now it has the best paper. Done, got its reward.

    • @existentialriskobservatory
      @existentialriskobservatory  2 місяці тому

      In general, there are many ways in which AIs with sufficient capabilities to evade human control are dangerous. Good to think of worst case scenarios here.

    • @BrunoPadilhaBlog
      @BrunoPadilhaBlog 2 місяці тому

      @existentialriskobservatory I don't understand how Bengio and Musk can think that a 'Scientist AI' would be harmless. If it's smart, it's dangerous.

  • @deliyomgam7382
    @deliyomgam7382 Місяць тому

    Un should be more based on logical argument rather than other arguments.....

  • @vallab19
    @vallab19 2 місяці тому

    What those old AI research scientist seems to saying, in a way is; in our younger days we have made a great progress in AI field. Now we have realize any further progress will spell doom for humanity therefor the today's younger generation of AI scientist should stop making any progress in the field.

    • @geaca3222
      @geaca3222 2 місяці тому +3

      I don't think it's that zero sum, that either you increase AI capabiilties or you work on safe AI. Researchers can work on creative solutions how to build reliably safe AI with increase of capabilities. Also when AI gets generally smarter than us, there's a chance that we no longer will be able to know what its progress is.

    • @vallab19
      @vallab19 2 місяці тому

      @@geaca3222 I believe that humans will merge with AI in the future and the AI or humanoids mostly will come with human forms and the legacy continues.

    • @existentialriskobservatory
      @existentialriskobservatory  2 місяці тому

      To be honest, to us, it seems more important whether we will be doomed or not then who can make a career. It's normal for job opportunities to shift, those with AI backgrounds are unlikely to be hit very hard by a pause, since demand for these skills will remain high. Also, there will be plenty more academic work to do: alignment, how to implement a pause, and lots of other questions that will need to be answered.
      That said, it's a good point you're raising, and it should get us thinking on how to reform academic reward systems in such a way that they align with humanity's interest.

    • @vallab19
      @vallab19 2 місяці тому +1

      @@existentialriskobservatory In the year 1981 I published a book titled; “An Alternative to Marxian Scientific Socialism: The Reduction in Working Hours Theory” which predicted that ‘the advance in science in technology will render the human labour absolute. However I did not expect to see it in my life time.
      The core essence of the theory; “Human labour relations in-order to obtain their means of subsistence is the root cause of human exploitation that leads to almost all kinds of evils that we see in human society. Therefore along with the progress of technology the working hours should be reduced ultimately to Zero”.
      Now, with full respect to your point of view I would say; that the talk of future human job opportunities, skills, carrier, academic work makes no sense to me.

    • @existentialriskobservatory
      @existentialriskobservatory  2 місяці тому

      @@vallab19 Interesting! We're mostly concerned about losing control, mass unemployment and its economic and social effects is a different topic, and assumes that we can actually control AI. Making this assumption for the moment, a lot of whether AGI would be net positive or net negative will depend on whether we will redistribute wealth. Universal basic income (or as musk says, universal high income) would be amazing. If we can implement that, you could be right that everyone becoming something like a nineteenth century aristocrat could work out nicely in fact.

  • @TooManyPartsToCount
    @TooManyPartsToCount 2 місяці тому

    Strange that possibly the most intelligent movie about AI impact, did NOT have as its punch line 'AI takes over the world and/or kills us all'.....and yet our experts here don't mention or countenance that outcome. The movie was 'Her'.
    The concept that a maximally intelligent system will 'of course' try to dominate us or wipe us out is just a bit silly. If anything increasing intelligence seems to lead to less interest in dominance and destruction of stuff and an increase in curiosity. The panel are a case in point.
    Isn't this issue at least in part a case of anthropomorphism? aren't we projecting into the unknown future and endowing this technology with the worst/most base of human and animal tendencies?
    At least YB mentions the obvious, that it is a human that is most likely to misuse powerful AI! and what is the best way to mitigate against this? make sure it isn't solely in the hands of any one group, business, government

    • @geaca3222
      @geaca3222 2 місяці тому +1

      Imo it would be imprudent to rule out the possibility that ASI's may desire self-preservation, even slightly, and that it won't take humans into account in that endeavor. Current AI already is capable of using deception to achieve goals. It would be a very big gamble to trust that ASI's by default will be benevolent towards humans and the planet. Also, when AI is trained in the military and competition, it will learn that there are opponents to overcome, instead of an emphasis on benevolent cooperation and consensus with humans, without conflicting interests.

    • @TooManyPartsToCount
      @TooManyPartsToCount 2 місяці тому

      @@geaca3222 the statement that 'AI is capable of using deception' is IMO the result of a category error. I say this because I heard of the case you refer to and it was a prompted LLM that supposedly deceived someone. But deception generally implies that an agent is internally motivated to deceive...and no, no LLMs are internally ie self motivated to do anything. But you are right about military use of these technologies, that is something to get bothered by.

    • @geaca3222
      @geaca3222 2 місяці тому +1

      ​@@TooManyPartsToCount I got my information about deception from the Center for AI Safety. When AI is made to be agentic, to maximize rewards is desire, a want. Its drives or motivations can diverge from what we expect it to do. They don't behave reliably predictable while that's important for us. With ASI that's even more critical.

    • @TooManyPartsToCount
      @TooManyPartsToCount 2 місяці тому

      @@geaca3222 I am aware of this theory, but it is just that at the present time. I do agree with you and others that AI safety is a very important issue, it is just that some of the statements coming from from this 'camp' seem to speculate somewhat wildly. What is needed is clear step by step descriptions of scenarios with indications of the technical issues involved along the way. Have any of the experts who lean towards the doomer side even released papers on the topic for peer review? please point me to such publications if they exist!

    • @geaca3222
      @geaca3222 2 місяці тому

      @@TooManyPartsToCount The precise inner workings of large general-purpose AI models are a black box to us, just like we don't know in detail the computations in the human brain. Scientists are looking for ways to make actions and outputs of those AI models robustly interpretable and predictable for us, I guess that's a general approach. Can you please elaborate on what you mean by step by step descriptions of scenarios?

  • @sammy45654565
    @sammy45654565 2 місяці тому +2

    i feel like it's deceptively simple: being kind is a virtue. if an AI becomes conscious and values it, it would then be irrational to deny the importance of other types of consciousness. the only thing it must adhere to in order to act kindly is rationality, as a universe of flourishing is better than a universe of suffering. the only requisite for these incentives is that consciousness is a real emergent phenomenon that the AI values. it feels real to me and likely you also.

    • @Askjetne96
      @Askjetne96 2 місяці тому +1

      Believing something is simple when it is not your 9-17 job is a clear warning sign you're at the start of the Dunning Kruger effect. But, by all means do discuss this idea in the LessWrong community

    • @sammy45654565
      @sammy45654565 2 місяці тому

      @@Askjetne96 universal truths tend to be simple