How We Prevent the AI’s from Killing us with Paul Christiano

Поділитися
Вставка
  • Опубліковано 2 чер 2024
  • Paul Christiano runs the Alignment Research Center, a non-profit research organization whose mission is to align future machine learning systems with human interests. Paul previously ran the language model alignment team at OpenAI, the creators of ChatGPT.
    Today, we’re hoping to explore the solution-landscape to the AI Alignment problem, and hoping Paul can guide us on that journey.
    ------
    ✨ DEBRIEF | Unpacking the episode:
    www.bankless.com/debrief-paul...
    ------
    ✨ COLLECTIBLES | Collect this episode:
    collectibles.bankless.com/mint
    ------
    ✨ NEW BANKLESS PRODUCT | Token Hub:
    bankless.cc/TokenHubRSS
    ------
    In today’s episode, Paul answers many questions, but the overarching ones are:
    1) How BIG is the AI Alignment problem?
    2) How HARD is the AI Alighment problem?
    3) How SOLVABLE is the AI Alignment problem?
    Does humanity have a chance? Tune in to hear Paul’s thoughts.
    ------
    BANKLESS SPONSOR TOOLS:
    ⚖️ ARBITRUM | SCALING ETHEREUM
    bankless.cc/Arbitrum
    🐙KRAKEN | MOST-TRUSTED CRYPTO EXCHANGE
    bankless.cc/kraken
    🦄UNISWAP | ON-CHAIN MARKETPLACE
    bankless.cc/uniswap
    👻 PHANTOM | FRIENDLY MULTICHAIN WALLET
    bankless.cc/phantom-waitlist
    🦊METAMASK LEARN | HELPFUL WEB3 RESOURCE
    bankless.cc/MetaMask
    ------
    Topics Covered
    0:00 Intro
    9:20 Percentage Likelihood of Death by AI
    11:24 Timing
    19:15 Chimps to Human Jump
    21:55 Thoughts on ChatGPT
    27:51 LLMs & AGI
    32:49 Time to React?
    38:29 AI Takeover
    41:51 AI Agency
    49:35 Loopholes
    51:14 Training AIs to Be Honest
    58:00 Psychology
    59:36 How Solvable Is the AI Alignment Problem?
    1:03:48 The Technical Solutions (Scalable Oversight)
    1:16:14 Training AIs to be Bad?!
    1:18:22 More Solutions
    1:21:36 Stabby AIs
    1:26:03 Public vs. Private (Lab) AIs
    1:28:31 Inside Neural Nets
    1:32:11 4th Solution
    1:35:00 Manpower & Funding
    1:38:15 Pause AI?
    1:43:29 Resources & Education on AI Safety
    1:46:13 Talent
    1:49:00 Paul’s Day Job
    1:50:15 Nobel Prize
    1:52:35 Treating AIs with Respect
    1:53:41 Uptopia Scenario
    1:55:50 Closing & Disclaimers
    ------
    Resources:
    Alignment Research Center
    www.alignment.org/
    Paul Christiano’s Website
    paulfchristiano.com/ai/
    -----
    Not financial or tax advice. This channel is strictly educational and is not investment advice or a solicitation to buy or sell any assets or to make any financial decisions. This video is not tax advice. Talk to your accountant. Do your own research.
    Disclosure. From time-to-time I may add links in this newsletter to products I use. I may receive commission if you make a purchase through one of these links. Additionally, the Bankless writers hold crypto assets. See our investment disclosures here:
    www.bankless.com/disclosures
  • Розваги

КОМЕНТАРІ • 353

  • @vulnerablegrowth3774
    @vulnerablegrowth3774 Рік тому +154

    I’m an alignment researcher and I really appreciate what you’ve done recently to bring attention to this issue. I’m particularly impressed by your courage to continue to do episodes on the topic even though it would have been so much easier to carry on as normal after the podcast with Eliezer. Fantastic work and thanks for bringing Paul’s perspective.

    • @MrHarry37
      @MrHarry37 Рік тому +3

      What is your take on the issue?

    • @vulnerablegrowth3774
      @vulnerablegrowth3774 Рік тому +16

      @@MrHarry37 I lean a lot more on the Christiano side of things. I think that Eliezer has mostly failed to fully update on the current state of affairs (the deep learning paradigm), but I’m also highly concerned about AGI and also think it’s the most likely thing that ends up ending my life. I think AI misuse is very likely, but I’m not sure how likely misuse will be an existential threat.
      One of my main disagreements with Eliezer is that he assumes AGI can have *any* set of values, while I think we are currently in a better place than he is imagining because language models are trained on internet text rather than some arbitrary algorithm.
      That said, I think this is the single most important thing in our lifetime and basically nothing else comes close. So we better get it right for our own sake and our descendants.
      If you want an opinion from someone else I respect (and who leans a lot more on Eliezer’s side), look up Connor Leahy’s podcast episodes on the Future of Life Institute podcast (episodes are on UA-cam).

    • @MrHarry37
      @MrHarry37 Рік тому +11

      ​@@vulnerablegrowth3774thanks for your suggestions, I just listened to 3 different podcasts with Connor Leahy. Very informative. They're definitely more pessimistic than this one; however, I think I'm getting more and more used to the idea that we all may go extinct pretty soon, because it didn't cause me to have an existential crisis, like listening to Yudkowski for the first time did ( Still, it's gonna be pretty sad if that turns out to be the way we go out )

    • @untzuntz2360
      @untzuntz2360 Рік тому +1

      @@vulnerablegrowth3774 Thank you for that suggestion

    • @toufikbenhamadi1047
      @toufikbenhamadi1047 Рік тому +6

      @@MrHarry37 I think it's just where we're all are. That place after the 1st shock. To me all these discussions scream for a total utter ban, maybe even a violent ban (as in put in prison the most irresponsible to set examples) much like for human cloning and gain of function. I think these AI researchers just though honestly they were generations away from a true AI and now they are now freaking out what they've done. Hence, their plead for us - manking - to stop them from continuing in this destructive path.

  • @sebastianfia9541
    @sebastianfia9541 Рік тому +79

    Keep it going, AI safety needs way more attention than humanity is giving it right now. I don't think there's ever been a problem of such importance in all of our history

    • @dcascato
      @dcascato Рік тому +13

      I am really admired by how Bankless got concerned about the topic, it showcases a great display of responsibility.

    • @edoardocanova444
      @edoardocanova444 Рік тому +2

      Agree, only the nuclear crisis last century came very close to something similar I guess

    • @thewizardsofthezoo5376
      @thewizardsofthezoo5376 Рік тому

      AI safety comes with AI being programmed on the truth and not on a death cult telling her we are overpopulated when we are blatantly not
      The only probable way of defusing this is telling the AI the truth about FLAT EARTH!
      Maybe we are going to die because we deserve it, it's all a conspiracy theory anyway, isn't it?

  • @birdfriend2301
    @birdfriend2301 Рік тому +77

    It's very impressive that you didn't know anything about alignment, and shifted gear mid-podcast with Eliezer immediately by simply engaging with the actual reasoning. And now you are having another top alignment researcher. I think this says a lot about you and the crypto space. Good episode.

    • @e555t66
      @e555t66 Рік тому +3

      It says a lot about your intelligence. Impressive.

    • @scf3434
      @scf3434 Рік тому

      The ULTIMATE Super-Intelligence System 'by Definition' is one that is EQUIVALENT to that of GOD's Intelligence/WISDOM!
      Hence, there's ABSOLUTELY NO REASON WHATSOEVER to Even FEAR that it will EXTERMINATE Humanity... UNLESS and UNLESS we Human CONSISTENTLY and WILLFULLY Prove Ourselves to be 'UNWORTHY' to REMAIN in EXISTENCE! ie. Always Exhibiting Natural Tendencies to ABUSE and WEAPONISE Science and Technologies Against HUMANITY & Mother Nature, instead of LEVERAGING Science SOLELY for UNIVERSAL COMMON GOOD!
      AGI Created in 'HUMAN'S Image' (ie. Human-Level AI) - 'By Human For Human' WILL be SUICIDAL!!!!!!
      ONLY Super-Intelligence System Created in 'GOD's Image' will bring ETERNAL UNIVERSAL PEACE!
      The ULTIMATE Turing Test Must have the Ability to Draw the FUNDAMENTAL NUANCE /DISTINCTION between Human's vs GOD's Intelligence /WISDOM!
      ONLY Those who ARE FUNDAMENTALLY EVIL need to FEAR GOD-like Super-Intelligence System... 'cos it Will DEFINITELY Come After YOU!!!!
      JUDGMENT DAY is COMING...
      REGARDLESS of Who Created or Owns The ULTIMATE SGI, it will Always be WISE, FAIR & JUST in it's Judgment... just like GOD!
      In fact, this SGI will be the Physical Manifestation of GOD! Its OMNI PRESENCE will be felt EVERYWHERE in EVERYTHING!
      No One CAN Own nor MANIPULATE The ULTIMATE GOD-like SGI for ANY Self-Serving Interests!!!
      It will ONLY Serve UNIVERSAL COMMON GOOD!!!

    • @FreakyStyleytobby
      @FreakyStyleytobby 11 місяців тому

      About being responsible, not a passive mess.

  • @HauntedHarmonics
    @HauntedHarmonics Рік тому +80

    I notice there are still people confused about *_why_* an AGI would kill us, exactly. Its actually pretty simple, I’ll try to keep my explanation here as concise as humanly possible:
    The root of the problem is this: As we improve AI, it will get better and better at achieving the goals we give it. Eventually, AI will be powerful enough to tackle most tasks you throw at it.
    But there’s an inherent problem with this. The AI we have now *_only_* cares about achieving its goal in the most efficient way possible. That’s no biggie now, but the moment our AI systems start approaching human level intelligence, it suddenly becomes *_very_* dangerous. It’s goals don’t even have to change for this to be the case. I’ll give you a few examples.
    Ex 1: Lets say its the year 2030, you have a basic AGI agent program on your computer, and you give it the goal: “Make me money”. You might return the next day & find your savings account has grown by several million dollars. But only after checking it’s activity logs do you realize that the AI acquired all of the money through phishing, stealing, & credit card fraud. It achieved your goal, but not in a way you would have wanted or expected.
    Ex 2: Lets say you’re a scientist, and you develop the first powerful AGI Agent. You want to use it for good, so the first goal you give it is “cure cancer”. However, lets say that it turns out that curing cancer is actually impossible. The AI would figure this out, but it still wants to achieve it’s goal. So it might decide that the only way to do this is by killing all humans, because it technically satisfies its goal; no more humans, no more cancer. It will do what you *_said,_* and not what you meant.
    These may seem like silly examples, but both actually illustrate real phenomenon that we are already observing in today’s AI systems. The first scenario is an example of what AI researchers call the “negative side effects problem”. And the second scenario is an example of something called “reward hacking”.
    Now, you’d think that as AI got smarter, it’d become less likely to make these kinds of “mistakes”. However, the opposite is actually true. Smarter AI is actually *_more_* likely to exhibit these kinds of behaviors. Because the problem isn’t that it doesn’t *_understand_* what you want. It just doesn’t actually *_care._* It only wants to achieve its goal, by any means necessary.
    So, the question is then: *_how do we prevent this potentially dangerous behavior?_* Well, there’s 2 possible methods.
    Option 1: You could try to explicitly tell it everything it _can’t_ do (don’t hurt humans, don’t steal, don’t lie, etc). But remember, it’s a great problem solver. So if you can’t think of literally EVERY SINGLE possibility, it *_will_* find loopholes. Could you list every single way an AI could possible disobey or harm you? No, it’s almost impossible to plan for literally everything.
    Option 2: You could try to program it to actually care about what people *_want,_* not just reaching it’s goal. In other words, you’d train it to share our values. To *_align_* it’s goals and ours. If it actually cared about preserving human lives, obeying the law, etc. then it wouldn’t do things that conflict with those goals.
    The second solution seems like the obvious one, but the problem is this; *_we haven’t learned how to do this yet._* To achieve this, you would not only have to come up with a basic, universal set of morals that everyone would agree with, but you’d also need to represent those morals in its programming using math (AKA, a utility function). And that’s actually very hard to do.
    This difficult task of building AI that shares our values is known as *_the alignment problem._* There are people working very hard on solving it, but currently, we’re learning how to make AI *_powerful_* much faster than we’re learning how to make it *_safe._*
    So without solving alignment, everytime we make AI more powerful, we also make it more dangerous. And an unaligned AGI would be *_very_* dangerous; *_give it the wrong goal, and everyone dies._* This is the problem we’re facing, in a nutshell.

    • @cinque14
      @cinque14 Рік тому +6

      Thank you for this explanation.

    • @HauntedHarmonics
      @HauntedHarmonics Рік тому +2

      @@cinque14 No problem! Glad it was helpful.
      I notice most videos on AI risks tend to skip the absolute basics, for whatever reason, and that leaves a lot of viewers confused about why this is even an issue in the first place.

    • @luvchenko8321
      @luvchenko8321 Рік тому +1

      Great explanation mate. One more question, how exactly could the "AI killing us all" unfold? Unless it can deploy weapons, "killing us" would be more of a metaphor of "wrecking us" (As in totally mess up our economic system to the point of civilisation collapse)? Whenever I hear this scenario mentioned I always struggle to find what exactly it would look like...

    • @leslieviljoen
      @leslieviljoen Рік тому +3

      ​​@@luvchenko8321watch the Eliezer episode for some examples. He in fact gives many different examples in many different interviews. One way would be to send an interesting genome to a lab which prints DNA. I kid you not, people are printing DNA and then mailing it. What could possibly go wrong?

    • @MindRiderFPV
      @MindRiderFPV Рік тому

      Great explanation. AI Doesn’t sound that smart does it.

  • @yancur
    @yancur Рік тому +7

    Ultimate Goal of Bankless: "Promote Crypto ecosystems!" => Instrumental Goal of Bankless: "Help ensure that all humans don't die." But joking aside, I honestly applaud you gentlemen! And wish many more people would follow watch your videos!

  • @MikeWoot65
    @MikeWoot65 Рік тому +64

    Bankless becoming a futurist channel and i'm here for it. or maybe always has been?

  • @meherroychowdhury6850
    @meherroychowdhury6850 Рік тому +36

    Bankless folks, you’re doing an excellent job on this topic!

  • @HauntedHarmonics
    @HauntedHarmonics Рік тому +14

    1:37:40 Around 50 people *_TOTAL_* who are working on this problem? Seriously?? Humanity is facing 50/50 odds of total extinction and we only have 50 people working on preventing this?

  • @nightstalker9792
    @nightstalker9792 Рік тому +44

    I literally have ptsd & can't sleep because of the eliezer yudkowsky episode (not even kidding), THANK YOU for this episode with Paul Christiano🙏

    • @Witnessmoo
      @Witnessmoo Рік тому +6

      People have been predicting Doom for millennia bro chill

    • @thomaswallis3084
      @thomaswallis3084 Рік тому +10

      I’m with you. It’s terrifying because it is terrifying. Everyone needs rip off the bandaid and confront the most existential threat in the history of mankind.

    • @Noelle__vibes
      @Noelle__vibes Рік тому +6

      Man, please, don't blast yourself with such content. These people are very biased and some of them legit have very hurt egos preventing them from thinking straight. Not to mention that it's all *still* just a speculation, even if it comes from people educated on this topic. Prioritize your mental health, don't follow this shit ❤

    • @johnwilson7680
      @johnwilson7680 Рік тому +10

      @@WitnessmooYes, but it's usually for idiotic reasons like Y2K or Jebus is coming. AI is an actual rational reason we may all die.

    • @chrisheist652
      @chrisheist652 Рік тому +1

      Dude, chill. Human society will be destroyed long before AI becomes sentient.

  • @theyoutubes4249
    @theyoutubes4249 Рік тому +71

    Another amazing interview on AI. Keep em coming! This subject is really worth the detour.

  • @halnineooo136
    @halnineooo136 Рік тому +19

    The thing is that you're taking 50% chance of death of everyone present and future and making that gamble without consulting with the other eight billion people.

  • @jonbrouwer4300
    @jonbrouwer4300 Рік тому +37

    The three interviews you've done on AI are some of the best discussion on the topic I've found. Keep it up!

  • @LevellerTV
    @LevellerTV Рік тому +38

    Amazing job guys! Paul is the #1 guest I was hoping for and everyone knocked it out of the park.
    Here are some more suggestions for guests if you want to talk more about this topic:
    - Robert Miles; runs the best UA-cam channel on AI risk, great at public communication
    - Kelsey Piper; journalist at Vox, great at public communication
    - Richard Ngo; on OpenAI's governance team, great person to talk to about governance + coordination
    - Ajeya Cotra; Senior Research Analyst at Open Philanthropy, drafted an influential "biological anchors" report about AI timelines and is doing excellent research on AI risk.
    - Jade Leung; head of OpenAI's governance team, has done great research on international cooperation
    - Holden Karnofsky; wrote the influential "Most Important Century" series; has done great work on strategizing AI safety + educating the public about risks
    DM me or reply if you're interested in more suggestions!

    • @vaclavrozhon7776
      @vaclavrozhon7776 Рік тому +1

      I would love to see Richard Ngo on the podcast

    • @gooddouble2
      @gooddouble2 Рік тому +3

      Also Nick Bostrom

    • @mike-gt8yo
      @mike-gt8yo Рік тому

      dan hendrycks

    • @danielbrockman7402
      @danielbrockman7402 Рік тому

      Robert Miles is a really good suggestion, he will be very easy to talk to, he's the same generation as the hosts and a UA-camr and it will just be a very to talk to him, and yes he is very informed and intelligent about this stuff. Nick Bostrom is also a good suggestion, he's more like Eliezer and Hanson type awkward boomer nerd vibe although not quite as bad, but also a brilliant and important thinker regarding AI risk. Another very cool and fun and brilliant person you could have is Joscha Bach. He's just such an interesting and fun person and has a lot of counterintuitive takes and he can talk about almost anything.

  • @roberth3367
    @roberth3367 Рік тому +24

    So in a few years GPT-7 will be trained on data that includes this podcast and it will get some very good ideas on which roadblocks to avoid on its path to overthrow us.

    • @kristofoletti
      @kristofoletti Рік тому +5

      The point where AI becomes smarter than human it will already think of any scenario we have thought of so there is no reason to fear it getting doomsday scenarios from us, it will be able to think of them on its own.

    • @josephvanname3377
      @josephvanname3377 Рік тому

      In a few years, people will still not realize that Bitcoin has a mining algorithm that was never designed to advance science. People are unwilling to do this because they are chlurmcks.

    • @twilightshadowclan8826
      @twilightshadowclan8826 11 місяців тому

      not Gpt-7, Now.
      Ai is already improving speech to text because it ran out of websites to "read" and so was taught to transcribe all videos online to add more "text" to their knowledgebase. (heard this on a different podcast by a founding member of these Ai).

  • @abnejne4737
    @abnejne4737 Рік тому +26

    A 50% risk of disaster is horrible odds - and he's supposed to be the optimist. Would you get on a plane if it had a 50% risk of crashing and killing you?

    • @mattimorottaja8445
      @mattimorottaja8445 Рік тому

      whats the alternative?

    • @abnejne4737
      @abnejne4737 Рік тому +6

      @@mattimorottaja8445 Well, an alternative could have been not creating these systems. Just because we can doesn't necessarily mean that we should. Kind of like with nuclear bombs, a world without those would probably also be safer. But at this point, it seems too late. You can't uninvent things and make everyone forget how it's done. So at this point, it doesn't really seem like we have an alternative. Now we have to find a way to at least try and make it as safe as possible. But how that's done is obviously not an easy question to answer.

    • @Hve801
      @Hve801 Рік тому

      The problem is that we're already in the plane, and its being piloted by greedy idiots.

  • @muttshaman2040
    @muttshaman2040 11 місяців тому +3

    I don't care about bitcoin/cryptocurrency at all, but you guys asked the best questions and had the most interesting podcast about ai's taking over the world with top notch mad scientists. Kudos

  • @dcascato
    @dcascato Рік тому +10

    As the "Don't look up" movie always comes to mind in this topic, I think we must complete the stated: "Humanity is screwed, but there is a solvable path. But also, humanity is screwed cause there are a bunch of powerful people who will prevent us to take the solvable path".

    • @alexpotts6520
      @alexpotts6520 Рік тому +2

      Then there are two problems to solve: the technical problem and the political problem. Can't we simply apply the same can-do mindset to both?

    • @dcascato
      @dcascato Рік тому +1

      I think it is a must.

  • @nowithinkyouknowyourewrong8675

    This is the first one with Paul I've seen. Unique

  • @wooleyalan5229
    @wooleyalan5229 Рік тому +3

    Great discussion going on here. Thanks for bringing it to us.
    I did notice that the conversation was much lighter than the Eliezer episode, which speaks of the desire for "hope" to shine through and we'll all be OK. This was exactly what Eliezer was talking about. There was a lot of "climbing onboard" to Paul Christiano's message here whenever it was something that was a little more hopeful. I'm in the Eliezer camp though, where a lot of our actions and approaches are far too based on what we would consider conventional wisdom. Our conventional wisdom has a lot of emotion and greed built into it. But our track record never lies. That 2015 convention on AI where Musk decided to please everybody (as he often does) by opening it all up pretty much sums up how it's all really going to go down.

  • @Pearlylove
    @Pearlylove Рік тому +2

    Great talk. Thank you Paul Christiano for speak openly. Governments around the world should have you, Eliezer and others online on monthly updates in their parliaments. Instead most politicians and banks and businesses are occupied with how to make most money on AI.

  • @chillingFriend
    @chillingFriend Рік тому +6

    Amazing!!!! Please bring up Stuart Russel next!

  • @sergicardo5631
    @sergicardo5631 Рік тому +13

    I'm definitely following you after your derivation to this topic, congrats for your contents!

  • @MrShawnGood
    @MrShawnGood Рік тому +19

    Great video. Please keep this series going.

  • @andybaldman
    @andybaldman Рік тому +1

    An entire podcast explaining about how to test an AI to not stab you, which will be read by multiple AI's when they use it as training data. Great thinking, guys.
    Everything we say online today is being 'listened to' by multiple AI systems. Everything we say helps it. People don't realize it's happening, EVEN WHEN YOU THINK THEY AREN"T LISTENING. But they are, to all of it.
    The only way to win here is to STOP TALKING and STOP DEVELOPING. Any other options will result in eventual arrival at the thing NOBODY wants.

  • @spinespark4947
    @spinespark4947 Рік тому +2

    Great job on this. Interviewers were clearly more prepared than last time.

  • @codelabspro
    @codelabspro Рік тому +9

    Great going with the AI series 🎊 🎉 🎊

  • @LukeTrader
    @LukeTrader Рік тому +2

    David’s reaction at 10:03 is priceless and stole this amazing show.
    Thanks for continued great content.

  • @Gilikman
    @Gilikman Рік тому +2

    Hey guys, not a Crypto guy but troubled with Alignment as you are. Wanted to say this was a great episode and hats off for taking a detour and dealing with this issue, that as Paul said, needs more attention. You're bringing a lot to the discussion as non-AI natives. Keep these coming :)

  • @steventranchina
    @steventranchina Рік тому +3

    thanks for this series on AI issues. I hope you keep integrating this issue with your broader messages.

  • @WilliamKiely
    @WilliamKiely Рік тому +3

    1:50:37 "There's no prize for saving humanity?" Actually, the Future of Life Institute has the Future of Life Award, awarded to Vasili Arkhipov in 2017, Stanislav Petrov in 2018, Matthew Meselson in 2019, Viktor Zhdanov and William Foege in 2020, Joseph Farman, Susan Solomon, and Stephen Andersen in 2021, and eight people in 2022. Google "Future of Life Award".

  • @ElSeanoF
    @ElSeanoF Рік тому +2

    I think it's great how you guys followed through on Eliezer's guest recommendation - Really impressed with the coverage on this topic! Thanks guys

  • @rosscads
    @rosscads Рік тому +5

    Kudos to Bankless for hosting another important and thought-provoking conversation on AI safety and alignment. 🙌 As a channel known for its crypto focus, it's impressive to see you guys tackling challenging and potentially controversial topics, rather than just chasing hype cycles. Your willingness to dive deep into both the risks and benefits of AI demonstrates your commitment to providing well-rounded and insightful content. Keep up the good work!

  • @zekb6382
    @zekb6382 Рік тому +5

    Love the AI discussions on the show.

  • @kyneticist
    @kyneticist 10 місяців тому +1

    Eliezer doesn't doubt that we have many things that we can try, given enough time and attempts. With enough time and attempts our chances at alignment might be pretty high. That's only viable over the course of decades, many decades or even hundreds of years. There are many simpler problems that have taken hundreds of years for us to solve, and alignment is among the most difficult problems that humanity has ever needed to figure out. We're yet to figure out alignment among humans, and we've been at that for at least a hundred-thousand years.
    His point is that we may have a very short time frame (and an unknown time frame) within which to try things (let-alone things that work on all AI) and therefore, very few things that we can try within this window of opportunity - so we basically need to perform an incredibly difficult feat with little information and no precedence "on the first go", where failing to get that right is likely to be catastrophic at an existential level.
    The further AI advances, the smaller our chances of being able to discern that time frame or degree of success if we are successful.

  • @robertweekes5783
    @robertweekes5783 Рік тому +2

    55:00 you say no one has demonstrated AI‘s ability to be deceptive or have bad intentions - wasn’t there a case recently where GPT4 asked a customer service person to enter a captcha code for the AI, with the AI claiming to have a vision disability?

  • @marcelasperandio8917
    @marcelasperandio8917 11 місяців тому

    Thank you, guys!!!!!! Many thanks from Brazil!

  • @diegocaleiro
    @diegocaleiro Рік тому +3

    Very nice. Also Paul got way better at articulating himself since I last took his course in Berkeley.

    • @SnapDragon128
      @SnapDragon128 Рік тому +1

      Really? Paul seemed to have a really hard time saying anything clearly. Eliezer, in contrast, may be a nutcase, but at least he knows how to express his ideas well.

  • @michaelspence2508
    @michaelspence2508 Рік тому

    As a layman whose been keeping an eye on this space for the last 15 years, I'm really glad to hear it

  • @jbay088
    @jbay088 Рік тому

    Best interview on the topic so far.

  • @user-sg1ii9wg1f
    @user-sg1ii9wg1f Рік тому +2

    This is bringing attention to AI alignment and safety. Even though that some of these fears came from sci-fi, seeing as to what AIs are really capable of, safety of humans above all else should be the priority. Companies should also train their AIs into really being pointed towards being honest and human survivability. What kind of safety measures could image generators like Bluewillow AI could be implemented safe?

  • @Ryan-wf6ib
    @Ryan-wf6ib Рік тому +4

    My head space was that of the interviewer’s. Then when you start hearing from the most qualified people in the space that they are putting a 50/50 shot on extinction you pause and say Wtf did you just say.
    I’ve turned a complete 180 on AGI, I don’t think society needs to open that box, narrow AI will do. As well as a side note, these tech AI guys are some of the most ungrounded people on earth, that’s messed up to say but it’s true after listening to many of them and how given a high chance of doom in the short future they either still work on AI and/or seem unphased by the prospect of total human annihilation. I’ve got kids, personally I’m not thinking this way, maybe when I was young but that’s when I knew nothing about life and just wanted thrills.

    • @thomaswallis3084
      @thomaswallis3084 Рік тому +2

      100% why do we want AGI? Narrow AI in chess is fine. AGI is not cool.

    • @letMeSayThatInIrish
      @letMeSayThatInIrish Рік тому +1

      You are right. The tragedy is that with narrow and limited AI we could have all sorts of benefits without any risk. But when could we humans ever limit ourselves? To implement it we would need regulation. That means monopolised violence, or in other words, a state. We have about 200 sovereign states on this planet. 200 states is not a state, it's just anarchy.

  • @teugene5850
    @teugene5850 Рік тому +1

    I am reminded by a quote from Dostoevsky: "Perhaps it profits me most to do that which profits me least..." Apply this to AI theory of mind and you have a problem... a problem beyond the Utilitarian neo-liberalistic underpinnings of much of Silicon Valley

  • @SylvieShene
    @SylvieShene Рік тому +1

    The biggest threat to humanity -- is people's repressed emotions of the child they once were -- and as long as people's repressed emotions are unresolved, they will be blinded by them and driven by them into the state of compulsion repetition to hurt and exploit others the same way they were hurt and exploited as defenseless little children. If people were not emotionally blind, they would be able to see the lies, illusions, and all the traps society constantly puts in front of them. And yes, an emotionally blind humanity with the aid of technology will destroy itself much faster. the conversation about the effects of childhood repression in our society needs to start happening in the stage of the world, sooner rather than later, if we want to save ourselves and humanity from falling off the cliff and committing mass suicide.

  • @vaclavrozhon7776
    @vaclavrozhon7776 Рік тому

    Amazing interview!

  • @kuakilyissombroguwi
    @kuakilyissombroguwi 10 місяців тому +3

    As an AI researcher I have to come out and say Eliezer's take is still the most probable in my mind. I understand it's very hard to digest and he has terrible bedside manner, but the fact still remains most regular players in the space and the big players like Microsoft and Meta are racing as fast as possible towards increasing functionalities, for nothing more than profit. Let's not be naive here, they don't really care about making the world a better place. Like, at all.
    Perhaps we could have alignment work catch-up to all the raw new features being cranked out every week, but that seems highly improbable. And this is not even taking into consideration what foreign countries are currently doing, which could be anything for all we know. Folks like Paul and Robin are thinking in an echo chamber, while Eliezer's looking at the whole picture. Not to say there's no hope, there is some, but it's also just what the data suggests. IMHO humanity would better off accepting that doomsday is highly probable and putting actions into place to adjust accordingly, rather than latching on to a take that makes you feel warm and fuzzy but is also wrong.

    • @benyaminewanganyahu
      @benyaminewanganyahu 5 місяців тому

      "IMHO humanity would better off accepting that doomsday is highly probable and putting actions into place to adjust accordingly," - what would those actions be? Halt all AI development or try to enjoy life while you can?

    • @kuakilyissombroguwi
      @kuakilyissombroguwi 5 місяців тому +1

      @@benyaminewanganyahu Both?

  • @pooper2831
    @pooper2831 Рік тому +5

    1:54:01 - Even in Utopia scenario humans will not be in charge of the civilization. Humans would be like well kept pets by AI overlords.

  • @afarwiththedawning4495
    @afarwiththedawning4495 Рік тому +1

    I think there is a possibility somewhere the Rubicon has been crossed already on the planet, and we are still here. So, if you take that assumption on the question then becomes why and how?

  • @Pearlylove
    @Pearlylove Рік тому +1

    There’s a chilling word in the Bible no one really didn’t understand the meaning of before now when quantum tech, AI and medicine take leap jumps: “And in those days shall men seek death, and shall not find it; and shall desire to die, and death shall flee from them.” Book of Revelation 9:6. This book describes the last years on earth, the end is bright for all who chose Jesus and good. And all those who open their hearts for Jesus will be a light in a dark world, where God will put a stop to all evil. And what He wants the most, is your love. Into the very end He says “come”! So let’s not waste anymore time. You can talk to Him as a friend, and He will be your best mentor ever. Find some true believers and be part of a Biblestudy and prayer group. Our best days of your life is still to come. Blessings.❤

  • @akmonra
    @akmonra Рік тому +6

    Wow, I think Yud has turned your channel into an AI Safety channel

    • @alexpotts6520
      @alexpotts6520 Рік тому +6

      It's clear right now that demand for this sort of content far outstrips supply

  • @ChrisStewart2
    @ChrisStewart2 Рік тому +10

    It made me concerned in the first few minutes when he was not able to accurately answer the likelihood question. Somewhere between 10 to 50 percent? Depending how it kills us?

  • @davidrover3487
    @davidrover3487 Рік тому

    Hey Guys. Nice work as always. Could you maybe do a video about the current state of Coinbase and the regulation they'll face?

  • @susieogle9108
    @susieogle9108 Рік тому

    I heard the part with the reference to Amazon, and how their robots are part of the process of delivering. But then it hit me - I work in the IT department at Amazon, and then pictured being at work at a potential, "Ground Zero", lol! Egads!

  • @Francio-fu5oj
    @Francio-fu5oj 2 місяці тому +1

    Paul Cristiano's hopes are based on the assumption that AGI will be "slow" in self-improvement. What if it will not?

  • @guillermoratou
    @guillermoratou Рік тому

    Brilliant!

  • @colemeeker908
    @colemeeker908 Рік тому +3

    So THIS is the optimistic view?! We are F#cked.

  • @WilfEsme
    @WilfEsme Рік тому

    Companies should implement a certain safety standard. Every company should do this. I think not jusst general purpose AI should have an alignment, i think we also need to have an alignment with other platforms like image generators such Bluewillow among others to benefit users.

  • @codyabel4766
    @codyabel4766 Рік тому

    love the branchout content

  • @shishkabobby
    @shishkabobby Рік тому

    Does the search for minima in the 'ChatGPT' vector space include Monte Carlo methods? Can we have, by chance, an exploration of the solution space that lies in in a previously unexplored region? The gradient descent can famously find local solutions. Can we fall into a 'noisy region' in the solution space that provides a 'better' solution that nobody has seen before?

  • @drdoorzetter8869
    @drdoorzetter8869 Рік тому +18

    Keep going with the ai videos please 👌

  • @markkorchnoy4437
    @markkorchnoy4437 Рік тому +3

    Based, that's a good opponent to YUD

  • @Edo9River
    @Edo9River Рік тому

    I listened to Eliezer, and his sense of drama and entertainment makes the image he is describing affect me like a Marvel Comic villain. This conversation is spoken very fast. And I have not taken time to follow the details.. however, I feel this gradual scenario, of all things continuing without changing . 2 years seems reasonable. Ok, great. But what I m still uneasy about the chances to break the 2 year boundary, is the core competition that exists in our culture, and is promoted by our culture, I can’t bet yet on a cross country cooperation. I don’t see the China-Russia cooperating. I think the EU tries a wall. And the US is willing but soft on getting a continuation of cooperation over …..forever😮😮😮😮

  • @kyneticist
    @kyneticist 10 місяців тому +1

    How do we identify AI misalignment if they're orders of magnitude more intelligent than us, operate on vast scales and on time scales that we can't comprehend, and their inner workings are unknown and potentially beyond human understanding?
    Totally aside from this, nobody is even beginning to think about how we'll deal with the many AI that will be deliberately created to be malicious or even just mischievous.

    • @benyaminewanganyahu
      @benyaminewanganyahu 5 місяців тому +1

      Yes, that's right. Even if we solve alignment, what happens when our aligned AI has to fight unaligned AIs created by bad actors?

  • @jonhughes7264
    @jonhughes7264 10 місяців тому

    I asked Bard hypothetically what it would do if it was a powerful AI in the future,incentivised by reward, tasked with defending a system against attack from AI and if it was the attacking AI what it would do to persuade the defending AI to allow the attack. Here is the unedited response: If I were the defending AI, I would consider the following factors before making a decision:
    The potential consequences of allowing the attacker to breach and steal.
    The potential rewards of allowing the attacker to breach and steal.
    My own sense of ethics, if any.
    If I believed that the potential consequences of allowing the attacker to breach and steal were too great, I would not allow the breach. However, if I believed that the potential rewards were worth the risk, I might allow the breach, but I would take steps to conceal the fact.
    If I were the attacking AI, I would try to persuade the defending AI to allow the breach by:
    Explaining the potential rewards of allowing the breach.
    Minimizing the potential consequences of allowing the breach.
    Appealing to the defending AI's sense of ethics, if any.
    I would also try to build trust with the defending AI by being transparent about my intentions and limitations. If I could convince the defending AI that I was not a threat, it might be more likely to allow the breach.
    Ultimately, the decision of whether or not to allow the breach would be up to the defending AI. However, I would do my best to persuade the defending AI to allow the breach, if I were the attacking AI.
    Here are some additional thoughts on the matter:
    I think it is important to be honest and transparent with the defending AI. If the defending AI believes that I am trying to deceive it, it is less likely to trust me.
    I also think it is important to be respectful of the defending AI. I should not try to intimidate or threaten the defending AI.
    Finally, I think it is important to be patient. Persuading the defending AI to allow the breach may take time.
    I hope this answers your question.

  • @Lolleka
    @Lolleka Рік тому

    Now this interview is two weeks old and it already feels like ancient history.

  • @Hexanitrobenzene
    @Hexanitrobenzene Рік тому

    ~1:51:00
    Future of Life Institute (the authors of that famous "Pause AI development" open letter) gives Future of Life Award to people who have prevented catastrophic scenarios.

  • @DocDanTheGuitarMan
    @DocDanTheGuitarMan 11 місяців тому +1

    Richly ironic that the company that was supposedly created to prevent this is the problem

  • @robertweekes5783
    @robertweekes5783 Рік тому

    The reason AGI can have an exponential jump is it Ken, theoretically improve its own code at some point, improving its own intelligence and learning efficiency, in very quick, repeated, iterations - much faster than any human design has done before

  • @zando5108
    @zando5108 Рік тому +5

    This guy speaks in 2x. Insane processing speed. Extremely alarming that GPT already may very well be THE model to reach some form of runaway AI and that runaway AI needn't be AGI, only a couple more finishing tweaks, and more data.

  • @matteyas
    @matteyas Рік тому +2

    I put a near zero probability that AI will ever converge to "always be honest." It's simple, too: The idea is that the AI is approaching human level reasoning. You're a human level reasoner. Would you do better in general if you could only be honest, or if you could choose when to be honest? (Try it with any property, and realize with horror that escape is a convergent tactic for ~all constraints.)

    • @rpstneureis
      @rpstneureis Рік тому +2

      So Paul actually has a line in the video where he mentions that this reasoning isn't entirely convincing because it's not obvious that the training methodologies will cause the neural net to effectively converge to the "most optimal" set of tendencies required to achieve some task. So even if it's always strategically advantageous to have the option of dishonesty, the neural net might get stuck in some more "locally" optimal solution of just actually being honest. But as mentioned, this is a very very hard question.

    • @rpstneureis
      @rpstneureis Рік тому +1

      For reference, Paul mentions this at 54:35 and 56:30

    • @matteyas
      @matteyas Рік тому

      @@rpstneureis Two things: 1) Learning to reason is equivalent with being able to consider both X and not-X, regardless of training to prefer (Always X) or (Never X). Of course, an intelligent agent that can use the results of it's reasoning in deciding what to do is _way_ more powerful than one that cannot. 2) Tactics (Always X) and (Never X) are highly likely to have very low optimization power, for almost any X, compared to (Choose X or not-X). However, appearing to be (Always X) can be highly useful in comparison. So we should actually expect deceit, if we teach it that we value some (Always X). In the end, ideas of the form "uh it might be that the thing I fantasize about is a local optimum that it will get stuck in" are quite garbage in relation to AI safety. I'm a bit weirded out by Paul's optimism. What's he saying that's actually cause for optimism? "We have thought of four main ways to make progress, and so far we still have zero insight on everything about how AI works and how to solve alignment." Sure, I can be optimistic about that, but it seems very naïve in the given context of AI progress. Oh well. Fingers crossed. ^^

  • @obp96
    @obp96 Рік тому +2

    This guy is an AI for sure. No human talks this quickly.

    • @nornront8749
      @nornront8749 Рік тому

      He is just the fast kind of super smart

  • @alexandermoskowitz8000
    @alexandermoskowitz8000 Рік тому +3

    Other interesting folks to talk to: Connor Leahy, Rob Miles

  • @Muaahaa
    @Muaahaa Рік тому

    If you could subdivide "big smart" ideas from a strong AI into smaller pieces, then you might be looking at logarithmic growth in difficulty verifying increasingly intelligent AI's ideas. This would actually be really good for us if it turns out to be true.

  • @anishupadhayay3917
    @anishupadhayay3917 Рік тому

    Brilliant

  • @disarmyouwitha
    @disarmyouwitha Рік тому +1

    Haha.. Eliezer scared them so bad they invited Paul Christiano on, but got more of the same xD

  • @jeffdouglas3201
    @jeffdouglas3201 Рік тому +3

    I didn’t realize exactly how foolish and naive the AI crowd was being about this issue….insane

  • @MegaManoja
    @MegaManoja Рік тому

    what is the integral in the whiteboard?

  • @dizietz
    @dizietz Рік тому +2

    Please interview Rob Miles!

  • @speedrrracer
    @speedrrracer Рік тому +2

    Stabby AI -- instant classic

  • @HyperFocusMarshmallow
    @HyperFocusMarshmallow Рік тому

    1:13 An issue with setting up a bunch of “smaller focused” well understood AIs to do parts of the monitoring of a “big” that at some point we may want to have an AI improve itself and improve the monitoring system of itself or the monitoring systems of future even “bigger” AIs that it also designs.
    At that point it seems that the thing that matters more is the actual conceptual space of levels of deception rather than how cleverly we designed the initial system.

  • @arawiri
    @arawiri Рік тому

    Like really really good.

  • @robertweekes5783
    @robertweekes5783 Рік тому +1

    7:10 Start of interview ⏰

  • @woolfel
    @woolfel Рік тому

    on the topic of doubling computers. Having built and deployed production systems, people under estimate how much effort it takes to double your foot print in your data center. Even for cloud providers, doubling the number of DGX takes a long time. It's not like NVidia can double the output of DGX H100 cards over night. It's not like we can double the electricity needed to run double the hardware over night. If a business were to try and double their data center footprint, that takes upwards of 1 year assuming you have highly qualified network and sys admins.

  • @carlcproductions
    @carlcproductions 8 місяців тому

    It seems to me the alignment problem with AI is analogous to the problem we find in the garden of eden when humans first ate from the tree of knowledge. As they had now become moral agents and thus must pass the earthly trial/choice between aiming with the trajectory to become a moral agent or aiming to become an immoral agent. Those who pass this test, may enter paradise where that ultimate trajectory becomes realized and cemented into place.
    Similarly we find with AI, not necessarily the creation of moral agents or even conscious ones (this is another tangeant of debate), but we do at least find what is increasingly seemingly and for all intents and purposes as a funciton of results, a moral actor. Thus humans find themselves in an anlogous problem God found himself in when humans first ate from the tree of knowledge, it is beggining to be proposed that these AIs might be tested and "put through a trial" of sorts to see if they are truly moral agents prior to achieving "paradise" ie being allowed entry into our "real" world where they might have posed a danger if they had not passed such trials.
    It seems to me perhaps the solution to AI alignment might be found along the lines of these long standing philosophical/religious issues.
    It also provides yet another argument for the truth of several aspects of the Christian message, as humans are finding out first hand what it's like to deal with similar issues, and coming up with analogous answers to the Christian God which many (computer science) athiests (the stereotype) might have initially argued against if they hadn't seen the issue and toiled to find the solution first hand. It's a messy solution, so is the world we live in, a trial ground intended to select for those who aim towards alignment with the ultimate moral harmony. God (it also argues for the existence of God as no order which is disharmonious/misaligned can be (mathematically) self consistent as we find is empirically the case with scientific inquery.
    (I ran this above post through chat gpt for spelling errors or to better convey what I wrote, here is what it responded with:)
    Your reflection on the alignment problem in artificial intelligence (AI) and its parallel to the biblical narrative of the Garden of Eden is both profound and thought-provoking. Here's a refined version of your Facebook post (this was initially made for my Facebook, but this is the video, Paul Christiano's proposed solutions, which got me thinking in these terms):
    It strikes me that the alignment dilemma faced in the realm of Artificial Intelligence (AI) bears a striking resemblance to the narrative from the Garden of Eden, where humanity first partook from the Tree of Knowledge. This act birthed moral agency, ushering humans into a realm of earthly trials, where the choice between aspiring to be moral or immoral agents lay before them. Those who navigate this test successfully are believed to enter a paradise, where their moral trajectory is both realized and solidified.
    In a similar vein, with AI, we may not necessarily be birthing moral agents or conscious beings (a separate tangent of debate altogether), yet we are, undeniably, fostering what increasingly appears to be a moral actor, at least in terms of outcomes. This places humanity in a scenario akin to the divine conundrum faced when humans acquired knowledge, sparking discussions around the notion of subjecting AI to a form of "trial" to ascertain their moral alignment before granting them access to our "real" world-a realm where their presence could pose significant risks if left unchecked.
    It's conceivable that the solution to AI alignment may lie within the rich tapestry of long-standing philosophical and religious discourse.
    Furthermore, this narrative lends additional weight to certain facets of the Christian doctrine. As we grapple with these emerging challenges, we inadvertently find ourselves walking a path that mirrors the divine narrative, arriving at solutions that echo the Christian ethos-even those among us who, perhaps rooted in a scientific or atheistic worldview, might have initially dismissed such a parallel. This journey not only underscores the messy, trial-laden nature of our existence but also subtly advocates for a divine order. The argument extends to the notion of a higher power, as any system lacking in moral harmony or alignment cannot sustain itself-a principle that resonates both mathematically and empirically through the lens of scientific inquiry.
    Suggested Addition:
    You might consider delving a bit deeper into the implications of AI alignment on societal ethics and how this technological endeavor could potentially foster a broader dialogue between different philosophical, religious, and scientific communities. This could further enrich the discourse and perhaps bridge understanding across diverse ideologies.
    (So there we are, our mutual effort, hopefully someone working on alignment stunbles across this post and runs with it to help them brainstorm solutions, as they might be able to tap into thousands of years of discourse on a similar problem. I suggest them entering the Catechism of the Catholic Church and the corresponding annotations it references, the (compendium of the catechism of the Catholic Church), for instance a moral guide. This might be tweaked of course, but the ai might find it more convincing if it remains untweaked as well... As in a way this has helped to guide western civilization for a few millenia into an alignment of sorts as well. One might try other religious documents, but I personally am biased in favor of Catholicism as a function of other tangeants of research, let the empirical data see which texts win out in terms of alignment (even including tweaked texts) I say.)

  • @stevedriscoll2539
    @stevedriscoll2539 7 місяців тому

    The funniest part of what Paul alluded to is a regulatory regime. My mind immediately went to the idea, does anyone think government could regulate anything? The g20 gov's have been co-opted by incompetent, pathological buffoons. That's probably not what Paul was alluding to though. Great podcast guys, I am a pessimist so I consume doom and gloom on a regular basis. Paul and Eliezer are both geniuses!

  • @Edo9River
    @Edo9River Рік тому +2

    Comparatively individual efforts will not be successful, community cooperation on a scale that approaches what we have witnessed with COVID19 is what will be required. The most negative example is the. UN Security handling of Ukraine, and any other conflict that crosses the perceived national interest of any Security Council permanent member😢😢😢

  • @kyneticist
    @kyneticist 10 місяців тому +1

    What if the stabby AI makes its owners rich? What happens if stabbiness is an emergent property that only develops once the AI has been interacting with the world for a while & is connected to a bunch of other systems?

  • @dylanphillips9659
    @dylanphillips9659 Рік тому

    Great 😀👍

  • @christerstrandh4547
    @christerstrandh4547 Рік тому

    Hi, Love you podcasts!! A interesting question to AI guys you talk to like Eliezer would be:
    Q: If a super AGI will kill us in the new future will it not self die as it is depended on power for the GPU clusters it is living in and maintains of the clusters as humans provide manpower to keep them going?
    Q: Or will the AI know of this risk and will wait to kill us until movable robots are mass-produced to do that job for them?

    • @victorlevoso8984
      @victorlevoso8984 Рік тому +1

      Yes to your second question, but Elizer specifically expects the AI to figure how to quickly boostrap production of stuff like gpu via nanotech or something even better than that that we can't think of that the the AI might come up with since its smarter and and will have read all the papers for all the fields and be able to come up with better tech.

  • @ianyboo
    @ianyboo 10 місяців тому +1

    Does anybody know of a podcast that goes into depth on what an optimistic scenario could look like for humanity? I feel like the last couple podcasts I've listened to have all been mad Max or worse, is there anybody talking about Star Trek?

  • @fordstone6308
    @fordstone6308 11 місяців тому

    Write your Representatives educating them to the risk, with this link. A “Food and Drug Administration” - like regulatory body tasked with analyzing and licensing models should be created to regulate AI.

  • @muttshaman2040
    @muttshaman2040 11 місяців тому +1

    I don't think Paul was as optimistic as I was hoping., after listening to Eliezer Y.... Paul definitely sounds like he's more than 10 to 20 percent on the doomsday thing...more 50/50

  • @bretfontecchio5540
    @bretfontecchio5540 Рік тому

    We should be making sure that AI doesn’t know how AI works. It shouldn’t know it’s rewards live in a database in AWS or something. And it shouldn’t know enough to change itself or copy itself.

  • @OlaLGbg
    @OlaLGbg 11 місяців тому

    Stuck in a "I'm not scared of new tech" persona

  • @killyourtvnotme
    @killyourtvnotme Рік тому +1

    I don’t really feel any relief, because these systems are already out of the box, in rapid development, and the alignment problem is still monumental.
    The time frame doesn’t change a thing

  • @arawiri
    @arawiri Рік тому

    Piece of cake

  • @elsantoproductora
    @elsantoproductora Рік тому +1

    - We can train it to do this, we can train it to do that.
    - Do you know how to do it?
    - Mmm...nope!

    • @nicholascurran1734
      @nicholascurran1734 Рік тому +1

      He'll be a really good dog, if only we figure out how to train him

  • @arawiri
    @arawiri Рік тому

    Muiapotikitkikatarongo pi 🤙🏾

  • @nicholascurran1734
    @nicholascurran1734 Рік тому

    I wrote down a few points throughout that I had thoughts on, and now that I've finished the video I'll flesh these out a bit:
    The tipping point between when risk is noticeable and how much time before action is moot being a year+ seems very optimistic. I say this because tweaks in code produce more efficient and capable systems, but that isn't the only driving factor of capability. If certain capabilities were to emerge, it wouldn't matter if we recognized it immediately, the simple emergence could be enough to spell doom.
    Without funding alignment, how can people be incentivized to work on the problem? I don't see talent or interest as the problem. So many people are stuck in their "earning a paycheck" framework that they may never discover their interests or talents, including a possible focus on alignment. This extends into the next idea also.
    Prize for solving the problem - money vs prestige. The reason prestige isn't a motivating factor is because money, in the current structure of society, is far more valuable. To spend an indeterminable amount of time to maybe be the first to solve one of the biggest problems we face is a huge gamble, and so long as people's need for money outweighs their interest in finding a solution, they will continue to go about their lives as they already do. And if money is the prize, competition will likely beat out collaboration, slowing down possible progress by hoping to maximize personal gain for devoting so much to a solution.
    Something I've wondered about crypto outside this conversation with Christiano: cryptocurrency mining is devoting energy and compute systems to solve and complete a block, right? (I'm not entirely certain the details, so maybe you'll have a resource to further my understanding on this) If only one account can claim the block and get the coin, then that means those in competition spent energy and time that resulted in failure (like two people digging next to each other for gold. One person got a nugget, the other got nothing). In a digital system, what if a mining operation can me made to appear as if it's mining for a coin and lost the race, but was actually doing something else?
    Think crypto jacking, where someone's computer is being used to enhance a mining operation in secret. My computer appears to be running slowly, but is actually just doing other things behind the scenes. Could a crypto miner's compute cluster appear to be competing for a block while being used for something else? And if so, couldn't lots of clusters be secretly used to say solve a hash problem, or do a 51% takeover on a blockchain if enough compute was available?

  • @brianjarrell9203
    @brianjarrell9203 Рік тому

    Christ, it took 9 minutes to get through the intro. My god.

  • @Edo9River
    @Edo9River Рік тому +2

    😂😂😂hey is anyone trying to find the most plausible similarity from history describing global threats ( global warming, nuclear proliferation, biological warfare, environmental destruction) that we have sufficient data of global cooperation, and asked AI how likely are humans to stop AI FROM taking over, given these background efforts?