The 10 Trillion Parameter AI Model With 300 IQ

Поділитися
Вставка

КОМЕНТАРІ • 95

  • @chapterme
    @chapterme 2 дні тому +5

    Chapters (Powered by ChapterMe) -
    0:00 Coming Up
    0:54 What models get unlocked with the biggest venture round ever?
    5:35 Some discoveries take a long time to actually be felt by regular people
    9:53 Distillation may be how most of us benefit
    14:26 o1 making previously impossible things possible
    21:17 The new Googles
    23:47 o1 makes the GPU needs even bigger
    25:44 Voice apps are fast growing
    27:05 Incumbents aren’t taking these innovations seriously
    31:52 Ten trillion parameters
    33:15 Outro

  • @thehari75
    @thehari75 3 дні тому +52

    Lol did sam altman tell yall to avoid talking about claude and comouter use update

    • @superresistant0
      @superresistant0 13 годин тому

      are you drunk? they said Claude has the fastest adoption in YC from 5 to 25% in few month

  • @hanchisun6164
    @hanchisun6164 3 дні тому +9

    One order of magnitude higher is referring to compute, not parameter size. The current frontier model (4o/pro/sonnet) is all much smaller than their previous sota (4/ultra/opus).
    GPT4-1.8T MoE might continue to be the largest model people train for a while.

  • @fil4dworldcomo623
    @fil4dworldcomo623 День тому

    Winning AI company capturing all the value is speculative, therefore an answer can also be speculative that the more any AI becomes powerful, the more complex society becomes, hence the value of being human to cope with new contexts to feed the "hungry AI" proportionately increases and hopefully it's fun.

  • @MicahBratt
    @MicahBratt 2 години тому

    Even if AI doesn't improve from here companies like Microsoft, palantir etc. are building out the AI eco systems that you will start seeing big companies using for all their backend systems. The average Joe will see the changes pretty fast I would think.

  • @ayakitokugawa
    @ayakitokugawa 13 годин тому

    Fact Check: Terence Tao's parents are first-generation immigrants from Hong Kong to Australia. Which makes him Australian-American after he naturalised for duo citizenship.

  • @artefactxinnovation
    @artefactxinnovation 3 дні тому +1

    Thanks for the inspiring sharing! Let's leverage the power of AI to improve human lives and building a more sustainable future!

  • @oiuhwoechwe
    @oiuhwoechwe 3 дні тому +3

    but can it code crysis?

  • @En1Gm4A
    @En1Gm4A Годину тому

    The girl misrepresented Fourier

  • @ZeroD4wn
    @ZeroD4wn 2 дні тому

    Feel like 01 would make more sense for Agents vs chat Interfaces. GPT-4 is good enough for chat-interfaces. The next step is Agents, 01 with chain of thought is perfect for building agents that can control operating systems to achieve outcomes.

    • @Gpt6master
      @Gpt6master День тому +1

      What do you think about the speed? Would that be the next barrier to overcome?

    • @chrism.1131
      @chrism.1131 17 годин тому

      @@Gpt6master the next Google, will be a company that scans all the new agents and combines the best of the best.

  • @UNCLESPARK
    @UNCLESPARK 2 дні тому

    This has me thinking when we’ll be able to run these large models on a usb drive size computer ….

    • @Gpt6master
      @Gpt6master День тому

      Good thinking but not likely until “AI” helps human invent the sub atomic circuits to replace today’s semiconductor technology.

  • @AdamDahdah.
    @AdamDahdah. 3 дні тому +3

    A great Notification

  • @Escape_The_Mundane
    @Escape_The_Mundane 2 дні тому

    How do you solve quantum coherence multibody physics problems. I think they said a practically useful quantum computer has 200,000 to 1,000,000 qubits. They will probably need AI's help to solve such hard problems.

  • @zerge69
    @zerge69 3 дні тому +5

    It's quite obvious that in a very few years there will be no software industry. Your AI will be your only application.

    • @kangaroomax8198
      @kangaroomax8198 2 дні тому +8

      The only people who say this build neither software nor AI.

    • @ikleveland
      @ikleveland 2 дні тому +1

      Yes, but how you interface with that ai will be important. So its important to build a platform or network or processing company now. People will need platforms & how they interface will be different in the future. Device startups realize this, as Steve Jobs said, the best UI is no UI .

    • @zerge69
      @zerge69 2 дні тому

      @ikleveland Temporarily. Soon the AI will build the required interface on the fly, and will manage all networking tasks. The era of software is coming to an end.

    • @zerge69
      @zerge69 2 дні тому

      @@kangaroomax8198 but I do, my friend. That's why I KNOW.

    • @SomeHobo
      @SomeHobo 2 дні тому +5

      It's quite obvious that you are not in the software industry.

  • @tyc00n
    @tyc00n 2 дні тому +3

    I love the runtime paradigm of o1 but I sure do struggle to find a use case where its better than a solid prompt template and 4o, quite the opposite infact o1 gives me worse results most of the time

    • @mattie3875
      @mattie3875 2 дні тому +1

      I do like it better for coding tougher problems and debugging problems.

  • @jasperstoj
    @jasperstoj 2 дні тому +5

    It can't do 90% of knowledge work today. It cant even do basic second year accounting journals right.

    • @ycombinator
      @ycombinator  2 дні тому +5

      Skill issue

    • @Brain4Brain
      @Brain4Brain День тому +1

      It can, you just don’t know how

    • @drawnyuun
      @drawnyuun День тому +3

      Yeah this whole thing was odd. We didn’t have o1 then maybe, but we do have it now and it’s not even as good as claude. Certainly not the leap they seem to claim it is.

    • @AaronBlox-h2t
      @AaronBlox-h2t День тому

      @@Brain4Brain What do you mean?

    • @Brain4Brain
      @Brain4Brain День тому

      @@AaronBlox-h2t RAG

  • @PSpace-j4r
    @PSpace-j4r 3 дні тому

    What does it mean for our healthcare system? Perhaps an AI doctor assistant to every human one?

  • @Brahvim
    @Brahvim 3 дні тому

    0:35 I heard that as "then Windows will just be whoever builds the best...".
    _Well..._

  • @knowtes-app
    @knowtes-app 3 дні тому +2

    Compute is indeed the king. Obviously NVIDIA had a head start but I think there's so much room for decentralized compute that's not being used. AI's future will depend on decentralization imho

    • @hongyihuang3560
      @hongyihuang3560 3 дні тому

      this is unfortunately nearly impossible. weight, activation, and gradient exchanges are prohibitively expensive over ethernet (both bandwidth and latency). despite looking compute intense, language models have a even larger bandwidth than compute problem

    • @knowtes-app
      @knowtes-app 3 дні тому

      @ that’s interesting. Are there any papers on this problem? Or any companies that are trying to solve it?

    • @AlexWilkinsonYYC
      @AlexWilkinsonYYC 2 дні тому

      I think you would have to build some pretty novel tech where you're really trying to "chunk" the stuff out so each decentralized comp would be handling almost like a "vertical slice" of the model and just sending the results back at the end, but I would suspect that may not be possible as it likely needs data from the other nodes to properly train each stage. It's probably possible to some degree, but I would suspect it's not advantageous enough to compete with the current solutions. The only way to find out is to try it I guess. Maybe there's something there. 🤷‍♂️

  • @saang
    @saang 3 дні тому

    Ant Financial's fundraise was larger?

  • @BlueBirdgg
    @BlueBirdgg 3 дні тому +7

    Btw, a little thing. I think you guys may (or not) want to hide your feet. They sometimes dangling take a little of the attention of the content... And thanks for the content.

    • @Enedee007
      @Enedee007 3 дні тому +7

      Take your eyes off them feet and focus on the content. I don’t even know they were there. I would have never seen them, focus!!!

    • @aglitchinthematrix157
      @aglitchinthematrix157 3 дні тому

      Hahahahahaha!!!

    • @BlueBirdgg
      @BlueBirdgg 3 дні тому

      @@Enedee007 I love to know the brand of their shoes, don't you?

    • @Enedee007
      @Enedee007 3 дні тому

      Sure, I do! 😝

    • @ElfProduct
      @ElfProduct 2 дні тому +1

      @@Enedee007no I completely agree with him, now he mentions it, the feet are incredibly distracting.

  • @GoNoGoDIY
    @GoNoGoDIY День тому

    Open AI 4.0+ cannot work effectively at the same time successfully in multiple windows. It cannot retain a topic if you come back later, it's like talking to a new person. I'm ADHD this feels like my failure when any request cannot be met. It frequently cuts me off even though I pay for plus to come back in two hours that's some bs. I'm creating gold, shot down by you used too much come back in two hours 🤣lame

  • @sanesanyo
    @sanesanyo День тому +1

    I really get bugged by the name OpenAI now when we know its so closed that they dont even disclose any model related information in their technical reports.

    • @tracy419
      @tracy419 22 години тому +1

      I feel bad for you. It seems like such a pointless thing to obsess over.
      It takes money to do this stuff, and unless you are ready to fund their requirements, you should probably just get over it and focus on something that actually matters.
      One of their goals was to provide AI to everyone, and considering anyone with an Internet connection and smartphone or computer can use it for free, they are following through.
      But they said!!! Blah blah blah....
      Move on and be thankful they put this technology out into the world so that it's not being hoarded by the elite.

  • @boonkiathan
    @boonkiathan 3 дні тому

    problem is that
    while compute poring over data can progressively build more and better the 'routine' apps and systems we know and use today
    you will be stuck with the 60pt system
    with no inhouse ability to reach 61points and
    you have to wait for AI to be able to generate code that can get to 61
    and nobody of even average talent would care to, understand or work on that pile of code that was conveniently generated in minutes, even bother
    to push it to 61 - and that is a latent threat to the system development ecosystem

  • @wakenupforbreakfasttoday
    @wakenupforbreakfasttoday 3 дні тому +2

    Can you have a ten trillion parameter model? Can you extract that many parameters from the dataset/language? I honestly don't know, but at some point you end up with a larger portion of parameters that have very few (if any?) instances, no? (ie most columns are all NA).

    • @WearyTimeTraveler
      @WearyTimeTraveler 3 дні тому

      You can, but there’s no evidence it will enable abilities at a relative scale and the hallucinations would be wild

    • @John4343sh
      @John4343sh 2 дні тому +1

      @@WearyTimeTraveler That is total bullshit.

  • @dannyisrael
    @dannyisrael 2 дні тому +1

    What are we expecting to gain from more parameters? I’m not feeling a lack for parameters.

    • @Brain4Brain
      @Brain4Brain День тому

      Scale. The bigger the model, the smarter and the more well it generalize

  • @funmif
    @funmif 3 дні тому

    Excellent comparison with Fourier transform!

  • @Sketching4Sanity
    @Sketching4Sanity 3 дні тому +1

    LOVE ✊🏿

  • @perbhatkumar7033
    @perbhatkumar7033 3 дні тому

    Is this mono audio 😭

  • @WearyTimeTraveler
    @WearyTimeTraveler 3 дні тому +23

    You’re massively underestimating the problems with AI’s inaccuracy. You act like scaling will fix all problems when it makes hallucinations worse and you’re just pouring money into more transformers instead of fixing the problems

    • @ycombinator
      @ycombinator  3 дні тому +19

      Read the scaling laws paper

    • @NewSmithsonian
      @NewSmithsonian 3 дні тому

      hallucinations are barely an issue

    • @jmoney4695
      @jmoney4695 3 дні тому +7

      Hallucinations are and will continue to be a problem, but if models continue to get smarter and cheaper, you can use a series of models to fact check the output.
      Just like how when I finish a paper, I have several people proof-read it. Probably closer to peer review because you can have models check the logic, check the facts and sources, etc.
      Hallucinations do not seem like an intractable problem - just a hard one.

    • @williamparrish2436
      @williamparrish2436 3 дні тому +5

      I'm convinced at this point that these comments come from people that don't know how to use AI. They ask murky, unclear questions without realizing it.

    • @ashleigh3021
      @ashleigh3021 3 дні тому +2

      They require the ability to hallucinate just as the human brain needs the ability to form hypotheses by free association. So most likely the result will be multiple models for each step in the chain - free association > hypothesis formation -> testing by falsification.

  • @BT-xm7ti
    @BT-xm7ti 3 дні тому

    Incredible video and discussion. High quality debates. Congrats for the work and study

  • @pauldannelachica2388
    @pauldannelachica2388 2 дні тому

    ❤❤❤❤❤

  • @angloland4539
    @angloland4539 2 дні тому

    ❤️☺️🍓

  • @petersuvara
    @petersuvara 2 дні тому

    Scaling like this doesn’t work.

    • @Brain4Brain
      @Brain4Brain День тому

      Says random guy on internet…

    • @petersuvara
      @petersuvara День тому

      @ says random bot on the internet.
      Do your research, everything points to scaling as a logarithmic curve, particularly for CNN and Large language model networks.

  • @kucchbecho
    @kucchbecho 3 дні тому

    what's the use of your AI, when you accept Arab funding, add policies that hide the any islamic things from GPT, AI. Literally, GPT dont talk directly but give excuses like Mullah of ISLAM to legtitimise the atrocities by founder