End of the silicon era. Processors of the future

Поділитися
Вставка

КОМЕНТАРІ • 809

  • @blinded6502
    @blinded6502 2 роки тому +1128

    Quantum computers are NOT a replacement for the classical computers, holy hell

    • @DocWolph
      @DocWolph 2 роки тому +90

      A Classic Computer (C-Comp) becomes the I/O device to talk to the Q-Comp.
      Input-> C-comp-> Q-comp-> (back to)C-Comp-> Output

    • @cebo494
      @cebo494 2 роки тому +347

      ​@@DocWolph It doesn't matter what the connection is, quantum computers simply are not meant to do the same things that classical computers do. That's what the guy meant by "are not a replacement". They are not better than classical computers for most of the tasks that the vast majority of people use computers for, like web browsing, text/image processing, gaming, etc. They have little to no values for consumers. They will likely forever remain in labs, datacenters, and industry for highly specialized tasks and research.

    • @fridolinkoch
      @fridolinkoch 2 роки тому +123

      Until someone discovers something you can do with them as a general person….just like when the internet arrived

    • @Galacticfungus
      @Galacticfungus 2 роки тому +59

      @@fridolinkoch There are a few Quantum ray tracing algorithms with papers now.

    • @DocWolph
      @DocWolph 2 роки тому +33

      @@cebo494
      Well at this point it is implied, at least, that you can not directly interface with a Q-Comp. That is, you need some kind of bridging software running from the C-Comp to the Q-Comp and back.
      This is regardless of application. Aside from scientific, engineering, security work, among other things, I can see Q-Comps being used for animation, rendering, or simulation (for example, I can see water sims taking only a few minutes rather than many hours or days for a FEW high quality frames with a Q-Comp) and that is not even the color of the tip of the iceberg of what is possible with Q-Comps.
      For most things people use Computers for, C-Comps will be at least enough.THIS is agreed. But there are things that are way bigger that even best current C-comp technology just is not adequate for. and Q-Comps at a price a small studio, or a very dedicated hobbyist, can buy to radically accelerate their sim work, animation, and rendering work (albeit the software may be coming a few years after the fact) is just what the Studio head asked for. And that is JUST animation.
      Anybody can become a "Doc Brown". pursue science, mathematics, engineering and MORE at home.
      Again, I generally agree with what you saying BUT you are not thinking broadly enough.

  • @DunnickFayuro
    @DunnickFayuro 2 роки тому +622

    Very good content. I just find the desync between your video and sound a bit annoying.

    • @PareshPatel-xc2vu
      @PareshPatel-xc2vu 2 роки тому +226

      I think it's dubbed, probably from Russian

    • @DunnickFayuro
      @DunnickFayuro 2 роки тому +21

      @@PareshPatel-xc2vu Nope. Just desync. The lips almost match the sound, with a little delay.

    • @namthainam
      @namthainam 2 роки тому +20

      Nope video is all ai generate from stock video so it's all jibberish. Might as well listen with your eyes closed

    • @samlovebutter
      @samlovebutter 2 роки тому +120

      @@DunnickFayuro it is dubbed from russian. The original channel called "Мой Компьютер"

    • @vadulme8640
      @vadulme8640 2 роки тому +90

      @@DunnickFayuro Yeah, it is dubbed. I am the guy who recorded the dub. I synced it the best I could to the original voice, but I could only do so much. And yes, I get it that it's annoying.

  • @cursedgamer2778
    @cursedgamer2778 2 роки тому +132

    I love these dual language channels, I don't know Russian and would have never have been able to understand your phenomenal video otherwise. Thank you, truly.

    • @tankerock
      @tankerock 2 роки тому +34

      No wonder, I was wondering why my video/audio was out of sync lol

    • @ysa1023
      @ysa1023 2 роки тому +1

      Ah so that's why his disgusting face seemed so familiar and audio was out of sync! He is one of the idiots that supports war in my country. He is one of the people that thinks I should die... Yeah, hope at least somebody is going to see this comment, and stop watching this genocide supporter.

    • @mka2
      @mka2  2 роки тому +18

      Спасибо друг! Это очень приятно читать! Привет из холодной России!

    • @headSoup
      @headSoup 2 роки тому +2

      @@mka2 привет друг! я из Американский

    • @headSoup
      @headSoup 2 роки тому +2

      @@mka2 я говорю просто 😢

  • @ajbowers44
    @ajbowers44 2 роки тому +389

    So glad UA-cam recommended this video. Nicely done, great balance of information and presentation without coming off overly optimistic or pessimistic.

    • @Andrew-rc3vh
      @Andrew-rc3vh 2 роки тому

      Yes indeed, so far it sounds intelligent. I'm hoping I'm not going to be lectured about Moore's law. I'm convinced that is a marketing meme for Intel.

    • @hyll6700
      @hyll6700 2 роки тому +3

      Quantum Computing isn't a replacement.

    • @aqua-bery
      @aqua-bery 2 роки тому +2

      Quantum computers can't replace regular computers. Yes they could calculate some very complex equations at a fraction of the speed. But they suck at doing simple calculations.

    • @hyll6700
      @hyll6700 2 роки тому

      @@Andrew-rc3vh Bruh 😂

    • @Andrew-rc3vh
      @Andrew-rc3vh 2 роки тому +1

      @@hyll6700 I did watch it to the end on this occasion and it was indeed no-nonsense. Good job.

  • @supriyochakrabortybme
    @supriyochakrabortybme 2 роки тому +77

    A very interesting compilation, the UA-cam algorithm recommended this hidden gem.

  • @yamatsukami987
    @yamatsukami987 2 роки тому +35

    also worth considering in the mean time, their are computers that can run on trits (-1, 0, 1) which requires some fundamental changes, but could theoretically be more effective. This could even be extended further though it gets less practical the more you add.

  • @kevinmitchell3168
    @kevinmitchell3168 2 роки тому +213

    You mentioned photonics use in communications but didn't mention photonics switching. They've been trying to make purely photonic chips for decades, and it's always just around the corner much like fusion. It's that research that led to photonics being integrated into silicon chips. It seems like that research was very close to yielding results but focus and funding got taken over by quantum computing. From all I've read purely photonic chips would run 100-1000 times faster than silicon at far lower power and heat. Hopefully as silicon reaches it's limitations there will be renewed funding and research for it.

    • @Luizfernando-dm2rf
      @Luizfernando-dm2rf 2 роки тому +30

      It's a bit sad really... Quantum computing while exciting, is very far from producing anything useful in terms of actual computing. There has been no more than 10 algorithims made to run on quantum computers for example.

    • @id104335409
      @id104335409 2 роки тому +14

      The comment I was searching for. I was hoping the video is on that topic. Not quantum computing.

    • @bdjfw2681
      @bdjfw2681 2 роки тому

      isn't fusion have focus and funding but still no result ? is it really that close for purely photonic chips ?

    • @dinozaurpickupline4221
      @dinozaurpickupline4221 2 роки тому +4

      light switches have their application in networking & data transfer

    • @jadedandbitter
      @jadedandbitter 2 роки тому +8

      Intel's working on it. Have a buddy that works in the photonics dept. Can never talk about where they're at though.

  • @tony_T_
    @tony_T_ 2 роки тому +106

    Silicon photonics is what I have the most hope in for in the next couple of decades. Imo, transitioning from electricity to light is just the most logical step forward. It will set moores law back by quite a bit, but the insane clock rate of the processors will make up for it. Most modern keyboards already use light to transmit signals.

    • @HeLrAiSiNg1
      @HeLrAiSiNg1 2 роки тому +1

      well light transmision exists for decades now unless you mean signals like hid interface that uses speed bandwidth like in hz its known that for ex most audio devices ps5 xbox etc have a light slot on the back to connect with optical fiber to send sound now a kb uses a microcontroller if its mechanical that works like microcomputers with speeds of some mhz so either at the usb there is a decoder that reads the signal or the computer itself can read it which i think its the first to be honestly but the photonic is promising concept but dont expect pcs out of the world sure there will be decent speeds we already have 6ghz on new gens cpu with 10+ cores and in reality the gains will be diminishing if for ex you play games you might just squeeze a bit more performance but thats it unless they make games with the most realistic graphics to look like real life which i think we have some decades to achieve such feat and at the end its also depend on gpus will they be made with the silicon photon ? and if yes when

    • @tony_T_
      @tony_T_ 2 роки тому +14

      @@HeLrAiSiNg1 Had a really hard time understanding what you typed out. The point is that the speed of light, being about 100x faster than electrons, will counteract the size increase of transistors and gates with its sheer speed. This would give Moore's law a good amount more headroom to keep progressing.

    • @dinozaurpickupline4221
      @dinozaurpickupline4221 2 роки тому +2

      @@HeLrAiSiNg1 I ALWAYS THOUGHT THERE WAS MORE TO GHZ STORY,CANT THE CPU MADE TO THINK ITS OPERATING ON HIGHER FREQUENCIES?

    • @HeLrAiSiNg1
      @HeLrAiSiNg1 2 роки тому

      @@tony_T_ dream on what i wrote were about light transmision exists for decades unless as the original post states keyboards with light transmision the logical thing is the usb is a decoder and the keyboard an encoder that transmits light with the information of what is pressed anyway it sure will bring some improvements but nothing super super wow that will drop jaws

    • @HeLrAiSiNg1
      @HeLrAiSiNg1 2 роки тому

      @@dinozaurpickupline4221 nope it cant ? You see a cpu has trillions of transistors gates etc the more small they make em to cramp more transistors everytime the less electricity they can stand if you see older cpus were able to hit 8ghz with nitro but newer is yet to hit anywhere that cause they cant moores law thing is thats apply everywhere not only pc cause overclock on cpu is by increasing voltages on cpu to gain speed that heats it if you for example put a 12v 2A the fan will be fine as long its 12v stable while its amperage draw is only 0.2A versus 2A but if voltage goes 12.1v it will heat a bit and slowly burn the same goes to cpus no mater what you do the only option is to give it more voltage for speed which means it need better cooling there is no magic programm the best you can improve is buy a rx 6900 and flash an rx 6900xt bios if it exist it will make it work a bit better cpu doesnt have anything stored that makes it work the way it does

  • @mlytle0
    @mlytle0 2 роки тому +37

    Silicon carbide is used in power transitors, handles high temperatures and very high frequencies, well above silicon. It already has a supply chain, and looks like a candidate.

  • @josgraha
    @josgraha 2 роки тому +27

    oh boy, quantum computers are even further out than tfets, optical compute, or memristors and you still need a von-neuman silicon computer to work with it. thanks so much for the great video, fantastic questions and background material. much appreciated

    • @ararune3734
      @ararune3734 2 роки тому +15

      Not only are they far away from any practical implementation, but they will simply never replace your standard computers in the way people think. Quantum computers will be better at performing certain tasks, but those tasks are not in our every day usage, so for your personal needs, you won't use a quantum computer, you'll stick to the standard computer.

  • @ryox82
    @ryox82 2 роки тому +9

    What I like about quantum computing is it could eventually be used to work the problems we face much faster, like disease cures, and silicon limitations.

    • @halrichard1969
      @halrichard1969 2 роки тому +1

      No one wants to cure disease. That should be obvious. If even half a dozen major diseases were cured, it would result in Trillions of dollars in losses for the Medical profession, big Pharma and medical hardware industries. They only want to "treat" diseases because that is where the profits are.

  • @tombouie
    @tombouie 2 роки тому +21

    Well-Done, summary:
    SILICON: Above vs below a 5nm gate width silicon is discrete on eor off vs statisticial on&off (aka tunnel field effect)
    GERMANIUM: Germanium Has much better lab performance but silicon is extraordinarily more pragmatic/practical (availability/cost, heat dissipation/tolerance, oxidation, freq band, etc). However germanium might be modified to improve its characteristics (ex: Molybdenite in development, lab germanane, etc).
    CARBON nano-tubes (in development): Graphene is a one-atom-thick layer of carbon atoms arranged in a hexagonal lattice. A carbon nano-tube is a tube of graphene
    GALLIUM NITRIDE: has some better performance than silicon & can be manufactured with silicon based equipment/industry
    Better CMOS: design based on statisticial on&off (aka tunnel field effect) instead of discrete on eor off to lower power/heat. Only work for graphene & at super low temps.
    MEMRISTOR: A memristor is an electrical component that limits or regulates the flow of electrical current in a circuit and remembers the amount of charge that has previously flowed through it. Memristors are important because they are non-volatile, meaning that they retain memory without power.
    OPTICAL COMPUTING:
    QUANTUM COMPUTING: In theory quantum computing can find least-worst solutions to problems no-matter the number of potential candidates (ex: NP problems like the traveling salement, decryption/password breaking, etc)

    • @Luizfernando-dm2rf
      @Luizfernando-dm2rf 2 роки тому +2

      I like how you said nothing about optical computing xD

    • @tombouie
      @tombouie 2 роки тому +2

      @@Luizfernando-dm2rf
      *I wasn't trying to show-off. These was just quick notes to remind me of technology tends I was fuzzy-on (thks to this good video). I'm a retired physicist & am already familar with optics & optical computing.
      *There is Trillion$ ... Trillion$ of semiconductor electronic infrastructure around the-world & comparatively optic/quantum infrastructures are very small. Quantum computer infrastructure is small but growing much faster than optic computing infrastructure. So we just might jump from semiconductor dominance to quantum dominance computing.

  • @thenoobgamer9225
    @thenoobgamer9225 2 роки тому +6

    This video is more informativ than I thought. I love the web for channels like this.

  • @shipsey2
    @shipsey2 2 роки тому +8

    Great video I found today I don't know if it's just me but the video and audio was out and not synced

    • @vadulme8640
      @vadulme8640 2 роки тому

      It's actually a dubbed video and we didn't have the orignal voice track without background music baked in, so yeah, doesn't look particularly stunning. We'll fix that with new vids. I mean the really new vids. Those that are already produced, will have to be dubbed the same way unfortunately.

  • @invertexyz
    @invertexyz 2 роки тому +9

    The next frontier will likely be 3D chips. We already do build them with a few layers, especially memory chips which are dozens of cell layers thick now. We will need to do the same with processors, and also switch to more thermally efficient materials to help avoid the increased heat per sqr unit as a processor gets thicker (building heat exchange tubes into the design itself can also drastically help combat this).
    A 3D design also opens up pathways for much more optimized computation methods that a chiplet design does not really lend itself well to as it scales. Chiplets primarily communicate to eachother through side channels, whereas a 3D chip would simply be one single complex chip of interconnected logic in all directions.
    The potential performance gains as we add layers is insane to think about, when you consider how thin a given layer of a processor is. You could have hundreds of thousands of layers eventually. A single processor rivaling billion dollar super computer warehouses of today.

    • @karotgamin790
      @karotgamin790 2 роки тому +1

      da cube chip

    • @dylanhecker6686
      @dylanhecker6686 2 роки тому +1

      Would the AMD 5800X3D be considered a 3D chip?

    • @invertexyz
      @invertexyz 2 роки тому +6

      @@dylanhecker6686 Not really, it's only the cache, which is essentially on-die RAM that is stacked, which was already being done for your SYSRAM too. The core logic is still single layer.

    • @dylanhecker6686
      @dylanhecker6686 2 роки тому +1

      @@invertexyz thanks!

    • @Pythoner
      @Pythoner Рік тому +1

      Well we already do have '3D chips'. A processor consists of many layers of circuits stacked on top of each other.

  • @pneumantic6297
    @pneumantic6297 2 роки тому +9

    I don't see quantum processors replacing processors but if they become cheap enough I can definitely see them becoming a new component of the computer. If they do replace anything it will probably be graphics cards. All in all, we probably won't see quantum computers hit mass market until probably decades. What is more likely to happen than most theories is the integration of supercomputers, powered by more rare metals, that use the internet to give you your computer as a service rather than owning a computer. The system we have will be a lot more stupid and only decode the sent information.

    • @nyalan8385
      @nyalan8385 2 роки тому

      Yeah I feel like cloud computing will be big, but hopefully the next big "breakthrough" will be just moving away from the x86 architecture. This should buy us a couple more decades at least before we have to start replacing silicon, or whatever other new innovation occurs during that time

    • @Jack-he8jv
      @Jack-he8jv Рік тому

      god i hope not, i hate this rent mentality.
      fuking WEF kikes will better write their wills if they try this.

  • @luketurnbull5648
    @luketurnbull5648 2 роки тому +5

    I just saw this is a dub channel of another. I didn’t know I was looking for content like this and I love it. I’ve only seen dub channels go to other languages from English so it’s really cool to see that it actually does work!

  • @khlorghaal
    @khlorghaal 2 роки тому +3

    "that is, if it continues to exist for us, in this line of events"
    that hit hard

  • @kykykyykykyk2851
    @kykykyykykyk2851 2 роки тому +5

    finally yt algorithm doing his job well

  • @lamebubblesflysohigh
    @lamebubblesflysohigh 2 роки тому +9

    The biggest problem with Quantum computer is the lack of software. Writing code for them is so complicated it practically eliminates 99% if not more people who currently make their living by writing code. That means less developers and thus less development. Unless coding will become heavily AI assisted.

    • @timothywilliams8530
      @timothywilliams8530 2 роки тому +2

      I'd say the biggest problem with them is that their cooling systems are the size of a room but, ya know.

    • @amentco8445
      @amentco8445 2 роки тому +1

      @@timothywilliams8530 Seeing how much of a mess it's been just to get quantum computers to the point they are now, I have doubts about all their pros in general.

    • @lamebubblesflysohigh
      @lamebubblesflysohigh 2 роки тому

      @@timothywilliams8530 My mom worked with a computer that took an entire room before I was born. Her company actually had 2 of those. Size, price, power consumption do not matter if it has a purpose... a profitable use. Current quantum computer have no use whatsoever... not because they are bad but because there is no software that would make them better than standard computers for them. And there is no software because there is only a handful of people who can write some and they they probably call each other by their given names ;-)

    • @abduking.
      @abduking. 24 дні тому

      i say the biggest problem is they barely exist....(YET). We still count qubits in the 100's and there quantum property dont last at all .
      Software is nothing i promise you as soon as quantum computers are truly viable software will be created asap.
      software has little to no effect on quantum computing development big corps like google are just throwing money on it like its just around the corner when its the furthest from that

  • @thebogsofmordor7356
    @thebogsofmordor7356 2 роки тому +2

    Ayyy I am your 1000th subscriber! Great video

  • @chrisglaze658
    @chrisglaze658 2 роки тому +16

    Really good run down of the limits of silicon. Great research and presentation!

    • @Polar_Onyx
      @Polar_Onyx 2 роки тому +1

      great research? most of the information given was a vast oversimplification or just outright wrong. I wish there wasn't so much dishonesty in tech channels. Being able to accurately depict advanced technology is difficult and it's much easier to make it sound fantastical and get more viewer retention

  • @RWBHere
    @RWBHere 2 роки тому +7

    4:56 Not quite true. I have some Germanium transistors here which were used at 10.7 MHz, and some other low power ones which were used at over 100 MHz. They were expensive, but very capable at low currents. Bipolar Silicon transistors have been made which would amplify at frequencies close to 20GHz (I made some of them, back in the 1980's.) For higher frequencies, JFET, then MOSFET, IGFET and other types are needed. But yes; Germanium transistors are definitely very limited at higher frequencies, in comparison to Silicon devices. Leakage currents are very problematic, and they lead to high noise levels. It would still be interesting to know how a Germanium MOSFET would perform at somewhat higher frequencies, however.

  • @KingSteven77
    @KingSteven77 2 роки тому +2

    Good video but the video and sound is not synced and it's annoying but other than that it's 10/10

  • @Daojyn
    @Daojyn 2 роки тому +8

    Very good video, but I have two questions:
    -I wonder if it took around a half a century to reach this performance on silicon/CMOS doesn't it mean that it would take around the same time for others technologies to catch up or do improvements already discovered accelerate the rest?
    -After the huge rise of chip price in the last years can we expect before others technologies are catching up, let's say in the next decades that chip prices will decrease a lot because of the silicon limit that meanwhile we will see a comeback of dual CPU and SLI/crossfire for PC to keep improving performances?

    • @Jack-he8jv
      @Jack-he8jv Рік тому

      amd chiplet design and intel vastly inferior big,little design is already a form of dual cpus.
      sli is very unlikely to ever come back due to how frustrating it was for everyone.
      for gpu its most likely gonna be software improvements.

    • @joefish6091
      @joefish6091 Рік тому

      RISC-V is the future.

  • @Notfound4747
    @Notfound4747 2 роки тому +5

    hey i realy liked the video nicely balanced and no stale moments, i got the random recomendation video in auto play and didn't regret a BIT XD

  • @xxamulyaxx
    @xxamulyaxx 2 роки тому +2

    UA-cam recommendations at its finest.
    Keep up.

  • @qm3ster
    @qm3ster 2 роки тому +3

    Dear Ivan. At 9:05, there's one frame where the cmos construction steps are shrinking, and the diagram there is not translated from Russian.
    Please enjoy it!

  • @ProDigit80
    @ProDigit80 2 роки тому +3

    Overheating becomes less of an issue with the reduction of transistor size. The only reason modern transistors are emanating more heat than older models, is because manufacturers make use of that reduction in heat output, by increasing cores, core frequency, and core complexity (more transistors per core).

    • @cj09beira
      @cj09beira 2 роки тому +1

      that's only happening because the transistors no longer use less power when Shrank at the same rate of the volume reduction which used to be the case now a 2x volume reduction only leads to 20-30% power improvement, which is not enough,

  • @meppeorga
    @meppeorga 2 роки тому +2

    We still got some ways to go with classical architecture, you can also gain some performance by improving how quickly we can access memory and how much of that fast memory we got, there are also architectural adjustments with concepts like RISC, bigLITTLE, vcash, new schedulers in OSs etc. so both on hardware and software level.
    With many new technologies being developed like Quantum, DNA-computing and Optical-computing, I don't think these will replace classic computer but with interconnects between them then they can work in tandem in the tasks they are best at.

  • @stormevans6897
    @stormevans6897 2 роки тому +1

    First time I've ever seen a channel translated like this, cool

  • @larrysal8866
    @larrysal8866 2 роки тому +3

    Can't wait to get my own silicone chip manufacturing system for cheap in a few decades B)

  • @squigglesmcjr199
    @squigglesmcjr199 2 роки тому +1

    Glad i found this channel premium content

  • @human_shaped
    @human_shaped 2 роки тому +6

    Nice job researching and pulling all these alternatives together.

  • @theosib
    @theosib 2 роки тому +2

    Years ago, I did a calculation to determine when the speed and power benefits of process scaling would no longer offset the process variation. I got 5nm for that number as well.

  • @GrandNecro
    @GrandNecro 2 роки тому +2

    i think just coming up with better architechture and firmware is the easier route for now than finding a replacement for silicon

  • @BadgerGamePlay
    @BadgerGamePlay 2 роки тому +1

    only ~800 subs oO such high production, nice video man cant wait to see more!

  • @WahteverYouWant
    @WahteverYouWant 2 роки тому +2

    Dam man you need to give credit to Eugene Khutoryanski for using his graphics

  • @onlinetech8077
    @onlinetech8077 2 роки тому +1

    Superb man👍💐👌♥️ #Mycomputer

  • @Wbfuhn
    @Wbfuhn 2 роки тому +3

    Look at it this way. Human DNA is 2nm. To achieve human DNA size would be not only astounding but also open the potential for micro robots. At some point all manufacturers will have no choice but to develop more cores to handle more tasks. In addition to this, it might be possible to develop cybernetics and simulate human nerves.

  • @JamesHill-vs4kn
    @JamesHill-vs4kn Рік тому +2

    Thank you! We support you.

  • @ouch1223
    @ouch1223 2 роки тому +1

    as an old viewer of yours, i'd like to wish u luck with ur new channel :D

  • @avetruetocaesar3463
    @avetruetocaesar3463 2 роки тому +3

    I wholeheartedly disagree with the concluding section of this video. It insinuates that quantum-cloud-compute backend reliant hardware, which are about as functional as a storage device with some added connectivity, but mostly hollow toys which will serve as GUIs is the future of computers. I reject such a plane of existence, even if there is simplicity, or even necessity, in it; such a paradigm shift would basically mark the end of personal computers and ownership of anything related to computers. Quantum computers may as well exist in their own space and continue to get more sophisticated; they will still require close-to-0-Kelvin temperatures to even function, so an average person cannot carry one such device, let alone drive one in a home setting in today's society. I cannot visualise a future in which these devices become so compact, efficient, easy-to-manage and affordable as to be the de facto standard for computers because somehow, for some godforsaken reason, they are the only way going forward. How dreadful.

    • @Luizfernando-dm2rf
      @Luizfernando-dm2rf 2 роки тому +2

      Ignoring the severe limitations Q-computing has as of today, the prospect of not being able to own a full computer is actual garbage. I'd rather stay with my classical slow piece of junk than having to trust big techs and cloud providers.

  • @MonkeyManMechanicals
    @MonkeyManMechanicals 2 роки тому +3

    Thank you. That was very informative with a thought processing progression that was spot on without someone knocking on your door. I've been imagining crystal tech ever since I learned about piezoelectric when I was 5. Magnetic cooling, heating and generators are my insanity. It's all about switches and ball bearings. Space and Storage. I still don't know why they just didn't cut the 0 into an 8??? Cheers

  • @gabriel2fan
    @gabriel2fan 2 роки тому +1

    Goood video man, best wishes from Switzerland

  • @GuigEspritDuSage
    @GuigEspritDuSage 2 роки тому +3

    According to some scientists, quantic computer may never become real fully capable computers and instead became at best some kind of accelerator or limited to very specific tasks.

  • @patelhitesh-ux2uf
    @patelhitesh-ux2uf 2 роки тому +2

    Very informative and explained in very interesting way. Thank you.

  • @florenciabruck
    @florenciabruck 2 роки тому +2

    Great information and you are very carismatic you can give tones of information without feeling overwelming.

  • @Mew178
    @Mew178 2 роки тому +2

    Good new channel subbed. Well done.

  • @PovertyHelping
    @PovertyHelping Місяць тому

    That's completely truth. Thanks so much for online classes. We really need enough resources. 🏭👨🏼‍💻🎮🤖🏡

  • @radomiami
    @radomiami 2 роки тому +4

    If I were to guess which one would become the first step in a post-silicon CMOS world, it would probably be a room temperature tunnel field effect transistor.

    • @SupaKoopaTroopa64
      @SupaKoopaTroopa64 2 роки тому

      I'd put my money on silicon/photonic hybrid processors. There are already ASICs out there using photonic processing which are hundreds of times more energy efficient than silicon equivalents. We already have the technology, it's just a matter of integrating them into a full CPU/GPU.

    • @absolute___zero
      @absolute___zero 2 роки тому +1

      ​@@SupaKoopaTroopa64 photons are mass-less packets of energy and thus it is very difficult to make them interact between themselves and with any piece of matter we are made or our chips are made of. that's the reason why we still use CMOS transistors with gate delay of 30 picoseconds instead of using light and get attosecond "gate delays" with "light transistors", if they would exist. Everybody has it very clear, creating a chip purely using light will make it millions of times faster than silicon chips. The reason why light is so fast, it is because it has no mass. But you can't make a flip-flop out of photons. That is why accelerating CMOS chips with analog-style logic (implemented using interference, like some startups are doing right now) is a bad idea. It is like putting a F1 engine on a bicycle and expecting it is going to do speeds of 400 km/hour. The bearings of the bicycle wheels will simply melt because they aren't designed for such high RPM (not talking even about aerodynamics here). It would be a good idea for a college student working in his garage, but not for a company that rises 100 million or so for this kind of project. I suggest you to learn quantum physics and chip-design before you invest. The rate of bankrupcy of new companies is like 9 to 10, if not more. Only understanding the physics of the computation you can make a fair judgement of which company is going right direction or not. There are always tons of options to invest but only few of them are real opportunities.

    • @SupaKoopaTroopa64
      @SupaKoopaTroopa64 2 роки тому

      @@absolute___zero I'm aware of these limitations. I'm just saying that production-ready photonics already exist, so they have a head start over many other technologies. Also, I don't plan on investing in any photonics startups, or even anything in the microprocessor industry right now, I was just saying that If I had to make a bet on which of these technologies would first appear in a consumer product, I'd go with photonics.

    • @absolute___zero
      @absolute___zero 2 роки тому

      @@SupaKoopaTroopa64 photonics of course has future, but not as co-processors, rather a standalone computing logic with only a few CMOS. But there are many other computing methods that have been developed by scientists but not exploited yet. For example, there are mechanical integrated circuits made of mechanical relays (they achieve 10 nanosecond times to switch on/off), or there are gold transistors working in vacuum like vacuum tubes, there are DNA computers, and so on... the next big idea will be a new computing paradigm. There is also lots of work to do in software to gain performance, for example, development of massively parallel operating system.

  • @Jalae
    @Jalae 2 роки тому +1

    i thought i was having a break in my consiousness when the mouth didn't match the sound

  • @diegoochoa572
    @diegoochoa572 2 роки тому +3

    Bro amazing content!

  • @19vangogh94
    @19vangogh94 2 роки тому +5

    Silicon era will continue for at least another century m8, anything else will most likely be 10x-1000x more expensive and used in niche applications.

    • @paulssnfuture2752
      @paulssnfuture2752 2 роки тому

      yeah, until any tech becomes consumer grade and priced like silicon/lower it won't be viable replacement anytime within the century

    • @cj09beira
      @cj09beira 2 роки тому +1

      @@paulssnfuture2752 which is very hard to do considering right now the base material is simply sand, and we got extremely good at making perfect crystals of Si, good luck beating either of those things.
      i think we will just have more and more silicon around the cpu ( more cache, integrated ram, accelerators etc)
      then we might get a cpu where one of the more important bits is a different faster material and the rest stays silicon

  • @antonnym214
    @antonnym214 Рік тому +1

    Check into computing with Ternary. It is mathematically more efficient than binary and requires just one extra voltage level. This could lead to more efficient and faster CPUs than we have today and some are already being built. All good wishes.

  • @devbites77
    @devbites77 2 роки тому +5

    Thanks. I enjoyed your fascinating journey into the chip realm, exploring the potential future.

  • @kaiperdaens7670
    @kaiperdaens7670 11 місяців тому +1

    During the video I was thinking what if we make a superconductor chip?

  • @John-vw4qz
    @John-vw4qz 2 роки тому +1

    Thanks sir. I never knew this information... Amazing developments..

  • @markvisser388
    @markvisser388 2 роки тому +3

    The Japanese have made it possible, and in mass. Smart move abandoning silicon 20 years ago to focus on developing GaN. Hands down!

  • @AzErLoc
    @AzErLoc 2 роки тому +3

    Silicon is also relatively easy to work with and process.

    • @leonel9110
      @leonel9110 2 роки тому

      Exactly and they’re working on 3nm already

  • @samirkumargupta3502
    @samirkumargupta3502 2 роки тому +1

    Awesome video, thanks.from India.

  • @jmr1920
    @jmr1920 2 роки тому +1

    good content. i really enjoyed it up to the end

  • @AlephCasara
    @AlephCasara 2 роки тому +2

    Excelent channel, you've got a brazilian subscriber here

    • @mka2
      @mka2  2 роки тому +1

      Olá! Obrigada!

  • @nathanwest2304
    @nathanwest2304 2 роки тому +2

    I really think that graphene is the right way to move forward, we don't know how or when, but ultimately they will be the solution

  • @DuckyThePilot
    @DuckyThePilot 2 роки тому +2

    your audio isnt lined up with the video.

  • @JakubHohn
    @JakubHohn 2 роки тому +3

    Thank you, I was lacking any news about memristors for about 8 years, so I stopped looking for updates. It is great to hear there is finally some progress

    • @mnomadvfx
      @mnomadvfx 2 роки тому +2

      The main problem with memristors is the success of and investment in NAND flash memory.
      The advent of 3D multi layer NAND devices allowed an otherwise dying technology a new lease on life in the middle of the last decade, and it is still pushing forward.
      Even though ReRAM / memristors would be far superior in power consumption, latency and speed they are still leagues behind NAND flash in density even at the 2D level, and at the 3D level it's not even worth mentioning.
      Weebit is making a headway, but it will be years longer before we see anything truly commercial come out of it.

  • @infinnite4938
    @infinnite4938 2 роки тому +2

    Nice video! I thought you would have 90k subs at least, but only 900!

  • @roboticroom7283
    @roboticroom7283 2 роки тому +2

    GE is used in military industry that's why. + Gallium arsenide is used in military because chips made from that materials has lowest fail rate and extended temperature limits.

  • @richardnorbron4969
    @richardnorbron4969 2 роки тому +1

    Great video. Thanks for posting.

  • @claudiohase296
    @claudiohase296 2 роки тому +3

    Very GOOD !!!!

  • @bakedbeings
    @bakedbeings 10 місяців тому +1

    There's good news coming out of georgia tech this week about silicon epigraphene transistors, compatible with current manufacturing methods 👍

  • @alexvincheu
    @alexvincheu 2 роки тому +1

    This thinking is what we humans do 🤔 and how we achieve it! Just by solving problems one by one. Then we wait for another or more problems and we solve them again, and again. That's the beauty in science and what our world represents.

  • @gogollepremier6969
    @gogollepremier6969 2 роки тому +1

    Glad to meet ypu Mihaïl

  • @tombouie
    @tombouie 2 роки тому +1

    Thks again & request you update your most excellent video once a ~quarter.

  • @ZMacZ
    @ZMacZ 10 місяців тому

    When it comes to silicon being a pillar of computing, that will never really go away.
    Why ? Mass computing. Basically, when wanting to do computing en masse, given that
    power usage is not a problem with enough silicon-based photovoltaics in mind,
    computing can still use mostly silicon-based computing for when computing something
    that requires so much computing power that using the more scarce resource based
    computing options become prohibitively expensive.
    It's like creating a very large computer that can be like a workhorse, while using the more
    scarce resource based computing option for small applications.
    The workhorse can be used as an off-site computing 'monster' to offload work on.
    This may seem rather inefficient, but given that electricity can easily be farmed with solar voltaics,
    the power requirements are a lesser drawback, while the mass computing becomes cheaper
    over time, and can yet benefit from increased performance.
    Once people go into space and build space stations, remote computing this way can be
    much cheaper than supplying the more scarce resource based computing options for all/everyone.
    The problem is not one about creating a few million small computing options that require scarce
    resources (and thus becoming expensive), but once the newer options need to be created
    by the billions, yearly.
    The amount of mining required to get enough of the scarcer resource would be bad
    for the environment in such a fashion that humanity would 'compute' itself to death.
    Since, computing to death does not compute, en masse computing with the cheaper
    silicon option makes more sense.
    Also notice that en masse computing is much more efficient. When a resource isn't used
    by someone, someone else may be allocated more computing resource aiding his or her
    task to complete more rapidly.
    When it comes to remote computing for things like gaming, this may also work well.
    Silicon will always stay the best and easiest resource for computing, since it's most abundant.
    Also, the new computing options will have their own size limitations, albeit somewhat smaller.
    This would be a fixed factor.
    4:28 Yes, the cost of fabrication of non silicon processors is much higher due to the scarcity
    of the required materials. This will never change. Also, due to the limitation of availability
    the newer processors would be fabricated in much smaller quantity, while being much more
    expensive. Due to rising cost of scarce materials the cost over time would rise, not become lower.
    For regular silicon-based processors, these would become cheaper over time.
    Combined with photovoltaics the increased power usage would not be the limiting factor.
    13:56 Photonics are best used for transfer of data, not computing. Extreme short relays,
    let's say between chip layers and external to chip pathways would be an option,
    since these can transfer data really fast, and yet use very little power,
    which results in less heat release.
    Photons travel light so to speak while electron travel is heavy. This means less heat release.
    There's one problem though, the creation of very small light emitters/receivers
    and very small fiber optics or such is much harder than a regular circuit.
    It's at this point in time at least not very useful. A similar thing is with the wattage per computing.
    If the wattage becomes low enough, further reduction of such becomes less useful,
    and with that only extreme high end computing purposes would be served by further reduction.
    Let's examine an example, where you can have a fully capable processor using a 10 cm by 10 cm
    solar panel for continuous usage. What would the difference make if it's then operable
    by an 8 x 8 cm solar panel instead, by increasing the cost tenfold ?
    It would mean less power used, yes, by as much as 33%, but then the tenfold cost ?
    So, at some point in time the resource cost will define the computing solution,
    rather than its capability, once again referring to the scarcity of materials.
    14:46 Photonics will not remove the limitations of distance. Any signal will still be limited to
    the speed of light, whether electronic or photonic based. Latency stays the same no matter
    which of these mediums are used. Basically, the best way to make processors smaller
    and faster, currently, is to provide on-chip memory, with comparatively slow memory being
    used for the externals. On chip memory would increase processing by removing much of
    the latency of memory instructions, like cache does, but then used on a larger scale,
    with only one cache and larger memory area. Then you'd have 16 GB base memory on the chip,
    with like 4-8 megs in cache. When a large-scale external reference would occur
    (like writing to SSD) the memory controller would copy the on-chip memory to the external one,
    subsequently transferring it to SSD, while optimizing the transfer speed, between on-chip
    and external memory. On chip the memory could have the same latency as regular cache,
    but transfer between the on-chip memory and external (DDR) memory would still be many
    times greater than the transfer speed between regular memory and SSD.
    Also, when using stuff like virtual memory on an SSD, this could then be replaced by
    the external (DDR) memory, increasing computational output per unit by that much as well.
    Knowing that most of the CPU's and GPU's time is wasted on waiting for memory return,
    you know that this transfer of operations from external memory to on-chip memory
    can make a big difference, even when the computational output of the processor
    is yet smaller. Solving each bottleneck in turn is also a way to speed up computing,
    in some cases making a bigger difference than making it smaller.

  • @captntunabeerd
    @captntunabeerd Рік тому +1

    Fantastic video ! Sub :) Only thing, maybe change the backround music for something more enjoyable than this danish mystery murder chasing scene :D

  • @antonnym214
    @antonnym214 Рік тому

    Very nice reporting. I subscribed immediately. Thank you!

  • @kwgm8578
    @kwgm8578 10 місяців тому

    Even if Moore's law may not apply to Si applications, they are still very useful in solving many of our computational applications at 16 and even 32 nm technologies.

  • @FardinMirza
    @FardinMirza 2 роки тому +1

    is it me or the audio is not synced to the video

  • @SomeRandomPiggo
    @SomeRandomPiggo 2 роки тому +3

    700 subscribers is way too low for a channel of this quality!

  • @kensuiki6791
    @kensuiki6791 2 роки тому +2

    Guess we should become computers and enter the Ghost in the Shell era.

  • @olagarto1917
    @olagarto1917 2 роки тому +3

    Photonics can also do prossessing via constructive/distructive wave interference

    • @agsystems8220
      @agsystems8220 2 роки тому

      Not really, as it messes with phase information. You can build simple gates, but as soon as you try to do something like an and gate either your 1+1 value is different amplitude from 1+0, or its phase is different. This is fine if you are just throwing it at a detector, but if you try to throw it into another gate built around interference it is not going to work right. There might be some quantum mechanical way to align the phases, but the naive approach does not work.

    • @SC-zq6cu
      @SC-zq6cu 2 роки тому

      Photonics cannot store memory yet as that would imply being able to trap light for very long periods of time which is still impossible.

    • @JorgetePanete
      @JorgetePanete 2 роки тому

      processing*

  • @Acharya_Slideshare
    @Acharya_Slideshare 2 роки тому

    Some channel does not need views they already have best views

  • @ctsd623
    @ctsd623 2 роки тому +5

    Re: "silicon being near the end of its life cycle," this is just sensationalistic journalism (or an uneducated remark). It's not going anywhere in our lifetimes due to the deeply entrenced and highly refined manufacturing process. More advanced materials will have a much higher and impractical cost of scale. The most certain outcome is slowly more advanced materials will be used in conjunction with silicon, but it isn't anywhere near ending its life cycle, not for a hundred years or more. We can even do photonic and quantum compute on silicon.

  • @projectw.a.a.p.f.t.a.d7762
    @projectw.a.a.p.f.t.a.d7762 2 роки тому +2

    I was watching how atom's are being controlled and manipulated on 2d materials. That would suggest we have a ways to go, regarding Moore's Laws future.

  • @aaronsmith593
    @aaronsmith593 Рік тому +1

    Silicone has so many uses. Pretty soon humans will completely turn into their favorite element.

  • @omarandati874
    @omarandati874 2 роки тому

    Continue your research for 6G you are almost there!

  • @ma2i485
    @ma2i485 2 роки тому

    Great video bro

  • @francisdebriey3609
    @francisdebriey3609 Рік тому

    Very good ... subscribed !

  • @MyReviews_karkan
    @MyReviews_karkan 2 роки тому

    I'm just waiting for a future without passwords.

  • @mr.electronx9036
    @mr.electronx9036 Рік тому +1

    I think the "next thing" will be photonic cpu or/and ai neuromorphic chiplet, which both are extryemly fast.
    The issue is still a silicon as a substrate. I heard that specific "glass" can be used even more better than silicon

  • @thetruthexperiment
    @thetruthexperiment 2 роки тому

    That’s a pretty sweet logo.

  • @linolafettatnet
    @linolafettatnet 2 роки тому +3

    Your voice is out of sync with the video, that is a bit jarring to watch. Otherwise nice and smooth presentation.

    • @linolafettatnet
      @linolafettatnet 2 роки тому +4

      @@dadarkmatterdude This makes sense, i did not even notice it. Great voice for the dub!

  • @SuperSnotgun
    @SuperSnotgun 2 роки тому +2

    Silicon still has a way to go before its phased out. The key is better packaging (things like 3D stacking).

    • @masoudfasihi4342
      @masoudfasihi4342 2 місяці тому

      3d stacking will exacerbate heat dissipation, though, so it is not a valid solution

  • @sctsmith
    @sctsmith Рік тому +1

    Still waiting for a so called quantum pc to actually do anything..

  • @fk319fk
    @fk319fk 2 роки тому +1

    This is what I like best about UA-cam, you got me thinking...
    1) The way we are doing things now is coming to an end.
    1a) we may get smaller features, but we are close to the limit
    1b) due to size, we may find different circuits that get better results, but again an incremental change
    1c) chiplets, and other die layouts, can give up more performance, but nothing major.
    2) We need to think about pipeline, Risc/Cisc/Vliw, process.
    2a) my personal opinion is we need to rethink compilers and CPU so they make optimal code, not optimizing compilers for a CPU
    3) The software process is broken
    3a) with ~8m lines of code for Linux and ~50m for Windows 10, something is wrong
    4) I am a product of 16-bit words and 64K of memory.
    4a) My OS was 4k
    4b) I could run a word processor or a spreadsheet
    4c) my machine was 10 MHz
    5) when we buy a new machine, we generally get all new things
    5a) New OS do NOT need to be backward compatible
    I really think we need to look at how we move the software to a CPU and fix that issue.

    • @cj09beira
      @cj09beira 2 роки тому +1

      that was a different time, with complexity many orders of magnitude simpler, no need to support a hundred different processors, and thousands of different IO devices, yes it has to be mostly backwards compatible or else you waste billions of dollars of software development for little to no gain.
      chiplets still has a long road ahead, we can go really far with it specially when we start stacking chiplets.

  • @johndoh5182
    @johndoh5182 2 роки тому +1

    Umm, the death of silicon as a base for ICs is GREATLY overstated.
    There is this thought about some END coming that just fails to think clearly about what will happen. So let's evaluate some timelines.
    Sure, silicon is going to hit a limit, but so would any other material. If you want an IC based on on/off transistors, there is only so small you can make the transistors, but once you hit that theoretical size, is that the death of silicon. The answer is no.
    YOu can measure CPU performance in multiple ways. One is raw performance. Raw performance is based on work done in a certain amount of time, but to calculate this you use multiple types of applications to calculate this performance. If I want a render machine, do I need a CPU that's great at gaming? The answer is no. I need something that's great at stream processing, and that could be a combination of CPUs and GPUs and accelerators, or a computer that integrates all three in a single circuit to run any type of stream processing programs. Maybe with the accelerators I can program it first before running the job, so the first part of the render application checks for an FPGA and if it exists it programs the accelerators before running the render tasks.
    Next, at the beginning, which is as far as I got because these kind of videos are silly, the talk is on chiplets and IPC (Instructions per clock cycle). What was skipped is making a process node much more efficient at running higher speeds. There's no law that says a CPU can't be clocked over 6GHz. In fact I think over the next 5 years TSMC and Intel will both have process nodes that run 6 - 6.5GHz and it won't be very high power consumption.
    Also what isn't mentioned is compute power in 3 dimensional space. We're getting to the point to where ICs can have multiple layers, and this is for different reasons, one being the shrinking of transistor size and every shrink usually comes with power reduction. I think it's safe to say that CPU design is going to involve 3D layouts and this will involve layering in such a way to where one layer isn't producing so much heat and the other layer can, so planning out the layout of circuits will become more complex.
    But here is the REAL meat and potatoes of this. Home users already have access to so much compute power it's crazy. A home user can now do with large servers had to do 10 years ago, and in another 5 years the amount of power that can exist in a laptop is competing with servers from a decade ago. The move to 2nm or Intel 20A which are the same thing is SO monumental that for home users there's no need to get past that. The issue won't be not having enough compute power, but having too much with rare exception. One area of growth is PC gaming. Graphics compute still has to improve to get high quality 4K gaming. But, we're there NOW. The GPUs that AMD and Nvidia are releasing right now are good enough for high quality 4K gaming, and this is using a 5nm node or modified 5nm node (TSMC N4). When semi-conductor companies get to 2nm the issue won't be not having enough transistors anymore. It will be hard to keep adding more compute units (graphics processors) and in fact for Nvidia GPUs they're using multiple types of compute in a single GPU, and that's basically the raw compute for making the image on the screen, a lighting and shading technique used for games called ray tracing for there are RT cores, and then the cores for AI, tensor cores. 2nm is SO small with SO MUCH transistor density that home computing just never needs to move to anything else, but there will be other nodes that come out with better transistor density, so like 18, 16A, etc.... and this can probably go all the way down to 10A. By this time we're about 2 decades from now.
    The REAL issue isn't going to be transistor density. It WILL be about making machines smarter, to where they can do more with the same number of transistors. IPC isn't a gimic. It's how you get more performance in a fixed time period from a CPU over an older CPU at the same clock speed.
    So, there are many tools for improving compute power, and they are NOT gimmicks to hide some flaw. It's PART of better engineering for computing.
    One is more processor cores. For home computing though you hit a limit because most programs home users use don't have so many processing threads and adding ever more cores is a total waste. For servers on the other hand, there are many types of server compute loads that can benefit for a LOT more cores.
    Another is clock speed. But this has to be solved while at the same time keeping power consumption down. THIS is a real issue and it's hard to solve, but it IS getting solved slowly, but every process node is different so you have to solve that problem for every process node.
    Another is instructions per clock (IPC). This also improves over time and one of the ways to solve this is simply to have more cache so the cores have more data and instructions next to the cores instead of having to access system memory to get data/instructions. But there are other ways to improve IPC too.
    Another is making better processors that can accelerate workloads. AMD, Intel and Nvidia are all working on this, and even Apple has done excellent work with acceleration and that's part of what makes the Apple M1 and M2 processors so fast and NOT the fact that it's based on ARM instead of X86-64.
    Another is being able to stack transistors which shrinks the area needed for the same circuit.
    Another IS chiplets. This helps with heat issues.
    The point is, yes silicon will hit a limit for transistor density for a 2 dimensional area, but that's a single problem, not THE problem. Solving the other problems I listed is more important.
    And lastly we can talk about how much silicon is needed. Well, as process nodes shrink transistor size over time, a single wafer can make a lot more ICs, so we are always improving the number of circuits that some from a fixed amount of silicon. It's not an issue at least not for another 50 years or so.

  • @SarcastSempervirens
    @SarcastSempervirens Рік тому

    great video, keep it up!

  • @lucifermorningstar8051
    @lucifermorningstar8051 2 роки тому

    amazing video. made me subscribe. ill watch more

  • @Kie-7077
    @Kie-7077 2 роки тому +1

    Cool, but the audio is badly skewed relative to the picture in this video.