State of ROCm 5.3 in 2022: 6x Mi210, 1petaflop, in the 2u Supermicro AS -2114GT-DNR

Поділитися
Вставка
  • Опубліковано 8 вер 2024
  • Wendell dives back into his new 2u Supermicro AS-2114GT-DNR server to talk more specifically about the 6 AMD Instinct MI210s held within! So many flops!
    Thanks Engenius for Sponsoring this video!
    Check out The ECW336 Here: www.engeniuste...
    **********************************
    Check us out online at the following places!
    linktr.ee/leve...
    IMPORTANT Any email lacking “level1techs.com” should be ignored and immediately reported to Queries@level1techs.com.
    -------------------------------------------------------------------------------------------------------------
    Intro and Outro Music: "Earth Bound" by Slynk
    Other Music: "Lively" by Zeeky Beats
    Edited by Autumn

КОМЕНТАРІ • 186

  • @kamrankazemi-far6420
    @kamrankazemi-far6420 Рік тому +58

    Being able to write code once and being able to run on either platform is so huge.

    • @ramanmono
      @ramanmono Рік тому +5

      Yes, Java promised to do this a gabillion years ago. Sadly I don't see any new tool getting any closer.

  • @jd_flick
    @jd_flick Рік тому +196

    I am really hoping AMD can make not using CUDA a reality

    • @harryshuman9637
      @harryshuman9637 Рік тому +20

      It's all up to the devs really.

    • @zacker150
      @zacker150 Рік тому +12

      I've given up on AMD gpus ever competing in compute.
      Hopefully Intel's OneAPI works out.

    • @RaaynML
      @RaaynML Рік тому +49

      @@zacker150 It's so weird to comment this on the same video in which you heard them currently competing in several of the top super computers

    • @LeDabe
      @LeDabe Рік тому +8

      @@RaaynML The AMD environment lacks tooling. Though a new tool, MIPerf, is coming and should play a similar role to the nsight compute nvidia provides

    • @youkofoxy
      @youkofoxy Рік тому +22

      They are trying hard, very hard, however the curse of the Ctrl-c Ctrl-v run too strong in the programming community.

  • @markpoint1351
    @markpoint1351 Рік тому +29

    my god Wendel you really made my day with that Shinning meme 🤣!!! thank you

  • @AI-xi4jk
    @AI-xi4jk Рік тому +33

    It would be cool to see just some torch benchmarks of some regular ML models vs 3090 and other Nvidia cards.

  • @stuartlunsford7556
    @stuartlunsford7556 Рік тому +59

    I really hope everyone starts pronouncing it Rock'em, like Rock'em Sock'em Robots. It's much more funner that way.

    • @scott2100
      @scott2100 Рік тому +5

      Same here, I thought that was just how it was pronounced

    • @bakedbeings
      @bakedbeings Рік тому +6

      People pronouncing it another way hadn't occured to me.

    • @jadesprite
      @jadesprite Рік тому +7

      Small m implies that it should be pronounced this way!

  • @iyke8913
    @iyke8913 Рік тому +40

    Wendell flips heavy server gear with ease and grace, meanwhile, ....... Linus drops everything.

  • @vtheofilis
    @vtheofilis Рік тому +10

    That Shining meme was pure gold.
    So, ROCm can help port CUDA stuff on OpenMP or whatever the open standard is called, in the data center side. I hope that it is also easier for desktop CUDA code to be ported, so that, for example, ANSYS can support AMD GPUS more easily.

    • @brenj
      @brenj Рік тому

      👍🏻

    • @hammerheadcorvette4
      @hammerheadcorvette4 Рік тому +1

      RocM (formerly HSA) has had tools to port CUDA workloads for years, but th presence and convenience of CUDA has been too strong for people to care. All it takes is an Open Source project and a company willing to change from the norm for whatever reason.

  • @Richardus33
    @Richardus33 Рік тому +19

    love this channel learned allot over the year, thanks Wendell!

  • @Gastell0
    @Gastell0 Рік тому +7

    12:47 - MI25 also supports SR-IOV, but there's no public documentation on how to actually utilize it

    • @wayland7150
      @wayland7150 Рік тому +2

      Tell us more please.

    • @2megaweeman
      @2megaweeman Рік тому

      @@wayland7150 I think @antonkovalenko is referencing the way you can flash the vbios of a wx9100 on a mi25 and use it for GPU task. I think the only way right now to do it after you flash is to use gpu-p(Hyper-v). Look for vega 64 GPU virtualization

    • @wayland7150
      @wayland7150 Рік тому

      @@2megaweeman Yeah, unfortunately the MI25 does not make sense for the homelab at the current price. Really wanting SR-IOV, it would make these cards worth a lot more than VEGA if someone smart could show us how to do that.

  • @stranglehold4713
    @stranglehold4713 Рік тому +5

    I regard yourself and Steve Burke as the two best voices in the computer hardware space. Your channel is treasure trove of information

  • @crookedtuna
    @crookedtuna Рік тому +27

    Been using ROCm on a 6700xt for stable diffusion and I'm shocked how well it performs considering it's not even a CDNA GPU.

    • @andrew_hd
      @andrew_hd Рік тому +1

      It's really cool tech to tinker with. I'm as well using 6700XT in SD. It's so nice to have 12 Gb vram.

    • @zabique
      @zabique Рік тому +2

      Could you recommend any tutorial how to make it work?

    • @chriswright8074
      @chriswright8074 Рік тому

      Most recent AMD consumer GPUs had support for it

    • @Ronoaldo
      @Ronoaldo Рік тому

      Do you happen to have any tutorials on running such models with customer GPUS. I have a 6800XT and would love to work on it. The far I got was using the default Docker container with Tensorflow, not sure if I'm on the right track? Thanks for any input.

  • @ll01dm
    @ll01dm Рік тому +10

    It's good to hear rocm has got easier to install. Back when I was using a Vega 56 I tried installing it. It was a nightmare. I gave up and just used a docker image.

  • @paxdriver
    @paxdriver Рік тому +12

    You should just try out stable-diffusion making 4k images instead of 1024x1024. The processing requirements scale quadratically as does pixel density with larger text to image generation so it's not feasible on normal human system, but the algorithm and walkthroughs are so organized anyone should be able to download the weights, set it up and get it running. You'd be the first with 4k diffusion, and you could even trying training it up to get better at faces and hands using that u2 sized sweet, sweet top rack candy 😍

  • @TheDoubleBee
    @TheDoubleBee Рік тому +14

    I work in the field of photogrammetry, a subset of computer vision, and I'm praying to whatever deity is willing to listen to make CUDA obsolete, but everything is moving so, so slow. Quite a while back I came across SYCL and I was mightily impressed, but it was in super early stages and I haven't checked back recently.
    Nvidia has had a horrible stranglehold on the whole computer vision industry for quite a while, but there might be some cracks showing given their recent open-sourcing of CV-CUDA libraries, which, you don't need me to point out, is an incredibly un-Nvidia move to pull - following their earlier and also un-Nvidia move of sort-of open-sourcing their driver for Linux.

    • @Pheatrix
      @Pheatrix Рік тому

      Nvidia started also updating their support for OpenCL. You are now not stuck forever on version 1.2 if you have a Nvidia GPU but can now use 3.0!
      Maybe you should have a look into OpenCL. It's pretty much CUDA but as an OpenStandard with support from all major vendors (for both GPU+CPU).
      It just needs publicity...

  • @Nobe_Oddy
    @Nobe_Oddy Рік тому +1

    OMG WENDELL!!!! @ 3:00 Is that Betty White as a ZOMBIE on your desk?!?!?!?! THATS AWESOME!!!! lmao!!!

  • @СусаннаСергеевна

    I like these little looks into Wendell's server room. It's basically my dream home setup. I've no clue what I'd do with it all, probably waste time putting pihole on kubernetes or something, but still.
    I'm actually really excited about new, improved ROCm. I've got torch running on a 6900XT so I can sort of do CUDA through ROCm already, but it's still missing an awful lot of features and performance compared to the Nvidia version, 99% of the time I'm better off just using an Nvidia card, even though my best Nvidia stuff is two generations behind RDNA2. I think consumer-accessible and actually fun machine learning things like Stable Diffusion is a great thing for this field, the more people who get into CUDA and ROCm, the more emphasis will be placed on accessible hardware with >8GB of GDDR and decent compute capabilities that are easy enough to use that even I could set it up.
    Unfortunately the reality is that, despite the advances they've made, AMD aren't really a competitor yet. Nvidia still has an enormous headstart, and breaking the "vendor lock-in" that CUDA so effectively creates is only the first step. AMD need to actually deliver competitive performance. They're in a good position to do that, chiplets are the future and Nvidia's monolithic dies are getting truly ridiculous (>600mm²!); AMDs yields are going to be far higher, which means they should be able to afford to stuff more cores into their products. That they aren't is somewhat baffling to me.

  • @tanmaypanadi1414
    @tanmaypanadi1414 Рік тому +1

    16:47 🤣 relentless execution

  • @mvanlierwalq
    @mvanlierwalq Рік тому +4

    Perhaps not the only reason, but the DOE's Energy Exascale Earth System Model (E3SM, the DOE climate model), requires big-time FP64 flops. AMD is, and has been for a while, WAY ahead of NVIDIA when it comes to FP64. Btw, running E3SM might be a good test. As far as I know, DOE has developed containerized versions of E3SM, and you should be able to download and run it (or a small chunk of it) on that machine.

    • @mvanlierwalq
      @mvanlierwalq Рік тому +1

      I'll add that traditionally climate and weather models have been written in Fortran. DOE has sunk a lot of effort into getting code refactored into C++ to be able to use GPUs. NASA instead has just stuck with CPUs in their machines. Big question where the field as a whole goes from here.

  • @ChristianHowell
    @ChristianHowell Рік тому +3

    Very good video... I think I know why everyone is rushing to support AMD... About 3 months or so ago I was watching a tech video about self driving and the gist was that full self driving will require around 2PFbf16 and if AMD hits their target with MI300 it will have around 2.5PF(QOPS?) as MI250X has 383TOPS with MI300 aiming for 8X the AI perf (from AMDs presentation)...
    That's exciting AF...

  • @jonteno
    @jonteno Рік тому +1

    going too be so fun watching you do vids on these! The enterprise side is so interesting atm!

  • @randomhkkid
    @randomhkkid Рік тому +27

    Would love to see stable diffusion performance on this machine. How large an image can you generate with the pooled gpu memory?

  • @LA-MJ
    @LA-MJ Рік тому +5

    N00b question. Can one test RocM on consumer RDNA2?

    • @tanmaypanadi1414
      @tanmaypanadi1414 Рік тому +1

      Asking the real questions. As far as I know No but I am sure someone will figure it out.

  • @Marc_Wolfe
    @Marc_Wolfe Рік тому +5

    Maybe in the future we can see what us poor people can still do with an MI25. I struggled for a little bit to get ROCM installed (apparently Vega support ended after ROCm 5.0 I think it was, specific versions of Linux too apparently), then I gave up and flashed it's vBIOS to a WX9100... after bashing my head off my keyboard to figure out the right buttons to press to get the flash to work... and realizing there were 2 BIOS chips that needed flashed.

    • @ewilliams28
      @ewilliams28 Рік тому +2

      I've seen those for less than $100 on eBay. I would really love to get one or two of those working for a VDI project that I'm working on. I really hate GRID.

    • @Marc_Wolfe
      @Marc_Wolfe Рік тому +1

      @@ewilliams28 Paid $80 plus tax for mine. I'd love a good excuse to use it for more than just gaming, but that was my main goal; so not a big concern, just nerd desires.

  • @NaumRusomarov
    @NaumRusomarov Рік тому +2

    modern fortran is still used even today for scientific computing. if you're a scientist who doesn't have time to deal with the quirks of c-languages, then fortran is really the best choice for you.

  • @Ronoaldo
    @Ronoaldo Рік тому

    16:41 This was amazing!!!😂

  • @matiasbrandolini
    @matiasbrandolini Рік тому +3

    Level 1? more like, level 2000. I didnt understand a word until I heard Fortran.. maybe because Im a COBOL programmer :)

  • @joshhua5
    @joshhua5 Рік тому +1

    I’ll set this up on my desktop tonight, been watching rocm for a while. Maybe I can finally retire the M40

  • @chooka003
    @chooka003 Рік тому

    I'd LOVE this for BOINC!!! "Drool"

  • @spinkey4842
    @spinkey4842 Рік тому +1

    0:48 AAAAHHHHHHHHHH!!!!!!!! him no want things plugged in his body

  • @linuxgeex
    @linuxgeex Рік тому +4

    ROCm is great because you can have the same machine learning setup on your workstation as on the supercomputer. This will succeed for the same reason that x86 succeeded and the same reason that Linux succeeded - accessibility by the masses. I believe the popular term these days is Democratisation.

  • @gsedej_MB
    @gsedej_MB Рік тому +2

    Great video. I would just like more broader (radeon cards) support. I eas playing with rocm since its release on rx480, but totaly lost interes with lack of rdna(1) support and even rx480 lost its official support. And all the details with pcie-atomics and almost none laptop dgpu and apu support.
    But again nice that they at least enterprice support.

  • @DarkReaper10
    @DarkReaper10 Рік тому +7

    Hi Wendell, I think you mistook Fortran for Cobol here. Fortran is used in science applications that get sent to HPC clusters, not really useful for finance.

    • @OGBhyve
      @OGBhyve Рік тому +2

      He definitely means Fortran here. Fortran, C, and C++ are the best supported languages for GPU programming. Those languages also have the OpenMP support he mentioned.

    • @DarkReaper10
      @DarkReaper10 Рік тому

      @@OGBhyve I know but his explanation that Fortran exists because of legacy finance applications is a Cobol backstory. I am a fellow HPC guy, I know Fortran very well.

    • @OGBhyve
      @OGBhyve Рік тому +3

      @@DarkReaper10 It's used in Finance too, but I see your point that it is more popular in scientific applications.

  • @ewilliams28
    @ewilliams28 Рік тому +1

    I would love to be able to use Instinct cards and be able to get rid of GRID as well.

  • @justwhyamerica
    @justwhyamerica Рік тому +1

    Patrick boyle runs a finance channel and might be willing to work with you on actually using openBB

  • @WolfgangWeidner
    @WolfgangWeidner Рік тому

    Important stuff, thank you!

  • @SxGaming3390
    @SxGaming3390 Рік тому +1

    The title is as appealing as the scientific names of most plants.

  • @garytill
    @garytill Рік тому +1

    Let's get that onto a 1ru tray.. nice.

  • @spuchoa
    @spuchoa Рік тому

    Great video!

  • @landwolf00
    @landwolf00 Рік тому +2

    Hi Wendell. Do you intend to benchmark rocm for pytorch? I'm very interested in this and it seems like it doesn't really exist on the web. As others have said, Cuda dependence is scary!

  • @ramanmono
    @ramanmono Рік тому +2

    So what's OneAPI and HIP? Now we need have 5 API's for example to run raytracing on GPU in Blender (nvidia optix and cuda, AMD HIP, Intel oneapi an mac metal). How will a small team or individual working on a piece of software that need GPU acceleration get that to work (decently optimized) with all mainstream platforms?

    • @Pheatrix
      @Pheatrix Рік тому

      They could usw OpenCL.
      An already existing API with Support from all major vendors for CPU and GPU computation (and everything else that implements it. e.g. FPGAs). It also supports all major OS (Windows Linux Mac and even Android just to name a few).
      I just don't get it why we need another standard that does the exact same thing.

    • @ramanmono
      @ramanmono Рік тому +1

      @@Pheatrix Yeah, but it's bad buggy and you could never close to the performance of Cuda. That is why it is abandoned. So seriously no dev is gonna use opencl for high performance gpgpu. Apple too completely removed support for it in favor of their own way better performing metal api.

    • @Pheatrix
      @Pheatrix Рік тому

      @@ramanmono
      Boinc, pretty much every cryptominer and a Lot of other programms use OpenCL.
      The performance gap between cuda and OpenCL ist there because Nvidia decided to only support up to OpenCL 1.2 however there are a lot of features that require at least 2.0
      Recently Nvidia bumped the supported version up to 3.0 so the performance gap should no longer be there.
      And the bugs: well every vendor hast to implement their own driver and compiler. AMD is known for buggy drivers and as I already said Nvidia pretty much abandoned OpenCL in favor for their proprietary solution.
      All of these problems are solvable. And with way less work than creating a completely new solution that solves the exact same solution

  • @Mr_Wh1
    @Mr_Wh1 Рік тому

    4:20 - A little server room ASMR for us all.

  • @denvera1g1
    @denvera1g1 Рік тому +1

    Get this man some Mi 250x

  • @RaspyYeti
    @RaspyYeti Рік тому +2

    Would it be possible for AMD to create it's own titan by having an RDNA die and an CDNA die in a SoC?
    Would they be able to use Async compute to feed the CDNA die and boost Raytracing calculations?

  • @builtofire1
    @builtofire1 Рік тому +3

    i guess Wendel has electricity bills

  • @Jack-qj2pr
    @Jack-qj2pr Рік тому +4

    One bug I found with ROCm is that it just doesn't work at all if you mix a Radeon Pro Duo Polaris with an RX Vega 64. It just doesn't detect anything if you mix cards. Pretty frustrating.

    • @TheKazragore
      @TheKazragore Рік тому +3

      I mean is mixing cards any sort of norm? Not making excuses (it not working sucks), merely pointing out that may not exactly be a priority usecase for fixes.

    • @Jack-qj2pr
      @Jack-qj2pr Рік тому +3

      @@TheKazragore I agree. I'd imagine with it being a relatively niche scenario, nobody would've tested it or even considered it.
      I just compiled ROCM again yesterday and my issue seems to have been fixed now, so happy days :)

  • @mrfilipelaureanoaguiar
    @mrfilipelaureanoaguiar Рік тому +1

    250v 20 Amps, at some point that could cook food or boil big amounts of water, that's super serial seriasly serial

  • @owlmostdead9492
    @owlmostdead9492 Рік тому

    The day CUDA is not the only option will be a good day

  • @sailorbob74133
    @sailorbob74133 Рік тому

    I'd love to see some follow up on this one.

  • @Pheatrix
    @Pheatrix Рік тому +3

    There already is an open standard for this:
    OpenCL !
    It runs on pretty much everything (including CPUs, FPGAS, and GPUs) and with OpenCL 3 you also get a newer version than 1.2 on Nvidia Devices.
    Why do we need a new standard if we can just use the one that already exists and has support from every major vendor?

  • @Veptis
    @Veptis Рік тому

    I supposed in the future we will look at Intel, their accelerator hardware (GPU Max?) And software stack (oneAPI) which includes all kinds of solutions. None of which seem finished tho.

  • @jannegrey593
    @jannegrey593 Рік тому +1

    OK. I hope to see also some more modern Radeon Instincts here. Unless the MI210 is one. IDK if AMD changes their names for those cards honestly, but I did hear about MI250 and MI300 - latter of which probably isn't out yet. I hope someone will educate me on this, because honestly quick google search has a lot of problems with sources that IDK if I should trust.

    • @KL-ky8fy
      @KL-ky8fy Рік тому +1

      it's the same architecture as mi250, they are both CDNA2, lunched in March this year,

    • @samuelschwager
      @samuelschwager Рік тому +1

      MI250 was launched 11/2021, MI210 03/2022, MI3xx is expected for 2023.

  • @BogdanTestsSoftware
    @BogdanTestsSoftware Рік тому

    Could you tell the wire monkey to wear hearing protection, so that they don't get hearing damage? You got me to laughing w/ tears about the #shining and AMD's relentless execution!

  • @hedrickwetshaves1997
    @hedrickwetshaves1997 Рік тому

    @Level1Techs Could you please explain all the different FP64 FP16 FP32 Int8 and is there anyway to compare them with each other?

  • @ChinchillaBONK
    @ChinchillaBONK Рік тому

    Hi, is it possible to do a basics video about ROCm ? Sorry to bother you and thanks.
    Also what are the differences in uses between EPYC , Threadripper CPUs and the many different GPUs like AMD Instinct ones Vs Nvidia A6000?

  • @Marc_Wolfe
    @Marc_Wolfe Рік тому

    17:02 Doom 2016 LOL

  • @kortaffel
    @kortaffel Рік тому

    Why are they only supporting OpenCL on Instinct?
    Why don't we have Vulkan or a new VulkanCompute version available?
    I heard OpenCL is stuck

  • @Yandarval
    @Yandarval Рік тому

    Every time I see Wendel go into the Server room. All I can think of is, where is your hearing protection, Wendell?

  • @danielsmith6834
    @danielsmith6834 Рік тому

    As for why Oak Ridge chose AMD for Frontier -- my guess is that Nvidia has massively optimised their silicon for AI workloads, where AMD has targeted more general GPGPU compute workloads. For a general purpose HPC system, FP64 is critical. Looking at the relative FP64 performance (especially FP64/W) shows how wide the gap is. Why Facebook/Meta are looking to switch? Given I'd imagine most of their workload is AI/ML, that's a much tougher puzzle.

    • @duckrutt
      @duckrutt Рік тому

      I don't see Meta swapping vendors but I can see them bringing up their cool new software every time they need to buy a batch of Tesla cards.

  • @scottxiong5844
    @scottxiong5844 Рік тому +1

    MM laser...it is fine. :D xD

  • @cedrust4111
    @cedrust4111 Рік тому

    @Level1Techs does Nivida or Intel have a direct competitor against the instinct accelerator?

  • @philhacker2405
    @philhacker2405 Рік тому

    Blender would be Fun.

  • @LeDabe
    @LeDabe Рік тому +1

    rocprof is soon to be hidden under an GUI called MIperf that has yet to be released by AMD but is available on Crusher (a TDS of frontier)

    • @LeDabe
      @LeDabe Рік тому

      it will provide information similar to what Nsight compute does. Imo tooling was one of the last big problem with working with AMD cards.

  • @NKG416
    @NKG416 Рік тому

    i don't know shit about HPC, but it seems everyone likes opensource
    it kind of felt i bought stuff from the right company

  • @Nec89
    @Nec89 Рік тому +1

    IM SUPER SERIAL GUYS! CONSOLE CABLES ARE REAL!!1!

  • @dawwdd
    @dawwdd Рік тому +7

    Intel CPUs are working excellent with PyTorch and it should be easy to join new GPUs considering oneapi, amd not so much lets hope it changes in near future and amd software will get better performance and some stability.
    I don't know anyone who use AMD over NVIDIA in Machine/Deep learning right now cause of ROCm extremely poor quality and problems with consumer gpu not working with ROCM at all so you can't develop locally, but there are few folks works with scientific computation focus mostly in HPC that use AMD for Float64 calculations.

    • @RobHickswm
      @RobHickswm Рік тому +3

      I use ROCm over Cuda sometimes. I've benchmarked a fair amount of tensorflow code for my research and it is neck and neck with last gen hardware (Radeon VII's vs A/P100's) . It is very easy to get it running particularly if you use the ROCm docker images for your tool of choice. And the tensorflow/jax code just runs with no modifications.

    • @dawwdd
      @dawwdd Рік тому +2

      @@RobHickswm Cool but Tensorflow isn't PyTorch I tested 3090 with Radeons in close price points and they are always few times slower maybe in the extremely high end datacenter they are close enough but I haven't any AMD card to test it.

    • @RobHickswm
      @RobHickswm Рік тому

      @@dawwdd I've only tested the Radeon VII (which uses HBM2 memory like the datacenter cards) and for things I'm doing (not canned ML benchmarks) it is as fast/faster than the Nvidias with a few exceptions here and there depending on the op. You're right. Not pytorch, just jax and tensorflow.

  • @garrettkajmowicz
    @garrettkajmowicz Рік тому +1

    Why hasn't AMD upstreamed their TensorFlow support?

    • @intoeleven
      @intoeleven Рік тому

      ROCm has supported tensorflow repo at their GitHub

    • @garrettkajmowicz
      @garrettkajmowicz Рік тому

      @@intoeleven Yes. They have a fork of TensorFlow. Which is why I've asked why they haven't upstreamed it. If it isn't mainline, it doesn't really matter that much.

    • @intoeleven
      @intoeleven Рік тому

      @@garrettkajmowicz They are upstreaming and syncing it constantly. Their own fork is for customers.

  • @dgo4490
    @dgo4490 Рік тому +2

    Come on, trading? Is that the best usage for this hardware?

  • @NavinF
    @NavinF Рік тому +1

    No mention of consumer AMD GPUs? It kinda feels like AMD doesn't care about ml. Researchers use CUDA because it's officially supported on their desktops.

    • @Cooe.
      @Cooe. 10 місяців тому

      They aren't going after individual researchers... 🤦 They want super computers, data centers, and multinational companies where it's MUCH easier, more efficient, and more profitable to gain market-share. And it's working. And RDNA cards did eventually get ROCm support though.

    • @NavinF
      @NavinF 10 місяців тому

      @@Cooe. Meh. Many off the shelf models require CUDA for at least one layer. Still makes no sense to use AMD for machine learning

    • @Cooe.
      @Cooe. 10 місяців тому

      @@NavinF Massive data centers aren't using off the shelf models, ya dingleberry... 🤦

    • @Cooe.
      @Cooe. 10 місяців тому

      @@NavinF Also, ROCm lets you run CUDA code anyways even if you're lazy (even though you won't get quuuuuuite the performance you would running it natively w/ the same FLOPS on Nvidia).

  • @camofelix
    @camofelix Рік тому

    All they have to do is make hip *checks notes* not shit
    It’s still a PITA to work with

  • @engineeranonymous
    @engineeranonymous Рік тому +1

    In my humble opinion AMD should focus on unified memory architecture like Apple M series CPU's. You can not offload a lot of computations to GPU because the memory transfer requirement simply kills your gains. An unified architecture will make every operation as a target for acceleration and Nvidia has no answer for this since they only make GPU's. AMD CPU's with built in GPU's can break benchmarks for both Intel and Nvidia.
    Correction :
    I'm such a fool. HBM unified memory will come to AMD in 2023 for datacenters with MI300 in 2023. They announced it in Financial Analyst Day 2022. I can't believe I missed it.

    • @tanmaypanadi1414
      @tanmaypanadi1414 Рік тому

      xylinx might be the able to help with accelerators but it's a few years off before we see any applications in the consumer realm.

    • @jesh879
      @jesh879 Рік тому

      You realize AMD was the one who created the HSA foundation right? HSA was demonstrated before Zen 1 existed. When AMD moves on this, no one will be doing it better.

    • @engineeranonymous
      @engineeranonymous Рік тому

      @@jesh879 Yeah I know but HSA only includes cache coherency (that's what I understand from v1.2 of standard) but Apples implementation goes beyond what AMD or Intel called UMA. In M1 CPU and GPU share the same ram and can use it utilize when needed.

  • @synt4x.93
    @synt4x.93 Рік тому +1

    Did the title change? Or am i high.

    • @Level1Techs
      @Level1Techs  Рік тому +3

      Title changed. Views are low and we're hoping the title change will fix it ~Editor Autumn

    • @synt4x.93
      @synt4x.93 Рік тому +2

      @@Level1Techs Great video, as always.

    • @Level1Techs
      @Level1Techs  Рік тому +2

      Thanks!

    • @tanmaypanadi1414
      @tanmaypanadi1414 Рік тому

      Let the clicks and engagement rise up.

    • @tanmaypanadi1414
      @tanmaypanadi1414 Рік тому

      @@Level1Techs Is there any way to get notifications as soon as the video drops?
      discord notifications work for me for some channels , is there something similar on the forums for us free tier folks other than UA-cam .

  • @maximusoptimus2000
    @maximusoptimus2000 Рік тому

    Just compare it with supercomputers from about 20 years ago

  • @thesunexpress
    @thesunexpress Рік тому

    Do a dnetc run on it?

  • @dracleirbag5838
    @dracleirbag5838 Рік тому

    What does it cost

  • @linuxgeex
    @linuxgeex Рік тому +1

    Cloud managed IOT can go straight to hell. They should ship an app that runs on your phone and provides an API that the IOT gear detects, and let you pair with bluetooth or with a button and extreme close range (easy to detect with the WiFi or BT hardware.) After that you should be able to manage it from the same app running on your PC, and you should be able to install a PKI signature onto the IOT device which forever locks it to a cert under your control, so it can't be hijacked, not even by your child/spouse/roommate/landlord etc.

    • @snowwsquire
      @snowwsquire Рік тому

      iot is dumb, internet protocol is overkill for a lightbulb, matter over thread is the future. zwave/zigbee for right now

  • @bryantallen703
    @bryantallen703 Рік тому

    but, can 1 MI250 run 64 instances of CRYSIS 64-bit

  • @Misiek-oc7bu
    @Misiek-oc7bu Рік тому

    but can it run crysis

  • @zyxwvutsrqponmlkh
    @zyxwvutsrqponmlkh Рік тому

    But can it run Cuda?

  • @Nobe_Oddy
    @Nobe_Oddy Рік тому +1

    Wendell is gonna suddenly disappear and we won't hear from him for 6 months and it'll turn out that while making his video about using the Supermicro on the stock-martket with-in 5 minutes of turning it on he managed to become 3 richest man on the planet and spent the last 6 months on HIS private private island LOL :D

  • @luridlogic
    @luridlogic Рік тому

    Can rocm be setup in Debian rather than Ubuntu?

    • @squirrel6687
      @squirrel6687 Рік тому +2

      Anything can. I run Debian Bookworm with everything from PerceptiLabs, Anaconda with all the juices even with SecureBoot and Nvidia and their CUDA Toolkit. Once up and running, no upgrade hell as with Ubuntu.

  • @WiihawkPL
    @WiihawkPL Рік тому +1

    now they should make an ai accelerator that doesn't cost a kidney

    • @Jake9066
      @Jake9066 Рік тому +3

      Sorry, "AI accelerator" contains two $-add words, so $$$ instead of $

  • @johnferrell1962
    @johnferrell1962 Рік тому

    Should I get this or the 4090?

  • @Rintse
    @Rintse Рік тому

    This title will get clicked by no one who is not a serious enthusiast/nerd.

  • @starfleetactual1909
    @starfleetactual1909 Рік тому

    Greek

  • @evrythingis1
    @evrythingis1 Рік тому +3

    Maybe Intel and Nvidia will learn that they shouldn't rely on being a monopoly for their success.

    • @HellsPerfectSpawn
      @HellsPerfectSpawn Рік тому

      What are you blabbering about Intel provides more open source code to Linux then all the other PC players combined

    • @evrythingis1
      @evrythingis1 Рік тому +1

      ​@@HellsPerfectSpawn Yeah, totally of their own accord, not because their monopoly was so severe that that literally had to after years of ILLEGALLY doing MSFT's bidding.

    • @HellsPerfectSpawn
      @HellsPerfectSpawn Рік тому

      @@evrythingis1 ??? What mental gymnastics are you jumping through mate?

    • @evrythingis1
      @evrythingis1 Рік тому

      @@HellsPerfectSpawn Do you not know anything at all about Intel's history of Antitrust violations!?

    • @HellsPerfectSpawn
      @HellsPerfectSpawn Рік тому

      @@evrythingis1 Again what kind of mental hoops are you jumping through. Are you trying to suggest that because Intel got sued in Europe it suddenly found a reason to go open source??

  • @stephenreaves3205
    @stephenreaves3205 Рік тому +2

    first?

  • @codejockey216
    @codejockey216 Рік тому

    Second, haha

  • @rtkevans
    @rtkevans Рік тому

    Dude wth is that framed picture on your desk??? Looks satanic…

  • @marcusaurelius6607
    @marcusaurelius6607 Рік тому

    and now it’s may 2023 and nobody cares about ML on amd cards. unless it’s a drop-in replacement, nobody will migrate their massive ML tech stacks to eh, what do you call it.. radeon?

  • @FLOODOFSINS
    @FLOODOFSINS Рік тому +1

    It's a shame this guy doesn't have any kids. He has so much knowledge crammed inside his head.

    • @tanmaypanadi1414
      @tanmaypanadi1414 Рік тому +2

      The UA-cam channel is his baby

    • @nathanlowery1141
      @nathanlowery1141 Рік тому

      We are his spawn

    • @Onihikage
      @Onihikage Рік тому

      He doesn't need children to leave a legacy. _We_ are his legacy.

    • @Blacklands
      @Blacklands Рік тому

      Well, he has a forum and a UA-cam channel...! He's teaching many more people than just the kids he doesn't have!

    • @FLOODOFSINS
      @FLOODOFSINS Рік тому

      @@Blacklands a forum is way better than having your own child and seeing your legacy live on along with everything that you can pass on to him besides tech stuff. You're so wise, maybe he can put that on his tombstone "I have a forum"