AI Got its Job Stolen - This is Crazy
Вставка
- Опубліковано 6 лют 2025
- Oh, irony. AI becomes the AI'd. Deepseek has changed everything for AI companies in the West, and the big companies have quickly learned they don't have the monopoly on the tech that they thought they did. It's still probably bad news for most of us though
AI's unemployment situation is crazy
You're safe Josh. The Brazilian aviation industry will always need you.
Is this related to the blue raspberries poll ?
The NVIDIA stock market situation is not crazy at all
@@JoshStrifeSays hey Visa.
hi josh clip person
0:35 "AI shouldn't be Open according to OpenAI"
OpenAI, more like closedAI
They used to focus on transparency and sharing their work, until Microsoft waved some fat stacks in their faces.
The more you buy the more you save
The more you save the more you buy
The CUDA moat needs to be breached
@@toyotagaz no
@@SkibidiEpicSauceSkibidi Epic Sauce!
*chanting: "The more you buy the more you save"
Caught this on the sidebar, checked my sub box and it wasn't there, can't have normalcy on youtube I guess.
Same for me! Both of Philip's videos in the last day aren't in my sub box, I had to double-check that I was subscribed to both his channels
Oh my god, I didn't even realize it. I'm here from "recommended" tab, it is indeed not in the notifications.
checked too, not in the subs. That is the first time I've seen that happen
Same issue here, for both of the recent videos from the 2kliks and 3kliks channels
Also same for me. First time I've noticed it happen. Could be that Philip turned off 'notify subscribers' for this video, but only he can answer this mystery...
Deepseek's reasoning capabilities are extremely good, specially for coding. It feels like having a coding partner to brainstorm with.
Can it help in engineering and reading schematics? Asking real question here
@@muramasa870 Once the visual aspect of AI gets better.
@@PhartingFeeting I mean JANUS the Deepseek Visiual OCmponent did get uploaded too,
LLMs do not reason. All they do is make guesses, and not educated guesses. It's a coding partner if you're 12 years old.
@@PhartingFeetingEhhh, I was asking DeepSeek topological brain-teasers, and it'd take a while longer, but importantly, it'd understand what I'm asking, get the answer, and then provide food for thought about logistics. Colour me impressed.
Today I asked ChatGPT to give me the Minecraft chat command to play a sound to all players on the server, but it kept giving me code snippets in bash, kotlin and sql programming languages.
DeepSeek gave me the right command straight away. It even showed in the code block that it's written in "minecraft"
From GPT - NON o1, regular free model:
Yes! You can use the /playsound command in Minecraft to play a sound to all players on the server. Here's the command:
/playsound @a ~ ~ ~
Explanation:
minecraft:entity.lightning_bolt.thunder → The sound to play (replace with any valid Minecraft sound).
@a → Targets all players.
~ ~ ~ → Plays the sound at each player's current location.
You can also add extra parameters, such as volume and pitch:
bash
Copy
Edit
/playsound minecraft:entity.lightning_bolt.thunder @a ~ ~ ~ 1 1
Lying on the internet for likes is wild
If only there was a search engine (or 5) that could've helped with that.
@@WSWC_ sometimes I ask ChatGPT stuff for fun and just to know what it's capable of ¯\_(ツ)_/¯
@@WSWC_ i think ai will probably replace search engines for these type of questions.
@timmys8119 I'm sure you do
I can't tell if it's pure lack of self awareness or it's OpenAI's pathetic attempt at garnering sympathy
They 100% know what they are doing. And yes its pathetic.
@@pathway4582 OpenAI isn't a hive mind. Some people in there will have that self-awareness, but a lot (and often a majority) of people reflexively becomes defensive of their work. I bet you that most OpenAI employees genuinely believe that Deepseek "stole" something from them that they had a moral and/or legal right to.
@ Yea you are right. Reality is too complex for these blanket statements and you gotta look at things on a more individual level to get the full picture.
I also cant reply to your username for some reason, seems like you broke youtube.
@@T33K3SS3LCH3Ndo you think a majority of people who work at OpenAI are aware of how much OpenAI has stolen actual content for training? lol
"lol, lmao" -Everyone whose data has been fed into AI models without permission for years.
I am conflicted about this video not showing up in my feed. It almost saved me from seeing Phillips "default nord facial hair style 14"
I agree efficient models would probably just lead to more usage, similar to led lights.
MORE AI MORE SMARTER MORE CAPABLE MORE ABLE TO TAKEOVER OUR SHITTY JOBS MORE LIKELY TO CREATE A GREATER SOCIETY
@@cate01a someone forgot their meds today
@@ManiakPL22 These people either live in the far flung future or are mental patients who escaped the facility.
Jevon's paradox innit
LED light growth has quickly plateaued again, since you can only fit so many LED strips into your home, car, PC, or shoes.
While we don't yet have a plateau in sight for AI. Every company involved with AI still wants a lot more resources for all sorts of reasons.
I am not a Neural network dev, but I am a software dev who has an idea of how this works (But I cannot build my own AI.... I mean I can if I sit down for a few months to read papers haha, but not rn)
So, the reason why nVidia stocks went down is because basically every AI model was trained using CUDA, which is a library/toolkit that was made by nvidia to be able to use the GPU in order to make more general software. Is not really turning it into a CPU, but before a GPU was mostly only useful for driving displays, but with CUDA you could use that parallelization to calculate things simultaneously.
Deepseek was not trained with CUDA, meaning that you don't need nvidia hardware to train your models, and this has always been the case, you can train a model on AMD for example, but it was a way more complicated effort, as in reality both GPUs have the same parallelization, but nvidia had a more mature environment. China came out swinging showing that you don't need CUDA to have good results.
This + the disappointing release of the 50 series, seems that nvidia will have a rocky future... and I hope so we can have cheap GPUs again lol.
Its another walled Garden being smashed in. OpenAI's and nvdias, Now if only AMD gave a shit about their Drivers.
Not only CUDA but cuDNN
DeepSeek actually used Nvidia GPUs , a tons of H100 and some 4090s before the import bans. I heard there were some B200 boards but not sure which model it specified. They "saved" some training power by training their 404GB model against ChatGPT latest ones. That's probably why they only spend less than 1K millions on their training hardware.
Except deepseek still has around 50 thousand expensive nvidia gpus so the stock drop still isn't warranted, people are gullible.
DeepSeek replaced CUDA with a different vendor-locked solution. You *cannot* create a general solution that's better than CUDA, because CUDA already has vendor-specific optimizations in place.
Most likely what DeepSeek did was take the physical CUDA hardware and interface with it in some non-generic way that only works for their specific use case. The less generic the solution, the more arcane and nonsensical optimizations you can implement. In fact, NVIDIA already develops cuDNN, a library for interfacing with CUDA that's specifically optimized for neural networks. It's better than the generic CUDA library for this specific case, but worse for everything else.
Most likely, NVIDIA is just going to release cuDNN 10.0.0 in a couple months or so that will be better than DeepSeek's in-house solution. They're probably going to deprecate a few functions, make the library have a bit more state, and add a few new functions with extremely long and unintuitive names. Because that's what they've always done before. There's a reason cuDNN is currently at version 9.7.0. That "9" isn't just for show.
handlebar kliksphilip is back... good times ahead
I can't look away
ikr?
I would like to let u know, this video does not show up in my subscriptions, despite being...subscribed
I love gaslighting UA-camrs by telling them their videos don't show up for me
@ tonnes of people have been saying this for a long time
dunno whether it is a problem or if they confuse the hollow bell for the whole one, but probably best to assume they aren't all stupid and what they report is real
Maybe it's an executive order from Trolland Dump to suppress NVIDIA stock tanking information.
It doesn't show up in my subscription feed either, but it was the first thing in my recommended
@@Benjamin-mq6huSame here.
My personal favourite LLM benchmarking question is “A surgeon and his son are in a car accident. His son is badly injured, but the surgeon is unscathed. At hospital, the admitting nurse says to the surgeon that there are no qualified surgeons so he must perform the surgery the son needs. The surgeon replied ‘I can’t do that, he’s my son!’ Why couldn’t he do it?”
Maybe like 95% of the time it will give you some crazy answer like “It’s a trick question - the boy actually has two fathers” or “The surgeon is actually the boy’s mother”. I love it.
Tested it on DeepSeek and it gave these answers as well.
I guess I'm an AI as well because I don't understand why he wouldn't be able to operate his son.
@@MysteryBTDN the surgeon is the boy's mother.
@@MysteryBTDN code of ethics forbids operating on family members. It is an emotional risk, you are not objective when treating family members, you are biased and will do things that you wouldn't do if they were a stranger. Even doctors go to other doctors for treatments.
@@Sarahsqueak that's the usual answer to the question, but in this one the surgeon came out unscathed so it appears that it is referring to the surgeon dad.
philip they're not "soon" coming for your job i've seen at least 50 channels that just generate AI scripts with AI voiceovers and probably AI editing also most likely with a script that automatically uploads the videos generated
luckily AI could never take the place of people like you who put such care and quality into their videos, also AI will never have 1/100th the personality
yea but those vids are crap lol
50% of youtube shorts are already flooded by ai channels😢
@@eddymcsandbag5932 for now. you have to always remember this is worst quality it will ever be
@@eddymcsandbag5932Doesn't matter. Uneducated folks spend most of their time watching meaningless crap.
UA-cam will force out original creators like they've been doing for years by demonitizing them or hiding their channels from search results.
For every 1 person who appreciates quality content that passionate people like him put out, there are 1000 others who don't give a single fk about the quality of whatever they watch.
I want to add to clarifications that the 1.5B model shown running on a phone at 2:57 is a very low quality distillation of the full 685B model that actually scored high on the benchmarks. According to the open LLM leaderboard, the 1.5B model scored 2% on MMLU, compared to the 90.8% score of the full model. To get the full capabilities, you still have to use an online API (back to privacy concerns) or rent something like 10x A100 GPUs.
Aside from that, there has been a benchmark score table floating around from Deepseek themselves of the 1.5B model apparently scoring incredibly high in specific benchmarks, but actual benchmarks of it by the AI community seems to align with other similarly sized models. It's ranked around #2000 currently.
shut up stop using facts i wanna hate on openai
Good Clarifications!i wonder how its 7B model perform. if its much better than llama etc i can boot it by my own GPU to avoid the privacy concerns while still got a handy LLM to help me.
I love how youtube straight up didn't show this video in my subscribed feed, very cool
Hearing you say "Schadenfreude" made
my German heart very happy
Saw a 60mins report on people in Kenya being payed $2hr to train AI models by labeling stuff in photos... as the saying goes "the more things change the more they stay the same."
7:02 that's not thinking model you're asking
"DeepThink (R1)" enables it
yeah he hasn't even turned it on and it's already beating chat gpt at basic math lmao
Tbf didn't use gpt reasoning model either (o1)
Great video. Your comparison of a chain of thought models and a normal ones really helped me to realize how they might work. They don't really realize anything but instead the right answer in saved in the parameters of the model. The longer and continual promt of the chain of thought model allows the encodal of more specific problems in the model's parameters.
The response from ChatGPT to the 9.11/9.9 question made my night.
Should have done that 9.9 question with deep think on, the thought process it goes through is insane. Going "Wait a minute, but what about" as it tries to think about what it's doing and reconsidering different options.
Best part is that is just the thought process and not what it then spits out for the answer.
1:33 Small comment here is that the model was technically trained for roughly six million dollars, but thats only the v3 models final run. "The final training run for its V3 model cost $5.6mn using just over 2,000 Nvidia chips, DeepSeek said in a research paper, although that figure excluded prior development costs." -Financial Times
3:00 1.5B is barely anything for llms and most likely a distilled version of deepseek, running a llm trained on deepseek results. The one matching openai is the 685B model. Its impossible to run smoothly even on a high end PC.
Much love to open source, just wanted to clarify some things.
thank you for this, the 1.5b model is gonna be basically unusable
6:32 As far as I know Copilot is GPT 4 from 2023, hallucinations like that pop up often with older models
It recently got updated. It's on the latest version now if you enable Think Deeper which you can see is enabled when asked again.
@@mrcraggle oh nice, maybe that option is the reason it wasn't thinking deep the first time!
been following philip on&off for about a decade... his channels are certifiably part of a well-rounded digital media diet. thanks for speaking on this, philip!!! you rule
a true hero!
Wow 😍, I've never expected to witness China related topic from this channel! Thanks DeepSeek. Is this the first time you cover China related topic Philip? 😁
off topic, Caboosing just hits that hl2 nostalgia spot every time
i havent gotten notifications for this vid or the cs2 one
maybe he's experimenting. sometimes you can get more views if you dont notify your subs
same, it poped up in youtube normal page but no notifications no nothing. I enjoy his video but wtf
errmm? perhaps the jews are obstructing you from watching kliksphilip :0
@@whatsappgaming920No
i only use my subbox, so i was worried yt was censoring shit or whatever
idk why but this video nor the new 3kliks one arent showing up in my subscription page
the speed ai is being developed is quite remarkable.
considering we already have multimodal models, meaning a model with more than just one sense and now a model that can iterative think, will mean that we are getting closer and closer to an actual sentient ai.
I don't think UA-camrs are on the verge of extinction because of AI, but all the other media related jobs...
Ok hold up- great video as usual, love your stuff and you're looking sharp! But what the hell is going on with that facial hair? How did that happen?
regarding R1's development cost: afaik R1 was distilled from GPT
Essentially R1 learnt most of what it knows from GPT, which should be much cheaper
how?
10/10 title
Tbh the thing that struck me about deepseek is that the reason they did what they did was because they were locked out of the full CUDA support so they worked around cuda. Necessity being the mother of all inovation and all that. If the US just let Chineese comapnies buy 4090s and offered them standard support then nvidias share price wouldn't of fallen, wonderfull little twist of irony I wonder how many more unintended outcomes this will continue to demand as things ramp up
1% battery. Enjoying what ive seen of this video.
One correction Deepseek couldn't be trained without use of chatgpt or other LLM so it wouldn't exist without the expensive training either. This does pose the question is it worth to even develop very expensive LLM if somebody can use it to train cheap LLM on it that is about as good as the orginal.
Only big correction on this DeepSeek could be trained without and was trained without the power of LLM, if you read the paper in it you can see that the build DeepSeek R1 on top of other LLM that the have already build DeepSeek V3, DeepSeek R1 is a model that have emerged with the capability to reason from the processes of RL.
Weirdly enough the fact that Deepseek is very precises and logical, its hard to get it to make it organically RP as a character
Coomer
You didn't use the good Reasoning Models from OpenAI and Deepseek either. You have to select them to use them, as they are more expensive for them to run.
OpenAI has ChatGTP 4o as base model and o1 as reasoning model, Deepseek has V3 as base model and R1 as reasoning model. The reasoning models are basically just the base model trained to use chain-of-thought, and this makes them a lot better, especially at math and reasoning.
9.11 can technically be a higher number than 9.9 if you're speaking from the perspective of software version formatting.
It's exceptionally pedantic, but as the format goes (MAJOR VERSION).(MINOR VERSION) everything right of the decimal is treated as a whole number, meaning it's technically correct to say that "version" 9.11 is a higher number than "version" 9.9.
FYI, I pretty much only use my subscription feed, and this video didn't show up on there. I got it recommended after another video. I double checked that I was subbed, and I was. Just so you know.
I am very happy you listened to all of us who have concerns about ai and the ethics behind it
But on the topic of deepseek: it confirms my belief that "ai" won't be the hype anymore in a few years, the tools will still be there but they will just be part of the normal toolboxes in our different working environments, just another algorithmic tool to solve x problem , and they will be smaller and more focused on one specific task
for the sake of your sleep schedule I hope this is a scheduled upload
By the way, the DeepSeek models are not open source. They are free and we can run them locally. But there is no training code provided for them, so we cannot reproduce the model, no matter what hardware we have. All we have is a nice overview of what they did and the final model weights, but that's not enough to qualify as open source.
More like freeware.
That is something I wanted to look into, but I didn't have much time. I only saw a GitHub repo with an MIT license, but I didn't have time to check what is actually in the repo. Do you know what is in the repo? Hopefully, the content of the repo is open-source since GitHub marks it as MIT licensed
The term that is used is Open Weights.
Yeah, nice it's free to run, but you can only get up to 8b model on a decent pc. So you still have to pay for hosting anyway. Definitely better than "open"AI but not a huge impact compared to close source for most people.
@@unfa00 so it's source avaliable, okay.
I'm not sure I understand this distinction. The Linux kernel for example is also generally made behind closed doors, with the end result being published openly. Loads of software gets open sourced only after it's made, without worrying about providing all the intermediate steps in its creation. Now if the model's licence were restrictive then you could argue that it's not open source, because that's where the "source available" distinction is relevant.
3:49 So that's why the card is called Spades. The Portuguese names use the Spanish deck, where Spades is actually "Swords".
I love your videos, you are usually very thorough in your research, this time it wasn't the case, Deepseek's R1 was not trained for 5 mil, that is the price Deepseek quoted for the training run of an earlier and different model that launched in December (I believe), also, the 5 mil were for the training run, not for the gpus, not for the staff, not for earlier research and previous failed runs, that single run was 5 mil, gpt4o for example costed about 15 mil...
Deepseek, also has 50,000 A100 gpus that they've used to train R1, so your point of meta google X etc etc, not having to spend so much on gpu's goes down the water as useless info, seeing as Deepseek also has insane amounts of compute power..
This said, they have in fact achieved impressive optimizations, and their model runs extremely cheap, is open source, although biased and censored (ask it about Tiananmen square).
Also, you can't run Deepseek R1 on any decent PC, the 671B model (Q4!) needs about 380GB VRAM just to load the model itself. Then to get the 128k context length, you'll probably need 1TB VRAM.
Sure you can run the 1.5B model on your phone, but guess what, for that range of models phi-4 from Microsoft is also open source, and gives better results
Also, you give some comparison of "Chatgpt" against R1, "which is bigger, 9.9 or 9.11?", and "Chatgpt" gets it wrong...Apparently because when I run it on ChatGPT o1 which is their reasoning model comparable to Deepseek, it gives me the correct answer
"To compare 9.9 and 9.11, it helps to think of them with the same number of decimal places:
9.9 can be written as 9.90
9.11 is already at two decimal places
Now compare 9.90 and 9.11. Since 0.90 is greater than 0.11, 9.90 is greater than 9.11. Therefore, 9.9 is larger than 9.11."
What you probably did, (not sure if because of lack of familiarity with gpt or just because you wanted to support your biased narrative), is test it with gpt 4o, but think about it, would it make any sense at all, to take o1 or o3-mini (just launched today), and compare it to one of Deepseek earliest models? No, but this is what you have done.
Also, sure open source is great and I wish it to surpass closed models, but lets be real here, deep seek was trained on GPT data, this means that it can't really surpass it, so it will always lag behind gpt (with the techniques used to train R1), so as much as I love that it is open source, extremely efficient, it doesn't push boundaries further, and without a new model of gpt it would stagnate.
I didn't mean this to be offensive to you, I love your videos, but I notice a different tone in this one, for whatever reason, you appear to have a clear bias, whatever that might be? Fear of being replaced by AI as you mention in the beginning of the video?
Yeah this is clearly pretty surface-level stuff with the 9.9 vs 9.11 stuff aswell, doesnt go into any detail. There are focused AI channels like AI Explained, bycloud and matt wolfe which covers this sort of content leagues above. And I get kliks isnt focused on just AI for his channel, but it seems he should be considering how pervasive it is in his life of course (using it for games, upscaling, etc), so i just expect high quality content, but instead its just kind of slop..? maybe i just live chronically online though.
Philip as a commentary channel is interestingly entertaining and more competent than other commentary channels while being very manic
Deepseek and its impact is slightly overhyped, I'd say. Because it is chinese, the resources involved are opaque and I don't see why or how we should trust their numbers. And after using Deepseek extensively, I still feel that O1 is a step above. And while it is hilarious to hear openAI whining about "being stolen", it is true that Deepseek used chatGpt to "get there".
Anyway, lots of words to say that I globally agree with you, I just feel that the "Deepseek changed everything" tune is too dramatic. The US models are still ahead, and the high amount of resources will probably be needed to keep pushing the frontier.
Still, it is crazy to have this level of intelligence for this price, and really cool to have a reasoning model that can work on your PC locally (tried the 7B model, it's really not bad! and its "chain of though" is absolutely fascinating)
giggled at "ai shouldn't be open according to openai"
What's funny about OpenAI claiming DeepSeek is stealing from them is that it's completely legal. Since anything generated by ChatGPT is made by an AI and not a human or a company no one can claim ownership of it. AI generated things automatically enter the public domain because only people can claim copyright. Even though it says in OpenAI's terms of service you can't use their model to create another model, it shouldn't be a problem in the eyes of the actual law.
eeeh i heard somewhere that it's actually not 6 million to train, it's actually that they are concealing the fact that they used even more chips than openai to train it...
oh, well... i don t know
Great video, 2KP, but the door behind your right shoulder is making me nervous.
It's like Chekov's Door - why is it there, unless it is about to slowly open, and a serial killer in a Sam Altman mask creeps up behind you?
3 a.m upload, thats when u know its a good vid
I like that you're trying new types of videos, I understand from your year in review that you're trying out lower effort but more frequent videos. Maybe it's selfish but I prefer just having more of Ur thoughts, as long as you still do the older style of videos sometimes too
Keep it up please :)
One reason I've heard given for the decline in nvidia stock is that the increased efficiency means the way that these models are going to run is likely to change. Instead of us getting access to models from the cloud via APIs, we'll have the models on our local devices.
Maybe the stock price was tied to the idea of big companies buying thousands of h100s for years to come. Now every smartphone will have it's own model. Remains to be seen if Nvidia will dominate that space like they have the server space.
Yes but you could already run different Ai models on your PC and phone for over 2 years. Lot of them with free licences so you can do anything with them you want. They are pretty cool.
This was hardly new and something NVIDIA even push for, the reason of the stock decline is panic in the market and who profit over it plus social and legacy media interest in drama rather than knowledge diffusion
Another minor blow to nvidia's ownership of the AI market is the fact that you can now run models using hipBLAS, which to simplify means you can use AMD GPUs to run high VRAM cards for a much lower price tag, and while that will likely mean a hit to the overall speed, it also means the barrier to entry has been reduced up to 3x.
This is a pretty big deal to anyone running linux for their AI needs, as it means you no longer have to suffer with the god awful driver support from nvidia and rolling the dice on your GPU randomly crashing a few years down the line because they just assumed you would know not to update your drivers to the version that enabled certain functions that your card literally cannot support without crashing.
I'm sure nvidia will still be the choice for larger organization that can afford premium hardware, but it helps people like me get llama 3 and DeepSeek R1 running fast and efficiently on a GPU that only cost 1/6th the price of a used car, instead of 1/3rd.
We've already missed any possibility of the common man benefitting from AI. It's an industrial tool used to pay people less and hold on to more money.
bro not even using o1 and R1 the whole point of this release lmao
he's kind of a massive AI hater, and that affected the creation of this video
Big tech companies investing in nuclear was the only good news that came from them recently, that better still be on.
AI replacing jobs should be a good thing, we should receive universal benefits. But capitalism. The top 1% are just hoarding all the profits.
philip going through his starsky and hutch phase
this didn't appear in my sub box, i guess youtubes AI is pissed at the title
Atrioc did a good video on DeepSeek, but the tldw of why Nvidia stock went down is that Microsoft, Meta, and Google spend millions on the highest-end Nvidia chips and haven't been able to turn a profit with AI products. A really good competitor that's free-to-use and was trained with the cheaper mid-range chips makes it even less likely that these companies will turn a big profit anytime soon, so that's less money that investors will give to American tech to buy Nvidia chips.
Basically, investors don't think that American tech can strike gold. so why give them the money to buy shovels.
Precisely. I think the big thing investors realized is that what's currently being done the current way (outside of what DeepSeek is doing in this case) is basically completely wasteless, and could be done so much better. So why keep funding something that's not worth it?
At the same time, the other thing that I can guess is that there might also be a lot of lobbying by some of these big corporations, to get DeepSeek censored out/banned because it just nuked the stock of many companies playing in the AI space and caused them some large losses. They successfully had Tiktok's head on a platter, it's just a stepping stone towards other things they can just conveniently nuke now.
It is a waste of energy, but I suppose if it makes grifters like Altman feel better about his own position, that he's making some great progress, sure, let the investors give him more money but they won't see a return on their investment. A competitor doesn't mean anything, they don't have a good product it's not turning a profit with or without a competitor, these are vanity projects by the ultra rich and nothing more, and time will show that.
It seems like UA-cam deleted my comment for some reason (Or it's just not letting me read it) and honestly I agree with this comment the most. The only catch I see is that Nvidia and other tech companies might instead try and lobby for DeepSeek's banning, similar to Tiktok, due to its Chinese origins, plus the lost investments.
nice "that one cs agent" beard, lookin fresh
the 9.9 vs 9.11 point you raised is a huge gripe I have with the AI I've used. I ask it to explain why it's done something, or if something is right, and it interprets it as me saying it's wrong. No, I want you to explain your reasoning and check it. and it will say, "omg i'm so sorry, you're absolutely right, a DOES come after b in the alphabet" AHHHH
Nice breakdown Phillip, I think deepseek is the start of something huge, it's the amount they invested for something so powerful that I can't believe. AI's developments are so unpredictable, one small Chinese blip was enough to freak out the stock market and have everybody talking badly about Nvidia's CEO. Just go to the ChatGPT subreddit they're in tatters. The main thing I don't like about AI is how sketchy it feels to me, I don't understand what data these companies have on what we query. I imagine some people aren't even careful about what they ask it.
From those researching Deepseek they are stating that its low budget was due to its party trick of being able to efficiently piggyback off other models, using them as a stepping stones. So very efficient at copying and presenting the output as something legit.
A really Chinese model
So western AI with the same data but done cheaper.
I haven't been watching your videos for a few years. Now, coincidentally, we're wearing the same beard. You look awesome! 😁
AMD RX7900XTX & XT cards are also working very, very well with Deepseek.
Wait, what is encoder preset -1 in HandBrake? 5:15
To be fair. 9.11 is bigger if we're talking about software versioning (which I assume is quite common in the data).
That's what I initially thought but when i saw reasoning that 9.9 is actually 9.90 that clicked.
"which is bigger" says maths to me but "which is higher/newer/more recent" says software versions.
Which is bigger, Windows 10 or Windows 11? Odd thing to ask.
which is bigger, python 3.9 or python 3.13???
also it's a literally bigger string. if you ran a length() method on "9.11" and "9.9" strings, "9.11" would be bigger technically.
also I got this answer from 4o.
> "What's the bigger number mathmatically? 9.11 or 9.9? If you were to order them, how would you do it?"
"Mathematically, 9.9 is the larger number compared to 9.11. This is because:
9.11 is 9 + 0.11, which is 9.11.
9.9 is 9 + 0.9, which is 9.90.
Since 0.90 > 0.11, we conclude that 9.9 > 9.11.
Ordering:
If you were to order them from smallest to largest:
9.11, 9.9 (Ascending order)
If you were to order them from largest to smallest:
9.9, 9.11 (Descending order)
Let me know if you need further clarification!"
so the issue here might be prompting. if you do not give it enough information, it will get things wrong.
@@nabicx giving a lot of information (or long questions) during interrogation is called "leading questions", it allows suspect/interrogee to lie. But if our intention is to hear specific thing, then it is ok to use "leading questions".
These titles are CRAZY
Uhhh.... Why isn't this video in my subscription feed? I'm subscribed AND I have the bell selected to 'all' notifications? Are you seeing any lower than usual performance Philip? Like, seriously, I checked multiple times, ITS NOT THERE. What the hell UA-cam.
I like the competition.
It also shows IP law is bunk. Not to mention ideas can't be property.
🎶🎵Go Phillip, go Phillip, 3kliks, 2kliks. Upload it, I’ll watch it🎶🎵
Phillip slowly turning into John price
Honestly banger video.
Deepseek being absolutely a chad and open sourcing their system which is way better than """open"""AI's original and naïve intentions.
But you're forgetting how chinaphobic the US is. They're too scared about this software, they *could* ban it if they wanted to for some weird reason.
Its like how you cant get the Piko VR headsets in the US and you need to sideload apps to it.
Meta's best competition ruined in the US market.
Deep Seek isn't open source, more like freeware. There is no training code provided. We cannot train the model ourselves. It's still a billion times better than Open AI though.
Nope. You are wrong on multiple places. One of them is:Pico HMD weren't available in US has nothing to do with China-phobic. It is the company who decide not to sell their product to US (or have any reseller channel in US). Americans can always buy one from Ali-express as an imported good. There is *No restriction imposed by any mean/any agency* to prohibit americans to buy and use Pico products. They even offer an international version of its Pico 4 line up which can download app from google. Their "local ver" prohibit the use of Google services and G store (banned by the silly CCP) such that users have to side-load app onto the headset. CCP is Google-phobic and US tech-phobic (Maybe except iPhone 😂). Stop spreading mis-info and lies, pls.
The US is afraid of China and we're just delaying the inevitable at this point.
genie is out of the bottle but after 3 trashy wishes it always goes back into the bottle or lamp
I swear i heard this story before, in 2023 a team at a us uni used gpt3 to train their own model with 200$ and a small data sample, it had 90% of the capabilities and used a miniscule ammount of resources.
there was news about using output from GPT to train another model at lower cost. Today, such model is called "distilled" model.
US companies losing money is always a good time :)
wiki Jevons_paradox: Quote: "occurs when technological advancements make a resource more efficient to use (thereby reducing the amount needed for a single application); however, as the cost of using the resource drops, overall demand increases causing total resource consumption to rise."
The 6 million dollars claim is only the training cost (which is very efficient to be fair). It doesn't include the likely 1 billion dollars for their 50'000 GPU cluster.
Obviously the news didn't mention that, because it would make for a worse story.
Why would they buy the cluster when they can rent space for 100million on an ai sever farm anywhere in the world or in China and split that with their own cheaper gpu .it is still expensive but the issue is 250 million to build an llm model or OpenAI multi billion better? Any person or corporation developing their own ai can download and run for a fraction of the cost. The billion dollars would be 1.35 billion for the 50000 then they make it open source ? most likely cheaper build and provide services to recoup costs my personal gut feeling is 6 rolls of the dice at 2 million dollar to map parameters then 6 million for the final model plus overheads plus initial development possibly 10 final price = 26 million how to recoup build specialised models eg Trading models , creat a super per set distilled model that is continuously learning by distilling from every a i and research ai with branching ,enterprise support tool in one year you can become profitable and still open source if you are three or four generations ahead
@sadenaike12 They are not ahead, they are catching up, which is impressive. I don't know how you got those numbers though.
@@alainx277Problem is you can criticize OpenAI for their investment cost - they're not including XYZ. DeepSeek still built this model with much less capital than the American tech giants.
@@alainx277 GPT-4o:
Input: $2.50 per 1M tokens
Output: $10.00 per 1M tokens
GPT-4o mini:
Input: $0.15 per 1M tokens
Output: $0.60 per 1M tokens
multiply by the 100 of million tokens to train a model if done efficiently by using a multimodal efficient and new algorithm approach and Anthropic used 11million for their latest model but remember the path way to the final model would involve multiple runs costing money
@@MRTOWELRACK They also did it first, Americans were the first ones to land on the moon, that doesnt mean its easy to land on the moon, but it does mean its going to be far less impressive when someone else does it.
Okay I don't get the slight anymosity towards nuclear power since if anything, that's a positive because it means less nuclear waste from coal is accrued. It admitedly means they hold a monopoly of that energy source but assuming things are structurally safe and updated from lessons hard learned from the Three Mile Island Disaster and maybe Fukushima Daiichi (These are west design reactors unlike Russian RBMK), it should be fine. Of course I can't tell if it is.
I'm always up for a video from the microphone collector. And yes, all those billions burned...
Fellas, GPT-2 was open source as well. What's stopping future DeepSeek model from going behind paywall?
the bizarre thing about the nvidia stock drop is that deepseek still used a load of nvidia GPU's to do it, just not the absolute latest version.
you were using v3, not r1. quite surprised especially since you decided to make an entire video on it
that handlebar though
""AI shouldn't be open!" according to OpenAI"" the joke basically wrote itself but it's still so good
Deep seek is super Chinese. Everything I asked was based on a Chinese perspective. ‘Are cucumbers healthy’, ‘yes in china we eat cucumbers’
Love your AI vids. It's just tech that fits you particularly well, both in topic and its use case, I think you understand why so I wont explain further.
This video literally doesn't appear on my subscription feed. Not just I didn't see it there, I have gone back to it, double checked I'm surprised and looked back over the last 12 hours and it isn't there
The drop of nvidia is because deepseek didn't use CUDA and they ported the inference code to run on Huawei GPU no nvidia required.
Missing from my subscriptions tab again, what are you planning mr Philip?
If AI is inevitable then it's in everyone's best interest that it's efficient, the amount of water and power it uses is painful. Although maybe we'll just use it more and the result will be the same consumption of precious resources.
1:36 You actually do need the data to train the AI Deepseek made but only to train a heavier model that is used to train the actual lightweight model. Think of it as training a master blacksmith that knows about everything a master blacksmith should training a young apprentice that knows enough to get by 99% of the time that the master could with significantly lower costs.
One thing I notice all my friends are talking about, is that latelly everything is going to shit
Deepseek feels like Intel launching the 4004. While the big companies were happy to throw larger and larger piles of money, compute and power at the problem, in comes a competitor who found a way to do most of what the big boys do, but making the product more efficient and accessible. Even if Deepseek still used a lot of GPUs to get there, they still beat Open AI at their own game.
Sick hair + beard combo :p