To try everything Brilliant has to offer-free-for a full 30 days, visit brilliant.org/bycloud/. You’ll also get 20% off an annual premium subscription. I have a few more interesting ways of how people use model distillation, lmk if yall would like to see more creative use of this idea!
@@Invuska out of curiosity what are you using to check the transcript? and does it separate text by individual (like if the video had multiple people talking)
Very good overview of Distillation. One thing I felt was kinda missing here: Why is distillation better than just training the student on the original dataset? It is because the teacher model gives better feedback via it's outputs. For example an LLM outputs a probability distribution of what the next token could be "60% chance the next letter is an A, 20% chance it is a B". This makes it way easier for the student to learn than having to figure out these probabilities by itself from many examples where the next letter is either A or B.
Bigger models are better in finding pattens and correlations because their "brains" have more "neurons". Small models are dumber to find correlations as good as big models ... but teacher "big model" can explain those correlations to a small model that could understand it. A small model couldn't grasp a complex patten by itself but after explaining by a "big model" can understand it anyway. Something like that
@@Gebsfrom404 It's less of memorizing and more of being told what is important to understand or learn. That's why you maintain huge amounts of the performance UP TO SOME LEVEL. At some point the compression is too lossy and the distilled model just kinda.. sucks.
@@Gebsfrom404 more like, derive vs understand. You need more knowledge and intelligence to derive those patterns, but not necessarily the same complexity to understand them
Half of the training data...more like answers...the question that are need to be fed to the model needs humans to take care of the data accuracy and governance so in short questions or initial prompts need to be produced by humans ( carefully crafted) the answers are produced by teacher model and rl training of student model to match the answers of teacher model
Thanks, great video! It would be interesting to hear details of how and why distillation process affects finetuning, to better understand the efforts and troubles of stable diffusion / flux fine tunning community.
Once you make the video on quantization, could you give some thoughts on quantization vs model size performance? i.e. given a fixed vram, should I go for a highly quantized 70B or a less quantized 8B for example.
In my experience there's little point in doing inference above 8 bit precision. However I often find a highly quantized 4 bit model behaves very similar to the same model in 8 bit with a high temperature. So if you value creativity over accuracy go with a highly quantized model. So According to this logic, I think 4 bit might be a bit too sloppy if you exclusively value accuracy. I don't really know though. Just do some tests I guess. Maybe compare a 4 bit 70b model vs the corresponding 8 bit 35b model. Just please don't go running an 8b model at 16 bit precision unless you have no other viable options. It would just be a waste of resources.
Literally just talking with a friend that I have had worse responses out of 4o then I have ever had before and he brought up it could be distillation, then this video popped on my feed haha
Ok, then why the hell didn't llama release something between 8B and 70B for 3.1? (at least Mistral graced us with 12B and 22B, those are fantastic btw).
Qwen 2.5 14b and 32b were recently released and they are even better than Mistral 22b (according to both real humans reviews and good benchmarks like MMLU-Pro)
Human teacher model: hmm, this text is likely 89% generated with chatgpt. Student must be cheating. F-. Failed. Expelled. AI Teacher Model: hmm, this is 89% similar to the best ChatGPT outputs, respect. A+. Passed. You’ve now graduated lil student model 🤭🥰
I was jumpscared at 2:58 by that minecraft damage noise. i thought i was getting hit in my game. I don't have minecraft open. I've barely played in years.
I was right , ever since they brought feb2022 update they started degrading the original chatgpt , i started protesting in their servers but i got banned
When will ChatGPT start replacing platforms like Masterclass and Coursera? After all, when the ChatGPT promo briefly demonstrated its language teaching capabilities, stocks for companies like Babel and others took a hit.
The stock market is driven by people with no technical experience following whatever is hyped the most right now. It’s a terrible gauge of how practical these technologies actually are.
I wonder why GPT-4o seems to perform better than GPT-4, when you look at arena leaderboards and stuff. I thought in theory, 4o is just “4 Turbo Version 2”? Is it because 4o is being used more for ChatGPT, which results in more first hand training data, thus being able to improve faster?
Can someone please teach me how to make videos like this? The sheer amount of memes and B-roll in his videos makes me think that he might spend 50% of his time finding memes for each video. Is there any way to automate this process of searching for memes and b-roll?
So, a drip by drip dumb down? This is gonna be leading to funny stuff later down the line. I think they're doing this to get rid of the devs creating it in order to splurge the money elsewhere. Either that or they're really strapped for cash. Given enough distillations passed down like matryoshka, there will only be distortions. I never thought that the downfall of AI would be due to human greed🤣
I have for sure noticed that GPT4 is much better, halucinates less and actually gets things done, pretty much the day GPT4o was released, and I haven't seen anything to convince me otherwise till thos day. At least the new "pause to think model" actually is able to give a response that isn't pulled out of it's ass
@@AB-wf8ek The student models not being fully trained is what provides an LLM that is not as educated as the teacher model. I don't think LLMs existed back then but I would need to check my compSci textbook to see when they were invented.
To try everything Brilliant has to offer-free-for a full 30 days, visit brilliant.org/bycloud/. You’ll also get 20% off an annual premium subscription.
I have a few more interesting ways of how people use model distillation, lmk if yall would like to see more creative use of this idea!
hi , in this cideo could you please mention that it is not actually RL but RLHF, that distinction is actually important to not ignore. thanks
drink a shot every time he says "Model"
I’m drunk already
He says model 70 times according to the transcript. Nah, I'm good lmao 😂
@@Invuska out of curiosity what are you using to check the transcript?
and does it separate text by individual (like if the video had multiple people talking)
meow
purr
@@leoj8633 there's a button for it in the video description
Very good overview of Distillation. One thing I felt was kinda missing here: Why is distillation better than just training the student on the original dataset? It is because the teacher model gives better feedback via it's outputs. For example an LLM outputs a probability distribution of what the next token could be "60% chance the next letter is an A, 20% chance it is a B". This makes it way easier for the student to learn than having to figure out these probabilities by itself from many examples where the next letter is either A or B.
Bigger models are better in finding pattens and correlations because their "brains" have more "neurons".
Small models are dumber to find correlations as good as big models ... but teacher "big model" can explain those correlations to a small model that could understand it.
A small model couldn't grasp a complex patten by itself but after explaining by a "big model" can understand it anyway.
Something like that
@@mirek190 Sounds like understand vs. memorize.
@@Gebsfrom404 It's less of memorizing and more of being told what is important to understand or learn. That's why you maintain huge amounts of the performance UP TO SOME LEVEL. At some point the compression is too lossy and the distilled model just kinda.. sucks.
@@Gebsfrom404 more like, derive vs understand.
You need more knowledge and intelligence to derive those patterns, but not necessarily the same complexity to understand them
@@TragicGFuel exactly
Teacher model: _generates your training data_
Half of the training data...more like answers...the question that are need to be fed to the model needs humans to take care of the data accuracy and governance so in short questions or initial prompts need to be produced by humans ( carefully crafted) the answers are produced by teacher model and rl training of student model to match the answers of teacher model
I dare you to find a better channel about technical AI... There's none. Amazing job as always bycloud !👏👏👏
Subscribe to AI explained and Bycloud , and you're basically SET!
@@AfifFarhaticame here to say AI explained too.
4o gotta be distilled af
To be honest for me it seems that ChatGPT is given better answers and is faster. I don't pay plus, but I am enjoying the process
@@andreilucasgoncalves1416 I feel the opposite
GPT 4o is distilled. GPT 4o mini is distille AF
Moonshine
Gpt4o is much dumber than Gpt4
WE DISTILLING OUR MODELS WITH THIS ONE 🗣️🗣️🗣️👏👏👏🔥🔥🔥🔥🔥🔥
Thanks, great video! It would be interesting to hear details of how and why distillation process affects finetuning, to better understand the efforts and troubles of stable diffusion / flux fine tunning community.
it would be awesome a video about quantization!
Once you make the video on quantization, could you give some thoughts on quantization vs model size performance? i.e. given a fixed vram, should I go for a highly quantized 70B or a less quantized 8B for example.
In my experience there's little point in doing inference above 8 bit precision. However I often find a highly quantized 4 bit model behaves very similar to the same model in 8 bit with a high temperature. So if you value creativity over accuracy go with a highly quantized model. So According to this logic, I think 4 bit might be a bit too sloppy if you exclusively value accuracy. I don't really know though. Just do some tests I guess. Maybe compare a 4 bit 70b model vs the corresponding 8 bit 35b model.
Just please don't go running an 8b model at 16 bit precision unless you have no other viable options. It would just be a waste of resources.
Thank you very much for this video! Just wondering, have you done the video about 1.58 bit already?
There is definitely quality degradation with model distillation.
This especially noticeable with diffusion models.
Literally just talking with a friend that I have had worse responses out of 4o then I have ever had before and he brought up it could be distillation, then this video popped on my feed haha
Nice video👍
Ok, then why the hell didn't llama release something between 8B and 70B for 3.1? (at least Mistral graced us with 12B and 22B, those are fantastic btw).
Qwen 2.5 14b and 32b were recently released and they are even better than Mistral 22b (according to both real humans reviews and good benchmarks like MMLU-Pro)
Just what I needed 🤝🏾🗿
Loved the vid, as always
Human teacher model: hmm, this text is likely 89% generated with chatgpt. Student must be cheating. F-. Failed. Expelled.
AI Teacher Model: hmm, this is 89% similar to the best ChatGPT outputs, respect. A+. Passed. You’ve now graduated lil student model 🤭🥰
This one made me lol'd. Very informative. Thanks!
They could make money from LLMs by making voice mode uncensored for 18+
Can you makes sure to update your MoE research? Seems promising. Maybe also AI hierarchy stuff?
I was jumpscared at 2:58 by that minecraft damage noise. i thought i was getting hit in my game. I don't have minecraft open. I've barely played in years.
Brilliant video.
System prompt makes anthropic response output really SHIT
didn't understand anything. Could you please repeat or reupload?
I think if you press the "join" button beside "subscribe" you will instantly understand the video
Real 😭
I was right , ever since they brought feb2022 update they started degrading the original chatgpt , i started protesting in their servers but i got banned
Yeah, you basically explained why your sponsor "Brilliant" isn't the best way to learn.
Woah that hits hard, wasn‘t the best choice for this video 😂
When will ChatGPT start replacing platforms like Masterclass and Coursera? After all, when the ChatGPT promo briefly demonstrated its language teaching capabilities, stocks for companies like Babel and others took a hit.
it’s harder but not impossible. The main selling point for them is UX, chatgpt can’t accommodate those as well rn
The stock market is driven by people with no technical experience following whatever is hyped the most right now. It’s a terrible gauge of how practical these technologies actually are.
Oh how I miss release week chatgpt-4
As a non-native speaker of English, I am responsible for saying that I didn't understand anything.
I wonder why GPT-4o seems to perform better than GPT-4, when you look at arena leaderboards and stuff.
I thought in theory, 4o is just “4 Turbo Version 2”?
Is it because 4o is being used more for ChatGPT, which results in more first hand training data, thus being able to improve faster?
Nice
now that we have o1 the previous 4o is crapp
o1 mini and 4o large are about the same size
@@erkinalp The compute for 4o is almost unbearable for coding now that o1 is available :D
@@erkinalp Where can i read more about this?
Can someone please teach me how to make videos like this? The sheer amount of memes and B-roll in his videos makes me think that he might spend 50% of his time finding memes for each video. Is there any way to automate this process of searching for memes and b-roll?
community: What is distillation? Let's finetune Flux anyway. *No full "checkpoints" since then 🎉
Who is going to open source 300k GPUs to train state of the art models? 😅 We should already be very glad with what we have.
ChatGPT-4 is better than it’s the newer models…
it got so much worse
So, a drip by drip dumb down?
This is gonna be leading to funny stuff later down the line.
I think they're doing this to get rid of the devs creating it in order to splurge the money elsewhere.
Either that or they're really strapped for cash.
Given enough distillations passed down like matryoshka, there will only be distortions.
I never thought that the downfall of AI would be due to human greed🤣
So AI goes to college.
All this and in future people are gonna make CS:GO cat edits with this
I have for sure noticed that GPT4 is much better, halucinates less and actually gets things done, pretty much the day GPT4o was released, and I haven't seen anything to convince me otherwise till thos day.
At least the new "pause to think model" actually is able to give a response that isn't pulled out of it's ass
too fast... please... distill your pace
There is "playback" button. Put it on 0.5x and stop whining
Sorry , I had to stop watching after the 20th dictation of “model” in the first 3 minutes.
We're getting scammed
Gpt 3.5
GPT 4 (costs money)
GPT 4 Turbo (worse)
GPT 4o (worse)
GPT 4o mini (worse)
o1 (not necessarily worse, but even smaller) way too expensive
o1 preview (even smaller)
o1 mini (super small)
Parameter count
3.5: 175B ($3/4 per M)
4: 1.7T ($30/60 per M)
4 Turbo: 500B? ($10/30 per M)
4o: 250B? ($5/15 per M)
4o mini: 8B? ($0.15/0.60 per M)
We should've been on GPT 4.5 which had 17T and anticipating GPT 5 which has 175T parameters.
Instead they giving 8B parameter models
@@ckq 4o is 880B, 4 turbo and 4o mini are 220B
Matryoshka Dolls
@@ckq 8B ones are what free tier users are routed through when they invoke the "GPTs" tools, not directly selectable through the menu.
guys , Im just starting out as an AI enthusiast making similar content
would appreciate your feedback
"A disciple is not above his teacher, but everyone when he is fully trained will be like his teacher."
- Jesus, Luke 6:40
I didn't realize JC was an LLM developer
@@AB-wf8ek The student models not being fully trained is what provides an LLM that is not as educated as the teacher model.
I don't think LLMs existed back then but I would need to check my compSci textbook to see when they were invented.
gpt4-o1 preview is a massive upgrade
bro is onto nothing 🔥🔥🔥
What about just pruning and fine tuning? @bycloud