Thanks for doing the live coding exercise and leaving the mistakes in. Sometimes. I wonder if it's only me who has experiences like that😅 Thanks again!
Love your videos, been following for a while. Have you done any benchmarking to compare the Flan-T5 based ALPACA to the LLaMa based ALPACA? Curious to know what the final performance of the Flan-T5 open model is in comparison to the original ALPACA model
Have to build the channel first, then activate monetization (guess currently its about 5cents per video per month), all additional expenses for Azure 640GB Alpaca benchmarks are second in list.
Your videos are a really great man. Am pretty new to these LLMs and starting to get around the idea of this AI revolution. What would be the best way to reach out to you if I had any questions in mind🤔?
This is amazing. Huge thanks for this. Any chance you've got a link to the notebook? Btw, I'm getting a CUDA out of memory error on a machine with 24GB of GPU RAM. Any chance you've got a pointer I might be missing?
Welp, I just noticed you're training a t5-large and I was trying a t5-xlarge so it looks like I answered my own question there :) Anywho, if you happen to have the notebook link that would be cool. Again, this is awesome.
I am huge fan of your videos. i am non programmer but have big interest in this stuff. I wanna request to actually give us a walkthrough on how to prepare our own data like some image pdf files to convert into vector embeddings and use contextual injections to make our own bots even for personal purposes. One more thing I am confused about is how to prepare examples to train our AI to do specific tasks like writing a report in a specified manner and use the specified vocabulary. Is there any written resource where I could learn that? Please reply. 😊
If you are not a programmer like you say then watching videos to code your own AI system has a steep learning curve. If you want to start, get familiar with the transformer architecture of AI models, every i/o question will then be answered.
If you run the small PLM on a free COLAB T4, nothing. But for bigger LLMs I have a new video upcoming with 8bit quantization and PEFT-LoRA which would allow for a faster and cheaper fine-tuning.
Thanks for doing the live coding exercise and leaving the mistakes in.
Sometimes. I wonder if it's only me who has experiences like that😅
Thanks again!
Why dont you give your colab code link too? :\
Great videos ! The new cool kid in the neighbourhood
Do you have the Google collab notebook for this?
Next time i hope the video about Alpaca lora
Same, it would be great to see a breakdown on how to train one
Love your videos, been following for a while.
Have you done any benchmarking to compare the Flan-T5 based ALPACA to the LLaMa based ALPACA? Curious to know what the final performance of the Flan-T5 open model is in comparison to the original ALPACA model
Have to build the channel first, then activate monetization (guess currently its about 5cents per video per month), all additional expenses for Azure 640GB Alpaca benchmarks are second in list.
Your videos are a really great man. Am pretty new to these LLMs and starting to get around the idea of this AI revolution. What would be the best way to reach out to you if I had any questions in mind🤔?
great video mate. Sorry to ask but can you provide the COLAB? Thanks once again for your knowledge. Cheers
As mentioned, code sequences are available as python files to download.
This is amazing. Huge thanks for this. Any chance you've got a link to the notebook? Btw, I'm getting a CUDA out of memory error on a machine with 24GB of GPU RAM. Any chance you've got a pointer I might be missing?
Welp, I just noticed you're training a t5-large and I was trying a t5-xlarge so it looks like I answered my own question there :) Anywho, if you happen to have the notebook link that would be cool. Again, this is awesome.
can you please share your colab code, im beginner in this and finding this very difficult to start with. Thanks :)
I am huge fan of your videos. i am non programmer but have big interest in this stuff. I wanna request to actually give us a walkthrough on how to prepare our own data like some image pdf files to convert into vector embeddings and use contextual injections to make our own bots even for personal purposes. One more thing I am confused about is how to prepare examples to train our AI to do specific tasks like writing a report in a specified manner and use the specified vocabulary. Is there any written resource where I could learn that? Please reply. 😊
If you are not a programmer like you say then watching videos to code your own AI system has a steep learning curve. If you want to start, get familiar with the transformer architecture of AI models, every i/o question will then be answered.
Can we use it for text classification
Please do this for the new Dolly Dataset. That would be epic
I didn't understand a THING ! What is this ? What is it used for ? CONTEXT PLEASE !
Easy.
Flan-T5 LLM is explained here: ua-cam.com/video/SHMsdAPo2Ls/v-deo.html
Alpaca is explained here: ua-cam.com/video/j6dqO2dSF9c/v-deo.html
Enjoy.
The channel IS the context. Not all series have to do the "Previously on 24" thing, Jack Bauer.
How much $$$ did the training cost you?
If you run the small PLM on a free COLAB T4, nothing. But for bigger LLMs I have a new video upcoming with 8bit quantization and PEFT-LoRA which would allow for a faster and cheaper fine-tuning.