INSANELY Fast AI Cold Call Agent- built w/ Groq
Вставка
- Опубліковано 1 чер 2024
- What exactly is Groq LPU? I will take you through a real example of building a real time AI cold call agent with the speed of Groq
🔗 Links
- Follow me on twitter: / jasonzhou1993
- Join my AI email list: www.ai-jason.com/
- My discord: / discord
- Vapi AI: vapi.ai/
- Groq: groq.com/
- RelevanceAI: relevanceai.com/
⏱️ Timestamps
0:00 Intro
1:07 CPU vs GPU vs LPU
8:45 What is LPU
10:27 Use cases Groq unlock
13:42 Tutorial: Build sales agent with voice AI
16:20 Demo: Voice AI
17:54 Setup Phone number for AI agent
19:00 Integrate voice AI into existing WhatsApp sales agent
23:58 Demo
👋🏻 About Me
My name is Jason Zhou, a product designer who shares interesting AI experiments & products. Email me if you need help building AI apps! ask@ai-jason.com
#groq #gpt5 #whisper #whisperkit #mixtral #gpt4turbo #gpt4 #ai #artificialintelligence #tutorial #stepbystep #openai #llm #chatgpt #largelanguagemodels #largelanguagemodel #bestaiagent #chatgpt #agentgpt #agent #autogen #autogpt #openai - Наука та технологія
What are the use cases you want to see me building with Groq?
Are you working for groq now?
I need a ai girlfriend
Personal agent that ‘sees’ zpwhat you do on your computer / phone and helps with it. (By sending a screenshot to it)
Doing literally anything requiring intelligence beyond a basic best-case simple script.
I have worked with a non-profit that helps with Fair housing problems. I think a good use case would be receiving calls for a business and helping the customers understand if they are having a real fair housing problem.
Would Groq be able to receive phone calls?
My first thought is how can we use this for scam baiting? We just need an elderly person's voice option to make the call and then prompt the AI to waste the scammers time talking about gift card activation codes.
Until the AI conjures up real credit card information from within its data and then some unfortunate persons life savings are gone 😢
@@venim1103 yeah nah mate.
i like this idea!
i suppose it is the other way around^^ natural speaking "people" will now scam old persons
It still costs tokens tho
Yes because we all want more cold calls from sales bots.
came here to also say this. Yech... Leave the calling to the humans, everything automated should have been an email.
Sure but what about more cold calls from better sales bots?
@@hiandrewfisher
Sales bot or human, what ever company still thinks in our time that cold calling is the way to go, is beyond the point of saving, and it should go bankrupt, for its own stupidity. The bots will just speed up that process.
@@nikolaizaicev9297i make 100k a year off of coldcalls
@@nikolaizaicev9297amen to that
This is one true gem of a video that focusses more on the use case. Thank you for breaking down the concepts really well and showing us demo of it's capabilities
🎯 Key Takeaways for quick navigation:
00:32 *🧠 Introduction to Groq's LPU (Large Language Model Processing Unit)*
- Introduction to Groq's LPU architecture designed specifically for AI inference.
- Explanation of the need for LPU in large language model inference.
- Comparison between LPU and other processing units like CPU and GPU.
05:37 *🔍 Comparison between CPU and GPU*
- Description of CPU as the central processing unit and its limitations in parallel computing.
- Explanation of GPU architecture, parallel computing power, and its expansion beyond gaming.
- Illustration of the difference between CPU and GPU through a painting demonstration.
06:05 *🔄 Limitations of GPU in Large Language Model Inference*
- Discussion on the limitations of GPU in handling large language model inference.
- Explanation of the complexities in achieving sequential execution on GPU.
- Overview of the latency issues and the need for complex control mechanisms.
09:47 *🚀 Groq's LPU Architecture and Performance Benefits*
- Introduction to Groq's LPU architecture designed for sequential tasks and low latency.
- Explanation of the simplified architecture and shared memory advantages.
- Discussion on the predictability and performance gains achieved with Groq's LPU.
11:37 *🗣️ Applications of Fast Inference Speeds*
- Exploration of potential applications such as real-time voice AI for natural conversations.
- Discussion on the reduction of latency enabling smoother interactions.
- Demonstration of real-time voice AI and its impact on user experience.
13:17 *🖼️ Utilization in Image and Video Processing*
- Highlighting the effectiveness of Groq for real-time image and video processing.
- Demonstration of image processing capabilities for various applications.
- Discussion on unlocking consumer-facing use cases with fast inference speeds.
14:40 *🤖 Building Real-time Voice AI with Groq*
- Discussion on building outbound sales agents using real-time voice AI.
- Introduction to platforms like Vee for integrating voice AI into applications.
- Demonstration of setting up a real-time voice AI assistant using Groq's model.
00:00 *📞 Setting Up Real-time Voice AI Cold Call Agent*
- Setting up a real-time voice AI cold call agent using Groq technology.
- Integration of voice AI capabilities into existing agent systems.
- Configuring API calls and server URLs for seamless communication between systems.
19:18 *🛠️ Integrating Real-time Voice AI with Existing Agent Systems*
- Demonstrates how to integrate real-time voice AI with existing agent systems.
- Setting up agent tools for making phone calls and receiving transcriptions.
- Configuring metadata and webhooks for seamless communication between platforms.
20:41 *📞 Configuring Call Functionality and AI Assistant*
- Configuring call functionality within agent systems for real-time voice AI interaction.
- Setting up dynamic message generation and personalized interactions.
- Defining schemas, URLs, and metadata for effective communication between systems.
Made with HARPA AI
This is really interesting. Thanks for the sharing Jason.
Loved this Jason!!! Thank you
You're incredible. Thanks for this Demo, Jason Sensei.
I wonder how many "Nigerian Prince" this thing could run in parallel? 🤔🤭
😂😂😂😂😂😂
Thanks for this awesome content, first time on your page but this is great and simple to follow and understand!
Another awesome video with great presentation and overview ,i give your video's example to many to make them understand how to educate viewer abour particular thing and tell about what,why,how and then implement things in easy way possible.
Keep feeding us quality content buddy :-))
So not quite there yet or reliable enough but getting closer. Thanks for these insights!
These are amazing use cases!! Lowering the barriers of entry to do high quality business associated with big companies!!
Thanks Jason
Thanks Jason for the good work.
Creating a UI questionnaire for non coder types to build applications to solve problems. Mostly business applications that might otherwise require a developer or consultant.
the thumbnail of this video is really cool, the text looks like it sticks out.
Great share. Seriously grateful for creators like you!
Many thanks for never bothering to define what LPU is an actual acronym for.
17:17 That is so fast and seamless. Super cool.
wow man this is incredible... holy molly!
amazing stuff 💯
Good stuff! Keep it up
Really great synopsis
I can't trust anything anymore! The demo in the end is very impressive
This is so powerful but also scary, what the world will look like in 12 month, when all the communication are driven by AI?
you would be busy scratching your balls, while AI does everything else.
Its easy to see this will replace all callcenters very soon. I assume they originally developed this chip for the new Tesla Autopilot software, that is mainly AI/video based.
I loved your video!
Thank you for detailed, informative content 10/10
WOW! Amazing tutorial. Top 3 I've watched ever! Keep up the great work! 🎉
what other two?
Tell us what other two asap ! Why are you treatening us like that.
I would say Trelis Research has good content youtube.com/@TrelisResearch?si=oM1o4NaE30h2nI4y and learning wise all of Lex Fridman youtube.com/@lexfridman?si=yHJb1O-mzDYqS6c1
Seems like my replies to the questions were deleted by UA-cam 😑
Can't wait to try this on some use cases I have in mind :D Great video as usual ;)
sexy Ai girlfriend 😘😍
The phone number thing is interesting... makes me fantasize about being able to have this as a replacement for the "leave a message after the beep" answering machines for your mobile if you don't get a call. A lot of people find leaving a message without having a conversation really awkward, so if you could instead connect to an AI assistant like this that actually talks to you, you could leave better messages, and the AI can summarize the conversation and leave you a txt message of the contents, or just leave their own summarized voice message.
nobody listens to answerfone messages, not since abourt 2007 id say haha
That’s a super amazing idea. Build it! You will become rich lol
You just described an AI secretary and yes this would be an amazing tool. Build it !!
With all this current technology it is possible to create a really cool AI girlfriend. And highly customizable.
@@abandonedmuse Launched it today and I'm still not rich lol
Excellent video! Keep up the good work.
Great explanation and example. Thank you very much.
Love this Jason, keep'em coming !!
Thanks, Jason for the great work!
Thanks a lot mate!
@@AIJasonZ use the ai to order pizza
This is awesome. I've seen a bunch of Voice AIs and all of them have terrible latency issues as well as obvious AI voices. Using Groq to get the latency way down and custom voices with PlayHT solves both issues. Thanks for sharing!
Awesome tutorial! The output seems to be conversation-aware. How can I train the voicebot so it will handle questions, and scripted answers the way I want it to? Would this be done in Groq? Your fitness caller did a great job and asked relevant questions to qualify you and give her an idea of where to go with the conversation...and the focus was on helping you and sales. Keep up the great work! I'm going to watch your video on how you built AI Agents for Research.
2:55 "In every frame 2 million pixels have to be generated"
This guy broke down graphics in a way that made sense, for the first time in 20 years.
Good for you ✌
isnt true dough, it just needs to get the pixels who are changing. And you dont render every pixel alone, but in object for object.
@@danielchoritz1903 In graphics you are rendering every pixel. You're talking about video codecs, whole different ballgame.
How wonderful, this is bound to improve trust among people and all of our lives. This is the best thing that science has wrought since industrialized warfare. Thank you, technology.
Helluva a channel you’re growing here big dawg! Keep up the good work!
Thank you for the background information reviews you provide.
Thank you for covering this, we are building AI Applications using groq. Fast, cheap, and reliable.
Hesu, Jason
The best channel
You grew so much
Since the first video
I love this moment
Where I am like,
Opening the feed,
Oh okay, Jason released a new video,
"Well, it's probably _Good As Always_".
...
Proceed to watch
...
ABSOLUTE PERFECTION
HANDS DOWN
MAJESTIC
INFORMATION BOILED DOWN LIKE A
BOOSTED MONKEY ANIMAL YOU ARE
NEVER HAVE I SEEN
THINGS PUT IN THAT MANNER TOGETHER
MUCH HARMONY
STRONG BALANCE
RESONANCE LEVEL?
DEEeeeeee
eeeeeeeeeee
eeeeeP.
From the Bottom of my heart,
With Love & Respect
Ivan
hey great video - can you do a full walkthrough of relevanceai and how you set that agent up as its not possible to follow from your video as looks like you had some pre defined steps in there thanks or drop and drop a link to the code you used to build this? thanks
Well done. It could be helpfull for custommer support actions
Awesome Video
Wow. Great information. Thank you for sharing!
good demo
Really insightful video boss
Thank you.
1:31 "I haven't do exercise at all for the past 3...or 6 months..." 😂
Awesome job man
❤ great video, great explanation!
Amazing video Jason.
Great video. Thanks for sharing
I'll have to try this. I managed to get very fast, close to realtime speech with the chatgpt api using a few queues and a local text to speech. The slowest part was the actual speech to text processing i believe. I was using whisper before they added all the new upgrades to the gpt api (this was when gpt3.5 just came out basically).
It just processed two sentences to speech and put out the audio while it provessed the next sentences. The issue was that twilio made it very difficult to work with this since I needed to make it a stream and that required some realtime communication protocol that worked over phone, so i just stopped and had my own little chat assistant. Im a weeb. It was an anime girl ai assistant.
We did this too, some of the audio engines even give an output that tells you the realtime factor -> if it's less than one, it means you can generate the sentences faster than they can be spoken! Basically we used a queue and pipe-lining to reduce the mean time to first output.
I don't think you need these LPU things unless you're trying to use an online service that just bulk process a bunch of sentences.
super @@ultimape
great video
Even if you are misleading with the idle cut times on the demo its impressive.
Great video, would be awesome if you could make one video of building a wrapper like this from scratch 😀
This was really good. Thank you J! I have a use case for HR.
This is why I subscribed!
Thanks a lot. appreciate the details. Very helpful
God my manifestation skills went through the roof this time. Only 7 minutes from process start until this video magically materialized.
yeah you’re gonna blow up bro 🎉 great video
Hi Jason , great content , I just have one remark concerning the demo , the video is being cut it would be really nice if it was left intact just to have an idea of the latency , otherwise nice video
Brilliant video bravo 👍
amazing video Jason
that intro was gold
awesome@
Amazing video!!tysm
Amazing!
Sick
Awesome video, would recommend blurring out your whatsapp next time tho, just to make sure no pranksters mess with your friends.
Agreed!
I woke up to this video. My day just got 10x better 😄.
I'm as a web dev, new to the field of AI LLMs. Just found Groq yesterday and tried with API access. but didn't understand what the heck is a LPU thing.but here we are! Thanks again!
As a non-dev, I am _so_ looking forward to tools like these.
Another amazing video!
I loved the Crysis reference hahaha
Great Video! Thank you! Ist IT possible only for WhatsApp to write with the customers?
good job
Can’t wait to watch this
This is amazing!
Really interesting video, thank-you!
This is awesome thanks 🙏
The highly-anticipated tool use (aka function calling) feature for Groq API was released last week!
Synthflow is pretty cool too
They even added vocal fry to the woman’s voice for realism. * slow clap *
I think it'd be very cool to use this for on-demand mini language lessons. Imagine before you go into any situation where you will be able to use your target language you can set up a quick call with the AI and have it role-play a conversation with you. And you could iteratively improve your language skills per situation. And have transcripts to further work on with your flesh and blood language teacher.
Nice India will love this...
Hi Jason! I am impress about the content of the video. Do you know how much could cost a call for an Agent with that setup?
Eleven labs conversational voice is so good for this. You should do it.
Great video, but just to clarify: GPU is Graphics Processing Unit not General Purpose Unit
Interesting ai, gonna give it a whirl on monday with my turbo api keys
This workflow is insane for CRM.
This is amazing. Where can we get the code and stuff that you do your channel?
The Sales Agencies after watching this video: „Ah f*** this sh*t, let‘s learn some new skills“
😂😂😂😂😂😂😂😂😂
Hey Jason thank you for sharing ! Any ressources on connecting relevance ai to WhatsApp Business?
Why are you edited your final demo to make responces appear faster than they actually are?
I can’t wait for this technology to get better. I need AI agents to for sales 😊
It's good enough now, why wait.
As far as I know from the All in Podcast, “Groq” isn’t particularly made to be the LPU or language processing unit. It was build as a very parallel processor and had little use case until it was a perfect fit for LLMs.
The brown skinned dude from the podcast owning a stake in the “Groq” company,
also explained, that they didn’t have a compiler as in Nvidias Cuda, thus they build one in the last year.
As the company was working on the idea for a while. It is more like the use case fits the product.
LLMs definitely don’t exist long enough, that it was specifically made for it.
So even as the LPU might be an adequate description right now:
It rather looks like the chip picked up that profession, when growing up/maturing.
Perfect timing interval for success:
-Later and we would see another chip taking the spotlight, even if a little later.
-earlier and the company might have bankrupted, if no use case were to be found
The company wasn't built for LLM's , mostly for providing processors specifically for Machine Learning use cases. The LLM wave was just something they were uniquely in a strong position to pursue, so they made a small natural pivot.
Great Work!!!
It will be something when ai can interrupt into a conversation correctly.