Like the actual detaill on the tool calling for function calling . Would appreaciate another tool calling demo (with more tools) with gpt 4o but using a the new langchain tool calling approach so that we can then swap any llm foundational model. Would be good to see if that abstraction makes it easier in other wayus as well once get more complex.
Hello Friend! Thanks a lot for the video. Your colab has text questioning, function calling and image questioning, but i would like to send a pdf and ask about it. Would you tell me the correct portion of the code to send a pdf file? I tried this but it didn't work: response = client.chat.completions.create( model=MODEL, messages=[ {"role": "user", "content": [ {"type": "text", "text": "Please summarize this pdf in bullet points."}, {"type": "pdf", "data": pdf_data} ]} ], temperature=0.0, ) Thanks bro!
🎯 Key Takeaways for quick navigation: 🌆 Das GPT-40-Modell kann Text und Bilder verarbeiten, um Antworten zu generieren. ⚡️️ Die Verarbeitungsgeschwindigkeit von GPT-40 ist schneller als die von GP4-Turbo. 🔍 Das Modell kann auch Funktionen aufrufen, um bestimmte Aufgaben auszuführen. 📊 Das Modell kann JSON-Antworten generieren und Bilder verarbeiten. 💻 GPT-40 kann als API verwendet werden, um es in eigenen Python-Skripten zu integrieren. 🎉 Das Modell kann auch Emotionen aus Bildern erkennen und beschreiben. 👥 Das Modell kann Funktionen aufrufen, um bestimmte Aufgaben auszuführen, wie zum Beispiel das Abrufen von NBA-Spielständen. 🔓 GPT-40 ist noch nicht in der Lage, Videos direkt zu verarbeiten, aber es gibt Möglichkeiten, Bilder aus Videos zu extrahieren und dann zu verarbeiten. Made with HARPA AI
The fact that the information cut off date is September 2021 suggests that it's based on GPT 3.5 data. I sense that OpenAI may be being mendacious. For example, it's seems unlikely that 4o would be so much faster than 4 given the claims of superior abilities. I think OpenAI has more explaining to do.
@@engineerprompt I considered this, so I tested it against information that was time based to confirm that it's knowledge base doesn't know about information after Sept 2021.
If you are interested in learning more about how to build robust RAG applications, check out this course: prompt-s-site.thinkific.com/courses/rag
finally someone use simple encoder to process image with brief explanation here your subs man!
good video mate, thanks. it helped me a lot
Thank you for the video! Your video helped solve the problem :)
Excellent info, thank you. What tool did you use for screen capture and camera following mouse cursor?
Like the actual detaill on the tool calling for function calling . Would appreaciate another tool calling demo (with more tools) with gpt 4o but using a the new langchain tool calling approach so that we can then swap any llm foundational model. Would be good to see if that abstraction makes it easier in other wayus as well once get more complex.
that's on my list
hello can this help me edit images, remove the background from my images which are on my WordPress database
Make video on voice and videos input as well.
I have tried it, bad performance just like gpt4v
Please create a video on how to process video as a series of images
Curious what python developer environment that is from 4 mins +?
It's google colab
Hello Friend! Thanks a lot for the video. Your colab has text questioning, function calling and image questioning, but i would like to send a pdf and ask about it. Would you tell me the correct portion of the code to send a pdf file? I tried this but it didn't work:
response = client.chat.completions.create(
model=MODEL,
messages=[
{"role": "user", "content": [
{"type": "text", "text": "Please summarize this pdf in bullet points."},
{"type": "pdf", "data": pdf_data}
]}
],
temperature=0.0,
)
Thanks bro!
🎯 Key Takeaways for quick navigation:
🌆 Das GPT-40-Modell kann Text und Bilder verarbeiten, um Antworten zu generieren.
⚡️️ Die Verarbeitungsgeschwindigkeit von GPT-40 ist schneller als die von GP4-Turbo.
🔍 Das Modell kann auch Funktionen aufrufen, um bestimmte Aufgaben auszuführen.
📊 Das Modell kann JSON-Antworten generieren und Bilder verarbeiten.
💻 GPT-40 kann als API verwendet werden, um es in eigenen Python-Skripten zu integrieren.
🎉 Das Modell kann auch Emotionen aus Bildern erkennen und beschreiben.
👥 Das Modell kann Funktionen aufrufen, um bestimmte Aufgaben auszuführen, wie zum Beispiel das Abrufen von NBA-Spielständen.
🔓 GPT-40 ist noch nicht in der Lage, Videos direkt zu verarbeiten, aber es gibt Möglichkeiten, Bilder aus Videos zu extrahieren und dann zu verarbeiten.
Made with HARPA AI
Please CrewAI with gpt4o manage and haiku assistants. And debugging it. Thanks!
The fact that the information cut off date is September 2021 suggests that it's based on GPT 3.5 data. I sense that OpenAI may be being mendacious. For example, it's seems unlikely that 4o would be so much faster than 4 given the claims of superior abilities. I think OpenAI has more explaining to do.
I am not sure if the model actually knows their cutoff date. It's most probably hallucinating
@@engineerprompt I considered this, so I tested it against information that was time based to confirm that it's knowledge base doesn't know about information after Sept 2021.
Why I cannot use gpt4o??
are you on the PLUS plan for free plan?
@@engineerprompt free
@@engineerprompt is it necessary to buy Plus plan for gpt 4o access ?
processing an image converted to b64 is expensive?(above 100000 tokens)
You can greatly reduce the image size and quality to cut down the tokens on API call, yet still achieve decent results
@@kristianlavigne8270 maybe 70000 tokens is a good option for you, not for me
Please create a video on how to process video as a series of images