- 253
- 360 492
AI Makerspace
Приєднався 28 січ 2023
Join Dr. Greg & The Wiz every week at AI Makerspace, the world's leading community for people who want to 🏗️Build, 🚢Ship, and 🚀Share production Large Language Model (LLM) applications!
AI Makerspace Transformation of the Week - Tshwanelo Matsane #totw #genai #aim
In this Transformation of the Week, I chat with Tshwanelo Matsane, a strategy consultant and recent graduate from the AI Engineering Bootcamp, who wanted to build his own applications, and be his company's AI thought-leader.
Learn bleeding-edge Gen AI concepts, best practices, and tools with the pros at AI Makerspace. Come Build, Ship, and Share with the best AI practitioners in the world. Upskill in less than 10 weeks with the highest rated bootcamp on Maven.
aimakerspace.io
#genai #learnai #totw #buildshipshare #aim
Learn bleeding-edge Gen AI concepts, best practices, and tools with the pros at AI Makerspace. Come Build, Ship, and Share with the best AI practitioners in the world. Upskill in less than 10 weeks with the highest rated bootcamp on Maven.
aimakerspace.io
#genai #learnai #totw #buildshipshare #aim
Переглядів: 26
Відео
LLM Engineering or AI Engineering - which are you more interested in?
Переглядів 11916 годин тому
LLM Engineering or AI Engineering - which are you more interested in? We heard from so many beginners in our community asking, “so what’s the difference?” Wonder no more! Dr. Greg of AI Makerspace is breaks it down in simple terms to give you a quick explanation. #llme #genai #aie Learn bleeding-edge Gen AI concepts, best practices, and tools with the pros at AI Makerspace. Come Build, Ship, an...
AI Makerspace Transformation of the Week - Laura Funderburk #totw #genai #aim
Переглядів 6214 днів тому
In this Transformation of the Week, I chat with Laura Funderburk of Bytewax, a graduate from the AI Makerspace LLM Operations cohort who started off in the call center, and is now an AI Engineer writing a book on Gen AI! Learn bleeding-edge Gen AI concepts, best practices, and tools with the pros at AI Makerspace. Come Build, Ship, and Share with the best AI practitioners in the world. Upskill ...
AI Makerspace Transformation of the Week - Pano Evangeliou #totw #genai #aim
Переглядів 11921 день тому
In this Transformation of the Week, I chat with Pano Evangeliou, a graduate from the AI Makerspace LLM Operations cohort who didn’t let a language barrier or time zones, stand in his way of Building, Shipping, and Sharing his way to becoming a Senior AI Engineer! Learn bleeding-edge Gen AI concepts, best practices, and tools with the pros at AI Makerspace. Come Build, Ship, and Share with the b...
LLM Engineering Challenge Walkthrough with The Wiz
Переглядів 40821 день тому
If you're looking to get started learning LLM Engineering, you're in the right place! The Wiz is here to walk you through, step-by-step, exactly how to complete the challenge. If you're a beginner, this is exactly how the homeworks will feel for you! If you're more advanced, we're going to make sure that the homeworks feel quite a bit more challenging! The LLM Engineering Challenge is designed ...
CareCompanion: AI Support for Dementia Caregivers
Переглядів 13321 день тому
This AI-powered chatbot provides practical support and resources to caregivers of individuals with chronic illnesses like dementia. By answering common questions, providing emotional support, and connecting caregivers to local resources, the chatbot helps caregivers navigate daily challenges. GitHub: github.com/angelachapman/ai-care-companion Slides: docs.google.com/presentation/d/1zrOrrBTq2Cd5...
Clinical Trial Accelerator
Переглядів 8821 день тому
The application reads a clinical trial protocol and creates an informed consent document using parallel agents. GitHub: github.com/mdean77a/clinical-trials/tree/main Slides: github.com/mdean77a/clinical-trials/blob/main/experimental/ClinicalTrialAccelerator.pdf Connect with the team! Mike: www.linkedin.com/in/mike-dean-301813227 Jeeva: www.linkedin.com/in/jeevalearn/ Apply for an upcoming cohor...
CareDash: Simplify Your Healthcare Journey
Переглядів 14021 день тому
CareDash: Simplify Your Healthcare Journey
Health Plan Chatbot using LangChain and Google Vertex AI
Переглядів 9721 день тому
Health Plan Chatbot using LangChain and Google Vertex AI
ChipPath - Uncover your future in Semiconductors
Переглядів 4821 день тому
ChipPath - Uncover your future in Semiconductors
AI Makerspace Transformation of the Week - Nitin Gupta #totw #genai #aim
Переглядів 10728 днів тому
AI Makerspace Transformation of the Week - Nitin Gupta #totw #genai #aim
AI Makerspace Transformation of the Week - Mert Bozkir #totw #genai #aim
Переглядів 101Місяць тому
AI Makerspace Transformation of the Week - Mert Bozkir #totw #genai #aim
AI Makerspace Transformation of the Week - Anna Tucker #totw #genai #aim
Переглядів 106Місяць тому
AI Makerspace Transformation of the Week - Anna Tucker #totw #genai #aim
AI Makerspace Transformation of the Week - Garret G. of Deepwriter AI #totw #genai #aim
Переглядів 101Місяць тому
AI Makerspace Transformation of the Week - Garret G. of Deepwriter AI #totw #genai #aim
LLM Engineering: The Foundations Cohort 3
Переглядів 354Місяць тому
LLM Engineering: The Foundations Cohort 3
AI Makerspace Transformation of the Week - Katerina Gawthorpe #totw #genai #aim
Переглядів 144Місяць тому
AI Makerspace Transformation of the Week - Katerina Gawthorpe #totw #genai #aim
AI Makerspace Transformation of the Week - Raul Salles de Padua #totw #genai #aim
Переглядів 1052 місяці тому
AI Makerspace Transformation of the Week - Raul Salles de Padua #totw #genai #aim
3N1 (Neural Nomad Nexus): Manager Sales Assistant
Переглядів 2043 місяці тому
3N1 (Neural Nomad Nexus): Manager Sales Assistant
ReportWiz - An Intelligent Business Reporting Assistant
Переглядів 2053 місяці тому
ReportWiz - An Intelligent Business Reporting Assistant
Healthcare Technology Management LLM (HTM-LLM)
Переглядів 1683 місяці тому
Healthcare Technology Management LLM (HTM-LLM)
AI Engineering Cohort 03 - Demo Day - Mental Mind Bot
Переглядів 1343 місяці тому
AI Engineering Cohort 03 - Demo Day - Mental Mind Bot
This was quite an amazing (and unexpected) GPU "deep dive" 🤯. Thanks for the care you put into all of what you share with us!
Glad you enjoyed it @tanguero2k7 … we did too! We’ll keep going to wherever exploration of the LLM Edge takes us!
Flash Attention - AIM Event: colab.research.google.com/drive/1-OPCDWnK3sQQncg6gm0bSMr9Nmx3lAil?usp=sharing Event Slides: www.canva.com/design/DAGXB3D4Yd0/ufpIRKB21NdBjzCwNYJOXA/view?DAGXB3D4Yd0&
Dr Greg and the Wiz, what a dream team! Thanks for your wonderful, truly wonderful efforts. I am certainly getting better following you guys. God bless.
AI engineering +1, but eventually will be interested in getting more into detail and understand LLM engineering better.
The most classic pattern we've seen from our community!
very useful! Thank you for your sharing
Dr Greg, you are absolutely incredible teacher. Keep it up!
Replit: replit.com/@replit/Anthropic-Computer-Use Event Slides: www.canva.com/design/DAGWX1fpgkA/AYTV-idbZfIoz-0uoosImQ/view?DAGWX1fpgkA&
hey did you called up the large action model live session?
It was a great video and RLHF was explained perfectly, thanks!
Thanks!
Thank you for boldly diving into these highly technical topics in plain English! Your videos are a gift to the community
Thank you for introducing our VPTQ project. VPTQ is an ongoing project and we welcome suggestions from everyone. We have currently open-sourced the inference code and algorithm code. : D
3. Inference Performance: Thank you both for helping me explain this :) Currently, our inference code is just a very naive inference example based on Torch. We haven't done any optimizations specifically for inference yet. I am currently trying to integrate VPTQ into vLLM. I believe this will greatly enhance the inference speed with vLLM. I am currently discussing this in the vLLM Slack channel and with the sig-quant group. Stay tuned!
2. Entropy: I am very interested in using Shannon entropy and information theory to quantitatively analyze the impact of model quantization. In fact, an unresolved academic issue in the quantization field is how to quantitatively describe the effects of quantization on the final loss, or even model accuracy, especially since there are many nonlinear layers involved. The GPTQ/VPTQ series of algorithms, which are based on second-order optimization, actually have very strong assumptions (as analyzed in the paper WoodFisher). These assumptions simplify the impact of quantization on model loss to the impact on proxy error, thus simplifying the algorithm. I believe this is a more fundamental problem. Currently, I am still conceptualizing and do not yet have a clear idea on how to perform quantitative analysis.
I'll try to address the last few questions from the video: 1. The results of VPTQ on smaller models. I have actually tried compressing smaller models like Qwen2.5 1B-3B, and achieved good results around 3-4 bits. I will release some of these results later. Indeed, on smaller models, compressing the model weights is quite challenging. I am currently working on improvements and will continue to provide updates. Thank you.
VPTQ - Vector Post-Training Quanization: colab.research.google.com/drive/1yItAepbYh9HVs3SEBqtZclc5AY4aJIAa?usp=sharing Event Slides: www.canva.com/design/DAGVuEs7C_s/cohyWXMzy7TD_MzR8-MA_g/view?DAGVuEs7C_s&
Hi Greg and Wiz. Great tutorial. I am actually applying it to my own application. I was wondering what would you sugges to do if the whole document size is large more than 700 pages. It want be passed in the contextual chunking function. If I take the 500 pages around the chhunk, the chaching wont work. Please can you advice? Thanks Saurabh
I would build some metadata, in this case, like a summary/outline and use that to generate contextual augments.
I'm super happy for this video supernova was on my radar last month.
I am trying to understand the unique concepts of this paper. It sounds like this is a workflow of agents and programmatic validators to synthetically generate DPO data. Is the system self learning as well?
It can be online, yes.
Hey this is quite useful.can you help me in how large action model works? The recent Claude computer or ominiparser and lavauge model or how the rabbit mq works.can you help with the codes refference or implementation.thank you
We're planning to cover computer use in an upcoming event soon - probably nov 13. Stay tuned!
Great information here! Thanks for making it public. I think you're going to get a sizeable community around you because of these live streams. Q: where in the code is prompt caching evoked?
Caching is offered by Anthropic's endpoint by default - and is being taken advantage of under the hood here.
Calibrated reward: github.com/huggingface/trl/pull/2155/files Mixture of judges: github.com/huggingface/trl/pull/2159/files CGPO Trainer (single task single objective): github.com/huggingface/trl/pull/2190/files Event Slides: www.canva.com/design/DAGVDvGDG54/kEflcFEuGxDKMTYb6Rj2vA/view?DAGVDvGDG54&
at 00:24:15 you give a formula for faithfulness, think it is flawed a bit. Should be (#Claims from the answer which exist in the context) / (#claims in answer). Otherwise there could be >1 result.
Can you be more specific about what the flaw is? Also, why do you choose the word "exist" rather than "inferred from?" --Here's what appears to be true from the documentation: -- "To calculate this a set of claims from the generated answer is first identified. Then each one of these claims are cross checked with given context to determine if it can be inferred from given context or not." Three steps to the calculation: 1. Break generated answer into statements 2. For each statement, verify if it can be inferred 3. Calculate Faithfulness! It seems that the condition "if (and only if) it can be inferred from the context" will keep the faithfulness calculation from going higher than 1.0
Great work here Richard and Gil - loved the demo
Love this guys. Great job!
hello!! thank you a lot for the videos! what is the best way to interact in sort of chat engine of chat loop with a workflow?
Can you expand on your request?
@@AI-Makerspace thank you for answering! I'm curious about the best practices for building a chat engine or chatbot that can interact in a continuous loop with a workflow. Currently, we are receiving one response at a time from the workflow, but I was wondering if we could enhance this by buffering the "chatmemory" and keep on with the conversation. Should this be achieved with a loop? I feel like I remember a llamaIndex or Langchain tool that kept the chat engine running, but I might be mistaken, maybe I was just re-querying. Also, how can I ensure other workflows share the same context? Additionally, is it possible to store interactions as vectorized semantic and episodic memories in a vector database, allowing the system to recall past conversations and in the future query from those memories and the RAG? and maybe do some type of reranking.
from llama_index import SimpleDirectoryReader, VectorStoreIndex from colorama import Fore, Style, init init(autoreset=True) def chat(): print(f"{Fore.CYAN}Loading documents...") index = VectorStoreIndex.from_documents( SimpleDirectoryReader("./data").load_data() ) chat_engine = index.as_chat_engine() print(f"{Fore.GREEN}Ready! Type 'quit' to exit ") while True: query = input(f"{Fore.GREEN}You: {Style.RESET_ALL}").strip() if query.lower() == 'quit': break if query: print(f"{Fore.BLUE}Assistant: {Style.RESET_ALL}{chat_engine.chat(query)} ") if __name__ == "__main__": try: chat() except Exception as e: print(f"{Fore.RED}Error: {e}")
Great video, but using text that the model was already trained on is a bad test case
Agreed! We typically stick with easy to consume toy-examples, however!
Hi! I really appreciated your video. BTW, I wrote an article titled "SWARMing Conversational AI: Integrating No-Code and Code in Agent-Based Workflows," which you can find online. I would love to hear your feedback on my perspective (SWARM emphasis on blending no-code instructions with hardcoded conversational steps. Thanks! Giorgio
Definitely will check it out!
OpenAI Swarm - Multi-Agent: colab.research.google.com/drive/1NumpfFNIPxsyjmruJ3jzyxxX2HY8V0MO?usp=sharing Event Slides: www.canva.com/design/DAGUZ0A-Zpc/uctbkE6-rHzlRfjxVFPAlg/view?DAGUZ0A-Zpc&
Are the slides available?
Sure thing, just pinned the slides and notebook in a comment!
Congratulations dude
Great video! I’m excited to dive into contextual retrieval next week. When it comes to productionizing hybrid retrieval with BM25, I’m considering using Elasticsearch, any other recommendations? My main concern with hybrid retrieval is the added complexity it brings to the production.
Elasticsearch is a great tool for this!
🚨MERT ALERT
Great video and demo as always! I learn much from your content. The contextual retrieval paper said if your corpus is less than 200k tokens, just skip rag and dump the entire corpus into the prompt for every question, and they will cache it (but only for a short time) and just use long context Q&A. I didn’t see them publish any metrics comparing long context to rag, so I take it with a grain of salt. They do want customers to spend as many tokens as possible... But I’m very intrigued at the same time. Maybe you could do a video comparing the two methods? That would be amazing research.
Great insights and instincts @Sean! We'll keep the content recommendation in mind for sure! This is farthest we've gotten on Long-Context and Evaluation for the big window LLMs: ua-cam.com/users/liveBrwhbjh3boU?si=V24z6pagQ0EQ8Ms1
thanks :)
Would the results be even better when combined with semantic chunking? Answer: research.trychroma.com/evaluating-chunking
RAG-ception 0:55 - Context of the contextually generated chunks. Got it...got..it.......got it....ok wait what? Need to watch the whole thing.
Re; Would the results be even better when combined with semantic chunking? For more on Semantic Chunking strategies: research.trychroma.com/evaluating-chunking
Contextual Retrieval: colab.research.google.com/drive/1KGVxiwc2zoY9v6f3IQfs8qJIZeGeMKAq?usp=sharing Event Slides: www.canva.com/design/DAGTv5ofV8g/-wFTZpoCu8yYzseTb_kx2g/view?DAGTv5ofV8g&
The Ragas part of code in the notebook is not working. Could you fix it?
Very interesting
Where this complete video, id like to understand this loss fuction and the matrix hessian
Hey Givanildo! The full event is here: ua-cam.com/users/livexmaG4al2A6E?si=bdHM0wzlll5XkXWJ To learn more about loss functions, check out this one! ua-cam.com/users/liveiB8FWR9aD5Q?si=4oABKIf-DDNQQv1R
Top
GPTQ - AIMS: colab.research.google.com/drive/1iZQ_Byo9F-bM6IGywtb3sN9TIEcE_Mqd?usp=sharing Event Slides: www.canva.com/design/DAGTF8vsIIM/9qbXt5T-pvt-KIHDtC_4nA/view?DAGTF8vsIIM&
Thanks for the great video. Subscribed! Question - We saw here that "similar pairs" were trained where the pair implies a (question , context). Is it possible to get good results by fine-tuning on a "similar questions" dataset i.e. (question1, question2) and the difference between those 2 questions is usually one word/phrase. So question1 would have the full-form of an entity; and question2 the acronym of the same entity. Reason I'm doing this is that I'm storing a mix of questions and contexts in my vector Database. If the user's query matches a question then I look up the corresponding answer (static answer that almost never changes - so no LLM required). If the match is a context instead, then LLM generation takes over.
Yes, that is a decent way to approach that problem.
How do I get a scholarship?
We don't currently have scholarships available @nazmuss! We are working to get our business model right and to grow our partnerships in the US so we can best serve our community members around the world in the long-term moving forward! In short, stay tuned!
Where does ground truth come from? Is this a human annotated property? I understand the ground truth in RAGAS refers to the correct answer to the question. It's typically used for the context_recall metric. But how to we get this? Human in the loop? LLM generated? More documents from the retrieval? Thank you?
"Ground Truth" can come from any of these sources! Of course, getting it straight from the people who perform whatever tasks you're automating is the right idea, but this can be very expensive. In the case of RAGAS the "Ground Truth" is represented by the output you get when you provide [question, retrieved context] pairs as input to a generator. That is, we are not actually using a RAG system, but passing "correct" [question, context] pairs as input. These are "correct" because they were synthetically generated and are known to be correct; see Synthetic Test Data Generation: docs.ragas.io/en/stable/concepts/testset_generation.html Note that Ground Truth is different than "Answer" because "Answer" actually uses the RAG application that you're building, while "Ground Truth" passes [question, context] pairs in direclty.
LOVE LOVE LOVE the Snatch Dags meme
lovely
Thanks Andres!
Bros DSE awareness is 0