AI Makerspace
AI Makerspace
  • 253
  • 360 492
AI Makerspace Transformation of the Week - Tshwanelo Matsane #totw #genai #aim
In this Transformation of the Week, I chat with Tshwanelo Matsane, a strategy consultant and recent graduate from the AI Engineering Bootcamp, who wanted to build his own applications, and be his company's AI thought-leader.
Learn bleeding-edge Gen AI concepts, best practices, and tools with the pros at AI Makerspace. Come Build, Ship, and Share with the best AI practitioners in the world. Upskill in less than 10 weeks with the highest rated bootcamp on Maven.
aimakerspace.io
#genai #learnai #totw #buildshipshare #aim
Переглядів: 26

Відео

LLM Engineering or AI Engineering - which are you more interested in?
Переглядів 11916 годин тому
LLM Engineering or AI Engineering - which are you more interested in? We heard from so many beginners in our community asking, “so what’s the difference?” Wonder no more! Dr. Greg of AI Makerspace is breaks it down in simple terms to give you a quick explanation. #llme #genai #aie Learn bleeding-edge Gen AI concepts, best practices, and tools with the pros at AI Makerspace. Come Build, Ship, an...
AI Makerspace Transformation of the Week - Laura Funderburk #totw #genai #aim
Переглядів 6214 днів тому
In this Transformation of the Week, I chat with Laura Funderburk of Bytewax, a graduate from the AI Makerspace LLM Operations cohort who started off in the call center, and is now an AI Engineer writing a book on Gen AI! Learn bleeding-edge Gen AI concepts, best practices, and tools with the pros at AI Makerspace. Come Build, Ship, and Share with the best AI practitioners in the world. Upskill ...
AI Makerspace Transformation of the Week - Pano Evangeliou #totw #genai #aim
Переглядів 11921 день тому
In this Transformation of the Week, I chat with Pano Evangeliou, a graduate from the AI Makerspace LLM Operations cohort who didn’t let a language barrier or time zones, stand in his way of Building, Shipping, and Sharing his way to becoming a Senior AI Engineer! Learn bleeding-edge Gen AI concepts, best practices, and tools with the pros at AI Makerspace. Come Build, Ship, and Share with the b...
LLM Engineering Challenge Walkthrough with The Wiz
Переглядів 40821 день тому
If you're looking to get started learning LLM Engineering, you're in the right place! The Wiz is here to walk you through, step-by-step, exactly how to complete the challenge. If you're a beginner, this is exactly how the homeworks will feel for you! If you're more advanced, we're going to make sure that the homeworks feel quite a bit more challenging! The LLM Engineering Challenge is designed ...
CareCompanion: AI Support for Dementia Caregivers
Переглядів 13321 день тому
This AI-powered chatbot provides practical support and resources to caregivers of individuals with chronic illnesses like dementia. By answering common questions, providing emotional support, and connecting caregivers to local resources, the chatbot helps caregivers navigate daily challenges. GitHub: github.com/angelachapman/ai-care-companion Slides: docs.google.com/presentation/d/1zrOrrBTq2Cd5...
Clinical Trial Accelerator
Переглядів 8821 день тому
The application reads a clinical trial protocol and creates an informed consent document using parallel agents. GitHub: github.com/mdean77a/clinical-trials/tree/main Slides: github.com/mdean77a/clinical-trials/blob/main/experimental/ClinicalTrialAccelerator.pdf Connect with the team! Mike: www.linkedin.com/in/mike-dean-301813227 Jeeva: www.linkedin.com/in/jeevalearn/ Apply for an upcoming cohor...
PropSentinel
Переглядів 8221 день тому
PropSentinel
CareDash: Simplify Your Healthcare Journey
Переглядів 14021 день тому
CareDash: Simplify Your Healthcare Journey
LitPilot
Переглядів 11021 день тому
LitPilot
Marketing Content Enhancer
Переглядів 6021 день тому
Marketing Content Enhancer
Health Plan Chatbot using LangChain and Google Vertex AI
Переглядів 9721 день тому
Health Plan Chatbot using LangChain and Google Vertex AI
Agentic Real Estate Humberto Gerardo
Переглядів 3221 день тому
Agentic Real Estate Humberto Gerardo
Smart Lease Navigator
Переглядів 3221 день тому
Smart Lease Navigator
Publicus
Переглядів 7421 день тому
Publicus
ChipPath - Uncover your future in Semiconductors
Переглядів 4821 день тому
ChipPath - Uncover your future in Semiconductors
AI Makerspace Transformation of the Week - Nitin Gupta #totw #genai #aim
Переглядів 10728 днів тому
AI Makerspace Transformation of the Week - Nitin Gupta #totw #genai #aim
AI Makerspace Transformation of the Week - Mert Bozkir #totw #genai #aim
Переглядів 101Місяць тому
AI Makerspace Transformation of the Week - Mert Bozkir #totw #genai #aim
AI Makerspace Transformation of the Week - Anna Tucker #totw #genai #aim
Переглядів 106Місяць тому
AI Makerspace Transformation of the Week - Anna Tucker #totw #genai #aim
Take the LLME 03 Challenge today!
Переглядів 248Місяць тому
Take the LLME 03 Challenge today!
AI Makerspace Transformation of the Week - Garret G. of Deepwriter AI #totw #genai #aim
Переглядів 101Місяць тому
AI Makerspace Transformation of the Week - Garret G. of Deepwriter AI #totw #genai #aim
LLM Engineering: The Foundations Cohort 3
Переглядів 354Місяць тому
LLM Engineering: The Foundations Cohort 3
AI Makerspace Transformation of the Week - Katerina Gawthorpe #totw #genai #aim
Переглядів 144Місяць тому
AI Makerspace Transformation of the Week - Katerina Gawthorpe #totw #genai #aim
AI Makerspace Transformation of the Week - Raul Salles de Padua #totw #genai #aim
Переглядів 1052 місяці тому
AI Makerspace Transformation of the Week - Raul Salles de Padua #totw #genai #aim
Drug Alert AI
Переглядів 1993 місяці тому
Drug Alert AI
The AI Engineering Bootcamp Cohort 4
Переглядів 4603 місяці тому
The AI Engineering Bootcamp Cohort 4
3N1 (Neural Nomad Nexus): Manager Sales Assistant
Переглядів 2043 місяці тому
3N1 (Neural Nomad Nexus): Manager Sales Assistant
ReportWiz - An Intelligent Business Reporting Assistant
Переглядів 2053 місяці тому
ReportWiz - An Intelligent Business Reporting Assistant
Healthcare Technology Management LLM (HTM-LLM)
Переглядів 1683 місяці тому
Healthcare Technology Management LLM (HTM-LLM)
AI Engineering Cohort 03 - Demo Day - Mental Mind Bot
Переглядів 1343 місяці тому
AI Engineering Cohort 03 - Demo Day - Mental Mind Bot

КОМЕНТАРІ

  • @tanguero2k7
    @tanguero2k7 День тому

    This was quite an amazing (and unexpected) GPU "deep dive" 🤯. Thanks for the care you put into all of what you share with us!

    • @AI-Makerspace
      @AI-Makerspace День тому

      Glad you enjoyed it @tanguero2k7 … we did too! We’ll keep going to wherever exploration of the LLM Edge takes us!

  • @AI-Makerspace
    @AI-Makerspace 2 дні тому

    Flash Attention - AIM Event: colab.research.google.com/drive/1-OPCDWnK3sQQncg6gm0bSMr9Nmx3lAil?usp=sharing Event Slides: www.canva.com/design/DAGXB3D4Yd0/ufpIRKB21NdBjzCwNYJOXA/view?DAGXB3D4Yd0&

  • @mmasa1
    @mmasa1 2 дні тому

    Dr Greg and the Wiz, what a dream team! Thanks for your wonderful, truly wonderful efforts. I am certainly getting better following you guys. God bless.

  • @MrMoonsilver
    @MrMoonsilver 7 днів тому

    AI engineering +1, but eventually will be interested in getting more into detail and understand LLM engineering better.

    • @AI-Makerspace
      @AI-Makerspace 7 днів тому

      The most classic pattern we've seen from our community!

  • @thuhuongnguyen6543
    @thuhuongnguyen6543 7 днів тому

    very useful! Thank you for your sharing

  • @mentorships3309
    @mentorships3309 9 днів тому

    Dr Greg, you are absolutely incredible teacher. Keep it up!

  • @AI-Makerspace
    @AI-Makerspace 9 днів тому

    Replit: replit.com/@replit/Anthropic-Computer-Use Event Slides: www.canva.com/design/DAGWX1fpgkA/AYTV-idbZfIoz-0uoosImQ/view?DAGWX1fpgkA&

  • @rakeshkumarrout2629
    @rakeshkumarrout2629 9 днів тому

    hey did you called up the large action model live session?

  • @fredflintstone7924
    @fredflintstone7924 13 днів тому

    It was a great video and RLHF was explained perfectly, thanks!

  • @lucindalinde4198
    @lucindalinde4198 13 днів тому

    Thank you for boldly diving into these highly technical topics in plain English! Your videos are a gift to the community

  • @yangwang8232
    @yangwang8232 16 днів тому

    Thank you for introducing our VPTQ project. VPTQ is an ongoing project and we welcome suggestions from everyone. We have currently open-sourced the inference code and algorithm code. : D

  • @yangwang8232
    @yangwang8232 16 днів тому

    3. Inference Performance: Thank you both for helping me explain this :) Currently, our inference code is just a very naive inference example based on Torch. We haven't done any optimizations specifically for inference yet. I am currently trying to integrate VPTQ into vLLM. I believe this will greatly enhance the inference speed with vLLM. I am currently discussing this in the vLLM Slack channel and with the sig-quant group. Stay tuned!

  • @yangwang8232
    @yangwang8232 16 днів тому

    2. Entropy: I am very interested in using Shannon entropy and information theory to quantitatively analyze the impact of model quantization. In fact, an unresolved academic issue in the quantization field is how to quantitatively describe the effects of quantization on the final loss, or even model accuracy, especially since there are many nonlinear layers involved. The GPTQ/VPTQ series of algorithms, which are based on second-order optimization, actually have very strong assumptions (as analyzed in the paper WoodFisher). These assumptions simplify the impact of quantization on model loss to the impact on proxy error, thus simplifying the algorithm. I believe this is a more fundamental problem. Currently, I am still conceptualizing and do not yet have a clear idea on how to perform quantitative analysis.

  • @yangwang8232
    @yangwang8232 16 днів тому

    I'll try to address the last few questions from the video: 1. The results of VPTQ on smaller models. I have actually tried compressing smaller models like Qwen2.5 1B-3B, and achieved good results around 3-4 bits. I will release some of these results later. Indeed, on smaller models, compressing the model weights is quite challenging. I am currently working on improvements and will continue to provide updates. Thank you.

  • @AI-Makerspace
    @AI-Makerspace 16 днів тому

    VPTQ - Vector Post-Training Quanization: colab.research.google.com/drive/1yItAepbYh9HVs3SEBqtZclc5AY4aJIAa?usp=sharing Event Slides: www.canva.com/design/DAGVuEs7C_s/cohyWXMzy7TD_MzR8-MA_g/view?DAGVuEs7C_s&

  • @aryansaurabhbhardwaj
    @aryansaurabhbhardwaj 18 днів тому

    Hi Greg and Wiz. Great tutorial. I am actually applying it to my own application. I was wondering what would you sugges to do if the whole document size is large more than 700 pages. It want be passed in the contextual chunking function. If I take the 500 pages around the chhunk, the chaching wont work. Please can you advice? Thanks Saurabh

    • @AI-Makerspace
      @AI-Makerspace 18 днів тому

      I would build some metadata, in this case, like a summary/outline and use that to generate contextual augments.

  • @PhiliBuster-i7t
    @PhiliBuster-i7t 19 днів тому

    I'm super happy for this video supernova was on my radar last month.

  • @MegaClockworkDoc
    @MegaClockworkDoc 22 дні тому

    I am trying to understand the unique concepts of this paper. It sounds like this is a workflow of agents and programmatic validators to synthetically generate DPO data. Is the system self learning as well?

  • @rakeshkumarrout2629
    @rakeshkumarrout2629 22 дні тому

    Hey this is quite useful.can you help me in how large action model works? The recent Claude computer or ominiparser and lavauge model or how the rabbit mq works.can you help with the codes refference or implementation.thank you

    • @AI-Makerspace
      @AI-Makerspace 22 дні тому

      We're planning to cover computer use in an upcoming event soon - probably nov 13. Stay tuned!

  • @hoopNscoops
    @hoopNscoops 22 дні тому

    Great information here! Thanks for making it public. I think you're going to get a sizeable community around you because of these live streams. Q: where in the code is prompt caching evoked?

    • @AI-Makerspace
      @AI-Makerspace 18 днів тому

      Caching is offered by Anthropic's endpoint by default - and is being taken advantage of under the hood here.

  • @AI-Makerspace
    @AI-Makerspace 23 дні тому

    Calibrated reward: github.com/huggingface/trl/pull/2155/files Mixture of judges: github.com/huggingface/trl/pull/2159/files CGPO Trainer (single task single objective): github.com/huggingface/trl/pull/2190/files Event Slides: www.canva.com/design/DAGVDvGDG54/kEflcFEuGxDKMTYb6Rj2vA/view?DAGVDvGDG54&

  • @MarkDavisRocks
    @MarkDavisRocks 23 дні тому

    at 00:24:15 you give a formula for faithfulness, think it is flawed a bit. Should be (#Claims from the answer which exist in the context) / (#claims in answer). Otherwise there could be >1 result.

    • @AI-Makerspace
      @AI-Makerspace 21 день тому

      Can you be more specific about what the flaw is? Also, why do you choose the word "exist" rather than "inferred from?" --Here's what appears to be true from the documentation: -- "To calculate this a set of claims from the generated answer is first identified. Then each one of these claims are cross checked with given context to determine if it can be inferred from given context or not." Three steps to the calculation: 1. Break generated answer into statements 2. For each statement, verify if it can be inferred 3. Calculate Faithfulness! It seems that the condition "if (and only if) it can be inferred from the context" will keep the faithfulness calculation from going higher than 1.0

  • @niting1978
    @niting1978 24 дні тому

    Great work here Richard and Gil - loved the demo

  • @richardgower4890
    @richardgower4890 24 дні тому

    Love this guys. Great job!

  • @johnini
    @johnini 29 днів тому

    hello!! thank you a lot for the videos! what is the best way to interact in sort of chat engine of chat loop with a workflow?

    • @AI-Makerspace
      @AI-Makerspace 29 днів тому

      Can you expand on your request?

    • @johnini
      @johnini 29 днів тому

      @@AI-Makerspace thank you for answering! I'm curious about the best practices for building a chat engine or chatbot that can interact in a continuous loop with a workflow. Currently, we are receiving one response at a time from the workflow, but I was wondering if we could enhance this by buffering the "chatmemory" and keep on with the conversation. Should this be achieved with a loop? I feel like I remember a llamaIndex or Langchain tool that kept the chat engine running, but I might be mistaken, maybe I was just re-querying. Also, how can I ensure other workflows share the same context? Additionally, is it possible to store interactions as vectorized semantic and episodic memories in a vector database, allowing the system to recall past conversations and in the future query from those memories and the RAG? and maybe do some type of reranking.

    • @johnini
      @johnini 29 днів тому

      from llama_index import SimpleDirectoryReader, VectorStoreIndex from colorama import Fore, Style, init init(autoreset=True) def chat(): print(f"{Fore.CYAN}Loading documents...") index = VectorStoreIndex.from_documents( SimpleDirectoryReader("./data").load_data() ) chat_engine = index.as_chat_engine() print(f"{Fore.GREEN}Ready! Type 'quit' to exit ") while True: query = input(f"{Fore.GREEN}You: {Style.RESET_ALL}").strip() if query.lower() == 'quit': break if query: print(f"{Fore.BLUE}Assistant: {Style.RESET_ALL}{chat_engine.chat(query)} ") if __name__ == "__main__": try: chat() except Exception as e: print(f"{Fore.RED}Error: {e}")

  • @MegaClockworkDoc
    @MegaClockworkDoc 29 днів тому

    Great video, but using text that the model was already trained on is a bad test case

    • @AI-Makerspace
      @AI-Makerspace 29 днів тому

      Agreed! We typically stick with easy to consume toy-examples, however!

  • @solyarisoftware
    @solyarisoftware 29 днів тому

    Hi! I really appreciated your video. BTW, I wrote an article titled "SWARMing Conversational AI: Integrating No-Code and Code in Agent-Based Workflows," which you can find online. I would love to hear your feedback on my perspective (SWARM emphasis on blending no-code instructions with hardcoded conversational steps. Thanks! Giorgio

  • @AI-Makerspace
    @AI-Makerspace Місяць тому

    OpenAI Swarm - Multi-Agent: colab.research.google.com/drive/1NumpfFNIPxsyjmruJ3jzyxxX2HY8V0MO?usp=sharing Event Slides: www.canva.com/design/DAGUZ0A-Zpc/uctbkE6-rHzlRfjxVFPAlg/view?DAGUZ0A-Zpc&

  • @scitechtalktv9742
    @scitechtalktv9742 Місяць тому

    Are the slides available?

    • @AI-Makerspace
      @AI-Makerspace Місяць тому

      Sure thing, just pinned the slides and notebook in a comment!

  • @enespacalar
    @enespacalar Місяць тому

    Congratulations dude

  • @danielusvyat
    @danielusvyat Місяць тому

    Great video! I’m excited to dive into contextual retrieval next week. When it comes to productionizing hybrid retrieval with BM25, I’m considering using Elasticsearch, any other recommendations? My main concern with hybrid retrieval is the added complexity it brings to the production.

    • @AI-Makerspace
      @AI-Makerspace Місяць тому

      Elasticsearch is a great tool for this!

  • @cmagganas
    @cmagganas Місяць тому

    🚨MERT ALERT

  • @seanbergman8927
    @seanbergman8927 Місяць тому

    Great video and demo as always! I learn much from your content. The contextual retrieval paper said if your corpus is less than 200k tokens, just skip rag and dump the entire corpus into the prompt for every question, and they will cache it (but only for a short time) and just use long context Q&A. I didn’t see them publish any metrics comparing long context to rag, so I take it with a grain of salt. They do want customers to spend as many tokens as possible... But I’m very intrigued at the same time. Maybe you could do a video comparing the two methods? That would be amazing research.

    • @AI-Makerspace
      @AI-Makerspace Місяць тому

      Great insights and instincts @Sean! We'll keep the content recommendation in mind for sure! This is farthest we've gotten on Long-Context and Evaluation for the big window LLMs: ua-cam.com/users/liveBrwhbjh3boU?si=V24z6pagQ0EQ8Ms1

  • @micbab-vg2mu
    @micbab-vg2mu Місяць тому

    thanks :)

  • @AI-Makerspace
    @AI-Makerspace Місяць тому

    Would the results be even better when combined with semantic chunking? Answer: research.trychroma.com/evaluating-chunking

  • @AmanBansil
    @AmanBansil Місяць тому

    RAG-ception 0:55 - Context of the contextually generated chunks. Got it...got..it.......got it....ok wait what? Need to watch the whole thing.

  • @AI-Makerspace
    @AI-Makerspace Місяць тому

    Re; Would the results be even better when combined with semantic chunking? For more on Semantic Chunking strategies: research.trychroma.com/evaluating-chunking

  • @AI-Makerspace
    @AI-Makerspace Місяць тому

    Contextual Retrieval: colab.research.google.com/drive/1KGVxiwc2zoY9v6f3IQfs8qJIZeGeMKAq?usp=sharing Event Slides: www.canva.com/design/DAGTv5ofV8g/-wFTZpoCu8yYzseTb_kx2g/view?DAGTv5ofV8g&

    • @weiwennie40
      @weiwennie40 9 днів тому

      The Ragas part of code in the notebook is not working. Could you fix it?

  • @givanildogramacho441
    @givanildogramacho441 Місяць тому

    Very interesting

  • @givanildogramacho441
    @givanildogramacho441 Місяць тому

    Where this complete video, id like to understand this loss fuction and the matrix hessian

    • @AI-Makerspace
      @AI-Makerspace Місяць тому

      Hey Givanildo! The full event is here: ua-cam.com/users/livexmaG4al2A6E?si=bdHM0wzlll5XkXWJ To learn more about loss functions, check out this one! ua-cam.com/users/liveiB8FWR9aD5Q?si=4oABKIf-DDNQQv1R

  • @andresrubio2015
    @andresrubio2015 Місяць тому

    Top

  • @AI-Makerspace
    @AI-Makerspace Місяць тому

    GPTQ - AIMS: colab.research.google.com/drive/1iZQ_Byo9F-bM6IGywtb3sN9TIEcE_Mqd?usp=sharing Event Slides: www.canva.com/design/DAGTF8vsIIM/9qbXt5T-pvt-KIHDtC_4nA/view?DAGTF8vsIIM&

  • @saurabhahlawat425
    @saurabhahlawat425 Місяць тому

    Thanks for the great video. Subscribed! Question - We saw here that "similar pairs" were trained where the pair implies a (question , context). Is it possible to get good results by fine-tuning on a "similar questions" dataset i.e. (question1, question2) and the difference between those 2 questions is usually one word/phrase. So question1 would have the full-form of an entity; and question2 the acronym of the same entity. Reason I'm doing this is that I'm storing a mix of questions and contexts in my vector Database. If the user's query matches a question then I look up the corresponding answer (static answer that almost never changes - so no LLM required). If the match is a context instead, then LLM generation takes over.

    • @AI-Makerspace
      @AI-Makerspace Місяць тому

      Yes, that is a decent way to approach that problem.

  • @nazmusas
    @nazmusas Місяць тому

    How do I get a scholarship?

    • @AI-Makerspace
      @AI-Makerspace Місяць тому

      We don't currently have scholarships available @nazmuss! We are working to get our business model right and to grow our partnerships in the US so we can best serve our community members around the world in the long-term moving forward! In short, stay tuned!

  • @yerson557
    @yerson557 Місяць тому

    Where does ground truth come from? Is this a human annotated property? I understand the ground truth in RAGAS refers to the correct answer to the question. It's typically used for the context_recall metric. But how to we get this? Human in the loop? LLM generated? More documents from the retrieval? Thank you?

    • @AI-Makerspace
      @AI-Makerspace Місяць тому

      "Ground Truth" can come from any of these sources! Of course, getting it straight from the people who perform whatever tasks you're automating is the right idea, but this can be very expensive. In the case of RAGAS the "Ground Truth" is represented by the output you get when you provide [question, retrieved context] pairs as input to a generator. That is, we are not actually using a RAG system, but passing "correct" [question, context] pairs as input. These are "correct" because they were synthetically generated and are known to be correct; see Synthetic Test Data Generation: docs.ragas.io/en/stable/concepts/testset_generation.html Note that Ground Truth is different than "Answer" because "Answer" actually uses the RAG application that you're building, while "Ground Truth" passes [question, context] pairs in direclty.

  • @cmagganas
    @cmagganas Місяць тому

    LOVE LOVE LOVE the Snatch Dags meme

  • @andres.yodars
    @andres.yodars Місяць тому

    lovely

  • @Warbaman
    @Warbaman Місяць тому

    Bros DSE awareness is 0