37% Better Output with 15 Lines of Code - Llama 3 8B (Ollama) & 70B (Groq)

Поділитися
Вставка
  • Опубліковано 12 лип 2024
  • To try everything Brilliant has to offer-free-for a full 30 days, visit brilliant.org/AllAboutAI . You’ll also get 20% off an annual premium subscription.
    37% Better Output with 15 Lines of Code - Llama 3 8B (Ollama) & 70B (Groq)
    GitHub Project:
    github.com/AllAboutAI-YT/easy...
    👊 Become a member and get access to GitHub and Code:
    / allaboutai
    🤖 Great AI Engineer Course:
    scrimba.com/learn/aiengineer?...
    📧 Join the newsletter:
    www.allabtai.com/newsletter/
    🌐 My website:
    www.allabtai.com
    In this video I try to improve a known problem when using RAG in local model like Llama 3 8B on ollama. This local RAG system was improved by just adding around 15 lines of code. Feel free to share and rate on GitHub :)
    00:00 Llama 3 Improved RAG Intro
    02:01 Problem / Soulution
    03:05 Brilliant.org
    04:26 How this works
    12:05 Llama 3 70B Groq
    15:12 Conclusion
  • Наука та технологія

КОМЕНТАРІ • 32

  • @AllAboutAI
    @AllAboutAI  2 місяці тому +3

    Brilliant: To try everything Brilliant has to offer-free-for a full 30 days, visit brilliant.org/AllAboutAI . You’ll also get 20% off an annual premium subscription.

  • @MattJonesYT
    @MattJonesYT 2 місяці тому +8

    Another approach to this is to just ask for the simple llm to hallucinate an answer to the current chat. That answer will not be correct but it will probably have the phrases needed for the RAG system to find the needed excerpts. There's a technical term for this idea which I can't remember but I came across it on the TwoSetAI channel which has a lot of similar tricks

    • @robboerman9378
      @robboerman9378 2 місяці тому +3

      HyDE, Hypothetical Document Embeddings. Works very well and easy to implement. Similarity search on a vector database using a hallucinated answer to the question instead of the question usually gives better similarity

    • @AllAboutAI
      @AllAboutAI  2 місяці тому +1

      yes this is nice, thnx :)

    • @kenhtinhthuc
      @kenhtinhthuc 2 місяці тому +2

      RAG is a bit too much of an exact match because it is based on concepts and similar concepts. Therefore no match, no return. HyDE makes the search a bit more fuzzy by expanding the query and introducing more concepts. It would be good to have an evaluator to check on the faithfuness of retrieval and the relevance of the ouputs to the original query.

  • @ASchnacky
    @ASchnacky 2 місяці тому +5

    Dolphin-llama3 & Groq-llama3
    are awesome! Well done!

    • @ByZaMo64
      @ByZaMo64 2 місяці тому

      how are they different?

  • @MarcShade
    @MarcShade 2 місяці тому +5

    dolphin-llama3:8b-v2.9-fp16 is so good as an assistant!

    • @ASchnacky
      @ASchnacky 2 місяці тому +1

      Dolphin-llama3 & Groq-llama3

  • @futureworldhealing
    @futureworldhealing 2 місяці тому +2

    best AI python coding channel hands down

  • @pec8377
    @pec8377 2 місяці тому +1

    @AllAboutAi the issue is it makes the assumption that the question is related to the content passed, which is not always the case in a conversation. Like suddenly you talk about something else, let's say "How are you", it will be rewritten to be aligned to the precedent context, which is not what you want.. then you need to implement some more mechanism or tweak your prompt to only rephrase when the question seems to be linked to the past. Many discussions about this..

  • @nic-ori
    @nic-ori 2 місяці тому +3

    👍👍👍Thanks! Useful information.

  • @Edoras5916
    @Edoras5916 2 місяці тому

    direct, didactic almost verbatim in my book, explanation. excellent

  • @technolus5742
    @technolus5742 2 місяці тому +1

    Great job

  • @akimezra7178
    @akimezra7178 2 місяці тому

    Bruuuuuuh, just found this channel, you sure you're human?!?! Wish i had 5% of your brain.... thank you so much for your work! Im learning so much!!

    • @SeattleShelby
      @SeattleShelby Місяць тому

      You just need a bigger neck beard. It’s all in the neck beard.

  • @realorfake4765
    @realorfake4765 2 місяці тому +1

    based on your experience, why is olama better than LMStudio?

  • @samyio4256
    @samyio4256 Місяць тому

    How is the retrieval so fast? Did you cut the loading time for context out of the video?

  • @vetonrushiti19
    @vetonrushiti19 4 дні тому

    can you please give the code of 70b model?

  • @elsondasilva8636
    @elsondasilva8636 2 місяці тому +1

    💎💎🌟💎💎💎💎

  • @iamisobe
    @iamisobe 2 місяці тому +1

    first

  • @monstercameron
    @monstercameron 2 місяці тому +1

    What about doing the same for the output? One pass is the internal voice, compare it to the promo to see if matches up and a second pass for any corections. Like giving LLMs an inner voice like we do.

  • @buttpub
    @buttpub 2 місяці тому +1

    the problem and solution is that your setup is stateless

    • @AllAboutAI
      @AllAboutAI  2 місяці тому

      interesting, will look into

    • @buttpub
      @buttpub 2 місяці тому

      @@AllAboutAIllms such as those built on transformer architectures, are fundamentally stateless, meaning they do not inherently maintain information about previous inputs across separate input sequences like recurrent neural networks. however; they can emulate state-like behavior through the use of positional and specialized embeddings that incorporate contextual information within a given sequence, processing data in a stateless manner, the autoregressive nature of many llms allows them to generate text by sequentially predicting the next token based on the accumualted outputs, mimicking a form of statefulness. allowing them to handle extensive and complex sequences effectively, tho each processing step inherently lacks a continuous internal state beyond its immediate inputs.