Emerging architectures for LLM applications

Поділитися
Вставка
  • Опубліковано 15 жов 2024
  • Everything from training models from scratch and fine-tuning open-source models to using hosted APIs, with a particular emphasis on the design pattern of in-context learning.
    Key topics we'll cover during the session include:
    Data preprocessing and embedding, focusing on the role of contextual data, embeddings, and vector databases in creating effective LLM applications.
    Strategies for prompt construction and retrieval, which are becoming increasingly complex and critical for product differentiation.
    Prompt execution and inference, analyzing the leading language model providers, their models, and tools used for logging, tracking, and evaluation of LLM outputs.
    Hosting solutions for LLMs, comparing the common solutions and emerging tools for easier and more efficient hosting of LLM applications.
    Whether you're a seasoned AI professional, a developer beginning your journey with LLMs, or simply an enthusiast interested in the applications of AI, this webinar offers valuable insights that can help you navigate the rapidly evolving landscape of LLMs.
    Follow along with the slides here go.superwise.a...

КОМЕНТАРІ • 33

  • @MattHabermehl
    @MattHabermehl Рік тому +14

    4k views and only 2 comments. This is the best UA-cam video I've seen by far on these strategies. Great content - thank you so much for sharing your expertise!

  • @investigativeinterviewing4617
    @investigativeinterviewing4617 Рік тому +18

    This is one of the best webinars I have seen on this topic. Great slides and presenters!

  • @williampourmajidi4710
    @williampourmajidi4710 Рік тому +9

    🎯 Key Takeaways for quick navigation:
    00:00 📚 Introduction to the topic of emerging architectures for LLM applications.
    01:54 🧐 Why focus on LLM architectures.
    04:02 📊 Audience poll on LLM use cases.
    05:17 🧠 Retrieval Augmented Generation (RAG) as a design pattern.
    08:05 💡 Advanced techniques in RAG and architectural considerations.
    14:40 📦 Orchestration and addressing complex tasks with LLMs.
    23:53 🧩 LLMs in Intermediate Summarization
    26:43 📊 Monitoring in LLM Architecture
    32:04 🛠️ LLM Agents and Tools
    39:05 🔄 Improving LLM Inference Speed
    49:26 🛡️ OpenAI's ChatGPT and its relevance in the field,
    50:12 🌐 Evolution of ChatGPT and the AI landscape,
    51:09 💼 OpenAI's models and their resource allocation,
    52:16 🏢 Factors influencing model choice: Engineering, economy, and legal considerations,
    Made with HARPA AI

  • @vakman9497
    @vakman9497 Рік тому

    I was very pleased to see how well everything was broken down! I was also shook to see a lot of the architecture strategies were things we were already implementing at our company so I'm happy to see we are on the right track 😅

  • @maria-wh3km
    @maria-wh3km 2 місяці тому

    it was awesome, thanks guys, keep up the good work.

  • @afederici75
    @afederici75 Рік тому +3

    This vieo was great! Thank you so much.

  • @sunnychopper6663
    @sunnychopper6663 Рік тому +1

    Really informative video. It will be interesting to see how different layers are formed throughout the coming months. Given the complexities of RAG, it'd be interesting to see hosted solutions that can offer competitive pricing on a RAG engine.

  • @todd-alex
    @todd-alex Рік тому +2

    Very informative. Several layers of LLM architectures need to be simplified like this. Maybe a standard for XAI should be developed based on a simplified architectural stack like this for LLMs.

  • @dr-maybe
    @dr-maybe Рік тому +3

    Very interesting, thanks for sharing

  • @vikassalaria24
    @vikassalaria24 Рік тому +2

    Really great presentation.Keep up the good work

  • @IsraelDavid-z8g
    @IsraelDavid-z8g Рік тому

    Wonderful video, learns a lot, thanks. This vieo was great! Thank you so much..

  • @mayurpatilprince2936
    @mayurpatilprince2936 Рік тому

    Informative video ... Waiting for next video :)

  • @MMABeijing
    @MMABeijing Рік тому

    That was very nice, thank you all

  • @_rjlynch
    @_rjlynch Рік тому

    Very informative, thanks!

  • @MengGe-s8l
    @MengGe-s8l Рік тому +2

    Wonderful video, learns a lot, thanks

  • @hidroman1993
    @hidroman1993 Рік тому

    So informative, looking forward to seeing more

  • @zhw7635
    @zhw7635 Рік тому +2

    Nice to see these topics covered, these come up as soon as I was attempting to implement something with llms

  • @VaibhavPatil-rx7pc
    @VaibhavPatil-rx7pc Рік тому +1

    Excellent detailed information thanks, please share slide details,

    • @superwiseai
      @superwiseai  Рік тому +1

      Thank you!
      You can access the slides here - go.superwise.ai/hubfs/PDF%20assets/LLM%20Architectures_8.8.2023.pdf

  • @billykotsos4642
    @billykotsos4642 Рік тому

    Great talk !

  • @vladimirobellini6128
    @vladimirobellini6128 9 місяців тому

    great ideas txs!

  • @RiazLaghari
    @RiazLaghari 8 місяців тому

    Great!

  • @HodgeLukeCEO
    @HodgeLukeCEO Рік тому +3

    Can you make the slides available? I have an issue seeing them and following along.

    • @superwiseai
      @superwiseai  Рік тому +1

      No problem here you go - go.superwise.ai/hubfs/PDF%20assets/LLM%20Architectures_8.8.2023.pdf

  • @GigaFro
    @GigaFro Рік тому

    Can someone provide an example of how one might introduce time as a factor in the embedding?

    • @serkanserttop1
      @serkanserttop1 Рік тому

      It would be in a meta field that you use to filter results, not in the vector embeddings itself.

  • @Aidev7876
    @Aidev7876 Рік тому

    Honestly. Not huge value for 55 minutes,,,

    • @k.8597
      @k.8597 Рік тому

      these videos seldom are.. lol.

  • @chirusikar
    @chirusikar 10 місяців тому

    Total gibberish in this video