What is the LLM's Context Window ?

Поділитися
Вставка
  • Опубліковано 11 вер 2024
  • 📹 VIDEO TITLE 📹
    What is the LLM's Context Window ?
    ✍️VIDEO DESCRIPTION ✍️
    In this video, we delve into the concept of the LLM Context Window and discuss how to optimize your system it for improved performance in cases where your input text exceeds the context window limit. Learn the importance of understanding and adjusting the LLM Context Window to enhance the accuracy and efficiency of your artificial intelligence models. Stay tuned to elevate your AI skills with expert insights and practical tips.
    🧑‍💻GITHUB URL 🧑‍💻
    No code samples for this video
    📽OTHER NEW MACHINA VIDEOS REFERENCED IN THIS VIDEO 📽
    What is the LLM’s Context Window? - • What is the LLM's Cont...
    What is LLM Chain of Thought Prompting? - • What is LLM Chain of T...
    Algorithms for Search Similarity - • Algorithms for Search ...
    How LLMs use Vector Databases - • How LLMs use Vector Da...
    What are LLM Embeddings ? - • What are LLM Embeddings ?
    How LLM’s are Driven by Vectors - • How LLM’s are Driven b...
    What is 0, 1, and Few Shot LLM Prompting ? - • What is 0, 1, and Few ...
    What are the LLM’s Top-P and TopK ? - • What are the LLM’s Top...
    What is the LLM’s Temperature ? - • What is the LLM’s Temp...
    What is LLM Prompt Engineering ? - • What is LLM Prompt Eng...
    What is LLM Tokenization? - • What is LLM Tokenizati...
    What is the LangChain Framework? - • What is the LangChain ...
    CoPilots vs AI Agents - • AI CoPilots versus AI ...
    What is an AI PC ? - • What is an AI PC ?
    What are AI HyperScalers? - • What are AI HyperScalers?
    What is LLM Fine-Tuning ? - • What is LLM Fine-Tuning ?
    What is LLM Pre-Training? - • What is LLM Pre-Training?
    AI ML Training versus Inference - • AI ML Training versus ...
    What is meant by AI ML Model Training Corpus? - • What is meant by AI ML...
    What is AI LLM Multi-Modality? - • What is AI LLM Multi-M...
    What is an LLM ? - • What is an LLM ?
    Predictive versus Generative AI ? - • Predictive versus Gene...
    What is a Foundation Model ? - • What is a Foundation M...
    What is AI, ML, Neural Networks and Deep Learning? - • What is AI, ML, Neural...
    AWS Lambda + Amazon Polly #001100 - • AWS Lambda + AWS Polly...
    AWS Lambda + Amazon Rekognition #001102 - • AWS Lambda + AWS Rekog...
    AWS Lambda + Amazon Comprehend #001103 - • AWS Lambda + AWS Compr...
    Why can’t you have AI driven Text Extraction? #001106 - • Why can’t you have AI ...
    Which Amazon ML / AI Service should you Use ? #001110 - • Which Amazon ML / AI S...
    Why can’t I do Generative AI in AWS? #001112 - • Why can’t I do Generat...
    Why care about Foundation Models? #001113 • Why care about Foundat...
    Why play in Amazon Bedrock playgrounds? #001114 • Why play in Amazon Bed...
    Get a ChatGPT API Key Now! #001000 - • Get a ChatGPT API Key ...
    AWS Lambda + ChatGPT API #001001 - • AWS Lambda + ChatGPT A...
    Lambda + ChatGPT + DynamoDb #001002 - • Lambda + ChatGPT + Dyn...
    Your own Custom AWS Website + ChatGPT API (part 1 of 5) #001003 - • Your own Custom AWS We...
    Your own Custom AWS Website + ChatGPT API (part 2 of 5) #001004 - • Your own Custom AWS We...
    Your own Custom AWS Website + ChatGPT API (part 3 of 5) #001005 - • Your own Custom AWS We...
    Your own Custom AWS Website + ChatGPT API (part 4 of 5) #001006 - • Your own Custom AWS We...
    Your own Custom AWS Website + ChatGPT API (part 5 of 5) #001007 - • Your own Custom AWS We...
    🔠KEYWORDS 🔠
    #LLM
    #LargeLanguageModel
    #NLP
    #NaturalLanguageProcessing
    #DataScience
    #MachineLearning
    #DataAnalysis
    #DeepLearning
    #LanguageModels
    #AI
    #ArtificialIntelligence
    #ContextWindow
    #NeuralNetworks
    #DeepLearning
    #DeepNeuralNetworks

КОМЕНТАРІ • 27

  • @aritzolaba
    @aritzolaba Місяць тому +2

    Cristal clear explanation. Thanks! more please :)

    • @NewMachina-CloudAI
      @NewMachina-CloudAI  Місяць тому

      You got it! Working to make each video better and better….

  • @veerabalajayaraj4459
    @veerabalajayaraj4459 9 днів тому

    Best explanation !

    • @NewMachina-CloudAI
      @NewMachina-CloudAI  9 днів тому

      Glad it was helpful! Trying to determine better with each video …. Thanks for feedback…

  • @rubenolguin5504
    @rubenolguin5504 4 дні тому

    Great content!

    • @NewMachina-CloudAI
      @NewMachina-CloudAI  День тому

      Thank you for your feedback… trying to get better with each video…. 🙏

  • @ParthivShah
    @ParthivShah 5 днів тому

    thanks.

    • @NewMachina-CloudAI
      @NewMachina-CloudAI  5 днів тому

      Glad you liked it…. Working to get better with each video… let me know if you have any ideas for videos …🙏

  • @ramakrishnay9887
    @ramakrishnay9887 Місяць тому +1

    Thanks for the explanation. Does it mean that the context window is common or saperate for both input and output?

    • @NewMachina-CloudAI
      @NewMachina-CloudAI  Місяць тому

      For LLM's, context window is for input tokens... There is normally, a LLM setting, called "maxLength" or something similar, that controls the maximum number of tokens, will be generated for a response... Thanks for feedback and question ....

    • @paultparker
      @paultparker 28 днів тому

      @@NewMachina-CloudAI i’m going to disagree here. I believe the context window typically includes both the LLM input and output, especially in a chat session like your examples. This is in most cases primarily how the LLM knows what it said before.

    • @NewMachina-CloudAI
      @NewMachina-CloudAI  27 днів тому +1

      @paulparker You are right... I was going through some documentation that ambigious about this ... and assumed it didn't include input.. I have found quite serveral other documents aligned with context window including both input and output. Thanks for helping clarify this ...

    • @paultparker
      @paultparker 23 дні тому

      @@NewMachina-CloudAI you’re welcome!

  • @paultparker
    @paultparker 28 днів тому

    The question on tooling is a good question. In my personal case, I don’t know enough here to know what tool I would prefer to use: my inclinations would be VS Code, and/or notebooks, but I don’t really understand Jupiter notebooks to be honest having never used them. I believe Colab and the like use notebooks?

    • @NewMachina-CloudAI
      @NewMachina-CloudAI  27 днів тому +1

      I am likely going to be showing examples just running in VSC, and maybe some in AWS Cloud using Lambda's and will likely do a simple one with Jupyter Notebooks to see how viewers like it ... thanks for providing your feedback on this....

    • @NewMachina-CloudAI
      @NewMachina-CloudAI  27 днів тому

      @paulparker Check out the frameworks LangChain and LlamaIndex.. I think these two open source frameworks will continue to get more traction... I am working on some videos in this area next ... I would interested if you have an opinion or thoughts on this frameworks ... not urgent, I suspect your are busy as we all are... but if you get a chance to check these out, let me know what you think ...

    • @paultparker
      @paultparker 23 дні тому

      @@NewMachina-CloudAI I thought that there was a different successor to Laingchain, and llama index doesn’t sound right. But I have not had time to mess with doing any of this myself.

    • @NewMachina-CloudAI
      @NewMachina-CloudAI  21 день тому

      @paulparker Are you maybe thinking about LangGraph or LangServe? Looks like there are some additional extensions to LangChain... some are driven by LangChain while others are from other teams.... Still getting a sense of all of these...

  • @paultparker
    @paultparker 28 днів тому +1

    Can you substantiate the claim that LLM providers do this primarily to make the models cheaper to run? I ask this because my understanding is that this is actually how the models work and have worked since the initial research. So it seems incorrect to say that this is an optimization chosen for performance at scale.

    • @NewMachina-CloudAI
      @NewMachina-CloudAI  27 днів тому +1

      Thanks for reaching out with your question... Can I get a quick clarification ... In the video "What is the LLM's Context Window", are you talking about the line "While larger 'context windows', improve the LLM’s performance, on longer text blocks, they also demand, more computational resources" ... I wanted to make sure I was following up on the the same part of the video you were were inquiring about ...

    • @paultparker
      @paultparker 23 дні тому +1

      @@NewMachina-CloudAI no, I think that was towards the beginning of the video, whereas I think what I am remember was towards the end.
      Yes, currently larger context windows require quadratically, more computation. However, there is a new approach that just came out for infinite context, windows. We will have to see if that is any good. Most errors in this post come from Siri’s broken dictation.

    • @NewMachina-CloudAI
      @NewMachina-CloudAI  21 день тому +1

      Ok, I will look into that... if you have a reference to this approach on infinite context please share.... New stuff happening quickly ...

  • @debojitmandal8670
    @debojitmandal8670 Місяць тому

    Then i have another question what's this concept
    Say for eg Annie loves jam but she hates bread and she also loves fruits
    So if i say context window is 2
    So i take 2 words towards left and 2 words towards right as input
    So for eg Annie loves but she is the input and jam is the output
    My second question is what's the difference bw context length and context windows to me what ever you explained sounded like context length rather then window so please help me to clarify

    • @NewMachina-CloudAI
      @NewMachina-CloudAI  Місяць тому

      Yes, context window, is measured in tokens. If Context Window is 2, then you could get one token and one token out.
      For second question, I should have been consistent, and used Context Window throughtout... for this topic, context length is the same as the Context window and is measured in tokens.

    • @NewMachina-CloudAI
      @NewMachina-CloudAI  Місяць тому

      Thanks for taking the time in asking me these questions ....

    • @debojitmandal8670
      @debojitmandal8670 Місяць тому

      @@NewMachina-CloudAI but sir what I have studied is context window and length are different contex window is the small amount window where your focus is but it's being interchanged very often.

    • @NewMachina-CloudAI
      @NewMachina-CloudAI  21 день тому

      Ahh.. I see what you are saying... I will try to be more precise with my terminology as well... thank you for sharing..