Abonia Sojasingarayar
Abonia Sojasingarayar
  • 29
  • 26 558
Welcome and Join the Journey - AI, ML, DataScience, GenerativeAI, ComputerVision | Channel Trailer
🌟 Welcome to our Channel
🎥 In this trailer, I’ll give you a glimpse of what this channel is all about.
Thank you for being here-I can’t wait to share this journey with you!
___________________________________________________________________________
🔔 Get our Newsletter and Featured Articles: abonia1.github.io/newsletter/
🔗 Linkedin: www.linkedin.com/in/aboniasojasingarayar/
🔗 Find me on Github: github.com/Abonia1
🔗 Medium Articles: medium.com/@abonia
Переглядів: 75

Відео

Easily Run Hugging Face GGUF Models Locally with Ollama #LLM #HuggingFace #GGUFModels #Ollama#asitop
Переглядів 13721 день тому
In this video, we'll dive into the latest and most efficient method for running Hugging Face GGUF models with Ollama. We’ll download the MistralLite model from the Hugging Face hub, run it locally, and monitor resource and memory usage using asitop.The tutorial could cover: 1.What GGUF models are and their use cases. 2.Why use Ollama for local inference. 3.Step-by-step setup and execution. 4.Re...
Part 2 | MLOps On GitHub | Deploy and Automate ML Workflow |Using GitHub Actions and CML for CI & CD
Переглядів 287Місяць тому
Comprehensive tutorial on using GitHub Actions and Continuous Machine Learning (CML) to automate machine learning workflows! In this video, we’ll walk through the complete process of setting up a CI/CD pipeline for a machine learning project. By the end, you’ll be able to create and deploy automated workflows, monitor model performance, and collaborate seamlessly with your team! ⭐️ Contents ⭐️ ...
Part 1 | MLOps On GitHub | Deploy and Automate ML Workflow |Using GitHub Actions and CML for CI& CD
Переглядів 360Місяць тому
Comprehensive tutorial on using GitHub Actions and Continuous Machine Learning (CML) to automate machine learning workflows! In this video, we’ll walk through the complete process of setting up a CI/CD pipeline for a machine learning project focused on churn prediction. By the end, you’ll be able to create and deploy automated workflows, monitor model performance, and collaborate seamlessly wit...
Clustering using Embedding - KMeans - PCA - Visualization
Переглядів 1632 місяці тому
Clustering using Embedding - A Hands-on Guide to Web Scraping, Text Embedding, and KMeans Clustering with Python This tutorial demonstrates the clustering of random Wikipedia articles using HF embedding and K-means clustering. It showcases the entire pipeline from data collection to visualization of clusters. ⭐️ Methodology and Contents ⭐️ 0:00 Introduction 04:27 Fetch random Wikipedia articles...
Running a Streamlit App from Google Colab - Serve an LLM app in Colab
Переглядів 4582 місяці тому
How to build and deploy a powerful sentiment analysis web app using Streamlit and DistilBERT, a state-of-the-art transformer model fine-tuned for sentiment classification. We'll walk through setting up the model, building the web interface, and deploying it with localtunnel in Google Colab. Whether you're analyzing customer reviews or social media posts, this tutorial will help you create a fas...
Table Extraction from PDF using Camelot - Tabula - PDFPlumber #PDFTableExtraction #Hands-On
Переглядів 3173 місяці тому
Python provides strong libraries which allows smart table extraction from PDF , offering flexibility, automation, and handling of various PDF formats. We will explore the following Python libraries that were specifically developed for easier table extraction: 1. Camelot 2. Tabula 3. Pdfplumber ⭐️ Contents ⭐️ 0:00 Introduction 03:22 Setup and Installation 04:41 Camelot 8:30 Tabula 10:38 PDFplumb...
📚 Book Review - Mastering NLP from foundations for LLMs
Переглядів 2103 місяці тому
Book Review - Mastering NLP from foundations for LLMs 🔔 Newsletter and Featured Articles: abonia1.github.io/newsletter/ 🔗 Linkedin: / aboniasojasingarayar 🔗 Find me on Github : github.com/Abonia1 🔗 Medium Articles: / abonia 🔗 Substack AI Magazine: aboniasojasingarayar.substack.com/
Running Ollama in Colab (Free Tier) - Step by Step Tutorial
Переглядів 2,4 тис.4 місяці тому
Ollama empowers you to leverage powerful large language models (LLMs) like Llama2,Llama3,Phi3 etc. without needing a powerful local machine. Google Colab’s free tier provides a cloud environment perfectly suited for running these resource-intensive models. This tutorial details setting up and running Ollama on the free version of Google Colab, allowing you to explore the capabilities of LLMs wi...
PandasAI and Ollama running locally
Переглядів 5194 місяці тому
PandasAI represents a major advancement in data analysis, effectively bridging the gap between traditional coding methods and intuitive natural language interactions. By automating repetitive tasks, generating accurate insights, and seamlessly transforming data, it enables users to extract maximum value from their data without needing extensive coding. ⭐️ Contents ⭐️ 00:00 Introduction to Panda...
SAM 2 Segment Anything - Image and Video Segmentation #computervision #objectsegmentation #sam #meta
Переглядів 5525 місяців тому
Advanced model for comprehensive object segmentation in both images and videos. It features a unified, promptable model architecture that excels in processing complex visual data in real time and supports zero-shot generalization. ✨ Key Features ✨ ✅ Unified Model Architecture ✅ Real-Time Performance ✅ Zero-Shot Generalization ✅ Interactive Refinement ✅ Advanced Visual Handling 🌟 Content 🌟 00:00...
📚 Book Review - Transformers for Natural Language Processing and Computer Vision - 3rd Edition
Переглядів 3115 місяців тому
Book Review: Transformers for Natural Language Processing and Computer Vision - Third Edition 🔔 Newsletter and Featured Articles: abonia1.github.io/newsletter/ 🔗 Linkedin: www.linkedin.com/in/aboniasojasingarayar/ 🔗 Find me on Github : github.com/Abonia1 🔗 Medium Articles: medium.com/@abonia 🔗 Substack AI Magazine: aboniasojasingarayar.substack.com/
Fine-Tuning YOLOv10 for Object Detection on a Custom Dataset #yolo #finetuning
Переглядів 1,7 тис.6 місяців тому
YOLOv10 is a new generation in the YOLO series for real-time end-to-end object detection. It aims to improve both the performance and efficiency of YOLO models by eliminating the need for non-maximum suppression (NMS) and comprehensively optimizing the model architecture. In this tutorial, we will explore its architecture and how to fine-tune it to detect cancer cells for cancer diagnosis. ⭐️ C...
Building and Testing a Multi-Modal Retriever - Hands-On #llamaindex #CLIPembeddings #Image-TextIndex
Переглядів 1577 місяців тому
Multi-Modal Retrieval using BGE text embedding and CLIP image embedding for Wikipedia Articles. In this tutorial, you will: 1. Understand the basics of multi-modal retrieval 2. Implement BGE text embedding 3. Integrate CLIP image embedding 4. Build a multi-modal retriever for Wikipedia articles 5. Visualize the result Access the code and notebook used in this tutorial: gist.github.com/Abonia1/1...
Anylabeling - Image Annotation Tool - ObjectDetection and Instance Segmenation #Computervision #YOLO
Переглядів 5277 місяців тому
Anylabeling - Image Annotation Tool - ObjectDetection and Instance Segmenation #Computervision #YOLO
Top LLM and Deep Learning Inference Engines - Curated List
Переглядів 2428 місяців тому
Top LLM and Deep Learning Inference Engines - Curated List
Summarization with LangChain using LLM - Stuff - Map_reduce - Refine
Переглядів 9388 місяців тому
Summarization with LangChain using LLM - Stuff - Map_reduce - Refine
Deploying a Retrieval-Augmented Generation (RAG) in AWS Lambda
Переглядів 3 тис.9 місяців тому
Deploying a Retrieval-Augmented Generation (RAG) in AWS Lambda
Build and Deploy LLM Application in AWS Lambda - BedRock - LangChain
Переглядів 8 тис.9 місяців тому
Build and Deploy LLM Application in AWS Lambda - BedRock - LangChain
Run Ollama with Langchain Locally - Local LLM
Переглядів 2 тис.10 місяців тому
Run Ollama with Langchain Locally - Local LLM
LLMLingua - Prompt Compression for LLM Use Cases 🔥
Переглядів 36211 місяців тому
LLMLingua - Prompt Compression for LLM Use Cases 🔥
What is RAG (Retrieval-Augmented Generation)?
Переглядів 25611 місяців тому
What is RAG (Retrieval-Augmented Generation)?
BERTScore Explained in 5 minutes
Переглядів 1,9 тис.Рік тому
BERTScore Explained in 5 minutes
Must read LLM and AI Research Papers of 2023 🔥
Переглядів 208Рік тому
Must read LLM and AI Research Papers of 2023 🔥

КОМЕНТАРІ

  • @HelloIamLauraa
    @HelloIamLauraa День тому

    thank youu I really liked it, very good explained

  • @zameerahmed1775
    @zameerahmed1775 15 днів тому

    ur a bit fast .......... pls slow down ur pace.

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 15 днів тому

      Hi Zameer Ahmed , Ah sorry about it, Sure will have this in my mind. thanks for your kind feedback and wish you an advance happy new year 🎉🙂

  • @leelasankar1
    @leelasankar1 16 днів тому

    Thank you very much such a great video

  • @יהונתןבוגנים-מ6כ
    @יהונתןבוגנים-מ6כ 16 днів тому

    Question about lambda function. If I decide in x time to change the rag mechanism, and I change my code. All this after I pushed a Docker image once and created a lambda function. What steps should I take? Does every change in my code require me to redeploy? Retag an image, upload to ecr, push it, create a new lambda function? I would appreciate help

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 16 днів тому

      Hello, yes, every code change requires rebuilding the Docker image, pushing it to ECR and then explicitly updating the Lambda function configuration to use the new image URI. This ensures you're deploying a complete, tested version each time. Use version tags on your ECR images (e.g v1.0) for better tracking.

  • @namankumarmuktha4507
    @namankumarmuktha4507 20 днів тому

    Life Saver..

  • @sanreetkaurmann4796
    @sanreetkaurmann4796 23 дні тому

    Oh my god, may god bless you forever and ever, you literally have no idea how this video helped... Thank you so so very muchhhhh!!!✨✨✨✨✨✨ Hope you have an amazinggg dayyy aheadd✨✨😇😇❤❤

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 22 дні тому

      Hi Sanreet, Happy to help and I'm glad it helped. Thank you so much for your kind words 🙂

  • @apocart8426
    @apocart8426 24 дні тому

    hi thank you very much for this insightful video. I was also wondering if we could run image generation models such as flux on google collab as well.

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 24 дні тому

      Hi, yes, absolutely we can. Sample notebook: colab.research.google.com/github/NielsRogge/Transformers-Tutorials/blob/master/Flux/Run_Flux_on_an_8GB_machine.ipynb Hope this helps.

    • @apocart8426
      @apocart8426 23 дні тому

      @@AboniaSojasingarayar You're amazing thankyou

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 23 дні тому

      Happy to help.

    • @apocart8426
      @apocart8426 21 день тому

      @@AboniaSojasingarayar hey so i tried this out, one of the issues im having is that this exhausts all of the free resources of colab. so if you know any good model that we can use on colab and how to use it then please do share

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 21 день тому

      You may try using a quantized model and GPU runtime. To install such model , recently published a tutorial on how we can pull gguf quantized model into ollama: ua-cam.com/video/8MjS0aOV8tE/v-deo.htmlsi=J5fCzL6lw_zRGJ2p - Additionally try using subprocess and threading for better performance.sample code here: def run_ollama(): subprocess.Popen(["ollama", "serve"]) ollama_thread = threading.Thread(target=run_ollama) ollama_thread.start() Hope this helps.

  • @zerofive3699
    @zerofive3699 Місяць тому

    👍

  • @marvinacklin792
    @marvinacklin792 2 місяці тому

    What language in this?

  • @Bumbblyfestyle
    @Bumbblyfestyle 2 місяці тому

    Very useful

  • @johannes7856
    @johannes7856 2 місяці тому

    Nice Tutorial, thanks. 😊

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 2 місяці тому

      Thank you so much! 😊 I’m glad you found the tutorial helpful!

    • @johannes7856
      @johannes7856 2 місяці тому

      @@AboniaSojasingarayar Do you know if there is a tool that can convert the Annoted json from the Anylabeling tool to the yolo format?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 2 місяці тому

      @johannes7856 Hi Johannes, You may try following library github.com/rooneysh/Labelme2YOLO If not you can convert any labelling json to coco json and again convert it to yolo using the above library. Hope this helps.

  • @DevSingh-v2h
    @DevSingh-v2h 2 місяці тому

    can you please share the colab notebook

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 2 місяці тому

      Sure, here it is: gist.github.com/Abonia1/fc442374e1c20c86db8effbf95d93eb6

  • @khlifimohamedrayen1303
    @khlifimohamedrayen1303 2 місяці тому

    Thank you very much for this tutorial! I was having many problems running the ollama server on colab without the colabxterm... You're such a life saver!

  • @Bumbblyfestyle
    @Bumbblyfestyle 2 місяці тому

    Good info 😊

  • @mohamadadhikasuryahaidar7652
    @mohamadadhikasuryahaidar7652 3 місяці тому

    thanks for the tutorial

  • @anandrajgt3602
    @anandrajgt3602 3 місяці тому

    Please post a video regarding github actions

  • @Nabeel27
    @Nabeel27 3 місяці тому

    I get error: Runtime.ImportModuleError: Unable to import module 'lambda_function': Error importing numpy: you should not try to import numpy from its source directory; please exit the numpy source tree, and relaunch your python interpreter from there. followed all steps as in your video.

    • @Nabeel27
      @Nabeel27 3 місяці тому

      Looks like I had to setup the lambda as arm64 and the layer (created on mac Docker) also as arm64. Next, it also requires Bedrock setup and access request to llama model to use. llama 2 is no longer available, have to request llama 3 8B or something else.

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 3 місяці тому

      Hello Nabeel, Are you still facing the above issue?

    • @Nabeel27
      @Nabeel27 3 місяці тому

      @@AboniaSojasingarayar Thank you so much for following up! the error I am getting now is this: "errorMessage": "Error raised by bedrock service: An error occurred (AccessDeniedException) when calling the InvokeModel operation: User: arn:aws:sts::701934491353:assumed-role/test_demo-role-sfu6wu6d/test_demo is not authorized to perform: bedrock:InvokeModel on resource: arn:aws:bedrock:us-east-1::foundation-model/meta.llama3-8b-instruct-v1:0 because no identity-based policy allows the bedrock:InvokeModel action",

    • @Nabeel27
      @Nabeel27 3 місяці тому

      @@AboniaSojasingarayar I was able to solve it. I got the permission to use llama3 and also had to update role permissions to use Bedrock.

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 3 місяці тому

      @@Nabeel27 Great 🎉

  • @Bumbblyfestyle
    @Bumbblyfestyle 3 місяці тому

  • @zerofive3699
    @zerofive3699 3 місяці тому

    Awesome abo keep up the good work

  • @enia123
    @enia123 3 місяці тому

    thank you I was studying something related, but my computer's performance was very poor due to lack of money. I had a problem with ollama not working in Colab, but it was resolved! thank you I would like to test a model created in Colab. Is there a way to temporarily run it as a web service?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 3 місяці тому

      Most welcome. Great and glad to hear that finally it worked. 1. Of course we can use the flask API and ColabCode package to serve your mode via endpoint in ngrok temporary URL. github.com/abhishekkrthakur/colabcode 2. And another way is using flask and flask-ngrok. pypi.org/project/flask-ngrok/ pypi.org/project/Flask-API/ Sample code for reference: from flask import Flask from flask_ngrok import run_with_ngrok app = Flask(__name__) run_with_ngrok(app) @app.route("/") def home(): return "Hello World" app.run() If needed I'll try to do a tuto on this topic in future. Hope this helps:)

    • @enia123
      @enia123 3 місяці тому

      @@AboniaSojasingarayar thank you Have a nice day~

  • @tapiaomars
    @tapiaomars 3 місяці тому

    Hi, its possible integrate DynamoDB for store and retrive context of last user prompts in lambda function?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 3 місяці тому

      Hello, Yes , DynamoDB, S3, or in-memory storage depending on requirements. Each piece of context is associated with a user ID, ensuring that contexts are isolated per user with conversation ID. Hope this helps.

    • @tapiaomars
      @tapiaomars 3 місяці тому

      @@AboniaSojasingarayar Thanks, I'll try it and let you know how it goes.

  • @ziaullah2115
    @ziaullah2115 3 місяці тому

    please create one video for breast cancer detection in yolov10 model

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 2 місяці тому

      Absolutely, I’ll work on getting it ready shortly. If there are specific areas you want me to concentrate on, just let me know! Also, do you have any custom dataset you'd like to use for this tutorial? Thanks

  • @iroudayaradjcalingarayar317
    @iroudayaradjcalingarayar317 3 місяці тому

    Super

  • @VenkatesanVenkat-fd4hg
    @VenkatesanVenkat-fd4hg 3 місяці тому

    Great discussion....

  • @mayshowgunmore5269
    @mayshowgunmore5269 4 місяці тому

    Hi I'm trying to run these processes, but in this video 12:36 how to create and execute the file named ".env" , it always show Error , I can't figure it out. Thanks!

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 3 місяці тому

      Hello, You can use local VScode or any IDE to create .env New file -> name it as .env And add your API key as follows: ROBOFLOW_API_KEY=your_api_key Once done drag and drop it in colab. Hope this helps.

  • @VenkatesanVenkat-fd4hg
    @VenkatesanVenkat-fd4hg 4 місяці тому

    Great share, insightful share as always...Are u using obs studio for recording....by Senior Data Scientist....

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 4 місяці тому

      Glad it helped. Not really! Just using the built-in recording and iMovie to edit it.

  • @alvaroaraujo7945
    @alvaroaraujo7945 4 місяці тому

    Hey, Abonia..Thanks for the amazing content. I just had one issue though: on executing the 'map_reduce_outputs' function, I had the ConnectionRefusedError: [Errno 61]. Hope someone know what it is

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 4 місяці тому

      @@alvaroaraujo7945 Hello , thanks for your kind words. It may be related to your ollama serve.Are you sure Ollama is running ?

  • @machinelearningzone.6230
    @machinelearningzone.6230 4 місяці тому

    Nice explanation and walkthrough. Could you provide the link to the code repo for this exercise.

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 4 місяці тому

      Glad it helped. As mentioned in the description, you can find the code and explanation in this article walkthrough. medium.com/@abonia/deploying-a-rag-application-in-aws-lambda-using-docker-and-ecr-08e246a7c515

  • @zerofive3699
    @zerofive3699 4 місяці тому

    It is very helpful mam , it is useful on impliying

  • @zerofive3699
    @zerofive3699 4 місяці тому

    Nice video mam

  • @user-wr4yl7tx3w
    @user-wr4yl7tx3w 4 місяці тому

    Can we simply rely on open source only without using Amazon? What if it is just prototyping?

  • @World-um5vo
    @World-um5vo 5 місяців тому

    Hi, Thank you for the video, So if we want to fine tune the model and evaluate it for videos, then how to do it ?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 5 місяців тому

      Your most welcome. Here I have introduced basic usage of SAM 2 models. If you want to evaluate your finetuned model you may try mean IoU score, for a set of predictions and targets or DICE, precision, recall, and mAP.

  • @Basant5911
    @Basant5911 5 місяців тому

    streaming does't work via doing this. I wrote code from scratch without langchain.

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 5 місяців тому

      @@Basant5911 can you share your code base and error or issue that you are facing currently please?

  • @DenisRothman
    @DenisRothman 5 місяців тому

    ❤Thank you for this fantastic educational video on my book!!! 🎉

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 5 місяців тому

      @@DenisRothman Thank you for your kind words. I'm grateful for the opportunity to review the book and share my thoughts. Your work is well-deserved and truly one of the most insightful books I've read.

  • @MohamedMohamed-xf7wh
    @MohamedMohamed-xf7wh 5 місяців тому

    You used a webpage as a data source for the RAG app, what If I add pdf file instead of the webpage as a data source, how can I deploy it in aws lambda?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 5 місяців тому

      To build RAG with pdf in AWS ecosystem, you need to follow steps that involve uploading the PDF to an S3 bucket, extracting text from the PDF, and then integrating this data with your RAG application.

    • @MohamedMohamed-xf7wh
      @MohamedMohamed-xf7wh 5 місяців тому

      @@AboniaSojasingarayar Can I locally extract text from pdf and build vector DB locally using vscode and then build the docker image and push it to ECR AWS like what you did in the video?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 5 місяців тому

      @@MohamedMohamed-xf7wh Yes, you can locally extract text from PDF files, build a vector database and then prepare your application for deployment on AWS Lambda by building a Docker image and pushing it to ECR. But which vector db are you using? It can be accessible with API?

    • @MohamedMohamed-xf7wh
      @MohamedMohamed-xf7wh 5 місяців тому

      @@AboniaSojasingarayar FAISS .. what is the problem with vector db?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 5 місяців тому

      @@MohamedMohamed-xf7wh Great!

  • @htayaung3812
    @htayaung3812 6 місяців тому

    Really Nice! Keep going. You deserve more subscribers.

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 5 місяців тому

      @@htayaung3812 Thank you so much for your support! I'm working to bring more tutorials.

  • @raulpradodantas9386
    @raulpradodantas9386 6 місяців тому

    Save my life to create lambda layers... I have been trying for days. TKS!

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 5 місяців тому

      @@raulpradodantas9386 Glad to hear that! You most welcome.

  • @SidSid-kp4ij
    @SidSid-kp4ij 6 місяців тому

    Hi I'm trying to run my trained model with interface to webcam but getting error can you share any insight on it

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 6 місяців тому

      @@SidSid-kp4ij Hello Sid, Sure can you post your error message here please?

  • @gk4457
    @gk4457 7 місяців тому

    All the best

  • @RajuSubramaniam-ho6kd
    @RajuSubramaniam-ho6kd 7 місяців тому

    Thanks for the video. Very useful for me as I am new to AWS lambda and bedrock. Can you please upload the lambda function source code? Thanks again!

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 7 місяців тому

      Glad it helped. Sure you can find the code and complete the article on this topic in the description. In any way here is the link to the code : medium.com/@abonia/build-and-deploy-llm-application-in-aws-cca46c662749

  • @jannatbellouchi3908
    @jannatbellouchi3908 7 місяців тому

    Which version of BERT is it used in BERTScore ?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 7 місяців тому

      As we are using lang= "en" so it uses roberta-large. We can also customize it using the model_type param of BERScorer class For default model for other languages,find it here: github.com/Tiiiger/bert_score/blob/master/bert_score/utils.py

  • @jagadeeshprasad5252
    @jagadeeshprasad5252 7 місяців тому

    hey great content. please continue to do more videos and real time projects. Thanks

  • @zerofive3699
    @zerofive3699 7 місяців тому

    Awesome mam , very easy to understand

  • @NJ-hn8yu
    @NJ-hn8yu 7 місяців тому

    Hi Abonia, thanks for sharing. I am facing this error . can you please tell how to resolve it "errorMessage": "Unable to import module 'lambda_function': No module named 'langchain_community'",

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 7 місяців тому

      Hello, You are most welcome. You must prepare your ZIP file with all the necessary packages. You can refer to the instructions starting at the 09:04

  • @humayounkhan7946
    @humayounkhan7946 7 місяців тому

    Hi Abonia, thanks for the thorough guide, but i'm abit confused with the lambda_layer.zip file, why did you have to create it through docker? is there an easier way to provide the dependencies in a zip file without going through docker? Thanks in advance!

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 7 місяців тому

      Hi Humayoun Khan, Yes we can but Docker facilitates the inclusion of the runtime interface client for Python, making the image compatible with AWS Lambda. Also it ensures a consistent and reproducible environment for Lambda function's dependencies. This is crucial for avoiding discrepancies between development, testing, and production environments. Hope this helps.

  • @evellynnicolemachadorosa2666
    @evellynnicolemachadorosa2666 8 місяців тому

    hello! Thanks for the video. I am from Brazil. What would you recommend for large documents, averaging 150 pages? I tried map-reduce, but the inference time was 40 minutes. Are there any tips for these very long documents?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 8 місяців тому

      Thanks for you kind words and glad this helped. Implement a strategy that combines semantic chunking with K-means clustering to address the model’s contextual limitations. By employing efficient clustering techniques, we can extract key passages effectively, thereby reducing the overhead associated with processing large volumes of text. This approach not only significantly lowers costs by minimizing the number of tokens processed but also mitigates the recency and primacy effects inherent in LLMs, ensuring a balanced consideration of all text segments.

    • @VirtualMachine-d8x
      @VirtualMachine-d8x 3 місяці тому

      ​@@AboniaSojasingarayar Video was great and very useful.. can you make the small video on this clustering method using embedding ?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 3 місяці тому

      @@VirtualMachine-d8x Sure will do, happy to hear from you again. Thanks for the feedback.

  • @Coff03
    @Coff03 8 місяців тому

    Did you use OpenAI API key here?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 8 місяців тому

      Here we use open-source Mixtral from ollama.But, yes we can use OpenAI models as well.

  • @MishelMichel
    @MishelMichel 8 місяців тому

    Very informatics nd Your voice very clear dr

  • @fkeb37e9w0
    @fkeb37e9w0 8 місяців тому

    Can we use openai and chromadb on aws??

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 8 місяців тому

      Yes we can! In the below tutorial I have demonstrated how we can create and deploy lambda layer via container for larger dependencies : ua-cam.com/video/gicsb9p7uj4/v-deo.htmlsi=F_X7-6YCAb0Kz3Jc

    • @fkeb37e9w0
      @fkeb37e9w0 8 місяців тому

      @@AboniaSojasingarayar yes but can this be done without eks or containers?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 8 місяців тому

      Yes! You can try it by creating a custom lambda layer.If you face issue try to use only the required libraries and remove any unnecessary dependencies from your zip file.Hope this helps.

  • @vijaygandhi7313
    @vijaygandhi7313 8 місяців тому

    In the abstractive summarization use-case, usually a lot of focus is given to the LLMs being used and its performance. Limitations of LLM including context length and ways to overcome this issue are often overlooked. Its important to make sure that our application is scalable when dealing with large document sizes. Thank you for this great and insightful video.

    • @AboniaSojasingarayar
      @AboniaSojasingarayar 8 місяців тому

      Thank you Vijay Gandhi, for your insightful comment! You've raised an excellent point about the importance of considering the limitations of LLMs in the context of abstractive summarization, especially regarding their context length and scalability issues when dealing with large documents. Indeed, one of the significant challenges in using LLMs for abstractive summarization is their inherent limitation in processing long texts due to the maximum token limit imposed by these models. This constraint can be particularly problematic when summarizing lengthy documents or articles, where the full context might not fit within the model's capacity.