- 29
- 26 558
Abonia Sojasingarayar
India
Приєднався 23 сер 2013
👋 Hi there!
Here we share learning and insights of industry level AI, ML, DS solution and make complex concepts simple and captivating.
Let's learn, and grow together.
Here we share learning and insights of industry level AI, ML, DS solution and make complex concepts simple and captivating.
Let's learn, and grow together.
Welcome and Join the Journey - AI, ML, DataScience, GenerativeAI, ComputerVision | Channel Trailer
🌟 Welcome to our Channel
🎥 In this trailer, I’ll give you a glimpse of what this channel is all about.
Thank you for being here-I can’t wait to share this journey with you!
___________________________________________________________________________
🔔 Get our Newsletter and Featured Articles: abonia1.github.io/newsletter/
🔗 Linkedin: www.linkedin.com/in/aboniasojasingarayar/
🔗 Find me on Github: github.com/Abonia1
🔗 Medium Articles: medium.com/@abonia
🎥 In this trailer, I’ll give you a glimpse of what this channel is all about.
Thank you for being here-I can’t wait to share this journey with you!
___________________________________________________________________________
🔔 Get our Newsletter and Featured Articles: abonia1.github.io/newsletter/
🔗 Linkedin: www.linkedin.com/in/aboniasojasingarayar/
🔗 Find me on Github: github.com/Abonia1
🔗 Medium Articles: medium.com/@abonia
Переглядів: 75
Відео
Easily Run Hugging Face GGUF Models Locally with Ollama #LLM #HuggingFace #GGUFModels #Ollama#asitop
Переглядів 13721 день тому
In this video, we'll dive into the latest and most efficient method for running Hugging Face GGUF models with Ollama. We’ll download the MistralLite model from the Hugging Face hub, run it locally, and monitor resource and memory usage using asitop.The tutorial could cover: 1.What GGUF models are and their use cases. 2.Why use Ollama for local inference. 3.Step-by-step setup and execution. 4.Re...
Part 2 | MLOps On GitHub | Deploy and Automate ML Workflow |Using GitHub Actions and CML for CI & CD
Переглядів 287Місяць тому
Comprehensive tutorial on using GitHub Actions and Continuous Machine Learning (CML) to automate machine learning workflows! In this video, we’ll walk through the complete process of setting up a CI/CD pipeline for a machine learning project. By the end, you’ll be able to create and deploy automated workflows, monitor model performance, and collaborate seamlessly with your team! ⭐️ Contents ⭐️ ...
Part 1 | MLOps On GitHub | Deploy and Automate ML Workflow |Using GitHub Actions and CML for CI& CD
Переглядів 360Місяць тому
Comprehensive tutorial on using GitHub Actions and Continuous Machine Learning (CML) to automate machine learning workflows! In this video, we’ll walk through the complete process of setting up a CI/CD pipeline for a machine learning project focused on churn prediction. By the end, you’ll be able to create and deploy automated workflows, monitor model performance, and collaborate seamlessly wit...
Clustering using Embedding - KMeans - PCA - Visualization
Переглядів 1632 місяці тому
Clustering using Embedding - A Hands-on Guide to Web Scraping, Text Embedding, and KMeans Clustering with Python This tutorial demonstrates the clustering of random Wikipedia articles using HF embedding and K-means clustering. It showcases the entire pipeline from data collection to visualization of clusters. ⭐️ Methodology and Contents ⭐️ 0:00 Introduction 04:27 Fetch random Wikipedia articles...
Running a Streamlit App from Google Colab - Serve an LLM app in Colab
Переглядів 4582 місяці тому
How to build and deploy a powerful sentiment analysis web app using Streamlit and DistilBERT, a state-of-the-art transformer model fine-tuned for sentiment classification. We'll walk through setting up the model, building the web interface, and deploying it with localtunnel in Google Colab. Whether you're analyzing customer reviews or social media posts, this tutorial will help you create a fas...
Table Extraction from PDF using Camelot - Tabula - PDFPlumber #PDFTableExtraction #Hands-On
Переглядів 3173 місяці тому
Python provides strong libraries which allows smart table extraction from PDF , offering flexibility, automation, and handling of various PDF formats. We will explore the following Python libraries that were specifically developed for easier table extraction: 1. Camelot 2. Tabula 3. Pdfplumber ⭐️ Contents ⭐️ 0:00 Introduction 03:22 Setup and Installation 04:41 Camelot 8:30 Tabula 10:38 PDFplumb...
📚 Book Review - Mastering NLP from foundations for LLMs
Переглядів 2103 місяці тому
Book Review - Mastering NLP from foundations for LLMs 🔔 Newsletter and Featured Articles: abonia1.github.io/newsletter/ 🔗 Linkedin: / aboniasojasingarayar 🔗 Find me on Github : github.com/Abonia1 🔗 Medium Articles: / abonia 🔗 Substack AI Magazine: aboniasojasingarayar.substack.com/
Running Ollama in Colab (Free Tier) - Step by Step Tutorial
Переглядів 2,4 тис.4 місяці тому
Ollama empowers you to leverage powerful large language models (LLMs) like Llama2,Llama3,Phi3 etc. without needing a powerful local machine. Google Colab’s free tier provides a cloud environment perfectly suited for running these resource-intensive models. This tutorial details setting up and running Ollama on the free version of Google Colab, allowing you to explore the capabilities of LLMs wi...
PandasAI and Ollama running locally
Переглядів 5194 місяці тому
PandasAI represents a major advancement in data analysis, effectively bridging the gap between traditional coding methods and intuitive natural language interactions. By automating repetitive tasks, generating accurate insights, and seamlessly transforming data, it enables users to extract maximum value from their data without needing extensive coding. ⭐️ Contents ⭐️ 00:00 Introduction to Panda...
SAM 2 Segment Anything - Image and Video Segmentation #computervision #objectsegmentation #sam #meta
Переглядів 5525 місяців тому
Advanced model for comprehensive object segmentation in both images and videos. It features a unified, promptable model architecture that excels in processing complex visual data in real time and supports zero-shot generalization. ✨ Key Features ✨ ✅ Unified Model Architecture ✅ Real-Time Performance ✅ Zero-Shot Generalization ✅ Interactive Refinement ✅ Advanced Visual Handling 🌟 Content 🌟 00:00...
📚 Book Review - Transformers for Natural Language Processing and Computer Vision - 3rd Edition
Переглядів 3115 місяців тому
Book Review: Transformers for Natural Language Processing and Computer Vision - Third Edition 🔔 Newsletter and Featured Articles: abonia1.github.io/newsletter/ 🔗 Linkedin: www.linkedin.com/in/aboniasojasingarayar/ 🔗 Find me on Github : github.com/Abonia1 🔗 Medium Articles: medium.com/@abonia 🔗 Substack AI Magazine: aboniasojasingarayar.substack.com/
Fine-Tuning YOLOv10 for Object Detection on a Custom Dataset #yolo #finetuning
Переглядів 1,7 тис.6 місяців тому
YOLOv10 is a new generation in the YOLO series for real-time end-to-end object detection. It aims to improve both the performance and efficiency of YOLO models by eliminating the need for non-maximum suppression (NMS) and comprehensively optimizing the model architecture. In this tutorial, we will explore its architecture and how to fine-tune it to detect cancer cells for cancer diagnosis. ⭐️ C...
Building and Testing a Multi-Modal Retriever - Hands-On #llamaindex #CLIPembeddings #Image-TextIndex
Переглядів 1577 місяців тому
Multi-Modal Retrieval using BGE text embedding and CLIP image embedding for Wikipedia Articles. In this tutorial, you will: 1. Understand the basics of multi-modal retrieval 2. Implement BGE text embedding 3. Integrate CLIP image embedding 4. Build a multi-modal retriever for Wikipedia articles 5. Visualize the result Access the code and notebook used in this tutorial: gist.github.com/Abonia1/1...
Anylabeling - Image Annotation Tool - ObjectDetection and Instance Segmenation #Computervision #YOLO
Переглядів 5277 місяців тому
Anylabeling - Image Annotation Tool - ObjectDetection and Instance Segmenation #Computervision #YOLO
Top LLM and Deep Learning Inference Engines - Curated List
Переглядів 2428 місяців тому
Top LLM and Deep Learning Inference Engines - Curated List
Summarization with LangChain using LLM - Stuff - Map_reduce - Refine
Переглядів 9388 місяців тому
Summarization with LangChain using LLM - Stuff - Map_reduce - Refine
Deploying a Retrieval-Augmented Generation (RAG) in AWS Lambda
Переглядів 3 тис.9 місяців тому
Deploying a Retrieval-Augmented Generation (RAG) in AWS Lambda
Build and Deploy LLM Application in AWS Lambda - BedRock - LangChain
Переглядів 8 тис.9 місяців тому
Build and Deploy LLM Application in AWS Lambda - BedRock - LangChain
Run Ollama with Langchain Locally - Local LLM
Переглядів 2 тис.10 місяців тому
Run Ollama with Langchain Locally - Local LLM
LLMLingua - Prompt Compression for LLM Use Cases 🔥
Переглядів 36211 місяців тому
LLMLingua - Prompt Compression for LLM Use Cases 🔥
What is RAG (Retrieval-Augmented Generation)?
Переглядів 25611 місяців тому
What is RAG (Retrieval-Augmented Generation)?
Must read LLM and AI Research Papers of 2023 🔥
Переглядів 208Рік тому
Must read LLM and AI Research Papers of 2023 🔥
thank youu I really liked it, very good explained
You are most welcome and glad it helped.
ur a bit fast .......... pls slow down ur pace.
Hi Zameer Ahmed , Ah sorry about it, Sure will have this in my mind. thanks for your kind feedback and wish you an advance happy new year 🎉🙂
Thank you very much such a great video
Glad that it helped:)
Question about lambda function. If I decide in x time to change the rag mechanism, and I change my code. All this after I pushed a Docker image once and created a lambda function. What steps should I take? Does every change in my code require me to redeploy? Retag an image, upload to ecr, push it, create a new lambda function? I would appreciate help
Hello, yes, every code change requires rebuilding the Docker image, pushing it to ECR and then explicitly updating the Lambda function configuration to use the new image URI. This ensures you're deploying a complete, tested version each time. Use version tags on your ECR images (e.g v1.0) for better tracking.
Life Saver..
Glad to hear that it was helpful.
Oh my god, may god bless you forever and ever, you literally have no idea how this video helped... Thank you so so very muchhhhh!!!✨✨✨✨✨✨ Hope you have an amazinggg dayyy aheadd✨✨😇😇❤❤
Hi Sanreet, Happy to help and I'm glad it helped. Thank you so much for your kind words 🙂
hi thank you very much for this insightful video. I was also wondering if we could run image generation models such as flux on google collab as well.
Hi, yes, absolutely we can. Sample notebook: colab.research.google.com/github/NielsRogge/Transformers-Tutorials/blob/master/Flux/Run_Flux_on_an_8GB_machine.ipynb Hope this helps.
@@AboniaSojasingarayar You're amazing thankyou
Happy to help.
@@AboniaSojasingarayar hey so i tried this out, one of the issues im having is that this exhausts all of the free resources of colab. so if you know any good model that we can use on colab and how to use it then please do share
You may try using a quantized model and GPU runtime. To install such model , recently published a tutorial on how we can pull gguf quantized model into ollama: ua-cam.com/video/8MjS0aOV8tE/v-deo.htmlsi=J5fCzL6lw_zRGJ2p - Additionally try using subprocess and threading for better performance.sample code here: def run_ollama(): subprocess.Popen(["ollama", "serve"]) ollama_thread = threading.Thread(target=run_ollama) ollama_thread.start() Hope this helps.
👍
What language in this?
We used python to create this LLM app.
Very useful
Glad it was helpful :)
Nice Tutorial, thanks. 😊
Thank you so much! 😊 I’m glad you found the tutorial helpful!
@@AboniaSojasingarayar Do you know if there is a tool that can convert the Annoted json from the Anylabeling tool to the yolo format?
@johannes7856 Hi Johannes, You may try following library github.com/rooneysh/Labelme2YOLO If not you can convert any labelling json to coco json and again convert it to yolo using the above library. Hope this helps.
can you please share the colab notebook
Sure, here it is: gist.github.com/Abonia1/fc442374e1c20c86db8effbf95d93eb6
Thank you very much for this tutorial! I was having many problems running the ollama server on colab without the colabxterm... You're such a life saver!
You are most welcome! Glad it helped.
Good info 😊
Glad it helped 🙂
thanks for the tutorial
Happy to help
Please post a video regarding github actions
Sure! Thanks for your suggestion.
I get error: Runtime.ImportModuleError: Unable to import module 'lambda_function': Error importing numpy: you should not try to import numpy from its source directory; please exit the numpy source tree, and relaunch your python interpreter from there. followed all steps as in your video.
Looks like I had to setup the lambda as arm64 and the layer (created on mac Docker) also as arm64. Next, it also requires Bedrock setup and access request to llama model to use. llama 2 is no longer available, have to request llama 3 8B or something else.
Hello Nabeel, Are you still facing the above issue?
@@AboniaSojasingarayar Thank you so much for following up! the error I am getting now is this: "errorMessage": "Error raised by bedrock service: An error occurred (AccessDeniedException) when calling the InvokeModel operation: User: arn:aws:sts::701934491353:assumed-role/test_demo-role-sfu6wu6d/test_demo is not authorized to perform: bedrock:InvokeModel on resource: arn:aws:bedrock:us-east-1::foundation-model/meta.llama3-8b-instruct-v1:0 because no identity-based policy allows the bedrock:InvokeModel action",
@@AboniaSojasingarayar I was able to solve it. I got the permission to use llama3 and also had to update role permissions to use Bedrock.
@@Nabeel27 Great 🎉
❤
Awesome abo keep up the good work
Thanks!
thank you I was studying something related, but my computer's performance was very poor due to lack of money. I had a problem with ollama not working in Colab, but it was resolved! thank you I would like to test a model created in Colab. Is there a way to temporarily run it as a web service?
Most welcome. Great and glad to hear that finally it worked. 1. Of course we can use the flask API and ColabCode package to serve your mode via endpoint in ngrok temporary URL. github.com/abhishekkrthakur/colabcode 2. And another way is using flask and flask-ngrok. pypi.org/project/flask-ngrok/ pypi.org/project/Flask-API/ Sample code for reference: from flask import Flask from flask_ngrok import run_with_ngrok app = Flask(__name__) run_with_ngrok(app) @app.route("/") def home(): return "Hello World" app.run() If needed I'll try to do a tuto on this topic in future. Hope this helps:)
@@AboniaSojasingarayar thank you Have a nice day~
Hi, its possible integrate DynamoDB for store and retrive context of last user prompts in lambda function?
Hello, Yes , DynamoDB, S3, or in-memory storage depending on requirements. Each piece of context is associated with a user ID, ensuring that contexts are isolated per user with conversation ID. Hope this helps.
@@AboniaSojasingarayar Thanks, I'll try it and let you know how it goes.
please create one video for breast cancer detection in yolov10 model
Absolutely, I’ll work on getting it ready shortly. If there are specific areas you want me to concentrate on, just let me know! Also, do you have any custom dataset you'd like to use for this tutorial? Thanks
Super
Glad it helped
Great discussion....
Thank you Venkatesan. I'm glad you enjoyed the discussion.
Hi I'm trying to run these processes, but in this video 12:36 how to create and execute the file named ".env" , it always show Error , I can't figure it out. Thanks!
Hello, You can use local VScode or any IDE to create .env New file -> name it as .env And add your API key as follows: ROBOFLOW_API_KEY=your_api_key Once done drag and drop it in colab. Hope this helps.
Great share, insightful share as always...Are u using obs studio for recording....by Senior Data Scientist....
Glad it helped. Not really! Just using the built-in recording and iMovie to edit it.
Hey, Abonia..Thanks for the amazing content. I just had one issue though: on executing the 'map_reduce_outputs' function, I had the ConnectionRefusedError: [Errno 61]. Hope someone know what it is
@@alvaroaraujo7945 Hello , thanks for your kind words. It may be related to your ollama serve.Are you sure Ollama is running ?
Nice explanation and walkthrough. Could you provide the link to the code repo for this exercise.
Glad it helped. As mentioned in the description, you can find the code and explanation in this article walkthrough. medium.com/@abonia/deploying-a-rag-application-in-aws-lambda-using-docker-and-ecr-08e246a7c515
It is very helpful mam , it is useful on impliying
Nice video mam
Can we simply rely on open source only without using Amazon? What if it is just prototyping?
Yes, we can use open source completely.
Hi, Thank you for the video, So if we want to fine tune the model and evaluate it for videos, then how to do it ?
Your most welcome. Here I have introduced basic usage of SAM 2 models. If you want to evaluate your finetuned model you may try mean IoU score, for a set of predictions and targets or DICE, precision, recall, and mAP.
streaming does't work via doing this. I wrote code from scratch without langchain.
@@Basant5911 can you share your code base and error or issue that you are facing currently please?
❤Thank you for this fantastic educational video on my book!!! 🎉
@@DenisRothman Thank you for your kind words. I'm grateful for the opportunity to review the book and share my thoughts. Your work is well-deserved and truly one of the most insightful books I've read.
You used a webpage as a data source for the RAG app, what If I add pdf file instead of the webpage as a data source, how can I deploy it in aws lambda?
To build RAG with pdf in AWS ecosystem, you need to follow steps that involve uploading the PDF to an S3 bucket, extracting text from the PDF, and then integrating this data with your RAG application.
@@AboniaSojasingarayar Can I locally extract text from pdf and build vector DB locally using vscode and then build the docker image and push it to ECR AWS like what you did in the video?
@@MohamedMohamed-xf7wh Yes, you can locally extract text from PDF files, build a vector database and then prepare your application for deployment on AWS Lambda by building a Docker image and pushing it to ECR. But which vector db are you using? It can be accessible with API?
@@AboniaSojasingarayar FAISS .. what is the problem with vector db?
@@MohamedMohamed-xf7wh Great!
Really Nice! Keep going. You deserve more subscribers.
@@htayaung3812 Thank you so much for your support! I'm working to bring more tutorials.
Save my life to create lambda layers... I have been trying for days. TKS!
@@raulpradodantas9386 Glad to hear that! You most welcome.
Hi I'm trying to run my trained model with interface to webcam but getting error can you share any insight on it
@@SidSid-kp4ij Hello Sid, Sure can you post your error message here please?
All the best
Thanks for the video. Very useful for me as I am new to AWS lambda and bedrock. Can you please upload the lambda function source code? Thanks again!
Glad it helped. Sure you can find the code and complete the article on this topic in the description. In any way here is the link to the code : medium.com/@abonia/build-and-deploy-llm-application-in-aws-cca46c662749
Which version of BERT is it used in BERTScore ?
As we are using lang= "en" so it uses roberta-large. We can also customize it using the model_type param of BERScorer class For default model for other languages,find it here: github.com/Tiiiger/bert_score/blob/master/bert_score/utils.py
hey great content. please continue to do more videos and real time projects. Thanks
Glad it helped. Sure I am already on it.
Awesome mam , very easy to understand
Hi Abonia, thanks for sharing. I am facing this error . can you please tell how to resolve it "errorMessage": "Unable to import module 'lambda_function': No module named 'langchain_community'",
Hello, You are most welcome. You must prepare your ZIP file with all the necessary packages. You can refer to the instructions starting at the 09:04
Hi Abonia, thanks for the thorough guide, but i'm abit confused with the lambda_layer.zip file, why did you have to create it through docker? is there an easier way to provide the dependencies in a zip file without going through docker? Thanks in advance!
Hi Humayoun Khan, Yes we can but Docker facilitates the inclusion of the runtime interface client for Python, making the image compatible with AWS Lambda. Also it ensures a consistent and reproducible environment for Lambda function's dependencies. This is crucial for avoiding discrepancies between development, testing, and production environments. Hope this helps.
hello! Thanks for the video. I am from Brazil. What would you recommend for large documents, averaging 150 pages? I tried map-reduce, but the inference time was 40 minutes. Are there any tips for these very long documents?
Thanks for you kind words and glad this helped. Implement a strategy that combines semantic chunking with K-means clustering to address the model’s contextual limitations. By employing efficient clustering techniques, we can extract key passages effectively, thereby reducing the overhead associated with processing large volumes of text. This approach not only significantly lowers costs by minimizing the number of tokens processed but also mitigates the recency and primacy effects inherent in LLMs, ensuring a balanced consideration of all text segments.
@@AboniaSojasingarayar Video was great and very useful.. can you make the small video on this clustering method using embedding ?
@@VirtualMachine-d8x Sure will do, happy to hear from you again. Thanks for the feedback.
Did you use OpenAI API key here?
Here we use open-source Mixtral from ollama.But, yes we can use OpenAI models as well.
Very informatics nd Your voice very clear dr
Glad it helped!
Can we use openai and chromadb on aws??
Yes we can! In the below tutorial I have demonstrated how we can create and deploy lambda layer via container for larger dependencies : ua-cam.com/video/gicsb9p7uj4/v-deo.htmlsi=F_X7-6YCAb0Kz3Jc
@@AboniaSojasingarayar yes but can this be done without eks or containers?
Yes! You can try it by creating a custom lambda layer.If you face issue try to use only the required libraries and remove any unnecessary dependencies from your zip file.Hope this helps.
In the abstractive summarization use-case, usually a lot of focus is given to the LLMs being used and its performance. Limitations of LLM including context length and ways to overcome this issue are often overlooked. Its important to make sure that our application is scalable when dealing with large document sizes. Thank you for this great and insightful video.
Thank you Vijay Gandhi, for your insightful comment! You've raised an excellent point about the importance of considering the limitations of LLMs in the context of abstractive summarization, especially regarding their context length and scalability issues when dealing with large documents. Indeed, one of the significant challenges in using LLMs for abstractive summarization is their inherent limitation in processing long texts due to the maximum token limit imposed by these models. This constraint can be particularly problematic when summarizing lengthy documents or articles, where the full context might not fit within the model's capacity.