More on Full Stack Data Science👇 👉Series Playlist: ua-cam.com/play/PLz-ep5RbHosWmAt-AMK0MBgh3GeSvbCmL.html 💻Example Code: github.com/ShawhinT/UA-cam-Blog/tree/main/full-stack-data-science/ml-engineering
One of the best aspects of AWS Elastic Cloud is how seamlessly everything comes together, whether you're using FastAPI or Docker. It's all integrated beautifully.
This is fantastic stuff as I’m pulling out my hair on this same step. You have the right idea for the next video, but I think the next one after that is making the chat interface publicly accessible
Great suggestion Brian! There are several ways one can do this. The simplest and cheapest would be hosting it via HuggingFace Spaces: huggingface.co/spaces/launch However, for this specific use case the most practical option would be to embed it into my Squarespace website. I'll need to do some more digging to see the best way to do that.
@@ShawhinTalebi The streamlit cloud has been my go-to so far. I am toying with creating a react front end template but would like to see what others are doing
Hi! Great video! Thanks for sharing with us! Do you have any recommendation on what to study next to learn more about these topics or to go deeper in it?
Hey Shaw! what a beautiful content. I followed all the steps from ML APP from scratch till deployment but at last moment, I don;t have AWS free tier account as they still ask to enter the debit card details for having free access, so can you please tell me another way round to cope up with this ?
Glad you liked it! You could try deploying to railway (railway.app/). I just used them for a project and don't think I needed to input credit card info.
i have been following this video, and everything went perfectly fine, until the container deployment on AWS (i think because I am working with an LLM and therefore i need GPU) what should i do in this situation?
Good question. This should be similar to the Gradio example shown 25:54. This blog post might be helpful: blog.streamlit.io/create-a-search-engine-with-streamlit-and-google-sheets/
Great tutorial... Im still not able to connect to the API unfortunately (This site can’t be reached,refused to connect.) even though I followed the network config steps you explained...
To confirm, you added inbound rules in the VPC dashboard to allow all incoming traffic from your IP? Does the IP listed in the inbound rules match yours?
I have a DL model which takes about 5 mins and 3gb GPU to process the query and to return result. I need to handle 5 queries per minute and I have a GPU with 8gb in GCP. How can I deploy such a model without memory leakage and I should be able to use the GPU at its full potential?
How big is that model? Do you have GPU parallelization enabled? If it takes 5 min and 3GB to do one query with parallelization, the model may be too big to meet those technical constraints.
that was an awesome video, I have a task for one click ml model deployment on aws, azure and GCP, like one click on aws and other click on azure. CAn u please guide me shortly the roadmap...!
Thank you Shaw for making so many amazing videos. quick question from this video.. where exactly are you making a connection between your dockerhub and AWS ECS? Is it where you mention the url of the image? what if someone has a similar image name (shawhint/yt-search-demo) or that's not possible? Sorry if its a dumb question 😐
More on Full Stack Data Science👇
👉Series Playlist: ua-cam.com/play/PLz-ep5RbHosWmAt-AMK0MBgh3GeSvbCmL.html
💻Example Code: github.com/ShawhinT/UA-cam-Blog/tree/main/full-stack-data-science/ml-engineering
One of the best aspects of AWS Elastic Cloud is how seamlessly everything comes together, whether you're using FastAPI or Docker. It's all integrated beautifully.
One of the best video available. on the internet for deployment.
Glad it was helpful :)
This is such a great video, no nonsense straight to the point!
Fantastic video. Thank you for this!
It was amazing ...love this type of content❤
Super detailed explanation, thank you!
This was so simplified. thank you Shawhin
great video ! very informative . Thanks a lot :)
Hey Shawn, videos on FastAPI and Docker from you would be great.
Thanks for the suggestion! I'll add it to my queue :)
I like your video, it deserves much more views
This is fantastic stuff as I’m pulling out my hair on this same step.
You have the right idea for the next video, but I think the next one after that is making the chat interface publicly accessible
Great suggestion Brian! There are several ways one can do this. The simplest and cheapest would be hosting it via HuggingFace Spaces: huggingface.co/spaces/launch
However, for this specific use case the most practical option would be to embed it into my Squarespace website. I'll need to do some more digging to see the best way to do that.
@@ShawhinTalebi The streamlit cloud has been my go-to so far. I am toying with creating a react front end template but would like to see what others are doing
Hi! Great video! Thanks for sharing with us! Do you have any recommendation on what to study next to learn more about these topics or to go deeper in it?
Happy to give recs. What's your end goal?
Thank you so much, such videos are really very helpful
Glad to hear :)
well done , Thanks !
Hey Shaw!
what a beautiful content. I followed all the steps from ML APP from scratch till deployment but at last moment, I don;t have AWS free tier account as they still ask to enter the debit card details for having free access, so can you please tell me another way round to cope up with this ?
Glad you liked it! You could try deploying to railway (railway.app/). I just used them for a project and don't think I needed to input credit card info.
i have been following this video, and everything went perfectly fine, until the container deployment on AWS (i think because I am working with an LLM and therefore i need GPU)
what should i do in this situation?
Self-hosting LLMs can get tricky. I'd recommend using an API like OpenAI, TogetherAI, or the like if possible.
how can we integrate streamlit to make the UI to get input and send to model and display the output here ?
Good question. This should be similar to the Gradio example shown 25:54.
This blog post might be helpful: blog.streamlit.io/create-a-search-engine-with-streamlit-and-google-sheets/
Thank you!
Great tutorial... Im still not able to connect to the API unfortunately (This site can’t be reached,refused to connect.) even though I followed the network config steps you explained...
To confirm, you added inbound rules in the VPC dashboard to allow all incoming traffic from your IP?
Does the IP listed in the inbound rules match yours?
I have a DL model which takes about 5 mins and 3gb GPU to process the query and to return result. I need to handle 5 queries per minute and I have a GPU with 8gb in GCP. How can I deploy such a model without memory leakage and I should be able to use the GPU at its full potential?
How big is that model? Do you have GPU parallelization enabled?
If it takes 5 min and 3GB to do one query with parallelization, the model may be too big to meet those technical constraints.
where does K8S fit in in this pipeline?
Good question. In my experience, Kubernetes is rarely used in DS/ML, so I wouldn't worry learning it if your just getting started.
Bro create a video for handling post and get request and multiple endpoints using fast api dockerize and ECR and aws lambda functions
Great suggestion. I'll add that to my list!
that was an awesome video, I have a task for one click ml model deployment on aws, azure and GCP, like one click on aws and other click on azure. CAn u please guide me shortly the roadmap...!
Thanks for your comment. Sorry I'm not sure what you mean by "one click ml model deployment". Could you share more details?
Thank you Shaw for making so many amazing videos. quick question from this video.. where exactly are you making a connection between your dockerhub and AWS ECS? Is it where you mention the url of the image? what if someone has a similar image name (shawhint/yt-search-demo) or that's not possible? Sorry if its a dumb question 😐
Good question! Yes, exactly. No one will have a similar image name because the first part will be your unique DockerHub username.