More on Full Stack Data Science👇 👉Series Playlist: ru-vid.com/group/PLz-ep5RbHosWmAt-AMK0MBgh3GeSvbCmL 💻Example Code: github.com/ShawhinT/RU-vid-Blog/tree/main/full-stack-data-science/ml-engineering
One of the best aspects of AWS Elastic Cloud is how seamlessly everything comes together, whether you're using FastAPI or Docker. It's all integrated beautifully.
This is fantastic stuff as I’m pulling out my hair on this same step. You have the right idea for the next video, but I think the next one after that is making the chat interface publicly accessible
Great suggestion Brian! There are several ways one can do this. The simplest and cheapest would be hosting it via HuggingFace Spaces: huggingface.co/spaces/launch However, for this specific use case the most practical option would be to embed it into my Squarespace website. I'll need to do some more digging to see the best way to do that.
@@ShawhinTalebi The streamlit cloud has been my go-to so far. I am toying with creating a react front end template but would like to see what others are doing
Hey Shaw! what a beautiful content. I followed all the steps from ML APP from scratch till deployment but at last moment, I don;t have AWS free tier account as they still ask to enter the debit card details for having free access, so can you please tell me another way round to cope up with this ?
Glad you liked it! You could try deploying to railway (railway.app/). I just used them for a project and don't think I needed to input credit card info.
Great tutorial... Im still not able to connect to the API unfortunately (This site can’t be reached,refused to connect.) even though I followed the network config steps you explained...
To confirm, you added inbound rules in the VPC dashboard to allow all incoming traffic from your IP? Does the IP listed in the inbound rules match yours?
Good question. This should be similar to the Gradio example shown 25:54. This blog post might be helpful: blog.streamlit.io/create-a-search-engine-with-streamlit-and-google-sheets/
I have a DL model which takes about 5 mins and 3gb GPU to process the query and to return result. I need to handle 5 queries per minute and I have a GPU with 8gb in GCP. How can I deploy such a model without memory leakage and I should be able to use the GPU at its full potential?
How big is that model? Do you have GPU parallelization enabled? If it takes 5 min and 3GB to do one query with parallelization, the model may be too big to meet those technical constraints.
that was an awesome video, I have a task for one click ml model deployment on aws, azure and GCP, like one click on aws and other click on azure. CAn u please guide me shortly the roadmap...!
Thank you Shaw for making so many amazing videos. quick question from this video.. where exactly are you making a connection between your dockerhub and AWS ECS? Is it where you mention the url of the image? what if someone has a similar image name (shawhint/yt-search-demo) or that's not possible? Sorry if its a dumb question 😐