4.9 • 848 Ratings
🗓️ 18 January 2021
⏱️ 53 minutes
🔗️ Recording | iTunes | RSS
🧾️ Download transcript
Machine learning model deployment on the cloud is typically handled with solutions like AWS SageMaker for end-to-end training and inference as a REST endpoint, AWS Batch for cost-effective on-demand batch jobs using Docker containers, and AWS Lambda for low-usage, serverless inference without GPU support. Storage and infrastructure options such as AWS EFS are essential for managing large model artifacts, while new tools like Cortex offer open source alternatives with features like cost savings and scale-to-zero for resource management.
Click on a timestamp to play from that location
0:00.0 | You're listening to Machine Learning Applied, and in this episode, we're going to talk about |
0:03.9 | machine learning hosting solutions out there with a special emphasis on serverless technology. |
0:10.0 | In the last Machine Learning Applied episode, we talked about various tech stack solutions, |
0:14.8 | especially for serverless solutions for hosting your server. |
0:18.9 | The Amplify Stack in particular uses behind the hood |
0:22.9 | AWS Lambda. And AWS Lambda is a very popular solution for serverless architectures. |
0:29.0 | AWS Lambda lets you write a single code function, whether it's Node.js or Python, and then |
0:34.4 | that function in conjunction with AWS API Gateway, will be exposed as a rest endpoint that your client can call. |
0:42.5 | Very powerful and saves you a lot of time and heartache. |
0:45.4 | Now, we want to achieve this level of functionality when deploying our machine learning models. |
0:50.9 | So in this episode, when I'm talking about machine learning deployment solutions, |
0:55.3 | I want to push towards as much of a serverless solution for deploying your machine learning |
1:00.1 | models as possible. Now, I want to preface by saying my own experience in machine learning |
1:04.7 | hosting is slightly unlimited, and in my experience trying to find the best serverless machine learning hosting technology |
1:12.6 | available, I've found that the providers are somewhat limited. |
1:16.1 | And this makes sense. |
1:17.5 | We've been deploying regular servers, app servers, using Fast API and Node.js for a very |
1:23.5 | long time. |
1:24.1 | And so hosting providers like AWS and GCP have had a lot of time to sit |
1:29.2 | with this problem in order to come up with solutions like AWS Lambda, where machine learning |
1:34.9 | is a more recently popular technology. And so they haven't had as much time to really perfect |
1:41.1 | this space. So you will find that even within a single cloud hosting provider |
... |
Please login to see the full transcript.
Disclaimer: The podcast and artwork embedded on this page are from OCDevel, and are the property of its owner and not affiliated with or endorsed by Tapesearch.
Generated transcripts are the property of OCDevel and are distributed freely under the Fair Use doctrine. Transcripts generated by Tapesearch are not guaranteed to be accurate.
Copyright © Tapesearch 2025.