https://github.com/dudeperf3ct/11-cortex-deploy
https://github.com/dudeperf3ct/11-cortex-deploy
aws-lambda cortex docker fastapi mlops model-serving transformers
Last synced: 3 months ago
JSON representation
- Host: GitHub
- URL: https://github.com/dudeperf3ct/11-cortex-deploy
- Owner: dudeperf3ct
- Created: 2022-01-14T07:28:56.000Z (over 3 years ago)
- Default Branch: main
- Last Pushed: 2022-01-15T17:16:04.000Z (over 3 years ago)
- Last Synced: 2024-12-31T10:17:49.854Z (5 months ago)
- Topics: aws-lambda, cortex, docker, fastapi, mlops, model-serving, transformers
- Language: Python
- Homepage:
- Size: 28.3 KB
- Stars: 1
- Watchers: 2
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: Readme.md
Awesome Lists containing this project
README
# Cortex
[Cortex](https://www.cortex.dev) is built to deploy, manage, and scale machine learning models in production for AWS. It provides features such as
- Serverless workloads
- Automated cluster management
- CI/CD and observability integrationsCortex supports 4 different ways to build scalable API :
1. [Realtime](https://docs.cortex.dev/workloads/realtime/example): create APIs that respond to requests in real-time.
2. [Async](https://docs.cortex.dev/workloads/async/example): create APIs that respond to requests asynchronously.
3. [Batch](https://docs.cortex.dev/workloads/batch/example): create APIs that run distributed batch jobs.
4. [Task](https://docs.cortex.dev/workloads/task/example): create APIs that run jobs on-demand.Cortex requires only two configuration file to deploy the application. Cortex creates a cluster from `cluster.yaml` file including a s3 bucket and cloudwatch log group. The Cortex cluster runs on an EKS (Kubernetes) cluster in a dedicated VPC on your AWS account. Each individual API contains `cortex.yaml` to deploy different types of workloads.
In this exercise, transformers sentiment classifier application is deployed using Cortex two different APIs.
1. [Realtime API](realtime/Readme.md)
2. [Async API](async/Readme.md)Cortex is super :rocket: With just 2 commands, 2 configuration files and right amount of patience, the application is deployed seamlessly without modifications to the application.
Further Readings:
We just barely scratch the surface by deploying a simple application. Cortex provides different features such `Traffic Splitter`, `Autoscaling`, `Update strategy`, `node groups`, etc. Cortex [cli](https://docs.cortex.dev/clients/cli) and [python client](https://docs.cortex.dev/clients/python) provides easy way to handle cortex updates/upgrades to deployments.
- Cortex Documentation : [Docs](https://docs.cortex.dev/)