You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+13-7Lines changed: 13 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,8 @@
1
1
# Deploy machine learning models in production
2
2
3
-
Cortex is an open source platform that takes machine learning models—trained with nearly any framework—and turns them into production web APIs in one command. <br>
3
+
Cortex is an open source platform for deploying machine learning models—trained with nearly any framework—as production web services.
4
+
5
+
<br>
4
6
5
7
<!-- Set header Cache-Control=no-cache on the S3 object metadata (see https://help.github.com/en/articles/about-anonymized-image-urls) -->
status up-to-date available requested last update avg latency
83
-
live 1 1 1 8s 123ms
88
+
live 1 1 1 8s 24ms
84
89
85
90
class count
86
91
positive 8
@@ -91,7 +96,7 @@ negative 4
91
96
92
97
## How it works
93
98
94
-
The CLI sends configuration and code to the cluster every time you run `cortex deploy`. Each model is loaded into a Docker container, along with any Python packages and request handling code. The model is exposed as a web service using Elastic Load Balancing (ELB), Flask, TensorFlow Serving, and ONNX Runtime. The containers are orchestrated on Elastic Kubernetes Service (EKS) while logs and metrics are streamed to CloudWatch.
99
+
The CLI sends configuration and code to the cluster every time you run `cortex deploy`. Each model is loaded into a Docker container, along with any Python packages and request handling code. The model is exposed as a web service using Elastic Load Balancing (ELB), TensorFlow Serving, and ONNX Runtime. The containers are orchestrated on Elastic Kubernetes Service (EKS) while logs and metrics are streamed to CloudWatch.
95
100
96
101
<br>
97
102
@@ -101,4 +106,5 @@ The CLI sends configuration and code to the cluster every time you run `cortex d
101
106
-[Sentiment analysis](https://github.com/cortexlabs/cortex/tree/0.11/examples/tensorflow/sentiment-analysis) in TensorFlow with BERT
102
107
-[Image classification](https://github.com/cortexlabs/cortex/tree/0.11/examples/tensorflow/image-classifier) in TensorFlow with Inception
103
108
-[Text generation](https://github.com/cortexlabs/cortex/tree/0.11/examples/pytorch/text-generator) in PyTorch with DistilGPT2
104
-
-[Iris classification](https://github.com/cortexlabs/cortex/tree/0.11/examples/xgboost/iris-classifier) in XGBoost / ONNX
109
+
-[Reading comprehension](https://github.com/cortexlabs/cortex/tree/0.11/examples/pytorch/text-generator) in PyTorch with ELMo-BiDAF
110
+
-[Iris classification](https://github.com/cortexlabs/cortex/tree/0.11/examples/sklearn/iris-classifier) in scikit-learn
Copy file name to clipboardExpand all lines: examples/sklearn/iris-classifier/README.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -48,7 +48,7 @@ $ python3 trainer.py
48
48
49
49
<br>
50
50
51
-
## Define a predictor
51
+
## Implement a predictor
52
52
53
53
1. Create another Python file `predictor.py`.
54
54
2. Add code to load and initialize your pickled model.
@@ -100,7 +100,7 @@ You can skip dependencies that are [pre-installed](../../../docs/deployments/pre
100
100
101
101
<br>
102
102
103
-
## Define a deployment
103
+
## Configure a deployment
104
104
105
105
Create a `cortex.yaml` file and add the configuration below. A `deployment` specifies a set of resources that are deployed together. An `api` provides a runtime for inference and makes our `predictor.py` implementation available as a web service that can serve real-time predictions:
0 commit comments