You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/blog/articles/2022-07-21-KServe-0.9-release.md
+10-10Lines changed: 10 additions & 10 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -17,8 +17,8 @@ KServe has the unique strength to build the distributed inference graph with its
17
17
The graph router is deployed behind an HTTP endpoint and can be scaled dynamically based on request volume. The InferenceGraph supports four different types of routing nodes: **Sequence**, **Switch**, **Ensemble**, **Splitter**.
18
18
19
19
20
-
-**Sequence Node**: It allows users to define multiple Steps with InferenceServices or Nodes as routing targets in a sequence. The Steps are executed in sequence and the request/response from the previous step and be passed to the next step as input based on configuration.
21
-
-**Switch Node**: It allows users to define routing conditions and select a step to execute if it matches the condition. The response is returned as soon as it finds the first step that matches the condition. If no condition is matched, the graph returns the original request.
20
+
-**Sequence Node**: It allows users to define multiple `Steps` with `InferenceServices` or `Nodes` as routing targets in a sequence. The `Steps` are executed in sequence and the request/response from the previous step and be passed to the next step as input based on configuration.
21
+
-**Switch Node**: It allows users to define routing conditions and select a `Step` to execute if it matches the condition. The response is returned as soon as it finds the first step that matches the condition. If no condition is matched, the graph returns the original request.
22
22
-**Ensemble Node**: A model ensemble requires scoring each model separately and then combines the results into a single prediction response. You can then use different combination methods to produce the final result. Multiple classification trees, for example, are commonly combined using a "majority vote" method. Multiple regression trees are often combined using various averaging techniques.
23
23
-**Splitter Node**: It allows users to split the traffic to multiple targets using a weighted distribution.
24
24
@@ -63,7 +63,8 @@ spec:
63
63
data: $request
64
64
condition: "[@this].#(predictions.0==\"dog\")"
65
65
```
66
-
Currently the `Serverless` deployment mode is supported with inference graphs. You can try it out following the [tutorial](https://kserve.github.io/website/master/modelserving/inference_graph/image_pipeline/) here.
66
+
67
+
Currently `InferenceGraph` is supported with the `Serverless` deployment mode. You can try it out following the [tutorial](https://kserve.github.io/website/master/modelserving/inference_graph/image_pipeline/).
67
68
68
69
69
70
## InferenceService API for ModelMesh
@@ -83,7 +84,7 @@ storage:
83
84
parameters: # Parameters to override the default values inside the common secret.
84
85
bucket: example-models
85
86
```
86
-
Learn more [here](https://kserve.github.io/website/master/modelserving/inference_graph/image_pipeline/).
87
+
Learn more [here](https://github.com/kserve/kserve/tree/release-0.9/docs/samples/storage/storageSpec).
87
88
88
89
89
90
@@ -122,10 +123,10 @@ spec:
122
123
123
124
## Other New Features:
124
125
125
-
- Support serving MLFlow model via MLServer serving runtime.
126
-
- Support unified autoscaling target and metric fields for InferenceService components with both Serverless and RawDeployment mode.
127
-
- Support InferenceService ingress class and url domain template configuration for RawDeployment mode.
128
-
- ModelMesh now has a default OpenVINO Model Server ServingRuntime.
126
+
- Support [serving MLFlow model format](https://kserve.github.io/website/0.9/modelserving/v1beta1/mlflow/v2/) via MLServer serving runtime.
127
+
- Support [unified autoscaling target and metric fields](https://kserve.github.io/website/0.9/modelserving/autoscaling/autoscaling/) for InferenceService components with both Serverless and RawDeployment mode.
128
+
- Support [InferenceService ingress class and url domain template configuration](https://kserve.github.io/website/0.9/admin/kubernetes_deployment/) for RawDeployment mode.
129
+
- ModelMesh now has a default [OpenVINO Model Server](https://github.com/openvinotoolkit/model_server) ServingRuntime.
129
130
130
131
131
132
## What’s Changed?
@@ -136,14 +137,13 @@ spec:
136
137
- Update MLServer serving runtime to 1.0.0
137
138
138
139
139
-
140
140
## Join the community
141
141
142
142
- Visit our [Website](https://kserve.github.io/website/) or [GitHub](https://github.com/kserve)
143
143
- Join the Slack ([#kserve](https://kubeflow.slack.com/join/shared_invite/zt-n73pfj05-l206djXlXk5qdQKs4o1Zkg#/))
144
144
- Attend our community meeting by subscribing to the [KServe calendar](https://wiki.lfaidata.foundation/display/kserve/calendars).
145
145
- View our [community github repository](https://github.com/kserve/community) to learn how to make contributions. We are excited to work with you to make KServe better and promote its adoption!
146
146
147
-
Thank you for using or checking out KServe!
147
+
Thank you for contributing or checking out KServe!
0 commit comments