Skip to main content

Serverless Installation Guide

info

Serverless Deployment is recommended primarily for predictive inference workloads.

KServe Serverless installation enables autoscaling based on request volume and supports scale down to and from zero. It also supports revision management and canary rollout based on revisions.

Applicability for Predictive Inference

Serverless deployment is particularly well-suited for predictive inference workloads because:

  • Predictive inference typically has shorter response times that work well with Knative's concurrency model
  • CPU-based models can efficiently scale to zero when not in use
  • Knative's request-based scaling aligns with the traffic patterns of many predictive workloads
  • Canary deployments and revisions enable safe updates to predictive models

For generative inference workloads that typically require GPU resources and have longer processing times, the Raw Kubernetes Deployment approach is recommended.

Kubernetes 1.30 is the minimally required version and please check the following recommended Knative, Istio versions for the corresponding Kubernetes version.

Kubernetes VersionRecommended Istio VersionRecommended Knative Version
1.301.22,1.231.15,1.16
1.311.24,1.251.16,1.17
1.321.25,1.261.17,1.18

1. Install Knative Serving

Please refer to Knative Serving install guide.

tip

If you are looking to use PodSpec fields such as nodeSelector, affinity or tolerations which are now supported in the v1beta1 API spec, you need to turn on the corresponding feature flags in your Knative configuration.

warning

Knative 1.13.1 requires Istio 1.20+, gRPC routing does not work with previous Istio releases, see release notes.

2. Install Networking Layer

The recommended networking layer for KServe is Istio as currently it works best with KServe, please refer to the Istio install guide. Alternatively you can also choose other networking layers like Kourier or Contour, see how to install Kourier with KServe guide.

3. Install Cert Manager

The minimally required Cert Manager version is 1.15.0 and you can refer to Cert Manager.

note

Cert manager is required to provision webhook certs for production grade installation, alternatively you can run self signed certs generation script.

4. Install KServe

Install using Helm

Install KServe CRDs

helm install kserve-crd oci://ghcr.io/kserve/charts/kserve-crd --version v0.12.0

Install KServe Resources

helm install kserve oci://ghcr.io/kserve/charts/kserve --version v0.12.0

Install using YAML

Install KServe CRDs and Controller, --server-side option is required as the InferenceService CRD is large, see this issue for details.

kubectl apply --server-side -f https://github.com/kserve/kserve/releases/download/v0.12.0/kserve.yaml

Install KServe Built-in ClusterServingRuntimes

kubectl apply --server-side -f https://github.com/kserve/kserve/releases/download/v0.12.0/kserve-cluster-resources.yaml
note

ClusterServingRuntimes are required to create InferenceService for built-in model serving runtimes with KServe v0.8.0 or higher.