mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-04-13 03:27:03 +08:00
[Doc] add KubeAI to serving integrations (#10837)
Signed-off-by: Sam Stoelinga <sammiestoel@gmail.com>
This commit is contained in:
parent
8b59631855
commit
7406274041
17
docs/source/serving/deploying_with_kubeai.rst
Normal file
17
docs/source/serving/deploying_with_kubeai.rst
Normal file
@ -0,0 +1,17 @@
|
||||
.. _deploying_with_kubeai:
|
||||
|
||||
Deploying with KubeAI
|
||||
=====================
|
||||
|
||||
`KubeAI <https://github.com/substratusai/kubeai>`_ is a Kubernetes operator that enables you to deploy and manage AI models on Kubernetes. It provides a simple and scalable way to deploy vLLM in production. Functionality such as scale-from-zero, load based autoscaling, model caching, and much more is provided out of the box with zero external dependencies.
|
||||
|
||||
|
||||
Please see the Installation Guides for environment specific instructions:
|
||||
|
||||
* `Any Kubernetes Cluster <https://www.kubeai.org/installation/any/>`_
|
||||
* `EKS <https://www.kubeai.org/installation/eks/>`_
|
||||
* `GKE <https://www.kubeai.org/installation/gke/>`_
|
||||
|
||||
Once you have KubeAI installed, you can
|
||||
`configure text generation models <https://www.kubeai.org/how-to/configure-text-generation-models/>`_
|
||||
using vLLM.
|
||||
@ -6,6 +6,7 @@ Integrations
|
||||
|
||||
run_on_sky
|
||||
deploying_with_kserve
|
||||
deploying_with_kubeai
|
||||
deploying_with_triton
|
||||
deploying_with_bentoml
|
||||
deploying_with_cerebrium
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user