Deploying Dynamo Inference Graphs to Kubernetes — Dynamo
Title: Deploying Dynamo Inference Graphs to Kubernetes#
URL Source: https://docs.nvidia.com/dynamo/archive/0.2.0/guides/dynamo_deploy/README.html?userAgent=PromptingBot%2F1.0.0
Published Time: Fri, 23 May 2025 15:10:50 GMT
Markdown Content: This guide provides an overview of the different deployment options available for Dynamo inference graphs in Kubernetes environments.
Deployment Options#
Dynamo provides two distinct deployment paths, each serving different use cases:
1. 🚀 Dynamo Cloud Kubernetes Platform [PREFERRED]#
The Dynamo Cloud Platform (deploy/dynamo/helm/) provides a managed deployment experience:
-
Contains the infrastructure components required for the Dynamo cloud platform
-
Used when deploying with the
dynamo deployCLI commands -
Provides a managed deployment experience
For detailed instructions on using the Dynamo Cloud Platform, see:
-
Dynamo Cloud Platform Guide: walks through installing and configuring the Dynamo cloud components on your Kubernetes cluster.
2. Manual Deployment with Helm Charts#
The manual deployment path (deploy/Kubernetes/) is available for users who need more control over their deployments:
-
Used for manually deploying inference graphs to Kubernetes
-
Contains Helm charts and configurations for deploying individual inference pipelines
-
Provides full control over deployment parameters
-
Requires manual management of infrastructure components
-
Documentation:
-
Manual Helm Deployment Guide: detailed instructions on manual deployment
-
[Deploying Dynamo Inference Graphs to Kubernetes using Helm](manual_helm_deployment.md#Deploying Dynamo Inference Graphs to Kubernetes using Helm): all-in-one script
-
Getting Started#
- For Dynamo Cloud Platform:
* Follow the [Dynamo Cloud Platform Guide](https://docs.nvidia.com/dynamo/archive/0.2.0/guides/dynamo_deploy/dynamo_cloud.html.md)
* Deploy a Hello World pipeline using the [Operator Deployment Guide](https://docs.nvidia.com/dynamo/archive/0.2.0/guides/dynamo_deploy/operator_deployment.html.md)
* Deploy a Dynamo LLM pipeline to Kubernetes [Deploy LLM Guide](https://docs.nvidia.com/dynamo/archive/0.2.0/examples/llm_deployment.html.md#deploy-to-kubernetes)
2. For Manual Deployment:
* Follow the [Manual Helm Deployment Guide](https://docs.nvidia.com/dynamo/archive/0.2.0/guides/dynamo_deploy/manual_helm_deployment.html.md)
Example Deployment#
See the [Hello World example](…/…/examples/hello_world.md#Deploying to and Running the Example in Kubernetes) for a complete walkthrough of deploying a simple inference graph.
See the LLM example for a complete walkthrough of deploying a production-ready LLM inference pipeline to Kubernetes.
Links/Buttons:
- Skip to main content
- document.write(
<img src="../../_static/nvidia-logo-horiz-rgb-wht-for-screen.svg" class="logo__image only-dark" alt="Dynamo - Home"/>); Dynamo - GitHub
- Quick Start
- Support Matrix
- High Level Architecture
- Disaggregated Serving
- KV Cache Manager
- KV Cache Routing
- SDK
- Python
- Dynamo CLI Overview
- dynamo run
- dynamo serve
- dynamo deploy
- Hello World
- LLM Deployment
- Multinode
- #
- Dynamo Cloud Platform Guide
- Operator Deployment Guide
- Manual Helm Deployment Guide
- Deploy LLM Guide
- Privacy Policy
- Manage My Privacy
- Do Not Sell or Share My Data
- Terms of Service
- Accessibility
- Corporate Policies
- Product Security
- Contact