Building Dynamo (dynamo build) — Dynamo
Title: Building Dynamo (dynamo build) — Dynamo
URL Source: https://docs.nvidia.com/dynamo/archive/0.2.1/guides/dynamo_build.html?userAgent=PromptingBot%2F1.0.0
Published Time: Thu, 05 Jun 2025 21:23:32 GMT
Markdown Content: Skip to main content
Back to top- [x] - [x]
Ctrl+K
Search Ctrl+K
Search Ctrl+K
Table of Contents
Architecture & Features
Dynamo Command Line Interface
- CLI Overview
- Running Dynamo (dynamo run)
- Serving Inference Graphs (dynamo serve)
- Building Dynamo (dynamo build)
- Deploying Inference Graphs (dynamo deploy)
Usage Guides
- Writing Python Workers in Dynamo
- Disaggregation and Performance Tuning
- KV Cache Router Performance Tuning
- Planner Benchmark Example
Deployment Guides
- Dynamo Cloud Kubernetes Platform
- Deploying Dynamo Inference Graphs to Kubernetes using the Dynamo Cloud Platform
- Manual Helm Deployment
- Minikube Setup Guide
API
Examples
-
Building...
Building Dynamo (dynamo build)#
This guide explains how to use the dynamo build command to containerize Dynamo inference graphs (pipelines) for deployment.
dynamo build is a command-line tool that helps containerize inference graphs created with Dynamo SDK. Run dynamo build --containerize to build a stand-alone Docker container that encapsulates your entire inference graph. This image can then be shared and run standalone.
Note
This experimental feature is tested on the examples in the examples/ directory. You need to make some modifications. Pay particular attention if your inference graph introduces custom dependencies.
Building a containerized inference graph#
The basic workflow for using dynamo build includes:
#. Defining your inference graph and testing locally with dynamo serve #. Specifying a base image for your inference graph. More on this below. #. Running dynamo build to build a containerized inference graph
Basic Usage#
dynamo build <graph_definition> --containerize
previous Serving Inference Graphs (dynamo serve)next Deploying Inference Graphs to Kubernetes (dynamo deploy)
On this page
Privacy Policy | Manage My Privacy | Do Not Sell or Share My Data | Terms of Service | Accessibility | Corporate Policies | Product Security | Contact
Copyright © 2025-2025, NVIDIA Corporation.
Links/Buttons:
- Skip to main content
- document.write(
<img src="../_static/nvidia-logo-horiz-rgb-wht-for-screen.svg" class="logo__image only-dark" alt="Dynamo - Home"/>); Dynamo - GitHub
- Support Matrix
- Getting Started
- High Level Architecture
- Distributed Runtime
- Disaggregated Serving
- KV Block Manager
- Motivation
- KVBM Architecture
- Understanding KVBM components
- KVBM Further Reading
- KV Cache Routing
- Planner
- CLI Overview
- Running Dynamo (dynamo run)
- Serving Inference Graphs (dynamo serve)
- Building Dynamo (dynamo build)
- Deploying Inference Graphs (dynamo deploy)
- Writing Python Workers in Dynamo
- Disaggregation and Performance Tuning
- KV Cache Router Performance Tuning
- Planner Benchmark Example
- Dynamo Cloud Kubernetes Platform
- Deploying Dynamo Inference Graphs to Kubernetes using the Dynamo Cloud Platform
- Manual Helm Deployment
- Minikube Setup Guide
- Dynamo SDK
- Python API
- Hello World Example
- LLM Deployment Examples
- Multinode Examples
- LLM Deployment Examples using TensorRT-LLM
- #
- Privacy Policy
- Manage My Privacy
- Do Not Sell or Share My Data
- Terms of Service
- Accessibility
- Corporate Policies
- Product Security
- Contact