Welcome to NVIDIA Dynamo — NVIDIA Dynamo Documentation

Last updated: 11/7/2025

Title: Welcome to NVIDIA Dynamo — NVIDIA Dynamo Documentation

URL Source: https://docs.nvidia.com/dynamo/latest/index.html

Published Time: Fri, 07 Nov 2025 17:51:53 GMT

Markdown Content: Welcome to NVIDIA Dynamo#

The NVIDIA Dynamo Platform is a high-performance, low-latency inference framework designed to serve all AI models—across any framework, architecture, or deployment scale.

💎 Discover the latest developments!

This guide is a snapshot of a specific point in time. For the latest information, examples, and Release Assets, see the Dynamo GitHub repository.

Quickstart#

Get started with Dynamo locally in just a few commands:

1. Install Dynamo

Install uv (recommended Python package manager)

curl -LsSf https://astral.sh/uv/install.sh | sh

Create virtual environment and install Dynamo

uv venv venv source venv/bin/activate

Use prerelease flag to install RC versions of flashinfer and/or other dependencies

uv pip install --prerelease=allow "ai-dynamo[sglang]" # or [vllm], [trtllm]

2. Start etcd/NATS

Fetch and start etcd and NATS using Docker Compose

VERSION=$(uv pip show ai-dynamo | grep Version | cut -d' ' -f2) curl -fsSL -o docker-compose.yml https://raw.githubusercontent.com/ai-dynamo/dynamo/refs/tags/v${VERSION}/deploy/docker-compose.yml docker compose -f docker-compose.yml up -d

3. Run Dynamo

Start the OpenAI compatible frontend (default port is 8000)

python -m dynamo.frontend

In another terminal, start an SGLang worker

python -m dynamo.sglang --model-path Qwen/Qwen3-0.6B

4. Test your deployment

curl localhost:8000/v1/chat/completions
-H "Content-Type: application/json"
-d '{"model": "Qwen/Qwen3-0.6B", "messages": [{"role": "user", "content": "Hello!"}], "max_tokens": 50}'

Links/Buttons: