Welcome to NVIDIA Dynamo — NVIDIA Dynamo Documentation
Title: Welcome to NVIDIA Dynamo — NVIDIA Dynamo Documentation
URL Source: https://docs.nvidia.com/dynamo/latest/index.html
Published Time: Fri, 07 Nov 2025 17:51:53 GMT
Markdown Content: Welcome to NVIDIA Dynamo#
The NVIDIA Dynamo Platform is a high-performance, low-latency inference framework designed to serve all AI models—across any framework, architecture, or deployment scale.
💎 Discover the latest developments!
This guide is a snapshot of a specific point in time. For the latest information, examples, and Release Assets, see the Dynamo GitHub repository.
Quickstart#
Get started with Dynamo locally in just a few commands:
1. Install Dynamo
Install uv (recommended Python package manager)
curl -LsSf https://astral.sh/uv/install.sh | sh
Create virtual environment and install Dynamo
uv venv venv source venv/bin/activate
Use prerelease flag to install RC versions of flashinfer and/or other dependencies
uv pip install --prerelease=allow "ai-dynamo[sglang]" # or [vllm], [trtllm]
2. Start etcd/NATS
Fetch and start etcd and NATS using Docker Compose
VERSION=$(uv pip show ai-dynamo | grep Version | cut -d' ' -f2) curl -fsSL -o docker-compose.yml https://raw.githubusercontent.com/ai-dynamo/dynamo/refs/tags/v${VERSION}/deploy/docker-compose.yml docker compose -f docker-compose.yml up -d
3. Run Dynamo
Start the OpenAI compatible frontend (default port is 8000)
python -m dynamo.frontend
In another terminal, start an SGLang worker
python -m dynamo.sglang --model-path Qwen/Qwen3-0.6B
4. Test your deployment
curl localhost:8000/v1/chat/completions
-H "Content-Type: application/json"
-d '{"model": "Qwen/Qwen3-0.6B",
"messages": [{"role": "user", "content": "Hello!"}],
"max_tokens": 50}'
Links/Buttons: