|
| 1 | +--- |
| 2 | +title: Supporting GPU provisioning and orchestration on Nebius |
| 3 | +date: 2025-04-11 |
| 4 | +description: "TBA" |
| 5 | +slug: nebius.md |
| 6 | +image: https://github.com/dstackai/static-assets/blob/main/static-assets/images/dstack-nebius-v2.png?raw=true |
| 7 | +categories: |
| 8 | + - Cloud fleets |
| 9 | + - NVIDIA |
| 10 | +--- |
| 11 | + |
| 12 | +# Supporting GPU provisioning and orchestration on Nebius |
| 13 | + |
| 14 | +As demand for GPU compute continues to scale, open-source tools tailored for AI workloads are becoming critical to |
| 15 | +developer velocity and efficiency. |
| 16 | +`dstack` is an open-source orchestrator purpose-built for AI infrastructure—offering a lightweight, container-native |
| 17 | +alternative to Kubernetes and Slurm. |
| 18 | + |
| 19 | +<img src="https://github.com/dstackai/static-assets/blob/main/static-assets/images/dstack-nebius-v2.png?raw=true" width="630"/> |
| 20 | + |
| 21 | +Today, we’re announcing native integration with [Nebius :material-arrow-top-right-thin:{ .external }](https://nebius.com/){:target="_blank"}, |
| 22 | +offering a streamlined developer experience for teams using GPUs for AI workloads. |
| 23 | +<!-- more --> |
| 24 | + |
| 25 | +## About Nebius |
| 26 | + |
| 27 | +Nebius provides cloud GPUs, |
| 28 | +offering high-performance clusters at competitive prices. This pricing is achieved through custom-designed hardware, |
| 29 | +partnerships with Original Design Manufacturers (ODMs), and infrastructure team expertise. |
| 30 | + |
| 31 | +Nebius offers various NVIDIA GPUs, including the L40S, H100, H200, GB200, NVL72, and B200 models, available on-demand |
| 32 | +and through reserved instances. Their data centers are located across Europe, with planned expansions into the US. |
| 33 | + |
| 34 | +## Why dstack |
| 35 | + |
| 36 | +Kubernetes offers flexibility, but its complexity is often unnecessary—especially for use cases like interactive |
| 37 | +development or multi-stage training. |
| 38 | +Slurm is excellent for batch scheduling but lacks native support for dev environments, real-time inference, and |
| 39 | +multi-user orchestration. |
| 40 | + |
| 41 | +`dstack` fills the gap: a developer-friendly platform with native GPU support across dev environments, tasks, and |
| 42 | +long-running services—without the operational overhead. |
| 43 | + |
| 44 | +## Getting started |
| 45 | + |
| 46 | +To use `dstack` with Nebius, configure your `nebius` backend: |
| 47 | + |
| 48 | +1. Log in to your [Nebius AI Cloud :material-arrow-top-right-thin:{ .external }](https://console.eu.nebius.com/){:target="_blank"} account. |
| 49 | +2. Navigate to `Access`, and select `Service Accounts`. |
| 50 | +3. Create a new service account, assign it to the `editors` group, and upload an authorized key. |
| 51 | + |
| 52 | +Then, configure the backend via `~/.dstack/server/config.yml`: |
| 53 | + |
| 54 | +<div editor-title="~/.dstack/server/config.yml"> |
| 55 | + |
| 56 | +```yaml |
| 57 | +projects: |
| 58 | + - name: main |
| 59 | + backends: |
| 60 | + - type: nebius |
| 61 | + creds: |
| 62 | + type: service_account |
| 63 | + service_account_id: serviceaccount-e002dwnbz81sbvg2bs |
| 64 | + public_key_id: publickey-e00fciu5rkoteyzo69 |
| 65 | + private_key_file: ~/path/to/key.pem |
| 66 | +``` |
| 67 | +
|
| 68 | +</div> |
| 69 | +
|
| 70 | +Now, proceed with installing and starting the `dstack` server: |
| 71 | + |
| 72 | +<div class="termy"> |
| 73 | + |
| 74 | +```shell |
| 75 | +$ pip install "dstack[nebius]" |
| 76 | +$ dstack server |
| 77 | +``` |
| 78 | + |
| 79 | +</div> |
| 80 | + |
| 81 | +For more details, refer to [Installation](../../docs/installation/index.md). |
| 82 | + |
| 83 | +Use the `dstack` CLI to |
| 84 | +manage [dev environments](../../docs/concepts/dev-environments.md), [tasks](../../docs/concepts/tasks.md), |
| 85 | +and [services](../../docs/concepts/services.md). |
| 86 | + |
| 87 | +<div class="termy"> |
| 88 | + |
| 89 | +```shell |
| 90 | +$ dstack apply -f .dstack.yml |
| 91 | +
|
| 92 | + # BACKEND REGION RESOURCES SPOT PRICE |
| 93 | + 1 nebius eu-north1 8xCPU, 32GB, 1xL40S (48GB) no $1.5484 |
| 94 | + 2 nebius eu-north1 16xCPU, 200GB, 1xH100 (80GB) no $2.95 |
| 95 | + 3 nebius eu-north1 16xCPU, 200GB, 1xH200 (141GB) no $3.5 |
| 96 | + ... |
| 97 | + Shown 3 of 7 offers, $28 max |
| 98 | + |
| 99 | + Override the run? [y/n]: |
| 100 | +``` |
| 101 | + |
| 102 | +</div> |
| 103 | + |
| 104 | +The new `nebius` backend supports CPU and GPU instances, [fleets](../../docs/concepts/fleets.md), |
| 105 | +[distributed tasks](../../docs/concepts/tasks.md#distributed-tasks), and more. |
| 106 | + |
| 107 | +> Support for [network volumes](../../docs/concepts/volumes.md#network-volumes) and accelerated cluster |
| 108 | +interconnects is coming soon. |
| 109 | + |
| 110 | +!!! info "What's next?" |
| 111 | + 1. Check [Quickstart](../../docs/quickstart.md) |
| 112 | + 2. Sign up with [Nebius AI Cloud :material-arrow-top-right-thin:{ .external }](https://console.eu.nebius.com/){:target="_blank"} |
| 113 | + 3. Read about [dev environments](../../docs/concepts/dev-environments.md), |
| 114 | + [tasks](../../docs/concepts/tasks.md), [services](../../docs/concepts/services.md), |
| 115 | + and [fleets](../../docs/concepts/fleets.md) |
| 116 | + 4. Join [Discord :material-arrow-top-right-thin:{ .external }](https://discord.gg/u8SmfwPpMd){:target="_blank"} |
0 commit comments