You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Data sources in Run:ai allow you to connect additional storage systems to your Run:ai projects, enabling seamless access to datasets required for your workloads. Run:ai supports various types of data sources, including PVC ([Persistent Volume Claim](https://kubernetes.io/docs/concepts/storage/persistent-volumes/)), NFS, S3-compatible storage, and more.
3
+
4
+
When a Run:ai project is created, a default data source (PVC) is automatically set up for this project. You can find the data source under the "Data Sources", listed as `pvc-<dashr_project_shortcode>`:
5
+
6
+

7
+
8
+
This dedicated persistent storage can be accessable from any workload running within the same project. The default mount path inside the container is `/scratch/<dashr_project_shortcode>`. It is especially useful for storing intermediate results, model checkpoints, and [uploading data to the cluster](./data_transfer.md) from external sources (*e.g.* RDS).
Define the necessary information for your workload:
19
20
20
21
* Under "Projects" select the project it will be linked to
21
-
* Under "Templates" select "Start from sratch" (do not use any existing template)
22
+
* Under "Templates" select "Start from sratch" (*i.e.* do not use any existing template)
23
+
* Provide a descriptive name for the workload
24
+
22
25

23
-
* Under "Environment" select the Jupyter Lab container environment you want to run
26
+
27
+
* Select an environment to create the container. The SIH team has prepared a [pre-built image](https://hub.docker.com/r/sydneyinformaticshub/dgx-interactive-jupyterlab) (`sydneyinformatics/dgx-interactive-jupyterlab`) with Jupyter Lab and commonly used data science packages installed.
* Under "Compute resource" select the resources required.
30
+
31
+
* Select the amount of compute resources to run the workload. In this tutorial, we will select the `small-fraction` option that requires 1 H200 GPU with 10% of its memory (~14GB).
There are other optional components you can add to a workload depending on the needs of your task. These include:
35
+
* Configure the [data source](./data_sources.md) to be mounted to the container. Here we select the default PVC created for the project. The mount path inside the container is set to `/scratch/<dashr_project_shortcode>`.
36
+
37
+

38
+
39
+
* Lastly, Click on "CREATE WORKLOAD" to submit the workload to the cluster.
40
+
41
+
## Step 3: Connect to Jupyter Lab
42
+
43
+
When the status changes to "Running", you can access the Jupyter Lab interface by selecting "Jupyter" under "CONNECT".
44
+
45
+

46
+
47
+
## (Optional) Step 4: Inspect system logs
48
+
You can review the system logs to access details about event history, workload metrics, and real-time container output. This information is especially useful for debugging issues when a workload fails to start.
49
+
50
+

29
51
30
-
* Volume (i.e. temporary data storage)
31
-
* Data Sources (e.g. PVCs)
32
-
* Other general settings
33
-

34
52
35
-
## Optional: Create a workload from a template
36
-
Besides the project allocation, all the other workload components can be populated from a pre-defined template:
37
-

0 commit comments