Skip to content

Commit 1020103

Browse files
authored
Merge pull request #12 from Sydney-Informatics-Hub/dgx-workshop-updates
More edits for the workshop
2 parents 7df6c58 + 7b829c9 commit 1020103

9 files changed

Lines changed: 36 additions & 14 deletions

fig/default_data_source.png

151 KB
Loading

fig/workload_comp_resource.png

-10.3 KB
Loading

fig/workload_connect_jupyter.png

96.9 KB
Loading

fig/workload_datasource.png

91.8 KB
Loading

fig/workload_definition.png

-2.53 KB
Loading

fig/workload_environment.png

16.4 KB
Loading

fig/workload_logs.png

378 KB
Loading

notebooks/data_sources.md

Lines changed: 8 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1,8 @@
1-
# Data Sources
1+
# Data Sources
2+
Data sources in Run:ai allow you to connect additional storage systems to your Run:ai projects, enabling seamless access to datasets required for your workloads. Run:ai supports various types of data sources, including PVC ([Persistent Volume Claim](https://kubernetes.io/docs/concepts/storage/persistent-volumes/)), NFS, S3-compatible storage, and more.
3+
4+
When a Run:ai project is created, a default data source (PVC) is automatically set up for this project. You can find the data source under the "Data Sources", listed as `pvc-<dashr_project_shortcode>`:
5+
6+
![Default Data Source](../fig/default_data_source.png)
7+
8+
This dedicated persistent storage can be accessable from any workload running within the same project. The default mount path inside the container is `/scratch/<dashr_project_shortcode>`. It is especially useful for storing intermediate results, model checkpoints, and [uploading data to the cluster](./data_transfer.md) from external sources (*e.g.* RDS).

notebooks/jupyter_tutorial.md

Lines changed: 28 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -6,32 +6,47 @@ Generally, the minimum requirements you need before creating the workload includ
66
* Being granted permission to an active project
77
* An [environment](environments.md) to run such job
88
* Have created a data source, e.g. a PVC, to store your input and output data
9-
* Understand the compute resources you need to run the job and have the option available under Compute Resources
9+
* Understand the compute resources you need to run the job and have the option available under "Compute Resources"
1010

1111
In this tutorial, we will create a simple Jupyter Lab workload that allows you to run Jupyter notebooks interactively on the SIH GPU cluster.
1212

1313
## Step 1: Create a workload
1414
Navigate to the Workloads section of the platform and click on the "NEW WORKLOAD" button. Select "Workspace" from the dropdown menu.
15+
1516
![New workload](../fig/workload_create_workspace.png)
1617

1718
## Step 2: Configure the workload from scratch
18-
Fill in the necessary details for your workload:
19+
Define the necessary information for your workload:
1920

2021
* Under "Projects" select the project it will be linked to
21-
* Under "Templates" select "Start from sratch" (do not use any existing template)
22+
* Under "Templates" select "Start from sratch" (*i.e.* do not use any existing template)
23+
* Provide a descriptive name for the workload
24+
2225
![Project and Template](../fig/workload_definition.png)
23-
* Under "Environment" select the Jupyter Lab container environment you want to run
26+
27+
* Select an environment to create the container. The SIH team has prepared a [pre-built image](https://hub.docker.com/r/sydneyinformaticshub/dgx-interactive-jupyterlab) (`sydneyinformatics/dgx-interactive-jupyterlab`) with Jupyter Lab and commonly used data science packages installed.
28+
2429
![Software environment](../fig/workload_environment.png)
25-
* Under "Compute resource" select the resources required.
30+
31+
* Select the amount of compute resources to run the workload. In this tutorial, we will select the `small-fraction` option that requires 1 H200 GPU with 10% of its memory (~14GB).
32+
2633
![Compute resource](../fig/workload_comp_resource.png)
2734

28-
There are other optional components you can add to a workload depending on the needs of your task. These include:
35+
* Configure the [data source](./data_sources.md) to be mounted to the container. Here we select the default PVC created for the project. The mount path inside the container is set to `/scratch/<dashr_project_shortcode>`.
36+
37+
![Data resource](../fig/workload_datasource.png)
38+
39+
* Lastly, Click on "CREATE WORKLOAD" to submit the workload to the cluster.
40+
41+
## Step 3: Connect to Jupyter Lab
42+
43+
When the status changes to "Running", you can access the Jupyter Lab interface by selecting "Jupyter" under "CONNECT".
44+
45+
![Connect to the Jupyter Lab interface](../fig/workload_connect_jupyter.png)
46+
47+
## (Optional) Step 4: Inspect system logs
48+
You can review the system logs to access details about event history, workload metrics, and real-time container output. This information is especially useful for debugging issues when a workload fails to start.
49+
50+
![Workload logs](../fig/workload_logs.png)
2951

30-
* Volume (i.e. temporary data storage)
31-
* Data Sources (e.g. PVCs)
32-
* Other general settings
33-
![](../fig/workload_additional_setups.png)
3452

35-
## Optional: Create a workload from a template
36-
Besides the project allocation, all the other workload components can be populated from a pre-defined template:
37-
![Create workload from existing template](../fig/workload_template.png)

0 commit comments

Comments
 (0)