Skip to content

Commit ff985c3

Browse files
committed
Add H200 to user guide
1 parent 8a5ada5 commit ff985c3

File tree

3 files changed

+5
-1
lines changed

3 files changed

+5
-1
lines changed

docs/hardware/gpu.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -12,6 +12,7 @@ The UBELIX GPU-based hardware partition currently consists of about 20 compute n
1212
| 8 | 8x NVIDIA RTX4090 | 128 (2x64) | 24 GB | 100Gb/s |
1313
| 1 | 6x NVIDIA A100 | 128 (2x64) | 80 GB | 100Gb/s |
1414
| 5 | 8x NVIDIA H100 | 128 (2x64) | 96 GB | 100Gb/s |
15+
| 2 | 8x NVIDIA H200 | 128 (2x64) | 141 GB | 100Gb/s |
1516

1617
See the [Slurm partitions page][slurm-partitions] for an overview of options
1718
for allocating these nodes.
@@ -43,6 +44,7 @@ follows:
4344
| Nvidia RTX 4090 | 16 | 90GB |
4445
| Nvidia A100 | 20 | 80GB |
4546
| Nvidia H100 | 16 | 90GB |
47+
| Nvidia H200 | 16 | 90GB |
4648

4749
## Network
4850

docs/runjobs/partitions.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ We are currently operating the following partitions:
99
| --------- | -------- | ---------- | ---------------- | ------------- |
1010
|**epyc2 (default)** | single and multi-core |AMD Epyc2 2x64 cores <br> AMD Epyc4 2x96 cores | 1TB <br> 1.5TB | 1TB |
1111
| bdw | full nodes only (x*20cores) | Intel Broadwell 2x10 cores | 156GB | 1TB |
12-
| gpu | GPU <br> (8 GPUs per node, <br> varying CPUs) | Nvidia RTX 3090 <br> Nvidia RTX 4090 <br> Nvidia A100 <br> Nvidia H100 | 24GB <br> 24GB <br> 80GB <br> 96GB | 1.92TB <br> 1.92TB <br> 1.92TB <br> 1.92TB |
12+
| gpu | GPU <br> (8 GPUs per node, <br> varying CPUs) | Nvidia RTX 3090 <br> Nvidia RTX 4090 <br> Nvidia A100 <br> Nvidia H100 <br> Nvidia H200 | 24GB <br> 24GB <br> 80GB <br> 96GB <br> 141GB | 1.92TB <br> 1.92TB <br> 1.92TB <br> 1.92TB <br> 1.92TB|
1313
| gpu-invest | GPU | *see gpu partition* | | |
1414
| icpu-*investor* | single and multi-core | *see epyc2 partition* | | |
1515

docs/runjobs/scheduled-jobs/gpus.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,7 @@ architecture and use one of the following `--gres` option to select it.
1919
| Nvidia Geforce RTX 4090 | `--gres=gpu:rtx4090:<number_of_gpus>` |
2020
| Nvidia A100 | `--gres=gpu:a100:<number_of_gpus>` |
2121
| Nvidia H100 | `--gres=gpu:h100:<number_of_gpus>` |
22+
| Nvidia H200 | `--gres=gpu:h200:<number_of_gpus>` |
2223

2324
Alternatively, you may use the `--gpus`, `--gpus-per-node` and
2425
`--gpus-per-tasks` otions. Note that the GPU type still needs to be specified
@@ -48,6 +49,7 @@ In the past, we observed that GPU resources were often left unused because some
4849
| Nvidia RTX 4090 | 16 | 90GB |
4950
| Nvidia A100 | 20 | 80GB |
5051
| Nvidia H100 | 16 | 90GB |
52+
| Nvidia H200 | 16 | 90GB |
5153

5254
If you submit a GPU job that requests more resources than are available per GPU, your job will be rejected. If your job requires more CPU and memory resources, you may choose to allocate additional GPUs even if these additional GPUs remain unused by your application.
5355

0 commit comments

Comments
 (0)