From fa63b915d308cbbc4b8ed01e0a2fe8d3b1156cb1 Mon Sep 17 00:00:00 2001 From: Martin Schroschk <martin.schroschk@tu-dresden.de> Date: Wed, 6 Dec 2023 15:06:16 +0100 Subject: [PATCH] Refine table with cluster overview --- .../jobs_and_resources/hardware_overview.md | 21 +++++++++---------- 1 file changed, 10 insertions(+), 11 deletions(-) diff --git a/doc.zih.tu-dresden.de/docs/jobs_and_resources/hardware_overview.md b/doc.zih.tu-dresden.de/docs/jobs_and_resources/hardware_overview.md index 65d000c9b..7ec5a6364 100644 --- a/doc.zih.tu-dresden.de/docs/jobs_and_resources/hardware_overview.md +++ b/doc.zih.tu-dresden.de/docs/jobs_and_resources/hardware_overview.md @@ -1,7 +1,6 @@ # HPC Resources -TODO: Update this introduction - +<!--TODO: Update this introduction--> HPC resources in ZIH systems comprise the *High Performance Computing and Storage Complex* and its extension *High Performance Computing – Data Analytics*. In total it offers scientists about 60,000 CPU cores and a peak performance of more than 1.5 quadrillion floating point @@ -18,18 +17,18 @@ users and the ZIH. will have five homogeneous clusters with their own Slurm instances and with cluster specific login nodes running on the same CPU. -With the installation and start of operation of the [new HPC system Barnard(#barnard), +With the installation and start of operation of the [new HPC system Barnard](#barnard), quite significant changes w.r.t. HPC system landscape at ZIH follow. The former HPC system Taurus is partly switched-off and partly split up into separate clusters. In the end, from the users' perspective, there will be **five separate clusters**: -| Name | Description | Year| DNS | -| --- | --- | --- | --- | -| **Barnard** | CPU cluster |2023| `n[1001-1630].barnard.hpc.tu-dresden.de` | -| **Romeo** | CPU cluster |2020| `i[8001-8190].romeo.hpc.tu-dresden.de` | -| **Alpha Centauri** | GPU cluster | 2021| `i[8001-8037].alpha.hpc.tu-dresden.de` | -| **Julia** | single SMP system |2021| `smp8.julia.hpc.tu-dresden.de` | -| **Power** | IBM Power/GPU system |2018| `ml[1-29].power9.hpc.tu-dresden.de` | +| Name | Description | Year of Installation | DNS | +| ----------------------------------- | ----------------------| -------------------- | --- | +| [`Barnard`](#barnard) | CPU cluster | 2023 | `n[1001-1630].barnard.hpc.tu-dresden.de` | +| [`Alpha Centauri`](#alpha-centauri) | GPU cluster | 2021 | `i[8001-8037].alpha.hpc.tu-dresden.de` | +| [`Julia`](#julia) | Single SMP system | 2021 | `smp8.julia.hpc.tu-dresden.de` | +| [`Romeo`](#romeo) | CPU cluster | 2020 | `i[8001-8190].romeo.hpc.tu-dresden.de` | +| [`Power`](#power9) | IBM Power/GPU cluster | 2018 | `ml[1-29].power9.hpc.tu-dresden.de` | All clusters will run with their own [Slurm batch system](slurm.md) and job submission is possible only from their respective login nodes. @@ -145,7 +144,7 @@ CPUs. ## Alpha Centauri The cluster **Alpha Centauri** (short: **Alpha**) by NEC provides AMD Rome CPUs and NVIDIA A100 GPUs -and designed for AI and ML tasks. +and is designed for AI and ML tasks. - 34 nodes, each with - 8 x NVIDIA A100-SXM4 Tensor Core-GPUs -- GitLab