Skip to content
Snippets Groups Projects
hardware_overview.md 8.63 KiB
Newer Older
Ulf Markwardt's avatar
Ulf Markwardt committed
HPC resources in ZIH systems comprise the *High Performance Computing and Storage Complex* and its
extension *High Performance Computing – Data Analytics*. In total it offers scientists
Martin Schroschk's avatar
Martin Schroschk committed
about 100,000 CPU cores and a peak performance of more than 1.5 quadrillion floating point
Ulf Markwardt's avatar
Ulf Markwardt committed
operations per second. The architecture specifically tailored to data-intensive computing, Big Data
analytics, and artificial intelligence methods with extensive capabilities for energy measurement
and performance monitoring provides ideal conditions to achieve the ambitious research goals of the
users and the ZIH.

Martin Schroschk's avatar
Martin Schroschk committed
HPC resources at ZIH comprise a total of **six systems**:
| Name                                | Description           | Year of Installation | DNS |
| ----------------------------------- | ----------------------| -------------------- | --- |
Ulf Markwardt's avatar
Ulf Markwardt committed
| [`Capella`](#capella)               | GPU cluster           | 2024                 | `c[1-244].capella.hpc.tu-dresden.de` |
| [`Barnard`](#barnard)               | CPU cluster           | 2023                 | `n[1001-1630].barnard.hpc.tu-dresden.de` |
| [`Alpha Centauri`](#alpha-centauri) | GPU cluster           | 2021                 | `i[8001-8037].alpha.hpc.tu-dresden.de` |
| [`Julia`](#julia)                   | Single SMP system     | 2021                 | `julia.hpc.tu-dresden.de` |
| [`Romeo`](#romeo)                   | CPU cluster           | 2020                 | `i[8001-8190].romeo.hpc.tu-dresden.de` |
| [`Power9`](#power9)                 | IBM Power/GPU cluster | 2018                 | `ml[1-29].power9.hpc.tu-dresden.de` |

All clusters will run with their own [Slurm batch system](slurm.md) and job submission is possible
only from their respective login nodes.
Martin Schroschk's avatar
Martin Schroschk committed
## Architectural Re-Design 2023
Martin Schroschk's avatar
Martin Schroschk committed
Over the last decade we have been running our HPC system of high heterogeneity with a single
Slurm batch system. This made things very complicated, especially to inexperienced users.  With
the replacement of the Taurus system by the cluster [Barnard](#barnard) we
**now create homogeneous clusters with their own Slurm instances and with cluster specific login
nodes** running on the same CPU.  Job submission will be possible only from within the cluster
(compute or login node).
Martin Schroschk's avatar
Martin Schroschk committed
All clusters will be integrated to the new InfiniBand fabric and have then the same access to
the shared filesystems. This recabling will require a brief downtime of a few days.
Ulf Markwardt's avatar
Ulf Markwardt committed
![Architecture overview 2023](../jobs_and_resources/misc/architecture_2024.png)
Martin Schroschk's avatar
Martin Schroschk committed
{: align=center}
Martin Schroschk's avatar
Martin Schroschk committed
### Compute Systems
Martin Schroschk's avatar
Martin Schroschk committed
All compute clusters now act as separate entities having their own
login nodes of the same hardware and their very own Slurm batch systems. The different hardware,
e.g. Romeo and Alpha Centauri, is no longer managed via a single Slurm instance with
corresponding partitions. Instead, you as user now chose the hardware by the choice of the
correct login node.
Martin Schroschk's avatar
Martin Schroschk committed
The login nodes can be used for smaller interactive jobs on the clusters. There are
restrictions in place, though, wrt. usable resources and time per user. For larger
computations, please use interactive jobs.
Martin Schroschk's avatar
Martin Schroschk committed
### Storage Systems
Martin Schroschk's avatar
Martin Schroschk committed
For an easier grasp on the major categories (size, speed), the
work filesystems now come with the names of animals.
Martin Schroschk's avatar
Martin Schroschk committed
#### Permanent Filesystems
Martin Schroschk's avatar
Martin Schroschk committed
We now have `/home` and `/software` in a Lustre filesystem. Snapshots
and tape backup are configured. (`/projects` remains the same until a recabling.)
Martin Schroschk's avatar
Martin Schroschk committed
The Lustre filesystem `/data/walrus` is meant for larger data with a slow
access. It is installed to replace `/warm_archive`.
Martin Schroschk's avatar
Martin Schroschk committed
#### Work Filesystems
Martin Schroschk's avatar
Martin Schroschk committed
In the filesystem market with new players  it is getting more and more
complicated to identify the best suited filesystem for a specific use case. Often,
only tests can find the best setup for a specific workload.
Martin Schroschk's avatar
Martin Schroschk committed
* `/data/horse` - 20 PB - high bandwidth (Lustre)
* `/data/octopus` - 0.5 PB - for interactive usage (Lustre) - to be mounted on Alpha Centauri
* `/data/weasel` - 1 PB - for high IOPS (WEKA)  - coming 2024.
Martin Schroschk's avatar
Martin Schroschk committed
#### Difference Between "Work" And "Permanent"
Martin Schroschk's avatar
Martin Schroschk committed
A large number of changing files is a challenge for any backup system. To protect
our snapshots and backup from work data,
`/projects` cannot be used for temporary data on the compute nodes - it is mounted read-only.
Martin Schroschk's avatar
Martin Schroschk committed
For `/home`, we create snapshots and tape backups. That's why working there,
with a high frequency of changing files is a bad idea.
Martin Schroschk's avatar
Martin Schroschk committed
Please use our data mover mechanisms to transfer worthy data to permanent
storages or long-term archives.
## Login and Dataport Nodes
Ulf Markwardt's avatar
Ulf Markwardt committed

- Login-Nodes
    - Individual for each cluster. See sections below.
- 2 Data-Transfer-Nodes
    - 2 servers without interactive login, only available via file transfer protocols
      (`rsync`, `ftp`)
    - `dataport[3-4].hpc.tu-dresden.de`
    - IPs: 141.30.73.\[4,5\]
    - Further information on the usage is documented on the site
      [dataport Nodes](../data_transfer/dataport_nodes.md)
- *outdated*: 2 Data-Transfer-Nodes `taurusexport[3-4].hrsk.tu-dresden.de`
    - DNS Alias `taurusexport.hrsk.tu-dresden.de`
    - 2 Servers without interactive login, only available via file transfer protocols
Ulf Markwardt's avatar
Ulf Markwardt committed
      (`rsync`, `ftp`)
    - available as long as outdated filesystems (e.g. `scratch`) are accessible

The cluster `Barnard` is a general purpose cluster by Bull. It is based on Intel Sapphire Rapids
    - 2 x Intel Xeon Platinum 8470 (52 cores) @ 2.00 GHz, Multithreading enabled
    - 512 GB RAM (8 x 32 GB DDR5-4800 MT/s per socket)
    - 12 nodes provide 1.8 TB local storage on NVMe device at `/tmp`
    - All other nodes are diskless and have no or very limited local storage (i.e. `/tmp`)
- Login nodes: `login[1-4].barnard.hpc.tu-dresden.de`
- Hostnames: `n[1001-1630].barnard.hpc.tu-dresden.de`
Martin Schroschk's avatar
Martin Schroschk committed
- Operating system: Red Hat Enterpise Linux 8.7
- Further information on the usage is documented on the site [CPU Cluster Barnard](barnard.md)
The cluster `Alpha Centauri` (short: `Alpha`) by NEC provides AMD Rome CPUs and NVIDIA A100 GPUs
and is designed for AI and ML tasks.
    - 8 x NVIDIA A100-SXM4 Tensor Core-GPUs
    - 2 x AMD EPYC CPU 7352 (24 cores) @ 2.3 GHz, Multithreading available
    - 1 TB RAM (16 x 32 GB DDR4-2933 MT/s per socket)
    - 3.5 TB local memory on NVMe device at `/tmp`
- Login nodes: `login[1-2].alpha.hpc.tu-dresden.de`
- Hostnames: `i[8001-8037].alpha.hpc.tu-dresden.de`
Martin Schroschk's avatar
Martin Schroschk committed
- Operating system: Rocky Linux 8.7
Martin Schroschk's avatar
Martin Schroschk committed
- Further information on the usage is documented on the site [GPU Cluster Alpha Centauri](alpha_centauri.md)
Ulf Markwardt's avatar
Ulf Markwardt committed
## Capella

The cluster `Capella` by MEGWARE provides AMD Genoa CPUs and NVIDIA H100 GPUs
and is designed for AI and ML tasks.

- 144 nodes, each with
    - 4 x NVIDIA H100-SXM5 Tensor Core-GPUs
    - 2 x AMD EPYC CPU 9334 (32 cores) @ 2.7 GHz, Multithreading disabled
    - 768 GB RAM (24 x 32 GB TruDDR5, 4800 MHz)
    - 800 GB local memory on NVMe device at `/tmp`
- Login nodes: `login[1-2].capella.hpc.tu-dresden.de`
- Hostnames: `c[1-144]].capella.hpc.tu-dresden.de`
- Operating system: Alma Linux 9.4

The cluster `Romeo` is a general purpose cluster by NEC based on AMD Rome CPUs.
- 192 nodes, each with
    - 2 x AMD EPYC CPU 7702 (64 cores) @ 2.0 GHz, Multithreading available
    - 512 GB RAM (8 x 32 GB DDR4-3200 MT/s per socket)
    - 200 GB local memory on SSD at `/tmp`
- Login nodes: `login[1-2].romeo.hpc.tu-dresden.de`
- Hostnames: `i[7001-7190].romeo.hpc.tu-dresden.de`
- Operating system: Rocky Linux 8.7
Martin Schroschk's avatar
Martin Schroschk committed
- Further information on the usage is documented on the site [CPU Cluster Romeo](romeo.md)
The cluster `Julia` is a large SMP (shared memory parallel) system by HPE based on Superdome Flex
    - 32 x Intel(R) Xeon(R) Platinum 8276M CPU @ 2.20 GHz (28 cores)
    - 47 TB RAM (12 x 128 GB DDR4-2933 MT/s per socket)
- Configured as one single node
- 48 TB RAM (usable: 47 TB - one TB is used for cache coherence protocols)
- 370 TB of fast NVME storage available at `/nvme/<projectname>`
- Login node: `julia.hpc.tu-dresden.de`
- Hostname: `julia.hpc.tu-dresden.de`
Martin Schroschk's avatar
Martin Schroschk committed
- Further information on the usage is documented on the site [SMP System Julia](julia.md)
Martin Schroschk's avatar
Martin Schroschk committed
## Power9
The cluster `Power9` by IBM is based on Power9 CPUs and provides NVIDIA V100 GPUs.
`Power9` is specifically designed for machine learning (ML) tasks.
- 32 nodes, each with
    - 2 x IBM Power9 CPU (2.80 GHz, 3.10 GHz boost, 22 cores)
    - 256 GB RAM (8 x 16 GB DDR4-2666 MT/s per socket)
    - 6 x NVIDIA VOLTA V100 with 32 GB HBM2
    - NVLINK bandwidth 150 GB/s between GPUs and host
- Login nodes: `login[1-2].power9.hpc.tu-dresden.de`
- Hostnames: `ml[1-29].power9.hpc.tu-dresden.de`
Martin Schroschk's avatar
Martin Schroschk committed
- Further information on the usage is documented on the site [GPU Cluster Power9](power9.md)