From d1f16e4c49f0c6fb6f162c42e98ae159835e8abb Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Sebastian=20D=C3=B6bel?= <sebastian.doebel@tu-dresden.de> Date: Tue, 5 Nov 2024 14:02:41 +0100 Subject: [PATCH] Apply 8 suggestion(s) to 5 file(s) Co-authored-by: Bert Wesarg <bert.wesarg@tu-dresden.de> --- doc.zih.tu-dresden.de/docs/access/jupyterhub.md | 2 +- doc.zih.tu-dresden.de/docs/access/ssh_login.md | 2 +- doc.zih.tu-dresden.de/docs/data_lifecycle/working.md | 2 +- doc.zih.tu-dresden.de/docs/jobs_and_resources/capella.md | 2 +- doc.zih.tu-dresden.de/docs/jobs_and_resources/power9.md | 8 ++++---- 5 files changed, 8 insertions(+), 8 deletions(-) diff --git a/doc.zih.tu-dresden.de/docs/access/jupyterhub.md b/doc.zih.tu-dresden.de/docs/access/jupyterhub.md index 790e8b92c..bc54e752c 100644 --- a/doc.zih.tu-dresden.de/docs/access/jupyterhub.md +++ b/doc.zih.tu-dresden.de/docs/access/jupyterhub.md @@ -109,7 +109,7 @@ At the following table it's possible to see what is available at each cluster. |----------------------|--------|----|---------|--------|------------|--------|-----| | Alpha | - | OK | OK | OK | OK* | OK* | - | | Barnard | OK | OK | OK | OK | OK* | OK* | - | -|Capella | OK | OK | OK | OK | - | - | - | +| Capella | OK | OK | OK | OK | - | - | - | | Romeo | - | OK | OK | OK | OK* | OK* | - | | VIS | OK | OK | OK | OK | OK* | OK* | OK | diff --git a/doc.zih.tu-dresden.de/docs/access/ssh_login.md b/doc.zih.tu-dresden.de/docs/access/ssh_login.md index 73ac764eb..00cf3b7cb 100644 --- a/doc.zih.tu-dresden.de/docs/access/ssh_login.md +++ b/doc.zih.tu-dresden.de/docs/access/ssh_login.md @@ -119,7 +119,7 @@ for more information on Dataport nodes. In the above `.ssh/config` file, the HPC system `Barnard` is chosen as an example. The very same settings can be made for individuall or all ZIH systems, e.g. `Capella`, `Alpha`, - `Julia`,`Romeo` etc. + `Julia`, `Romeo` etc. ## X11-Forwarding diff --git a/doc.zih.tu-dresden.de/docs/data_lifecycle/working.md b/doc.zih.tu-dresden.de/docs/data_lifecycle/working.md index a00a400e1..d7c5d9a2e 100644 --- a/doc.zih.tu-dresden.de/docs/data_lifecycle/working.md +++ b/doc.zih.tu-dresden.de/docs/data_lifecycle/working.md @@ -10,7 +10,7 @@ performance and permanence. | `Lustre` | `/data/walrus` | 20 PB | global | Only accessible via [Workspaces](workspaces.md). For moderately low bandwidth, low IOPS. Mounted read-only on compute nodes. | | `WEKAio` | `/data/weasel` | 1 PB | global (w/o Power) | *Coming 2024!* For high IOPS | | `ext4` | `/tmp` | 95 GB | node local | Systems: tbd. Is cleaned up after the job automatically. | -| `WEKAio` | `/data/cat` | 1 PB | shared on Capella | For high IOPS. Only available on [Capella](../jobs_and_resources/capella.md). | +| `WEKAio` | `/data/cat` | 1 PB | only Capella | For high IOPS. Only available on [Capella](../jobs_and_resources/capella.md). | ## Recommendations for Filesystem Usage diff --git a/doc.zih.tu-dresden.de/docs/jobs_and_resources/capella.md b/doc.zih.tu-dresden.de/docs/jobs_and_resources/capella.md index 54e620fe7..6b82a999c 100644 --- a/doc.zih.tu-dresden.de/docs/jobs_and_resources/capella.md +++ b/doc.zih.tu-dresden.de/docs/jobs_and_resources/capella.md @@ -27,7 +27,7 @@ Although all other [filesystems](../data_lifecycle/file_systems.md) ### Modules The easiest way is using the [module system](../software/modules.md). -All software available from the module system has been specifically build for the cluster `Alpha` +All software available from the module system has been deliberately build for the cluster `Alpha` i.e., with optimization for Zen4 (Genoa) microarchitecture and CUDA-support enabled. To check the available modules for `Capella`, use the command diff --git a/doc.zih.tu-dresden.de/docs/jobs_and_resources/power9.md b/doc.zih.tu-dresden.de/docs/jobs_and_resources/power9.md index 4f8217175..600174f79 100644 --- a/doc.zih.tu-dresden.de/docs/jobs_and_resources/power9.md +++ b/doc.zih.tu-dresden.de/docs/jobs_and_resources/power9.md @@ -12,7 +12,7 @@ partition `power` within the now decommissioned `Taurus` system. With the decomm - 32 nodes, each with - 2 x IBM Power9 CPU (2.80 GHz, 3.10 GHz boost, 22 cores) - 256 GB RAM (8 x 16 GB DDR4-2666 MT/s per socket) - - 6 x NVIDIA VOLTA V100 with 32 GB HBM2 + - 6 x NVIDIA Tesla V100 with 32 GB HBM2 - NVLINK bandwidth 150 GB/s between GPUs and host - Login nodes: `login[1-2].power9.hpc.tu-dresden.de` - Hostnames: `ml[1-29].power9.hpc.tu-dresden.de` @@ -25,14 +25,14 @@ partition `power` within the now decommissioned `Taurus` system. With the decomm If you want to use containers on `Power9`, please refer to the page [Singularity for Power9 Architecuture](../software/singularity_power9.md). -The compute nodes of the cluster `power9` are built on the base of +The compute nodes of the cluster `Power9` are built on the base of [Power9 architecture](https://www.ibm.com/it-infrastructure/power/power9) from IBM. The system was created for AI challenges, analytics and working with data-intensive workloads and accelerated databases. The main feature of the nodes is the ability to work with the [NVIDIA Tesla V100](https://www.nvidia.com/en-gb/data-center/tesla-v100/) GPU with **NV-Link** support that allows a total bandwidth with up to 300 GB/s. Each node on the -cluster `power9` has 6x Tesla V-100 GPUs. You can find a detailed specification of the cluster in our +cluster `Power9` has six Tesla V100 GPUs. You can find a detailed specification of the cluster in our [Power9 documentation](../jobs_and_resources/hardware_overview.md). @@ -43,5 +43,5 @@ cluster `power9` has 6x Tesla V-100 GPUs. You can find a detailed specification ### Power AI -There are tools provided by IBM, that work on cluster `power9` and are related to AI tasks. +There are tools provided by IBM, that work on cluster `Power9` and are related to AI tasks. For more information see our [Power AI documentation](power_ai.md). -- GitLab