From 9432ea8cd249c4fb79ee27b1e25d8728fc4af96d Mon Sep 17 00:00:00 2001 From: Martin Schroschk <martin.schroschk@tu-dresden.de> Date: Wed, 30 Jun 2021 09:21:50 +0200 Subject: [PATCH] HardwareTaurus: Fix checks --- .../docs/use_of_hardware/HardwareTaurus.md | 46 +++++++++---------- 1 file changed, 23 insertions(+), 23 deletions(-) diff --git a/doc.zih.tu-dresden.de/docs/use_of_hardware/HardwareTaurus.md b/doc.zih.tu-dresden.de/docs/use_of_hardware/HardwareTaurus.md index 30aef5ccc..cc76897e3 100644 --- a/doc.zih.tu-dresden.de/docs/use_of_hardware/HardwareTaurus.md +++ b/doc.zih.tu-dresden.de/docs/use_of_hardware/HardwareTaurus.md @@ -13,19 +13,19 @@ - Direct access to these nodes is granted via IP whitelisting (contact <hpcsupport@zih.tu-dresden.de>) - otherwise use TU Dresden VPN. -# AMD Rome CPUs + NVIDIA A100 +## AMD Rome CPUs + NVIDIA A100 -- 32 nodes, each with - - 8 x NVIDIA A100-SXM4 - - 2 x AMD EPYC CPU 7352 (24 cores) @ 2.3 GHz, MultiThreading - disabled - - 1 TB RAM - - 3.5 TB /tmp local NVMe device -- Hostnames: taurusi\[8001-8034\] -- SLURM partition `alpha` -- dedicated mostly for ScaDS-AI +- 32 nodes, each with + - 8 x NVIDIA A100-SXM4 + - 2 x AMD EPYC CPU 7352 (24 cores) @ 2.3 GHz, MultiThreading + disabled + - 1 TB RAM + - 3.5 TB /tmp local NVMe device +- Hostnames: taurusi\[8001-8034\] +- SLURM partition `alpha` +- dedicated mostly for ScaDS-AI -# Island 7 - AMD Rome CPUs +## Island 7 - AMD Rome CPUs - 192 nodes, each with - 2x AMD EPYC CPU 7702 (64 cores) @ 2.0GHz, MultiThreading @@ -34,18 +34,18 @@ - 200 GB /tmp on local SSD local disk - Hostnames: taurusi\[7001-7192\] - SLURM partition `romeo` -- more information under [RomeNodes](RomeNodes) +- more information under [RomeNodes](RomeNodes.md) -# Large SMP System HPE Superdome Flex +## Large SMP System HPE Superdome Flex - 32 x Intel(R) Xeon(R) Platinum 8276M CPU @ 2.20GHz (28 cores) - 47 TB RAM - currently configured as one single node - Hostname: taurussmp8 - SLURM partition `julia` -- more information under [HPE SD Flex](SDFlex) +- more information under [HPE SD Flex](SDFlex.md) -# IBM Power9 Nodes for Machine Learning +## IBM Power9 Nodes for Machine Learning For machine learning, we have 32 IBM AC922 nodes installed with this configuration: @@ -57,7 +57,7 @@ configuration: - SLURM partition `ml` - Hostnames: taurusml\[1-32\] -# Island 4 to 6 - Intel Haswell CPUs +## Island 4 to 6 - Intel Haswell CPUs - 1456 nodes, each with 2x Intel(R) Xeon(R) CPU E5-2680 v3 (12 cores) @ 2.50GHz, MultiThreading disabled, 128 GB SSD local disk @@ -72,7 +72,7 @@ configuration: - 44 nodes with 10.67 GB RAM per core (256 GB total): taurusi\[4189-4232\] - SLURM Partition `haswell` -- [Node topology](%ATTACHURL%/i4000.png) +- [Node topology] **todo** %ATTACHURL%/i4000.png ### Extension of Island 4 with Broadwell CPUs @@ -83,28 +83,28 @@ configuration: - Hostname: taurusi\[4233-4264\] - SLURM partition `broadwell` -# Island 2 Phase 2 - Intel Haswell CPUs + NVIDIA K80 GPUs +## Island 2 Phase 2 - Intel Haswell CPUs + NVIDIA K80 GPUs - 64 nodes, each with 2x Intel(R) Xeon(R) CPU E5-E5-2680 v3 (12 cores) @ 2.50GHz, MultiThreading Disabled, 64 GB RAM (2.67 GB per core), 128 GB SSD local disk, 4x NVIDIA Tesla K80 (12 GB GDDR RAM) GPUs - Hostname: taurusi2\[045-108\] - SLURM Partition `gpu` -- [Node topology](%ATTACHURL%/i4000.png) (without GPUs) +- [Node topology] **todo %ATTACHURL%/i4000.png** (without GPUs) -# SMP Nodes - up to 2 TB RAM +## SMP Nodes - up to 2 TB RAM - 5 Nodes each with 4x Intel(R) Xeon(R) CPU E7-4850 v3 (14 cores) @ 2.20GHz, MultiThreading Disabled, 2 TB RAM - Hostname: `taurussmp[3-7]` - SLURM Partition `smp2` - - [Node topology](%ATTACHURL%/smp2.png) + - [Node topology] **todo** %ATTACHURL%/smp2.png ----# Island 2 Phase 1 - Intel Sandybridge CPUs + NVIDIA K20x GPUs +## Island 2 Phase 1 - Intel Sandybridge CPUs + NVIDIA K20x GPUs - 44 nodes, each with 2x Intel(R) Xeon(R) CPU E5-2450 (8 cores) @ 2.10GHz, MultiThreading Disabled, 48 GB RAM (3 GB per core), 128 GB SSD local disk, 2x NVIDIA Tesla K20x (6 GB GDDR RAM) GPUs - Hostname: `taurusi2[001-044]` - SLURM Partition `gpu1` -- [Node topology](%ATTACHURL%/i2000.png) (without GPUs +- [Node topology] **todo** %ATTACHURL%/i2000.png (without GPUs) -- GitLab