Skip to content
Snippets Groups Projects
Commit 9432ea8c authored by Martin Schroschk's avatar Martin Schroschk
Browse files

HardwareTaurus: Fix checks

parent dbbce07b
No related branches found
No related tags found
3 merge requests!322Merge preview into main,!319Merge preview into main,!123HardwareTaurus: Fix checks
......@@ -13,19 +13,19 @@
- Direct access to these nodes is granted via IP whitelisting (contact
<hpcsupport@zih.tu-dresden.de>) - otherwise use TU Dresden VPN.
# AMD Rome CPUs + NVIDIA A100
## AMD Rome CPUs + NVIDIA A100
- 32 nodes, each with
- 8 x NVIDIA A100-SXM4
- 2 x AMD EPYC CPU 7352 (24 cores) @ 2.3 GHz, MultiThreading
disabled
- 1 TB RAM
- 3.5 TB /tmp local NVMe device
- Hostnames: taurusi\[8001-8034\]
- SLURM partition `alpha`
- dedicated mostly for ScaDS-AI
- 32 nodes, each with
- 8 x NVIDIA A100-SXM4
- 2 x AMD EPYC CPU 7352 (24 cores) @ 2.3 GHz, MultiThreading
disabled
- 1 TB RAM
- 3.5 TB /tmp local NVMe device
- Hostnames: taurusi\[8001-8034\]
- SLURM partition `alpha`
- dedicated mostly for ScaDS-AI
# Island 7 - AMD Rome CPUs
## Island 7 - AMD Rome CPUs
- 192 nodes, each with
- 2x AMD EPYC CPU 7702 (64 cores) @ 2.0GHz, MultiThreading
......@@ -34,18 +34,18 @@
- 200 GB /tmp on local SSD local disk
- Hostnames: taurusi\[7001-7192\]
- SLURM partition `romeo`
- more information under [RomeNodes](RomeNodes)
- more information under [RomeNodes](RomeNodes.md)
# Large SMP System HPE Superdome Flex
## Large SMP System HPE Superdome Flex
- 32 x Intel(R) Xeon(R) Platinum 8276M CPU @ 2.20GHz (28 cores)
- 47 TB RAM
- currently configured as one single node
- Hostname: taurussmp8
- SLURM partition `julia`
- more information under [HPE SD Flex](SDFlex)
- more information under [HPE SD Flex](SDFlex.md)
# IBM Power9 Nodes for Machine Learning
## IBM Power9 Nodes for Machine Learning
For machine learning, we have 32 IBM AC922 nodes installed with this
configuration:
......@@ -57,7 +57,7 @@ configuration:
- SLURM partition `ml`
- Hostnames: taurusml\[1-32\]
# Island 4 to 6 - Intel Haswell CPUs
## Island 4 to 6 - Intel Haswell CPUs
- 1456 nodes, each with 2x Intel(R) Xeon(R) CPU E5-2680 v3 (12 cores)
@ 2.50GHz, MultiThreading disabled, 128 GB SSD local disk
......@@ -72,7 +72,7 @@ configuration:
- 44 nodes with 10.67 GB RAM per core (256 GB total):
taurusi\[4189-4232\]
- SLURM Partition `haswell`
- [Node topology](%ATTACHURL%/i4000.png)
- [Node topology] **todo** %ATTACHURL%/i4000.png
### Extension of Island 4 with Broadwell CPUs
......@@ -83,28 +83,28 @@ configuration:
- Hostname: taurusi\[4233-4264\]
- SLURM partition `broadwell`
# Island 2 Phase 2 - Intel Haswell CPUs + NVIDIA K80 GPUs
## Island 2 Phase 2 - Intel Haswell CPUs + NVIDIA K80 GPUs
- 64 nodes, each with 2x Intel(R) Xeon(R) CPU E5-E5-2680 v3 (12 cores)
@ 2.50GHz, MultiThreading Disabled, 64 GB RAM (2.67 GB per core),
128 GB SSD local disk, 4x NVIDIA Tesla K80 (12 GB GDDR RAM) GPUs
- Hostname: taurusi2\[045-108\]
- SLURM Partition `gpu`
- [Node topology](%ATTACHURL%/i4000.png) (without GPUs)
- [Node topology] **todo %ATTACHURL%/i4000.png** (without GPUs)
# SMP Nodes - up to 2 TB RAM
## SMP Nodes - up to 2 TB RAM
- 5 Nodes each with 4x Intel(R) Xeon(R) CPU E7-4850 v3 (14 cores) @
2.20GHz, MultiThreading Disabled, 2 TB RAM
- Hostname: `taurussmp[3-7]`
- SLURM Partition `smp2`
- [Node topology](%ATTACHURL%/smp2.png)
- [Node topology] **todo** %ATTACHURL%/smp2.png
---# Island 2 Phase 1 - Intel Sandybridge CPUs + NVIDIA K20x GPUs
## Island 2 Phase 1 - Intel Sandybridge CPUs + NVIDIA K20x GPUs
- 44 nodes, each with 2x Intel(R) Xeon(R) CPU E5-2450 (8 cores) @
2.10GHz, MultiThreading Disabled, 48 GB RAM (3 GB per core), 128 GB
SSD local disk, 2x NVIDIA Tesla K20x (6 GB GDDR RAM) GPUs
- Hostname: `taurusi2[001-044]`
- SLURM Partition `gpu1`
- [Node topology](%ATTACHURL%/i2000.png) (without GPUs
- [Node topology] **todo** %ATTACHURL%/i2000.png (without GPUs)
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment