From 9432ea8cd249c4fb79ee27b1e25d8728fc4af96d Mon Sep 17 00:00:00 2001
From: Martin Schroschk <martin.schroschk@tu-dresden.de>
Date: Wed, 30 Jun 2021 09:21:50 +0200
Subject: [PATCH] HardwareTaurus: Fix checks

---
 .../docs/use_of_hardware/HardwareTaurus.md    | 46 +++++++++----------
 1 file changed, 23 insertions(+), 23 deletions(-)

diff --git a/doc.zih.tu-dresden.de/docs/use_of_hardware/HardwareTaurus.md b/doc.zih.tu-dresden.de/docs/use_of_hardware/HardwareTaurus.md
index 30aef5ccc..cc76897e3 100644
--- a/doc.zih.tu-dresden.de/docs/use_of_hardware/HardwareTaurus.md
+++ b/doc.zih.tu-dresden.de/docs/use_of_hardware/HardwareTaurus.md
@@ -13,19 +13,19 @@
 -   Direct access to these nodes is granted via IP whitelisting (contact
     <hpcsupport@zih.tu-dresden.de>) - otherwise use TU Dresden VPN.
 
-# AMD Rome CPUs + NVIDIA A100
+## AMD Rome CPUs + NVIDIA A100
 
--   32 nodes, each with
-    -   8 x NVIDIA A100-SXM4
-    -   2 x AMD EPYC CPU 7352 (24 cores) @ 2.3 GHz, MultiThreading
-        disabled
-    -   1 TB RAM
-    -   3.5 TB /tmp local NVMe device
--   Hostnames: taurusi\[8001-8034\]
--   SLURM partition `alpha`
--   dedicated mostly for ScaDS-AI
+- 32 nodes, each with
+  -   8 x NVIDIA A100-SXM4
+  -   2 x AMD EPYC CPU 7352 (24 cores) @ 2.3 GHz, MultiThreading
+      disabled
+  -   1 TB RAM
+  -   3.5 TB /tmp local NVMe device
+- Hostnames: taurusi\[8001-8034\]
+- SLURM partition `alpha`
+- dedicated mostly for ScaDS-AI
 
-# Island 7 - AMD Rome CPUs
+## Island 7 - AMD Rome CPUs
 
 -   192 nodes, each with
     -   2x AMD EPYC CPU 7702 (64 cores) @ 2.0GHz, MultiThreading
@@ -34,18 +34,18 @@
     -   200 GB /tmp on local SSD local disk
 -   Hostnames: taurusi\[7001-7192\]
 -   SLURM partition `romeo`
--   more information under [RomeNodes](RomeNodes)
+-   more information under [RomeNodes](RomeNodes.md)
 
-# Large SMP System HPE Superdome Flex
+## Large SMP System HPE Superdome Flex
 
 -   32 x Intel(R) Xeon(R) Platinum 8276M CPU @ 2.20GHz (28 cores)
 -   47 TB RAM
 -   currently configured as one single node
     -   Hostname: taurussmp8
 -   SLURM partition `julia`
--   more information under [HPE SD Flex](SDFlex)
+-   more information under [HPE SD Flex](SDFlex.md)
 
-# IBM Power9 Nodes for Machine Learning
+## IBM Power9 Nodes for Machine Learning
 
 For machine learning, we have 32 IBM AC922 nodes installed with this
 configuration:
@@ -57,7 +57,7 @@ configuration:
 -   SLURM partition `ml`
 -   Hostnames: taurusml\[1-32\]
 
-# Island 4 to 6 - Intel Haswell CPUs
+## Island 4 to 6 - Intel Haswell CPUs
 
 -   1456 nodes, each with 2x Intel(R) Xeon(R) CPU E5-2680 v3 (12 cores)
     @ 2.50GHz, MultiThreading disabled, 128 GB SSD local disk
@@ -72,7 +72,7 @@ configuration:
     -   44 nodes with 10.67 GB RAM per core (256 GB total):
         taurusi\[4189-4232\]
 -   SLURM Partition `haswell`
--   [Node topology](%ATTACHURL%/i4000.png)
+-   [Node topology] **todo** %ATTACHURL%/i4000.png
 
 ### Extension of Island 4 with Broadwell CPUs
 
@@ -83,28 +83,28 @@ configuration:
 -   Hostname: taurusi\[4233-4264\]
 -   SLURM partition `broadwell`
 
-# Island 2 Phase 2 - Intel Haswell CPUs + NVIDIA K80 GPUs
+## Island 2 Phase 2 - Intel Haswell CPUs + NVIDIA K80 GPUs
 
 -   64 nodes, each with 2x Intel(R) Xeon(R) CPU E5-E5-2680 v3 (12 cores)
     @ 2.50GHz, MultiThreading Disabled, 64 GB RAM (2.67 GB per core),
     128 GB SSD local disk, 4x NVIDIA Tesla K80 (12 GB GDDR RAM) GPUs
 -   Hostname: taurusi2\[045-108\]
 -   SLURM Partition `gpu`
--   [Node topology](%ATTACHURL%/i4000.png) (without GPUs)
+-   [Node topology] **todo %ATTACHURL%/i4000.png** (without GPUs)
 
-# SMP Nodes - up to 2 TB RAM
+## SMP Nodes - up to 2 TB RAM
 
 -   5 Nodes each with 4x Intel(R) Xeon(R) CPU E7-4850 v3 (14 cores) @
     2.20GHz, MultiThreading Disabled, 2 TB RAM
     -   Hostname: `taurussmp[3-7]`
     -   SLURM Partition `smp2`
-    -   [Node topology](%ATTACHURL%/smp2.png)
+    -   [Node topology] **todo** %ATTACHURL%/smp2.png
 
----# Island 2 Phase 1 - Intel Sandybridge CPUs + NVIDIA K20x GPUs
+## Island 2 Phase 1 - Intel Sandybridge CPUs + NVIDIA K20x GPUs
 
 -   44 nodes, each with 2x Intel(R) Xeon(R) CPU E5-2450 (8 cores) @
     2.10GHz, MultiThreading Disabled, 48 GB RAM (3 GB per core), 128 GB
     SSD local disk, 2x NVIDIA Tesla K20x (6 GB GDDR RAM) GPUs
 -   Hostname: `taurusi2[001-044]`
 -   SLURM Partition `gpu1`
--   [Node topology](%ATTACHURL%/i2000.png) (without GPUs
+-   [Node topology] **todo** %ATTACHURL%/i2000.png (without GPUs)
-- 
GitLab