From 218d82c92e78f7f4caefbdba7bb646d00fbdba99 Mon Sep 17 00:00:00 2001
From: Maximilian Sander <maximilian.sander1@tu-dresden.de>
Date: Mon, 13 May 2024 08:32:56 +0200
Subject: [PATCH] Add RAM speeds for each partition

---
 .../docs/jobs_and_resources/hardware_overview.md       | 10 +++++-----
 1 file changed, 5 insertions(+), 5 deletions(-)

diff --git a/doc.zih.tu-dresden.de/docs/jobs_and_resources/hardware_overview.md b/doc.zih.tu-dresden.de/docs/jobs_and_resources/hardware_overview.md
index 3bf43c019..c8d49918b 100644
--- a/doc.zih.tu-dresden.de/docs/jobs_and_resources/hardware_overview.md
+++ b/doc.zih.tu-dresden.de/docs/jobs_and_resources/hardware_overview.md
@@ -110,7 +110,7 @@ CPUs.
 
 - 630 nodes, each with
     - 2 x Intel Xeon Platinum 8470 (52 cores) @ 2.00 GHz, Multithreading enabled
-    - 512 GB RAM
+    - 512 GB RAM (8 x 32 GB DDR5-4800 MT/s per socket)
     - 12 nodes provide 1.8 TB local storage on NVMe device at `/tmp`
     - All other nodes are diskless and have no or very limited local storage (i.e. `/tmp`)
 - Login nodes: `login[1-4].barnard.hpc.tu-dresden.de`
@@ -126,7 +126,7 @@ and is designed for AI and ML tasks.
 - 34 nodes, each with
     - 8 x NVIDIA A100-SXM4 Tensor Core-GPUs
     - 2 x AMD EPYC CPU 7352 (24 cores) @ 2.3 GHz, Multithreading available
-    - 1 TB RAM
+    - 1 TB RAM (16 x 32 GB DDR4-2933 MT/s per socket)
     - 3.5 TB local memory on NVMe device at `/tmp`
 - Login nodes: `login[1-2].alpha.hpc.tu-dresden.de`
 - Hostnames: `i[8001-8037].alpha.hpc.tu-dresden.de`
@@ -139,7 +139,7 @@ The cluster `Romeo` is a general purpose cluster by NEC based on AMD Rome CPUs.
 
 - 192 nodes, each with
     - 2 x AMD EPYC CPU 7702 (64 cores) @ 2.0 GHz, Multithreading available
-    - 512 GB RAM
+    - 512 GB RAM (8 x 32 GB DDR4-3200 MT/s per socket)
     - 200 GB local memory on SSD at `/tmp`
 - Login nodes: `login[1-2].romeo.hpc.tu-dresden.de`
 - Hostnames: `i[7001-7190].romeo.hpc.tu-dresden.de`
@@ -153,7 +153,7 @@ architecture.
 
 - 1 node, with
     - 32 x Intel(R) Xeon(R) Platinum 8276M CPU @ 2.20 GHz (28 cores)
-    - 47 TB RAM
+    - 47 TB RAM (12 x 128 GB DDR4-2933 MT/s per socket)
 - Configured as one single node
 - 48 TB RAM (usable: 47 TB - one TB is used for cache coherence protocols)
 - 370 TB of fast NVME storage available at `/nvme/<projectname>`
@@ -168,7 +168,7 @@ The cluster `Power9` by IBM is based on Power9 CPUs and provides NVIDIA V100 GPU
 
 - 32 nodes, each with
     - 2 x IBM Power9 CPU (2.80 GHz, 3.10 GHz boost, 22 cores)
-    - 256 GB RAM DDR4 2666 MHz
+    - 256 GB RAM (8 x 16 GB DDR4-2666 MT/s per socket)
     - 6 x NVIDIA VOLTA V100 with 32 GB HBM2
     - NVLINK bandwidth 150 GB/s between GPUs and host
 - Login nodes: `login[1-2].power9.hpc.tu-dresden.de`
-- 
GitLab