diff --git a/doc.zih.tu-dresden.de/docs/jobs_and_resources/hardware_overview.md b/doc.zih.tu-dresden.de/docs/jobs_and_resources/hardware_overview.md index 97017331eee0bcd291c2199ff16d11c594586c53..021f3cca5e8fa22ce843f36fe007e5f24f33feb8 100644 --- a/doc.zih.tu-dresden.de/docs/jobs_and_resources/hardware_overview.md +++ b/doc.zih.tu-dresden.de/docs/jobs_and_resources/hardware_overview.md @@ -100,7 +100,7 @@ storages or long-term archives. For about one month, the new cluster Barnard, and the old cluster Taurus will run side-by-side - both with their respective filesystems. We provide a comprehensive -[description of the migration to Barnard](migration_to_barnard.md). +[description of the migration to Barnard](barnard.md). <! -- The follwing figure provides a graphical overview of the overall process (red: user action @@ -191,8 +191,7 @@ The cluster **Romeo** is a general purpose cluster by NEC based on AMD Rome CPUs - 2 x AMD EPYC CPU 7702 (64 cores) @ 2.0 GHz, Multithreading available - 512 GB RAM - 200 GB local memory on SSD at `/tmp` -- Hostnames: `i[7001-7190].romeo.hpc.tu-dresden.de` (after - [recabling phase](architecture_2023.md#migration-phase)]) +- Hostnames: `i[7001-7190].romeo.hpc.tu-dresden.de` - Login nodes: `login[1-2].romeo.hpc.tu-dresden.de` - Further information on the usage is documented on the site [CPU Cluster Romeo](romeo.md) @@ -229,8 +228,7 @@ architecture. - Configured as one single node - 48 TB RAM (usable: 47 TB - one TB is used for cache coherence protocols) - 370 TB of fast NVME storage available at `/nvme/<projectname>` -- Hostname: `smp8.julia.hpc.tu-dresden.de` (after - [recabling phase](architecture_2023.md#migration-phase)]) +- Hostname: `smp8.julia.hpc.tu-dresden.de` - Further information on the usage is documented on the site [SMP System Julia](julia.md) ??? note "Maintenance from November 27 to December 12" @@ -266,12 +264,11 @@ The cluster **Power9** by IBM is based on Power9 CPUs and provides NVIDIA V100 G - 256 GB RAM DDR4 2666 MHz - 6 x NVIDIA VOLTA V100 with 32 GB HBM2 - NVLINK bandwidth 150 GB/s between GPUs and host -- Hostnames: `ml[1-29].power9.hpc.tu-dresden.de` (after - [recabling phase](architecture_2023.md#migration-phase)]) +- Hostnames: `ml[1-29].power9.hpc.tu-dresden.de` (after recabling phase; expected January '24) - Login nodes: `login[1-2].power9.hpc.tu-dresden.de` - Further information on the usage is documented on the site [GPU Cluster Power9](power9.md) -??? note "Maintenance from November 27 to December 12" +??? note "Maintenance" The recabling will take place from November 27 to December 12. After the maintenance, the Power9 system reappears as a stand-alone cluster that can be reached via @@ -283,6 +280,6 @@ The cluster **Power9** by IBM is based on Power9 CPUs and provides NVIDIA V100 G `/home` on Barnard](barnard.md#data-management-and-data-transfer). The old work filesystems `/lustre/scratch` and `/lustre/ssd will` be turned off on January 1 - 2024 for good (no data access afterwards!). The new work filesystem available on the Power9 - system will be `/horse`. Please + 2024 for good (no data access afterwards!). The only work filesystem available on the Power9 + system will be `/beegfs`. Please [migrate your working data to `/horse`](barnard.md#data-migration-to-new-filesystems).