diff --git a/doc.zih.tu-dresden.de/docs/jobs_and_resources/partitions_and_limits.md b/doc.zih.tu-dresden.de/docs/jobs_and_resources/partitions_and_limits.md index fae72f74a208a41c5eb731295729439c193fa9ff..71702293bbb83f3f51959418d616a80a743342f0 100644 --- a/doc.zih.tu-dresden.de/docs/jobs_and_resources/partitions_and_limits.md +++ b/doc.zih.tu-dresden.de/docs/jobs_and_resources/partitions_and_limits.md @@ -63,7 +63,8 @@ to request it. ZIH systems comprise different sets of nodes with different amount of installed memory which affect where your job may be run. To achieve the shortest possible waiting time for your jobs, you should -be aware of the limits shown in the table [Partitions and limits table](../jobs_and_resources/partitions_and_limits.md#slurm-partitions). +be aware of the limits shown in the +[Partitions and limits table](../jobs_and_resources/partitions_and_limits.md#slurm-partitions). ## Slurm Partitions @@ -71,35 +72,37 @@ The available compute nodes are grouped into logical (possibly overlapping) sets **partitions**. You can submit your job to a certain partition using the Slurm option `--partition=<partition-name>`. -Some nodes have Multithreading (SMT) enabled, so for every physical core allocated -(e.g., with `SLURM_HINT=nomultithread`), you will always get `MB per Core`*`number of threads`, -because the memory of the other threads is allocated implicitly, too. +Some partitions/nodes have Simultaneous Multithreading (SMT) enabled. You request for this +additional threads using the Slurm option `--hint=multithread` or by setting the environment +varibale `SLURM_HINT=multithread`. Besides the usage of the threads to speed up the computations, +the memory of the other threads is allocated implicitly, too, and you will allways get +`Memory per Core`*`number of threads` as memory pledge. Some partitions have a *interactive* counterpart for interactive jobs. The corresponding partitions -is suffixed with `-interactive` (e.g. `ml-interactive`) and have the same configuration. +are suffixed with `-interactive` (e.g. `ml-interactive`) and have the same configuration. -There is also a meta partition `haswell`, which contain partition `haswell64`, -`haswell256` and `smp2`and this is also the default partition. If you specify no partition or +There is also a meta partition `haswell`, which contains the partitions `haswell64`, +`haswell256` and `smp2`. `haswell` is also the default partition. If you specify no partition or partition `haswell` a Slurm plugin will choose the partition which fits to your memory requirements. There are some other partitions, which are not specified in the table above, but those partitions should not be used directly. <!-- partitions_and_limits_table --> -| Partition | Nodes | # Nodes | Cores per Node (SMT) | RAM per Core (SMT) (MB) | RAM per Node (MB) | GPUs per Node | -|:--------|:------|--------:|---------------:|------------:|------------:|--------------:| -| gpu2 | taurusi[2045-2103] | 59 | 24 | 2,583 | 62,000 | 4 | -| gpu2-interactive | taurusi[2045-2103] | 59 | 24 | 2,583 | 62,000 | 4 | -| haswell | taurusi[6001-6604],taurussmp[3-7] | 609 | | | | | -| haswell64 | taurusi[6001-6540,6559-6604] | 586 | 24 | 2,541 | 61,000 | none | -| haswell256 | taurusi[6541-6558] | 18 | 24 | 10,583 | 254,000 | none | -| interactive | taurusi[6605-6612] | 8 | 24 | 2,541 | 61,000 | none | -| smp2 | taurussmp[3-7] | 5 | 56 | 36,500 | 2,044,000 | none | -| hpdlf | taurusa[3-16] | 14 | 12 | 7,916 | 95,000 | 3 | -| ml | taurusml[3-32] | 30 | 44 (SMT: 176) | 1,443 | 254,000 | 6 | -| ml-interactive | taurusml[1-2] | 2 | 44 (SMT: 176) | 1,443 | 254,000 | 6 | -| romeo | taurusi[7003-7192] | 190 | 128 (SMT: 256) | 1,972 | 505,000 | none | -| romeo-interactive | taurusi[7001-7002] | 2 | 128 (SMT: 256) | 1,972 | 505,000 | none | -| julia | taurussmp8 | 1 | 896 | 54,006 | 48,390,000 | none | -| alpha | taurusi[8003-8034] | 32 | 48 (SMT: 96) | 10,312 | 990,000 | 8 | -| alpha-interactive | taurusi[8001-8002] | 2 | 48 (SMT: 96) | 10,312 | 990,000 | 8 | +| Partition | Nodes | # Nodes | Cores per Node | Threads per Core | Memory per Node [in MB] | Memory per Core [in MB] | GPUs per Node +|:--------|:------|--------:|---------------:|------------:|------------:|--------------:|--------------:| +| gpu2 | taurusi[2045-2103] | 59 | 24 | 1 | 62,000 | 2,583 | 4 | +| gpu2-interactive | taurusi[2045-2103] | 59 | 24 | 1 | 62,000 | 2,583 | 4 | +| haswell | taurusi[6001-6604],taurussmp[3-7] | 609 | | | | | | +| haswell64 | taurusi[6001-6540,6559-6604] | 586 | 24 | 1 | 61,000 | 2,541 | | +| haswell256 | taurusi[6541-6558] | 18 | 24 | 1 | 254,000 | 10,583 | | +| interactive | taurusi[6605-6612] | 8 | 24 | 1 | 61,000 | 2,541 | | +| smp2 | taurussmp[3-7] | 5 | 56 | 1 | 2,044,000 | 36,500 | | +| hpdlf | taurusa[3-16] | 14 | 12 | 1 | 95,000 | 7,916 | 3 | +| ml | taurusml[3-32] | 30 | 44 | 4 | 254,000 | 1,443 | 6 | +| ml-interactive | taurusml[1-2] | 2 | 44 | 4 | 254,000 | 1,443 | 6 | +| romeo | taurusi[7003-7192] | 190 | 128 | 2 | 505,000 | 1,972 | | +| romeo-interactive | taurusi[7001-7002] | 2 | 128 | 2 | 505,000 | 1,972 | | +| julia | taurussmp8 | 1 | 896 | 1 | 48,390,000 | 54,006 | | +| alpha | taurusi[8003-8034] | 32 | 48 | 2 | 990,000 | 10,312 | 8 | +| alpha-interactive | taurusi[8001-8002] | 2 | 48 | 2 | 990,000 | 10,312 | 8 | {: summary="Partitions and limits table" align="bottom"}