Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
hpc-compendium
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Wiki
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Deploy
Releases
Package registry
Container Registry
Model registry
Operate
Terraform modules
Monitor
Incidents
Service Desk
Analyze
Value stream analytics
Contributor analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Terms and privacy
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
ZIH
hpcsupport
hpc-compendium
Commits
9432ea8c
Commit
9432ea8c
authored
3 years ago
by
Martin Schroschk
Browse files
Options
Downloads
Patches
Plain Diff
HardwareTaurus: Fix checks
parent
dbbce07b
No related branches found
No related tags found
3 merge requests
!322
Merge preview into main
,
!319
Merge preview into main
,
!123
HardwareTaurus: Fix checks
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
doc.zih.tu-dresden.de/docs/use_of_hardware/HardwareTaurus.md
+23
-23
23 additions, 23 deletions
doc.zih.tu-dresden.de/docs/use_of_hardware/HardwareTaurus.md
with
23 additions
and
23 deletions
doc.zih.tu-dresden.de/docs/use_of_hardware/HardwareTaurus.md
+
23
−
23
View file @
9432ea8c
...
...
@@ -13,19 +13,19 @@
-
Direct access to these nodes is granted via IP whitelisting (contact
<hpcsupport@zih.tu-dresden.de>
) - otherwise use TU Dresden VPN.
# AMD Rome CPUs + NVIDIA A100
#
# AMD Rome CPUs + NVIDIA A100
-
32 nodes, each with
-
8 x NVIDIA A100-SXM4
-
2 x AMD EPYC CPU 7352 (24 cores) @ 2.3 GHz, MultiThreading
disabled
-
1 TB RAM
-
3.5 TB /tmp local NVMe device
-
Hostnames: taurusi
\[
8001-8034
\]
-
SLURM partition
`alpha`
-
dedicated mostly for ScaDS-AI
-
32 nodes, each with
-
8 x NVIDIA A100-SXM4
-
2 x AMD EPYC CPU 7352 (24 cores) @ 2.3 GHz, MultiThreading
disabled
-
1 TB RAM
-
3.5 TB /tmp local NVMe device
-
Hostnames: taurusi
\[
8001-8034
\]
-
SLURM partition
`alpha`
-
dedicated mostly for ScaDS-AI
# Island 7 - AMD Rome CPUs
#
# Island 7 - AMD Rome CPUs
-
192 nodes, each with
-
2x AMD EPYC CPU 7702 (64 cores) @ 2.0GHz, MultiThreading
...
...
@@ -34,18 +34,18 @@
-
200 GB /tmp on local SSD local disk
-
Hostnames: taurusi
\[
7001-7192
\]
-
SLURM partition
`romeo`
-
more information under
[
RomeNodes
](
RomeNodes
)
-
more information under
[
RomeNodes
](
RomeNodes
.md
)
# Large SMP System HPE Superdome Flex
#
# Large SMP System HPE Superdome Flex
-
32 x Intel(R) Xeon(R) Platinum 8276M CPU @ 2.20GHz (28 cores)
-
47 TB RAM
-
currently configured as one single node
-
Hostname: taurussmp8
-
SLURM partition
`julia`
-
more information under
[
HPE SD Flex
](
SDFlex
)
-
more information under
[
HPE SD Flex
](
SDFlex
.md
)
# IBM Power9 Nodes for Machine Learning
#
# IBM Power9 Nodes for Machine Learning
For machine learning, we have 32 IBM AC922 nodes installed with this
configuration:
...
...
@@ -57,7 +57,7 @@ configuration:
-
SLURM partition
`ml`
-
Hostnames: taurusml
\[
1-32
\]
# Island 4 to 6 - Intel Haswell CPUs
#
# Island 4 to 6 - Intel Haswell CPUs
-
1456 nodes, each with 2x Intel(R) Xeon(R) CPU E5-2680 v3 (12 cores)
@ 2.50GHz, MultiThreading disabled, 128 GB SSD local disk
...
...
@@ -72,7 +72,7 @@ configuration:
-
44 nodes with 10.67 GB RAM per core (256 GB total):
taurusi
\[
4189-4232
\]
-
SLURM Partition
`haswell`
-
[
Node topology
]
(
%ATTACHURL%/i4000.png
)
-
[Node topology]
**todo**
%ATTACHURL%/i4000.png
### Extension of Island 4 with Broadwell CPUs
...
...
@@ -83,28 +83,28 @@ configuration:
-
Hostname: taurusi
\[
4233-4264
\]
-
SLURM partition
`broadwell`
# Island 2 Phase 2 - Intel Haswell CPUs + NVIDIA K80 GPUs
#
# Island 2 Phase 2 - Intel Haswell CPUs + NVIDIA K80 GPUs
-
64 nodes, each with 2x Intel(R) Xeon(R) CPU E5-E5-2680 v3 (12 cores)
@ 2.50GHz, MultiThreading Disabled, 64 GB RAM (2.67 GB per core),
128 GB SSD local disk, 4x NVIDIA Tesla K80 (12 GB GDDR RAM) GPUs
-
Hostname: taurusi2
\[
045-108
\]
-
SLURM Partition
`gpu`
-
[
Node topology
]
(
%ATTACHURL%/i4000.png
)
(
without
GPUs)
-
[Node topology]
**todo
%ATTACHURL%/i4000.png
**
(without GPUs)
# SMP Nodes - up to 2 TB RAM
#
# SMP Nodes - up to 2 TB RAM
-
5 Nodes each with 4x Intel(R) Xeon(R) CPU E7-4850 v3 (14 cores) @
2.
20GHz, MultiThreading Disabled, 2 TB RAM
-
Hostname:
`taurussmp[3-7]`
-
SLURM Partition
`smp2`
-
[
Node topology
]
(
%ATTACHURL%/smp2.png
)
-
[Node topology]
**todo**
%ATTACHURL%/smp2.png
---
# Island 2 Phase 1 - Intel Sandybridge CPUs + NVIDIA K20x GPUs
#
# Island 2 Phase 1 - Intel Sandybridge CPUs + NVIDIA K20x GPUs
-
44 nodes, each with 2x Intel(R) Xeon(R) CPU E5-2450 (8 cores) @
2.
10GHz, MultiThreading Disabled, 48 GB RAM (3 GB per core), 128 GB
SSD local disk, 2x NVIDIA Tesla K20x (6 GB GDDR RAM) GPUs
-
Hostname:
`taurusi2[001-044]`
-
SLURM Partition
`gpu1`
-
[
Node topology
]
(
%ATTACHURL%/i2000.png
)
(
without
GPUs
-
[Node topology]
**todo**
%ATTACHURL%/i2000.png (without GPUs
)
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment