Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
S
Slurm
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Build
Pipelines
Jobs
Pipeline schedules
Artifacts
Deploy
Releases
Package Registry
Model registry
Operate
Environments
Terraform modules
Monitor
Incidents
Service Desk
Analyze
Value stream analytics
Contributor analytics
CI/CD analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Terms and privacy
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
tud-zih-energy
Slurm
Commits
29317df4
Commit
29317df4
authored
12 years ago
by
Morris Jette
Browse files
Options
Downloads
Patches
Plain Diff
Major updates to list of installation sites
parent
6f094a96
No related branches found
No related tags found
No related merge requests found
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
doc/html/slurm.shtml
+34
-23
34 additions, 23 deletions
doc/html/slurm.shtml
with
34 additions
and
23 deletions
doc/html/slurm.shtml
+
34
−
23
View file @
29317df4
<!--#include virtual="header.txt"-->
<h1>S
LURM
Workload Manager</h1>
<h1>S
lurm
Workload Manager</h1>
<p>S
LURM
is an open-source workload manager designed for Linux clusters of
<p>S
lurm
is an open-source workload manager designed for Linux clusters of
all sizes.
It provides three key functions.
First it allocates exclusive and/or non-exclusive access to resources
...
...
@@ -12,7 +12,7 @@ Second, it provides a framework for starting, executing, and monitoring work
Finally, it arbitrates contention for resources by managing a queue of
pending work. </p>
<p>S
LURM
's design is very modular with dozens of optional plugins.
<p>S
lurm
's design is very modular with dozens of optional plugins.
In its simplest configuration, it can be installed and configured in a
couple of minutes (see <a href="http://www.linux-mag.com/id/7239/1/">
Caos NSA and Perceus: All-in-one Cluster Software Stack</a>
...
...
@@ -28,7 +28,7 @@ world-class computer centers and rely upon a
or supporting sophisticated
<a href="priority_multifactor.html">job prioritization</a> algorithms.</p>
<p>While other workload managers do exist, S
LURM
is unique in several
<p>While other workload managers do exist, S
lurm
is unique in several
respects:
<ul>
<li><b>Scalability</b>: It is designed to operate in a heterogeneous cluster
...
...
@@ -55,52 +55,63 @@ can specify size and time limit ranges.</li>
help identify load imbalances and other anomalies.</li>
</ul></p>
<p>S
LURM
provides workload management on many of the most powerful computers in
<p>S
lurm
provides workload management on many of the most powerful computers in
the world including:
<ul>
<li><a href="https://asc.llnl.gov/computing_resources/sequoia/">Sequoia</a>,
a BlueGene/Q system at <a href="https://www.llnl.gov">LLNL</a>
an <a href="http://www.ibm.com">IBM</a> BlueGene/Q system at
<a href="https://www.llnl.gov">Lawrence Livermore National Laboratory</a>
with 1.6 petabytes of memory, 96 racks, 98,304 compute nodes, and 1.6
million cores, with a peak performance of over 20 Petaflops.</li>
million cores, with a peak performance of over 20 Petaflops.</li>
<li><a href="http://www.tacc.utexas.edu/stampede">Stampede</a> at the
<a href="http://www.tacc.utexas.edu">Texas Advanced Computing Center/University of Texas</a>
is a <a herf="http://www.dell.com">Dell</a> with over
80,000 <a href="http://www.intel.com">Intel</a> Xeon cores,
Intel Phi co-processors, plus
128 <a href="http://www.nvidia.com">NVIDIA</a> GPUs
delivering 2.66 Petaflops.</li>
<li><a href="http://www.nytimes.com/2010/10/28/technology/28compute.html?_r=1&partner=rss&emc=rss">
Tianhe-1A</a> designed by
<a href="http://english.nudt.edu.cn">The National University of Defense Technology (NUDT)</a>
in China with 14,336 Intel CPUs and 7,168 NVDIA Tesla M2050 GPUs, with a peak performance of 2.507 Petaflops.</li>
in China with 14,336 Intel CPUs and 7,168 NVDIA Tesla M2050 GPUs,
with a peak performance of 2.507 Petaflops.</li>
<li><a href="http://www-hpc.cea.fr/en/complexe/tgcc-curie.htm">TGCC
Curie</a>, owned by
GENCI and operated in
to
the TGCC by
CEA, Curie
is offering 3 different fractions
of x86-64 computing resources
for addressing a wide range of scientific
challenges and offering
an aggregate peak performance of 2 PetaFlops.</li>
<li><a href="http://www-hpc.cea.fr/en/complexe/tgcc-curie.htm">TGCC
Curie</a>,
owned by <a href="http://www.genci.fr">
GENCI
</a>
and operated in the TGCC by
<a href="http://www.cea.fr">CEA</a>, Curie
is offering 3 different fractions
of x86-64 computing resources
for addressing a wide range of scientific
challenges and offering
an aggregate peak performance of 2 PetaFlops.</li>
<li><a href="http://www.wcm.bull.com/internet/pr/rend.jsp?DocId=567851&lang=en">
Tera 100</a> at <a href="http://www.cea.fr">CEA</a>
with 140,000 Intel Xeon 7500 processing cores, 300TB of
central memory and a theoretical computing power of 1.25 Petaflops.</li>
<li><a href="http://hpc.msu.ru/?q=node/59">Lomonosov</a>, a
<a href="http://www.t-platforms.com">T-Platforms</a> system at
<a href="http://hpc.msu.ru">Moscow State University Research Computing Center</a>
with 52,168 Intel Xeon processing cores and 8,840 NVIDIA GPUs.</li>
<li><a href="http://compeng.uni-frankfurt.de/index.php?id=86">LOEWE-CSC</a>,
a combined CPU-GPU Linux cluster
at
<a href="http://csc.uni-frankfurt.de">The Center for Scientific
Computing (CSC)</a>
of the Goethe University Frankfurt, Germany,
a combined CPU-GPU Linux cluster
at
<a href="http://csc.uni-frankfurt.de">The Center for Scientific
Computing (CSC)</a>
of the Goethe University Frankfurt, Germany,
with 20,928 AMD Magny-Cours CPU cores (176 Teraflops peak
performance) plus 778 ATI Radeon 5870 GPUs (2.1 Petaflops peak
performance single precision and 599 Teraflops double precision) and
QDR Infiniband interconnect.</li>
<li><a href="https://asc.llnl.gov/computing_resources/sequoia/">Dawn</a>,
a BlueGene/P system at <a href="https://www.llnl.gov">LLNL</a>
with 147,456 PowerPC 450 cores with a peak
performance of 0.5 Petaflops.</li>
<li><a href="http://www.cscs.ch/compute_resources">Rosa</a>,
a CRAY XT5 at the <a href="http://www.cscs.ch">Swiss National Supercomputer Centre</a>
a <a href="http://www.cray.com">Cray</a> XT5 at the
<a href="http://www.cscs.ch">Swiss National Supercomputer Centre</a>
named after Monte Rosa in the Swiss-Italian Alps, elevation 4,634m.
3,688 AMD hexa-core Opteron @ 2.4 GHz, 28.8 TB DDR2 RAM, 290 TB Disk,
9.6 GB/s interconnect bandwidth (Seastar).</li>
</ul>
<p style="text-align:center;">Last modified
2 Octo
ber 2012</p>
<p style="text-align:center;">Last modified
7 Decem
ber 2012</p>
<!--#include virtual="footer.txt"-->
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment