Skip to content
Snippets Groups Projects
Commit d9e35c96 authored by Moe Jette's avatar Moe Jette
Browse files

General updates to web pages.

parent 3de706e1
No related branches found
No related tags found
No related merge requests found
......@@ -4,41 +4,10 @@
<h2>Index</h2>
<ul>
<li><a href="#20">SLURM Version 2.0, May 2009</a></li>
<li><a href="#21">SLURM Version 2.1, January 2010</a></li>
<li><a href="#22">SLURM Version 2.2, available in late 2010</a></li>
<li><a href="#23">SLURM Version 2.3 and beyond</a></li>
</ul>
<h2><a name="20">Major Updates in SLURM Version 2.0</a></h2>
<p>SLURM Version 2.0 was released in May 2009.
Major enhancements include:
<ul>
<li>Sophisticated <a href="priority_multifactor.html">job prioritization
plugin</a> is now available.
Jobs can be prioritized based upon their age, size and/or fair-share resource
allocation using hierarchical bank accounts.</li>
<li>An assortment of <a href="resource_limits.html">resource limits</a>
can be imposed upon individual users and/or hierarchical bank accounts
such as maximum job time limit, maximum job size, and maximum number of
running jobs.</li>
<li><a href="reservations.html">Advanced reservations</a> can be made to
insure resources will be available when needed.</li>
<li>Idle nodes can now be completely <a href="power_save.html">powered
down</a> when idle and automatically restarted when their is work
available.</li>
<li>Jobs in higher priority partitions (queues) can automatically
<a href="preempt.html">preempt</a> jobs in lower priority queues.
The preempted jobs will automatically resume execution upon completion
of the higher priority job.</li>
<li>Specific cores are allocated to jobs and jobs steps in order to effective
preempt or gang schedule jobs.</li>
<li>A new configuration parameter, <i>PrologSlurmctld</i>, can be used to
support the booting of different operating systems for each job.</li>
<li>Added switch topology configuration options to optimize job resource
allocation with respect to communication performance.</li>
<li>Automatic <a href="checkpoint_blcr.html">Checkpoint/Restart using BLCR</a>
is now available.</li>
<li><a href="#23">SLURM Version 2.3, available in 2011</a></li>
<li><a href="#24">SLURM Version 2.4 and beyond</a></li>
</ul>
<h2><a name="21">Major Updates in SLURM Version 2.1</a></h2>
......@@ -79,21 +48,31 @@ different versions of the commands and deamons to interoperate.</li>
<li>Permit SLURM commands to operate between clusters (e.g. status jobs on a
different cluster or submit a job on one cluster to run on another).</li>
<li>Major enhancements for high-throughput computing. Job throughput
rates now exceed 120,000 jobs per hour.</li>
rates now exceed 120,000 jobs per hour with bursts of job submissions at
several times that rate.</li>
</ul>
<h2><a name="23">Major Updates in SLURM Version 2.3</a></h2>
<p>SLURM Version 2.3 release is planned in 2011.
Major enhancements currently planned include:
<ul>
<li>Support for Cray computers (integration with ALPS/BASIL).</li>
<li>Support for BlueGene/Q computers.</li>
<li>Integration with FlexLM license management.</li>
<li>Numerous enhancements to advanced resource reservations.</li>
</ul>
<h2><a name="23">Major Updates in SLURM Version 2.3 and beyond</a></h2>
<p> Detailed plans for release dates and contents of future SLURM releases have
not been finalized. Anyone desiring to perform SLURM development should notify
<a href="mailto:slurm-dev@lists.llnl.gov">slurm-dev@lists.llnl.gov</a>
<h2><a name="23">Major Updates in SLURM Version 2.4 and beyond</a></h2>
<p> Detailed plans for release dates and contents of additional SLURM releases
have not been finalized. Anyone desiring to perform SLURM development should
notify <a href="mailto:slurm-dev@lists.llnl.gov">slurm-dev@lists.llnl.gov</a>
to coordinate activities. Future development plans includes:
<ul>
<li>Support for BlueGene/Q systems.</li>
<li>Add Kerberos credential support including credential forwarding
and refresh.</li>
<li>Provide a web-based SLURM administration tool.</li>
</ul>
<p style="text-align:center;">Last modified 30 August 2010</p>
<p style="text-align:center;">Last modified 31 August 2010</p>
<!--#include virtual="footer.txt"-->
......@@ -17,31 +17,27 @@ In its simplest configuration, it can be installed and configured in a
couple of minutes (see <a href="http://www.linux-mag.com/id/7239/1/">
Caos NSA and Perceus: All-in-one Cluster Software Stack</a>
by Jeffrey B. Layton).
More complex configurations rely upon a
More complex configurations can satisfy the job scheduling needs of
world-class computer centers and rely upon a
<a href="http://www.mysql.com/">MySQL</a> database for archiving
<a href="accounting.html">accounting</a> records, managing
<a href="resource_limits.html">resource limits</a> by user or bank account,
or supporting sophisticated
<a href="priority_multifactor.html">job prioritization</a> algorithms.
SLURM also provides an Applications Programming Interface (API) for
integration with external schedulers such as
<a href="http://www.clusterresources.com/pages/products/maui-cluster-scheduler.php">
The Maui Scheduler</a> or
<a href="http://www.clusterresources.com/pages/products/moab-cluster-suite.php">
Moab Cluster Suite</a>.</p>
<a href="priority_multifactor.html">job prioritization</a> algorithms.</p>
<p>While other resource managers do exist, SLURM is unique in several
respects:
<ul>
<li>Its source code is freely available under the
<a href="http://www.gnu.org/licenses/gpl.html">GNU General Public License</a>.</li>
<li>It is designed to operate in a heterogeneous cluster with up to 65,536 nodes
and hundreds of thousands of processors.</li>
<li>It can sustain a throughput rate of over 120,000 jobs per hour.</li>
<li>It can sustain a throughput rate of over 120,000 jobs per hour with
bursts of job submissions at several times that rate.</li>
<li>Its source code is freely available under the
<a href="http://www.gnu.org/licenses/gpl.html">GNU General Public License</a>.</li>
<li>It is portable; written in C with a GNU autoconf configuration engine.
While initially written for Linux, other UNIX-like operating systems should
be easy porting targets.</li>
<li>SLURM is highly tolerant of system failures, including failure of the node
<li>It is highly tolerant of system failures, including failure of the node
executing its control functions.</li>
<li>A plugin mechanism exists to support various interconnects, authentication
mechanisms, schedulers, etc. These plugins are documented and simple enough
......@@ -82,10 +78,9 @@ and a three-dimensional torus interconnect.</li>
<a href="http://www.bull.com">Bull</a>.
It is also distributed and supported by
<a href="http://www.adaptivecomputing.com">Adaptive Computing</a>,
<a href="http://www.infiscale.com">Infiscale</a>,
<a href="http://www.ibm.com">IBM</a> and
<a href="http://www.sun.com">Sun Microsystems</a>.</p>
<a href="http://www.infiscale.com">Infiscale</a> and
<a href="http://www.ibm.com">IBM</a>.</p>
<p style="text-align:center;">Last modified 30 August 2010</p>
<p style="text-align:center;">Last modified 31 August 2010</p>
<!--#include virtual="footer.txt"-->
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment