diff --git a/doc/html/slurm.shtml b/doc/html/slurm.shtml
index 85051ac2b26bc9529155ae4a9d8c37928978697c..5c1c55fe1d6de561527acf22c9f198ff4ece95be 100644
--- a/doc/html/slurm.shtml
+++ b/doc/html/slurm.shtml
@@ -1,8 +1,8 @@
 <!--#include virtual="header.txt"-->
 
-<h1>SLURM Workload Manager</h1>
+<h1>Slurm Workload Manager</h1>
 
-<p>SLURM is an open-source workload manager designed for Linux clusters of
+<p>Slurm is an open-source workload manager designed for Linux clusters of
 all sizes.
 It provides three key functions.
 First it allocates exclusive and/or non-exclusive access to resources
@@ -12,7 +12,7 @@ Second, it provides a framework for starting, executing, and monitoring work
 Finally, it arbitrates contention for resources by managing a queue of
 pending work. </p>
 
-<p>SLURM's design is very modular with dozens of optional plugins.
+<p>Slurm's design is very modular with dozens of optional plugins.
 In its simplest configuration, it can be installed and configured in a
 couple of minutes (see <a href="http://www.linux-mag.com/id/7239/1/">
 Caos NSA and Perceus: All-in-one Cluster Software Stack</a>
@@ -28,7 +28,7 @@ world-class computer centers and rely upon a
 or supporting sophisticated
 <a href="priority_multifactor.html">job prioritization</a> algorithms.</p>
 
-<p>While other workload managers do exist, SLURM is unique in several
+<p>While other workload managers do exist, Slurm is unique in several
 respects:
 <ul>
 <li><b>Scalability</b>: It is designed to operate in a heterogeneous cluster
@@ -55,52 +55,63 @@ can specify size and time limit ranges.</li>
 help identify load imbalances and other anomalies.</li>
 </ul></p>
 
-<p>SLURM provides workload management on many of the most powerful computers in
+<p>Slurm provides workload management on many of the most powerful computers in
 the world including:
 <ul>
 <li><a href="https://asc.llnl.gov/computing_resources/sequoia/">Sequoia</a>,
-a BlueGene/Q system at <a href="https://www.llnl.gov">LLNL</a>
+an <a href="http://www.ibm.com">IBM</a> BlueGene/Q system at
+<a href="https://www.llnl.gov">Lawrence Livermore National Laboratory</a>
 with 1.6 petabytes of memory, 96 racks, 98,304 compute nodes, and 1.6
-  million cores, with a peak performance of over 20 Petaflops.</li>
+million cores, with a peak performance of over 20 Petaflops.</li>
+
+<li><a href="http://www.tacc.utexas.edu/stampede">Stampede</a> at the
+<a href="http://www.tacc.utexas.edu">Texas Advanced Computing Center/University of Texas</a>
+is a <a herf="http://www.dell.com">Dell</a> with over
+80,000 <a href="http://www.intel.com">Intel</a> Xeon cores,
+Intel Phi co-processors, plus
+128 <a href="http://www.nvidia.com">NVIDIA</a> GPUs
+delivering 2.66 Petaflops.</li>
 
 <li><a href="http://www.nytimes.com/2010/10/28/technology/28compute.html?_r=1&partner=rss&emc=rss">
 Tianhe-1A</a> designed by 
 <a href="http://english.nudt.edu.cn">The National University of Defense Technology (NUDT)</a>
-in China with 14,336 Intel CPUs and 7,168 NVDIA Tesla M2050 GPUs, with a peak performance of 2.507 Petaflops.</li>
+in China with 14,336 Intel CPUs and 7,168 NVDIA Tesla M2050 GPUs,
+with a peak performance of 2.507 Petaflops.</li>
 
-<li><a href="http://www-hpc.cea.fr/en/complexe/tgcc-curie.htm">TGCC
-    Curie</a>, owned by GENCI and operated into the TGCC by CEA, Curie
-    is offering 3 different fractions of x86-64 computing resources
-    for addressing a wide range of scientific challenges and offering
-    an aggregate peak performance of 2 PetaFlops.</li>
+<li><a href="http://www-hpc.cea.fr/en/complexe/tgcc-curie.htm">TGCC Curie</a>,
+owned by <a href="http://www.genci.fr">GENCI</a> and operated in the TGCC by
+<a href="http://www.cea.fr">CEA</a>, Curie is offering 3 different fractions
+of x86-64 computing resources for addressing a wide range of scientific
+challenges and offering an aggregate peak performance of 2 PetaFlops.</li>
 
 <li><a href="http://www.wcm.bull.com/internet/pr/rend.jsp?DocId=567851&lang=en">
 Tera 100</a> at <a href="http://www.cea.fr">CEA</a>
 with 140,000 Intel Xeon 7500 processing cores, 300TB of 
 central memory and a theoretical computing power of 1.25 Petaflops.</li>
 
+<li><a href="http://hpc.msu.ru/?q=node/59">Lomonosov</a>, a
+<a href="http://www.t-platforms.com">T-Platforms</a> system at
+<a href="http://hpc.msu.ru">Moscow State University Research Computing Center</a> 
+with 52,168 Intel Xeon processing cores and 8,840 NVIDIA GPUs.</li>
+
 <li><a href="http://compeng.uni-frankfurt.de/index.php?id=86">LOEWE-CSC</a>,
-a combined CPU-GPU Linux cluster
-at <a href="http://csc.uni-frankfurt.de">The Center for Scientific
-Computing (CSC)</a> of the Goethe University Frankfurt, Germany,
+a combined CPU-GPU Linux cluster at
+<a href="http://csc.uni-frankfurt.de">The Center for Scientific Computing (CSC)</a>
+of the Goethe University Frankfurt, Germany,
 with 20,928 AMD Magny-Cours CPU cores (176 Teraflops peak
 performance) plus 778 ATI Radeon 5870 GPUs (2.1 Petaflops peak
 performance single precision and 599 Teraflops double precision) and
 QDR Infiniband interconnect.</li>
 
-<li><a href="https://asc.llnl.gov/computing_resources/sequoia/">Dawn</a>,
-a BlueGene/P system at <a href="https://www.llnl.gov">LLNL</a>
-with 147,456 PowerPC 450 cores with a peak
-performance of 0.5 Petaflops.</li>
-
 <li><a href="http://www.cscs.ch/compute_resources">Rosa</a>,
-a CRAY XT5 at the <a href="http://www.cscs.ch">Swiss National Supercomputer Centre</a>
+a <a href="http://www.cray.com">Cray</a> XT5 at the
+<a href="http://www.cscs.ch">Swiss National Supercomputer Centre</a>
 named after Monte Rosa in the Swiss-Italian Alps, elevation 4,634m.
 3,688 AMD hexa-core Opteron @ 2.4 GHz, 28.8 TB DDR2 RAM, 290 TB Disk,
 9.6 GB/s interconnect bandwidth (Seastar).</li>
 
 </ul>
 
-<p style="text-align:center;">Last modified 2 October 2012</p>
+<p style="text-align:center;">Last modified 7 December 2012</p>
 
 <!--#include virtual="footer.txt"-->