From 352476f4e7fa6e9f3700c0ba1952d9d35cdd07ed Mon Sep 17 00:00:00 2001 From: Morris Jette <jette@schedmd.com> Date: Tue, 4 Oct 2011 13:40:01 -0700 Subject: [PATCH] Correct faq.html numbering --- doc/html/faq.shtml | 14 +++++++------- 1 file changed, 7 insertions(+), 7 deletions(-) diff --git a/doc/html/faq.shtml b/doc/html/faq.shtml index ca707b1bfa8..2110ab5becf 100644 --- a/doc/html/faq.shtml +++ b/doc/html/faq.shtml @@ -88,7 +88,7 @@ SLURM? Why does the DAKOTA program not run with SLURM?</a></li> core files?</a></li> <li><a href="#limit_propagation">Is resource limit propagation useful on a homogeneous cluster?</a></li> -<li<a href="#clock">Do I need to maintain synchronized clocks +<li><a href="#clock">Do I need to maintain synchronized clocks on the cluster?</a></li> <li><a href="#cred_invalid">Why are "Invalid job credential" errors generated?</a></li> @@ -1396,7 +1396,7 @@ address instead of the correct address and make it so the communication doesn't work. Solution is to either remove this line or set a different nodeaddr that is known by your other nodes.</p> -<p><a name="stop_sched"><b>38. How can I stop SLURM from scheduling jobs?</b></a></br> +<p><a name="stop_sched"><b>39. How can I stop SLURM from scheduling jobs?</b></a></br> You can stop SLURM from scheduling jobs on a per partition basis by setting that partition's state to DOWN. Set its state UP to resume scheduling. For example: @@ -1405,7 +1405,7 @@ $ scontrol update PartitionName=foo State=DOWN $ scontrol update PartitionName=bar State=UP </pre></p> -<p><a name="scontrol_multi_jobs"><b>39. Can I update multiple jobs with a +<p><a name="scontrol_multi_jobs"><b>40. Can I update multiple jobs with a single <i>scontrol</i> command?</b></a></br> No, but you can probably use <i>squeue</i> to build the script taking advantage of its filtering and formatting options. For example: @@ -1413,7 +1413,7 @@ advantage of its filtering and formatting options. For example: $ squeue -tpd -h -o "scontrol update jobid=%i priority=1000" >my.script </pre></p> -<p><a name="amazon_ec2"><b>40. Can SLURM be used to run jobs on +<p><a name="amazon_ec2"><b>41. Can SLURM be used to run jobs on Amazon's EC2?</b></a></br> <p>Yes, here is a description of use SLURM use with <a href="http://aws.amazon.com/ec2/">Amazon's EC2</a> courtesy of @@ -1437,7 +1437,7 @@ which I then copy over the /usr/local on the first instance and NFS export to all other instances. This way I have persistent home directories and a very simple first-login script that configures the virtual cluster for me.</p> -<p><a name="core_dump"><b>41. If a SLURM daemon core dumps, where can I find the +<p><a name="core_dump"><b>42. If a SLURM daemon core dumps, where can I find the core file?</b></a></br> <p>For <i>slurmctld</i> the core file will be in the same directory as its log files (<i>SlurmctldLogFile</i>) iif configured using an fully qualified @@ -1453,7 +1453,7 @@ Otherwise it will be found in directory used for saving state occurs. It will either be in spawned job's working directory on the same location as that described above for the <i>slurmd</i> daemon.</p> -<p><a name="totalview"><b>42. How can TotalView be configured to operate with +<p><a name="totalview"><b>43. How can TotalView be configured to operate with SLURM?</b></a></br> <p>The following lines should also be added to the global <i>.tvdrc</i> file for TotalView to operate with SLURM: @@ -1470,7 +1470,7 @@ dset TV::parallel_configs { } </pre></p> -<p><a name="git_patch"><b>43. How can a patch file be generated from a SLURM +<p><a name="git_patch"><b>44. How can a patch file be generated from a SLURM commit in github?</b></a></br> <p>Find and open the commit in github then append ".patch" to the URL and save the resutling file. For an example, see: -- GitLab