diff --git a/doc/html/accounting.shtml b/doc/html/accounting.shtml index 94674f70497b0c822fb274679152754a0bcd490e..254c2a9c506d634bcd09f35f73eeb209d4c7f73e 100644 --- a/doc/html/accounting.shtml +++ b/doc/html/accounting.shtml @@ -316,10 +316,7 @@ user names and optional partition name.) <b>Without AccountingStorageEnforce being set (the default behavior) jobs will be executed based upon policies configured in Slurm on each cluster.</b> -<br> -It is advisable to run without the 'limits' option set when running a -scheduler on top of Slurm like Moab that does not update in real -time its limits per association.</li> +</li> <li><b>AccountingStorageHost</b>: The name or address of the host where SlurmDBD executes</li> @@ -715,7 +712,7 @@ as deleted. If an entity has existed for less than 1 day, the entity will be removed completely. This is meant to clean up after typographic errors.</p> -<p style="text-align:center;">Last modified 6 April 2015</p> +<p style="text-align:center;">Last modified 15 December 2016</p> <!--#include virtual="footer.txt"--> diff --git a/doc/html/configurator.html.in b/doc/html/configurator.html.in index 869ab18f5247ca61e961de6704054a2f45460145..ed9ecdbfa05cda95927b43b194ab31e8fea1cbf0 100644 --- a/doc/html/configurator.html.in +++ b/doc/html/configurator.html.in @@ -195,10 +195,8 @@ function displayfile() "#DefMemPerCPU=0 <br>" + "FastSchedule=" + get_radio_value(document.config.fast_schedule) + "<br>" + "#MaxMemPerCPU=0 <br>" + - "#SchedulerRootFilter=1 <br>" + "#SchedulerTimeSlice=30 <br>" + "SchedulerType=sched/" + get_radio_value(document.config.sched_type) + "<br>" + - get_field("SchedulerPort",document.config.scheduler_port) + "<br>" + "SelectType=select/" + get_radio_value(document.config.select_type) + "<br>" + get_select_type_params() + "<br>" + "# <br>" + @@ -488,13 +486,6 @@ Select one value for <B>SchedulerType</B>:<BR> FIFO with backfill<BR> <input type="radio" name="sched_type" value="builtin"> <B>Builtin</B>: First-In First-Out (FIFO)<BR> -<input type="radio" name="sched_type" value="wiki"> <B>Wiki</B>: Wiki interface -to Maui (configuration parameter <B>SchedulerPort</B> must specified)<BR> -<input type="radio" name="sched_type" value="wiki2"> <B>Wiki2</B>: Wiki interface -to Moab (configuration parameter <B>SchedulerPort</B> must specified)<BR> -<P> -<input type="text" name="scheduler_port" value="7321"> <B>SchedulerPort</B>: scheduler -communications port (used by Wiki and Wiki2 only) <P> Define what node configuration (sockets, cores, memory, etc.) should be used. Using values defined in the configuration file will provide faster scheduling.<BR> @@ -816,5 +807,5 @@ before terminating all remaining tasks. A value of zero indicates unlimited wait </FORM> <HR> <a href="disclaimer.html" target="_blank" class="privacy">Legal Notices</a><br> -Last modified 3 February 2015</P> +Last modified 15 December 2016</P> </BODY> diff --git a/doc/html/faq.shtml b/doc/html/faq.shtml index 8b2ff3577e8521fa451ea864dc003298c738cb3b..fb0014aae8d7c2ddc7678d5ee5b5f8a47f7340b7 100644 --- a/doc/html/faq.shtml +++ b/doc/html/faq.shtml @@ -131,8 +131,6 @@ indicate?</a></li> <li><a href="#task_prolog">How could I automatically print a job's Slurm job ID to its standard output?</a></li> -<li><a href="#moab_start">I run Slurm with the Moab or Maui scheduler. - How can I start a job under Slurm without the scheduler?</a></li> <li><a href="#orphan_procs">Why are user processes and <i>srun</i> running even though the job is supposed to be completed?</a></li> <li><a href="#slurmd_oom">How can I prevent the <i>slurmd</i> and @@ -286,13 +284,7 @@ time limit (that associated with the partition), and the ability to backfill sch jobs will be limited. The backfill scheduler does not alter job specifications of required or excluded nodes, so jobs which specify nodes will substantially reduce the effectiveness of backfill scheduling. See the <a href="#backfill"> -backfill</a> section for more details. If the scheduler type is <b>wiki</b>, -this represents -<a href="http://www.clusterresources.com/pages/products/maui-cluster-scheduler.php"> -The Maui Scheduler</a> or -<a href="http://www.clusterresources.com/pages/products/moab-cluster-suite.php"> -Moab Cluster Suite</a>. -Please refer to its documentation for help. For any scheduler, you can check priorities +backfill</a> section for more details. For any scheduler, you can check priorities of jobs using the command <span class="commandline">scontrol show job</span>.</p> <p><a name="sharing"><b>4. Why does the srun --overcommit option not permit multiple jobs @@ -1653,27 +1645,7 @@ then fi </pre> -<p><a name="moab_start"><b>34. I run Slurm with the Moab or Maui scheduler. -How can I start a job under Slurm without the scheduler?</b></a></br> -When Slurm is configured to use the Moab or Maui scheduler, all submitted -jobs have their priority initialized to zero, which Slurm treats as a held -job. The job only begins when Moab or Maui decide where and when to start -the job, setting the required node list and setting the job priority to -a non-zero value. To circumvent this, submit your job using a Slurm or -Moab command then manually set its priority to a non-zero value (must be -done by user root). For example:</p> -<pre> -$ scontrol update jobid=1234 priority=1000000 -</pre> -<p>Note that changes in the configured value of <i>SchedulerType</i> only -take effect when the <i>slurmctld</i> daemon is restarted (reconfiguring -Slurm will not change this parameter. You will also manually need to -modify the priority of every pending job. -When changing to Moab or Maui scheduling, set every job priority to zero. -When changing from Moab or Maui scheduling, set every job priority to a -non-zero value (preferably fairly large, say 1000000).</p> - -<p><a name="orphan_procs"><b>35. Why are user processes and <i>srun</i> +<p><a name="orphan_procs"><b>34. Why are user processes and <i>srun</i> running even though the job is supposed to be completed?</b></a></br> Slurm relies upon a configurable process tracking plugin to determine when all of the processes associated with a job or job step have completed. @@ -1682,7 +1654,7 @@ Those plugins dependent upon process group IDs or parent process IDs are not reliable. See the <i>ProctrackType</i> description in the <i>slurm.conf</i> man page for details. We rely upon the sgi_job for most systems.</p> -<p><a name="slurmd_oom"><b>36. How can I prevent the <i>slurmd</i> and +<p><a name="slurmd_oom"><b>35. How can I prevent the <i>slurmd</i> and <i>slurmstepd</i> daemons from being killed when a node's memory is exhausted?</b></a></br> You can set the value in the <i>/proc/self/oom_adj</i> for @@ -1691,7 +1663,7 @@ daemon with the <i>SLURMD_OOM_ADJ</i> and/or <i>SLURMSTEPD_OOM_ADJ</i> environment variables set to the desired values. A value of -17 typically will disable killing.</p> -<p><a name="ubuntu"><b>37. I see my host of my calling node as 127.0.1.1 +<p><a name="ubuntu"><b>36. I see my host of my calling node as 127.0.1.1 instead of the correct IB address. Why is that?</b></a></br> Some systems by default will put your host in the /etc/hosts file as something like</p> @@ -1709,7 +1681,7 @@ name from the "127.0.1.1" as well as the "127.0.0.1" addresses in the /etc/hosts file. It is ok if they point to localhost, but not the actual name of the node.</p> -<p><a name="stop_sched"><b>38. How can I stop Slurm from scheduling jobs?</b></a></br> +<p><a name="stop_sched"><b>37. How can I stop Slurm from scheduling jobs?</b></a></br> You can stop Slurm from scheduling jobs on a per partition basis by setting that partition's state to DOWN. Set its state UP to resume scheduling. For example: @@ -1718,7 +1690,7 @@ $ scontrol update PartitionName=foo State=DOWN $ scontrol update PartitionName=bar State=UP </pre></p> -<p><a name="scontrol_multi_jobs"><b>39. Can I update multiple jobs with a +<p><a name="scontrol_multi_jobs"><b>38. Can I update multiple jobs with a single <i>scontrol</i> command?</b></a></br> No, but you can probably use <i>squeue</i> to build the script taking advantage of its filtering and formatting options. For example: @@ -1726,7 +1698,7 @@ advantage of its filtering and formatting options. For example: $ squeue -tpd -h -o "scontrol update jobid=%i priority=1000" >my.script </pre></p> -<p><a name="amazon_ec2"><b>40. Can Slurm be used to run jobs on +<p><a name="amazon_ec2"><b>39. Can Slurm be used to run jobs on Amazon's EC2?</b></a></br> <p>Yes, here is a description of use Slurm use with <a href="http://aws.amazon.com/ec2/">Amazon's EC2</a> courtesy of @@ -1750,7 +1722,7 @@ which I then copy over the /usr/local on the first instance and NFS export to all other instances. This way I have persistent home directories and a very simple first-login script that configures the virtual cluster for me.</p> -<p><a name="core_dump"><b>41. If a Slurm daemon core dumps, where can I find the +<p><a name="core_dump"><b>40. If a Slurm daemon core dumps, where can I find the core file?</b></a></br> <p>For <i>slurmctld</i>, the core file will be in the same directory as its log files (<i>SlurmctldLogFile</i>) if configured using an fully qualified @@ -1788,7 +1760,7 @@ Oct 15 11:31:20 knc abrtd: Directory 'ccpp-2015-10-15-11:31:20-21477' creation d in the core_pattern. For example:<br> sysctl kernel.core_pattern=/tmp/core.%e.%p</p> -<p><a name="totalview"><b>42. How can TotalView be configured to operate with +<p><a name="totalview"><b>41. How can TotalView be configured to operate with Slurm?</b></a></br> <p>The following lines should also be added to the global <i>.tvdrc</i> file for TotalView to operate with Slurm: @@ -1823,7 +1795,7 @@ dset TV::parallel_configs { } !--> -<p><a name="git_patch"><b>43. How can a patch file be generated from a Slurm +<p><a name="git_patch"><b>42. How can a patch file be generated from a Slurm commit in github?</b></a></br> <p>Find and open the commit in github then append ".patch" to the URL and save the resulting file. For an example, see: @@ -1831,7 +1803,7 @@ the resulting file. For an example, see: https://github.com/SchedMD/slurm/commit/91e543d433bed11e0df13ce0499be641774c99a3.patch</a> </p> -<p><a name="enforce_limits"><b>44. Why are the resource limits set in the +<p><a name="enforce_limits"><b>43. Why are the resource limits set in the database not being enforced?</b></a></br> In order to enforce resource limits, set the value of <b>AccountingStorageEnforce</b> in each cluster's slurm.conf configuration @@ -1839,7 +1811,7 @@ file appropriately. If <b>AccountingStorageEnforce</b> does not contains an option of "limits", then resource limits will not be enforced on that cluster. See <a href="resource_limits.html">Resource Limits</a> for more information.</p> -<p><a name="restore_priority"><b>45. After manually setting a job priority +<p><a name="restore_priority"><b>44. After manually setting a job priority value, how can it's priority value be returned to being managed by the priority/multifactor plugin?</b></a></br> Hold and then release the job as shown below.</p> @@ -1848,14 +1820,14 @@ $ scontrol hold <jobid> $ scontrol release <jobid> </pre> -<p><a name="health_check_example"><b>46. Does any one have an example node +<p><a name="health_check_example"><b>45. Does any one have an example node health check script for Slurm?</b></a></br> Probably the most comprehensive and lightweight health check tool out there is <a href="https://github.com/mej/nhc">Node Health Check</a>. It has integration with Slurm as well as Torque resource managers.</p> -<p><a name="add_nodes"><b>47. What process should I follow to add nodes to Slurm?</b></a></br> +<p><a name="add_nodes"><b>46. What process should I follow to add nodes to Slurm?</b></a></br> The slurmctld daemon has a multitude of bitmaps to track state of nodes and cores in the system. Adding nodes to a running system would require the slurmctld daemon to rebuild all of those bitmaps, which the developers feel would be safer to do by @@ -1876,7 +1848,7 @@ NOTE: Jobs submitted with srun, and that are waiting for an allocation, prior to new nodes being added to the slurm.conf can fail if the job is allocated one of the new nodes. -<p><a name="licenses"><b>48. Can Slurm be configured to manage licenses?</b></a></br> +<p><a name="licenses"><b>47. Can Slurm be configured to manage licenses?</b></a></br> Slurm is not currently integrated with FlexLM, but it does provide for the allocation of global resources called licenses. Use the Licenses configuration parameter in your slurm.conf file (e.g. "Licenses=foo:10,bar:20"). @@ -1887,7 +1859,7 @@ without restarting the slurmctld daemon, but it is possible to dynamically reserve licenses and remove them from being available to jobs on the system (e.g. "scontrol update reservation=licenses_held licenses=foo:5,bar:2").</p> -<p><a name="salloc_default_command"><b>49. Can the salloc command be configured to +<p><a name="salloc_default_command"><b>48. Can the salloc command be configured to launch a shell on a node in the job's allocation?</b></a></br> Yes, just use the SallocDefaultCommand configuration parameter in your slurm.conf file as shown below.</p> @@ -1902,11 +1874,11 @@ SallocDefaultCommand="srun -n1 -N1 --mem-per-cpu=0 --gres=craynetwork:0 --pty -- </pre> </p> -<p><a name="upgrade"><b>50. What should I be aware of when upgrading Slurm?</b></a></br> +<p><a name="upgrade"><b>49. What should I be aware of when upgrading Slurm?</b></a></br> See the Quick Start Administrator Guide <a href="quickstart_admin.html#upgrade">Upgrade</a> section for details.</p> -<p><a name="torque"><b>51. How easy is it to switch from PBS or Torque to Slurm?</b></a></br> +<p><a name="torque"><b>50. How easy is it to switch from PBS or Torque to Slurm?</b></a></br> A lot of users don't even notice the difference. Slurm has wrappers available for the mpiexec, pbsnodes, qdel, qhold, qrls, qstat, and qsub commands (see contribs/torque in the distribution and the @@ -1927,7 +1899,7 @@ The plugin will be installed in<br> <install_directory>/lib/slurm/spank_pbs.so<br> See the SPANK man page for configuration details.</p> -<p><a name="sssd"><b>52. I am having trouble using SSSD with Slurm.</b></a></br> +<p><a name="sssd"><b>51. I am having trouble using SSSD with Slurm.</b></a></br> SSSD or System Security Services Deamon does not allow enumeration of group members by default. Note that enabling enumeration in large environments might not be feasible. However, as of version 16.05 slurm @@ -1943,7 +1915,7 @@ enumerate = True case_sensitive = False </pre> -<p><a name="ha_db"><b>53. How critical is configuring high availability for my +<p><a name="ha_db"><b>52. How critical is configuring high availability for my database?</b></a></br> <ul> <li>Consider if you really need mysql failover. Short outage of slurmdbd is not @@ -1962,7 +1934,7 @@ IP between master/backup mysql servers and:</li> </ul> </ul> -<p><a name="sql"><b>54.How can I use double quotes in MySQL queries?</b></a></br> +<p><a name="sql"><b>53.How can I use double quotes in MySQL queries?</b></a></br> Execute: <pre> SET session sql_mode='ANSI_QUOTES'; @@ -1972,7 +1944,7 @@ SET session sql_mode='ANSI_QUOTES'; show columns from "tux_assoc_table" where Field='is_def'; </pre> -<p><a name="reboot"><b>55. Why is a compute node down with the reason set to +<p><a name="reboot"><b>54. Why is a compute node down with the reason set to "Node unexpectedly rebooted"?</b></a></br> This is indicative of the slurmctld daemon running on the cluster's head node as well as the slurmd daemon on the compute node when the compute node reboots. @@ -1980,7 +1952,7 @@ If you want to prevent this condition from setting the node into a DOWN state then configure ReturnToService to 2. See the slurm.conf man page for details. Otherwise use the scontrol or sview to manually return the node to service.</p> -<p><a name="reqspec"><b>56. How can a job which has exited with a specific exit +<p><a name="reqspec"><b>55. How can a job which has exited with a specific exit code be requeued?</b></a></br> Slurm supports requeue in hold with a <b>SPECIAL_EXIT</b> state using the command:</p> @@ -2056,7 +2028,7 @@ we touch the file /tmp/myfile, then release the job which will finish in COMPLETE state. </p> -<p><a name="user_account"><b>57. Can a user's account be changed in the database?</b></a></br> +<p><a name="user_account"><b>56. Can a user's account be changed in the database?</b></a></br> A user's account can not be changed directly. A new association needs to be created for the user with the new account. Then the association with the old account can be deleted.</p> @@ -2066,13 +2038,13 @@ sacctmgr create user name=adam cluster=tux account=physics sacctmgr delete user name=adam cluster=tux account=chemistry </pre> -<p><a name="mpi_perf"><b>58. What might account for MPI performance being below +<p><a name="mpi_perf"><b>57. What might account for MPI performance being below the expected level?</b></a><br> Starting the slurmd daemons with limited locked memory can account for this. Adding the line "ulimit -l unlimited" to <i>/etc/sysconfig/slurm</i> file can fix this.</p> -<p><a name="state_info"><b>59. How could some jobs submitted immediately before +<p><a name="state_info"><b>58. How could some jobs submitted immediately before the slurmctld daemon crashed be lost?</b></a><br> Any time the slurmctld daemon or hardware fails before state information reaches disk can result in lost state. @@ -2084,7 +2056,7 @@ information is cached in memory rather than written to disk when the node fails. The interval between state saves being written to disk can be configured at build time by defining SAVE_MAX_WAIT to a different value than five.</p> -<p><a name="delete_partition"><b>60. How do I safely remove partitions? +<p><a name="delete_partition"><b>59. How do I safely remove partitions? </b></a><br> Partitions should be removed using the "scontrol delete PartitionName=<partition>" command. This is because @@ -2097,7 +2069,7 @@ Removing a partition from the slurm.conf and restarting will cancel any existing jobs that reference the removed partitions. </p> -<p><a name="cpu_freq"><b>61. Why is Slurm unable to set the CPU frequency for +<p><a name="cpu_freq"><b>60. Why is Slurm unable to set the CPU frequency for jobs?</b></a><br> First check that Slurm is configured to bind jobs to specific CPUs by making sure that TaskPlugin is configured to either affinity or cgroup. @@ -2115,7 +2087,7 @@ https://bugzilla.kernel.org/show_bug.cgi?id=57141</a> and<br> <a href="http://unix.stackexchange.com/questions/121410/setting-cpu-governor-to-on-demand-or-conservative"> http://unix.stackexchange.com/questions/121410/setting-cpu-governor-to-on-demand-or-conservative</a>.</p> -<p><a name="mic_config"><b>62. How can Slurm be configured to support Intel +<p><a name="mic_config"><b>61. How can Slurm be configured to support Intel Xeon Phi (MIC)?</b></a><br> Users should see the <a href="#mic">Xeon Phi use information</a> above. Slurm configuration details for Xeon Phi offload support are available @@ -2142,14 +2114,14 @@ install the latest MPSS and Slurm packages from yum/zypper, add new MICs (via console utility or GUI), add MICs to Slurm queues if necessary, restart the host, use MICs via Slurm.</p> -<p><a name="cluster_acct"><b>63. When adding a new cluster, how can the Slurm cluster +<p><a name="cluster_acct"><b>62. When adding a new cluster, how can the Slurm cluster configuration be copied from an existing cluster to the new cluster?</b></a><br> Accounts need to be configured the cluster. An easy way to copy information from an existing cluster is to use the sacctmgr command to dump that cluster's information, modify it using some editor, the load the new information using the sacctmgr command. See the sacctmgr man page for details, including an example.</p> -<p><a name="cray_dvs"><b>64. How can I update Slurm on a Cray DVS file system +<p><a name="cray_dvs"><b>63. How can I update Slurm on a Cray DVS file system without rebooting the nodes?</b></a><br> The problem with DVS caching is related to the fact that the dereferenced value of /opt/slurm/default symlink is cached in the DVS attribute cache, and that @@ -2176,7 +2148,7 @@ before starting the slurmds again.</p> <p>(Information courtesy of Douglas Jacobsen, NERSC, Lawrence Berkeley National Laboratory)</p> -<p><a name="dbd_rebuild"><b>65. How can I rebuild the database hierarchy?</b></a><br> +<p><a name="dbd_rebuild"><b>64. How can I rebuild the database hierarchy?</b></a><br> If you see errors of this sort:</p> <pre> error: Can't find parent id 3358 for assoc 1504, this should never happen. @@ -2187,7 +2159,7 @@ administrator error in directly modifying the database. In order to rebuild the database information, start the slurmdbd daemon with the "-R" option followed by an optional comma separated list of cluster names to operate on.</p> -<p><a name="routing queue"><b>66. How can a routing queue be configured?</b></a><br> +<p><a name="routing queue"><b>65. How can a routing queue be configured?</b></a><br> A job submit plugin is designed to have access to a job request from a user, plus information about all of the available system partitions/queue. An administrator can write a C plugin or LUA script to set an incoming job's @@ -2199,6 +2171,6 @@ Also see the available job submit plugins distributed with Slurm for examples <p class="footer"><a href="#top">top</a></p> -<p style="text-align:center;">Last modified 19 July 2016</p> +<p style="text-align:center;">Last modified 15 December 2016</p> <!--#include virtual="footer.txt"--> diff --git a/doc/html/man_index.shtml b/doc/html/man_index.shtml index d0ab80ce810c66267087e3867d7f9d36720af770..b225bb143fc28caec9b194d1a918ec2717ba81ac 100644 --- a/doc/html/man_index.shtml +++ b/doc/html/man_index.shtml @@ -46,7 +46,6 @@ Documentation for other versions of Slurm is distributed with the code</b></p> <tr><td><a href="slurm.conf.html">slurm.conf</a></td><td>Slurm configuration file</td></tr> <tr><td><a href="slurmdbd.conf.html">slurmdbd.conf</a></td><td>Slurm Database Daemon (SlurmDBD) configuration file</td></tr> <tr><td><a href="topology.conf.html">topology.conf</a></td><td>Slurm configuration file for defining the network topology</td></tr> -<tr><td><a href="wiki.conf.html">wiki.conf</a></td><td>Slurm configuration file for wiki and wiki2 scheduler plugins</td></tr> </table> <h2>Daemons and Other</h2> @@ -59,6 +58,6 @@ Documentation for other versions of Slurm is distributed with the code</b></p> </table> -<p style="text-align:center;">Last modified 8 December 2015</p> +<p style="text-align:center;">Last modified 15 December 2016</p> <!--#include virtual="footer.txt"--> diff --git a/doc/html/quickstart_admin.shtml b/doc/html/quickstart_admin.shtml index b5fb827deb9ea0174251c1801b41508ade7918e9..4efea29a6f74691e6a9a2892888934dc89cce37d 100644 --- a/doc/html/quickstart_admin.shtml +++ b/doc/html/quickstart_admin.shtml @@ -339,12 +339,6 @@ backfill scheduling does require users to specify job time limits.</li> <li><i>gang</i> time-slices jobs in the same partition/queue and can be used to preempt jobs from lower-priority queues in order to execute jobs in higher priority queues.</li> -<li><i>wiki</i> is an interface for use with -<a href="http://www.clusterresources.com/pages/products/maui-cluster-scheduler.php"> -The Maui Scheduler</a></li> -<li><i>wiki2</i> is an interface for use with the -<a href="http://www.clusterresources.com/pages/products/moab-cluster-suite.php"> -Moab Cluster Suite</a> </ul> <p>For more information about scheduling options see @@ -849,6 +843,6 @@ options such as mysql and gui tools via a configuration menu.</p> </pre> <p class="footer"><a href="#top">top</a></p> -<p style="text-align:center;">Last modified 17 August 2016</p> +<p style="text-align:center;">Last modified 15 December 2016</p> <!--#include virtual="footer.txt"--> diff --git a/doc/html/reservations.shtml b/doc/html/reservations.shtml index 0cb0549c799c1ebeb2fe46c1df7c392c517ff0ec..8d2ab5218ada653471a23d23aba7e131e4a4bcc0 100644 --- a/doc/html/reservations.shtml +++ b/doc/html/reservations.shtml @@ -27,10 +27,6 @@ The <i>scontrol</i>, <i>smap</i> and <i>sview</i> commands can be used to view reservations. The man pages for the various commands contain details.</p> -<p>Note for users of Maui or Moab schedulers: <br> -Maui and Moab are not integrated with Slurm's resource reservation system, -but should use their own advanced reservation system.</p> - <h2>Reservation Creation</h2> <p>One common mode of operation for a reservation would be to reserve @@ -433,7 +429,7 @@ considering the initiation of jobs. This will prevent the initiation of some jobs which would complete execution before a reservation given fewer jobs to time-slice with.</p> -<p style="text-align: center;">Last modified 28 January 2016</p> +<p style="text-align: center;">Last modified 15 December 2016</p> <!--#include virtual="footer.txt"--> diff --git a/doc/html/resource_limits.shtml b/doc/html/resource_limits.shtml index d59401556c7e112084d22cbc9431c036ba8b1835..1f0f1dffd754b459901232eaea2a40b7d7af4b6b 100644 --- a/doc/html/resource_limits.shtml +++ b/doc/html/resource_limits.shtml @@ -5,10 +5,6 @@ <p>Familiarity with Slurm's <a href="accounting.html">Accounting</a> web page is strongly recommended before use of this document.</p> -<p><b>Note for users of Maui or Moab schedulers:</b><br> -Maui and Moab are not integrated with Slurm's resource limits, -but should use their own resource limits mechanisms.</p> - <h2>Hierarchy</h2> <p>Slurm's hierarchical limits are enforced in the following order @@ -97,10 +93,6 @@ user names and optional partition name.) Without AccountingStorageEnforce being set (the default behavior) jobs will be executed based upon policies configured in Slurm on each cluster. -<br> -It is advisable to run without the 'limits' option set when running a -scheduler on top of Slurm, like Moab, that does not update in real -time its limits per association. </p> <h2>Tools</h2> @@ -262,6 +254,6 @@ data maintained in the Slurm database. More information can be found in the <a href="priority_multifactor.html">priority/multifactor</a> plugin description.</p> -<p style="text-align: center;">Last modified 19 October 2015</p> +<p style="text-align: center;">Last modified 15 December 2016</p> </ul></body></html> diff --git a/doc/html/schedplugins.shtml b/doc/html/schedplugins.shtml index 4856db853dfe193b5942e048295d68d986466815..eb64a697d35f84be18056efea7ce03461313c188 100644 --- a/doc/html/schedplugins.shtml +++ b/doc/html/schedplugins.shtml @@ -10,9 +10,6 @@ scheduler plugins.</p> <p>It is noteworthy that two different models are used for job scheduling. The <b>backfill</b> scheduler let. Slurm establish the initial job priority and can periodically alter job priorities to change their order within the queue. -The <b>wiki</b> scheduler establishes an initial priority of zero (held) for -all jobs. These jobs only begin execution when the <b>wiki</b> scheduler -explicitly raises their priority (releasing them). Developers may use the model that best fits their needs. Note that a separate <a href="selectplugins.html">node selection plugin</a> is available for controlling that aspect of scheduling.</p> @@ -28,15 +25,6 @@ abbreviation for the type of scheduler. We recommend, for example:</p> scheduling services. This is the default behavior and implements first-in-first-out scheduling.</li> <li><b>backfill</b> — Raise the priority of jobs if doing so results in their starting earlier without any delay in the expected initiation time of any higher priority job.</li> -<li><b>wiki</b> — Uses -<a href="http://www.clusterresources.com/pages/products/maui-cluster-scheduler.php"> -The Maui Scheduler</a> (Wiki version) -as an external entity to control Slurm job scheduling.</li> -<li><b>wiki2</b> — Uses -<a href="http://www.clusterresources.com/pages/products/moab-cluster-suite.php"> -Moab Cluster Suite</a> as an external entity to control Slurm job scheduling. -Note that wiki2 is an expanded version of the wiki plugin with additional -functions supported specifically for Moab.</li> </ul> <p><span class="commandline">const char plugin_name[]</span><br> @@ -186,6 +174,6 @@ specific error code.</p> or NULL if no description found in this plugin.</p> <p class="footer"><a href="#top">top</a></p> -<p style="text-align:center;">Last modified 27 March 2015</p> +<p style="text-align:center;">Last modified 15 December 2016</p> <!--#include virtual="footer.txt"--> diff --git a/doc/html/troubleshoot.shtml b/doc/html/troubleshoot.shtml index 8697401749e7d71b17b4b96d551e887a5db4a904..de860ba0527e2a95e98d9df106ce894599d395dc 100644 --- a/doc/html/troubleshoot.shtml +++ b/doc/html/troubleshoot.shtml @@ -105,13 +105,6 @@ reduce the effectiveness of backfill scheduling. See the <a href="faq.html#backfill">backfill documentation</a> for more details.</li> -<li>If the scheduler type is <i>wiki</i>, this represents -<a href="http://www.clusterresources.com/pages/products/maui-cluster-scheduler.php"> -The Maui Scheduler</a> or -<a href="http://www.clusterresources.com/pages/products/moab-cluster-suite.php"> -Moab Cluster Suite</a>. -Please refer to its documentation for help.</li> -</ul> <p class="footer"><a href="#top">top</a></p> @@ -305,6 +298,6 @@ partition 000. </ol> <p class="footer"><a href="#top">top</a></p> -<p style="text-align:center;">Last modified 15 April 2015</p> +<p style="text-align:center;">Last modified 15 December 2016</p> <!--#include virtual="footer.txt"--> diff --git a/doc/man/man5/Makefile.am b/doc/man/man5/Makefile.am index 6f29f48046438112d50b32a6d565aeaba2f039b9..4f60a36850eb1414610562423c0eafb27b566989 100644 --- a/doc/man/man5/Makefile.am +++ b/doc/man/man5/Makefile.am @@ -12,8 +12,7 @@ man5_MANS = \ nonstop.conf.5 \ slurm.conf.5 \ slurmdbd.conf.5 \ - topology.conf.5 \ - wiki.conf.5 + topology.conf.5 EXTRA_DIST = $(man5_MANS) @@ -31,8 +30,7 @@ html_DATA = \ nonstop.conf.html \ slurm.conf.html \ slurmdbd.conf.html \ - topology.conf.html \ - wiki.conf.html + topology.conf.html MOSTLYCLEANFILES = ${html_DATA} diff --git a/doc/man/man5/Makefile.in b/doc/man/man5/Makefile.in index af7b6edf929d5f11d95b83780cf9c527dca98a01..ff19c717f5b092e0c2698b5d3fad37efd25804e8 100644 --- a/doc/man/man5/Makefile.in +++ b/doc/man/man5/Makefile.in @@ -469,8 +469,7 @@ man5_MANS = \ nonstop.conf.5 \ slurm.conf.5 \ slurmdbd.conf.5 \ - topology.conf.5 \ - wiki.conf.5 + topology.conf.5 EXTRA_DIST = $(man5_MANS) $(am__append_1) @HAVE_MAN2HTML_TRUE@html_DATA = \ @@ -485,8 +484,7 @@ EXTRA_DIST = $(man5_MANS) $(am__append_1) @HAVE_MAN2HTML_TRUE@ nonstop.conf.html \ @HAVE_MAN2HTML_TRUE@ slurm.conf.html \ @HAVE_MAN2HTML_TRUE@ slurmdbd.conf.html \ -@HAVE_MAN2HTML_TRUE@ topology.conf.html \ -@HAVE_MAN2HTML_TRUE@ wiki.conf.html +@HAVE_MAN2HTML_TRUE@ topology.conf.html @HAVE_MAN2HTML_TRUE@MOSTLYCLEANFILES = ${html_DATA} @HAVE_MAN2HTML_TRUE@SUFFIXES = .html diff --git a/doc/man/man5/slurm.conf.5 b/doc/man/man5/slurm.conf.5 index 89f58f72a28068368bc0d907df571edcde549b04..c34b9b541034ac7d4c45f26047595323665d4fd9 100644 --- a/doc/man/man5/slurm.conf.5 +++ b/doc/man/man5/slurm.conf.5 @@ -655,9 +655,6 @@ including state, job ids and allocated nodes counter. .TP \fBTriggers\fR Slurmctld triggers -.TP -\fBWiki\fR -Sched/wiki and wiki2 communications .RE .TP @@ -1690,9 +1687,6 @@ scheduling be enabled (i.e, "PreemptMode=SUSPEND,GANG"). .TP 12 \fBOFF\fR is the default value and disables job preemption and gang scheduling. -This is the only option compatible with \fBSchedulerType=sched/wiki\fR -or \fBSchedulerType=sched/wiki2\fR (used by Maui and Moab respectively, -which provide their own job preemption functionality). .TP \fBCANCEL\fR always cancel the job. @@ -1859,9 +1853,8 @@ Cleared on the first day of each year at time 00:00. \fBPriorityType\fR This specifies the plugin to be used in establishing a job's scheduling priority. Supported values are "priority/basic" (jobs are prioritized -by order of arrival, also suitable for sched/wiki and sched/wiki2), -"priority/multifactor" (jobs are prioritized based upon size, age, -fair\-share of allocation, etc). +by order of arrival), "priority/multifactor" (jobs are prioritized based +upon size, age, fair\-share of allocation, etc). Also see \fBPriorityFlags\fR for configuration options. The default value is "priority/basic". @@ -2794,12 +2787,6 @@ priority job. \fBsched/hold\fR To hold all newly arriving jobs if a file "/etc/slurm.hold" exists otherwise use the built\-in FIFO scheduler -.TP -\fBsched/wiki\fR -For the Wiki interface to the Maui Scheduler -.TP -\fBsched/wiki2\fR -For the Wiki interface to the Moab Cluster Suite .RE .TP @@ -5203,4 +5190,4 @@ details. \fBgetrlimit\fR (2), \fBgres.conf\fR(5), \fBgroup\fR (5), \fBhostname\fR (1), \fBscontrol\fR(1), \fBslurmctld\fR(8), \fBslurmd\fR(8), \fBslurmdbd\fR(8), \fBslurmdbd.conf\fR(5), \fBsrun(1)\fR, -\fBspank(8)\fR, \fBsyslog\fR (2), \fBtopology.conf\fR(5), \fBwiki.conf\fR(5) +\fBspank(8)\fR, \fBsyslog\fR (2), \fBtopology.conf\fR(5) diff --git a/doc/man/man5/wiki.conf.5 b/doc/man/man5/wiki.conf.5 deleted file mode 100644 index aab5453249e0e370ccb427f211a7a06731e09cdd..0000000000000000000000000000000000000000 --- a/doc/man/man5/wiki.conf.5 +++ /dev/null @@ -1,206 +0,0 @@ -.TH "wiki.conf" "5" "Slurm Configuration File" "April 2015" "Slurm Configuration File" - -.SH "NAME" -wiki.conf \- Slurm configuration file for wiki and wiki2 scheduler plugins -.SH "DESCRIPTION" -\fBwiki.conf\fP is an ASCII file which describes wiki and wiki2 -scheduler specific Slurm configuration information. -The file location can be modified at system build time using the -DEFAULT_SLURM_CONF parameter or at execution time by setting the SLURM_CONF -environment variable. The file will always be located in the -same directory as the \fBslurm.conf\fP file. -.LP -Parameter names are case insensitive. -Any text following a "#" in the configuration file is treated -as a comment through the end of that line. -Changes to the configuration file take effect upon restart of -Slurm daemons, daemon receipt of the SIGHUP signal, or execution -of the command "scontrol reconfigure" unless otherwise noted. -.LP -The overall configuration parameters available include: - -.TP -\fBAuthKey\fR -Authentication key for communications. -This should be no more than a 32\-bit unsigned integer and match the -KEY configured in the \fBmoab\-private.cnf\fR file (for the Moab Scheduler) -or the \fB\-\-with-key=\fR value used to configure the Maui Scheduler. - -.TP -\fBEHost\fR -Name the computer on which Moab server executes. -It is used in establishing a communications path for event notification. -By default \fBEHost\fR will be identical in value to the -\fBControlAddr\fR configured in slurm.conf. -Not applicable to wiki plugin, only the wiki2 plugin. - -.TP -\fBEHostBackup\fR -Name the computer on which the backup Moab server executes. -It is used in establishing a communications path for event notification. -By default \fBEHostBackup\fR will be identical in value to the -\fBBackupAddr\fR configured in slurm.conf. -Not applicable to wiki plugin, only the wiki2 plugin. - -.TP -\fBEPort\fR -Port to be used to notify Moab of events (job submitted to Slurm, -job terminates, etc.). -This numeric value should match EPORT configured in the -\fBmoab.cnf\fR file. -Not applicable to wiki plugin, only the wiki2 plugin. - -.TP -\fBExcludePartitions\fR -Identifies partitions whose jobs are to be scheduled directly -by Slurm rather than Moab/Maui. -This only effects jobs which are submitted using Slurm -commands (i.e. srun, salloc or sbatch, NOT msub from Moab). -These jobs will be scheduled on a First\-Come\-First\-Served -basis directly by Slurm. -Note that Slurm recognizes jobs submitted via msub based -upon the value \fBFirstJobId\fR configured in \fIslurm.conf\fR. -Set the values \fBMINJOBID\fR and \fBMAXJOBID\fR in \fImoab.cfg\fR -accordingly. -For example MINJOBID=1, MAXJOBID=65535 and FirstJobId=65536. -Jobs submitted using msub will have job ID values in the range -of 1 and 65535 while jobs submitted directly using Slurm commands -will have a job ID of 65536 or higher. -Moab/Maui controls for resource reservation, fair share -scheduling, etc. will not apply to the initiation of these jobs. -While Moab/Maui will not control the initiation of jobs in these -partitions, it will account for and report the jobs. -If more than one partition is to be scheduled directly by -Slurm, use a comma separator between their names. -This may provide faster response times than Moab/Maui scheduling. - -.TP -\fBHidePartitionJobs\fR -Identifies partitions whose jobs are not to be reported to Moab/Maui. -These jobs will not be accounted for or otherwise visible to Moab/Maui. -Any partitions listed here must also be listed in \fBExcludePartitions\fR. -This only effects jobs which are submitted using Slurm commands (i.e. -\fIsrun\fR, \fIsalloc\fR or \fIsbatch\fR, NOT \fImsub\fR from Moab). -If more than one partition is to have its jobs hidden, use a comma -separator between their names. - -.TP -\fBHidePartitionNodes\fR -Identifies partitions whose nodes are not to be reported to Moab/Maui. -Nodes which are configured only in hidden partitions will not be accounted -for or otherwise visible to Moab/Maui. -Any partitions included here will normally also be included in the -\fBExcludePartitions\fR and \fBHidePartitionJobs\fR parameters. -If more than one partition is to have its jobs hidden, use a comma -separator between their names. - -.TP -\fBHostFormat\fR -Controls the format of host lists exchanged between Slurm and Moab. -The default value is "0". -Not applicable to wiki plugin, only the wiki2 plugin. - -.RS -.TP -\fB0\fR -No data compression. Each host name is listed individually. -.TP -\fB1\fR -Slurm hostlist expressions are exchanged with task counts -(e.g. "tux[0\-16]*2") in job state information and job -initiation requests. -.TP -\fB2\fR -Slurm hostlist expressions are used to report node state information. -.RE - -.TP -\fBJobAggregationTime\fR -This is used to avoid notifying Moab of large numbers of events -occurring about the same time. -If an event occurs within this number of seconds since Moab was -last notified of an event, another notification is not sent. -This should be an integer number of seconds. -The default value is 10 seconds. -The value should match JOBAGGREGATIONTIME configured in the -\fBmoab.cnf\fR file. -Not applicable to wiki plugin, only the wiki2 plugin. - -.TP -\fBJobPriority\fR -Controls initial job priority. -The default value is "hold". -Not applicable to wiki plugin, only the wiki2 plugin. -.RS -.TP -\fBrun\fR -Job permitted to run directly under Slurm's control -.TP -\fBhold\fR -Hold all incoming jobs until Moab or Maui tell them to run -.RE - -.SH "EXAMPLE" -.LP -.br -################################################################## -.br -# Slurm's configuration file for sched/wiki plugin -.br -################################################################## -.br -# -.br -# Hold all incoming jobs until Moab tells them to run -.br -JobPriority=hold -.br -# Slurm directly schedules jobs in the debug and test partitions -.br -ExcludePartitions=debug,test -.br -# -.br -# Moab encryption key, matches KEY in moab\-private.cfg -.br -AuthKey=1234 -.br -# -.br -# Moab event notification port, matches EPORT in moab.cfg -.br -EPort=15017 -.br -# Moab event notification hosts, where Moab executes -.br -EHost=tux0 -.br -EHostBackup=tux1 -.br -# Moab event notification throttle, matches JOBAGGREGATIONTIME -.br -# in moab.cfg (integer value in seconds) -.br -JobAggregationTime=15 - -.SH "COPYING" -Copyright (C) 2006-2007 The Regents of the University of California. -Produced at Lawrence Livermore National Laboratory (cf, DISCLAIMER). -CODE\-OCEC\-09\-009. All rights reserved. -.LP -This file is part of Slurm, a resource management program. -For details, see <https://slurm.schedmd.com/>. -.LP -Slurm is free software; you can redistribute it and/or modify it under -the terms of the GNU General Public License as published by the Free -Software Foundation; either version 2 of the License, or (at your option) -any later version. -.LP -Slurm is distributed in the hope that it will be useful, but WITHOUT ANY -WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS -FOR A PARTICULAR PURPOSE. See the GNU General Public License for more -details. - -.SH "SEE ALSO" -.LP -\fBslurm.conf\fR(5)