Newer
Older
.TH SCONTROL "1" "March 2003" "scontrol 0.1" "Slurm components"
.SH "NAME"
scontrol \- Used view and modify Slurm configuration and state.
.SH "SYNOPSIS"
\fBscontrol\fR [\fIOPTIONS\fR...] [\fICOMMAND\fR...]
.SH "DESCRIPTION"
\fBscontrol\fR is used to view or modify Slurm configuration including: job,
job step, node, partition, and overall system configuration. Most of the
commands can only be executed by user root. If an attempt to view or modify
configuration information is made by an unauthorized user, an error message
will be printed and the requested action will not occur. If no command is
entered on the execute line, \fBscontrol\fR will operate in an interactive
mode and prompt for input. It will continue prompting for input and executing
commands until explicitly terminated. If a command is entered on the execute
line, \fBscontrol\fR will execute that command and terminate. All commands
and options are case-insensitive, although node names and partition names
are case-sensitive (node names "LX" and "lx" are distinct).
.TP
.TP
\fB-h\fR
Print a help message describing the usage of scontrol.
.TP
\fB-q\fR
Print no warning or informational messages, only fatal error messages.
.TP
\fB-v\fR
Print detailed event logging. This includes time-stamps on data structures,
record counts, etc.
.TP
\fIabort\fP
Instruct the Slurm controller to terminate immediately and generate a core file.
.TP
\fIexit\fP
Terminate the execution of scontrol.
.TP
\fIhelp\fP
Display a description of scontrol options and commands.
.TP
\fIpid2jid\fP \fIPROC_ID\fP
Print the Slurm job id corresponding to the supplied process id, \fIPROC_ID\fP,
on the current node.
.TP
\fIping\fP
Ping the primary and secondary slurmctld daemon and report if
they are responding.
.TP
\fIquiet\fP
Print no warning or informational messages, only fatal error messages.
.TP
\fIquit\fP
Terminate the execution of scontrol.
.TP
\fIreconfigure\fP
Instruct all Slurm daemons to re-read the configuration file.
This mechanism would be used to modify configuration parameters (Epilog,
Prolog, SlurmctldLogFile, SlurmdLogFile, etc.) register the physical
addition or removal of nodes from the cluster or recognize the change
of a node's configuration, such as the addition of memory or processors.
The Slurm controller (slurmctld) forwards the request all other daemons
(slurmd daemon on each compute node). Running jobs continue execution.
Display the state of the specified entity with the specified identification.
\fIENTITY\fP may be \fIconfig\fP, \fIdaemons\fP, \fIjob\fP, \fInode\fP,
\fIpartition\fP or \fIstep\fP.
\fIID\fP can be used to identify a specific element of the identified
entity: the configuration parameter name, job ID, node name, partition name,
or job step ID for entities \fIconfig\fP, \fIjob\fP, \fInode\fP, \fIpartition\fP,
and \fIstep\fP respectively.
Multiple node names may be specified using simple node range expressions
(e.g. "lx[10-20]"). All other \fIID\fP values must identify a single
element. The job step ID is of the form "job_id.step_id", (e.g. "1234.1").
By default, all elements of the entity type specified are printed.
Instruct all Slurm daemons to save current state and terminate.
The Slurm controller (slurmctld) forwards the request all other daemons
(slurmd daemon on each compute node).
\fIupdate\fP \fISPECIFICATION\fP
Update job, node or partition configuration per the supplied specification.
\fISPECIFICATION\fP is in the same format as the Slurm configuration file
and the output of the \fIshow\fP command described above. It may be desirable
to execute the \fIshow\fP command (described above) on the specific entity
you which to update, then use cut-and-paste tools to enter updated configuration
values to the \fIupdate\fP. Note that while most configuration values can be
changed using this command, not all can be changed using this mechanism. In
particular, the hardware configuration of a node or the physical addition or
removal of nodes from the cluster may only be accomplished through editing
the Slurm configuration file and executing the \fIreconfigure\fP command
(described above).
.TP
\fIverbose\fP
Print detailed event logging. This includes time-stamps on data structures,
record counts, etc.
.TP
\fIversion\fP
Display the version number of scontrol being executed.
.TP
\fI!!\fP
Repeat the last command executed.
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
.TP
\fBSPECIFICATIONS FOR UPDATE COMMAND, JOBS\fR
.TP
\fIContiguous\fP=<yes|no>
Set the job's requirement for contiguous (consecutive) nodes to be allocated.
Possible values are"YES" and "NO".
.TP
\fIFeatures\fP=<features>
Set the job's required features on nodes specified value. Multiple values
may be comma separated if all features are required (AND operation) or
separated by "|" if any of the specified features are required (OR operation).
.TP
\fIJobId\fP=<id>
Identify the job to be updated. This specification is required.
.TP
\fIMinMemory\fP=<megabytes>
Set the job's minimum real memory required per nodes to the specified value.
.TP
\fIMinProcs\fP=<count>
Set the job's minimum number of processors per nodes to the specified value.
.TP
\fIMinTmpDisk\fP=<megabytes>
Set the job's minimum temporary disk space required per nodes to the specified value.
.TP
\fIName\fP=<name>
Set the job's name to the specified value.
.TP
\fIPartition\fP=<name>
Set the job's partition to the specified value.
.TP
\fIPriority\fP=<minutes>
Set the job's priority to the specified value.
.TP
\fIReqNodeList\fP=<nodes>
Set the job's list of required node. Multiple node names may be specified using
simple node range expressions (e.g. "lx[10-20]").
.TP
\fIReqNodes\fP=<count>
Set the job's count of required nodes to the specified value.
.TP
\fIReqProcs\fP=<count>
Set the job's count of required processors to the specified value.
.TP
\fIShared\fP=<yes|no>
Set the job's ability to share nodes with other jobs. Possible values are
"YES" and "NO".
.TP
\fITimeLimit\fP=<minutes>
Set the job's time limit to the specified value.
.TP
\fBSPECIFICATIONS FOR UPDATE COMMAND, NODES\fR
.TP
\fINodeName\fP=<name>
Identify the node(s) to be updated. Multiple node names may be specified using
simple node range expressions (e.g. "lx[10-20]"). This specification is required.
.TP
\fIState\fP=<state>
Identify the state to be assigned to the node. Possible values are "NoResp",
"DOWN", "IDLE", "DRAIN", "DRAINED", "DRAINING", and "ALLOCATED".
To drain a node specify "DRAIN", "DRAINED", or "DRAINING".
SLURM will automatically set it to the appropriate value of either
"DRAINING" or "DRAINED" depending if the node is allocated or not.
The "NoResp" state will only set the "NoResp" flag for a node without
changing its underlying state.
.TP
\fBSPECIFICATIONS FOR UPDATE COMMAND, PARTITIONS\fR
.TP
\fIAllowGroups\fP=<name>
Identify the user groups which may use this partition. Multiple groups
may be specified in a comma separated list.
.TP
\fIDefault\fP=<yes|no>
Specify if this partition is to be used by jobs which do not explicitly
identify a partition to use. Possible values are"YES" and "NO".
.TP
\fINodes\fP=<name>
Identify the node(s) to be associated with this partition. Multiple node names
may be specified using simple node range expressions (e.g. "lx[10-20]").
Note that jobs may only be associated with one partition at any time.
.TP
\fIPartitionName\fP=<name>
Identify the partition to be updated. This specification is required.
.TP
\fIRootOnly\fP=<yes|no>
Specify if only allocation requests initiated by user root will be satisfied.
This can be used to restrict control of the partition to some meta-scheduler.
Possible values are"YES" and "NO".
.TP
\fIShared\fP=<yes|no|force>
Specify if nodes in this partition can be shared by multiple jobs.
Possible values are"YES", "NO" and "FORCE".
.TP
\fIState\fP=<up|down>
Specify if jobs can be allocated nodes in this partition.
Possible values are"UP" and "DOWN".
If a partition allocated nodes to running jobs, those jobs will continue
execution even after the partition's state is set to "DOWN". The jobs
.TP
\fIMaxNodes\fP=<count>
Set the maximum number of nodes which will be allocated to any single job
in the partition. Specify a number or "INFINITE".
.TP
\fIMinNodes\fP=<count>
Set the minimum number of nodes which will be allocated to any single job
in the partition.
.SH "EXAMPLE"
.eo
.br
# scontrol
.br
scontrol: show part class
.br
PartitionName=class MaxTime=30 MaxNodes=2 TotalNodes=10
TotalCPUs=160 RootOnly=NO Default=NO Shared=NO State=UP
.br
Nodes=lx[0031-0040] AllowGroups=students
scontrol: update PartitionName=class MaxTime=99 MaxNodes=4
JobId=65539 UserId=1500 JobState=PENDING TimeLimit=100
Priority=100 Partition=batch Name=job01 NodeList=(null)
StartTime=0 EndTime=0 Shared=0 ReqProcs=1000
.br
ReqNodes=400 Contiguous=1 MinProcs=4 MinMemory=1024
MinTmpDisk=2034ReqNodeList=lx[3000-3003]
.br
Features=(null) JobScript=/bin/hostname
.br
scontrol: update JobId=65539 TimeLimit=200 Priority=500
.br
scontrol: quit
.ec
.SH "COPYING"
Copyright (C) 2002 The Regents of the University of California.
Produced at Lawrence Livermore National Laboratory (cf, DISCLAIMER).
UCRL-CODE-2002-040.
.LP
This file is part of SLURM, a resource management program.
For details, see <http://www.llnl.gov/linux/slurm/>.
.LP
SLURM is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free
Software Foundation; either version 2 of the License, or (at your option)
any later version.
.LP
SLURM is distributed in the hope that it will be useful, but WITHOUT ANY
WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
details.
.SH "FILES"
.LP
/etc/slurm.conf
\fBscancel\fR(1), \fBsinfo\fR(1), \fBsqueue\fR(1),
\fBslurm_load_ctl_conf\fR(3), \fBslurm_load_jobs\fR(3), \fBslurm_load_node\fR(3),
\fBslurm_load_partitions\fR(3),
\fBslurm_reconfigure\fR(3), \fBslurm_shutdown\fR(3),
\fBslurm_update_job\fR(3), \fBslurm_update_node\fR(3), \fBslurm_update_partition\fR(3)