diff --git a/doc/man/man1/sacctmgr.1 b/doc/man/man1/sacctmgr.1 index 6a70748f78d2be5d5c257f61ce71eb67574293a0..0b785d7f492b97444253500c00b9876f16b58272 100644 --- a/doc/man/man1/sacctmgr.1 +++ b/doc/man/man1/sacctmgr.1 @@ -239,7 +239,7 @@ To clear a previously set value use the modify command with a new value of \-1. .TP \fIGrpCPUMins\fP=<max cpu minutes> Maximum number of CPU minutes running jobs are able to be allocated in -aggregate for this association and all association which are children +aggregate for this association and all associations which are children of this association. To clear a previously set value use the modify command with a new value of \-1. @@ -258,7 +258,7 @@ they are able to run inside the limit. .TP \fIGrpCPUs\fP=<max cpus> Maximum number of CPUs running jobs are able to be allocated in aggregate for -this association and all association which are children of this association. +this association and all associations which are children of this association. To clear a previously set value use the modify command with a new value of \-1. .P @@ -268,26 +268,26 @@ Resource plugin. .TP \fIGrpJobs\fP=<max jobs> Maximum number of running jobs in aggregate for -this association and all association which are children of this association. +this association and all associations which are children of this association. To clear a previously set value use the modify command with a new value of \-1. .TP \fIGrpNodes\fP=<max nodes> Maximum number of nodes running jobs are able to be allocated in aggregate for -this association and all association which are children of this association. +this association and all associations which are children of this association. To clear a previously set value use the modify command with a new value of \-1. .TP \fIGrpSubmitJobs\fP=<max jobs> Maximum number of jobs which can be in a pending or running state at any time -in aggregate for this association and all association which are children of +in aggregate for this association and all associations which are children of this association. To clear a previously set value use the modify command with a new value of \-1. .TP \fIGrpWall\fP=<max wall> Maximum wall clock time running jobs are able to be allocated in aggregate for -this association and all association which are children of this association. +this association and all associations which are children of this association. To clear a previously set value use the modify command with a new value of \-1. .P NOTE: This limit is not enforced if set on the root association of a @@ -373,7 +373,7 @@ Valid <operator> values include: Set \fIQosLevel\fP to the specified value. .TP \fB+=\fR -Add the specified <qos> value to the current \fIQosLevel\fP . +Add the specified <qos> value to the current \fIQosLevel\fP. .TP \fB\-=\fR Remove the specified <qos> value from the current \fIQosLevel\fP. @@ -534,34 +534,34 @@ Number used in conjunction with other accounts to determine job priority. .TP \fIGrpCPUMins\fP Maximum number of CPU minutes running jobs are able to be allocated in -aggregate for this association and all association which are children +aggregate for this association and all associations which are children of this association. .TP \fIGrpCPUs\fP Maximum number of CPUs running jobs are able to be allocated in aggregate for -this association and all association which are children of this association. +this association and all associations which are children of this association. .TP \fIGrpJobs\fP Maximum number of running jobs in aggregate for -this association and all association which are children of this association. +this association and all associations which are children of this association. .TP \fIGrpNodes\fP Maximum number of nodes running jobs are able to be allocated in aggregate for -this association and all association which are children of this association. +this association and all associations which are children of this association. .TP \fIGrpSubmitJobs\fP Maximum number of jobs which can be in a pending or running state at any time -in aggregate for this association and all association which are children of +in aggregate for this association and all associations which are children of this association. .TP \fIGrpWall\fP Maximum wall clock time running jobs are able to be allocated in aggregate for -this association and all association which are children of this association. +this association and all associations which are children of this association. .TP \fIID\fP @@ -569,8 +569,8 @@ The id of the association. .TP \fILFT\fP -Associations are kept in a hierarchy this is the left most -spot in the hierarchy. When used with the RGT variable all +Associations are kept in a hierarchy: this is the left most +spot in the hierarchy. When used with the RGT variable, all associations with a LFT inside this LFT and before the RGT are children of this association. @@ -620,8 +620,8 @@ The numeric values of valid QOS\' for this association. .TP \fIRGT\fP -Associations are kept in a hierarchy this is the right most -spot in the hierarchy. When used with the LFT variable all +Associations are kept in a hierarchy: this is the right most +spot in the hierarchy. When used with the LFT variable, all associations with a LFT inside this RGT and after the LFT are children of this association. @@ -1261,7 +1261,7 @@ cluster. Comments can be put into the file with the # character. Each line of information must begin with one of the four titles; \fBCluster, Parent, Account or User\fP. Following the title is a space, dash, space, entity value, then specifications. Specifications are colon separated. If any -variable such as Organization has a space in it surround the name with +variable such as Organization has a space in it, surround the name with single or double quotes. To create a file of associations one can run @@ -1298,35 +1298,35 @@ associations on this cluster. These options are as follows... .TP \fIGrpCPUMins=\fP Maximum number of CPU hours running jobs are able to -be allocated in aggregate for this association and all association +be allocated in aggregate for this association and all associations which are children of this association. (NOTE: This limit is not currently enforced in SLURM. You can still set this, but have to wait for future versions of SLURM before it is enforced.) .TP \fIGrpCPUs=\fP Maximum number of CPUs running jobs are able to be -allocated in aggregate for this association and all association which +allocated in aggregate for this association and all associations which are children of this association. (NOTE: This limit is not currently enforced in SLURM. You can still set this, but have to wait for future versions of SLURM before it is enforced.) .TP \fIGrpJobs=\fP Maximum number of running jobs in aggregate for this -association and all association which are children of this association. +association and all associations which are children of this association. .TP \fIGrpNodes=\fP Maximum number of nodes running jobs are able to be -allocated in aggregate for this association and all association which +allocated in aggregate for this association and all associations which are children of this association. .TP \fIGrpSubmitJobs=\fP Maximum number of jobs which can be in a pending or running state at any time in aggregate for this association and all -association which are children of this association. +associations which are children of this association. .TP \fIGrpWall=\fP Maximum wall clock time running jobs are able to be -allocated in aggregate for this association and all association which +allocated in aggregate for this association and all associations which are children of this association. .TP \fIFairShare=\fP @@ -1372,35 +1372,35 @@ A brief description of the account. .TP \fIGrpCPUMins=\fP Maximum number of CPU hours running jobs are able to -be allocated in aggregate for this association and all association +be allocated in aggregate for this association and all associations which are children of this association. (NOTE: This limit is not currently enforced in SLURM. You can still set this, but have to wait for future versions of SLURM before it is enforced.) .TP \fIGrpCPUs=\fP Maximum number of CPUs running jobs are able to be -allocated in aggregate for this association and all association which +allocated in aggregate for this association and all associations which are children of this association. (NOTE: This limit is not currently enforced in SLURM. You can still set this, but have to wait for future versions of SLURM before it is enforced.) .TP \fIGrpJobs=\fP Maximum number of running jobs in aggregate for this -association and all association which are children of this association. +association and all associations which are children of this association. .TP \fIGrpNodes=\fP Maximum number of nodes running jobs are able to be -allocated in aggregate for this association and all association which +allocated in aggregate for this association and all associations which are children of this association. .TP \fIGrpSubmitJobs=\fP Maximum number of jobs which can be in a pending or running state at any time in aggregate for this association and all -association which are children of this association. +associations which are children of this association. .TP \fIGrpWall=\fP Maximum wall clock time running jobs are able to be -allocated in aggregate for this association and all association which +allocated in aggregate for this association and all associations which are children of this association. .TP \fIFairShare=\fP diff --git a/doc/man/man1/scontrol.1 b/doc/man/man1/scontrol.1 index b99cd317a15ada18be053d2d06ea1b8568bd8429..4026048125e219d78551ba3f85254273b9e084b0 100644 --- a/doc/man/man1/scontrol.1 +++ b/doc/man/man1/scontrol.1 @@ -782,7 +782,7 @@ Format is the same as for MaxTime. .TP \fIDisableRootJobs\fP=<yes|no> -Specify if jobs can be executed as user roo. +Specify if jobs can be executed as user root. Possible values are "YES" and "NO". .TP diff --git a/doc/man/man1/srun.1 b/doc/man/man1/srun.1 index 5c519cf5916e38ad283e225cf5feff88ca695c34..0aef9884489f157950e31107e8eb026cb51bf874 100644 --- a/doc/man/man1/srun.1 +++ b/doc/man/man1/srun.1 @@ -1417,7 +1417,7 @@ Same as \fB\-\-core\fR Same as \fB\-\-cpu_bind\fR .TP \fBSLURM_CPUS_PER_TASK\fR -Same as \fB\-c, \-\-ncpus\-per\-task\fR +Same as \fB\-c, \-\-cpus\-per\-task\fR .TP \fBSLURM_DEBUG\fR Same as \fB\-v, \-\-verbose\fR diff --git a/doc/man/man5/slurm.conf.5 b/doc/man/man5/slurm.conf.5 index 62195d16a9df9f06db5e2b9754c45972539d282e..522d954f844c903e4f5905e83f7d492102c7bfac 100644 --- a/doc/man/man5/slurm.conf.5 +++ b/doc/man/man5/slurm.conf.5 @@ -2576,7 +2576,7 @@ for each job) plus \fBPrologSlurmctld\fR and \fBEpilogSlurmctld\fR NOTE: Standard output and error messages are normally not preserved. Explicitly write output and error messages to an appropriate location -if you which to preserve that information. +if you wish to preserve that information. NOTE: The Prolog script is ONLY run on any individual node when it first sees a job step from a new allocation; it does not diff --git a/doc/man/man8/spank.8 b/doc/man/man8/spank.8 index 6fe599d9fc22045584ccbbb85d2c3e88f77782f3..34eef669e570615c936e62f18c87b9a5c541f7a1 100644 --- a/doc/man/man8/spank.8 +++ b/doc/man/man8/spank.8 @@ -345,7 +345,7 @@ plugin's functions will cause \fBslurmd\fR to terminate the job, while \fIoptional\fR plugins only cause a warning. .LP If a fully\-qualified path is not specified for a plugin, then the -currently configure \fIPluginDir\fR in \fBslurm.conf\fR(5) is searched. +currently configured \fIPluginDir\fR in \fBslurm.conf\fR(5) is searched. .LP \fBSPANK\fR plugins are stackable, meaning that more than one plugin may be placed into the config file. The plugins will simply be called diff --git a/slurm/slurm.h.in b/slurm/slurm.h.in index 89fe1b30e1e93185711601bb3ab72decf722d32b..eb682c2820d0f12d74d628a9c53c58f22683de22 100644 --- a/slurm/slurm.h.in +++ b/slurm/slurm.h.in @@ -490,7 +490,7 @@ enum node_states { #define NODE_STATE_FAIL 0x2000 /* node is failing, do not allocate * new work */ #define NODE_STATE_POWER_UP 0x4000 /* restore power or otherwise - * configure a a node */ + * configure a node */ #define NODE_STATE_MAINT 0x8000 /* node in maintenance reservation */ /* used to define the size of the credential.signature size @@ -1355,7 +1355,7 @@ typedef struct node_info { dynamic_plugin_data_t *select_nodeinfo; /* opaque data structure, * use * slurm_get_select_nodeinfo() - * to access conents */ + * to access contents */ } node_info_t; typedef struct node_info_msg { @@ -1440,7 +1440,7 @@ typedef struct resource_allocation_response_msg { dynamic_plugin_data_t *select_jobinfo; /* opaque data structure, * use * slurm_get_select_jobinfo() - * to access conents */ + * to access contents */ } resource_allocation_response_msg_t; typedef struct job_alloc_info_response_msg { @@ -1455,7 +1455,7 @@ typedef struct job_alloc_info_response_msg { dynamic_plugin_data_t *select_jobinfo; /* opaque data structure, * use * slurm_get_select_jobinfo() - * to access conents */ + * to access contents */ } job_alloc_info_response_msg_t; typedef struct partition_info_msg { diff --git a/src/api/allocate.c b/src/api/allocate.c index dd277926bb4b1c75988412b01c2b9841d7f9e0a1..ca4c9c5a91e43b7a41546e8159ed81a419db772e 100644 --- a/src/api/allocate.c +++ b/src/api/allocate.c @@ -244,7 +244,7 @@ slurm_allocate_resources_blocking (const job_desc_msg_t *user_req, } break; case RESPONSE_RESOURCE_ALLOCATION: - /* Yay, the controller has acknowledge our request! But did + /* Yay, the controller has acknowledged our request! But did we really get an allocation yet? */ resp = (resource_allocation_response_msg_t *) resp_msg.data; if (resp->node_cnt > 0) { diff --git a/src/slurmctld/node_mgr.c b/src/slurmctld/node_mgr.c index f28dee2fff323e0f43b347ece0035c9bf7f411c6..cdc2a5e385b26711145fc0f57ca945002c2912cd 100644 --- a/src/slurmctld/node_mgr.c +++ b/src/slurmctld/node_mgr.c @@ -2203,7 +2203,7 @@ void set_node_down (char *name, char *reason) node_ptr = find_node_record (name); if (node_ptr == NULL) { - error ("node_not_resp unable to find node %s", name); + error ("set_node_down unable to find node %s", name); return; } diff --git a/src/slurmctld/read_config.c b/src/slurmctld/read_config.c index 1c76a65d4260965cf2cac44ad29691fd84586d79..81538b1d69a46ae7a2497618155884bf3eed27dc 100644 --- a/src/slurmctld/read_config.c +++ b/src/slurmctld/read_config.c @@ -926,7 +926,7 @@ static int _restore_node_state(int recover, old_node_ptr->reason = NULL; } if (recover == 2) { - /* NOTE: features in node record just a placeholder for + /* NOTE: features in node record just a placeholder * for restore_node_features() to set up new config * records. */ xfree(node_ptr->features);