- May 12, 2014
-
-
Morris Jette authored
-
Puenlap Lee authored
Also correct related documentation
-
Hongjia Cao authored
Completing nodes is removed when calling _try_sched() for a job, which is the case in select_nodes(). If _try_sched() thinks the job can run now but select_nodes() returns ESLURM_NODES_BUSY, the backfill loop will be ended.
-
- May 09, 2014
-
-
Danny Auble authored
-
Danny Auble authored
-
- May 08, 2014
-
-
Morris Jette authored
Fix sinfo -R to print each down/drained node once, rather than once per partition. This was broken in the sinfo change to process each partition's information in a separate pthread.
-
Morris Jette authored
Correct sinfo --sort fields to match documentation: E => Reason, H -> Reason Time (new), R -> Partition Name, u/U -> Reason user (new)
-
- May 07, 2014
-
-
Morris Jette authored
Without this patch, jobs with an infinite time limit would have their preemption GraceTime ignored.
-
Morris Jette authored
related to bug 789
-
Danny Auble authored
-
Danny Auble authored
them.
-
- May 06, 2014
-
-
Morris Jette authored
-
Danny Auble authored
-
Danny Auble authored
-
Morris Jette authored
-
Morris Jette authored
In slurm.spec file, replace "Requires cray-MySQL-devel-enterprise" with "Requires mysql-devel" per David Gloe.
-
- May 05, 2014
-
-
Danny Auble authored
-
Danny Auble authored
-
Danny Auble authored
Related to bug 771
-
Morris Jette authored
In version 14.03.2 was using "slurm_<jobid>_4294967294.out" due to error in job array logic.
-
Danny Auble authored
cnode counts.
-
- May 02, 2014
-
-
Danny Auble authored
-
Danny Auble authored
This is for bug 775
-
Danny Auble authored
-
Danny Auble authored
-
- May 01, 2014
-
-
Danny Auble authored
regression from 2a674aee
-
Danny Auble authored
-
Danny Auble authored
is running.
-
Danny Auble authored
-
- Apr 30, 2014
-
-
David Bigagli authored
together.
-
Morris Jette authored
Switch/nrt - Properly track usage of CAU and RDMA resources with multiple tasks per compute node. Previous logic would allocate resources once per task and then deallocate once per node, leaking CMA and RDMA resources and preventing their use by future jobs.
-
Morris Jette authored
If a job is held, then only release it with the "scontrol release <jobid>" command rather than a simple reset of the job's priority. This is needed to support job arrays better. Otherwise a priority reset of a job array would free all requeued/held jobs from that job array rather than leaving them held.
-
- Apr 28, 2014
-
-
Danny Auble authored
-
Danny Auble authored
in 2.0 :)
-
Morris Jette authored
Previously partition priority was only considered when used as a component of a job's priority with the priority/multifactor plugin. Now the partition priority is considered first, as documented, and the job priority is considered second. bug 764
-
- Apr 26, 2014
-
-
Stuart Midgley authored
Add --priority option to the salloc, sbatch and srun commands.
-
Danny Auble authored
This code was originally put here to enforce checks to make sure jobs didn't go over the limit. If they didn't request the amount then we set the limit and worked off that as if it were a request. If we do this now we could get jobs deigned which would cancel the job at submit with a very unrelated note as to why the job failed. Since we now check this these limits after the node selection this isn't needed.
-
Danny Auble authored
-
- Apr 25, 2014
-
-
Morris Jette authored
In addition to accepting a job ID argument to the hold and release commands also accept a job name (e.g. "scontrol hold my.bash")
-
Morris Jette authored
Add a job's exit state (COMPLETED, FAILED, etc) and exit code to email message bug 737
-