- Jul 02, 2015
-
-
Morris Jette authored
Add association usage information to "scontrol show cache" command output.
-
- Jul 01, 2015
-
-
Brian Christiansen authored
When submitting a job with srun -n# the job may be allocated more than # because the job was given the whole core or socket (eg. CR_CORE, CR_SOCKET). sacct showed only what the step used and not the allocation. This commit shows the job and the step if job and step cpus are different.
-
Morris Jette authored
Major re-write of the sreport command to support --tres job option and permit users to select specific tracable resources to generate reports for. For most reports, each TRES is listed on a separate line of output with its name. The default TRES type is "cpu" to minimize changes to output.
-
- Jun 30, 2015
-
-
Thomas Cadeau authored
Bug 1745
-
Brian Christiansen authored
This reverts commit 3f91f4b2.
-
- Jun 29, 2015
-
-
Nathan Yee authored
Bug 1745
-
David Bigagli authored
-
- Jun 26, 2015
-
-
Danny Auble authored
-
Brian Christiansen authored
Bug 1746
-
- Jun 25, 2015
-
-
Morris Jette authored
-
Morris Jette authored
-
- Jun 24, 2015
-
-
David Bigagli authored
-
Morris Jette authored
-
- Jun 23, 2015
-
-
David Bigagli authored
-
- Jun 22, 2015
-
-
Morris Jette authored
Updates of existing bluegene advanced reservations did not work at all. Some multi-core configurations resulting in an abort due to creating core_bitmaps for the reservation that only had one bit per node rather than one bit per core. These bugs were introduced in commit 5f258072
-
David Bigagli authored
-
David Bigagli authored
-
- Jun 19, 2015
-
-
David Bigagli authored
-
- Jun 15, 2015
-
-
Morris Jette authored
Logic was assuming the reservation had a node bitmap which was being used to check for overlapping jobs. If there is no node bitmap (e.g. a licenses only reservation), an abort would result.
-
- Jun 12, 2015
-
-
Brian Christiansen authored
Bug 1739
-
Brian Christiansen authored
Bug 1743
-
Brian Christiansen authored
Bug 1743
-
- Jun 11, 2015
-
-
Brian Christiansen authored
Bug 1733
-
- Jun 10, 2015
-
-
Morris Jette authored
-
- Jun 09, 2015
-
-
David Bigagli authored
-
Morris Jette authored
1. I submit a first job that uses 1 GPU: $ srun --gres gpu:1 --pty bash $ echo $CUDA_VISIBLE_DEVICES 0 2. while the first one is still running, a 2-GPU job asking for 1 task per node waits (and I don't really understand why): $ srun --ntasks-per-node=1 --gres=gpu:2 --pty bash srun: job 2390816 queued and waiting for resources 3. whereas a 2-GPU job requesting 1 core per socket (so just 1 socket) actually gets GPUs allocated from two different sockets! $ srun -n 1 --cores-per-socket=1 --gres=gpu:2 -p testk --pty bash $ echo $CUDA_VISIBLE_DEVICES 1,2 With this change #2 works the same way as #3. bug 1725
-
Brian Christiansen authored
Bug 1572
-
Brian Christiansen authored
Bug 1572
-
- Jun 05, 2015
-
-
Danny Auble authored
of 4.
-
Danny Auble authored
Only going to do this in the master as it may affect scripts. This reverts commit 454f78e6. Conflicts: NEWS
-
- Jun 04, 2015
-
-
David Bigagli authored
-
David Bigagli authored
-
- Jun 03, 2015
-
-
Morris Jette authored
switch/cray: Refine logic to set PMI_CRAY_NO_SMP_ENV environment variable. Rather than testing for the task distribution option, test the actual task IDs to see fi they are monotonically increasing across all nodes. Based upon idea from Brian Gilmer (Cray).
-
Morris Jette authored
Add srun --accel-bind option to control how tasks are bound to GPUs and NIC Generic RESources (GRES). Based in part upon work by Matthieu Ospici (ATOS). gres/nic plugin modified to set OMPI_MCA_btl_openib_if_include environment variable based upon allocated devices (usable with OpenMPI and Melanox). Reset GRES env vars after task affinity set
-
- Jun 02, 2015
-
-
Danny Auble authored
-
Danny Auble authored
afterward cause a divide by zero error.
-
Danny Auble authored
corruption if thread uses the pointer basing validity off the id. Bug 1710
-
- Jun 01, 2015
-
-
Morris Jette authored
If an salloc or srun command is executed on a "front-end" configuration, that job will be assigned a slurmd shepherd daemon on the same host as used to execute the command when possible rather than an slurmd daemon on an arbitrary front-end node.
-
David Bigagli authored
-
- May 30, 2015
-
-
Danny Auble authored
-