Skip to content
Snippets Groups Projects
  1. Sep 23, 2013
  2. Aug 13, 2013
    • jette's avatar
      select/cons_res - Add test for zero node allocation · e180d341
      jette authored
      I don't see how this could happen, but it might explain something
      reported by Harvard University. In any case, this could prevent
      an infinite loop if the task distribution funciton is passed a
      job allocation with zero nodes.
      e180d341
    • jette's avatar
      select/cons_res - Avoid extraneous "oversubscribe" error messages · 302d8b3f
      jette authored
      This problem was reported by Harvard University and could be
      reproduced with a command line of "srun -N1 --tasks-per-node=2 -O id".
      With other job types, the error message could be logged many times
      for each job. This change logs the error once per job and only if
      the job request does not include the -O/--overcommit option.
      302d8b3f
  3. Jul 05, 2013
  4. Jun 28, 2013
  5. Jun 25, 2013
  6. Jun 21, 2013
  7. Jun 19, 2013
  8. Jun 12, 2013
  9. Jun 11, 2013
  10. Jun 10, 2013
  11. Jun 06, 2013
  12. Jun 05, 2013
  13. Jun 04, 2013
  14. Jun 03, 2013
    • jette's avatar
      Fix for job step allocation with required hostlist and exclusive option · 523b1992
      jette authored
      Previously if the required node has no available CPUs left, then other
      nodes in the job allocation would be used
      523b1992
    • Hongjia Cao's avatar
      restore max_nodes of desc to NO_VAL when checkpointing job · f82e0fb8
      Hongjia Cao authored
      We're having some trouble getting our slurm jobs to successfully
      restart after a checkpoint.  For this test, I'm using sbatch and a
      simple, single-threaded executable.  Slurm is 2.5.4, blcr is 0.8.5.
      I'm submitting the job using sbatch:
      
      $ sbatch -n 1 -t 12:00:00 bin/bowtie-ex.sh
      
      I am able to create the checkpoint and vacate the node:
      
      $ scontrol checkpoint create 137
      .... time passes ....
      $ scontrol vacate 137
      
      At that point, I see the checkpoint file from blcr in the current
      directory and the checkpoint file from Slurm
      in /var/spool/slurm-llnl/checkpoint.  However, when I attempt to
      restart the job:
      
      $ scontrol checkpoint restart 137
      scontrol_checkpoint error: Node count specification invalid
      
      In slurmctld's log (at level 7) I see:
      
      [2013-05-29T12:41:08-07:00] debug2: Processing RPC: REQUEST_CHECKPOINT(restart) from uid=*****
      [2013-05-29T12:41:08-07:00] debug3: Version string in job_ckpt header is JOB_CKPT_002
      [2013-05-29T12:41:08-07:00] _job_create: max_nodes == 0
      [2013-05-29T12:41:08-07:00] _slurm_rpc_checkpoint restart 137: Node count specification invalid
      f82e0fb8
  15. May 30, 2013
  16. May 29, 2013
  17. May 23, 2013
  18. May 22, 2013
Loading