Skip to content
Snippets Groups Projects
  1. Jan 09, 2015
  2. Jan 08, 2015
  3. Jan 07, 2015
  4. Jan 06, 2015
  5. Jan 05, 2015
  6. Jan 02, 2015
  7. Jan 01, 2015
  8. Dec 31, 2014
  9. Dec 30, 2014
  10. Dec 29, 2014
  11. Dec 26, 2014
  12. Dec 24, 2014
  13. Dec 23, 2014
    • Morris Jette's avatar
      Fix bad job array task ID value · 48016f86
      Morris Jette authored
      Prevent invalid job array task ID value if a task is started using gang
      scheduling (i.e. the task starts in a SUSPENDED state). The task ID gets
      set to NO_VAL and the task string is also cleared.
      48016f86
    • Morris Jette's avatar
      2da84715
    • Morris Jette's avatar
      Prevent gang resume of suspended job · 161d0336
      Morris Jette authored
      Prevent a job manually suspended from being resumed by gang scheduler once
      free resources are available.
      bug 1335
      161d0336
    • Dorian Krause's avatar
      set node state RESERVED on maint reservation delete · cf846644
      Dorian Krause authored
      we have hit the following problem that seems to be present in Slurm
      slurm-14-11-2-1 and previous versions. When a node is reserved and an
      overlapping maint reservation is created and later deleted the scontrol
      output will report the node as IDLE rather than RESERVED:
      
      + scontrol show node node1
      + grep State
         State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol create reservation starttime=now duration=120 user=usr01000
      nodes=node1 ReservationName=X
      Reservation created: X
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=RESERVED ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol create reservation starttime=now duration=120 user=usr01000
      nodes=ALL flags=maint,ignore_jobs ReservationName=Y
      Reservation created: Y
      + sleep 10
      + grep State
      + scontrol show nodes node1
         State=MAINT ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol delete ReservationName=Y
      + sleep 10
      + scontrol show nodes node1
      + grep State
      *   State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1*
      + scontrol delete ReservationName=X
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1
      
      Note that the after the deletion of reservation "X" the State=IDLE
      instead of State=RESERVED. I think that the delete_resv() function in
      slurmctld/reservation.c should call set_node_maint_mode(true) like
      update_resv() does. With the patch pasted at the end of this e-mail I
      get the following output which matches my expectation:
      
      + scontrol show node node1
      + grep State
         State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol create reservation starttime=now duration=120 user=usr01000
      nodes=node1 ReservationName=X
      Reservation created: X
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=RESERVED ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol create reservation starttime=now duration=120 user=usr01000
      nodes=ALL flags=maint,ignore_jobs ReservationName=Y
      Reservation created: Y
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=MAINT ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol delete ReservationName=Y
      + sleep 10
      + scontrol show nodes node1
      + grep State
      *   State=RESERVED ThreadsPerCore=1 TmpDisk=0 Weight=1*
      + scontrol delete ReservationName=X
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1
      
      Thanks,
      Dorian
      cf846644
Loading