- Nov 22, 2013
-
-
David Bigagli authored
-
David Bigagli authored
-
- Nov 20, 2013
-
-
David Bigagli authored
entire parallel job.
-
- Nov 18, 2013
-
-
Morris Jette authored
Conflicts: doc/html/faq.shtml
-
Morris Jette authored
The time/resource allocation matrix is rebuilt on each job exit, which severely impacts performance at large counts of running jobs (say >10k jobs).
-
Morris Jette authored
-
Michel Hummel authored
Logic in backfill scheduler did not change when array_task_id was changed from 16 to 32-bit
-
- Nov 16, 2013
-
-
Phil Eckert authored
-
Chrysovalantis Paschoulas authored
-
- Nov 15, 2013
-
-
Rod Schultz authored
limits are configured as 0.
-
Morris Jette authored
bug 511
-
Morris Jette authored
-
Morris Jette authored
Add ability to clear a node's DRAIN flag using scontrol or sview by setting it's state to "UNDRAIN". The node's base state (e.g. "DOWN" or "IDLE") will not be changed. bug 514
-
- Nov 14, 2013
-
-
Danny Auble authored
-
Danny Auble authored
-
Morris Jette authored
bug 511
-
Morris Jette authored
-
Danny Auble authored
-
Danny Auble authored
-
Danny Auble authored
-
Morris Jette authored
-
Danny Auble authored
-
Morris Jette authored
-
- Nov 13, 2013
-
-
Morris Jette authored
-
Danny Auble authored
modular.
-
Morris Jette authored
-
Morris Jette authored
This makes it simpler to enable detailed debugging for reservations. This includes more information than we probably want to see with the DebugFlag=reservation and would be only for developer debugging
-
Morris Jette authored
This might have worked fine for core reservations or when there are sufficient idle nodes to use, the the select_g_resv_test() function clears the node bitmap for nodes that it can not use and the reservation create logic did not restore that bitmap after a failed resource selection attempt. This logic restores the node bitmap on a failed call to select_g_resv_test() so we can add nodes to the bitmap of available nodes rather than having it repeatedly cleared. The logic also adds some performance enhancements that I will add to in the next commit.
-
Morris Jette authored
-
jette authored
-
jette authored
-
David Bigagli authored
-
Morris Jette authored
This fixes a bug where a system is enforcing memory limits and the job already has a step running on some of the nodes then tries to start another step using some of those nodes. For example wwith DefMemPerNode configured and the select plugin enforcing memory limits, try: salloc -N2 bash $ srun -N1 sleep 10& $ srun -N2 hostname Without this patch, the second srun would fail instead of pend.
-
- Nov 12, 2013
-
-
Danny Auble authored
on a task level if any task hit it the check will be triggered)
-
David Bigagli authored
-
Danny Auble authored
-
David Bigagli authored
-
Troy Baer authored
-
David Bigagli authored
-
Danny Auble authored
-