diff --git a/slurm/slurm.h.in b/slurm/slurm.h.in
index 2260a563996ad26f172b5da95d64aed462698fea..d6aaf9adcb305cf4b5a38d1bbbaa8a951b8d3f6f 100644
--- a/slurm/slurm.h.in
+++ b/slurm/slurm.h.in
@@ -6,32 +6,32 @@
  *  Produced at Lawrence Livermore National Laboratory (cf, DISCLAIMER).
  *  Written by Morris Jette <jette1@llnl.gov>, et. al.
  *  CODE-OCEC-09-009. All rights reserved.
- *  
+ *
  *  This file is part of SLURM, a resource management program.
  *  For details, see <https://computing.llnl.gov/linux/slurm/>.
  *  Please also read the included file: DISCLAIMER.
- *  
+ *
  *  SLURM is free software; you can redistribute it and/or modify it under
  *  the terms of the GNU General Public License as published by the Free
  *  Software Foundation; either version 2 of the License, or (at your option)
  *  any later version.
  *
- *  In addition, as a special exception, the copyright holders give permission 
+ *  In addition, as a special exception, the copyright holders give permission
  *  to link the code of portions of this program with the OpenSSL library under
- *  certain conditions as described in each individual source file, and 
- *  distribute linked combinations including the two. You must obey the GNU 
- *  General Public License in all respects for all of the code used other than 
- *  OpenSSL. If you modify file(s) with this exception, you may extend this 
- *  exception to your version of the file(s), but you are not obligated to do 
+ *  certain conditions as described in each individual source file, and
+ *  distribute linked combinations including the two. You must obey the GNU
+ *  General Public License in all respects for all of the code used other than
+ *  OpenSSL. If you modify file(s) with this exception, you may extend this
+ *  exception to your version of the file(s), but you are not obligated to do
  *  so. If you do not wish to do so, delete this exception statement from your
- *  version.  If you delete this exception statement from all source files in 
+ *  version.  If you delete this exception statement from all source files in
  *  the program, then also delete it here.
- *  
+ *
  *  SLURM is distributed in the hope that it will be useful, but WITHOUT ANY
  *  WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
  *  FOR A PARTICULAR PURPOSE.  See the GNU General Public License for more
  *  details.
- *  
+ *
  *  You should have received a copy of the GNU General Public License along
  *  with SLURM; if not, write to the Free Software Foundation, Inc.,
  *  51 Franklin Street, Fifth Floor, Boston, MA 02110-1301  USA.
@@ -104,11 +104,11 @@ BEGIN_C_DECLS
   /*
    * 'bool' can be implemented in a variety of ways.
    * C++ may define it one way.
-   * <stdbool.h> may declare 'bool' to be a different type, such as 
+   * <stdbool.h> may declare 'bool' to be a different type, such as
    * an enum which is not necessarily compatible with C++.
    *
-   * If your compiler can't resolve 'bool', just define 
-   * CRAPPY_COMPILER at build time in order for SLURM to 
+   * If your compiler can't resolve 'bool', just define
+   * CRAPPY_COMPILER at build time in order for SLURM to
    * define it's own version of bool.
    */
 
@@ -150,7 +150,7 @@ BEGIN_C_DECLS
    typedef struct job_resources job_resources_t;
 #endif
 
-/* Define select_jobinfo_t, select_nodeinfo_t below 
+/* Define select_jobinfo_t, select_nodeinfo_t below
  * to avoid including extraneous slurm headers */
 #ifndef __select_jobinfo_t_defined
 #  define  __select_jobinfo_t_defined	/* Opaque data for select plugins */
@@ -164,7 +164,7 @@ BEGIN_C_DECLS
    typedef struct jobacctinfo jobacctinfo_t;     /* opaque data type */
 #endif
 
-/* Define allocation_msg_thread_t below to avoid including extraneous 
+/* Define allocation_msg_thread_t below to avoid including extraneous
  * slurm headers */
 #ifndef __allocation_msg_thread_t_defined
 #  define  __allocation_msg_thread_t_defined
@@ -185,8 +185,8 @@ BEGIN_C_DECLS
 #define SLURM_VERSION_MINOR(a)   (((a) >>  8) & 0xff)
 #define SLURM_VERSION_MICRO(a)    ((a)        & 0xff)
 
-/* Define the API's version. Update in META as needed. 
- * Also defined in config.h. 
+/* Define the API's version. Update in META as needed.
+ * Also defined in config.h.
  * High-order byte is major version. Update when existing APIs change.
  * Middle byte     is minor version. Update when new functions are added.
  * Low-order byte  is micro version. Update on patches and bug fixes. */
@@ -207,8 +207,8 @@ BEGIN_C_DECLS
 /* Job step ID of batch scripts */
 #define SLURM_BATCH_SCRIPT (0xfffffffe)
 
-/* last entry must be JOB_END, keep in sync with job_state_string and 
- *	job_state_string_compact. values may be ORed with JOB_STATE_FLAGS 
+/* last entry must be JOB_END, keep in sync with job_state_string and
+ *	job_state_string_compact. values may be ORed with JOB_STATE_FLAGS
  *	below.  */
 enum job_states {
 	JOB_PENDING,		/* queued waiting for initiation */
@@ -237,8 +237,8 @@ enum job_states {
 
 #define NICE_OFFSET 10000	/* offset for job's nice value */
 
-/* Reason for job to be pending rather than executing or reason for job 
- * failure. If multiple reasons exists, only one is given for the sake of 
+/* Reason for job to be pending rather than executing or reason for job
+ * failure. If multiple reasons exists, only one is given for the sake of
  * system efficiency */
 enum job_state_reason {
 /* Reasons for job to be pending */
@@ -343,7 +343,7 @@ enum select_print_mode {
 };
 
 enum select_node_cnt {
-	SELECT_GET_NODE_SCALING,      /* Give scaling factor for node count */ 
+	SELECT_GET_NODE_SCALING,      /* Give scaling factor for node count */
 	SELECT_APPLY_NODE_MIN_OFFSET, /* Apply min offset to variable */
 	SELECT_APPLY_NODE_MAX_OFFSET, /* Apply max offset to variable */
 	SELECT_SET_NODE_CNT,	      /* Set altered node cnt */
@@ -400,7 +400,7 @@ typedef enum cpu_bind_type {	/* cpu binding type from --cpu_bind=... */
 	/* verbose can be set with any other flag */
 	CPU_BIND_VERBOSE   = 0x01, /* =v, */
 	/* the following auto-binding flags are mutually exclusive */
-	CPU_BIND_TO_THREADS= 0x02, /* =threads */ 
+	CPU_BIND_TO_THREADS= 0x02, /* =threads */
 	CPU_BIND_TO_CORES  = 0x04, /* =cores */
 	CPU_BIND_TO_SOCKETS= 0x08, /* =sockets */
 	CPU_BIND_TO_LDOMS  = 0x10, /* locality domains */
@@ -413,8 +413,8 @@ typedef enum cpu_bind_type {	/* cpu binding type from --cpu_bind=... */
 	CPU_BIND_LDRANK    = 0x200,/* =locality domain rank */
 	CPU_BIND_LDMAP	   = 0x400,/* =map_ldom:<list of locality domains> */
 	CPU_BIND_LDMASK	   = 0x800,/* =mask_ldom:<list of ldom masks> */
-	
-	/* the following is used only as a flag for expressing 
+
+	/* the following is used only as a flag for expressing
 	 * the contents of TaskPluginParams */
 	CPU_BIND_CPUSETS   = 0x8000
 } cpu_bind_type_t;
@@ -431,10 +431,10 @@ typedef enum mem_bind_type {    /* memory binding type from --mem_bind=... */
 	MEM_BIND_LOCAL  = 0x20	/* =local */
 } mem_bind_type_t;
 
-/* The last entry in node_states must be STATE_END, keep in sync with 
- * node_state_string. values may be ORed with NODE_STATE_FLAGS below. 
- * Node states typically alternate between NODE_STATE_IDLE and 
- * NODE_STATE_ALLOCATED. The NODE_STATE_COMPLETING flag will be set 
+/* The last entry in node_states must be STATE_END, keep in sync with
+ * node_state_string. values may be ORed with NODE_STATE_FLAGS below.
+ * Node states typically alternate between NODE_STATE_IDLE and
+ * NODE_STATE_ALLOCATED. The NODE_STATE_COMPLETING flag will be set
  * when jobs are in the process of terminating. */
 enum node_states {
 	NODE_STATE_UNKNOWN,	/* node's initial state, unknown */
@@ -448,9 +448,9 @@ enum node_states {
 };
 #define NODE_STATE_BASE       0x00ff
 #define NODE_STATE_FLAGS      0xff00
-#define NODE_RESUME           0x0100	/* Restore a DRAINED, DRAINING, DOWN 
-					 * or FAILING node to service (e.g. 
-					 * IDLE or ALLOCATED). Used in 
+#define NODE_RESUME           0x0100	/* Restore a DRAINED, DRAINING, DOWN
+					 * or FAILING node to service (e.g.
+					 * IDLE or ALLOCATED). Used in
 					 * slurm_update_node() request */
 #define NODE_STATE_DRAIN      0x0200	/* node do not new allocated work */
 #define NODE_STATE_COMPLETING 0x0400	/* node is completing allocated job */
@@ -458,7 +458,7 @@ enum node_states {
 #define NODE_STATE_POWER_SAVE 0x1000	/* node in power save mode */
 #define NODE_STATE_FAIL       0x2000	/* node is failing, do not allocate
 					 * new work */
-#define NODE_STATE_POWER_UP   0x4000	/* restore power or otherwise 
+#define NODE_STATE_POWER_UP   0x4000	/* restore power or otherwise
 					 * configure a a node */
 #define NODE_STATE_MAINT      0x8000	/* node in maintenance reservation */
 
@@ -530,17 +530,17 @@ typedef struct job_descriptor {	/* For submit, allocate, and update requests */
 	char *account;		/* charge to specified account */
 	uint16_t acctg_freq;	/* accounting polling interval (seconds) */
 	char *alloc_node;	/* node making resource allocation request
-				 * NOTE: Normally set by slurm_submit* or 
+				 * NOTE: Normally set by slurm_submit* or
 				 * slurm_allocate* function */
 	uint16_t alloc_resp_port; /* port to send allocation confirmation to */
 	uint32_t alloc_sid;	/* local sid making resource allocation request
-				 * NOTE: Normally set by slurm_submit* or 
+				 * NOTE: Normally set by slurm_submit* or
 				 * slurm_allocate* function */
 	uint32_t argc;		/* number of arguments to the script */
 	char **argv;		/* arguments to the script */
 	time_t begin_time;	/* delay initiation until this time */
 	uint16_t ckpt_interval;	/* periodically checkpoint this job */
-	char *ckpt_dir;  	/* directory to store checkpoint images */
+	char *ckpt_dir;	 	/* directory to store checkpoint images */
 	char *comment;		/* arbitrary comment (used by Moab scheduler) */
 	uint16_t contiguous;	/* 1 if job requires contiguous nodes,
 				 * 0 otherwise,default=0 */
@@ -548,20 +548,20 @@ typedef struct job_descriptor {	/* For submit, allocate, and update requests */
 	uint16_t cpu_bind_type;	/* see cpu_bind_type_t */
 	char *dependency;	/* syncrhonize job execution with other jobs */
 	time_t end_time;	/* time by which job must complete, used for
-				 * job update only now, possible deadline 
+				 * job update only now, possible deadline
 				 * scheduling in the future */
-	char **environment;	/* environment variables to set for job, 
+	char **environment;	/* environment variables to set for job,
 				 *  name=value pairs, one per line */
 	uint32_t env_size;	/* element count in environment */
 	char *exc_nodes;	/* comma separated list of nodes excluded
 				 * from job's allocation, default NONE */
-	char *features;		/* comma separated list of required features, 
+	char *features;		/* comma separated list of required features,
 				 * default NONE */
-	uint32_t group_id;	/* group to assume, if run as root. */ 
-	uint16_t immediate;	/* 1 if allocate to run or fail immediately, 
+	uint32_t group_id;	/* group to assume, if run as root. */
+	uint16_t immediate;	/* 1 if allocate to run or fail immediately,
 				 * 0 if to be queued awaiting resources */
 	uint32_t job_id;	/* job ID, default set by SLURM */
-	uint16_t kill_on_node_fail; /* 1 if node failure to kill job, 
+	uint16_t kill_on_node_fail; /* 1 if node failure to kill job,
 				 * 0 otherwise,default=1 */
 	char *licenses;		/* licenses required by the job */
 	uint16_t mail_type;	/* see MAIL_JOB_ definitions above */
@@ -570,20 +570,20 @@ typedef struct job_descriptor {	/* For submit, allocate, and update requests */
 	uint16_t mem_bind_type;	/* see mem_bind_type_t */
 	char *name;		/* name of the job, default "" */
 	char *network;		/* network use spec */
-	uint16_t nice;		/* requested priority change, 
+	uint16_t nice;		/* requested priority change,
 				 * NICE_OFFSET == no change */
-	uint32_t num_tasks;	/* number of tasks to be started, 
+	uint32_t num_tasks;	/* number of tasks to be started,
 				 * for batch only */
 	uint8_t open_mode;	/* out/err open mode truncate or append,
 				 * see OPEN_MODE_* */
 	uint16_t other_port;	/* port to send various notification msg to */
 	uint8_t overcommit;	/* over subscribe resources, for batch only */
-	char *partition;	/* name of requested partition, 
+	char *partition;	/* name of requested partition,
 				 * default in SLURM config */
 	uint16_t plane_size;	/* plane size when task_dist =
 				   SLURM_DIST_PLANE */
-	uint32_t priority;	/* relative priority of the job,  
-				 * explicitly set only for user root, 
+	uint32_t priority;	/* relative priority of the job,
+				 * explicitly set only for user root,
 				 * 0 == held (don't initiate) */
 	char *qos;		/* Quality of Service */
 	char *resp_host;	/* NOTE: Set by slurmctld */
@@ -602,7 +602,7 @@ typedef struct job_descriptor {	/* For submit, allocate, and update requests */
 	uint16_t task_dist;	/* see enum task_dist_state */
 	uint32_t time_limit;	/* maximum run time in minutes, default is
 				 * partition limit */
-	uint32_t user_id;	/* set only if different from current UID, 
+	uint32_t user_id;	/* set only if different from current UID,
 				 * can only be explicitly set by user root */
 	uint16_t warn_signal;	/* signal to send when approaching end time */
 	uint16_t warn_time;	/* time before end to send signal (seconds) */
@@ -610,15 +610,15 @@ typedef struct job_descriptor {	/* For submit, allocate, and update requests */
 
 	/* job constraints: */
 	uint16_t job_min_cpus;    /* minimum # CPUs per node, default=0 */
-	uint32_t job_min_memory;   /* minimum real memory per node OR 
+	uint32_t job_min_memory;   /* minimum real memory per node OR
 				    * real memory per CPU | MEM_PER_CPU,
 				    * default=0 (no limit) */
 	uint32_t job_min_tmp_disk; /* minimum tmp disk per node, default=0 */
-	uint32_t num_procs;	/* total count of processors required, 
+	uint32_t num_procs;	/* total count of processors required,
 				 * default=0 */
-	uint32_t min_nodes;	/* minimum number of nodes required by job, 
+	uint32_t min_nodes;	/* minimum number of nodes required by job,
 				 * default=0 */
-	uint32_t max_nodes;	/* maximum number of nodes usable by job, 
+	uint32_t max_nodes;	/* maximum number of nodes usable by job,
 				 * default=0 */
 	uint16_t min_sockets;	/* minimum number of sockets per node required
 				 * by job, default=0 */
@@ -674,7 +674,7 @@ typedef struct job_info {
 	time_t end_time;	/* time of termination, actual or expected */
 	char *exc_nodes;	/* comma separated list of excluded nodes */
 	int *exc_node_inx;	/* excluded list index pairs into node_table:
-				 * start_range_1, end_range_1, 
+				 * start_range_1, end_range_1,
 				 * start_range_2, .., -1  */
 	uint32_t exit_code;	/* exit code for job (status from wait call) */
 	char *features;		/* comma separated list of required features */
@@ -692,9 +692,9 @@ typedef struct job_info {
 	char *name;		/* name of the job */
 	char *network;		/* network specification */
 	char *nodes;		/* list of nodes allocated to job */
-	uint16_t nice;    	/* requested priority change */
+	uint16_t nice;	  	/* requested priority change */
 	int *node_inx;		/* list index pairs into node_table for *nodes:
-				 * start_range_1, end_range_1, 
+				 * start_range_1, end_range_1,
 				 * start_range_2, .., -1  */
 	uint16_t ntasks_per_core;/* number of tasks to invoke on each core */
 	uint16_t ntasks_per_node;/* number of tasks to invoke on each node */
@@ -704,14 +704,14 @@ typedef struct job_info {
 	uint32_t num_procs;	/* number of processors required by job */
 	char *partition;	/* name of assigned partition */
 	time_t pre_sus_time;	/* time job ran prior to last suspend */
-	uint32_t priority;	/* relative priority of the job, 
+	uint32_t priority;	/* relative priority of the job,
 				 * 0=held, 1=required nodes DOWN/DRAINED */
 	char *qos;		/* Quality of Service */
 	char *req_nodes;	/* comma separated list of required nodes */
-	int *req_node_inx;	/* required list index pairs into node_table: 
-				 * start_range_1, end_range_1, 
+	int *req_node_inx;	/* required list index pairs into node_table:
+				 * start_range_1, end_range_1,
 				 * start_range_2, .., -1  */
-        uint16_t requeue;       /* enable or disable job requeue option */
+	uint16_t requeue;       /* enable or disable job requeue option */
 	uint16_t restart_cnt;	/* count of job restarts */
 	char *resv_name;	/* reservation name */
 	select_jobinfo_t *select_jobinfo; /* opaque data type,
@@ -748,7 +748,7 @@ typedef struct slurm_step_layout {
 	uint16_t task_dist;	/* see enum task_dist_state */
 	/* Array (of length "node_cnt") of task ID arrays.  The length
 	 * of each subarray is designated by the corresponding value in
-         * the tasks array. */
+	 * the tasks array. */
 	uint32_t **tids;	/* host id => task id mapping */
 } slurm_step_layout_t;
 
@@ -761,8 +761,8 @@ typedef struct slurm_step_io_fds {
 } slurm_step_io_fds_t;
 
 #define SLURM_STEP_IO_FDS_INITIALIZER {{0, (uint32_t)-1, (uint32_t)-1},\
-                                   {1, (uint32_t)-1, (uint32_t)-1},\
-                                   {2, (uint32_t)-1, (uint32_t)-1}}
+				   {1, (uint32_t)-1, (uint32_t)-1},\
+				   {2, (uint32_t)-1, (uint32_t)-1}}
 
 typedef struct launch_tasks_response_msg {
 	uint32_t return_code;
@@ -787,8 +787,8 @@ typedef struct srun_ping_msg {
 } srun_ping_msg_t;
 
 typedef struct srun_job_complete_msg {
-        uint32_t job_id;        /* slurm job_id */
-        uint32_t step_id;       /* step_id or NO_VAL */
+	uint32_t job_id;        /* slurm job_id */
+	uint32_t step_id;       /* step_id or NO_VAL */
 } srun_job_complete_msg_t;
 
 typedef struct srun_timeout_msg {
@@ -818,10 +818,10 @@ typedef struct {
 	uint16_t ckpt_interval;	/* checkpoint interval in minutes */
 	uint32_t cpu_count;	/* number of required processors */
 	uint16_t exclusive;	/* 1 if CPUs not shared with other steps */
-	uint16_t immediate;	/* 1 if allocate to run or fail immediately, 
+	uint16_t immediate;	/* 1 if allocate to run or fail immediately,
 				 * 0 if to be queued awaiting resources */
 	uint32_t job_id;	/* job ID */
-	uint32_t mem_per_cpu;	/* memory required per CPU (MB), 
+	uint32_t mem_per_cpu;	/* memory required per CPU (MB),
 				 * use job limit if 0 */
 	char *ckpt_dir;		/* directory to store checkpoint image files */
 	char *name;		/* name of the job step */
@@ -853,7 +853,7 @@ typedef struct {
 	char **env;
 	char *cwd;
 	bool user_managed_io;
-	uint32_t msg_timeout; /* timeout set for sending message */    
+	uint32_t msg_timeout; /* timeout set for sending message */
 
 	/* START - only used if user_managed_io is false */
 	bool buffered_stdio;
@@ -907,7 +907,7 @@ typedef struct {
 } slurm_allocation_callbacks_t;
 
 typedef struct {
-	char *ckpt_dir;	        /* path to store checkpoint image files */
+	char *ckpt_dir;		/* path to store checkpoint image files */
 	uint16_t ckpt_interval;	/* checkpoint interval in minutes */
 	uint32_t job_id;	/* job ID */
 	char *name;		/* name of job step */
@@ -936,14 +936,14 @@ typedef struct job_step_info_response_msg {
 typedef struct node_info {
 	char *arch;		/* computer architecture */
 	uint16_t cores;         /* number of cores per CPU */
-	uint16_t cpus;		/* configured count of cpus running on 
+	uint16_t cpus;		/* configured count of cpus running on
 				 * the node */
 	char *features;		/* arbitrary list of features for node */
 	char *name;		/* node name */
 	uint16_t node_state;	/* see enum node_states */
 	char *os;		/* operating system currently running */
 	uint32_t real_memory;	/* configured MB of real memory on the node */
-	char *reason;   	/* reason for node being DOWN or DRAINING */
+	char *reason;		/* reason for node being DOWN or DRAINING */
 	uint16_t sockets;       /* number of sockets per node */
 	uint16_t threads;       /* number of threads per core */
 	uint32_t tmp_disk;	/* configured MB of total disk in TMP_FS */
@@ -984,7 +984,7 @@ typedef struct job_alloc_info_msg {
 
 typedef struct partition_info {
 	char *allow_alloc_nodes;/* list names of allowed allocating nodes */
-	char *allow_groups;	/* comma delimited list of groups, 
+	char *allow_groups;	/* comma delimited list of groups,
 				 * null indicates all */
 	uint16_t default_part;	/* 1 if this is default partition */
 	uint32_t default_time;	/* minutes, NO_VAL or INFINITE */
@@ -996,7 +996,7 @@ typedef struct partition_info {
 	uint32_t min_nodes;	/* per job */
 	char *name;		/* name of the partition */
 	int *node_inx;		/* list index pairs into node_table:
-				 * start_range_1, end_range_1, 
+				 * start_range_1, end_range_1,
 				 * start_range_2, .., -1  */
 	char *nodes;		/* list names of nodes in partition */
 	uint16_t priority;	/* scheduling priority for jobs */
@@ -1076,7 +1076,7 @@ typedef struct {
 typedef block_info_t update_block_msg_t;
 
 /*
- * slurm_print_block_info_msg - output information about all Bluegene 
+ * slurm_print_block_info_msg - output information about all Bluegene
  *	blocks based upon message as loaded using slurm_load_block
  * IN out - file to write to
  * IN info_ptr - block information message pointer
@@ -1086,7 +1086,7 @@ extern void slurm_print_block_info_msg PARAMS((
 	FILE *out, block_info_msg_t *info_ptr, int one_liner));
 
 /*
- * slurm_print_block_info - output information about a specific Bluegene 
+ * slurm_print_block_info - output information about a specific Bluegene
  *	block based upon message as loaded using slurm_load_block
  * IN out - file to write to
  * IN bg_info_ptr - an individual block information record pointer
@@ -1096,7 +1096,7 @@ extern void slurm_print_block_info PARAMS((
 	FILE *out, block_info_t *bg_info_ptr, int one_liner));
 
 /*
- * slurm_sprint_block_info - output information about a specific Bluegene 
+ * slurm_sprint_block_info - output information about a specific Bluegene
  *	block based upon message as loaded using slurm_load_block
  * IN bg_info_ptr - an individual partition information record pointer
  * IN one_liner - print as a single line if true
@@ -1107,26 +1107,26 @@ extern char *slurm_sprint_block_info PARAMS((
 	block_info_t * bg_info_ptr, int one_liner));
 
 /*
- * slurm_load_block_info - issue RPC to get slurm all node select plugin 
- *      information if changed since update_time 
+ * slurm_load_block_info - issue RPC to get slurm all node select plugin
+ *      information if changed since update_time
  * IN update_time - time of current configuration data
- * IN block_info_msg_pptr - place to store a node select configuration 
+ * IN block_info_msg_pptr - place to store a node select configuration
  *      pointer
  * RET 0 or a slurm error code
  * NOTE: free the response using slurm_free_block_info_msg
  */
 extern int slurm_load_block_info PARAMS((
-	time_t update_time, 
+	time_t update_time,
 	block_info_msg_t **block_info_msg_pptr));
 
 /*
- * slurm_free_block_info_msg - free buffer returned by 
+ * slurm_free_block_info_msg - free buffer returned by
  *	slurm_load_block
  * IN block_info_msg_pptr - data is freed and pointer is set to NULL
  * RET 0 or a slurm error code
  */
 extern int slurm_free_block_info_msg PARAMS((
-        block_info_msg_t **block_info_msg_pptr));
+	block_info_msg_t **block_info_msg_pptr));
 
 /* update state or remove block */
 extern int slurm_update_block PARAMS((update_block_msg_t *block_msg));
@@ -1208,7 +1208,7 @@ typedef struct reservation_name_msg {
 #define PREEMPT_MODE_OFF	0x0000	/* disable job preemption */
 #define PREEMPT_MODE_SUSPEND	0x0001	/* suspend jobs to preempt */
 #define PREEMPT_MODE_REQUEUE	0x0002	/* requeue or kill jobs to preempt */
-#define PREEMPT_MODE_CHECKPOINT	0x0003	/* checkpoint job to preempt, 
+#define PREEMPT_MODE_CHECKPOINT	0x0003	/* checkpoint job to preempt,
 					 * no automatic restart */
 #define PREEMPT_MODE_CANCEL	0x0004	/* always cancel the job */
 #define PREEMPT_MODE_GANG	0x8000	/* enable gang scheduling */
@@ -1243,7 +1243,7 @@ typedef struct slurm_ctl_conf {
 	uint32_t debug_flags;	/* see DEBUG_FLAG_* above for values */
 	uint32_t def_mem_per_cpu; /* default MB memory per allocated CPU */
 	uint16_t disable_root_jobs; /* if set then user root can't run jobs */
-	uint16_t enforce_part_limits;	/* if set, reject job exceeding 
+	uint16_t enforce_part_limits;	/* if set, reject job exceeding
 					 * partition size and/or time limits */
 	char *epilog;		/* pathname of job epilog */
 	uint32_t epilog_msg_time;  /* usecs for slurmctld to process an
@@ -1257,7 +1257,7 @@ typedef struct slurm_ctl_conf {
 	char * health_check_program;	/* pathname of health check program */
 	uint16_t inactive_limit;/* seconds of inactivity before a
 				 * inactive resource allocation is released */
-	uint16_t job_acct_gather_freq; /* poll frequency for job accounting 
+	uint16_t job_acct_gather_freq; /* poll frequency for job accounting
 					* gather plugins */
 	char *job_acct_gather_type; /* job accounting gather type */
 	char *job_ckpt_dir;	/* directory saving job record checkpoint */
@@ -1273,21 +1273,21 @@ typedef struct slurm_ctl_conf {
 	uint16_t job_requeue;	/* If set, jobs get requeued on node failre */
 	uint16_t kill_on_bad_exit; /* If set, the job will be
 				 * terminated immediately when one of
-				 * the processes is aborted or crashed */ 
-	uint16_t kill_wait;	/* seconds between SIGXCPU to SIGKILL 
+				 * the processes is aborted or crashed */
+	uint16_t kill_wait;	/* seconds between SIGXCPU to SIGKILL
 				 * on job termination */
 	char *licenses;		/* licenses available on this cluster */
 	char *mail_prog;	/* pathname of mail program */
 	uint16_t max_job_cnt;	/* maximum number of active jobs */
 	uint32_t max_mem_per_cpu; /* maximum MB memory per allocated CPU */
 	uint16_t max_tasks_per_node; /* maximum tasks per node */
-	uint16_t min_job_age;	/* COMPLETED jobs over this age (secs) 
-	                         * purged from in memory records */
+	uint16_t min_job_age;	/* COMPLETED jobs over this age (secs)
+				 * purged from in memory records */
 	char *mpi_default;	/* Default version of MPI in use */
 	char *mpi_params;	/* MPI parameters */
 	uint16_t msg_timeout;	/* message timeout */
 	uint32_t next_job_id;	/* next slurm generated job_id to assign */
-	char *node_prefix;      /* prefix of nodes in partition, only set in 
+	char *node_prefix;      /* prefix of nodes in partition, only set in
 				   bluegene clusters NULL otherwise */
 	uint16_t over_time_limit; /* job's time limit can be exceeded by this
 				   * number of minutes before cancellation */
@@ -1309,25 +1309,25 @@ typedef struct slurm_ctl_conf {
 	uint32_t priority_weight_js; /* weight for Job Size factor */
 	uint32_t priority_weight_part; /* weight for Partition factor */
 	uint32_t priority_weight_qos; /* weight for QOS factor */
-	uint16_t private_data;	/* block viewing of information, 
+	uint16_t private_data;	/* block viewing of information,
 				 * see PRIVATE_DATA_* */
 	char *proctrack_type;	/* process tracking plugin type */
 	char *prolog;		/* pathname of job prolog run by slurmd */
 	char *prolog_slurmctld;	/* pathname of job prolog run by slurmctld */
 	uint16_t propagate_prio_process; /* 1 if process priority should
-				          * be propagated */
-        char *propagate_rlimits;/* Propagate (all/specific) resource limits */
-        char *propagate_rlimits_except;/* Propagate all rlimits except these */
+					  * be propagated */
+	char *propagate_rlimits;/* Propagate (all/specific) resource limits */
+	char *propagate_rlimits_except;/* Propagate all rlimits except these */
 	char *resume_program;	/* program to make nodes full power */
 	uint16_t resume_rate;	/* nodes to make full power, per minute */
 	uint16_t resume_timeout;/* time required in order to perform a node
 				 * resume operation */
-	uint16_t resv_over_run;	/* how long a running job can exceed 
+	uint16_t resv_over_run;	/* how long a running job can exceed
 				 * reservation time */
-	uint16_t ret2service;	/* 1 return DOWN node to service at 
+	uint16_t ret2service;	/* 1 return DOWN node to service at
 				 * registration */
 	char *salloc_default_command; /* default salloc command */
-	char *sched_params;	/* SchedulerParameters OR 
+	char *sched_params;	/* SchedulerParameters OR
 				 * contents of scheduler plugin config file */
 	uint16_t sched_time_slice;	/* gang scheduler slice time, secs */
 	char *schedtype;	/* type of scheduler to use */
@@ -1337,7 +1337,7 @@ typedef struct slurm_ctl_conf {
 	char *select_type;	/* type of node selector to use */
 	void *select_conf_key_pairs; /* key-pair list which can be
 				      * listed with slurm_print_key_pairs */
-	uint16_t select_type_param; /* Parameters 
+	uint16_t select_type_param; /* Parameters
 				 * describing the select_type plugin */
 	uint32_t slurm_user_id;	/* uid of slurm_user_name */
 	char *slurm_user_name;	/* user that slurmctld runs as */
@@ -1347,14 +1347,14 @@ typedef struct slurm_ctl_conf {
 	char *slurmctld_logfile;/* where slurmctld error log gets written */
 	char *slurmctld_pidfile;/* where to put slurmctld pidfile         */
 	uint32_t slurmctld_port;/* default communications port to slurmctld */
-	uint16_t slurmctld_timeout;/* seconds that backup controller waits 
+	uint16_t slurmctld_timeout;/* seconds that backup controller waits
 				 * on non-responding primarly controller */
 	uint16_t slurmd_debug;	/* slurmd logging level */
 	char *slurmd_logfile;	/* where slurmd error log gets written */
 	char *slurmd_pidfile;   /* where to put slurmd pidfile           */
 	uint32_t slurmd_port;	/* default communications port to slurmd */
 	char *slurmd_spooldir;	/* where slurmd put temporary state info */
-	uint16_t slurmd_timeout;/* how long slurmctld waits for slurmd before 
+	uint16_t slurmd_timeout;/* how long slurmctld waits for slurmd before
 				 * considering node DOWN */
 	char *slurm_conf;	/* pathname of slurm config file */
 	char *srun_epilog;      /* srun epilog program */
@@ -1378,10 +1378,10 @@ typedef struct slurm_ctl_conf {
 	uint16_t track_wckey;    /* see if we are using wckey or not */
 	uint16_t tree_width;    /* number of threads per node to span */
 	char *unkillable_program; /* program run by the slurmstepd when
-                                   * processes in a job step are unkillable */
+				   * processes in a job step are unkillable */
 	uint16_t unkillable_timeout; /* time in seconds, after processes in a
-                                      * job step have been signalled, before
-                                      * they are considered "unkillable". */
+				      * job step have been signalled, before
+				      * they are considered "unkillable". */
 	uint16_t use_pam;	/* enable/disable PAM support */
 	char *version;		/* version of slurmctld */
 	uint16_t wait_time;	/* default job --wait time */
@@ -1416,7 +1416,7 @@ typedef struct submit_response_msg {
 typedef struct slurm_update_node_msg {
 	char *node_names;	/* comma separated list of required nodes */
 	uint16_t node_state;	/* see enum node_states */
-	char *reason;   	/* reason for node being DOWN or DRAINING */
+	char *reason;		/* reason for node being DOWN or DRAINING */
 	char *features;		/* new feature for node */
 	uint32_t weight;	/* new weight for node */
 } update_node_msg_t;
@@ -1466,8 +1466,8 @@ typedef struct trigger_info_msg {
 \*****************************************************************************/
 
 /*
- * slurm_init_job_desc_msg - initialize job descriptor with 
- *	default values 
+ * slurm_init_job_desc_msg - initialize job descriptor with
+ *	default values
  * OUT job_desc_msg - user defined job descriptor
  */
 extern void slurm_init_job_desc_msg PARAMS((job_desc_msg_t * job_desc_msg));
@@ -1484,7 +1484,7 @@ extern void slurm_init_job_desc_msg PARAMS((job_desc_msg_t * job_desc_msg));
  * NOTE: free the allocated using slurm_free_resource_allocation_response_msg
  */
 extern int slurm_allocate_resources PARAMS((
-	job_desc_msg_t * job_desc_msg , 
+	job_desc_msg_t * job_desc_msg ,
 	resource_allocation_response_msg_t ** job_alloc_resp_msg));
 
 /*
@@ -1499,7 +1499,7 @@ extern int slurm_allocate_resources PARAMS((
  *      the controller will put the job in the PENDING state.  If
  *      pending callback is not NULL, it will be called with the job_id
  *      of the pending job as the sole parameter.
- * 
+ *
  * RET allocation structure on success, NULL on error set errno to
  *	indicate the error (errno will be ETIMEDOUT if the timeout is reached
  *      with no allocation granted)
@@ -1540,7 +1540,7 @@ extern int slurm_allocation_lookup PARAMS((
 	uint32_t job_id, job_alloc_info_response_msg_t **info));
 
 /*
- * slurm_allocation_lookup_lite - retrieve minor info for an existing 
+ * slurm_allocation_lookup_lite - retrieve minor info for an existing
  *				  resource allocation
  * IN job_id - job allocation identifier
  * OUT resp - job allocation information
@@ -1568,9 +1568,9 @@ extern int slurm_allocation_lookup_lite PARAMS((
 extern char *slurm_read_hostfile PARAMS((char *filename, int n));
 
 /*
- * slurm_allocation_msg_thr_create - startup a message handler talking 
- * with the controller dealing with messages from the controller during an 
- * allocation. 
+ * slurm_allocation_msg_thr_create - startup a message handler talking
+ * with the controller dealing with messages from the controller during an
+ * allocation.
  * IN port - port we are listening for messages on from the controller
  * IN callbacks - callbacks for different types of messages
  * RET allocation_msg_thread_t * or NULL on failure
@@ -1579,9 +1579,9 @@ extern allocation_msg_thread_t *slurm_allocation_msg_thr_create PARAMS(
 	(uint16_t *port, const slurm_allocation_callbacks_t *callbacks));
 
 /*
- * slurm_allocation_msg_thr_destroy - shutdown the message handler talking 
+ * slurm_allocation_msg_thr_destroy - shutdown the message handler talking
  * with the controller dealing with messages from the controller during an
- * allocation. 
+ * allocation.
  * IN msg_thr - allocation_msg_thread_t pointer allocated with
  *              slurm_allocation_msg_thr_create
  */
@@ -1595,11 +1595,11 @@ extern void slurm_allocation_msg_thr_destroy PARAMS(
  * RET 0 on success, otherwise return -1 and set errno to indicate the error
  */
 extern int slurm_submit_batch_job PARAMS((
-	job_desc_msg_t * job_desc_msg, 
+	job_desc_msg_t * job_desc_msg,
 	submit_response_msg_t ** slurm_alloc_msg));
 
 /*
- * slurm_free_submit_response_response_msg - free slurm 
+ * slurm_free_submit_response_response_msg - free slurm
  *	job submit response message
  * IN msg - pointer to job submit response message
  * NOTE: buffer is loaded by slurm_submit_batch_job
@@ -1608,7 +1608,7 @@ extern void slurm_free_submit_response_response_msg PARAMS((
 		submit_response_msg_t *msg));
 
 /*
- * slurm_job_will_run - determine if a job would execute immediately if 
+ * slurm_job_will_run - determine if a job would execute immediately if
  *	submitted now
  * IN job_desc_msg - description of resource allocation request
  * RET 0 on success, otherwise return -1 and set errno to indicate the error
@@ -1639,7 +1639,7 @@ extern void slurm_free_sbcast_cred_msg PARAMS((job_sbcast_cred_msg_t * msg));
  * IN batch_flag - 1 to signal batch shell only, otherwise 0
  * RET 0 on success, otherwise return -1 and set errno to indicate the error
  */
-extern int slurm_kill_job PARAMS((uint32_t job_id, uint16_t signal, 
+extern int slurm_kill_job PARAMS((uint32_t job_id, uint16_t signal,
 			uint16_t batch_flag));
 
 /*
@@ -1649,7 +1649,7 @@ extern int slurm_kill_job PARAMS((uint32_t job_id, uint16_t signal,
  * IN signal  - signal number
  * RET 0 on success, otherwise return -1 and set errno to indicate the error
  */
-extern int slurm_kill_job_step PARAMS((uint32_t job_id, uint32_t step_id, 
+extern int slurm_kill_job_step PARAMS((uint32_t job_id, uint32_t step_id,
 				       uint16_t signal));
 
 /*
@@ -1677,7 +1677,7 @@ extern int slurm_signal_job_step PARAMS((uint32_t job_id, uint32_t step_id,
 \*****************************************************************************/
 
 /*
- * slurm_complete_job - note the completion of a job and all of its steps 
+ * slurm_complete_job - note the completion of a job and all of its steps
  * IN job_id - the job's id
  * IN job_return_code - the highest exit code of any task of the job
  * RET 0 on success, otherwise return -1 and set errno to indicate the error
@@ -1712,7 +1712,7 @@ extern int slurm_terminate_job_step PARAMS((
  *	SLURM TASK SPAWNING FUNCTIONS
 \*****************************************************************************/
 
-/* 
+/*
  * slurm_step_ctx_params_t_init - This initializes parameters
  *	in the structure that you will pass to slurm_step_ctx_create().
  *	This function will NOT allocate any new memory.
@@ -1722,7 +1722,7 @@ extern int slurm_terminate_job_step PARAMS((
 extern void slurm_step_ctx_params_t_init PARAMS((slurm_step_ctx_params_t *ptr));
 
 /*
- * slurm_step_ctx_create - Create a job step and its context. 
+ * slurm_step_ctx_create - Create a job step and its context.
  * IN step_params - job step parameters
  * RET the step context or NULL on failure with slurm errno set
  * NOTE: Free allocated memory using slurm_step_ctx_destroy.
@@ -1732,7 +1732,7 @@ extern slurm_step_ctx_t *slurm_step_ctx_create PARAMS((
 
 /*
  * slurm_step_ctx_create_no_alloc - Create a job step and its context without
- *                                  getting an allocation. 
+ *                                  getting an allocation.
  * IN step_params - job step parameters
  * IN step_id     - since we are faking it give me the id to use
  * RET the step context or NULL on failure with slurm errno set
@@ -1747,7 +1747,7 @@ slurm_step_ctx_create_no_alloc PARAMS((
  * IN ctx - job step context generated by slurm_step_ctx_create
  * RET SLURM_SUCCESS or SLURM_ERROR (with slurm_errno set)
  */
-extern int slurm_step_ctx_get PARAMS((slurm_step_ctx_t *ctx, 
+extern int slurm_step_ctx_get PARAMS((slurm_step_ctx_t *ctx,
 	int ctx_key, ...));
 
 /*
@@ -1757,7 +1757,7 @@ extern int slurm_step_ctx_get PARAMS((slurm_step_ctx_t *ctx,
  * OUT data - the requested data type
  * RET SLURM_SUCCESS or SLURM_ERROR (with slurm_errno set)
  */
-extern int slurm_jobinfo_ctx_get PARAMS((switch_jobinfo_t *jobinfo, 
+extern int slurm_jobinfo_ctx_get PARAMS((switch_jobinfo_t *jobinfo,
 	int data_type, void *data));
 
 /*
@@ -1781,7 +1781,7 @@ extern int slurm_step_ctx_daemon_per_node_hack PARAMS((slurm_step_ctx_t *ctx));
  */
 extern int slurm_step_ctx_destroy PARAMS((slurm_step_ctx_t *ctx));
 
-/* 
+/*
  * slurm_step_launch_params_t_init - initialize a user-allocated
  *      slurm_step_launch_params_t structure with default values.
  *	default values.  This function will NOT allocate any new memory.
@@ -1819,8 +1819,8 @@ extern void slurm_step_launch_wait_finish PARAMS((slurm_step_ctx_t *ctx));
  */
 extern void slurm_step_launch_abort PARAMS((slurm_step_ctx_t *ctx));
 
-/* 
- * Forward a signal to all those nodes with running tasks 
+/*
+ * Forward a signal to all those nodes with running tasks
  */
 extern void slurm_step_launch_fwd_signal PARAMS((slurm_step_ctx_t *ctx,
 	 int signo));
@@ -1830,24 +1830,24 @@ extern void slurm_step_launch_fwd_signal PARAMS((slurm_step_ctx_t *ctx,
 \*****************************************************************************/
 
 /*
- * slurm_api_version - Return a single number reflecting the SLURM API's 
- *	version number. Use the macros SLURM_VERSION_NUM, SLURM_VERSION_MAJOR, 
+ * slurm_api_version - Return a single number reflecting the SLURM API's
+ *	version number. Use the macros SLURM_VERSION_NUM, SLURM_VERSION_MAJOR,
  *	SLURM_VERSION_MINOR, and SLURM_VERSION_MICRO to work with this value
  * RET API's version number
  */
 extern long slurm_api_version PARAMS((void));
 
 /*
- * slurm_load_ctl_conf - issue RPC to get slurm control configuration  
- *	information if changed since update_time 
+ * slurm_load_ctl_conf - issue RPC to get slurm control configuration
+ *	information if changed since update_time
  * IN update_time - time of current configuration data
- * IN slurm_ctl_conf_ptr - place to store slurm control configuration 
+ * IN slurm_ctl_conf_ptr - place to store slurm control configuration
  *	pointer
  * RET 0 on success, otherwise return -1 and set errno to indicate the error
  * NOTE: free the response using slurm_free_ctl_conf
  */
 extern int slurm_load_ctl_conf PARAMS((
-	time_t update_time, 
+	time_t update_time,
 	slurm_ctl_conf_t  **slurm_ctl_conf_ptr));
 
 /*
@@ -1858,7 +1858,7 @@ extern int slurm_load_ctl_conf PARAMS((
 extern void slurm_free_ctl_conf PARAMS((slurm_ctl_conf_t* slurm_ctl_conf_ptr));
 
 /*
- * slurm_print_ctl_conf - output the contents of slurm control configuration 
+ * slurm_print_ctl_conf - output the contents of slurm control configuration
  *	message as loaded using slurm_load_ctl_conf
  * IN out - file to write to
  * IN slurm_ctl_conf_ptr - slurm control configuration pointer
@@ -1876,7 +1876,7 @@ extern void *slurm_ctl_conf_2_key_pairs PARAMS((
 	slurm_ctl_conf_t* slurm_ctl_conf_ptr));
 
 /*
- * slurm_load_slurmd_status - issue RPC to get the status of slurmd 
+ * slurm_load_slurmd_status - issue RPC to get the status of slurmd
  *	daemon on this machine
  * IN slurmd_status_ptr - place to store slurmd status information
  * RET 0 or -1 on error
@@ -1893,7 +1893,7 @@ extern void slurm_free_slurmd_status PARAMS((
 		slurmd_status_t* slurmd_status_ptr));
 
 /*
- * slurm_print_slurmd_status - output the contents of slurmd status 
+ * slurm_print_slurmd_status - output the contents of slurmd status
  *	message as loaded using slurm_load_slurmd_status
  * IN out - file to write to
  * IN slurmd_status_ptr - slurmd status pointer
@@ -1902,23 +1902,23 @@ void slurm_print_slurmd_status PARAMS((
 		FILE* out, slurmd_status_t * slurmd_status_ptr));
 
 /*
- * slurm_print_key_pairs - output the contents of key_pairs 
+ * slurm_print_key_pairs - output the contents of key_pairs
  *	which is a list of opaque data type config_key_pair_t
  * IN out - file to write to
  * IN key_pairs - List contianing key pairs to be printed
  * IN title - title of key pair list
  */
 void slurm_print_key_pairs PARAMS((
-                FILE* out, void* key_pairs, char *title));
+		FILE* out, void* key_pairs, char *title));
 
 /*****************************************************************************\
  *	SLURM JOB RESOURCES READ/PRINT FUNCTIONS
 \*****************************************************************************/
 
 /*
- * slurm_job_cpus_allocated_on_node_id - 
+ * slurm_job_cpus_allocated_on_node_id -
  *                        get the number of cpus allocated to a job
- *	                  on a node by node id
+ *			  on a node by node id
  * IN job_resrcs_ptr	- pointer to job_resources structure
  * IN node_id		- node id in allocation
  * RET cpus used or -1 on error
@@ -1927,9 +1927,9 @@ extern int slurm_job_cpus_allocated_on_node_id PARAMS(
 	(job_resources_t *job_resrcs_ptr, int node_id));
 
 /*
- * slurm_job_cpus_allocated_on_node - 
+ * slurm_job_cpus_allocated_on_node -
  *                        get the number of cpus allocated to a job
- *	                  on a node by node name
+ *			  on a node by node name
  * IN job_resrcs_ptr	- pointer to job_resources structure
  * IN node_name		- name of node
  * RET cpus used or -1 on error
@@ -1951,7 +1951,7 @@ extern void slurm_free_job_info_msg PARAMS((job_info_msg_t * job_buffer_ptr));
 /*
  * slurm_get_end_time - get the expected end time for a given slurm job
  * IN jobid     - slurm job id
- * end_time_ptr - location in which to store scheduled end time for job 
+ * end_time_ptr - location in which to store scheduled end time for job
  * RET 0 or -1 on error
  */
 extern int slurm_get_end_time PARAMS((uint32_t jobid, time_t *end_time_ptr));
@@ -1973,7 +1973,7 @@ extern int slurm_job_node_ready(uint32_t job_id);
 /*
  * slurm_load_job - issue RPC to get job information for one job ID
  * IN job_info_msg_pptr - place to store a job configuration pointer
- * IN job_id -  ID of job we want information about 
+ * IN job_id -  ID of job we want information about
  * IN show_flags - job filtering options
  * RET 0 or -1 on error
  * NOTE: free the response using slurm_free_job_info_msg
@@ -1982,8 +1982,8 @@ extern int slurm_load_job PARAMS((job_info_msg_t **resp, uint32_t job_id,
 	uint16_t show_flags));
 
 /*
- * slurm_load_jobs - issue RPC to get slurm all job configuration  
- *	information if changed since update_time 
+ * slurm_load_jobs - issue RPC to get slurm all job configuration
+ *	information if changed since update_time
  * IN update_time - time of current configuration data
  * IN job_info_msg_pptr - place to store a job configuration pointer
  * IN show_flags - job filtering options
@@ -1995,7 +1995,7 @@ extern int slurm_load_jobs PARAMS((
 	uint16_t show_flags));
 
 /*
- * slurm_notify_job - send message to the job's stdout, 
+ * slurm_notify_job - send message to the job's stdout,
  *	usable only by user root
  * IN job_id - slurm job_id or 0 for all jobs
  * IN message - arbitrary message
@@ -2004,7 +2004,7 @@ extern int slurm_load_jobs PARAMS((
 extern int slurm_notify_job PARAMS(( uint32_t job_id, char *message ));
 
 /*
- * slurm_pid2jobid - issue RPC to get the slurm job_id given a process_id 
+ * slurm_pid2jobid - issue RPC to get the slurm job_id given a process_id
  *	on this machine
  * IN job_pid - process_id of interest on this machine
  * OUT job_id_ptr - place to store a slurm job_id
@@ -2013,17 +2013,17 @@ extern int slurm_notify_job PARAMS(( uint32_t job_id, char *message ));
 extern int slurm_pid2jobid PARAMS(( pid_t job_pid, uint32_t * job_id_ptr )) ;
 
 /*
- * slurm_print_job_info - output information about a specific Slurm 
+ * slurm_print_job_info - output information about a specific Slurm
  *	job based upon message as loaded using slurm_load_jobs
  * IN out - file to write to
  * IN job_ptr - an individual job information record pointer
  * IN one_liner - print as a single line if true
  */
-extern void slurm_print_job_info PARAMS(( FILE*, job_info_t * job_ptr, 
+extern void slurm_print_job_info PARAMS(( FILE*, job_info_t * job_ptr,
 					  int one_liner ));
 
 /*
- * slurm_print_job_info_msg - output information about all Slurm 
+ * slurm_print_job_info_msg - output information about all Slurm
  *	jobs based upon message as loaded using slurm_load_jobs
  * IN out - file to write to
  * IN job_info_msg_ptr - job information message pointer
@@ -2033,18 +2033,18 @@ extern void slurm_print_job_info_msg PARAMS((
 	FILE * out, job_info_msg_t * job_info_msg_ptr, int one_liner ));
 
 /*
- * slurm_sprint_job_info - output information about a specific Slurm 
+ * slurm_sprint_job_info - output information about a specific Slurm
  *	job based upon message as loaded using slurm_load_jobs
  * IN job_ptr - an individual job information record pointer
  * IN one_liner - print as a single line if true
  * RET out - char * containing formatted output (must be freed after call)
  *           NULL is returned on failure.
  */
-extern char *slurm_sprint_job_info PARAMS(( job_info_t * job_ptr, 
-				            int one_liner ));
+extern char *slurm_sprint_job_info PARAMS(( job_info_t * job_ptr,
+					    int one_liner ));
 
 /*
- * slurm_update_job - issue RPC to a job's configuration per request, 
+ * slurm_update_job - issue RPC to a job's configuration per request,
  *	only usable by user root or (for some parameters) the job's owner
  * IN job_msg - description of job updates
  * RET 0 on success, otherwise return -1 and set errno to indicate the error
@@ -2056,13 +2056,13 @@ extern int slurm_update_job PARAMS(( job_desc_msg_t * job_msg )) ;
 \*****************************************************************************/
 
 /*
- * slurm_get_job_steps - issue RPC to get specific slurm job step   
+ * slurm_get_job_steps - issue RPC to get specific slurm job step
  *	configuration information if changed since update_time.
- *	a job_id value of NO_VAL implies all jobs, a step_id value of 
+ *	a job_id value of NO_VAL implies all jobs, a step_id value of
  *	NO_VAL implies all steps
  * IN update_time - time of current configuration data
  * IN job_id - get information for specific job id, NO_VAL for all jobs
- * IN step_id - get information for specific job step id, NO_VAL for all 
+ * IN step_id - get information for specific job step id, NO_VAL for all
  *	job steps
  * IN step_response_pptr - place to store a step response pointer
  * IN show_flags - job step filtering options
@@ -2070,12 +2070,12 @@ extern int slurm_update_job PARAMS(( job_desc_msg_t * job_msg )) ;
  * NOTE: free the response using slurm_free_job_step_info_response_msg
  */
 extern int slurm_get_job_steps PARAMS((
-	time_t update_time, uint32_t job_id, uint32_t step_id, 
+	time_t update_time, uint32_t job_id, uint32_t step_id,
 	job_step_info_response_msg_t **step_response_pptr,
 	uint16_t show_flags));
 
 /*
- * slurm_free_job_step_info_response_msg - free the job step 
+ * slurm_free_job_step_info_response_msg - free the job step
  *	information response message
  * IN msg - pointer to job step information response message
  * NOTE: buffer is loaded by slurm_get_job_steps.
@@ -2084,24 +2084,24 @@ extern void slurm_free_job_step_info_response_msg PARAMS((
 	job_step_info_response_msg_t * msg));
 
 /*
- * slurm_print_job_step_info_msg - output information about all Slurm 
+ * slurm_print_job_step_info_msg - output information about all Slurm
  *	job steps based upon message as loaded using slurm_get_job_steps
  * IN out - file to write to
  * IN job_step_info_msg_ptr - job step information message pointer
  * IN one_liner - print as a single line if true
  */
-extern void slurm_print_job_step_info_msg PARAMS(( 
-	FILE * out, job_step_info_response_msg_t * job_step_info_msg_ptr, 
+extern void slurm_print_job_step_info_msg PARAMS((
+	FILE * out, job_step_info_response_msg_t * job_step_info_msg_ptr,
 	int one_liner ));
 
 /*
- * slurm_print_job_step_info - output information about a specific Slurm 
+ * slurm_print_job_step_info - output information about a specific Slurm
  *	job step based upon message as loaded using slurm_get_job_steps
  * IN out - file to write to
  * IN job_ptr - an individual job step information record pointer
  * IN one_liner - print as a single line if true
  */
-extern void slurm_print_job_step_info PARAMS(( 
+extern void slurm_print_job_step_info PARAMS((
 	FILE * out, job_step_info_t * step_ptr, int one_liner ));
 
 /*
@@ -2114,17 +2114,17 @@ extern void slurm_print_job_step_info PARAMS((
  *   slurm_free_step_layout) on success, and NULL on error.
  */
 extern slurm_step_layout_t *slurm_job_step_layout_get PARAMS((uint32_t job_id,
-	                                                    uint32_t step_id));
+							    uint32_t step_id));
 /*
- * slurm_sprint_job_step_info - output information about a specific Slurm 
+ * slurm_sprint_job_step_info - output information about a specific Slurm
  *	job step based upon message as loaded using slurm_get_job_steps
  * IN job_ptr - an individual job step information record pointer
  * IN one_liner - print as a single line if true
  * RET out - char * containing formatted output (must be freed after call)
  *           NULL is returned on failure.
  */
-extern char *slurm_sprint_job_step_info PARAMS(( job_step_info_t * step_ptr, 
-					         int one_liner ));
+extern char *slurm_sprint_job_step_info PARAMS(( job_step_info_t * step_ptr,
+						 int one_liner ));
 
 void slurm_job_step_layout_free PARAMS((slurm_step_layout_t *layout));
 
@@ -2133,8 +2133,8 @@ void slurm_job_step_layout_free PARAMS((slurm_step_layout_t *layout));
 \*****************************************************************************/
 
 /*
- * slurm_load_node - issue RPC to get slurm all node configuration information 
- *	if changed since update_time 
+ * slurm_load_node - issue RPC to get slurm all node configuration information
+ *	if changed since update_time
  * IN update_time - time of current configuration data
  * IN node_info_msg_pptr - place to store a node configuration pointer
  * IN show_flags - node filtering options
@@ -2160,7 +2160,7 @@ extern void slurm_free_node_info_msg PARAMS((
  * IN node_info_msg_ptr - node information message pointer
  * IN one_liner - print as a single line if true
  */
-extern void slurm_print_node_info_msg PARAMS(( 
+extern void slurm_print_node_info_msg PARAMS((
 	FILE * out, node_info_msg_t * node_info_msg_ptr, int one_liner )) ;
 
 /*
@@ -2171,7 +2171,7 @@ extern void slurm_print_node_info_msg PARAMS((
  * IN node_scaling - number of nodes each node represents
  * IN one_liner - print as a single line if true
  */
-extern void slurm_print_node_table PARAMS(( 
+extern void slurm_print_node_table PARAMS((
 	FILE * out, node_info_t * node_ptr,
 	int node_scaling, int one_liner ));
 
@@ -2184,7 +2184,7 @@ extern void slurm_print_node_table PARAMS((
  * RET out - char * containing formatted output (must be freed after call)
  *           NULL is returned on failure.
  */
-extern char *slurm_sprint_node_table PARAMS(( node_info_t * node_ptr, 
+extern char *slurm_sprint_node_table PARAMS(( node_info_t * node_ptr,
 					      int node_scaling,
 					      int one_liner ));
 
@@ -2195,7 +2195,7 @@ extern char *slurm_sprint_node_table PARAMS(( node_info_t * node_ptr,
 void slurm_init_update_node_msg (update_node_msg_t * update_node_msg);
 
 /*
- * slurm_update_node - issue RPC to a node's configuration per request, 
+ * slurm_update_node - issue RPC to a node's configuration per request,
  *	only usable by user root
  * IN node_msg - description of node updates
  * RET 0 on success, otherwise return -1 and set errno to indicate the error
@@ -2208,17 +2208,17 @@ extern int slurm_update_node PARAMS(( update_node_msg_t * node_msg ));
 \*****************************************************************************/
 
 /*
- * slurm_load_topo - issue RPC to get slurm all switch topology configuration 
- *	information 
+ * slurm_load_topo - issue RPC to get slurm all switch topology configuration
+ *	information
  * IN node_info_msg_pptr - place to store a node configuration pointer
  * RET 0 or a slurm error code
  * NOTE: free the response using slurm_free_topo_info_msg
  */
-extern int slurm_load_topo PARAMS(( 
+extern int slurm_load_topo PARAMS((
 	topo_info_response_msg_t **topo_info_msg_pptr ));
 
 /*
- * slurm_free_topo_info_msg - free the switch topology configuration 
+ * slurm_free_topo_info_msg - free the switch topology configuration
  *	information response message
  * IN msg - pointer to switch topology configuration response message
  * NOTE: buffer is loaded by slurm_load_topo.
@@ -2226,14 +2226,14 @@ extern int slurm_load_topo PARAMS((
 extern void slurm_free_topo_info_msg PARAMS(( topo_info_response_msg_t *msg ));
 
 /*
- * slurm_print_topo_info_msg - output information about all switch topology 
- *	configuration information based upon message as loaded using 
+ * slurm_print_topo_info_msg - output information about all switch topology
+ *	configuration information based upon message as loaded using
  *	slurm_load_topo
  * IN out - file to write to
  * IN topo_info_msg_ptr - switch topology information message pointer
  * IN one_liner - print as a single line if not zero
  */
-extern void slurm_print_topo_info_msg PARAMS(( 
+extern void slurm_print_topo_info_msg PARAMS((
 	FILE * out, topo_info_response_msg_t *topo_info_msg_ptr, int one_liner )) ;
 
 /*
@@ -2245,7 +2245,7 @@ extern void slurm_print_topo_info_msg PARAMS((
  * RET out - char * containing formatted output (must be freed after call)
  *	   NULL is returned on failure.
  */
-extern void slurm_print_topo_record PARAMS((FILE * out, topo_info_t *topo_ptr, 
+extern void slurm_print_topo_record PARAMS((FILE * out, topo_info_t *topo_ptr,
 				    int one_liner ));
 
 /*****************************************************************************\
@@ -2279,18 +2279,18 @@ extern int slurm_get_select_nodeinfo PARAMS(
  *	SLURM PARTITION CONFIGURATION READ/PRINT/UPDATE FUNCTIONS
 \*****************************************************************************/
 
-/* 
- * slurm_init_part_desc_msg - initialize partition descriptor with 
- *	default values 
+/*
+ * slurm_init_part_desc_msg - initialize partition descriptor with
+ *	default values
  * IN/OUT update_part_msg - user defined partition descriptor
  */
 extern void slurm_init_part_desc_msg PARAMS((update_part_msg_t * update_part_msg ));
 
 /*
- * slurm_load_partitions - issue RPC to get slurm all partition configuration  
- *	information if changed since update_time 
+ * slurm_load_partitions - issue RPC to get slurm all partition configuration
+ *	information if changed since update_time
  * IN update_time - time of current configuration data
- * IN partition_info_msg_pptr - place to store a partition configuration 
+ * IN partition_info_msg_pptr - place to store a partition configuration
  *	pointer
  * IN show_flags - partitions filtering options
  * RET 0 or a slurm error code
@@ -2301,16 +2301,16 @@ extern int slurm_load_partitions PARAMS((
 	uint16_t show_flags));
 
 /*
- * slurm_free_partition_info_msg - free the partition information 
+ * slurm_free_partition_info_msg - free the partition information
  *	response message
  * IN msg - pointer to partition information response message
  * NOTE: buffer is loaded by slurm_load_partitions
  */
-extern void slurm_free_partition_info_msg PARAMS(( 
+extern void slurm_free_partition_info_msg PARAMS((
 	partition_info_msg_t * part_info_ptr ));
 
 /*
- * slurm_print_partition_info_msg - output information about all Slurm 
+ * slurm_print_partition_info_msg - output information about all Slurm
  *	partitions based upon message as loaded using slurm_load_partitions
  * IN out - file to write to
  * IN part_info_ptr - partitions information message pointer
@@ -2320,24 +2320,24 @@ extern void slurm_print_partition_info_msg PARAMS((
 	FILE * out, partition_info_msg_t * part_info_ptr, int one_liner ));
 
 /*
- * slurm_print_partition_info - output information about a specific Slurm 
+ * slurm_print_partition_info - output information about a specific Slurm
  *	partition based upon message as loaded using slurm_load_partitions
  * IN out - file to write to
  * IN part_ptr - an individual partition information record pointer
  * IN one_liner - print as a single line if true
  */
-extern void slurm_print_partition_info PARAMS(( 
+extern void slurm_print_partition_info PARAMS((
 	FILE *out , partition_info_t * part_ptr, int one_liner ));
 
 /*
- * slurm_sprint_partition_info - output information about a specific Slurm 
+ * slurm_sprint_partition_info - output information about a specific Slurm
  *	partition based upon message as loaded using slurm_load_partitions
  * IN part_ptr - an individual partition information record pointer
  * IN one_liner - print as a single line if true
  * RET out - char * with formatted output (must be freed after call)
  *           NULL is returned on failure.
  */
-extern char *slurm_sprint_partition_info PARAMS(( 
+extern char *slurm_sprint_partition_info PARAMS((
 		partition_info_t * part_ptr, int one_liner ));
 
 /*
@@ -2368,11 +2368,11 @@ extern int slurm_delete_partition PARAMS(( delete_part_msg_t * part_msg ));
 \*****************************************************************************/
 
 /*
- * slurm_init_resv_desc_msg - initialize reservation descriptor with 
- *	default values 
+ * slurm_init_resv_desc_msg - initialize reservation descriptor with
+ *	default values
  * OUT job_desc_msg - user defined partition descriptor
  */
-extern void slurm_init_resv_desc_msg PARAMS(( 
+extern void slurm_init_resv_desc_msg PARAMS((
 		resv_desc_msg_t * update_resv_msg ));
 /*
  * slurm_create_reservation - create a new reservation, only usable by user root
@@ -2384,7 +2384,7 @@ extern char * slurm_create_reservation PARAMS((
 		resv_desc_msg_t * resv_msg ));
 
 /*
- * slurm_update_reservation - modify an existing reservation, only usable by 
+ * slurm_update_reservation - modify an existing reservation, only usable by
  *	user root
  * IN resv_msg - description of reservation
  * RET 0 on success, otherwise return -1 and set errno to indicate the error
@@ -2392,64 +2392,64 @@ extern char * slurm_create_reservation PARAMS((
 extern int slurm_update_reservation PARAMS((resv_desc_msg_t * resv_msg));
 
 /*
- * slurm_delete_reservation - issue RPC to delete a reservation, only usable 
+ * slurm_delete_reservation - issue RPC to delete a reservation, only usable
  *	by user root
  * IN resv_msg - description of reservation to delete
  * RET 0 on success, otherwise return -1 and set errno to indicate the error
  */
-extern int slurm_delete_reservation PARAMS(( 
+extern int slurm_delete_reservation PARAMS((
 		reservation_name_msg_t * resv_msg ));
 
 /*
- * slurm_load_reservations - issue RPC to get all slurm reservation 
- *	configuration information if changed since update_time 
+ * slurm_load_reservations - issue RPC to get all slurm reservation
+ *	configuration information if changed since update_time
  * IN update_time - time of current configuration data
- * IN reserve_info_msg_pptr - place to store a reservation configuration 
+ * IN reserve_info_msg_pptr - place to store a reservation configuration
  *	pointer
  * RET 0 or a slurm error code
  * NOTE: free the response using slurm_free_reservation_info_msg
  */
-extern int slurm_load_reservations PARAMS(( time_t update_time, 
+extern int slurm_load_reservations PARAMS(( time_t update_time,
 		reserve_info_msg_t **resp ));
 
 /*
- * slurm_print_reservation_info_msg - output information about all Slurm 
+ * slurm_print_reservation_info_msg - output information about all Slurm
  *	reservations based upon message as loaded using slurm_load_reservation
  * IN out - file to write to
  * IN resv_info_ptr - reservation information message pointer
  * IN one_liner - print as a single line if true
  */
-void slurm_print_reservation_info_msg PARAMS(( FILE* out, 
+void slurm_print_reservation_info_msg PARAMS(( FILE* out,
 		reserve_info_msg_t * resv_info_ptr, int one_liner ));
 
 /*
- * slurm_print_reservation_info - output information about a specific Slurm 
+ * slurm_print_reservation_info - output information about a specific Slurm
  *	reservation based upon message as loaded using slurm_load_reservation
  * IN out - file to write to
  * IN resv_ptr - an individual reservation information record pointer
  * IN one_liner - print as a single line if true
  */
-void slurm_print_reservation_info PARAMS(( FILE* out, 
+void slurm_print_reservation_info PARAMS(( FILE* out,
 		reserve_info_t * resv_ptr, int one_liner ));
 
 /*
- * slurm_sprint_reservation_info - output information about a specific Slurm 
+ * slurm_sprint_reservation_info - output information about a specific Slurm
  *	reservation based upon message as loaded using slurm_load_reservations
  * IN resv_ptr - an individual reservation information record pointer
  * IN one_liner - print as a single line if true
  * RET out - char * containing formatted output (must be freed after call)
  *           NULL is returned on failure.
  */
-char *slurm_sprint_reservation_info PARAMS(( reserve_info_t * resv_ptr, 
+char *slurm_sprint_reservation_info PARAMS(( reserve_info_t * resv_ptr,
 		int one_liner ));
 
 /*
- * slurm_free_reservation_info_msg - free the reservation information 
+ * slurm_free_reservation_info_msg - free the reservation information
  *	response message
  * IN msg - pointer to reservation information response message
  * NOTE: buffer is loaded by slurm_load_reservation
  */
-extern void slurm_free_reservation_info_msg PARAMS(( 
+extern void slurm_free_reservation_info_msg PARAMS((
 	reserve_info_msg_t * resv_info_ptr ));
 
 /*****************************************************************************\
@@ -2465,14 +2465,14 @@ extern int slurm_ping PARAMS(( int primary ));
 
 /*
  * slurm_reconfigure - issue RPC to have Slurm controller (slurmctld)
- *	reload its configuration file 
+ *	reload its configuration file
  * RET 0 or a slurm error code
  */
 extern int slurm_reconfigure PARAMS(( void ));
 
 /*
  * slurm_shutdown - issue RPC to have Slurm controller (slurmctld)
- *	cease operations, both the primary and backup controller 
+ *	cease operations, both the primary and backup controller
  *	are shutdown.
  * IN options - 0: all slurm daemons are shutdown
  *              1: slurmctld generates a core file
@@ -2543,7 +2543,7 @@ extern int slurm_checkpoint_able PARAMS(( uint32_t job_id, uint32_t step_id,
  * IN step_id - job step on which to perform operation
  * RET 0 or a slurm error code
  */
-extern int slurm_checkpoint_disable PARAMS(( uint32_t job_id, 
+extern int slurm_checkpoint_disable PARAMS(( uint32_t job_id,
 		uint32_t step_id ));
 
 
@@ -2553,7 +2553,7 @@ extern int slurm_checkpoint_disable PARAMS(( uint32_t job_id,
  * IN step_id - job step on which to perform operation
  * RET 0 or a slurm error code
  */
-extern int slurm_checkpoint_enable PARAMS(( uint32_t job_id, 
+extern int slurm_checkpoint_enable PARAMS(( uint32_t job_id,
 					    uint32_t step_id ));
 
 /*
@@ -2565,9 +2565,9 @@ extern int slurm_checkpoint_enable PARAMS(( uint32_t job_id,
  * IN image_dir - directory to store image files
  * RET 0 or a slurm error code
  */
-extern int slurm_checkpoint_create PARAMS(( uint32_t job_id, 
-					    uint32_t step_id, 
-					    uint16_t max_wait, 
+extern int slurm_checkpoint_create PARAMS(( uint32_t job_id,
+					    uint32_t step_id,
+					    uint16_t max_wait,
 					    char *image_dir ));
 
 /*
@@ -2579,9 +2579,9 @@ extern int slurm_checkpoint_create PARAMS(( uint32_t job_id,
  * IN image_dir - directory to store image files
  * RET 0 or a slurm error code
  */
-extern int slurm_checkpoint_vacate PARAMS(( uint32_t job_id, 
-					    uint32_t step_id, 
-					    uint16_t max_wait, 
+extern int slurm_checkpoint_vacate PARAMS(( uint32_t job_id,
+					    uint32_t step_id,
+					    uint16_t max_wait,
 					    char *image_dir ));
 
 /*
@@ -2592,9 +2592,9 @@ extern int slurm_checkpoint_vacate PARAMS(( uint32_t job_id,
  * IN image_dir - directory to find checkpoint image files
  * RET 0 or a slurm error code
  */
-extern int slurm_checkpoint_restart PARAMS(( uint32_t job_id, 
-					     uint32_t step_id, 
-					     uint16_t stick, 
+extern int slurm_checkpoint_restart PARAMS(( uint32_t job_id,
+					     uint32_t step_id,
+					     uint16_t stick,
 					     char *image_dir ));
 
 /*
@@ -2607,10 +2607,10 @@ extern int slurm_checkpoint_restart PARAMS(( uint32_t job_id,
  * IN error_msg - error message, preserved for highest error_code
  * RET 0 or a slurm error code
  */
-extern int slurm_checkpoint_complete PARAMS(( uint32_t job_id, 
-					      uint32_t step_id, 
-					      time_t begin_time, 
-					      uint32_t error_code, 
+extern int slurm_checkpoint_complete PARAMS(( uint32_t job_id,
+					      uint32_t step_id,
+					      time_t begin_time,
+					      uint32_t error_code,
 					      char *error_msg ));
 
 /*
@@ -2624,11 +2624,11 @@ extern int slurm_checkpoint_complete PARAMS(( uint32_t job_id,
  * IN error_msg - error message, preserved for highest error_code
  * RET 0 or a slurm error code
  */
-extern int slurm_checkpoint_task_complete PARAMS(( uint32_t job_id, 
+extern int slurm_checkpoint_task_complete PARAMS(( uint32_t job_id,
 						   uint32_t step_id,
-						   uint32_t task_id, 
-						   time_t begin_time, 
-						   uint32_t error_code, 
+						   uint32_t task_id,
+						   time_t begin_time,
+						   uint32_t error_code,
 						   char *error_msg ));
 
 /*
@@ -2644,8 +2644,8 @@ extern int slurm_checkpoint_task_complete PARAMS(( uint32_t job_id,
  *	must be freed by the caller to prevent memory leak
  * RET 0 or a slurm error code
  */
-extern int slurm_checkpoint_error PARAMS(( uint32_t job_id, uint32_t step_id, 
-					   uint32_t *error_code, 
+extern int slurm_checkpoint_error PARAMS(( uint32_t job_id, uint32_t step_id,
+					   uint32_t *error_code,
 					   char **error_msg ));
 
 /*
@@ -2658,8 +2658,8 @@ extern int slurm_checkpoint_error PARAMS(( uint32_t job_id, uint32_t step_id,
  * IN nodelist: nodes to send the request
  * RET: 0 on success, non-zero on failure with errno set
  */
-extern int slurm_checkpoint_tasks PARAMS(( uint32_t job_id, uint16_t step_id, 
-					   time_t begin_time, char *image_dir, 
+extern int slurm_checkpoint_tasks PARAMS(( uint32_t job_id, uint16_t step_id,
+					   time_t begin_time, char *image_dir,
 					   uint16_t max_wait, char *nodelist));
 
 
@@ -2714,7 +2714,7 @@ extern hostlist_t slurm_hostlist_create PARAMS(( const char *hostlist ));
 /* slurm_hostlist_count():
  *
  * Return the number of hosts in hostlist hl.
- */ 
+ */
 extern int slurm_hostlist_count PARAMS((hostlist_t hl));
 
 /*
@@ -2726,8 +2726,8 @@ extern void slurm_hostlist_destroy PARAMS(( hostlist_t hl ));
 
 /* slurm_hostlist_find():
  *
- * Searches hostlist hl for the first host matching hostname 
- * and returns position in list if found. 
+ * Searches hostlist hl for the first host matching hostname
+ * and returns position in list if found.
  *
  * Returns -1 if host is not found.
  *
@@ -2740,14 +2740,14 @@ extern int slurm_hostlist_find PARAMS((hostlist_t hl, const char *hostname));
  *
  * The hosts argument may take the same form as in slurm_hostlist_create()
  *
- * Returns the number of hostnames inserted into the list, 
+ * Returns the number of hostnames inserted into the list,
  * or 0 on failure.
  */
 extern int slurm_hostlist_push PARAMS((hostlist_t hl, const char *hosts));
 
 /* slurm_hostlist_push_host():
  *
- * Push a single host onto the hostlist hl. 
+ * Push a single host onto the hostlist hl.
  * This function is more efficient than slurm_hostlist_push() for a single
  * hostname, since the argument does not need to be checked for ranges.
  *
@@ -2762,7 +2762,7 @@ extern int slurm_hostlist_push_host PARAMS((hostlist_t hl, const char *host));
  * or -1 if truncation occurred.
  *
  * The result will be NULL terminated.
- * 
+ *
  * slurm_hostlist_ranged_string() will write a bracketed hostlist representation
  * where possible.
  */
@@ -2783,7 +2783,7 @@ extern char * slurm_hostlist_shift PARAMS(( hostlist_t hl ));
 /* slurm_hostlist_uniq():
  *
  * Sort the hostlist hl and remove duplicate entries.
- * 
+ *
  */
 extern void slurm_hostlist_uniq PARAMS((hostlist_t hl));