/*****************************************************************************\ * gres_gpu.c - Support GPUs as a generic resources. ***************************************************************************** * Copyright (C) 2010 Lawrence Livermore National Security. * Produced at Lawrence Livermore National Laboratory (cf, DISCLAIMER). * Written by Morris Jette <jette1@llnl.gov> * CODE-OCEC-09-009. All rights reserved. * * This file is part of SLURM, a resource management program. * For details, see <http://www.schedmd.com/slurmdocs/>. * Please also read the included file: DISCLAIMER. * * SLURM is free software; you can redistribute it and/or modify it under * the terms of the GNU General Public License as published by the Free * Software Foundation; either version 2 of the License, or (at your option) * any later version. * * In addition, as a special exception, the copyright holders give permission * to link the code of portions of this program with the OpenSSL library under * certain conditions as described in each individual source file, and * distribute linked combinations including the two. You must obey the GNU * General Public License in all respects for all of the code used other than * OpenSSL. If you modify file(s) with this exception, you may extend this * exception to your version of the file(s), but you are not obligated to do * so. If you do not wish to do so, delete this exception statement from your * version. If you delete this exception statement from all source files in * the program, then also delete it here. * * SLURM is distributed in the hope that it will be useful, but WITHOUT ANY * WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS * FOR A PARTICULAR PURPOSE. See the GNU General Public License for more * details. * * You should have received a copy of the GNU General Public License along * with SLURM; if not, write to the Free Software Foundation, Inc., * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. \*****************************************************************************/ #if HAVE_CONFIG_H # include "config.h" # if STDC_HEADERS # include <string.h> # endif # if HAVE_SYS_TYPES_H # include <sys/types.h> # endif /* HAVE_SYS_TYPES_H */ # if HAVE_UNISTD_H # include <unistd.h> # endif # if HAVE_INTTYPES_H # include <inttypes.h> # else /* ! HAVE_INTTYPES_H */ # if HAVE_STDINT_H # include <stdint.h> # endif # endif /* HAVE_INTTYPES_H */ #else /* ! HAVE_CONFIG_H */ # include <sys/types.h> # include <unistd.h> # include <stdint.h> # include <string.h> #endif /* HAVE_CONFIG_H */ #ifdef HAVE_HWLOC # include <hwloc.h> #endif /* HAVE_HWLOC */ #include <stdio.h> #include <stdlib.h> #include <ctype.h> #include "slurm/slurm.h" #include "slurm/slurm_errno.h" #include "src/common/slurm_xlator.h" #include "src/common/bitstring.h" #include "src/common/env.h" #include "src/common/gres.h" #include "src/common/list.h" #include "src/common/xstring.h" /* * These variables are required by the generic plugin interface. If they * are not found in the plugin, the plugin loader will ignore it. * * plugin_name - A string giving a human-readable description of the * plugin. There is no maximum length, but the symbol must refer to * a valid string. * * plugin_type - A string suggesting the type of the plugin or its * applicability to a particular form of data or method of data handling. * If the low-level plugin API is used, the contents of this string are * unimportant and may be anything. SLURM uses the higher-level plugin * interface which requires this string to be of the form * * <application>/<method> * * where <application> is a description of the intended application of * the plugin (e.g., "auth" for SLURM authentication) and <method> is a * description of how this plugin satisfies that application. SLURM will * only load authentication plugins if the plugin_type string has a prefix * of "auth/". * * plugin_version - Specifies the version number of the plugin. This would * typically be the same for all plugins. */ const char plugin_name[] = "Gres GPU plugin"; const char plugin_type[] = "gres/gpu"; const uint32_t plugin_version = 110; static char gres_name[] = "gpu"; static int *gpu_devices = NULL; static int nb_available_files; /* * We could load gres state or validate it using various mechanisms here. * This only validates that the configuration was specified in gres.conf. * In the general case, no code would need to be changed. */ extern int node_config_load(List gres_conf_list) { int i, rc = SLURM_ERROR; ListIterator iter; gres_slurmd_conf_t *gres_slurmd_conf; int nb_gpu = 0; /* Number of GPUs in the list */ int available_files_index = 0; xassert(gres_conf_list); iter = list_iterator_create(gres_conf_list); if (iter == NULL) fatal("list_iterator_create: malloc failure"); while ((gres_slurmd_conf = list_next(iter))) { if (strcmp(gres_slurmd_conf->name, gres_name)) continue; rc = SLURM_SUCCESS; if (gres_slurmd_conf->file) nb_gpu++; } list_iterator_destroy(iter); gpu_devices = NULL; nb_available_files = -1; /* (Re-)Allocate memory if number of files changed */ if (nb_gpu != nb_available_files) { xfree(gpu_devices); /* No-op if NULL */ gpu_devices = (int *) xmalloc(sizeof(int) * nb_gpu); nb_available_files = nb_gpu; for (i = 0; i < nb_available_files; i++) gpu_devices[i] = -1; } iter = list_iterator_create(gres_conf_list); if (iter == NULL) fatal("list_iterator_create: malloc failure"); while ((gres_slurmd_conf = list_next(iter))) { if ((strcmp(gres_slurmd_conf->name, gres_name) == 0) && gres_slurmd_conf->file) { /* Populate gpu_devices array with number * at end of the file name */ for (i = 0; gres_slurmd_conf->file[i]; i++) { if (!isdigit(gres_slurmd_conf->file[i])) continue; gpu_devices[available_files_index] = atoi(gres_slurmd_conf->file + i); break; } available_files_index++; } } list_iterator_destroy(iter); if (rc != SLURM_SUCCESS) fatal("%s failed to load configuration", plugin_name); for (i = 0; i < nb_available_files; i++) info("gpu %d is device number %d", i, gpu_devices[i]); return rc; } /* * Set environment variables as appropriate for a job (i.e. all tasks) based * upon the job's GRES state. */ extern void job_set_env(char ***job_env_ptr, void *gres_ptr) { int i, len; char *dev_list = NULL; gres_job_state_t *gres_job_ptr = (gres_job_state_t *) gres_ptr; if ((gres_job_ptr != NULL) && (gres_job_ptr->node_cnt == 1) && (gres_job_ptr->gres_bit_alloc != NULL) && (gres_job_ptr->gres_bit_alloc[0] != NULL)) { len = bit_size(gres_job_ptr->gres_bit_alloc[0]); for (i=0; i<len; i++) { if (!bit_test(gres_job_ptr->gres_bit_alloc[0], i)) continue; if (!dev_list) dev_list = xmalloc(128); else xstrcat(dev_list, ","); if (gpu_devices && (gpu_devices[i] >= 0)) xstrfmtcat(dev_list, "%d", gpu_devices[i]); else xstrfmtcat(dev_list, "%d", i); } } if (dev_list) { env_array_overwrite(job_env_ptr,"CUDA_VISIBLE_DEVICES", dev_list); xfree(dev_list); } else { /* The gres.conf file must identify specific device files * in order to set the CUDA_VISIBLE_DEVICES env var */ error("gres/gpu unable to set CUDA_VISIBLE_DEVICES, " "no device files configured"); } } /* * Set environment variables as appropriate for a job (i.e. all tasks) based * upon the job step's GRES state. */ extern void step_set_env(char ***job_env_ptr, void *gres_ptr) { int i, len; char *dev_list = NULL; gres_step_state_t *gres_step_ptr = (gres_step_state_t *) gres_ptr; if ((gres_step_ptr != NULL) && (gres_step_ptr->node_cnt == 1) && (gres_step_ptr->gres_bit_alloc != NULL) && (gres_step_ptr->gres_bit_alloc[0] != NULL)) { len = bit_size(gres_step_ptr->gres_bit_alloc[0]); for (i=0; i<len; i++) { if (!bit_test(gres_step_ptr->gres_bit_alloc[0], i)) continue; if (!dev_list) dev_list = xmalloc(128); else xstrcat(dev_list, ","); if (gpu_devices && (gpu_devices[i] >= 0)) xstrfmtcat(dev_list, "%d", gpu_devices[i]); else xstrfmtcat(dev_list, "%d", i); } } if (dev_list) { env_array_overwrite(job_env_ptr,"CUDA_VISIBLE_DEVICES", dev_list); xfree(dev_list); } else { /* The gres.conf file must identify specific device files * in order to set the CUDA_VISIBLE_DEVICES env var */ error("gres/gpu unable to set CUDA_VISIBLE_DEVICES, " "no device files configured"); } } /* Send GRES information to slurmstepd on the specified file descriptor*/ extern void send_stepd(int fd) { int i; safe_write(fd, &nb_available_files, sizeof(int)); for (i = 0; i < nb_available_files; i++) safe_write(fd, &gpu_devices[i], sizeof(int)); return; rwfail: error("gres_plugin_send_stepd failed"); } /* Receive GRES information from slurmd on the specified file descriptor*/ extern void recv_stepd(int fd) { int i; safe_read(fd, &nb_available_files, sizeof(int)); if (nb_available_files > 0) gpu_devices = xmalloc(sizeof(int) * nb_available_files); for (i = 0; i < nb_available_files; i++) safe_read(fd, &gpu_devices[i], sizeof(int)); return; rwfail: error("gres_plugin_recv_stepd failed"); }