From 927f634d72a6fa2b78abf1db2341b5acbbd4b79e Mon Sep 17 00:00:00 2001 From: Moe Jette <jette1@llnl.gov> Date: Tue, 29 Mar 2011 02:23:59 +0000 Subject: [PATCH] update to cray admin web page --- doc/html/cray.shtml | 40 +++++++++++++++++++++++++++++++++++----- 1 file changed, 35 insertions(+), 5 deletions(-) diff --git a/doc/html/cray.shtml b/doc/html/cray.shtml index 315286769aa..077900ddbcb 100644 --- a/doc/html/cray.shtml +++ b/doc/html/cray.shtml @@ -91,7 +91,7 @@ http://code.google.com/p/munge/downloads/list</a></li> which has already the correct paths - in particular, the libmunge.* need to reside in <i>/usr/lib64</i> since SLURM plugins call them.</p> -<p>Install tar-ball by +<p>Install the tar-ball by executing: <ul> <li>scp $LIBROOT/munge_build-.*YYYY-MM-DD.tar.gz boot:</li> <li>ssh boot</li> @@ -111,13 +111,16 @@ reside in <i>/usr/lib64</i> since SLURM plugins call them.</p> <i>mkdir --mode=0711 -vp /var/lib/munge</i><br> <i>mkdir --mode=0700 -vp /var/log/munge</i><br> <i>mkdir --mode=0755 -vp /var/run/munge</i></li> -<li><i>module load slurm</i> as root, to get the right slurm paths</li> +<li><i>module load slurm</i> as root, to get the correct SLURM paths</li> <li><i>munged --key-file /opt/slurm/munge/etc/munge.key</i></li> <li>Try a "<i>munge -n</i>" to see if munged accepts input</li> </ol></p> -<p>When done, verify network connectivity by executing<br> - $ <i>munge -n |ssh other-login-host /opt/slurm/munge/bin/unmunge</i></p> +<p>When done, verify network connectivity by executing: +<ul> +<li>munge -n |ssh other-login-host /opt/slurm/munge/bin/unmunge</li> +</ul> + <p>If you decide to keep the installation, you may be interested in automating the process using an <i>init.d</i> script distributed with the Munge as @@ -221,8 +224,35 @@ SDBuser=alps_user SDBdb=XT5istanbul </pre> +<p>One additional configuration script can be used to insure that the slurmd +daemons execute with the highest resource limits possible, overriding default +limits on Suse systems. Depending upon what resource limits are propagated +from the user's environment, lower limits may apply to user jobs, but this +script will insure that higher limits are possible. Copy the file +<i>contribs/cray/etc_sysconfig_slurm</i> into <i>/etc/sysconfig/slurm</i> +for these limits to take effect. This script is executed from +<i>/etc/init.d/slurm</i>, which is typically executed to start the SLURM +daemons. An excerpt of <i>contribs/cray/etc_sysconfig_slurm</i>is shown +below.</p> + +<pre> +# +# /etc/sysconfig/slurm for Cray XT/XE systems +# +# Cray is SuSe-based, which means that ulimits from +# /etc/security/limits.conf will get picked up any time SLURM is +# restarted e.g. via pdsh/ssh. Since SLURM respects configured limits, +# this can mean that for instance batch jobs get killed as a result +# of configuring CPU time limits. Set sane start limits here. +# +# Values were taken from pam-1.1.2 Debian package +ulimit -t unlimited # max amount of CPU time in seconds +ulimit -d unlimited # max size of a process's data segment in KB +</pre> + + <p class="footer"><a href="#top">top</a></p> -<p style="text-align:center;">Last modified 20 February 2011</p></td> +<p style="text-align:center;">Last modified 28 March 2011</p></td> <!--#include virtual="footer.txt"--> -- GitLab