From febba6ab47bc7d67aaa81e4439724e8666dd4079 Mon Sep 17 00:00:00 2001 From: Jan Frenzel <jan.frenzel@tu-dresden.de> Date: Thu, 5 Aug 2021 18:07:09 +0200 Subject: [PATCH] Corrected a few spelling mistakes. --- .../docs/data_transfer/data_mover.md | 2 +- .../docs/jobs_and_resources/overview.md | 2 +- .../docs/jobs_and_resources/sd_flex.md | 6 ++--- .../docs/jobs_and_resources/slurm.md | 15 ++++++++---- .../docs/software/compilers.md | 24 +++++++++---------- .../docs/software/fem_software.md | 2 +- .../docs/software/mathematics.md | 8 +++---- .../docs/software/power_ai.md | 2 +- 8 files changed, 33 insertions(+), 28 deletions(-) diff --git a/doc.zih.tu-dresden.de/docs/data_transfer/data_mover.md b/doc.zih.tu-dresden.de/docs/data_transfer/data_mover.md index 56b3e3daa..856af9f30 100644 --- a/doc.zih.tu-dresden.de/docs/data_transfer/data_mover.md +++ b/doc.zih.tu-dresden.de/docs/data_transfer/data_mover.md @@ -54,7 +54,7 @@ Options for dtrsync: -a, --archive archive mode; equals -rlptgoD (no -H,-A,-X) - -r, --recursive recurse into directorie + -r, --recursive recurse into directories -l, --links copy symlinks as symlinks -p, --perms preserve permissions -t, --times preserve modification times diff --git a/doc.zih.tu-dresden.de/docs/jobs_and_resources/overview.md b/doc.zih.tu-dresden.de/docs/jobs_and_resources/overview.md index 55d5c40ad..142662727 100644 --- a/doc.zih.tu-dresden.de/docs/jobs_and_resources/overview.md +++ b/doc.zih.tu-dresden.de/docs/jobs_and_resources/overview.md @@ -29,7 +29,7 @@ ZIH uses the batch system Slurm for resource management and job scheduling. ??? note "Batch System" The batch system is the central organ of every HPC system users interact with its compute - resources. The batchsystem finds an adequate compute system (partition/island) for your compute + resources. The batch system finds an adequate compute system (partition/island) for your compute jobs. It organizes the queueing and messaging, if all resources are in use. If resources are available for your job, the batch system allocates and connects to these resources, transfers run-time environment, and starts the job. diff --git a/doc.zih.tu-dresden.de/docs/jobs_and_resources/sd_flex.md b/doc.zih.tu-dresden.de/docs/jobs_and_resources/sd_flex.md index 8f536f9cd..04624da4e 100644 --- a/doc.zih.tu-dresden.de/docs/jobs_and_resources/sd_flex.md +++ b/doc.zih.tu-dresden.de/docs/jobs_and_resources/sd_flex.md @@ -2,7 +2,7 @@ - Hostname: taurussmp8 - Access to all shared file systems -- SLURM partition `julia` +- Slurm partition `julia` - 32 x Intel(R) Xeon(R) Platinum 8276M CPU @ 2.20GHz (28 cores) - 48 TB RAM (usable: 47 TB - one TB is used for cache coherence protocols) @@ -14,7 +14,7 @@ There are 370 TB of NVMe devices installed. For immediate access for all project of fast NVMe storage is available at `/nvme/1/<projectname>`. For testing, we have set a quota of 100 GB per project on this NVMe storage.This is -With a more detailled proposal on how this unique system (large shared memory + NVMe storage) can +With a more detailed proposal on how this unique system (large shared memory + NVMe storage) can speed up their computations, a project's quota can be increased or dedicated volumes of up to the full capacity can be set up. @@ -26,7 +26,7 @@ full capacity can be set up. variables, so that OpenMPI uses shared memory instead of Infiniband for message transport. `export OMPI_MCA_pml=ob1;   export OMPI_MCA_mtl=^mxm` - Use `I_MPI_FABRICS=shm` so that Intel MPI doesn't even consider - using InfiniBand devices itself, but only shared-memory instead + using Infiniband devices itself, but only shared-memory instead ## Open for Testing diff --git a/doc.zih.tu-dresden.de/docs/jobs_and_resources/slurm.md b/doc.zih.tu-dresden.de/docs/jobs_and_resources/slurm.md index d45f5b34d..0c4d3d92a 100644 --- a/doc.zih.tu-dresden.de/docs/jobs_and_resources/slurm.md +++ b/doc.zih.tu-dresden.de/docs/jobs_and_resources/slurm.md @@ -80,7 +80,7 @@ one job per user. Please check the availability of nodes there with `sinfo -p in Slurm will forward your X11 credentials to the first (or even all) node for a job with the (undocumented) --x11 option. For example, an -interactive session for 1 hour with Matlab using eigth cores can be +interactive session for 1 hour with Matlab using eight cores can be started with: ```Shell Session @@ -190,7 +190,7 @@ specifically for their target. srun ./path/to/binary ``` -#### Multiple Programms Running Simultaneously in a Job +#### Multiple Programs Running Simultaneously in a Job In this short example, our goal is to run four instances of a program concurrently in a **single** batch script. Of course we could also start a batch script four times with sbatch but this is not @@ -251,8 +251,13 @@ other jobs, even if it doesn't use up all resources in the nodes: ```Bash #!/bin/bash -#SBATCH -J Benchmark<br />#SBATCH -p haswell<br />#SBATCH --nodes=2<br />#SBATCH --ntasks-per-node=2 -#SBATCH --cpus-per-task=8<br />#SBATCH --exclusive # ensure that nobody spoils my measurement on 2 x 2 x 8 cores<br />#SBATCH --mail-user=your.name@tu-dresden.de +#SBATCH -J Benchmark +#SBATCH -p haswell +#SBATCH --nodes=2 +#SBATCH --ntasks-per-node=2 +#SBATCH --cpus-per-task=8 +#SBATCH --exclusive # ensure that nobody spoils my measurement on 2 x 2 x 8 cores +#SBATCH --mail-user=your.name@tu-dresden.de #SBATCH --time=00:10:00 srun ./my_benchmark @@ -400,7 +405,7 @@ src="data:;base64,iVBORw0KGgoAAAANSUhEUgAAAvoAAADyCAIAAACzsfbGAAAABmJLR0QA/wD/AP ### Node Features for Selective Job Submission The nodes in our HPC system are becoming more diverse in multiple aspects: hardware, mounted -storage, software. The system administrators can decribe the set of properties and it is up to the +storage, software. The system administrators can describe the set of properties and it is up to the user to specify her/his requirements. These features should be thought of as changing over time (e.g. a file system get stuck on a certain node). diff --git a/doc.zih.tu-dresden.de/docs/software/compilers.md b/doc.zih.tu-dresden.de/docs/software/compilers.md index e6d56a032..19a70e463 100644 --- a/doc.zih.tu-dresden.de/docs/software/compilers.md +++ b/doc.zih.tu-dresden.de/docs/software/compilers.md @@ -37,7 +37,7 @@ optimization, data alignment and so on. You can list all available compiler opti `-help`. Reading the man-pages is a good idea, too. The user benefits from the (nearly) same set of compiler flags for optimization for the C,C++, and -Fortran-compilers. In the following table, only a couple of important compiler-dependant options are +Fortran-compilers. In the following table, only a couple of important compiler-dependent options are listed. For more detailed information, the user should refer to the man pages or use the option -help to list all options of the compiler. @@ -96,17 +96,17 @@ parallelism in the code. Therefore it is sometimes necessary to provide the compiler with some hints. Some possible directives are (Fortran style): -| | | -|--------------------------|-----------------------------------| -| `CDEC$ ivdep` | ignore assumed vector dependences | -| `CDEC$ swp` | try to software-pipeline | -| `CDEC$ noswp` | disable software-pipeline | -| `CDEC$ loop count (n)` | hint for optimization | -| `CDEC$ distribute point` | split this large loop | -| `CDEC$ unroll (n)` | unroll (n) times | -| `CDEC$ nounroll` | do not unroll | -| `CDEC$ prefetch a` | prefetch array a | -| `CDEC$ noprefetch a` | do not prefetch array a | +| | | +|--------------------------|------------------------------------| +| `CDEC$ ivdep` | ignore assumed vector dependencies | +| `CDEC$ swp` | try to software-pipeline | +| `CDEC$ noswp` | disable software-pipeline | +| `CDEC$ loop count (n)` | hint for optimization | +| `CDEC$ distribute point` | split this large loop | +| `CDEC$ unroll (n)` | unroll (n) times | +| `CDEC$ nounroll` | do not unroll | +| `CDEC$ prefetch a` | prefetch array a | +| `CDEC$ noprefetch a` | do not prefetch array a | The compiler directives are the same for `ifort` and `icc` . The syntax for C/C++ is like `#pragma ivdep`, `#pragma swp`, and so on. diff --git a/doc.zih.tu-dresden.de/docs/software/fem_software.md b/doc.zih.tu-dresden.de/docs/software/fem_software.md index 5c576c143..bd65ea983 100644 --- a/doc.zih.tu-dresden.de/docs/software/fem_software.md +++ b/doc.zih.tu-dresden.de/docs/software/fem_software.md @@ -149,7 +149,7 @@ runwb2 The ANSYS workbench (runwb2) can also be used in a batch script to start calculations (the solver, not GUI) from a workbench project into the background. To do so, you have to specify the -B parameter (for batch -mode), -F for your project file, and can then either add differerent +mode), -F for your project file, and can then either add different commands via -E parameters directly, or specify a workbench script file containing commands via -R. diff --git a/doc.zih.tu-dresden.de/docs/software/mathematics.md b/doc.zih.tu-dresden.de/docs/software/mathematics.md index fc5d7e894..9edba0288 100644 --- a/doc.zih.tu-dresden.de/docs/software/mathematics.md +++ b/doc.zih.tu-dresden.de/docs/software/mathematics.md @@ -16,7 +16,7 @@ interface capabilities within a document-like user interface paradigm. ### Fonts -To remotely use the graphical frontend you have to add the Mathematica fonts to the local +To remotely use the graphical frontend, you have to add the Mathematica fonts to the local fontmanager. #### Linux Workstation @@ -36,7 +36,7 @@ You have to add additional Mathematica fonts at your local PC If you use **Xming** as X-server at your PC (refer to [remote access from Windows](../access/ssh_mit_putty.md), follow these steps: -1. Create a new folder `Mathematica` in the diretory `fonts` of the installation directory of Xming +1. Create a new folder `Mathematica` in the directory `fonts` of the installation directory of Xming (mostly: `C:\\Programme\\Xming\\fonts\\`) 1. Extract the fonts archive into this new directory `Mathematica`. In result you should have the two directories `DBF` and `Type1`. @@ -56,7 +56,7 @@ C:\WINDOWS\Fonts ### Mathematica and Slurm -Please use the batchsystem Slurm for running calculations. This is a small example for a batch +Please use the batch system Slurm for running calculations. This is a small example for a batch script, that you should prepare and start with the command `sbatch <scriptname>`. The File `mathtest.m` is your input script that includes the calculation statements for Mathematica. The file `mathtest.output` will hold the results of your calculations. @@ -154,7 +154,7 @@ srun --pty matlab -nodisplay -r basename_of_your_matlab_script #NOTE: you must o many instances of your calculation as you'd like, since it does not need a license during runtime when compiled to a binary. -You can find detailled documentation on the Matlab compiler at +You can find detailed documentation on the Matlab compiler at [Mathworks' help pages](https://de.mathworks.com/help/compiler/). ### Using the MATLAB Compiler (mcc) diff --git a/doc.zih.tu-dresden.de/docs/software/power_ai.md b/doc.zih.tu-dresden.de/docs/software/power_ai.md index 7a03aa31e..dc0fa59b3 100644 --- a/doc.zih.tu-dresden.de/docs/software/power_ai.md +++ b/doc.zih.tu-dresden.de/docs/software/power_ai.md @@ -9,7 +9,7 @@ are valid for PowerAI version 1.5.4 - \<a href="<https://www.ibm.com/support/knowledgecenter/en/SS5SF7_1.5.3/welcome/welcome.htm>" target="\_blank" title="Landing Page">Landing Page\</a> (note that - you can select differnet PowerAI versions with the drop down menu + you can select different PowerAI versions with the drop down menu "Change Product or version") - \<a href="<https://developer.ibm.com/linuxonpower/deep-learning-powerai/>" -- GitLab