diff --git a/doc.zih.tu-dresden.de/docs/software/Visualization.md b/doc.zih.tu-dresden.de/docs/software/Visualization.md index 31c83ce5fce14717a0f957b623c2f9138e02e15c..2d7b0787d7e1652504c9d8f0d8baafa91f940b4d 100644 --- a/doc.zih.tu-dresden.de/docs/software/Visualization.md +++ b/doc.zih.tu-dresden.de/docs/software/Visualization.md @@ -4,12 +4,14 @@ [ParaView](https://paraview.org) is an open-source, multi-platform data analysis and visualization application. It is available on Taurus under -the `ParaView` [modules](RuntimeEnvironment#Modules) +the `ParaView` [modules](modules.md#modules-environment) - taurus$ module avail ParaView +```Bash +taurus$ module avail ParaView - ParaView/5.4.1-foss-2018b-mpi (D) ParaView/5.5.2-intel-2018a-mpi ParaView/5.7.0-osmesa - ParaView/5.4.1-intel-2018a-mpi ParaView/5.6.2-foss-2019b-Python-3.7.4-mpi ParaView/5.7.0 + ParaView/5.4.1-foss-2018b-mpi (D) ParaView/5.5.2-intel-2018a-mpi ParaView/5.7.0-osmesa + ParaView/5.4.1-intel-2018a-mpi ParaView/5.6.2-foss-2019b-Python-3.7.4-mpi ParaView/5.7.0 +``` The ParaView package comprises different tools which are designed to meet interactive, batch and in-situ workflows. @@ -27,51 +29,55 @@ invoked using this very mpiexec*** command. Make sure to not use *srun or mpiexec* from another MPI module, e.g., check what *mpiexec* is in the path: - taurus$ module load ParaView/5.7.0-osmesa - taurus$ which mpiexec - /sw/installed/ParaView/5.7.0-osmesa/bin/mpiexec +```Bash +taurus$ module load ParaView/5.7.0-osmesa +taurus$ which mpiexec +/sw/installed/ParaView/5.7.0-osmesa/bin/mpiexec +``` -The resources for the MPI processes have to be allocated via the Slurm -option *-c NUM* (not *-n*, as it would be usually for MPI processes). It -might be valuable in terms of runtime to bind/pin the MPI processes to -hardware. A convenient option is *-bind-to core*. All other options can -be obtained by *taurus$ mpiexec -bind-to -help* or from \<a -href="<https://wiki.mpich.org/mpich/index.php/Using_the_Hydra_Process_Manager#Process-core_Binding%7Cwiki.mpich.org>" -title="<https://wiki.mpich.org/mpich/index.php/Using_the_Hydra_Process_Manager#Process-core_Binding%7Cwiki.mpich.org>"><https://wiki.mpich.org/mpich/index.php/Using_the_Hydra_Process_Manager#Process-core_Binding%7Cwiki.mpich.org>\</a> +The resources for the MPI processes have to be allocated via the Slurm option *-c NUM* (not *-n*, as +it would be usually for MPI processes). It might be valuable in terms of runtime to bind/pin the MPI +processes to hardware. A convenient option is *-bind-to core*. All other options can be obtained by +*taurus$ mpiexec -bind-to -help* or from +[https://wiki.mpich.org/mpich/index.php/Using_the_Hydra_Process_Manager#Process-core_Binding%7Cwiki.mpich.org]. Jobfile - #!/bin/bash +```Bash +#!/bin/bash - #SBATCH -N 1 - #SBATCH -c 12 - #SBATCH --time=01:00:00 +#SBATCH -N 1 +#SBATCH -c 12 +#SBATCH --time=01:00:00 - # Make sure to only use ParaView - module purge - module load ParaView/5.7.0-osmesa +# Make sure to only use ParaView +module purge +module load ParaView/5.7.0-osmesa - pvbatch --mpi --force-offscreen-rendering pvbatch-script.py +pvbatch --mpi --force-offscreen-rendering pvbatch-script.py +``` -Interactive allocation via \`salloc\` +Interactive allocation via `salloc` - taurus$ salloc -N 1 -c 16 --time=01:00:00 bash - salloc: Pending job allocation 336202 - salloc: job 336202 queued and waiting for resources - salloc: job 336202 has been allocated resources - salloc: Granted job allocation 336202 - salloc: Waiting for resource configuration - salloc: Nodes taurusi6605 are ready for job +```Bash +taurus$ salloc -N 1 -c 16 --time=01:00:00 bash +salloc: Pending job allocation 336202 +salloc: job 336202 queued and waiting for resources +salloc: job 336202 has been allocated resources +salloc: Granted job allocation 336202 +salloc: Waiting for resource configuration +salloc: Nodes taurusi6605 are ready for job - # Make sure to only use ParaView - taurus$ module purge - taurus$ module load ParaView/5.7.0-osmesa +# Make sure to only use ParaView +taurus$ module purge +taurus$ module load ParaView/5.7.0-osmesa - # Go to working directory, e.g. workspace - taurus$ cd /path/to/workspace +# Go to working directory, e.g. workspace +taurus$ cd /path/to/workspace - # Execute pvbatch using 16 MPI processes in parallel on allocated resources - taurus$ pvbatch --mpi --force-offscreen-rendering pvbatch-script.py +# Execute pvbatch using 16 MPI processes in parallel on allocated resources +taurus$ pvbatch --mpi --force-offscreen-rendering pvbatch-script.py +``` ### Using GPUs @@ -83,41 +89,43 @@ ParaView/5.9.0-RC1-egl-mpi-Python-3.8, and pass the option Jobfile - #!/bin/bash +```Bash +#!/bin/bash - #SBATCH -N 1 - #SBATCH -c 12 - #SBATCH --gres=gpu:2 - #SBATCH --partition=gpu2 - #SBATCH --time=01:00:00 +#SBATCH -N 1 +#SBATCH -c 12 +#SBATCH --gres=gpu:2 +#SBATCH --partition=gpu2 +#SBATCH --time=01:00:00 - # Make sure to only use ParaView - module purge - module load ParaView/5.9.0-RC1-egl-mpi-Python-3.8 +# Make sure to only use ParaView +module purge +module load ParaView/5.9.0-RC1-egl-mpi-Python-3.8 - mpiexec -n $SLURM_CPUS_PER_TASK -bind-to core pvbatch --mpi --egl-device-index=$CUDA_VISIBLE_DEVICES --force-offscreen-rendering pvbatch-script.py - #or - pvbatch --mpi --egl-device-index=$CUDA_VISIBLE_DEVICES --force-offscreen-rendering pvbatch-script.py +mpiexec -n $SLURM_CPUS_PER_TASK -bind-to core pvbatch --mpi --egl-device-index=$CUDA_VISIBLE_DEVICES --force-offscreen-rendering pvbatch-script.py +#or +pvbatch --mpi --egl-device-index=$CUDA_VISIBLE_DEVICES --force-offscreen-rendering pvbatch-script.py +``` ## Interactive Mode There are different ways of using ParaView on the cluster: -- GUI via NICE DCV on a GPU node -- Client-/Server mode with MPI-parallel off-screen-rendering -- GUI via X forwarding +- GUI via NICE DCV on a GPU node +- Client-/Server mode with MPI-parallel off-screen-rendering +- GUI via X forwarding -#### Using the GUI via NICE DCV on a GPU node +### Using the GUI via NICE DCV on a GPU node -This option provides hardware accelerated OpenGL and might provide the -best performance and smooth handling. First, you need to open a DCV -session, so please follow the instructions under [virtual -desktops](Compendium.VirtualDesktops). Start a terminal (right-click on -desktop -> Terminal) in your virtual desktop session, then load the -ParaView module as usual and start the GUI: +This option provides hardware accelerated OpenGL and might provide the best performance and smooth +handling. First, you need to open a DCV session, so please follow the instructions under +[virtual desktops](VirtualDesktops.md). Start a terminal (right-click on desktop -> Terminal) in your +virtual desktop session, then load the ParaView module as usual and start the GUI: - taurus$ module load ParaView/5.7.0 - paraview +```Bash +taurus$ module load ParaView/5.7.0 +paraview +``` Since your DCV session already runs inside a job, i.e., it has been scheduled to a compute node, no `srun command` is necessary here. @@ -136,18 +144,19 @@ version of the ParaView module, which supports offscreen-rendering. Then, start the `pvserver` via `srun` in parallel using multiple MPI processes: - taurus$ module ParaView/5.7.0-osmesa - taurus$ srun -N1 -n8 --mem-per-cpu=2500 -p interactive --pty pvserver --force-offscreen-rendering - srun: job 2744818 queued and waiting for resources - srun: job 2744818 has been allocated resources - Waiting for client... - Connection URL: cs://taurusi6612.taurus.hrsk.tu-dresden.de:11111 - Accepting connection(s): taurusi6612.taurus.hrsk.tu-dresden.de:11111 +```Bash +taurus$ module ParaView/5.7.0-osmesa +taurus$ srun -N1 -n8 --mem-per-cpu=2500 -p interactive --pty pvserver --force-offscreen-rendering +srun: job 2744818 queued and waiting for resources +srun: job 2744818 has been allocated resources +Waiting for client... +Connection URL: cs://taurusi6612.taurus.hrsk.tu-dresden.de:11111 +Accepting connection(s): taurusi6612.taurus.hrsk.tu-dresden.de:11111 +``` -If the default port 11111 is already in use, an alternative port can be -specified via *`-sp=port`.*Once the resources are allocated, the -pvserver is started in parallel and connection information are -outputed.*\<br />* +If the default port 11111 is already in use, an alternative port can be specified via `-sp=port`. +*Once the resources are allocated, the pvserver is started in parallel and connection information +are outputed.* This contains the node name which your job and server runs on. However, since the node names of the cluster are not present in the public domain @@ -157,13 +166,17 @@ name to an IP address on the cluster: Suffix the nodename with **-mn** to get the management network (ethernet) address, and pass it to a lookup-tool like *host* in another SSH session: - taurus$ host taurusi6605-mn<br />taurusi6605-mn.taurus.hrsk.tu-dresden.de has address 172.24.140.229 +```Bash +taurus$ host taurusi6605-mn<br />taurusi6605-mn.taurus.hrsk.tu-dresden.de has address 172.24.140.229 +```Bash The SSH tunnel has to be created from the user's localhost. The following example will create a forward SSH tunnel to localhost on port 22222 (or what ever port is prefered): - localhost$ ssh -L 22222:10.10.32.228:11111 userlogin@cara.dlr.de +```Bash +localhost$ ssh -L 22222:10.10.32.228:11111 userlogin@cara.dlr.de +``` The final step is to start ParaView locally on your own machine and add the connection @@ -185,24 +198,23 @@ Pipeline Browser (instead of it saying builtin). You now are connected to the pvserver running on a Taurus node and can open files from the cluster's filesystems. -##### Caveats +#### Caveats -Connecting to the compute nodes will only work when you are **inside the -TUD campus network**, because otherwise, the private networks 172.24.\* -will not be routed. That's why you either need to use -[VPN](https://tu-dresden.de/zih/dienste/service-katalog/arbeitsumgebung/zugang_datennetz/vpn), -or, when coming via the ZIH login gateway (`login1.zih.tu-dresden.de`), -use an SSH tunnel. For the example IP address from above, this could -look like the following: +Connecting to the compute nodes will only work when you are **inside the TUD campus network**, +because otherwise, the private networks 172.24.\* will not be routed. That's why you either need to +use [VPN](https://tu-dresden.de/zih/dienste/service-katalog/arbeitsumgebung/zugang_datennetz/vpn), +or, when coming via the ZIH login gateway (`login1.zih.tu-dresden.de`), use an SSH tunnel. For the +example IP address from above, this could look like the following: - # Replace "user" with your login name, of course: - ssh -f -N -L11111:172.24.140.229:11111 user@login1.zih.tu-dresden.de +```Bash +# Replace "user" with your login name, of course: +ssh -f -N -L11111:172.24.140.229:11111 user@login1.zih.tu-dresden.de +``` -This line opens the port 11111 locally and tunnels it via `login1` to -the `pvserver` running on the Taurus node. Note that you then must -instruct your local ParaView client to connect to host `localhost` -instead. The recommendation, though, is to use VPN, which makes this -extra step unnecessary. +This line opens the port 11111 locally and tunnels it via `login1` to the `pvserver` running on the +Taurus node. Note that you then must instruct your local ParaView client to connect to host +`localhost` instead. The recommendation, though, is to use VPN, which makes this extra step +unnecessary. #### Using the GUI via X forwarding (not recommended)