Skip to content
Snippets Groups Projects
Commit 17dde5ca authored by Jan Frenzel's avatar Jan Frenzel Committed by Loc Nguyen Dang Duc
Browse files

Apply 12 suggestion(s) to 1 file(s)

parent 7feb142f
No related branches found
No related tags found
3 merge requests!322Merge preview into main,!319Merge preview into main,!248review cfd.md
...@@ -33,20 +33,20 @@ Example for OpenFOAM job script: ...@@ -33,20 +33,20 @@ Example for OpenFOAM job script:
#SBATCH --time=12:00:00 # walltime #SBATCH --time=12:00:00 # walltime
#SBATCH --ntasks=60 # number of processor cores (i.e. tasks) #SBATCH --ntasks=60 # number of processor cores (i.e. tasks)
#SBATCH --mem-per-cpu=500M # memory per CPU core #SBATCH --mem-per-cpu=500M # memory per CPU core
#SBATCH -J "Test" # job name #SBATCH -job-name="Test" # job name
#SBATCH --mail-user=mustermann@tu-dresden.de # email address (only tu-dresden) #SBATCH --mail-user=marie@tu-dresden.de # email address (only tu-dresden)
#SBATCH --mail-type=ALL #SBATCH --mail-type=ALL
OUTFILE="Output" OUTFILE="Output"
module load OpenFOAM module load OpenFOAM
source $FOAM_BASH source $FOAM_BASH
cd /scratch/<YOURUSERNAME> # work directory in /scratch...! cd /scratch/ws/1/marie-example-workspace # work directory using workspace
srun pimpleFoam -parallel > "$OUTFILE" srun pimpleFoam -parallel > "$OUTFILE"
``` ```
## Ansys CFX ## Ansys CFX
Ansys CFX is a powerful finite-volume-based program package for modelling general fluid flow in Ansys CFX is a powerful finite-volume-based program package for modeling general fluid flow in
complex geometries. The main components of the CFX package are the flow solver cfx5solve, the complex geometries. The main components of the CFX package are the flow solver cfx5solve, the
geometry and mesh generator cfx5pre, and the post-processor cfx5post. geometry and mesh generator cfx5pre, and the post-processor cfx5post.
...@@ -57,24 +57,24 @@ Example for CFX job script: ...@@ -57,24 +57,24 @@ Example for CFX job script:
#SBATCH --time=12:00 # walltime #SBATCH --time=12:00 # walltime
#SBATCH --ntasks=4 # number of processor cores (i.e. tasks) #SBATCH --ntasks=4 # number of processor cores (i.e. tasks)
#SBATCH --mem-per-cpu=1900M # memory per CPU core #SBATCH --mem-per-cpu=1900M # memory per CPU core
#SBATCH --mail-user=.......@tu-dresden.de # email address (only tu-dresden) #SBATCH --mail-user=marie@tu-dresden.de # email address (only tu-dresden)
#SBATCH --mail-type=ALL #SBATCH --mail-type=ALL
module load ANSYS module load ANSYS
cd /scratch/<YOURUSERNAME> # work directory in /scratch...! cd /scratch/ws/1/marie-example-workspace # work directory using workspace
cfx-parallel.sh -double -def StaticMixer.def cfx-parallel.sh -double -def StaticMixer.def
``` ```
## Ansys Fluent ## Ansys Fluent
Fluent need the hostnames and can be run in parallel like this: Fluent needs the host names and can be run in parallel like this:
```bash ```bash
#!/bin/bash #!/bin/bash
#SBATCH --time=12:00 # walltime #SBATCH --time=12:00 # walltime
#SBATCH --ntasks=4 # number of processor cores (i.e. tasks) #SBATCH --ntasks=4 # number of processor cores (i.e. tasks)
#SBATCH --mem-per-cpu=1900M # memory per CPU core #SBATCH --mem-per-cpu=1900M # memory per CPU core
#SBATCH --mail-user=.......@tu-dresden.de # email address (only tu-dresden) #SBATCH --mail-user=marie@tu-dresden.de # email address (only tu-dresden)
#SBATCH --mail-type=ALL #SBATCH --mail-type=ALL
module load ANSYS module load ANSYS
...@@ -83,28 +83,28 @@ nodeset -e $SLURM_JOB_NODELIST | xargs -n1 > hostsfile_job_$SLURM_JOBID.txt ...@@ -83,28 +83,28 @@ nodeset -e $SLURM_JOB_NODELIST | xargs -n1 > hostsfile_job_$SLURM_JOBID.txt
fluent 2ddp -t$SLURM_NTASKS -g -mpi=intel -pinfiniband -cnf=hostsfile_job_$SLURM_JOBID.txt < input.in fluent 2ddp -t$SLURM_NTASKS -g -mpi=intel -pinfiniband -cnf=hostsfile_job_$SLURM_JOBID.txt < input.in
``` ```
To use fluent interactive, please try: To use fluent interactively, please try:
```console ```console
marie@login$ module load ANSYS/19.2 marie@login$ module load ANSYS/19.2
marie@login$ srun -N 1 --cpus-per-task=4 --time=1:00:00 --pty --x11=first bash marie@login$ srun --nodes=1 --cpus-per-task=4 --time=1:00:00 --pty --x11=first bash
marie@login$ fluent & marie@login$ fluent &
``` ```
## STAR-CCM+ ## STAR-CCM+
Note: you have to use your own license in order to run STAR-CCM+ on Taurus, so you have to specify Note: you have to use your own license in order to run STAR-CCM+ on ZIH systems, so you have to specify
the parameters `-licpath` and `-podkey`, see the example below. the parameters `-licpath` and `-podkey`, see the example below.
Our installation provides a script `create_rankfile -f CCM` that generates a hostlist from the SLURM Our installation provides a script `create_rankfile -f CCM` that generates a host list from the
job environment that can be passed to starccm+, enabling it to run across multiple nodes. Slurm job environment that can be passed to `starccm+`, enabling it to run across multiple nodes.
```bash ```bash
#!/bin/bash #!/bin/bash
#SBATCH --time=12:00 # walltime #SBATCH --time=12:00 # walltime
#SBATCH --ntasks=32 # number of processor cores (i.e. tasks) #SBATCH --ntasks=32 # number of processor cores (i.e. tasks)
#SBATCH --mem-per-cpu=2500M # memory per CPU core #SBATCH --mem-per-cpu=2500M # memory per CPU core
#SBATCH --mail-user=.......@tu-dresden.de # email address (only tu-dresden) #SBATCH --mail-user=marie@tu-dresden.de # email address (only tu-dresden)
#SBATCH --mail-type=ALL #SBATCH --mail-type=ALL
module load STAR-CCM+ module load STAR-CCM+
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment