From a2e7a94148b88edbbb308d05f4b9743b58ae85a8 Mon Sep 17 00:00:00 2001 From: Taras Lazariv <taras.lazariv@tu-dresden.de> Date: Thu, 11 Nov 2021 16:16:08 +0100 Subject: [PATCH] Apply 1 suggestion(s) to 1 file(s) --- .../software/big_data_frameworks_spark.md | 34 +++++++++++++++++-- 1 file changed, 32 insertions(+), 2 deletions(-) diff --git a/doc.zih.tu-dresden.de/docs/software/big_data_frameworks_spark.md b/doc.zih.tu-dresden.de/docs/software/big_data_frameworks_spark.md index 8e8a87f47..2f087b8a9 100644 --- a/doc.zih.tu-dresden.de/docs/software/big_data_frameworks_spark.md +++ b/doc.zih.tu-dresden.de/docs/software/big_data_frameworks_spark.md @@ -119,8 +119,38 @@ short test runs, it is **recommended to launch your jobs in the background using that, you can conveniently put the parameters directly into the job file and submit it via `sbatch [options] <job file>`. -Please use a [batch job](../jobs_and_resources/slurm.md) similar to -[example-spark.sbatch](misc/example-spark.sbatch). +Please use a [batch job](../jobs_and_resources/slurm.md) with a configuration, similar to the example below: + +??? example "spark.sbatch" + ```bash + #!/bin/bash -l + #SBATCH --time=01:00:00 + #SBATCH --partition=haswell + #SBATCH --nodes=2 + #SBATCH --exclusive + #SBATCH --mem=60G + #SBATCH --job-name="example-spark" + + ml Spark/3.0.1-Hadoop-2.7-Java-1.8-Python-3.7.4-GCCcore-8.3.0 + + function myExitHandler () { + stop-all.sh + } + + #configuration + . framework-configure.sh spark $SPARK_HOME/conf + + #register cleanup hook in case something goes wrong + trap myExitHandler EXIT + + start-all.sh + + spark-submit --class org.apache.spark.examples.SparkPi $SPARK_HOME/examples/jars/spark-examples_2.12-3.0.1.jar 1000 + + stop-all.sh + + exit 0 + ``` ## Jupyter Notebook -- GitLab