From 34cde20969f1a5a829fbb8407fb9aee591f8147e Mon Sep 17 00:00:00 2001 From: Jan Frenzel <jan.frenzel@tu-dresden.de> Date: Thu, 11 Nov 2021 16:33:36 +0100 Subject: [PATCH] Removed example-spark.sbatch as it is included in big_data_frameworks_spark.md. --- .../software/big_data_frameworks_spark.md | 5 ++-- .../docs/software/misc/example-spark.sbatch | 27 ------------------- 2 files changed, 3 insertions(+), 29 deletions(-) delete mode 100644 doc.zih.tu-dresden.de/docs/software/misc/example-spark.sbatch diff --git a/doc.zih.tu-dresden.de/docs/software/big_data_frameworks_spark.md b/doc.zih.tu-dresden.de/docs/software/big_data_frameworks_spark.md index 2f087b8a9..3ef9d6f30 100644 --- a/doc.zih.tu-dresden.de/docs/software/big_data_frameworks_spark.md +++ b/doc.zih.tu-dresden.de/docs/software/big_data_frameworks_spark.md @@ -119,12 +119,13 @@ short test runs, it is **recommended to launch your jobs in the background using that, you can conveniently put the parameters directly into the job file and submit it via `sbatch [options] <job file>`. -Please use a [batch job](../jobs_and_resources/slurm.md) with a configuration, similar to the example below: +Please use a [batch job](../jobs_and_resources/slurm.md) with a configuration, similar to the +example below: ??? example "spark.sbatch" ```bash #!/bin/bash -l - #SBATCH --time=01:00:00 + #SBATCH --time=00:05:00 #SBATCH --partition=haswell #SBATCH --nodes=2 #SBATCH --exclusive diff --git a/doc.zih.tu-dresden.de/docs/software/misc/example-spark.sbatch b/doc.zih.tu-dresden.de/docs/software/misc/example-spark.sbatch deleted file mode 100644 index 2fcf3aa39..000000000 --- a/doc.zih.tu-dresden.de/docs/software/misc/example-spark.sbatch +++ /dev/null @@ -1,27 +0,0 @@ -#!/bin/bash -#SBATCH --time=00:03:00 -#SBATCH --partition=haswell -#SBATCH --nodes=1 -#SBATCH --exclusive -#SBATCH --mem=50G -#SBATCH -J "example-spark" - -ml Spark/3.0.1-Hadoop-2.7-Java-1.8-Python-3.7.4-GCCcore-8.3.0 - -function myExitHandler () { - stop-all.sh -} - -#configuration -. framework-configure.sh spark $SPARK_HOME/conf - -#register cleanup hook in case something goes wrong -trap myExitHandler EXIT - -start-all.sh - -spark-submit --class org.apache.spark.examples.SparkPi $SPARK_HOME/examples/jars/spark-examples_2.12-3.0.1.jar 1000 - -stop-all.sh - -exit 0 -- GitLab