From 34cde20969f1a5a829fbb8407fb9aee591f8147e Mon Sep 17 00:00:00 2001
From: Jan Frenzel <jan.frenzel@tu-dresden.de>
Date: Thu, 11 Nov 2021 16:33:36 +0100
Subject: [PATCH] Removed example-spark.sbatch as it is included in
 big_data_frameworks_spark.md.

---
 .../software/big_data_frameworks_spark.md     |  5 ++--
 .../docs/software/misc/example-spark.sbatch   | 27 -------------------
 2 files changed, 3 insertions(+), 29 deletions(-)
 delete mode 100644 doc.zih.tu-dresden.de/docs/software/misc/example-spark.sbatch

diff --git a/doc.zih.tu-dresden.de/docs/software/big_data_frameworks_spark.md b/doc.zih.tu-dresden.de/docs/software/big_data_frameworks_spark.md
index 2f087b8a9..3ef9d6f30 100644
--- a/doc.zih.tu-dresden.de/docs/software/big_data_frameworks_spark.md
+++ b/doc.zih.tu-dresden.de/docs/software/big_data_frameworks_spark.md
@@ -119,12 +119,13 @@ short test runs, it is **recommended to launch your jobs in the background using
 that, you can conveniently put the parameters directly into the job file and submit it via
 `sbatch [options] <job file>`.
 
-Please use a [batch job](../jobs_and_resources/slurm.md) with a configuration, similar to the example below:
+Please use a [batch job](../jobs_and_resources/slurm.md) with a configuration, similar to the
+example below:
 
 ??? example "spark.sbatch"
     ```bash
     #!/bin/bash -l
-    #SBATCH --time=01:00:00
+    #SBATCH --time=00:05:00
     #SBATCH --partition=haswell
     #SBATCH --nodes=2
     #SBATCH --exclusive
diff --git a/doc.zih.tu-dresden.de/docs/software/misc/example-spark.sbatch b/doc.zih.tu-dresden.de/docs/software/misc/example-spark.sbatch
deleted file mode 100644
index 2fcf3aa39..000000000
--- a/doc.zih.tu-dresden.de/docs/software/misc/example-spark.sbatch
+++ /dev/null
@@ -1,27 +0,0 @@
-#!/bin/bash
-#SBATCH --time=00:03:00
-#SBATCH --partition=haswell
-#SBATCH --nodes=1
-#SBATCH --exclusive
-#SBATCH --mem=50G
-#SBATCH -J "example-spark"
-
-ml Spark/3.0.1-Hadoop-2.7-Java-1.8-Python-3.7.4-GCCcore-8.3.0
-
-function myExitHandler () {
-	stop-all.sh
-}
-
-#configuration
-. framework-configure.sh spark $SPARK_HOME/conf
-
-#register cleanup hook in case something goes wrong
-trap myExitHandler EXIT
-
-start-all.sh
-
-spark-submit --class org.apache.spark.examples.SparkPi $SPARK_HOME/examples/jars/spark-examples_2.12-3.0.1.jar 1000
-
-stop-all.sh
-
-exit 0
-- 
GitLab