diff options
Diffstat (limited to 'vnfs/DAaaS/sample-apps/training/sample-spark-app')
-rw-r--r-- | vnfs/DAaaS/sample-apps/training/sample-spark-app/templates/SampleSparkApp.yaml | 4 | ||||
-rw-r--r-- | vnfs/DAaaS/sample-apps/training/sample-spark-app/values.yaml | 27 |
2 files changed, 8 insertions, 23 deletions
diff --git a/vnfs/DAaaS/sample-apps/training/sample-spark-app/templates/SampleSparkApp.yaml b/vnfs/DAaaS/sample-apps/training/sample-spark-app/templates/SampleSparkApp.yaml index f728f82e..1b12e479 100644 --- a/vnfs/DAaaS/sample-apps/training/sample-spark-app/templates/SampleSparkApp.yaml +++ b/vnfs/DAaaS/sample-apps/training/sample-spark-app/templates/SampleSparkApp.yaml @@ -2,7 +2,7 @@ apiVersion: "sparkoperator.k8s.io/v1beta1" kind: SparkApplication metadata: name: {{ .Values.nameOfTheSparkApp }} - namespace: {{ .Release.Namespace }} + namespace: {{ .Release.Namespace }} spec: type: {{ .Values.programmingLanguageType }} mode: {{ .Values.modeOfSparkApp | default "cluster" }} @@ -10,7 +10,7 @@ spec: imagePullPolicy: {{ .Values.imagePullPolicy | default "IfNotPresent" }} mainClass: {{ .Values.mainClassOfTheSparkApp }} mainApplicationFile: {{ .Values.mainApplicationFileOfTheSparkApp }} - arguments: + arguments: {{- range .Values.argumentsOfTheSparkProgram }} - {{ . }} {{ end }} diff --git a/vnfs/DAaaS/sample-apps/training/sample-spark-app/values.yaml b/vnfs/DAaaS/sample-apps/training/sample-spark-app/values.yaml index afb48d67..07457ac8 100644 --- a/vnfs/DAaaS/sample-apps/training/sample-spark-app/values.yaml +++ b/vnfs/DAaaS/sample-apps/training/sample-spark-app/values.yaml @@ -2,7 +2,6 @@ # This is a YAML-formatted file. # Declare variables to be passed into your templates. - #===========================KUBERNETES POD RELATED CONFIGs======================== image: spark-tf-keras-horo:latest imagePullPolicy: Never @@ -11,47 +10,33 @@ volumesName: test-volume hostpath: /tmp hostpathType: Directory - - #============================SPARK APP RELATED CONFIGs============================= - -nameOfTheSparkApp: spark-apache-logs2 +nameOfTheSparkApp: spark-apache-logs2 # Python or Scala supported. programmingLanguageType: Scala modeOfSparkApp: cluster mainClassOfTheSparkApp: ApacheLogAnalysis # can be http path, s3 path, minio path -mainApplicationFileOfTheSparkApp: https://github.com/mohanraj1311/ApacheLogAnalysisJar/raw/master/analysisofapachelogs_2.11-0.1.jar +mainApplicationFileOfTheSparkApp: https://github.com/mohanraj1311/ApacheLogAnalysisJar/raw/master/analysisofapachelogs_2.11-0.1.jar argumentsOfTheSparkProgram: - - hdfs://hdfs-1-namenode-1.hdfs-1-namenode.hdfs1.svc.cluster.local:8020/data/apache-logs - - + - hdfs://hdfs-1-namenode-1.hdfs-1-namenode.hdfs1.svc.cluster.local:8020/data/apache-logs #============================SPARK DRIVER RELATED CONFIGs========================= driverCores: 0.1 driverCoreLimit: 200m driverMemory: 1024m driverVolumeMountsName: test-volume -driverVolumeMountPath: /tmp - - +driverVolumeMountPath: /tmp #============================SPARK EXECUTOR RELATED CONFIGs======================= -executorCores: 1 -executorInstances: 1 +executorCores: 1 +executorInstances: 1 executorMemory: 512m executorVolumeMountsName: test-volume executorVolumeMountPath: /tmp - - #===========================HADOOP RELATED CONFIGs=============================== # config map of the hdfs hadoopConfigMap: hdfs-1-config - ################################################################################### - - - - |