Skip to content

Commit ab2b504

Browse files
committed
Merge branch '3-configure-node-selection' into handle-spark-app-termination
2 parents c7c18d5 + bb1f25e commit ab2b504

16 files changed

+10
-189
lines changed

.github/workflows/build.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -112,7 +112,7 @@ jobs:
112112
- uses: actions/checkout@2541b1294d2704b0964813337f33b291d3f8596b # tag=v3.0.2
113113
with:
114114
submodules: recursive
115-
- uses: EmbarkStudios/cargo-deny-action@3481b77dfd7b1d3c62bebdbfb57695131cd59c8f # tag=v1.2.17
115+
- uses: EmbarkStudios/cargo-deny-action@30ecad1d5873c1cc2cad10a33637371ca094768b # tag=v1.3.1
116116
with:
117117
command: check ${{ matrix.checks }}
118118

.pre-commit-config.yaml

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -33,8 +33,3 @@ repos:
3333
rev: 4.0.1
3434
hooks:
3535
- id: flake8
36-
37-
- repo: https://github.com/PyCQA/pylint
38-
rev: v2.13.9
39-
hooks:
40-
- id: pylint

Cargo.lock

Lines changed: 2 additions & 2 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

rust/crd/Cargo.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ edition = "2021"
55
license = "OSL-3.0"
66
name = "stackable-spark-k8s-crd"
77
repository = "https://github.com/stackabletech/spark-k8s-operator"
8-
version = "0.1.0"
8+
version = "0.2.0-nightly"
99

1010
[dependencies]
1111
stackable-operator = { git = "https://github.com/stackabletech/operator-rs.git", tag="0.19.0" }

rust/operator-binary/Cargo.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ edition = "2021"
55
license = "OSL-3.0"
66
name = "stackable-spark-k8s-operator"
77
repository = "https://github.com/stackabletech/spark-k8s-operator"
8-
version = "0.1.0"
8+
version = "0.2.0-nightly"
99

1010
[dependencies]
1111
stackable-operator = { git = "https://github.com/stackabletech/operator-rs.git", tag="0.19.0" }

tests/kuttl-test.yaml.jinja2

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -8,3 +8,4 @@ testDirs:
88

99
startKIND: false
1010
suppress: ["events"]
11+
parallel: 2

tests/templates/kuttl/node-selector/00-assert.yaml

Lines changed: 0 additions & 21 deletions
This file was deleted.

tests/templates/kuttl/node-selector/00-s3-secret.yaml

Lines changed: 0 additions & 9 deletions
This file was deleted.

tests/templates/kuttl/node-selector/00-s3-upload-container.yaml

Lines changed: 0 additions & 36 deletions
This file was deleted.

tests/templates/kuttl/node-selector/00-setup-minio.yaml

Lines changed: 0 additions & 15 deletions
This file was deleted.

tests/templates/kuttl/node-selector/01-prepare-bucket.yaml

Lines changed: 0 additions & 11 deletions
This file was deleted.

tests/templates/kuttl/node-selector/02-assert.yaml

Lines changed: 0 additions & 13 deletions
This file was deleted.

tests/templates/kuttl/node-selector/02-deps-volume.yaml

Lines changed: 0 additions & 48 deletions
This file was deleted.

tests/templates/kuttl/node-selector/10-assert.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ apiVersion: kuttl.dev/v1beta1
33
kind: TestAssert
44
metadata:
55
name: node-selector-assert
6-
timeout: 900
6+
timeout: 300
77
---
88
# The Job starting the whole process
99
apiVersion: spark.stackable.tech/v1alpha1

tests/templates/kuttl/node-selector/10-deploy-spark-app.yaml.j2

Lines changed: 3 additions & 25 deletions
Original file line numberDiff line numberDiff line change
@@ -6,41 +6,19 @@ metadata:
66
spec:
77
version: "1.0"
88
sparkImage: docker.stackable.tech/stackable/spark-k8s:{{ test_scenario['values']['spark'] }}-hadoop{{ test_scenario['values']['hadoop'][:-2] }}-stackable{{ test_scenario['values']['stackable'] }}
9+
sparkImagePullPolicy: IfNotPresent
910
mode: cluster
10-
mainClass: org.apache.spark.examples.SparkPi
11-
mainApplicationFile: s3a://my-bucket/spark-examples_2.12-{{ test_scenario['values']['spark'] }}.jar
12-
s3bucket:
13-
inline:
14-
bucketName: my-bucket
15-
connection:
16-
inline:
17-
host: test-minio
18-
port: 9000
19-
secretClass: minio-credentials
20-
volumes:
21-
- name: spark-pi-deps
22-
persistentVolumeClaim:
23-
claimName: spark-pi-pvc
24-
sparkConf:
25-
spark.hadoop.fs.s3a.aws.credentials.provider: "org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider"
26-
spark.hadoop.fs.s3a.path.style.access: "true"
27-
spark.driver.extraClassPath: "/dependencies/jars/hadoop-aws-3.2.0.jar:/dependencies/jars/aws-java-sdk-bundle-1.11.375.jar"
28-
spark.executor.extraClassPath: "/dependencies/jars/hadoop-aws-3.2.0.jar:/dependencies/jars/aws-java-sdk-bundle-1.11.375.jar"
11+
mainClass: org.apache.spark.examples.SparkALS
12+
mainApplicationFile: local:///stackable/spark/examples/jars/spark-examples_2.12-{{ test_scenario['values']['spark'] }}.jar
2913
driver:
3014
cores: 1
3115
coreLimit: "1200m"
3216
memory: "512m"
33-
volumeMounts:
34-
- name: spark-pi-deps
35-
mountPath: /dependencies
3617
nodeSelector:
3718
node: "1"
3819
executor:
3920
cores: 1
4021
instances: 1
4122
memory: "512m"
42-
volumeMounts:
43-
- name: spark-pi-deps
44-
mountPath: /dependencies
4523
nodeSelector:
4624
kubernetes.io/os: linux
Binary file not shown.

0 commit comments

Comments
 (0)