Skip to content
This repository was archived by the owner on Feb 16, 2024. It is now read-only.

Commit 23e111a

Browse files
authored
Document supported namespaces of stacks and demos (#274)
1 parent fa2c3d9 commit 23e111a

File tree

2 files changed

+24
-0
lines changed

2 files changed

+24
-0
lines changed

demos/demos-v2.yaml

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,7 @@ demos:
1111
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/airflow-scheduled-job/02-airflow-spark-clusterrolebinding.yaml
1212
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml
1313
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml
14+
supportedNamespaces: []
1415
hbase-hdfs-load-cycling-data:
1516
description: Copy data from S3 bucket to an HBase table
1617
stackableStack: hdfs-hbase
@@ -21,6 +22,7 @@ demos:
2122
manifests:
2223
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/hbase-hdfs-load-cycling-data/distcp-cycling-data.yaml
2324
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml
25+
supportedNamespaces: []
2426
nifi-kafka-druid-earthquake-data:
2527
description: Demo ingesting earthquake data into Kafka using NiFi, streaming it into Druid and creating a Superset dashboard
2628
documentation: https://docs.stackable.tech/stackablectl/stable/demos/nifi-kafka-druid-earthquake-data.html
@@ -37,6 +39,7 @@ demos:
3739
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml
3840
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml
3941
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml
42+
supportedNamespaces: ["default"]
4043
nifi-kafka-druid-water-level-data:
4144
description: Demo ingesting water level data into Kafka using NiFi, streaming it into Druid and creating a Superset dashboard
4245
documentation: https://docs.stackable.tech/stackablectl/stable/demos/nifi-kafka-druid-water-level-data.html
@@ -53,6 +56,7 @@ demos:
5356
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml
5457
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml
5558
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml
59+
supportedNamespaces: ["default"]
5660
spark-k8s-anomaly-detection-taxi-data:
5761
description: Demo loading New York taxi data into an S3 bucket and carrying out an anomaly detection analysis on it
5862
documentation: https://docs.stackable.tech/stackablectl/stable/demos/spark-k8s-anomaly-detection-taxi-data.html
@@ -68,6 +72,7 @@ demos:
6872
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/spark-k8s-anomaly-detection-taxi-data/load-test-data.yaml
6973
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml
7074
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml
75+
supportedNamespaces: []
7176
trino-iceberg:
7277
description: Demo containing Trino using Apache Iceberg as a S3 data lakehouse
7378
documentation: https://docs.stackable.tech/stackablectl/stable/demos/trino-iceberg.html
@@ -78,6 +83,7 @@ demos:
7883
- minio
7984
- s3
8085
manifests: []
86+
supportedNamespaces: []
8187
trino-taxi-data:
8288
description: Demo loading 2.5 years of New York taxi data into S3 bucket, creating a Trino table and a Superset dashboard
8389
documentation: https://docs.stackable.tech/stackablectl/stable/demos/trino-taxi-data.html
@@ -92,6 +98,7 @@ demos:
9298
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/trino-taxi-data/load-test-data.yaml
9399
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/trino-taxi-data/create-table-in-trino.yaml
94100
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/trino-taxi-data/setup-superset.yaml
101+
supportedNamespaces: []
95102
data-lakehouse-iceberg-trino-spark:
96103
description: Data lakehouse using Iceberg lakehouse on S3, Trino as query engine, Spark for streaming ingest and Superset for data visualization. Multiple datasources like taxi data, water levels in Germany, earthquakes, e-charging stations and more are loaded.
97104
documentation: https://docs.stackable.tech/stackablectl/stable/demos/data-lakehouse-iceberg-trino-spark.html
@@ -115,6 +122,7 @@ demos:
115122
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml
116123
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml
117124
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml
125+
supportedNamespaces: ["default"]
118126
jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data:
119127
description: Jupyterhub with PySpark and HDFS integration
120128
documentation: https://docs.stackable.tech/stackablectl/stable/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data.html
@@ -126,6 +134,7 @@ demos:
126134
- ny-taxi-data
127135
manifests:
128136
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml
137+
supportedNamespaces: []
129138
logging:
130139
description: Demo showing the logging stack in action
131140
documentation: https://docs.stackable.tech/stackablectl/stable/demos/logging.html
@@ -138,3 +147,4 @@ demos:
138147
- zookeeper
139148
manifests:
140149
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/logging/zookeeper.yaml
150+
supportedNamespaces: []

stacks/stacks-v2.yaml

Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -12,6 +12,7 @@ stacks:
1212
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/monitoring/grafana-dashboards.yaml
1313
- helmChart: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/_templates/prometheus.yaml
1414
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/_templates/prometheus-service-monitor.yaml
15+
supportedNamespaces: []
1516
parameters:
1617
- name: grafanaAdminPassword
1718
description: Password of the Grafana admin user
@@ -33,6 +34,7 @@ stacks:
3334
- helmChart: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/_templates/opensearch-dashboards.yaml
3435
- helmChart: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/_templates/vector-aggregator.yaml
3536
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/_templates/vector-aggregator-discovery.yaml
37+
supportedNamespaces: []
3638
parameters:
3739
- name: openSearchAdminPassword
3840
description: Password of the OpenSearch admin user
@@ -57,6 +59,7 @@ stacks:
5759
- helmChart: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/_templates/postgresql-airflow.yaml
5860
- helmChart: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/_templates/redis-airflow.yaml
5961
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/airflow/airflow.yaml
62+
supportedNamespaces: []
6063
parameters:
6164
- name: airflowAdminPassword
6265
description: Password of the Airflow admin user
@@ -99,6 +102,7 @@ stacks:
99102
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/data-lakehouse-iceberg-trino-spark/kafka.yaml
100103
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/data-lakehouse-iceberg-trino-spark/nifi.yaml
101104
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/nifi-kafka-druid-superset-s3/superset.yaml # Reuse
105+
supportedNamespaces: []
102106
parameters:
103107
- name: trinoAdminPassword
104108
description: Password of the Trino admin user
@@ -131,6 +135,7 @@ stacks:
131135
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/hdfs-hbase/zookeeper.yaml
132136
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/hdfs-hbase/hdfs.yaml
133137
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/hdfs-hbase/hbase.yaml
138+
supportedNamespaces: []
134139
parameters: []
135140
nifi-kafka-druid-superset-s3:
136141
description: Stack containing NiFi, Kafka, Druid, MinIO and Superset for data visualization
@@ -159,6 +164,7 @@ stacks:
159164
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/nifi-kafka-druid-superset-s3/druid.yaml
160165
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/nifi-kafka-druid-superset-s3/superset.yaml
161166
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/nifi-kafka-druid-superset-s3/nifi.yaml
167+
supportedNamespaces: []
162168
parameters:
163169
- name: nifiAdminPassword
164170
description: Password of the NiFI admin user
@@ -198,6 +204,7 @@ stacks:
198204
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/trino-superset-s3/trino.yaml
199205
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/spark-trino-superset-s3/trino-prediction-catalog.yaml
200206
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/trino-superset-s3/superset.yaml
207+
supportedNamespaces: []
201208
parameters:
202209
- name: trinoAdminPassword
203210
description: Password of the Trino admin user
@@ -234,6 +241,7 @@ stacks:
234241
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/trino-superset-s3/hive-metastore.yaml
235242
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/trino-superset-s3/trino.yaml
236243
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/trino-superset-s3/superset.yaml
244+
supportedNamespaces: []
237245
parameters:
238246
- name: trinoAdminPassword
239247
description: Password of the Trino admin user
@@ -268,6 +276,7 @@ stacks:
268276
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/trino-iceberg/s3-connection.yaml
269277
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/trino-iceberg/hive-metastores.yaml
270278
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/trino-iceberg/trino.yaml
279+
supportedNamespaces: []
271280
parameters:
272281
- name: trinoAdminPassword
273282
description: Password of the Trino admin user
@@ -293,7 +302,9 @@ stacks:
293302
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/jupyterhub-pyspark-hdfs/zookeeper.yaml
294303
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/jupyterhub-pyspark-hdfs/hdfs.yaml
295304
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/jupyterhub-pyspark-hdfs/serviceaccount.yaml
305+
# TODO Use patched JHub that created service for us from customer setup (ask Sebastian)
296306
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/jupyterhub-pyspark-hdfs/spark_driver_service.yaml
307+
supportedNamespaces: []
297308
parameters:
298309
- name: jupyterHubAdminPassword
299310
description: Password of the JupyterHub admin user
@@ -321,6 +332,7 @@ stacks:
321332
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/dual-hive-hdfs-s3/hdfs.yaml
322333
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/dual-hive-hdfs-s3/hive.yaml
323334
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/dual-hive-hdfs-s3/trino.yaml
335+
supportedNamespaces: []
324336
parameters:
325337
- name: minioAdminPassword
326338
description: Password of the Minio admin user
@@ -340,6 +352,7 @@ stacks:
340352
- ldap
341353
manifests:
342354
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/authentication/openldap-tls.yaml
355+
supportedNamespaces: ["default"]
343356
# TODO: Parameterize
344357
parameters: []
345358
openldap:
@@ -358,5 +371,6 @@ stacks:
358371
manifests:
359372
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/authentication/openldap-tls.yaml
360373
- plainYaml: https://raw.githubusercontent.com/stackabletech/stackablectl/main/stacks/authentication/openldap-tls-authenticationclass.yaml
374+
supportedNamespaces: []
361375
# TODO: Parameterize
362376
parameters: []

0 commit comments

Comments
 (0)