Skip to content

Commit a7d34d2

Browse files
NickLarsenNZxeniaperazvansiegfriedweber
authored
feat: Update demos for Release 24.11 (#116)
* chore(opa): bump to 0.67.0 Depends on: stackabletech/docker-images#797 * chore(opa): bump 0.67.1 * chore(opa): bump to 0.67.0 Depends on: stackabletech/docker-images#797 * chore: Bump Vector to 0.40.0 * chore(airflow): bump version to 2.9.3 * chore(trino): bump version to 455 * chore(spark): version 3.5.2 * Update create-spark-ingestion-job.yaml * chore: bump hdfs version to 3.4.0 * chore: bump nifi for 24.11 * chore: Upgrade the Vector aggregator * chore(druid): Bump to 30.0.0 * chore(pre-commit): Ignore braces lint for templates yaml files * chore: lint * Apply suggestions from code review Use images that exist. We can't use hdfs from 24.7 because there is a tool needed. For consistency, we use the same image for consistency * Apply suggestions from code review * chore: fix lint issue * Apply suggestions from code review * chore: fix lint issue * chore: disable the brace lint --------- Co-authored-by: xeniape <xenia.fischer@stackable.tech> Co-authored-by: Razvan-Daniel Mihai <84674+razvan@users.noreply.github.com> Co-authored-by: Siegfried Weber <mail@siegfriedweber.net>
1 parent f8ef88d commit a7d34d2

File tree

31 files changed

+67
-60
lines changed

31 files changed

+67
-60
lines changed

.yamllint.yaml

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,4 @@ rules:
77
check-keys: false
88
comments:
99
min-spaces-from-content: 1 # Needed due to https://github.com/adrienverge/yamllint/issues/443
10-
braces:
11-
ignore: |
12-
stacks/signal-processing/*
13-
stacks/_templates/*
10+
braces: disable # because the yaml files are templates which can have {{ ... }}

demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -43,13 +43,13 @@ data:
4343
stackable.tech/vendor: Stackable
4444
spec:
4545
sparkImage:
46-
productVersion: 3.5.1
46+
productVersion: 3.5.2
4747
mode: cluster
4848
mainApplicationFile: local:///stackable/spark/jobs/spark-ingest-into-lakehouse.py
4949
deps:
5050
packages:
51-
- org.apache.iceberg:iceberg-spark-runtime-3.5_2.12:1.5.0
52-
- org.apache.spark:spark-sql-kafka-0-10_2.12:3.5.1
51+
- org.apache.iceberg:iceberg-spark-runtime-3.5_2.12:1.6.1
52+
- org.apache.spark:spark-sql-kafka-0-10_2.12:3.5.2
5353
s3connection:
5454
reference: minio
5555
sparkConf:

demos/end-to-end-security/create-spark-report.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -55,7 +55,7 @@ data:
5555
name: spark-report
5656
spec:
5757
sparkImage:
58-
productVersion: 3.5.1
58+
productVersion: 3.5.2
5959
mode: cluster
6060
mainApplicationFile: local:///stackable/spark/jobs/spark-report.py
6161
deps:

demos/hbase-hdfs-load-cycling-data/distcp-cycling-data.yaml

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,14 +11,15 @@ spec:
1111
# We use 24.3.0 here which contains the distcp MapReduce components
1212
# This is not included in the 24.7 images and will fail.
1313
# See: https://github.com/stackabletech/docker-images/issues/793
14-
image: docker.stackable.tech/stackable/hadoop:3.3.4-stackable24.3.0
14+
image: docker.stackable.tech/stackable/hadoop:3.4.0-stackable0.0.0-dev
1515
env:
1616
- name: HADOOP_USER_NAME
1717
value: stackable
1818
- name: HADOOP_CONF_DIR
1919
value: "/stackable/conf/hdfs"
2020
- name: HADOOP_CLASSPATH
2121
value: "/stackable/hadoop/share/hadoop/tools/lib/*.jar"
22+
# yamllint disable-line rule:line-length
2223
command: ["bash", "-c", "bin/hdfs dfs -mkdir -p /data/raw && bin/hadoop distcp -D fs.s3a.aws.credentials.provider=org.apache.hadoop.fs.s3a.AnonymousAWSCredentialsProvider s3a://public-backup-nyc-tlc/cycling-tripdata/demo-cycling-tripdata.csv.gz hdfs://hdfs/data/raw"]
2324
volumeMounts:
2425
- name: config-volume-hdfs

demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,13 +8,15 @@ spec:
88
spec:
99
containers:
1010
- name: load-ny-taxi-data
11-
image: docker.stackable.tech/stackable/hadoop:3.3.4-stackable24.7.0
11+
image: docker.stackable.tech/stackable/hadoop:3.4.0-stackable0.0.0-dev
12+
# yamllint disable rule:line-length
1213
command: ["bash", "-c", "/stackable/hadoop/bin/hdfs dfs -mkdir -p /ny-taxi-data/raw \
1314
&& cd /tmp \
1415
&& for month in 2020-09; do \
1516
curl -O https://repo.stackable.tech/repository/misc/ny-taxi-data/fhvhv_tripdata_$month.parquet \
1617
&& /stackable/hadoop/bin/hdfs dfs -put fhvhv_tripdata_$month.parquet /ny-taxi-data/raw/; \
1718
done"]
19+
# yamllint enable rule:line-length
1820
volumeMounts:
1921
- name: hdfs-discovery-configmap
2022
mountPath: /hdfs

demos/signal-processing/DownloadAndWriteToDB.xml

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -61,7 +61,7 @@
6161
<bundle>
6262
<artifact>nifi-dbcp-service-nar</artifact>
6363
<group>org.apache.nifi</group>
64-
<version>1.21.0</version>
64+
<version>1.27.0</version>
6565
</bundle>
6666
<comments></comments>
6767
<descriptors>
@@ -258,7 +258,7 @@
258258
<bundle>
259259
<artifact>nifi-record-serialization-services-nar</artifact>
260260
<group>org.apache.nifi</group>
261-
<version>1.21.0</version>
261+
<version>1.27.0</version>
262262
</bundle>
263263
<comments></comments>
264264
<descriptors>
@@ -561,7 +561,7 @@
561561
</position>
562562
<height>88.0</height>
563563
<label>This flow downloads a dataset, writing it to a temporary table in TimescaleDB.
564-
This data is then written to the target table with the time offsets preserved,
564+
This data is then written to the target table with the time offsets preserved,
565565
but re-based to the current time. This means that the data can be displayed
566566
in Grafana as if it were being streamed, whereas in fact the dashboard moves
567567
through "future" data that has already been persisted.</label>
@@ -584,7 +584,7 @@ through "future" data that has already been persisted.</label>
584584
<bundle>
585585
<artifact>nifi-standard-nar</artifact>
586586
<group>org.apache.nifi</group>
587-
<version>1.21.0</version>
587+
<version>1.27.0</version>
588588
</bundle>
589589
<config>
590590
<backoffMechanism>PENALIZE_FLOWFILE</backoffMechanism>
@@ -1069,7 +1069,7 @@ through "future" data that has already been persisted.</label>
10691069
<bundle>
10701070
<artifact>nifi-standard-nar</artifact>
10711071
<group>org.apache.nifi</group>
1072-
<version>1.21.0</version>
1072+
<version>1.27.0</version>
10731073
</bundle>
10741074
<config>
10751075
<backoffMechanism>PENALIZE_FLOWFILE</backoffMechanism>
@@ -1223,7 +1223,7 @@ from conditions_temp;</value>
12231223
<bundle>
12241224
<artifact>nifi-standard-nar</artifact>
12251225
<group>org.apache.nifi</group>
1226-
<version>1.21.0</version>
1226+
<version>1.27.0</version>
12271227
</bundle>
12281228
<config>
12291229
<backoffMechanism>PENALIZE_FLOWFILE</backoffMechanism>

demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ data:
3737
name: spark-ad
3838
spec:
3939
sparkImage:
40-
productVersion: 3.5.1
40+
productVersion: 3.5.2
4141
mode: cluster
4242
mainApplicationFile: local:///spark-scripts/spark-ad.py
4343
deps:

demos/spark-k8s-anomaly-detection-taxi-data/load-test-data.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,7 @@ spec:
99
containers:
1010
- name: load-ny-taxi-data
1111
image: "bitnami/minio:2022-debian-10"
12+
# yamllint disable-line rule:line-length
1213
command: ["bash", "-c", "cd /tmp && for month in 2020-09 2020-10 2020-11 2020-12; do curl -O https://repo.stackable.tech/repository/misc/ny-taxi-data/fhvhv_tripdata_$month.parquet && mc --insecure alias set minio http://minio:9000/ $(cat /minio-s3-credentials/accessKey) $(cat /minio-s3-credentials/secretKey) && mc cp fhvhv_tripdata_$month.parquet minio/demo/ny-taxi-data/raw/ && mc mb --ignore-existing minio/prediction; done"]
1314
volumeMounts:
1415
- name: minio-s3-credentials

demos/trino-taxi-data/load-test-data.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,7 @@ spec:
99
containers:
1010
- name: load-ny-taxi-data
1111
image: "bitnami/minio:2024-debian-12"
12+
# yamllint disable-line rule:line-length
1213
command: ["bash", "-c", "cd /tmp && for month in 2020-01 2020-02 2020-03 2020-04 2020-05 2020-06 2020-07 2020-08 2020-09 2020-10 2020-11 2020-12 2021-01 2021-02 2021-03 2021-04 2021-05 2021-06 2021-07 2021-08 2021-09 2021-10 2021-11 2021-12 2022-01 2022-02 2022-03 2022-04; do curl -O https://repo.stackable.tech/repository/misc/ny-taxi-data/yellow_tripdata_$month.parquet && mc --insecure alias set minio http://minio:9000/ $(cat /minio-s3-credentials/accessKey) $(cat /minio-s3-credentials/secretKey) && mc cp yellow_tripdata_$month.parquet minio/demo/ny-taxi-data/raw/; done"]
1314
volumeMounts:
1415
- name: minio-s3-credentials

docs/modules/demos/pages/trino-iceberg.adoc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -78,7 +78,7 @@ As an alternative, you can use https://trino.io/download.html[trino-cli] by runn
7878

7979
[source,console]
8080
----
81-
$ java -jar ~/Downloads/trino-cli-451-executable.jar --user admin --insecure --password --server https://172.18.0.2:30856
81+
$ java -jar ~/Downloads/trino-cli-455-executable.jar --user admin --insecure --password --server https://172.18.0.2:30856
8282
----
8383

8484
Make sure to replace the server endpoint with the endpoint listed in the `stackablectl stacklet list` output.

0 commit comments

Comments
 (0)