1
1
---
2
2
demos :
3
- trino-taxi-data :
4
- description : Demo loading 2.5 years of New York taxi data into S3 bucket, creating a Trino table and a Superset dashboard
5
- documentation : https://docs.stackable.tech/stackablectl/stable/demos/trino-taxi-data.html
6
- stackableStack : trino-superset-s3
3
+ airflow-scheduled-job :
4
+ description : Activate a simple Airflow DAG to run continuously at a set interval
5
+ stackableStack : airflow
7
6
labels :
8
- - trino
9
- - superset
10
- - minio
11
- - s3
12
- - ny-taxi-data
7
+ - airflow
8
+ - job-scheduling
13
9
manifests :
14
- - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/trino-taxi-data/load-test-data.yaml
15
- - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/trino-taxi-data/create-table-in-trino.yaml
16
- - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/trino-taxi-data/setup-superset.yaml
17
- kafka-druid-water-level-data :
18
- description : Demo ingesting water level data into Kafka, streaming it into Druid and creating a Superset dashboard
19
- documentation : https://docs.stackable.tech/stackablectl/stable/demos/kafka-druid-water-level-data.html
20
- stackableStack : kafka-druid-superset-s3
10
+ - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/airflow-scheduled-job/enable-and-run-dag.yaml
11
+ hbase-hdfs-load-cycling-data :
12
+ description : Copy data from S3 bucket to an HBase table
13
+ stackableStack : hdfs-hbase
21
14
labels :
22
- - kafka
23
- - druid
24
- - superset
25
- - minio
26
- - s3
27
- - water-levels
15
+ - hbase
16
+ - hdfs
17
+ - cycling-tripdata
28
18
manifests :
29
- - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/kafka-druid-water-level-data/ingest-test-data.yaml
30
- - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/kafka-druid-water-level-data/create-druid-ingestion-job.yaml
31
- - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/kafka-druid-water-level-data/setup-superset.yaml
19
+ - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/hbase-hdfs-load-cycling-data/01-distcp-cycling-data.yaml
20
+ - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/hbase-hdfs-load-cycling-data/02-create-hfile-and-import-to-hbase.yaml
32
21
nifi-kafka-druid-earthquake-data :
33
22
description : Demo ingesting earthquake data into Kafka using NiFi, streaming it into Druid and creating a Superset dashboard
34
23
documentation : https://docs.stackable.tech/stackablectl/stable/demos/nifi-kafka-druid-earthquake-data.html
@@ -45,21 +34,33 @@ demos:
45
34
- plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml
46
35
- plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml
47
36
- plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml
48
- hbase-hdfs-load-cycling-data :
49
- description : Copy data from S3 bucket to an HBase table
50
- stackableStack : hdfs-hbase
37
+ nifi-kafka-druid-water-level-data :
38
+ description : Demo ingesting water level data into Kafka using NiFi, streaming it into Druid and creating a Superset dashboard
39
+ documentation : https://docs.stackable.tech/stackablectl/stable/demos/nifi-kafka-druid-water-level-data.html
40
+ stackableStack : nifi-kafka-druid-superset-s3
51
41
labels :
52
- - hbase
53
- - hdfs
54
- - cycling-tripdata
42
+ - nifi
43
+ - kafka
44
+ - druid
45
+ - superset
46
+ - minio
47
+ - s3
48
+ - water-levels
55
49
manifests :
56
- - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/hbase-hdfs-load-cycling-data/01-distcp-cycling-data.yaml
57
- - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/hbase-hdfs-load-cycling-data/02-create-hfile-and-import-to-hbase.yaml
58
- airflow-scheduled-job :
59
- description : Activate a simple Airflow DAG to run continously at a set interval
60
- stackableStack : airflow
50
+ - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml # TODO
51
+ - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml # TODO
52
+ - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml # TODO
53
+ trino-taxi-data :
54
+ description : Demo loading 2.5 years of New York taxi data into S3 bucket, creating a Trino table and a Superset dashboard
55
+ documentation : https://docs.stackable.tech/stackablectl/stable/demos/trino-taxi-data.html
56
+ stackableStack : trino-superset-s3
61
57
labels :
62
- - airflow
63
- - job-scheduling
58
+ - trino
59
+ - superset
60
+ - minio
61
+ - s3
62
+ - ny-taxi-data
64
63
manifests :
65
- - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/airflow-scheduled-job/enable-and-run-dag.yaml
64
+ - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/trino-taxi-data/load-test-data.yaml
65
+ - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/trino-taxi-data/create-table-in-trino.yaml
66
+ - plainYaml : https://raw.githubusercontent.com/stackabletech/stackablectl/main/demos/trino-taxi-data/setup-superset.yaml
0 commit comments