Skip to content

Commit 2426269

Browse files
committed
Integrate: Naming things. Adjust reference labels.
Let's also use canonical, non-prefixed variants for the reference labels. This patch aggressively prunes previous nomenclature to create less confusion for future authors. If anything breaks on this, it is now the right time to fix forward.
1 parent f6223a3 commit 2426269

File tree

15 files changed

+45
-54
lines changed

15 files changed

+45
-54
lines changed

docs/ingest/cdc/index.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -23,15 +23,15 @@ Native and specialized integration connectors for CrateDB, both managed and unma
2323
:gutter: 2
2424

2525
::::{grid-item-card} Amazon DynamoDB
26-
:link: aws-dynamodb
26+
:link: dynamodb
2727
:link-type: ref
2828
Load data from DynamoDB, a fully managed NoSQL database service provided by
2929
Amazon Web Services (AWS), which is designed for high-performance, scalable
3030
applications and offers key-value and document data structures.
3131
::::
3232

3333
::::{grid-item-card} Amazon Kinesis
34-
:link: aws-kinesis
34+
:link: kinesis
3535
:link-type: ref
3636
Load data from Amazon Kinesis Data Streams, a serverless streaming data service
3737
that simplifies the capture, processing, and storage of data streams at any scale.
@@ -56,7 +56,7 @@ both managed and unmanaged.
5656
:gutter: 2
5757

5858
::::{grid-item-card} AWS DMS
59-
:link: aws-dms
59+
:link: dms
6060
:link-type: ref
6161
Use AWS Database Migration Service (AWS DMS), a managed migration and replication
6262
service that helps move your database and analytics workloads between different

docs/ingest/etl/index.md

Lines changed: 31 additions & 31 deletions
Original file line numberDiff line numberDiff line change
@@ -27,29 +27,23 @@ outlines how to use them effectively. Additionally, see support for {ref}`cdc` s
2727

2828

2929
::::{grid-item-card} {material-outlined}`air;2em` Dataflow / Pipeline / Code-first
30-
- {ref}`apache-airflow`
30+
- {ref}`airflow`
3131

3232
Apache Airflow is an open-source software platform to programmatically author,
3333
schedule, and monitor workflows. Pipelines are defined in Python, allowing for
3434
dynamic pipeline generation and on-demand, code-driven pipeline invocation.
3535

36-
- {ref}`apache-flink`
37-
38-
Apache Flink is a programming framework and distributed processing engine for
39-
stateful computations over unbounded and bounded data streams, written in Java.
40-
41-
- {ref}`apache-nifi`
42-
43-
Apache NiFi is a dataflow system based on the concepts of flow-based programming.
44-
It supports powerful and scalable directed graphs of data routing, transformation,
45-
and system mediation logic.
46-
4736
- {ref}`dbt`
4837

4938
dbt is an SQL-first platform for transforming data in data warehouses using
5039
Python and SQL. The data abstraction layer provided by dbt-core allows the
5140
decoupling of the models on which reports and dashboards rely from the source data.
5241

42+
- {ref}`flink`
43+
44+
Apache Flink is a programming framework and distributed processing engine for
45+
stateful computations over unbounded and bounded data streams, written in Java.
46+
5347
- {ref}`kestra`
5448

5549
Kestra is an open-source workflow automation and orchestration toolkit with a rich
@@ -63,23 +57,29 @@ outlines how to use them effectively. Additionally, see support for {ref}`cdc` s
6357
the Singer specification. Singer is a composable open-source ETL framework and
6458
specification, including powerful data extraction and consolidation elements.
6559

60+
- {ref}`nifi`
61+
62+
Apache NiFi is a dataflow system based on the concepts of flow-based programming.
63+
It supports powerful and scalable directed graphs of data routing, transformation,
64+
and system mediation logic.
65+
6666
+++
6767
Use data pipeline programming frameworks and platforms.
6868
::::
6969

7070

7171
::::{grid-item-card} {material-outlined}`all_inclusive;2em` Low-code / No-code / Visual
72-
- {ref}`apache-hop`
73-
74-
Apache Hop aims to be the future of data integration. Visual development enables
75-
developers to be more productive than they can be through code.
76-
7772
- {ref}`estuary`
7873

7974
Estuary provides real-time data integration and modern ETL and ELT data pipelines
8075
as a fully managed solution. Estuary Flow is a real-time, reliable change data
8176
capture (CDC) solution.
8277

78+
- {ref}`hop`
79+
80+
Apache Hop aims to be the future of data integration. Visual development enables
81+
developers to be more productive than they can be through code.
82+
8383
- {ref}`n8n`
8484

8585
n8n is a workflow automation tool that helps you to connect any app with an API with
@@ -97,13 +97,13 @@ Use visual data flow and integration frameworks and platforms.
9797

9898

9999
::::{grid-item-card} {material-outlined}`storage;2em` Databases
100-
- {ref}`aws-dms`
100+
- {ref}`dms`
101101

102102
AWS DMS is a managed migration and replication service that helps move your
103103
database and analytics workloads between different kinds of databases quickly,
104104
securely, and with minimal downtime and zero data loss.
105105

106-
- {ref}`aws-dynamodb`
106+
- {ref}`dynamodb`
107107

108108
DynamoDB is a fully managed NoSQL database service provided by Amazon Web Services (AWS).
109109

@@ -132,13 +132,13 @@ Load data from database systems.
132132

133133

134134
::::{grid-item-card} {material-outlined}`fast_forward;2em` Streams
135-
- {ref}`apache-kafka`
135+
- {ref}`kafka`
136136

137137
Apache Kafka is an open-source distributed event streaming platform
138138
for high-performance data pipelines, streaming analytics, data integration,
139139
and mission-critical applications.
140140

141-
- {ref}`aws-kinesis`
141+
- {ref}`kinesis`
142142

143143
Amazon Kinesis Data Streams is a serverless streaming data service that simplifies
144144
the capture, processing, and storage of data streams at any scale, such as
@@ -186,7 +186,7 @@ Use serverless compute units for custom import tasks.
186186

187187
::::{grid-item-card} {material-outlined}`dataset;2em` Datasets
188188

189-
- {ref}`apache-iceberg`
189+
- {ref}`iceberg`
190190

191191
Apache Iceberg is an open table format for analytic datasets.
192192

@@ -202,25 +202,25 @@ Load data from datasets and open table formats.
202202
:::
203203

204204
:::{div}
205-
- {ref}`apache-airflow`
206-
- {ref}`apache-flink`
207-
- {ref}`apache-hop`
208-
- {ref}`apache-iceberg`
209-
- {ref}`apache-kafka`
210-
- {ref}`apache-nifi`
211-
- {ref}`aws-dynamodb`
212-
- {ref}`aws-kinesis`
213-
- {ref}`aws-dms`
205+
- {ref}`airflow`
214206
- {ref}`aws-lambda`
215207
- {ref}`azure-functions`
216208
- {ref}`dbt`
209+
- {ref}`dms`
210+
- {ref}`dynamodb`
217211
- {ref}`estuary`
212+
- {ref}`flink`
213+
- {ref}`hop`
214+
- {ref}`iceberg`
218215
- {ref}`influxdb`
216+
- {ref}`kafka`
219217
- {ref}`kestra`
218+
- {ref}`kinesis`
220219
- {ref}`meltano`
221220
- {ref}`mongodb`
222221
- {ref}`mysql`
223222
- {ref}`n8n`
223+
- {ref}`nifi`
224224
- {ref}`node-red`
225225
- {ref}`risingwave`
226226
- {ref}`sql-server`

docs/integrate/airflow/index.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,4 @@
11
(airflow)=
2-
(apache-airflow)=
32
(astronomer)=
43
# Airflow / Astronomer
54

docs/integrate/dbt/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ With dbt, anyone on your data team can safely contribute to production-grade dat
3030
pipelines.
3131

3232
The idea is that data engineers make source data available to an environment where
33-
dbt projects run, for example with [Debezium](#debezium) or with [Airflow](#apache-airflow).
33+
dbt projects run, for example with {ref}`debezium` or with {ref}`airflow`.
3434
Afterwards, data analysts can run their dbt projects against this data to produce models
3535
(tables and views) that can be used with a number of [BI tools](#bi-tools).
3636

docs/integrate/dms/index.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,4 @@
1-
(aws-dms)=
2-
(cdc-dms)=
1+
(dms)=
32
# DMS (AWS Database Migration Service)
43

54
:::{include} /_include/links.md

docs/integrate/dynamodb/index.md

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,4 @@
1-
(aws-dynamodb)=
2-
(cdc-dynamodb)=
1+
(dynamodb)=
32
# DynamoDB
43

54
:::{include} /_include/links.md
@@ -38,7 +37,7 @@ servers or infrastructure.
3837
:::{rubric} Related
3938
:::
4039
- [Amazon DynamoDB Streams]
41-
- {ref}`aws-kinesis`
40+
- {ref}`kinesis`
4241
- [Amazon Kinesis Data Streams]
4342
::::
4443

docs/integrate/estuary/index.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,4 @@
11
(estuary)=
2-
32
# Estuary
43

54
```{div} .float-right

docs/integrate/flink/index.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,3 @@
1-
(apache-flink)=
21
(flink)=
32
# Flink
43

docs/integrate/hop/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
(apache-hop)=
1+
(hop)=
22
# Hop
33

44
```{div} .float-right

docs/integrate/iceberg/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
(apache-iceberg)=
1+
(iceberg)=
22
# Iceberg
33

44
```{div} .float-right

0 commit comments

Comments
 (0)