2022-07-11 16:54:40 +02:00
|
|
|
---
|
|
|
|
title: Run Airflow Connector using the CLI
|
|
|
|
slug: /openmetadata/connectors/pipeline/airflow/cli
|
|
|
|
---
|
|
|
|
|
|
|
|
<ConnectorIntro connector="Airflow" goal="CLI"/>
|
|
|
|
|
|
|
|
<Requirements />
|
|
|
|
|
2022-08-18 11:43:45 +02:00
|
|
|
<PythonMod connector="Airflow" module="airflow" />
|
|
|
|
|
|
|
|
Note that this installs the same Airflow version that we ship in the Ingestion Container, which is
|
|
|
|
Airflow `2.3.3` from Release `0.12`.
|
|
|
|
|
|
|
|
The ingestion using Airflow version 2.3.3 as a source package has been tested against Airflow 2.3.3 and Airflow 2.2.5.
|
|
|
|
|
2022-07-11 16:54:40 +02:00
|
|
|
<MetadataIngestionServiceDev service="pipeline" connector="Airflow" goal="CLI"/>
|
|
|
|
|
|
|
|
<h4>Source Configuration - Service Connection</h4>
|
|
|
|
|
|
|
|
- **hostPort**: URL to the Airflow instance.
|
|
|
|
- **numberOfStatus**: Number of status we want to look back to in every ingestion (e.g., Past executions from a DAG).
|
|
|
|
- **connection**: Airflow metadata database connection. See
|
|
|
|
these [docs](https://airflow.apache.org/docs/apache-airflow/stable/howto/set-up-database.html)
|
|
|
|
for supported backends.
|
|
|
|
|
|
|
|
In terms of `connection` we support the following selections:
|
|
|
|
|
|
|
|
- `backend`: Should not be used from the UI. This is only applicable when ingesting Airflow metadata locally by running
|
|
|
|
the ingestion from a DAG. It will use the current Airflow SQLAlchemy connection to extract the data.
|
|
|
|
- `MySQL`, `Postgres`, `MSSQL` and `SQLite`: Pass the required credentials to reach out each of these services. We will
|
|
|
|
create a connection to the pointed database and read Airflow data from there.
|
|
|
|
|
|
|
|
<MetadataIngestionConfig service="pipeline" connector="Airflow" goal="CLI" />
|