mirror of
https://github.com/datahub-project/datahub.git
synced 2025-07-04 23:57:03 +00:00
2.0 KiB
2.0 KiB
import Tabs from '@theme/Tabs'; import TabItem from '@theme/TabItem';
DataFlow & DataJob
Why Would You Use DataFlow and DataJob?
The DataFlow and DataJob entities are used to represent data processing pipelines and jobs within a data ecosystem. They allow users to define, manage, and monitor the flow of data through various stages of processing, from ingestion to transformation and storage.
Goal Of This Guide
This guide will show you how to
- Create a DataFlow.
- Create a Datajob with a DataFlow.
Prerequisites
For this tutorial, you need to deploy DataHub Quickstart and ingest sample data. For detailed steps, please refer to Datahub Quickstart Guide.
Create DataFlow
{{ inline /metadata-ingestion/examples/library/create_dataflow.py show_path_as_comment }}
Create DataJob
DataJob must be associated with a DataFlow. You can create a DataJob by providing the DataFlow object or the DataFlow URN and its platform instance.
```python {{ inline /metadata-ingestion/examples/library/create_datajob.py show_path_as_comment }} ``` ```python {{ inline /metadata-ingestion/examples/library/create_datajob_with_flow_urn.py show_path_as_comment }} ```Read DataFlow
{{ inline /metadata-ingestion/examples/library/read_dataflow.py show_path_as_comment }}
Example Output
>> DataFlow name: example_dataflow
>> DataFlow platform: urn:li:dataPlatform:airflow
>> DataFlow description: airflow pipeline for production
Read DataJob
{{ inline /metadata-ingestion/examples/library/read_datajob.py show_path_as_comment }}
Example Output
>> DataJob name: example_datajob
>> DataJob Flow URN: urn:li:dataFlow:(airflow,PROD.example_dag,PROD)
>> DataJob description: example datajob