Docs: Brand Name Tag Updation (#22760)

* Docs:  Updation

* Docs:  Updation

* Docs:  Updation

---------

Co-authored-by: “Rounak <“rounakpreet.d@deuexsolutions.com”>
Co-authored-by: Prajwal214 <167504578+Prajwal214@users.noreply.github.com>
This commit is contained in:
Rounak Dhillon 2025-08-06 10:25:09 +05:30 committed by GitHub
parent 79189e49e0
commit 2b2a4090fc
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
37 changed files with 76 additions and 46 deletions

View File

@ -1,6 +1,6 @@
---
title: Airflow Connector | OpenMetadata Workflow Orchestration
description: Connect Apache Airflow to OpenMetadata effortlessly. Complete setup guide, configuration steps, and pipeline metadata extraction for seamless data lineage tracking.
title: Airflow Connector | `brandName` Workflow Orchestration
description: Connect Apache Airflow to `brandName` effortlessly. Complete setup guide, configuration steps, and pipeline metadata extraction for seamless data lineage tracking.
slug: /connectors/pipeline/airflow
---

View File

@ -1,5 +1,6 @@
---
title: Collate SaaS
title: Collate SaaS Setup Guide | Connect Database Services for Metadata Extraction
description: Learn how to configure database services in Collate SaaS, allow IP access, test connections, and begin metadata extraction using built-in connectors like Snowflake.
slug: /getting-started/day-1/collate-saas
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Run the ingestion from your Airflow
description: Learn how to run OpenMetadata ingestion workflows using Python, Docker, or Virtualenv operators in Airflow for secure and flexible metadata ingestion.
slug: /getting-started/day-1/hybrid-saas/airflow
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Managing Credentials
title: Managing Credentials Securely | Collate Ingestion Best Practices
description: Learn secure ways to manage credentials in Collate ingestion workflows using environment variables, Airflow connections, GitHub secrets, and existing services.
slug: /getting-started/day-1/hybrid-saas/credentials
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Run the ingestion from GCP Composer
title: Run the ingestion from GCP Composer
description: Learn to run Collate ingestion in GCP Composer using Python or KubernetesPod operators. Install packages or run containers for secure, scalable ingestion.
slug: /getting-started/day-1/hybrid-saas/gcp-composer
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Run the ingestion from GitHub Actions
description: Automate Collate ingestion with GitHub Actions. Set up YAML configs, secure credentials, schedule workflows, and send Slack alerts on failure.
slug: /getting-started/day-1/hybrid-saas/github-actions
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Hybrid Ingestion Runner
title: Hybrid Ingestion Runner | Secure Metadata Workflows in Your Cloud
description: Learn to configure and manage Hybrid Ingestion Runner to securely execute workflows in your cloud using AWS, Azure, or GCP secrets—without exposing credentials.
slug: /getting-started/day-1/hybrid-saas/hybrid-ingestion-runner
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Hybrid SaaS | Secure Metadata Ingestion for Collate
description: Learn how to securely ingest metadata using the Collate Ingestion Agent in your own infrastructure. Ideal for private networks and hybrid SaaS setups.
slug: /getting-started/day-1/hybrid-saas
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Run the ingestion from AWS MWAA
description: Set up Collate ingestion workflows on AWS MWAA using Python, ECS, or Virtualenv operators. Compare approaches and configure DAGs for secure metadata ingestion.
slug: /getting-started/day-1/hybrid-saas/mwaa
collate: true
---
@ -35,7 +36,7 @@ To install the package, we need to update the `requirements.txt` file from the M
openmetadata-ingestion[<plugin>]==x.y.z
```
Where `x.y.z` is the version of the OpenMetadata ingestion package. Note that the version needs to match the server version. If we are using the server at 1.3.1, then the ingestion package needs to also be 1.3.1.
Where `x.y.z` is the version of the Collate ingestion package. Note that the version needs to match the server version. If we are using the server at 1.3.1, then the ingestion package needs to also be 1.3.1.
The plugin parameter is a list of the sources that we want to ingest. An example would look like this `openmetadata-ingestion[mysql,snowflake,s3]==1.3.1`.
@ -78,7 +79,7 @@ def metadata_ingestion_workflow():
with DAG(
"redshift_ingestion",
default_args=default_args,
description="An example DAG which runs a OpenMetadata ingestion workflow",
description="An example DAG which runs a Collate ingestion workflow",
start_date=days_ago(1),
is_paused_upon_creation=False,
catchup=False,
@ -293,7 +294,7 @@ with DAG(
```
Note that depending on the kind of workflow you will be deploying, the YAML configuration will need to updated following
the official OpenMetadata docs, and the value of the `pipelineType` configuration will need to hold one of the following values:
the official Collate docs, and the value of the `pipelineType` configuration will need to hold one of the following values:
- `metadata`
- `usage`
@ -410,7 +411,7 @@ YAML config
with DAG(
"redshift_ingestion",
default_args=default_args,
description="An example DAG which runs a OpenMetadata ingestion workflow",
description="An example DAG which runs a Collate ingestion workflow",
start_date=days_ago(1),
is_paused_upon_creation=False,
catchup=False,
@ -419,7 +420,7 @@ with DAG(
task_id="ingest_redshift",
python_callable=metadata_ingestion_workflow,
requirements=['openmetadata-ingestion==1.0.5.0',
'apache-airflow==2.4.3', # note, v2.4.3 is the first version that does not conflict with OpenMetadata's 'tabulate' requirements
'apache-airflow==2.4.3', # note, v2.4.3 is the first version that does not conflict with Collate's 'tabulate' requirements
'apache-airflow-providers-amazon==6.0.0', # Amazon Airflow provider is necessary for MWAA
'watchtower',],
system_site_packages=False,
@ -430,7 +431,7 @@ with DAG(
Where you can update the YAML configuration and workflow classes accordingly. accordingly. Further examples on how to
run the ingestion can be found on the documentation (e.g., [Snowflake](/connectors/database/snowflake)).
You will also need to determine the OpenMetadata ingestion extras and Airflow providers you need. Note that the Openmetadata version needs to match the server version. If we are using the server at 0.12.2, then the ingestion package needs to also be 0.12.2. An example of the extras would look like this `openmetadata-ingestion[mysql,snowflake,s3]==0.12.2.2`.
You will also need to determine the Collate ingestion extras and Airflow providers you need. Note that the Openmetadata version needs to match the server version. If we are using the server at 0.12.2, then the ingestion package needs to also be 0.12.2. An example of the extras would look like this `openmetadata-ingestion[mysql,snowflake,s3]==0.12.2.2`.
For Airflow providers, you will want to pull the provider versions from [the matching constraints file](https://raw.githubusercontent.com/apache/airflow/constraints-2.4.3/constraints-3.7.txt). Since this example installs Airflow Providers v2.4.3 on Python 3.7, we use that constraints file.
Also note that the ingestion workflow function must be entirely self-contained as it will run by itself in the virtualenv. Any imports it needs, including the configuration, must exist within the function itself.

View File

@ -1,5 +1,6 @@
---
title: Day 1
title: Day 1 Setup Guide | Connect, Ingest & Collaborate in Collate
description: Start your Collate journey by connecting data sources, ingesting metadata, inviting users, and exploring key features. Everything you need for Day 1.
slug: /getting-started/day-1
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Getting Started
title: Getting Started with Collate | Setup & Onboarding Guide
description: Quickly set up your Collate environment in 30 minutes. Learn how to connect data sources, explore features, and begin your onboarding journey.
slug: /getting-started
collate: true
---

View File

@ -1,5 +1,5 @@
---
title: Data Profiler | OpenMetadata Data Profiling Guide
title: Data Profiler | `brandName` Data Profiling Guide
description: Explore profiling workflows including histogram metrics, null counts, and field-level health.
slug: /how-to-guides/data-quality-observability/profiler
---

View File

@ -1,6 +1,6 @@
---
title: Airflow Connector | OpenMetadata Workflow Orchestration
description: Connect Apache Airflow to OpenMetadata effortlessly. Complete setup guide, configuration steps, and pipeline metadata extraction for seamless data lineage tracking.
title: Airflow Connector | `brandName` Workflow Orchestration
description: Connect Apache Airflow to `brandName` effortlessly. Complete setup guide, configuration steps, and pipeline metadata extraction for seamless data lineage tracking.
slug: /connectors/pipeline/airflow
---

View File

@ -1,5 +1,6 @@
---
title: Collate SaaS
title: Collate SaaS Setup Guide | Connect Database Services for Metadata Extraction
description: Learn how to configure database services in Collate SaaS, allow IP access, test connections, and begin metadata extraction using built-in connectors like Snowflake.
slug: /getting-started/day-1/collate-saas
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Run the ingestion from your Airflow
description: Learn how to run OpenMetadata ingestion workflows using Python, Docker, or Virtualenv operators in Airflow for secure and flexible metadata ingestion.
slug: /getting-started/day-1/hybrid-saas/airflow
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Managing Credentials
title: Managing Credentials Securely | Collate Ingestion Best Practices
description: Learn secure ways to manage credentials in Collate ingestion workflows using environment variables, Airflow connections, GitHub secrets, and existing services.
slug: /getting-started/day-1/hybrid-saas/credentials
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Run the ingestion from GCP Composer
title: Run the ingestion from GCP Composer
description: Learn to run Collate ingestion in GCP Composer using Python or KubernetesPod operators. Install packages or run containers for secure, scalable ingestion.
slug: /getting-started/day-1/hybrid-saas/gcp-composer
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Run the ingestion from GitHub Actions
description: Automate Collate ingestion with GitHub Actions. Set up YAML configs, secure credentials, schedule workflows, and send Slack alerts on failure.
slug: /getting-started/day-1/hybrid-saas/github-actions
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Hybrid Ingestion Runner
title: Hybrid Ingestion Runner | Secure Metadata Workflows in Your Cloud
description: Learn to configure and manage Hybrid Ingestion Runner to securely execute workflows in your cloud using AWS, Azure, or GCP secrets—without exposing credentials.
slug: /getting-started/day-1/hybrid-saas/hybrid-ingestion-runner
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Hybrid SaaS | Secure Metadata Ingestion for Collate
description: Learn how to securely ingest metadata using the Collate Ingestion Agent in your own infrastructure. Ideal for private networks and hybrid SaaS setups.
slug: /getting-started/day-1/hybrid-saas
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Run the ingestion from AWS MWAA
description: Set up Collate ingestion workflows on AWS MWAA using Python, ECS, or Virtualenv operators. Compare approaches and configure DAGs for secure metadata ingestion.
slug: /getting-started/day-1/hybrid-saas/mwaa
collate: true
---
@ -35,7 +36,7 @@ To install the package, we need to update the `requirements.txt` file from the M
openmetadata-ingestion[<plugin>]==x.y.z
```
Where `x.y.z` is the version of the OpenMetadata ingestion package. Note that the version needs to match the server version. If we are using the server at 1.3.1, then the ingestion package needs to also be 1.3.1.
Where `x.y.z` is the version of the Collate ingestion package. Note that the version needs to match the server version. If we are using the server at 1.3.1, then the ingestion package needs to also be 1.3.1.
The plugin parameter is a list of the sources that we want to ingest. An example would look like this `openmetadata-ingestion[mysql,snowflake,s3]==1.3.1`.
@ -78,7 +79,7 @@ def metadata_ingestion_workflow():
with DAG(
"redshift_ingestion",
default_args=default_args,
description="An example DAG which runs a OpenMetadata ingestion workflow",
description="An example DAG which runs a Collate ingestion workflow",
start_date=days_ago(1),
is_paused_upon_creation=False,
catchup=False,
@ -293,7 +294,7 @@ with DAG(
```
Note that depending on the kind of workflow you will be deploying, the YAML configuration will need to updated following
the official OpenMetadata docs, and the value of the `pipelineType` configuration will need to hold one of the following values:
the official Collate docs, and the value of the `pipelineType` configuration will need to hold one of the following values:
- `metadata`
- `usage`
@ -410,7 +411,7 @@ YAML config
with DAG(
"redshift_ingestion",
default_args=default_args,
description="An example DAG which runs a OpenMetadata ingestion workflow",
description="An example DAG which runs a Collate ingestion workflow",
start_date=days_ago(1),
is_paused_upon_creation=False,
catchup=False,
@ -419,7 +420,7 @@ with DAG(
task_id="ingest_redshift",
python_callable=metadata_ingestion_workflow,
requirements=['openmetadata-ingestion==1.0.5.0',
'apache-airflow==2.4.3', # note, v2.4.3 is the first version that does not conflict with OpenMetadata's 'tabulate' requirements
'apache-airflow==2.4.3', # note, v2.4.3 is the first version that does not conflict with Collate's 'tabulate' requirements
'apache-airflow-providers-amazon==6.0.0', # Amazon Airflow provider is necessary for MWAA
'watchtower',],
system_site_packages=False,
@ -430,7 +431,7 @@ with DAG(
Where you can update the YAML configuration and workflow classes accordingly. accordingly. Further examples on how to
run the ingestion can be found on the documentation (e.g., [Snowflake](/connectors/database/snowflake)).
You will also need to determine the OpenMetadata ingestion extras and Airflow providers you need. Note that the Openmetadata version needs to match the server version. If we are using the server at 0.12.2, then the ingestion package needs to also be 0.12.2. An example of the extras would look like this `openmetadata-ingestion[mysql,snowflake,s3]==0.12.2.2`.
You will also need to determine the Collate ingestion extras and Airflow providers you need. Note that the Openmetadata version needs to match the server version. If we are using the server at 0.12.2, then the ingestion package needs to also be 0.12.2. An example of the extras would look like this `openmetadata-ingestion[mysql,snowflake,s3]==0.12.2.2`.
For Airflow providers, you will want to pull the provider versions from [the matching constraints file](https://raw.githubusercontent.com/apache/airflow/constraints-2.4.3/constraints-3.7.txt). Since this example installs Airflow Providers v2.4.3 on Python 3.7, we use that constraints file.
Also note that the ingestion workflow function must be entirely self-contained as it will run by itself in the virtualenv. Any imports it needs, including the configuration, must exist within the function itself.

View File

@ -1,5 +1,6 @@
---
title: Day 1
title: Day 1 Setup Guide | Connect, Ingest & Collaborate in Collate
description: Start your Collate journey by connecting data sources, ingesting metadata, inviting users, and exploring key features. Everything you need for Day 1.
slug: /getting-started/day-1
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Getting Started
title: Getting Started with Collate | Setup & Onboarding Guide
description: Quickly set up your Collate environment in 30 minutes. Learn how to connect data sources, explore features, and begin your onboarding journey.
slug: /getting-started
collate: true
---

View File

@ -1,5 +1,5 @@
---
title: Data Profiler | OpenMetadata Data Profiling Guide
title: Data Profiler | `brandName` Data Profiling Guide
description: Explore profiling workflows including histogram metrics, null counts, and field-level health.
slug: /how-to-guides/data-quality-observability/profiler
---

View File

@ -1,6 +1,6 @@
---
title: Airflow Connector | OpenMetadata Workflow Orchestration
description: Connect Apache Airflow to OpenMetadata effortlessly. Complete setup guide, configuration steps, and pipeline metadata extraction for seamless data lineage tracking.
title: Airflow Connector | `brandName` Workflow Orchestration
description: Connect Apache Airflow to `brandName` effortlessly. Complete setup guide, configuration steps, and pipeline metadata extraction for seamless data lineage tracking.
slug: /connectors/pipeline/airflow
---

View File

@ -1,5 +1,6 @@
---
title: Collate SaaS
title: Collate SaaS Setup Guide | Connect Database Services for Metadata Extraction
description: Learn how to configure database services in Collate SaaS, allow IP access, test connections, and begin metadata extraction using built-in connectors like Snowflake.
slug: /getting-started/day-1/collate-saas
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Run the ingestion from your Airflow
description: Learn how to run OpenMetadata ingestion workflows using Python, Docker, or Virtualenv operators in Airflow for secure and flexible metadata ingestion.
slug: /getting-started/day-1/hybrid-saas/airflow
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Managing Credentials
title: Managing Credentials Securely | Collate Ingestion Best Practices
description: Learn secure ways to manage credentials in Collate ingestion workflows using environment variables, Airflow connections, GitHub secrets, and existing services.
slug: /getting-started/day-1/hybrid-saas/credentials
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Run the ingestion from GCP Composer
title: Run the ingestion from GCP Composer
description: Learn to run Collate ingestion in GCP Composer using Python or KubernetesPod operators. Install packages or run containers for secure, scalable ingestion.
slug: /getting-started/day-1/hybrid-saas/gcp-composer
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Run the ingestion from GitHub Actions
description: Automate Collate ingestion with GitHub Actions. Set up YAML configs, secure credentials, schedule workflows, and send Slack alerts on failure.
slug: /getting-started/day-1/hybrid-saas/github-actions
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Hybrid Ingestion Runner
title: Hybrid Ingestion Runner | Secure Metadata Workflows in Your Cloud
description: Learn to configure and manage Hybrid Ingestion Runner to securely execute workflows in your cloud using AWS, Azure, or GCP secrets—without exposing credentials.
slug: /getting-started/day-1/hybrid-saas/hybrid-ingestion-runner
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Hybrid SaaS | Secure Metadata Ingestion for Collate
description: Learn how to securely ingest metadata using the Collate Ingestion Agent in your own infrastructure. Ideal for private networks and hybrid SaaS setups.
slug: /getting-started/day-1/hybrid-saas
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Run the ingestion from AWS MWAA
description: Set up Collate ingestion workflows on AWS MWAA using Python, ECS, or Virtualenv operators. Compare approaches and configure DAGs for secure metadata ingestion.
slug: /getting-started/day-1/hybrid-saas/mwaa
collate: true
---
@ -35,7 +36,7 @@ To install the package, we need to update the `requirements.txt` file from the M
openmetadata-ingestion[<plugin>]==x.y.z
```
Where `x.y.z` is the version of the OpenMetadata ingestion package. Note that the version needs to match the server version. If we are using the server at 1.3.1, then the ingestion package needs to also be 1.3.1.
Where `x.y.z` is the version of the Collate ingestion package. Note that the version needs to match the server version. If we are using the server at 1.3.1, then the ingestion package needs to also be 1.3.1.
The plugin parameter is a list of the sources that we want to ingest. An example would look like this `openmetadata-ingestion[mysql,snowflake,s3]==1.3.1`.
@ -78,7 +79,7 @@ def metadata_ingestion_workflow():
with DAG(
"redshift_ingestion",
default_args=default_args,
description="An example DAG which runs a OpenMetadata ingestion workflow",
description="An example DAG which runs a Collate ingestion workflow",
start_date=days_ago(1),
is_paused_upon_creation=False,
catchup=False,
@ -293,7 +294,7 @@ with DAG(
```
Note that depending on the kind of workflow you will be deploying, the YAML configuration will need to updated following
the official OpenMetadata docs, and the value of the `pipelineType` configuration will need to hold one of the following values:
the official Collate docs, and the value of the `pipelineType` configuration will need to hold one of the following values:
- `metadata`
- `usage`
@ -410,7 +411,7 @@ YAML config
with DAG(
"redshift_ingestion",
default_args=default_args,
description="An example DAG which runs a OpenMetadata ingestion workflow",
description="An example DAG which runs a Collate ingestion workflow",
start_date=days_ago(1),
is_paused_upon_creation=False,
catchup=False,
@ -419,7 +420,7 @@ with DAG(
task_id="ingest_redshift",
python_callable=metadata_ingestion_workflow,
requirements=['openmetadata-ingestion==1.0.5.0',
'apache-airflow==2.4.3', # note, v2.4.3 is the first version that does not conflict with OpenMetadata's 'tabulate' requirements
'apache-airflow==2.4.3', # note, v2.4.3 is the first version that does not conflict with Collate's 'tabulate' requirements
'apache-airflow-providers-amazon==6.0.0', # Amazon Airflow provider is necessary for MWAA
'watchtower',],
system_site_packages=False,
@ -430,7 +431,7 @@ with DAG(
Where you can update the YAML configuration and workflow classes accordingly. accordingly. Further examples on how to
run the ingestion can be found on the documentation (e.g., [Snowflake](/connectors/database/snowflake)).
You will also need to determine the OpenMetadata ingestion extras and Airflow providers you need. Note that the Openmetadata version needs to match the server version. If we are using the server at 0.12.2, then the ingestion package needs to also be 0.12.2. An example of the extras would look like this `openmetadata-ingestion[mysql,snowflake,s3]==0.12.2.2`.
You will also need to determine the Collate ingestion extras and Airflow providers you need. Note that the Openmetadata version needs to match the server version. If we are using the server at 0.12.2, then the ingestion package needs to also be 0.12.2. An example of the extras would look like this `openmetadata-ingestion[mysql,snowflake,s3]==0.12.2.2`.
For Airflow providers, you will want to pull the provider versions from [the matching constraints file](https://raw.githubusercontent.com/apache/airflow/constraints-2.4.3/constraints-3.7.txt). Since this example installs Airflow Providers v2.4.3 on Python 3.7, we use that constraints file.
Also note that the ingestion workflow function must be entirely self-contained as it will run by itself in the virtualenv. Any imports it needs, including the configuration, must exist within the function itself.

View File

@ -1,5 +1,6 @@
---
title: Day 1
title: Day 1 Setup Guide | Connect, Ingest & Collaborate in Collate
description: Start your Collate journey by connecting data sources, ingesting metadata, inviting users, and exploring key features. Everything you need for Day 1.
slug: /getting-started/day-1
collate: true
---

View File

@ -1,5 +1,6 @@
---
title: Getting Started
title: Getting Started with Collate | Setup & Onboarding Guide
description: Quickly set up your Collate environment in 30 minutes. Learn how to connect data sources, explore features, and begin your onboarding journey.
slug: /getting-started
collate: true
---

View File

@ -1,5 +1,5 @@
---
title: Data Profiler | OpenMetadata Data Profiling Guide
title: Data Profiler | `brandName` Data Profiling Guide
description: Explore profiling workflows including histogram metrics, null counts, and field-level health.
slug: /how-to-guides/data-quality-observability/profiler
---

View File

@ -1,5 +1,5 @@
---
title: Tests - UI Config | OpenMetadata Quality Config Guide
title: Tests - UI Config | `brandName` Quality Config Guide
description: Define UI tests to validate data quality during ingestion and enforce expectations at column or table level.
slug: /how-to-guides/data-quality-observability/quality/tests-ui
---