Ignoring revisions in .git-blame-ignore-revs. Click here to bypass and see the normal blame view.

32 lines
1.3 KiB
Markdown
Raw Normal View History

---
title: Spark Engine Configuration | OpenMetadata Spark Profiling Setup
description: Learn how to configure your profiler pipeline to use Spark Engine for distributed data profiling.
slug: /how-to-guides/data-quality-observability/profiler/spark-engine/configuration
collate: true
---
# Spark Engine Configuration
## Overview
There are two ways to configure Spark Engine in OpenMetadata:
{% inlineCalloutContainer %}
{% inlineCallout
color="violet-70"
bold="UI Configuration"
icon="MdAnalytics"
href="/how-to-guides/data-quality-observability/profiler/spark-engine/ui-configuration" %}
Configure Spark Engine through the OpenMetadata UI.
{% /inlineCallout %}
{% inlineCallout
icon="MdOutlineSchema"
bold="External Configuration"
href="/how-to-guides/data-quality-observability/profiler/spark-engine/external-configuration" %}
Configure Spark Engine using YAML files for external workflows.
{% /inlineCallout %}
{% /inlineCalloutContainer %}
{% note %}
Before configuring, ensure you have completed the [Spark Engine Prerequisites](/how-to-guides/data-quality-observability/profiler/spark-engine/prerequisites) and understand the [Partitioning Requirements](/how-to-guides/data-quality-observability/profiler/spark-engine/partitioning).
{% /note %}