
* Doc: SEO Title Updation * Doc: Data Discovery Gif Updation * Doc: SEO title updation * Doc: SEO Title Updation --------- Co-authored-by: “Rounak <“rounakpreet.d@deuexsolutions.com”>
3.7 KiB
title | slug |
---|---|
Kafka Connector | OpenMetadata Messaging Integration | /connectors/messaging/kafka |
{% connectorDetailsHeader name="Kafka" stage="PROD" platform="OpenMetadata" availableFeatures=["Topics", "Sample Data"] unavailableFeatures=[] / %}
In this section, we provide guides and references to use the Kafka connector.
Configure and schedule Kafka metadata and profiler workflows from the OpenMetadata UI:
{% partial file="/v1.7/connectors/ingestion-modes-tiles.md" variables={yamlPath: "/connectors/messaging/kafka/yaml"} /%}
Requirements
Connecting to Kafka does not require any previous configuration.
The ingestion of the Kafka topics' schema is done separately by configuring the Schema Registry URL. However, only the Bootstrap Servers information is mandatory.
Metadata Ingestion
{% partial file="/v1.7/connectors/metadata-ingestion-ui.md" variables={ connector: "Kafka", selectServicePath: "/images/v1.7/connectors/kafka/select-service.png", addNewServicePath: "/images/v1.7/connectors/kafka/add-new-service.png", serviceConnectionPath: "/images/v1.7/connectors/kafka/service-connection.png", } /%}
{% stepsContainer %} {% extraContent parentTagName="stepsContainer" %}
Connection Details
- Bootstrap Servers: List of brokers as comma separated values of broker
host
orhost:port
. Example:host1:9092,host2:9092
- Schema Registry URL: URL of the Schema Registry used to ingest the schemas of the topics.
- SASL Username: SASL username for use with the PLAIN and SASL-SCRAM mechanisms.
- SASL Password: SASL password for use with the PLAIN and SASL-SCRAM mechanisms.
- SASL Mechanism: SASL mechanism to use for authentication.
- Basic Auth User Info: Schema Registry Client HTTP credentials in the form of
username:password
. By default, user info is extracted from the URL if present. - Consumer Config: The accepted additional values for the consumer configuration can be found in the following link.
If you are using Confluent kafka and SSL encryption is enabled you need to addsecurity.protocol
as key andSASL_SSL
as value under Consumer Config - Schema Registry Config: The accepted additional values for the Schema Registry configuration can be found in the following link.
{% note %}
To ingest the topic schema Schema Registry URL
must be passed
{% /note %}
{% /extraContent %}
{% partial file="/v1.7/connectors/test-connection.md" /%}
{% partial file="/v1.7/connectors/messaging/configure-ingestion.md" /%}
{% partial file="/v1.7/connectors/ingestion-schedule-and-deploy.md" /%}
{% /stepsContainer %}
Securing Kafka Connection with SSL in OpenMetadata
To establish secure connections between OpenMetadata and Kafka, navigate to the Advanced Config
section. Here, you can provide the CA certificate used for SSL validation by specifying the caCertificate
. Alternatively, if both client and server require mutual authentication, you'll need to use all three parameters: ssl key
, ssl cert
, and caCertificate
. In this case, ssl_cert
is used for the client’s SSL certificate, ssl_key
for the private key associated with the SSL certificate, and caCertificate
for the CA certificate to validate the server’s certificate.
{% image src="/images/v1.7/connectors/ssl_kafka.png" alt="SSL Configuration" height="450px" caption="SSL Configuration" /%}