mirror of
https://github.com/open-metadata/OpenMetadata.git
synced 2025-07-16 13:41:37 +00:00
129 lines
5.1 KiB
Markdown
129 lines
5.1 KiB
Markdown
![]() |
---
|
||
|
title: Backup Metadata
|
||
|
slug: /deployment/upgrade/backup-metadata
|
||
|
---
|
||
|
|
||
|
# Backup Metadata
|
||
|
|
||
|
## Introduction
|
||
|
|
||
|
The goal of OpenMetadata is to enable company-wide collaboration around metadata. The more we use it, the more value
|
||
|
this brings to the table, which means that keeping the metadata safe can become a critical activity for our Disaster
|
||
|
Recovery practices.
|
||
|
|
||
|
While there are cloud services that feature automatic snapshots and replication, the metadata CLI
|
||
|
now allows all users to perform backups regardless of the underlying infrastructure.
|
||
|
|
||
|
## Installation
|
||
|
|
||
|
The CLI comes bundled in the base `openmetadata-ingestion` Python package. You can install it with:
|
||
|
|
||
|
```commandline
|
||
|
pip install openmetadata-ingestion
|
||
|
```
|
||
|
|
||
|
One of the `backup` features is to upload the generated backup to cloud storage (currently supporting S3). To use this,
|
||
|
you can instead install the package with the backup plugin:
|
||
|
|
||
|
```commandline
|
||
|
pip install "openmetadata-ingestion[backup]"
|
||
|
```
|
||
|
|
||
|
This tool acts as a wrapper around the powerful `mysqldump` utility with some commodity addons on top. `mysqldump` is part
|
||
|
of the `mysql-client` package and can be installed on your machine as:
|
||
|
|
||
|
- **macOS**: `brew install mysql-client`
|
||
|
- **Ubuntu**: `sudo apt-get install mysql-client`
|
||
|
|
||
|
## Backup CLI
|
||
|
|
||
|
After the installation, we can take a look at the different options to run the CLI:
|
||
|
|
||
|
```commandline
|
||
|
> metadata backup --help
|
||
|
Usage: metadata backup [OPTIONS]
|
||
|
|
||
|
Run a backup for the metadata DB. Requires mysqldump installed on the
|
||
|
host.
|
||
|
|
||
|
We can pass as many options as required with `-o <opt1>, -o <opt2> [...]`
|
||
|
|
||
|
To run the upload, provide the information as `--upload endpoint bucket
|
||
|
key` and properly configure the environment variables AWS_ACCESS_KEY_ID &
|
||
|
AWS_SECRET_ACCESS_KEY
|
||
|
|
||
|
Options:
|
||
|
-h, --host TEXT Host that runs the database [required]
|
||
|
-u, --user TEXT User to run the backup [required]
|
||
|
-p, --password TEXT Credentials for the user [required]
|
||
|
-d, --database TEXT Database to backup [required]
|
||
|
--port TEXT Database service port
|
||
|
--output PATH Local path to store the backup
|
||
|
--upload <TEXT TEXT TEXT>... S3 endpoint, bucket & key to upload the backup
|
||
|
file
|
||
|
-o, --options TEXT
|
||
|
```
|
||
|
|
||
|
### Database Connection
|
||
|
|
||
|
There is a set of four required parameters, the minimum required for us to access the database service and run the
|
||
|
backup: `host`, `user`, `password` and `database` to point to. Note that the user should have at least read access to the
|
||
|
database. By default, we'll try to connect through the port `3306`, but this can be overridden with the `--port` option.
|
||
|
|
||
|
### Output
|
||
|
|
||
|
The CLI will create a dump file that looks like `openmetadata_YYYYmmddHHMM_backup.sql`. This will help us identify the
|
||
|
date each backup was generated. We can also specify an output path, which we'll create if it does not exist, via
|
||
|
`--output`.
|
||
|
|
||
|
### Uploading to S3
|
||
|
|
||
|
We currently support uploading the backup files to S3. To run this, make sure to have `AWS_ACCESS_KEY_ID` and
|
||
|
`AWS_SECRET_ACCESS_KEY` as environment variables with permissions to the bucket that you'd like to point to. Afterwards,
|
||
|
we can just use `--upload <endpoint> <bucket> <key>` to have the CLI upload the file. In this case, you'll get both the
|
||
|
local dump file and the one in the cloud.
|
||
|
|
||
|
### mysqldump options
|
||
|
|
||
|
`mysqldump` allows many options when running the command, and some of them might be required in different infrastructures.
|
||
|
The `--options` parameters help us pass to `mysqldump` all of these required options via `-o <opt1>, -o <opt2> [...]`. An
|
||
|
example of this could be the default values we have used for them: `--protocol=tcp` and `--no-tablespaces`, which are
|
||
|
required to run the command pointing to the local Docker container with the database and the default `read-only` user
|
||
|
OpenMetadata provides in the Docker Compose.
|
||
|
|
||
|
### Trying it out
|
||
|
|
||
|
We can do a test locally preparing some containers:
|
||
|
|
||
|
1. `sh docker/run_local_docker.sh` to start the `docker compose` service.
|
||
|
2. `docker run -p 9000:9000 -p 9001:9001 minio/minio server /data --console-address ":9001"` to start minio, an object
|
||
|
storage S3 compatible.
|
||
|
3. Connect to [http://localhost:9001](http://localhost:9001) to reach the minio console and create a bucket
|
||
|
called `my-bucket`
|
||
|
4. Finally, we just need to prepare the environment variables as:
|
||
|
```
|
||
|
export AWS_ACCESS_KEY_ID=minioadmin
|
||
|
export AWS_SECRET_ACCESS_KEY=minioadmin
|
||
|
```
|
||
|
|
||
|
An example CLI call will look as:
|
||
|
|
||
|
```commandline
|
||
|
metadata backup -u openmetadata_user -p openmetadata_password \
|
||
|
-h localhost -d openmetadata_db --output=dir1/dir2 \
|
||
|
--upload http://localhost:9000 my-bucket backup/
|
||
|
```
|
||
|
|
||
|
And we'll get the following output:
|
||
|
|
||
|
```commandline
|
||
|
Creating OpenMetadata backup for localhost:3306/openmetadata_db...
|
||
|
Backup stored locally under dir1/dir2/openmetadata_202201250823_backup.sql
|
||
|
Uploading dir1/dir2/openmetadata_202201250823_backup.sql to http://localhost:9000/my-bucket/backup/openmetadata_202201250823_backup.sql...
|
||
|
```
|
||
|
|
||
|
If we now head to the minio console and check the `my-backup` bucket, we'll see our SQL dump in there.
|
||
|
|
||
|
<Image src="/images/deployment/backup/minio-example.png" alt="minio"/>
|
||
|
|