2021-03-05 00:12:12 -08:00
---
title: "Deploying with Docker"
hide_title: true
---
2019-09-08 20:25:58 -07:00
# Docker Images
2020-08-06 16:38:53 -07:00
## Prerequisites
You need to install [docker ](https://docs.docker.com/install/ ) and
[docker-compose ](https://docs.docker.com/compose/install/ ) (if using Linux; on Windows and Mac compose is included with
Docker Desktop).
Make sure to allocate enough hardware resources for Docker engine. Tested & confirmed config: 2 CPUs, 8GB RAM, 2GB Swap
area.
## Quickstart
2019-12-18 18:57:18 -08:00
The easiest way to bring up and test DataHub is using DataHub [Docker ](https://www.docker.com ) images
2020-06-01 09:36:51 -07:00
which are continuously deployed to [Docker Hub ](https://hub.docker.com/u/linkedin ) with every commit to repository.
2019-09-08 20:25:58 -07:00
2020-08-06 16:38:53 -07:00
You can easily download and run all these images and their dependencies with our
[quick start guide ](../docs/quickstart.md ).
DataHub Docker Images:
2022-07-22 19:43:48 +05:30
Do not use `latest` or `debug` tags for any of the image as those are not supported and present only due to leagcy reasons. Please use `head` or tags specific for versions like `v0.8.40` . For production we recommend using version specific tags not `head` .
* [linkedin/datahub-ingestion ](https://hub.docker.com/r/linkedin/datahub-ingestion/ ) - This contains the Python CLI. If you are looking for docker image for every minor CLI release you can find them under [acryldata/datahub-ingestion ](https://hub.docker.com/r/acryldata/datahub-ingestion/ ).
* [linkedin/datahub-gms ](https://hub.docker.com/repository/docker/linkedin/datahub-gms/ ).
* [linkedin/datahub-frontend-react ](https://hub.docker.com/repository/docker/linkedin/datahub-frontend-react/ )
* [linkedin/datahub-mae-consumer ](https://hub.docker.com/repository/docker/linkedin/datahub-mae-consumer/ )
* [linkedin/datahub-mce-consumer ](https://hub.docker.com/repository/docker/linkedin/datahub-mce-consumer/ )
* [acryldata/datahub-upgrade ](https://hub.docker.com/r/acryldata/datahub-upgrade/ )
* [linkedin/datahub-kafka-setup ](https://hub.docker.com/r/acryldata/datahub-kafka-setup/ )
* [linkedin/datahub-elasticsearch-setup ](https://hub.docker.com/r/linkedin/datahub-elasticsearch-setup/ )
* [acryldata/datahub-mysql-setup ](https://hub.docker.com/r/acryldata/datahub-mysql-setup/ )
* [acryldata/datahub-postgres-setup ](https://hub.docker.com/r/acryldata/datahub-postgres-setup/ )
* [acryldata/datahub-actions ](https://hub.docker.com/r/acryldata/datahub-actions ). Do not use `acryldata/acryl-datahub-actions` as that is deprecated and no longer used.
2019-09-08 20:25:58 -07:00
2020-08-06 16:38:53 -07:00
Dependencies:
2020-08-19 21:39:46 -07:00
* [Kafka, Zookeeper, and Schema Registry ](kafka-setup )
2020-08-08 04:58:55 -07:00
* [Elasticsearch ](elasticsearch-setup )
* [MySQL ](mysql )
2021-06-24 15:44:03 -07:00
* [(Optional) Neo4j ](neo4j )
2019-09-08 20:25:58 -07:00
2020-08-06 16:38:53 -07:00
### Ingesting demo data.
2019-12-16 12:05:42 -08:00
2021-07-29 15:32:23 -07:00
If you want to test ingesting some data once DataHub is up, use the `./docker/ingestion/ingestion.sh` script or `datahub docker ingest-sample-data` . See the [quickstart guide ](../docs/quickstart.md ) for more details.
2019-09-08 20:25:58 -07:00
2020-08-06 16:38:53 -07:00
## Using Docker Images During Development
See [Using Docker Images During Development ](../docs/docker/development.md ).
## Building And Deploying Docker Images
2020-08-08 04:58:55 -07:00
We use GitHub Actions to build and continuously deploy our images. There should be no need to do this manually; a
2020-08-06 16:38:53 -07:00
successful release on Github will automatically publish the images.
### Building images
2020-08-08 04:58:55 -07:00
> This is **not** our recommended development flow and most developers should be following the
2020-09-30 06:20:14 -07:00
> [Using Docker Images During Development](../docs/docker/development.md) guide.
2020-08-08 04:58:55 -07:00
2020-08-06 16:38:53 -07:00
To build the full images (that we are going to publish), you need to run the following:
```
COMPOSE_DOCKER_CLI_BUILD=1 DOCKER_BUILDKIT=1 docker-compose -p datahub build
```
This is because we're relying on builtkit for multistage builds. It does not hurt also set `DATAHUB_VERSION` to
something unique.
2021-02-10 16:10:08 -08:00