mirror of
https://github.com/open-metadata/OpenMetadata.git
synced 2025-08-22 07:58:06 +00:00
MINOR - 1.5.0 docs & update what's new (#17585)
* MINOR - Docs 1.5 upgrade * quickstart * docs * what's new
This commit is contained in:
parent
c5623d0254
commit
6b615f4518
@ -115,25 +115,73 @@ We believe this update will bring greater consistency and clarity to our version
|
||||
|
||||
## 1.5.0
|
||||
|
||||
### Secrets Manager
|
||||
### Multi Owners
|
||||
OpenMetadata allows a single user or a team to be tagged as owners for any data assets. In Release 1.5.0, we allow users to tag multiple individual owners or a single team. This will allow organizations to add ownership to multiple individuals without necessarily needing to create a team around them like previously.
|
||||
|
||||
- Starting with the release 1.5.0, the JWT Token for the bots will be sent to the Secrets Manager if you configured one.
|
||||
It won't appear anymore in your `dag_generated_configs` in Airflow.
|
||||
This is a backward incompatible change, if you are using APIs, please make sure the owner field is now changed to “owners”
|
||||
|
||||
### Python SDK
|
||||
|
||||
- `metadata insight` command removed.
|
||||
Since Data Insights application was moved to be an internal system application instead of relying on external pipelines the SDK command to run the pipeline was removed.
|
||||
### Import/Export Format
|
||||
To support the multi-owner format, we have now changed how we export and import the CSV file in glossary, services, database, schema, table, etc. The new format will be
|
||||
user:userName;team:TeamName
|
||||
|
||||
If you are importing an older file, please make sure to make this change.
|
||||
|
||||
### Pydantic V2
|
||||
The core of OpenMetadata are the JSON Schemas that define the metadata standard. These schemas are automatically translated into Java, Typescript, and Python code with Pydantic classes.
|
||||
|
||||
In this release, we have [migrated](https://docs.pydantic.dev/latest/migration/) the codebase from Pydantic V1 to Pydantic V2.
|
||||
|
||||
### Deployment Related Changes (OSS only)
|
||||
|
||||
`./bootstrap/bootstrap_storage.sh` **removed**
|
||||
|
||||
OpenMetadata community has built rolling upgrades to database schema and the data to make upgrades easier. This tool is now called as ./bootstrap/openmetadata-ops.sh and has been part of our releases since 1.3. The `bootstrap_storage.sh` doesn’t support new native schemas in OpenMetadata. Hence, we have deleted this tool from this release.
|
||||
|
||||
While upgrading, please refer to our Upgrade Notes in the documentation. Always follow the best practices provided there.
|
||||
|
||||
### Database Connection Pooling
|
||||
|
||||
- OpenMetadata leverages Jdbi for database operations like reading, writing, and deleting. In this release, we've introduced new configuration options to enhance connection pooling, enabling more efficient database usage, especially in resource-constrained environments.
|
||||
OpenMetadata uses Jdbi to handle database-related operations such as read/write/delete. In this release, we introduced additional configs to help with connection pooling, allowing the efficient use of a database with low resources.
|
||||
|
||||
If your cluster operates at a large scale, consider updating these default settings to optimize connection management.
|
||||
Please update the defaults if your cluster is running at a large scale to scale up the connections efficiently.
|
||||
|
||||
For detailed information on the new configurations, please refer to the [documentation](/deployment/database-connection-pooling).
|
||||
For the new configuration, please refer to the [doc](https://docs.open-metadata.org/latest/deployment/database-connection-pooling) here
|
||||
|
||||
### UI
|
||||
### Data Insights
|
||||
|
||||
### API
|
||||
The Data Insights application is meant to give you a quick glance at your data's state and allow you to take action based on the information you receive. To continue pursuing this objective, the application was completely refactored to allow customizability.
|
||||
|
||||
Part of this refactor was making Data Insights an internal application, no longer relying on an external pipeline. This means triggering Data Insights from the Python SDK will no longer be possible.
|
||||
|
||||
With this change you will need to run a backfill on the Data Insights for the last couple of days since the Data Assets data changed.
|
||||
|
||||
### UI Changes
|
||||
|
||||
#### New Explore Page
|
||||
|
||||
Explore page displays hierarchically organized data assets by grouping them into `services > database > schema > tables/stored procedures`. This helps users organically find the data asset they are looking for based on a known database or schema they were using. This is a new feature and changes the way the Explore page was built in previous releases.
|
||||
|
||||
#### Connector Schema Changes
|
||||
|
||||
In the latest release, several updates and enhancements have been made to the JSON schema across various connectors. These changes aim to improve security, configurability, and expand integration capabilities. Here's a detailed breakdown of the updates:
|
||||
|
||||
- **KafkaConnect**: Added `schemaRegistryTopicSuffixName` to enhance topic configuration flexibility for schema registries.
|
||||
- **GCS Datalake**: Introduced `bucketNames` field, allowing users to specify targeted storage buckets within the Google Cloud Storage environment.
|
||||
- **OpenLineage**: Added `saslConfig` to enhance security by enabling SASL (Simple Authentication and Security Layer) configuration.
|
||||
- **Salesforce**: Added sslConfig to strengthen the security layer for Salesforce connections by supporting SSL.
|
||||
- **DeltaLake**: Updated schema by moving metastoreConnection to a newly created `metastoreConfig.json` file. Additionally, introduced `configSource` to better define source configurations, with new support for `metastoreConfig.json` and `storageConfig.json`.
|
||||
- **Iceberg RestCatalog**: Removed clientId and `clientSecret` as mandatory fields, making the schema more flexible for different authentication methods.
|
||||
- **DBT Cloud Pipelines**: Added as a new connector to support cloud-native data transformation workflows using DBT.
|
||||
- **Looker**: Expanded support to include connections using GitLab integration, offering more flexible and secure version control.
|
||||
- **Tableau**: Enhanced support by adding capabilities for connecting with `TableauPublishedDatasource` and `TableauEmbeddedDatasource`, providing more granular control over data visualization and reporting.
|
||||
|
||||
### Include DDL
|
||||
During the Database Metadata ingestion, we can optionally pick up the DDL for both tables and views. During the metadata ingestion, we use the view DDLs to generate the View Lineage.
|
||||
|
||||
To reduce the processing time for out-of-the-box workflows, we are disabling the include DDL by default, whereas before, it was enabled, which potentially led to long-running workflows.
|
||||
|
||||
### Secrets Manager
|
||||
Starting with the release 1.5.0, the JWT Token for the bots will be sent to the Secrets Manager if you configured one. It won't appear anymore in your dag_generated_configs in Airflow.
|
||||
|
||||
### Python SDK
|
||||
The `metadata insight` command has been removed. Since Data Insights application was moved to be an internal system application instead of relying on external pipelines the SDK command to run the pipeline was removed.
|
||||
|
@ -1,7 +1,7 @@
|
||||
# 1.4.3 Release 🎉
|
||||
# 1.5.0 Release 🎉
|
||||
|
||||
{% note noteType="Tip" %}
|
||||
**June 15th, 2024**
|
||||
{% note noteType="Tip" %}
|
||||
**Aug 26th, 2024**
|
||||
{% /note %}
|
||||
|
||||
{% inlineCalloutContainer %}
|
||||
@ -10,13 +10,171 @@ color="violet-70"
|
||||
icon="celebration"
|
||||
bold="Upgrade OpenMetadata"
|
||||
href="/deployment/upgrade" %}
|
||||
Learn how to upgrade your OpenMetadata instance to 1.4.3!
|
||||
Learn how to upgrade your OpenMetadata instance to 1.5.0!
|
||||
{% /inlineCallout %}
|
||||
{% /inlineCalloutContainer %}
|
||||
|
||||
You can find the GitHub release [here](https://github.com/open-metadata/OpenMetadata/releases/tag/1.4.3-release).
|
||||
You can find the GitHub release [here](https://github.com/open-metadata/OpenMetadata/releases/tag/1.5.0-release).
|
||||
|
||||
## Improvements
|
||||
# Backward Incompatible Changes
|
||||
|
||||
- Fixed User Signup Flow Issue missing authorize.
|
||||
- Fixed vulnerabilities for azure-identity and msal4j.
|
||||
## Multi Owners
|
||||
OpenMetadata allows a single user or a team to be tagged as owners for any data assets. In Release 1.5.0, we allow users to tag multiple individual owners or a single team. This will allow organizations to add ownership to multiple individuals without necessarily needing to create a team around them like previously.
|
||||
|
||||
This is a backward incompatible change, if you are using APIs, please make sure the owner field is now changed to “owners”
|
||||
|
||||
## Import/Export Format
|
||||
To support the multi-owner format, we have now changed how we export and import the CSV file in glossary, services, database, schema, table, etc. The new format will be
|
||||
user:userName;team:TeamName
|
||||
|
||||
If you are importing an older file, please make sure to make this change.
|
||||
|
||||
## Pydantic V2
|
||||
The core of OpenMetadata are the JSON Schemas that define the metadata standard. These schemas are automatically translated into Java, Typescript, and Python code with Pydantic classes.
|
||||
|
||||
In this release, we have [migrated](https://docs.pydantic.dev/latest/migration/) the codebase from Pydantic V1 to Pydantic V2.
|
||||
|
||||
## Deployment Related Changes (OSS only)
|
||||
|
||||
`./bootstrap/bootstrap_storage.sh` **removed**
|
||||
|
||||
OpenMetadata community has built rolling upgrades to database schema and the data to make upgrades easier. This tool is now called as ./bootstrap/openmetadata-ops.sh and has been part of our releases since 1.3. The `bootstrap_storage.sh` doesn’t support new native schemas in OpenMetadata. Hence, we have deleted this tool from this release.
|
||||
|
||||
While upgrading, please refer to our Upgrade Notes in the documentation. Always follow the best practices provided there.
|
||||
|
||||
## Database Connection Pooling
|
||||
|
||||
OpenMetadata uses Jdbi to handle database-related operations such as read/write/delete. In this release, we introduced additional configs to help with connection pooling, allowing the efficient use of a database with low resources.
|
||||
|
||||
Please update the defaults if your cluster is running at a large scale to scale up the connections efficiently.
|
||||
|
||||
For the new configuration, please refer to the [doc](https://docs.open-metadata.org/latest/deployment/database-connection-pooling) here
|
||||
|
||||
## Data Insights
|
||||
|
||||
The Data Insights application is meant to give you a quick glance at your data's state and allow you to take action based on the information you receive. To continue pursuing this objective, the application was completely refactored to allow customizability.
|
||||
|
||||
Part of this refactor was making Data Insights an internal application, no longer relying on an external pipeline. This means triggering Data Insights from the Python SDK will no longer be possible.
|
||||
|
||||
With this change you will need to run a backfill on the Data Insights for the last couple of days since the Data Assets data changed.
|
||||
|
||||
## UI Changes
|
||||
|
||||
### New Explore Page
|
||||
|
||||
Explore page displays hierarchically organized data assets by grouping them into `services > database > schema > tables/stored procedures`. This helps users organically find the data asset they are looking for based on a known database or schema they were using. This is a new feature and changes the way the Explore page was built in previous releases.
|
||||
|
||||
### Connector Schema Changes
|
||||
|
||||
In the latest release, several updates and enhancements have been made to the JSON schema across various connectors. These changes aim to improve security, configurability, and expand integration capabilities. Here's a detailed breakdown of the updates:
|
||||
|
||||
- **KafkaConnect**: Added `schemaRegistryTopicSuffixName` to enhance topic configuration flexibility for schema registries.
|
||||
- **GCS Datalake**: Introduced `bucketNames` field, allowing users to specify targeted storage buckets within the Google Cloud Storage environment.
|
||||
- **OpenLineage**: Added `saslConfig` to enhance security by enabling SASL (Simple Authentication and Security Layer) configuration.
|
||||
- **Salesforce**: Added sslConfig to strengthen the security layer for Salesforce connections by supporting SSL.
|
||||
- **DeltaLake**: Updated schema by moving metastoreConnection to a newly created `metastoreConfig.json` file. Additionally, introduced `configSource` to better define source configurations, with new support for `metastoreConfig.json` and `storageConfig.json`.
|
||||
- **Iceberg RestCatalog**: Removed clientId and `clientSecret` as mandatory fields, making the schema more flexible for different authentication methods.
|
||||
- **DBT Cloud Pipelines**: Added as a new connector to support cloud-native data transformation workflows using DBT.
|
||||
- **Looker**: Expanded support to include connections using GitLab integration, offering more flexible and secure version control.
|
||||
- **Tableau**: Enhanced support by adding capabilities for connecting with `TableauPublishedDatasource` and `TableauEmbeddedDatasource`, providing more granular control over data visualization and reporting.
|
||||
|
||||
## Include DDL
|
||||
During the Database Metadata ingestion, we can optionally pick up the DDL for both tables and views. During the metadata ingestion, we use the view DDLs to generate the View Lineage.
|
||||
|
||||
To reduce the processing time for out-of-the-box workflows, we are disabling the include DDL by default, whereas before, it was enabled, which potentially led to long-running workflows.
|
||||
|
||||
## Secrets Manager
|
||||
Starting with the release 1.5.0, the JWT Token for the bots will be sent to the Secrets Manager if you configured one. It won't appear anymore in your dag_generated_configs in Airflow.
|
||||
|
||||
## Python SDK
|
||||
The `metadata insight` command has been removed. Since Data Insights application was moved to be an internal system application instead of relying on external pipelines the SDK command to run the pipeline was removed.
|
||||
|
||||
# What's New
|
||||
|
||||
## Data Observability with Anomaly Detection (Collate)
|
||||
|
||||
OpenMetadata has been driving innovation in Data Quality in Open Source. Many organizations are taking advantage of the following Data Quality features to achieve better-quality data
|
||||
|
||||
1. A Native Profiler to understand the shape of the data, freshness, completeness, volume, and ability to add your own metrics, including column level profiler over time-series and dashboards
|
||||
2. No-code data quality tests, deploy, collect results back to see it in a dashboard all within OpenMetadata
|
||||
3. Create alerts and get notified of Test results through email, Slack, NSteams, GChat, and Webhook
|
||||
4. Incident Manager to collaborate around test failures and visibility to downstream consumers of failures from upstream
|
||||
|
||||
In 1.5.0, we are bringing in **Anomaly Detection** based on AI to predict when an anomaly happens based on our learning historical data and automatically sending notifications to the owners of the table to warn them of the impending incidents
|
||||
|
||||
{% youtube videoId="BPuNC8vPcsw" start="0:00" end="1:01" width="560px" height="315px" /%}
|
||||
|
||||
## Enhanced Data Quality Dashboard (Collate)
|
||||
|
||||
We also have improved the Table Data quality dashboard to showcase the tests categorized and make it easy for everyone to consume. When there are issues, the new dashboard makes it easier to understand the Data Quality coverage of your tables and the possible impact each test failure has by organizing tests into different groups.
|
||||
|
||||
{% youtube videoId="bXcQBtZuyoU" start="0:00" end="2:10" width="560px" height="315px" /%}
|
||||
|
||||
## Freshness Data Quality Tests (Collate)
|
||||
Working with old data can lead to making wrong decisions. With the new Freshness test, you can validate that your data arrives at the right time. Freshness tests are a critical part of any data team's toolset. Bringing these tests together with lineage information and the Incident Manager, your team will be able to quickly detect issues related to missing data or stuck pipelines.
|
||||
|
||||
{% youtube videoId="QRcR3m9cCGo" start="0:00" end="1:09" width="560px" height="315px" /%}
|
||||
|
||||
## Data Diff Data Quality Tests
|
||||
Data quality checks are important not only within a single table but also between different tables. These data diff checks can ensure key data remains unchanged after transformation, or conversely, ensure that the transformations were actually performed.
|
||||
|
||||
We are introducing the **table difference data quality test** to validate that multiple appearances of the same information remain consistent. Note that the test allows you to specify which column to use as a key and which columns you want to compare, and even add filters in the data to give you more control over multiple use cases.
|
||||
|
||||
{% youtube videoId="oxZVS_UGrE4" start="0:00" end="2:22" width="560px" height="315px" /%}
|
||||
|
||||
## Domains RBAC & Subdomains
|
||||
OpenMetadata introduced Domains & Data Products in 1.3.0. Since then, many large organizations have started using Domains & Data Products to achieve better ownership and collaboration around domains that can span multiple teams.
|
||||
|
||||
In the 1.5.0 release, we added support for subdomains. This will help teams to organize into multiple subdomains within each domain.
|
||||
|
||||
### RBAC for Domains
|
||||
With the 1.5.0 release, we are adding more stricter controls around Domain. Now, teams, data assets, glossaries, and classification can have domain concepts and can get a policy such that only users within a domain can access the data within a domain. Domain owners can use Data Products to publish data products and showcase publicly available data assets from a specific domain.
|
||||
|
||||
This will help large companies to use a single OpenMetadata platform to unify all of their data and teams but also provide more stringent controls to segment the data between domains
|
||||
|
||||
{% youtube videoId="r-_HaewjgTQ" start="0:00" end="0:44" width="560px" height="315px" /%}
|
||||
|
||||
## Improved Explore Page & Data Asset Widget
|
||||
OpenMetadata, with its simple UI/UX and data collaboration features, is becoming more attractive to non-technical users as well. Data Governance teams are using OpenMetadata to add glossary terms and policies around metadata. Teams using Collate SaaS product are taking advantage of our Automations feature to gain productivity in their governance tasks.
|
||||
|
||||
Our new improved navigation on the Explore page will help users navigate hierarchically and find the data they are looking for. Users will see the data assets now grouped by `service name -> database -> schema -> tables/stored procedures`.
|
||||
|
||||
We are also making the discovery of data more accessible for users introducing a data asset widget, which will group the assets by platform type. This will help users find their data if they are working on a specific platform such as Looker or Snowflake they can easily click on that icon and get to the data.
|
||||
|
||||
{% youtube videoId="45ekUIRO1Ec" start="0:00" end="1:11" width="560px" height="315px" /%}
|
||||
|
||||
## Pipeline Status Widget
|
||||
We are also adding another widget you can use to customize the Landing Page of the User Personas in your organization.
|
||||
|
||||
With the Pipeline Status widget, Data Engineers can easily track the pipelines that are not behaving as expected. This widget, together with the obervability alerts that are already in place, will help your teams jump even faster to solving any issues in the platform.
|
||||
|
||||
## API as Data Asset
|
||||
The Internet runs using APIs, both producing and consuming data. Organizations today run many microservices and REST APIs to capture data from their users and update a transaction database in the backend.
|
||||
|
||||
On top of the many supported connectors across Databases, Dashboards, ML Models, etc. We believe that providing support for API Services as data assets will help to get the full picture of how the data is coming through from various services and landing into databases, going to warehouses and BI tools.
|
||||
|
||||
In 1.5.0 we are introducing APIs as another first-class entity. Teams can now capture API requests and responses payloads and use our column level lineage to capture the relation between APIs and any other asset in the platform.
|
||||
|
||||
{% youtube videoId="b9wrVnM3u80" start="0:00" end="0:33" width="560px" height="315px" /%}
|
||||
|
||||
## Glossary Improvements
|
||||
OpenMetadata supports multiple glossaries, an import/export and review process, and bulk asset tagging with glossary terms. Many teams are taking advantage of these features, and with an amazing open-source community, we are receiving great feedback on improving glossary functionality.
|
||||
|
||||
Here are some of the improvements coming in 1.5.0:
|
||||
1. Glossary Reviewers can be teams
|
||||
2. Updating a glossary will enforce a re-review
|
||||
3. Renaming the Glossary Term while it's under review will keep the task associated with it open
|
||||
|
||||
## Data Insights (Collate)
|
||||
The Data Insights application is meant to give you a quick glance of your data's state and allow you to take action based on the information you receive.
|
||||
|
||||
To continue pursuing this objective, the application was completely refactored to allow customizability. This is achieved by the possibility of now creating custom dashboards. On this release you can create charts based on your data assets metadata based on your needs.
|
||||
|
||||
## Ingestion Connectors
|
||||
80+ connectors to help teams to centralize metadata. We continue to push the boundaries of this mission, in
|
||||
- **Apache Flink** as a Pipeline Connector
|
||||
- **SAP ERP**, after a long and successful collaboration with our community and SAP experts
|
||||
- **Teradata** as a community contribution from [gbpy](https://github.com/gpby) to broaden the integration capabilities for enterprise-scale analytics and data management.
|
||||
- **GCS Storage Connector** as a community contribution from [Matt Chamberlin](https://github.com/MChamberlin)
|
||||
|
||||
**Full Changelog**: https://github.com/open-metadata/OpenMetadata/compare/1.4.0-release...1.5.0-release
|
||||
|
@ -22,7 +22,7 @@ This is a narrowed-down view of the database schema, focusing on the main tables
|
||||
you will find the columns that are `GENERATED`:
|
||||
|
||||
{% image
|
||||
src="/images/v1.5/main-concepts/db-diagram/db-diagram.png"
|
||||
src="/images/v1.5/main-concepts/backend-db/db-diagram.png"
|
||||
alt="Backend DB Simplified Diagram"
|
||||
caption="Backend DB Simplified Diagram"
|
||||
/%}
|
||||
|
@ -119,15 +119,15 @@ The latest version is at the top of the page
|
||||
You can use the curl or wget command as well to fetch the docker compose files from your terminal -
|
||||
|
||||
```commandline
|
||||
curl -sL -o docker-compose.yml https://github.com/open-metadata/OpenMetadata/releases/download/1.4.1-release/docker-compose.yml
|
||||
curl -sL -o docker-compose.yml https://github.com/open-metadata/OpenMetadata/releases/download/1.5.0-release/docker-compose.yml
|
||||
|
||||
curl -sL -o docker-compose-postgres.yml https://github.com/open-metadata/OpenMetadata/releases/download/1.4.1-release/docker-compose-postgres.yml
|
||||
curl -sL -o docker-compose-postgres.yml https://github.com/open-metadata/OpenMetadata/releases/download/1.5.0-release/docker-compose-postgres.yml
|
||||
```
|
||||
|
||||
```commandline
|
||||
wget https://github.com/open-metadata/OpenMetadata/releases/download/1.4.1-release/docker-compose.yml
|
||||
wget https://github.com/open-metadata/OpenMetadata/releases/download/1.5.0-release/docker-compose.yml
|
||||
|
||||
wget https://github.com/open-metadata/OpenMetadata/releases/download/1.4.1-release/docker-compose-postgres.yml
|
||||
wget https://github.com/open-metadata/OpenMetadata/releases/download/1.5.0-release/docker-compose-postgres.yml
|
||||
```
|
||||
|
||||
### 3. Start the Docker Compose Services
|
||||
@ -166,10 +166,10 @@ You can validate that all containers are up by running with command `docker ps`.
|
||||
```commandline
|
||||
❯ docker ps
|
||||
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
|
||||
470cc8149826 openmetadata/server:1.4.1 "./openmetadata-star…" 45 seconds ago Up 43 seconds 3306/tcp, 9200/tcp, 9300/tcp, 0.0.0.0:8585-8586->8585-8586/tcp openmetadata_server
|
||||
63578aacbff5 openmetadata/ingestion:1.4.1 "./ingestion_depende…" 45 seconds ago Up 43 seconds 0.0.0.0:8080->8080/tcp openmetadata_ingestion
|
||||
470cc8149826 openmetadata/server:1.5.0 "./openmetadata-star…" 45 seconds ago Up 43 seconds 3306/tcp, 9200/tcp, 9300/tcp, 0.0.0.0:8585-8586->8585-8586/tcp openmetadata_server
|
||||
63578aacbff5 openmetadata/ingestion:1.5.0 "./ingestion_depende…" 45 seconds ago Up 43 seconds 0.0.0.0:8080->8080/tcp openmetadata_ingestion
|
||||
9f5ee8334f4b docker.elastic.co/elasticsearch/elasticsearch:7.16.3 "/tini -- /usr/local…" 45 seconds ago Up 44 seconds 0.0.0.0:9200->9200/tcp, 0.0.0.0:9300->9300/tcp openmetadata_elasticsearch
|
||||
08947ab3424b openmetadata/db:1.4.1 "/entrypoint.sh mysq…" 45 seconds ago Up 44 seconds (healthy) 3306/tcp, 33060-33061/tcp openmetadata_mysql
|
||||
08947ab3424b openmetadata/db:1.5.0 "/entrypoint.sh mysq…" 45 seconds ago Up 44 seconds (healthy) 3306/tcp, 33060-33061/tcp openmetadata_mysql
|
||||
```
|
||||
|
||||
In a few seconds, you should be able to access the OpenMetadata UI at [http://localhost:8585](http://localhost:8585)
|
||||
|
@ -14,6 +14,316 @@ version. To see what's coming in next releases, please check our [Roadmap](/rele
|
||||
|
||||
{% partial file="/v1.5/releases/latest.md" /%}
|
||||
|
||||
# 1.4.8 Release
|
||||
|
||||
{% note noteType="Tip" %}
|
||||
**August 6th, 2024**
|
||||
{% /note %}
|
||||
|
||||
- Make `Include ddl` disabled by default
|
||||
- Made DDL configuration consistent with views
|
||||
- Fix user profile task listing.
|
||||
- Fix import/export UI flow ${CollateIconWithLinkMD}.
|
||||
- Improve SAML logging backend.
|
||||
- Add Unity Catalog Lineage Dialect.
|
||||
- Clean idle connections during ingestion.
|
||||
- Fix Databricks Describe Table during metadata ingestion.
|
||||
- Glossary Upload now shows permissions errors for non-owners.
|
||||
- Fix task not showing in the right panel after clicking, in the Activity Feed.
|
||||
- Support multiple dbt run_results.json for a single manifest for improved performance.
|
||||
|
||||
# 1.4.7 Release
|
||||
|
||||
{% note noteType="Tip" %}
|
||||
**August 6th, 2024**
|
||||
{% /note %}
|
||||
|
||||
You can find the GitHub release [here](https://github.com/open-metadata/OpenMetadata/releases/tag/1.4.7-release).
|
||||
|
||||
## Improvements
|
||||
|
||||
- Resolved issue with Azure login related to email principal claims.
|
||||
|
||||
# 1.4.6 Release
|
||||
|
||||
{% note noteType="Tip" %}
|
||||
**August 2nd, 2024**
|
||||
{% /note %}
|
||||
|
||||
You can find the GitHub release [here](https://github.com/open-metadata/OpenMetadata/releases/tag/1.4.6-release).
|
||||
|
||||
## Improvements
|
||||
|
||||
- Fix lineage PATCH API for ingestion.
|
||||
- Fix Trino Azure config secret masking.
|
||||
- Fix setuptools version due to yanked setuptools release.
|
||||
- Fix MSSQL busy connection error during test connection.
|
||||
- Fixed test case summary updates.
|
||||
- Fixed Test Suite indexing.
|
||||
- Fix repeated alerts being sent after no changes in the Entity.
|
||||
- Fixed an issue handling users with capital letters.
|
||||
- Centralize OIDC flow handling.
|
||||
- Fixed Ingestion Pipeline alert URL.
|
||||
|
||||
# 1.4.5 Release
|
||||
|
||||
{% note noteType="Tip" %}
|
||||
**July 19th, 2024**
|
||||
{% /note %}
|
||||
|
||||
You can find the GitHub release [here](https://github.com/open-metadata/OpenMetadata/releases/tag/1.4.5-release).
|
||||
|
||||
## Improvements
|
||||
|
||||
- Improve query filtering with prepared statements.
|
||||
- Bug fix in regex to match test case when using sampled data.
|
||||
- Bug fix in global profiler config for Snowflake, Redshift, and BigQuery.
|
||||
- Bug fix for Arg mismatch for DataModels in QlikSense.
|
||||
|
||||
# 1.4.4 Release
|
||||
|
||||
{% note noteType="Tip" %}
|
||||
**July 4th, 2024**
|
||||
{% /note %}
|
||||
|
||||
You can find the GitHub release [here](https://github.com/open-metadata/OpenMetadata/releases/tag/1.4.4-release).
|
||||
|
||||
## Improvements
|
||||
|
||||
- Introduced SSL for Salesforce
|
||||
- Fixed the claim mappings and usernames
|
||||
- Fixed issues in Salesforce connector
|
||||
- FIxed issues in Alation connector
|
||||
- Verified for changes in new env for claim mapping
|
||||
|
||||
# 1.4.3 Release
|
||||
|
||||
{% note noteType="Tip" %}
|
||||
**June 15th, 2024**
|
||||
{% /note %}
|
||||
|
||||
You can find the GitHub release [here](https://github.com/open-metadata/OpenMetadata/releases/tag/1.4.3-release).
|
||||
|
||||
## Improvements
|
||||
|
||||
- Fixed User Signup Flow Issue missing authorize.
|
||||
- Fixed vulnerabilities for azure-identity and msal4j.
|
||||
|
||||
# 1.4.2 Release
|
||||
|
||||
{% note noteType="Tip" %}
|
||||
**June 10th, 2024**
|
||||
{% /note %}
|
||||
|
||||
You can find the GitHub release [here](https://github.com/open-metadata/OpenMetadata/releases/tag/1.4.2-release).
|
||||
|
||||
## Enhancements
|
||||
|
||||
- In OpenMetadata, we support connecting the data assets to the knowledge articles. The knowledge articles that are pulled from the Alation connector have image URLs. We have enhanced the Alation connector to download and display the images in the Knowledge Articles.
|
||||
- Test cases can now be filtered by Service, Tag, and Tier.
|
||||
|
||||
## Changes
|
||||
|
||||
- One team or multiple users can be selected as reviewers for a Glossary term.,
|
||||
- Updated the openmetadata.yaml to remove WebAnalyticsHandler.,
|
||||
- Add appType as part of the schema in the ingestion pipeline.,
|
||||
- We now sanitize the Activity Feed editor content.
|
||||
|
||||
## Improvements
|
||||
|
||||
- Fixed the lineage view for tables with many columns.
|
||||
- Fixed an issue with updating the lineage edge descriptions.
|
||||
- Fixed an issue with Null Schema Field.
|
||||
- Fixed the glossary term review process issues.
|
||||
- Fixed the Kafka SSL connection arguments.
|
||||
- Fixed an issue with dbt ingestion pipeline that was occurring due to non enum values.
|
||||
- Fixed an issue with Announcements.
|
||||
- Fixed redirection issues for Tags and Glossary Terms.
|
||||
- Fixed a minor issue with filtering the Profiler.
|
||||
- Fixed the registration Issue with Event Handlers.
|
||||
- Fixed the sign-in issues with SAML.
|
||||
- Fixed issues with partition migration with Redshift services.
|
||||
- Fixed an issue with the Quicksight connector.
|
||||
- Fixed some minor issues with the user Profile page.
|
||||
- Fixed some issues with the Teams page.
|
||||
|
||||
# 1.4.1 Release
|
||||
|
||||
{% note noteType="Tip" %}
|
||||
**May 27th, 2024**
|
||||
{% /note %}
|
||||
|
||||
You can find the GitHub release [here](https://github.com/open-metadata/OpenMetadata/releases/tag/1.4.1-release).
|
||||
|
||||
In 1.4.1, we provide migration fixes on top of the 1.4.0 release. Do check out the 1.4.0 release highlights below.
|
||||
|
||||
# 1.4.0 Release 🎉
|
||||
|
||||
{% note noteType="Tip" %}
|
||||
**May 21th, 2024**
|
||||
|
||||
[OpenMetadata 1.4.0 Release](https://blog.open-metadata.org/openmetadata-release-1-4-0-f6fb11ec34d7)
|
||||
{% /note %}
|
||||
|
||||
You can find the GitHub release [here](https://github.com/open-metadata/OpenMetadata/releases/tag/1.4.0-release).
|
||||
|
||||
## Backward Incompatible Changes
|
||||
|
||||
### Tooling
|
||||
- Metadata Backup/Recovery is deprecated. No further support will be provided.
|
||||
- Users are advised to use database native tools to backup and store it in their object store for recovery.
|
||||
- `bootstrap/bootstrap_storage.sh` has been deprecated in favor of bootstrap/openmetadata-ops.sh
|
||||
|
||||
### UI
|
||||
- Activity has been improved. New update specific cards display critical information such as data quality test case updates, description, tag update or removal.
|
||||
- For Lineage, the Expand All button has been removed. A new Layers button is introduced at the bottom left corner. With the Layers button, you can add Column Level Lineage or Data Observability details to your Lineage view.
|
||||
- View Definition is now renamed as Schema Definition.
|
||||
- Adding Glossary Term view is improved. Now we show glossary terms hierarchically enabling a better understanding of how the terms are setup while adding it to a table or dashboard.
|
||||
- For Classification, users can set classification to be mutually exclusive only at the time of creation. Once created, you cannot change it back to mutually non-exclusive or vice-versa. This is to prevent conflicts of adding multiple tags that belong to same classification and later turning the mutually exclusive flag back to true.
|
||||
|
||||
### API
|
||||
- Table Schema's `ViewDefinition` is now renamed to `SchemaDefinition` to capture Tables' Create Schema.
|
||||
- Bulk Import API now creates entities if they are not present during the import.
|
||||
- Table's TestSuite is migrated to EntityReference. Previously it used to store entire payload of TestSuite.
|
||||
|
||||
## Automator (Collate only)
|
||||
|
||||
{% youtube videoId="zdh4yzHw4w0" start="0:00" end="2:29" width="560px" height="315px" /%}
|
||||
|
||||
- Easily maintain high-quality metadata at scale with automations. The Automator streamlines governance processes from ownership assignments to tagging, ensuring compliance and consistency.
|
||||
- You can update the properties of your assets by filtering by service, owner, domain, or any other supported property from the advanced search.
|
||||
- Easily see which assets have been selected by jumping to the Explore page in one click.
|
||||
- For tables, data models, topics, and search indexes, you can apply the action to their columns or fields.
|
||||
- We added support for the following actions: adding and removing owner, tier, domain, tags, glossary terms and descriptions, ML PII tagging, and propagation of tags and glossary terms through lineage.
|
||||
|
||||
## Bulk Upload Data Assets (Collate only)
|
||||
|
||||
{% youtube videoId="CXxDdS6AifY" start="0:00" end="2:19" width="560px" height="315px" /%}
|
||||
|
||||
- Bulk upload/download database, schema, and table entities from/into a CSV file for quick edition or creation.
|
||||
- Supports an inline editor to validate/update assets before performing the upload.
|
||||
- APIs are available in OSS.
|
||||
|
||||
## Data Quality Improvements
|
||||
{% youtube videoId="UNOHvBMVcYM" start="0:00" end="1:28" width="560px" height="315px" /%}
|
||||
|
||||
- The Table schema page now shows the Data Quality tests for each column.
|
||||
- Improved filtering options for test suite and test cases.
|
||||
- We have improved how the UI fetches the Data Quality details for improved performance.
|
||||
- We now compute Unique and Count in the same query to avoid inconsistency due to the high frequency of data insertion.
|
||||
- Fixed the issue with removing the test case description upon the test case display name change.
|
||||
- Support has been added for an empty string as a missing count.
|
||||
|
||||
## Data Profiler
|
||||
- Implemented a global profiler configuration page, allowing admin to exclude certain metric computations for specific data types.
|
||||
- Added profiler support for Redshift complex types and DynamoDB.
|
||||
- Fixed an issue with performing sum operations for large values in profiler ingestion.
|
||||
- Fixed the histogram unit's issues with scientific notation.
|
||||
|
||||
## Incident Manager
|
||||
- We now display a sample of failed rows for the latest failed test cases. Once the issue is resolved, the failed sample will be deleted. (Collate Only)
|
||||
- Fixed the Date time filter for the Incident Manager.
|
||||
- Notifications are sent for the tasks created by the Incident Manager.
|
||||
|
||||
## Lineage Improvements
|
||||
https://www.youtube.com/watch?v=KZdVb8DiHJs - Video on Column Lineage Search
|
||||
|
||||
- Column Lineage Search
|
||||
|
||||
{% youtube videoId="KZdVb8DiHJs" start="0:00" end="0:30" width="560px" height="315px" /%}
|
||||
|
||||
- Lineage Layers
|
||||
|
||||
{% youtube videoId="wtBMeLvA6Sw" start="0:00" end="0:43" width="560px" height="315px" /%}
|
||||
|
||||
- OpenMetadata already supports Column-level lineage, and now we have introduced Task-level lineage for Pipelines, Chart-level lineage for Dashboards, Feature-level lineage for ML Models, Field-level lineage for Topics, and columns for dashboard Data Models.
|
||||
- Automated column-level lineage is now supported for Tableau, Superset, QlikCloud, and QlikSense between Data Models and Tables.
|
||||
- The child nodes in a lineage graph are sorted in alphabetical order.
|
||||
- Improved the log of failed-to-parse queries.
|
||||
- Fixed an issue with automated column-level lineage overwriting the pipeline lineage and manual column lineage.
|
||||
- Snowflake & Databricks now support automated lineage between external tables and their origin storage container.
|
||||
- Lineage can be exported as a CSV file.
|
||||
- OpenMetadata spark agent now supports automated lineage between tables and their origin storage container.
|
||||
- Fixed an issue with parsing lineage queries for Redshift.
|
||||
- Now, we support pipeline as an edge between any two entity types.
|
||||
- We now parse PowerBi DAX files for lineage.
|
||||
- Support has been added for dynamic tables.
|
||||
|
||||
## Data Insights
|
||||
- Previously, the data insights reports displayed only the percentage coverage of ownership and description. Now, users can drill down to view the data assets with no owner or description.
|
||||
- Improved the UX for data insight filters.
|
||||
|
||||
## Cost Analysis (Collate Only)
|
||||
- Lifecycle data for Cost Analysis has been implemented for BigQuery, Snowflake, and Redshift.
|
||||
|
||||
## Custom Theme
|
||||
|
||||
{% youtube videoId="-NiU1flBHs0" start="0:00" end="1:02" width="560px" height="315px" /%}
|
||||
|
||||
- Previously supported adding logo, monogram, and favicon to your OpenMetadata instance.
|
||||
- Now, it supports customizing the theme with colors to suit your company branding.
|
||||
|
||||
## Landing Page Widgets (Collate Only)
|
||||
|
||||
{% youtube videoId="Kakfa-lYGOU" start="0:00" end="0:59" width="560px" height="315px" /%}
|
||||
|
||||
- A widget was added to list the pipelines that belong to a user or their team.
|
||||
- Added a Data Quality Widget to list the summary of data quality tests belonging to a user or their team.
|
||||
|
||||
## Ingestion Performance Improvements
|
||||
- Bigquery, Redshift, and Snowflake now support incremental metadata ingestions by scanning DML operations on the query history.
|
||||
- Database Services now support parallelizing the metadata ingestion at each schema.
|
||||
|
||||
## Connectors
|
||||
- Now supports a new connector for QlikCloud.
|
||||
- New Kafka Connect connector
|
||||
- We now parse complex protobuf schemas for Kafka
|
||||
- Improved model storage ingestion for Sagemaker and Mlflow.
|
||||
- Added an option to include or exclude drafts from dashboards.
|
||||
- Added an option to include or exclude paused pipelines in Airflow.
|
||||
- Revamped SSL support to allow users to upload the required certificates directly in the UI.
|
||||
- The character support has been enhanced for tag ingestion to include /.
|
||||
- In the Oracle connector, we rolled back to use all_ tables instead of dba_.
|
||||
- Added support for Azure auth in Trino.
|
||||
- For QlikSense, we have added an option to disable SSL validation.
|
||||
|
||||
## Custom Properties
|
||||
|
||||
{% youtube videoId="lZoSeKkErBk" start="0:00" end="1:07" width="560px" height="315px" /%}
|
||||
|
||||
- Custom Properties now allow linking other assets in the platform, such as Tables, Dashboards, etc. To enable this, create a Custom Property as an Entity Reference or Entity Reference List.
|
||||
|
||||
## Glossary
|
||||
- The glossary term parent can now be changed from the Details page.
|
||||
- On the data assets page, glossary terms are displayed by hierarchy.
|
||||
|
||||
## Alerts & Notification Improvements
|
||||
- The Activity Feed provides more contextual information, removing the need to move to entity pages.
|
||||
- Alerts give more accurate information about the entity, as well as conversations and tasks.
|
||||
|
||||
## Localization
|
||||
- Fixed localization issues in the confirmation logic for the delete function.
|
||||
- Fixed the search index language configuration.
|
||||
|
||||
## Roles
|
||||
- Now, roles can be inherited from the user configuration in SSO.
|
||||
|
||||
## Search
|
||||
- You can now filter by assets without a description or an owner.
|
||||
- Improved the match results for search results.
|
||||
|
||||
## Others
|
||||
- The description is auto-expanded when the data asset has no data and has the space to accommodate a lengthy description.
|
||||
- User email IDs have been masked and are only visible to Admins.
|
||||
- Users can filter Queries by owner, tag, and creation date in the UI.
|
||||
- Added a button in the Query Editor to copy the Query.
|
||||
- Improved Elasticsearch re-indexing.
|
||||
- Improved the charts based on custom metrics.
|
||||
- Improved the usage of the refresh token.
|
||||
- Redundant scroll bars have been removed from the UI.
|
||||
- Improved the bot role binding to provide more control over which roles are passed to the system bots.
|
||||
|
||||
|
||||
# 1.3.4 Release 🎉
|
||||
|
||||
{% note noteType="Tip" %}
|
||||
|
@ -10,5 +10,5 @@ major releases.
|
||||
|
||||
| Release | Release Date |
|
||||
|:--------|:-------------|
|
||||
| 1.4.x | 2024/05/22 |
|
||||
| 1.3.x | 2024/02/06 |
|
||||
| 1.5.x | 2024/08/26 |
|
||||
| 1.4.x | 2024/05/22 |
|
@ -718,7 +718,7 @@ API:
|
||||
{
|
||||
id: 43,
|
||||
version: 'v1.5.0',
|
||||
description: `Released on 23rd August 2024.`,
|
||||
description: `Released on 26th August 2024.`,
|
||||
features: [
|
||||
{
|
||||
title: `Data Observability with Anomaly Detection (Collate)`,
|
||||
@ -851,7 +851,13 @@ In the latest release, several updates and enhancements have been made to the JS
|
||||
**Include DDL:**
|
||||
|
||||
- During the Database Metadata ingestion, we can optionally pick up the DDL for both tables and views. During the metadata ingestion, we use the view DDLs to generate the View Lineage.
|
||||
- To reduce the processing time for out-of-the-box workflows, we are disabling the include DDL by default, whereas before, it was enabled, which potentially led to long-running workflows.`,
|
||||
- To reduce the processing time for out-of-the-box workflows, we are disabling the include DDL by default, whereas before, it was enabled, which potentially led to long-running workflows.
|
||||
|
||||
**Secrets Manager**
|
||||
Starting with the release 1.5.0, the JWT Token for the bots will be sent to the Secrets Manager if you configured one. It won't appear anymore in your dag_generated_configs in Airflow.
|
||||
|
||||
**Python SDK**
|
||||
The \`metadata insight\` command has been removed. Since Data Insights application was moved to be an internal system application instead of relying on external pipelines the SDK command to run the pipeline was removed.`,
|
||||
[`Data Observability with Anomaly Detection ${CollateIconWithLinkMD}`]: `OpenMetadata has been driving innovation in Data Quality in Open Source. Many organizations are taking advantage of the following Data Quality features to achieve better-quality data
|
||||
|
||||
1. A Native Profiler to understand the shape of the data, freshness, completeness, volume, and ability to add your own metrics, including column level profiler over time-series and dashboards
|
||||
|
Loading…
x
Reference in New Issue
Block a user