datahub/smoke-test/tests/database/test_database.py
2025-01-17 23:50:13 +05:30

33 lines
932 B
Python

import logging
import pytest
from datahub.emitter.mce_builder import make_dataset_urn
from tests.utilities.concurrent_openapi import run_tests
from tests.utils import delete_urns, wait_for_writes_to_sync
logger = logging.getLogger(__name__)
generated_urns = [make_dataset_urn("test", f"database_test_{i}") for i in range(0, 100)]
@pytest.fixture(scope="module")
def ingest_cleanup_data(graph_client, request):
print("removing test data before")
delete_urns(graph_client, generated_urns)
wait_for_writes_to_sync()
yield
print("removing test data after")
delete_urns(graph_client, generated_urns)
wait_for_writes_to_sync()
def test_mysql_deadlock_gap_locking(auth_session, ingest_cleanup_data):
# This generates concurrent batches with interleaved urn ids
run_tests(
auth_session,
fixture_globs=["tests/database/v3/mysql_gap_deadlock/*.json"],
num_workers=8,
)