graphrag/tests/verbs/test_extract_covariates.py
Nathan Evans 1df89727c3
Pipeline registration (#1940)
* Move covariate run conditional

* All pipeline registration

* Fix method name construction

* Rename context storage -> output_storage

* Rename OutputConfig as generic StorageConfig

* Reuse Storage model under InputConfig

* Move input storage creation out of document loading

* Move document loading into workflows

* Semver

* Fix smoke test config for new workflows

* Fix unit tests

---------

Co-authored-by: Alonso Guevara <alonsog@microsoft.com>
2025-06-12 16:14:39 -07:00

80 lines
3.3 KiB
Python

# Copyright (c) 2024 Microsoft Corporation.
# Licensed under the MIT License
from pandas.testing import assert_series_equal
from graphrag.config.create_graphrag_config import create_graphrag_config
from graphrag.config.enums import ModelType
from graphrag.data_model.schemas import COVARIATES_FINAL_COLUMNS
from graphrag.index.workflows.extract_covariates import (
run_workflow,
)
from graphrag.utils.storage import load_table_from_storage
from .util import (
DEFAULT_MODEL_CONFIG,
create_test_context,
load_test_table,
)
MOCK_LLM_RESPONSES = [
"""
(COMPANY A<|>GOVERNMENT AGENCY B<|>ANTI-COMPETITIVE PRACTICES<|>TRUE<|>2022-01-10T00:00:00<|>2022-01-10T00:00:00<|>Company A was found to engage in anti-competitive practices because it was fined for bid rigging in multiple public tenders published by Government Agency B according to an article published on 2022/01/10<|>According to an article published on 2022/01/10, Company A was fined for bid rigging while participating in multiple public tenders published by Government Agency B.)
""".strip()
]
async def test_extract_covariates():
input = load_test_table("text_units")
context = await create_test_context(
storage=["text_units"],
)
config = create_graphrag_config({"models": DEFAULT_MODEL_CONFIG})
llm_settings = config.get_language_model_config(
config.extract_claims.model_id
).model_dump()
llm_settings["type"] = ModelType.MockChat
llm_settings["responses"] = MOCK_LLM_RESPONSES
config.extract_claims.enabled = True
config.extract_claims.strategy = {
"type": "graph_intelligence",
"llm": llm_settings,
"claim_description": "description",
}
await run_workflow(config, context)
actual = await load_table_from_storage("covariates", context.output_storage)
for column in COVARIATES_FINAL_COLUMNS:
assert column in actual.columns
# our mock only returns one covariate per text unit, so that's a 1:1 mapping versus the LLM-extracted content in the test data
assert len(actual) == len(input)
# assert all of the columns that covariates copied from the input
assert_series_equal(actual["text_unit_id"], input["id"], check_names=False)
# make sure the human ids are incrementing
assert actual["human_readable_id"][0] == 1
assert actual["human_readable_id"][1] == 2
# check that the mock data is parsed and inserted into the correct columns
assert actual["covariate_type"][0] == "claim"
assert actual["subject_id"][0] == "COMPANY A"
assert actual["object_id"][0] == "GOVERNMENT AGENCY B"
assert actual["type"][0] == "ANTI-COMPETITIVE PRACTICES"
assert actual["status"][0] == "TRUE"
assert actual["start_date"][0] == "2022-01-10T00:00:00"
assert actual["end_date"][0] == "2022-01-10T00:00:00"
assert (
actual["description"][0]
== "Company A was found to engage in anti-competitive practices because it was fined for bid rigging in multiple public tenders published by Government Agency B according to an article published on 2022/01/10"
)
assert (
actual["source_text"][0]
== "According to an article published on 2022/01/10, Company A was fined for bid rigging while participating in multiple public tenders published by Government Agency B."
)