2024-04-14 03:48:24 +03:00
|
|
|
import copy
|
2024-05-22 21:22:17 +01:00
|
|
|
from typing import Any, Dict, List
|
2024-05-06 15:16:49 +01:00
|
|
|
from unittest.mock import MagicMock, patch
|
2024-04-14 03:48:24 +03:00
|
|
|
|
|
|
|
import pytest
|
|
|
|
|
2024-05-22 21:22:17 +01:00
|
|
|
from autogen.agentchat.contrib.capabilities.text_compressors import TextCompressor
|
|
|
|
from autogen.agentchat.contrib.capabilities.transforms import (
|
|
|
|
MessageHistoryLimiter,
|
|
|
|
MessageTokenLimiter,
|
|
|
|
TextMessageCompressor,
|
|
|
|
_count_tokens,
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
class _MockTextCompressor:
|
|
|
|
def compress_text(self, text: str, **compression_params) -> Dict[str, Any]:
|
|
|
|
return {"compressed_prompt": ""}
|
2024-04-14 03:48:24 +03:00
|
|
|
|
|
|
|
|
|
|
|
def get_long_messages() -> List[Dict]:
|
|
|
|
return [
|
|
|
|
{"role": "assistant", "content": [{"type": "text", "text": "are you doing?"}]},
|
|
|
|
{"role": "user", "content": "very very very very very very long string"},
|
|
|
|
{"role": "user", "content": "hello"},
|
|
|
|
{"role": "assistant", "content": [{"type": "text", "text": "there"}]},
|
|
|
|
{"role": "user", "content": "how"},
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
|
|
def get_short_messages() -> List[Dict]:
|
|
|
|
return [
|
|
|
|
{"role": "user", "content": "hello"},
|
|
|
|
{"role": "assistant", "content": [{"type": "text", "text": "there"}]},
|
2024-04-29 05:11:16 +03:00
|
|
|
{"role": "user", "content": "how are you"},
|
2024-04-14 03:48:24 +03:00
|
|
|
]
|
|
|
|
|
|
|
|
|
|
|
|
def get_no_content_messages() -> List[Dict]:
|
|
|
|
return [{"role": "user", "function_call": "example"}, {"role": "assistant", "content": None}]
|
|
|
|
|
|
|
|
|
2024-05-22 21:22:17 +01:00
|
|
|
def get_text_compressors() -> List[TextCompressor]:
|
|
|
|
compressors: List[TextCompressor] = [_MockTextCompressor()]
|
|
|
|
try:
|
|
|
|
from autogen.agentchat.contrib.capabilities.text_compressors import LLMLingua
|
|
|
|
|
|
|
|
compressors.append(LLMLingua())
|
|
|
|
except ImportError:
|
|
|
|
pass
|
|
|
|
|
|
|
|
return compressors
|
|
|
|
|
|
|
|
|
2024-04-14 03:48:24 +03:00
|
|
|
@pytest.fixture
|
|
|
|
def message_history_limiter() -> MessageHistoryLimiter:
|
|
|
|
return MessageHistoryLimiter(max_messages=3)
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.fixture
|
|
|
|
def message_token_limiter() -> MessageTokenLimiter:
|
|
|
|
return MessageTokenLimiter(max_tokens_per_message=3)
|
|
|
|
|
|
|
|
|
2024-04-29 05:11:16 +03:00
|
|
|
@pytest.fixture
|
|
|
|
def message_token_limiter_with_threshold() -> MessageTokenLimiter:
|
|
|
|
return MessageTokenLimiter(max_tokens_per_message=1, min_tokens=10)
|
|
|
|
|
|
|
|
|
2024-05-22 21:22:17 +01:00
|
|
|
def _filter_dict_test(
|
|
|
|
post_transformed_message: Dict, pre_transformed_messages: Dict, roles: List[str], exclude_filter: bool
|
|
|
|
) -> bool:
|
|
|
|
is_role = post_transformed_message["role"] in roles
|
|
|
|
if exclude_filter:
|
|
|
|
is_role = not is_role
|
|
|
|
|
|
|
|
if isinstance(post_transformed_message["content"], list):
|
|
|
|
condition = (
|
|
|
|
len(post_transformed_message["content"][0]["text"]) < len(pre_transformed_messages["content"][0]["text"])
|
|
|
|
if is_role
|
|
|
|
else len(post_transformed_message["content"][0]["text"])
|
|
|
|
== len(pre_transformed_messages["content"][0]["text"])
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
condition = (
|
|
|
|
len(post_transformed_message["content"]) < len(pre_transformed_messages["content"])
|
|
|
|
if is_role
|
|
|
|
else len(post_transformed_message["content"]) == len(pre_transformed_messages["content"])
|
|
|
|
)
|
|
|
|
|
|
|
|
return condition
|
|
|
|
|
|
|
|
|
2024-04-29 05:11:16 +03:00
|
|
|
# MessageHistoryLimiter
|
2024-04-14 03:48:24 +03:00
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.parametrize(
|
|
|
|
"messages, expected_messages_len",
|
|
|
|
[(get_long_messages(), 3), (get_short_messages(), 3), (get_no_content_messages(), 2)],
|
|
|
|
)
|
|
|
|
def test_message_history_limiter_apply_transform(message_history_limiter, messages, expected_messages_len):
|
|
|
|
transformed_messages = message_history_limiter.apply_transform(messages)
|
|
|
|
assert len(transformed_messages) == expected_messages_len
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.parametrize(
|
|
|
|
"messages, expected_logs, expected_effect",
|
|
|
|
[
|
|
|
|
(get_long_messages(), "Removed 2 messages. Number of messages reduced from 5 to 3.", True),
|
|
|
|
(get_short_messages(), "No messages were removed.", False),
|
|
|
|
(get_no_content_messages(), "No messages were removed.", False),
|
|
|
|
],
|
|
|
|
)
|
|
|
|
def test_message_history_limiter_get_logs(message_history_limiter, messages, expected_logs, expected_effect):
|
|
|
|
pre_transform_messages = copy.deepcopy(messages)
|
|
|
|
transformed_messages = message_history_limiter.apply_transform(messages)
|
|
|
|
logs_str, had_effect = message_history_limiter.get_logs(pre_transform_messages, transformed_messages)
|
|
|
|
assert had_effect == expected_effect
|
|
|
|
assert logs_str == expected_logs
|
|
|
|
|
|
|
|
|
|
|
|
# MessageTokenLimiter tests
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.parametrize(
|
|
|
|
"messages, expected_token_count, expected_messages_len",
|
2024-04-29 05:11:16 +03:00
|
|
|
[(get_long_messages(), 9, 5), (get_short_messages(), 5, 3), (get_no_content_messages(), 0, 2)],
|
2024-04-14 03:48:24 +03:00
|
|
|
)
|
|
|
|
def test_message_token_limiter_apply_transform(
|
|
|
|
message_token_limiter, messages, expected_token_count, expected_messages_len
|
|
|
|
):
|
2024-05-22 21:22:17 +01:00
|
|
|
transformed_messages = message_token_limiter.apply_transform(copy.deepcopy(messages))
|
2024-04-14 03:48:24 +03:00
|
|
|
assert (
|
|
|
|
sum(_count_tokens(msg["content"]) for msg in transformed_messages if "content" in msg) == expected_token_count
|
|
|
|
)
|
|
|
|
assert len(transformed_messages) == expected_messages_len
|
|
|
|
|
|
|
|
|
2024-05-22 21:22:17 +01:00
|
|
|
@pytest.mark.parametrize("messages", [get_long_messages(), get_short_messages()])
|
|
|
|
def test_message_token_limiter_with_filter(messages):
|
|
|
|
# Test truncating all messages except for user
|
|
|
|
message_token_limiter = MessageTokenLimiter(max_tokens_per_message=0, filter_dict={"role": "user"})
|
|
|
|
transformed_messages = message_token_limiter.apply_transform(copy.deepcopy(messages))
|
|
|
|
|
|
|
|
pre_post_messages = zip(messages, transformed_messages)
|
|
|
|
|
|
|
|
for pre_transform, post_transform in pre_post_messages:
|
|
|
|
assert _filter_dict_test(post_transform, pre_transform, ["user"], exclude_filter=True)
|
|
|
|
|
|
|
|
# Test truncating all user messages only
|
|
|
|
message_token_limiter = MessageTokenLimiter(
|
|
|
|
max_tokens_per_message=0, filter_dict={"role": "user"}, exclude_filter=False
|
|
|
|
)
|
|
|
|
transformed_messages = message_token_limiter.apply_transform(copy.deepcopy(messages))
|
|
|
|
|
|
|
|
pre_post_messages = zip(messages, transformed_messages)
|
|
|
|
for pre_transform, post_transform in pre_post_messages:
|
|
|
|
assert _filter_dict_test(post_transform, pre_transform, ["user"], exclude_filter=False)
|
|
|
|
|
|
|
|
|
2024-04-29 05:11:16 +03:00
|
|
|
@pytest.mark.parametrize(
|
|
|
|
"messages, expected_token_count, expected_messages_len",
|
|
|
|
[(get_long_messages(), 5, 5), (get_short_messages(), 5, 3), (get_no_content_messages(), 0, 2)],
|
|
|
|
)
|
|
|
|
def test_message_token_limiter_with_threshold_apply_transform(
|
|
|
|
message_token_limiter_with_threshold, messages, expected_token_count, expected_messages_len
|
|
|
|
):
|
|
|
|
transformed_messages = message_token_limiter_with_threshold.apply_transform(messages)
|
|
|
|
assert (
|
|
|
|
sum(_count_tokens(msg["content"]) for msg in transformed_messages if "content" in msg) == expected_token_count
|
|
|
|
)
|
|
|
|
assert len(transformed_messages) == expected_messages_len
|
|
|
|
|
|
|
|
|
2024-04-14 03:48:24 +03:00
|
|
|
@pytest.mark.parametrize(
|
|
|
|
"messages, expected_logs, expected_effect",
|
|
|
|
[
|
|
|
|
(get_long_messages(), "Truncated 6 tokens. Number of tokens reduced from 15 to 9", True),
|
|
|
|
(get_short_messages(), "No tokens were truncated.", False),
|
|
|
|
(get_no_content_messages(), "No tokens were truncated.", False),
|
|
|
|
],
|
|
|
|
)
|
|
|
|
def test_message_token_limiter_get_logs(message_token_limiter, messages, expected_logs, expected_effect):
|
|
|
|
pre_transform_messages = copy.deepcopy(messages)
|
|
|
|
transformed_messages = message_token_limiter.apply_transform(messages)
|
|
|
|
logs_str, had_effect = message_token_limiter.get_logs(pre_transform_messages, transformed_messages)
|
|
|
|
assert had_effect == expected_effect
|
|
|
|
assert logs_str == expected_logs
|
|
|
|
|
|
|
|
|
2024-05-22 21:22:17 +01:00
|
|
|
# TextMessageCompressor tests
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.parametrize("text_compressor", get_text_compressors())
|
|
|
|
def test_text_compression(text_compressor):
|
2024-05-06 15:16:49 +01:00
|
|
|
"""Test the TextMessageCompressor transform."""
|
|
|
|
|
2024-05-22 21:22:17 +01:00
|
|
|
compressor = TextMessageCompressor(text_compressor=text_compressor)
|
2024-05-06 15:16:49 +01:00
|
|
|
|
|
|
|
text = "Run this test with a long string. "
|
|
|
|
messages = [
|
2024-05-22 21:22:17 +01:00
|
|
|
{"role": "assistant", "content": [{"type": "text", "text": "".join([text] * 3)}]},
|
|
|
|
{"role": "role", "content": [{"type": "text", "text": "".join([text] * 3)}]},
|
|
|
|
{"role": "assistant", "content": [{"type": "text", "text": "".join([text] * 3)}]},
|
|
|
|
{"role": "assistant", "content": [{"type": "text", "text": "".join([text] * 3)}]},
|
2024-05-06 15:16:49 +01:00
|
|
|
]
|
|
|
|
|
2024-05-22 21:22:17 +01:00
|
|
|
transformed_messages = compressor.apply_transform([{"content": text}])
|
2024-05-06 15:16:49 +01:00
|
|
|
|
|
|
|
assert len(transformed_messages[0]["content"]) < len(text)
|
|
|
|
|
|
|
|
# Test compressing all messages
|
2024-05-22 21:22:17 +01:00
|
|
|
compressor = TextMessageCompressor(text_compressor=text_compressor)
|
|
|
|
transformed_messages = compressor.apply_transform(copy.deepcopy(messages))
|
2024-05-06 15:16:49 +01:00
|
|
|
|
2024-05-22 21:22:17 +01:00
|
|
|
pre_post_messages = zip(messages, transformed_messages)
|
|
|
|
for pre_transform, post_transform in pre_post_messages:
|
|
|
|
assert len(post_transform["content"][0]["text"]) < len(pre_transform["content"][0]["text"])
|
2024-05-06 15:16:49 +01:00
|
|
|
|
|
|
|
|
2024-05-22 21:22:17 +01:00
|
|
|
@pytest.mark.parametrize("messages", [get_long_messages(), get_short_messages()])
|
|
|
|
@pytest.mark.parametrize("text_compressor", get_text_compressors())
|
|
|
|
def test_text_compression_with_filter(messages, text_compressor):
|
|
|
|
# Test truncating all messages except for user
|
|
|
|
compressor = TextMessageCompressor(text_compressor=text_compressor, filter_dict={"role": "user"})
|
|
|
|
transformed_messages = compressor.apply_transform(copy.deepcopy(messages))
|
|
|
|
|
|
|
|
pre_post_messages = zip(messages, transformed_messages)
|
|
|
|
for pre_transform, post_transform in pre_post_messages:
|
|
|
|
assert _filter_dict_test(post_transform, pre_transform, ["user"], exclude_filter=True)
|
|
|
|
|
|
|
|
# Test truncating all user messages only
|
|
|
|
compressor = TextMessageCompressor(
|
|
|
|
text_compressor=text_compressor, filter_dict={"role": "user"}, exclude_filter=False
|
|
|
|
)
|
|
|
|
transformed_messages = compressor.apply_transform(copy.deepcopy(messages))
|
|
|
|
|
|
|
|
pre_post_messages = zip(messages, transformed_messages)
|
|
|
|
for pre_transform, post_transform in pre_post_messages:
|
|
|
|
assert _filter_dict_test(post_transform, pre_transform, ["user"], exclude_filter=False)
|
|
|
|
|
2024-05-06 15:16:49 +01:00
|
|
|
|
2024-05-22 21:22:17 +01:00
|
|
|
@pytest.mark.parametrize("text_compressor", get_text_compressors())
|
|
|
|
def test_text_compression_cache(text_compressor):
|
2024-05-06 15:16:49 +01:00
|
|
|
messages = get_long_messages()
|
|
|
|
mock_compressed_content = (1, {"content": "mock"})
|
|
|
|
|
|
|
|
with patch(
|
|
|
|
"autogen.agentchat.contrib.capabilities.transforms.TextMessageCompressor._cache_get",
|
|
|
|
MagicMock(return_value=(1, {"content": "mock"})),
|
|
|
|
) as mocked_get, patch(
|
|
|
|
"autogen.agentchat.contrib.capabilities.transforms.TextMessageCompressor._cache_set", MagicMock()
|
|
|
|
) as mocked_set:
|
2024-05-22 21:22:17 +01:00
|
|
|
compressor = TextMessageCompressor(text_compressor=text_compressor)
|
2024-05-06 15:16:49 +01:00
|
|
|
|
2024-05-22 21:22:17 +01:00
|
|
|
compressor.apply_transform(messages)
|
|
|
|
compressor.apply_transform(messages)
|
2024-05-06 15:16:49 +01:00
|
|
|
|
|
|
|
assert mocked_get.call_count == len(messages)
|
|
|
|
assert mocked_set.call_count == len(messages)
|
|
|
|
|
|
|
|
# We already populated the cache with the mock content
|
|
|
|
# We need to test if we retrieve the correct content
|
2024-05-22 21:22:17 +01:00
|
|
|
compressor = TextMessageCompressor(text_compressor=text_compressor)
|
|
|
|
compressed_messages = compressor.apply_transform(messages)
|
2024-05-06 15:16:49 +01:00
|
|
|
|
|
|
|
for message in compressed_messages:
|
|
|
|
assert message["content"] == mock_compressed_content[1]
|
|
|
|
|
|
|
|
|
2024-04-14 03:48:24 +03:00
|
|
|
if __name__ == "__main__":
|
|
|
|
long_messages = get_long_messages()
|
|
|
|
short_messages = get_short_messages()
|
2024-04-29 05:11:16 +03:00
|
|
|
no_content_messages = get_no_content_messages()
|
2024-05-06 15:16:49 +01:00
|
|
|
msg_history_limiter = MessageHistoryLimiter(max_messages=3)
|
|
|
|
msg_token_limiter = MessageTokenLimiter(max_tokens_per_message=3)
|
|
|
|
msg_token_limiter_with_threshold = MessageTokenLimiter(max_tokens_per_message=1, min_tokens=10)
|
2024-04-29 05:11:16 +03:00
|
|
|
|
|
|
|
# Test Parameters
|
|
|
|
message_history_limiter_apply_transform_parameters = {
|
|
|
|
"messages": [long_messages, short_messages, no_content_messages],
|
|
|
|
"expected_messages_len": [3, 3, 2],
|
|
|
|
}
|
|
|
|
|
|
|
|
message_history_limiter_get_logs_parameters = {
|
|
|
|
"messages": [long_messages, short_messages, no_content_messages],
|
|
|
|
"expected_logs": [
|
|
|
|
"Removed 2 messages. Number of messages reduced from 5 to 3.",
|
|
|
|
"No messages were removed.",
|
|
|
|
"No messages were removed.",
|
|
|
|
],
|
|
|
|
"expected_effect": [True, False, False],
|
|
|
|
}
|
|
|
|
|
|
|
|
message_token_limiter_apply_transform_parameters = {
|
|
|
|
"messages": [long_messages, short_messages, no_content_messages],
|
|
|
|
"expected_token_count": [9, 5, 0],
|
|
|
|
"expected_messages_len": [5, 3, 2],
|
|
|
|
}
|
|
|
|
|
|
|
|
message_token_limiter_with_threshold_apply_transform_parameters = {
|
|
|
|
"messages": [long_messages, short_messages, no_content_messages],
|
|
|
|
"expected_token_count": [5, 5, 0],
|
|
|
|
"expected_messages_len": [5, 3, 2],
|
|
|
|
}
|
|
|
|
|
|
|
|
message_token_limiter_get_logs_parameters = {
|
|
|
|
"messages": [long_messages, short_messages, no_content_messages],
|
|
|
|
"expected_logs": [
|
|
|
|
"Truncated 6 tokens. Number of tokens reduced from 15 to 9",
|
|
|
|
"No tokens were truncated.",
|
|
|
|
"No tokens were truncated.",
|
|
|
|
],
|
|
|
|
"expected_effect": [True, False, False],
|
|
|
|
}
|
2024-04-14 03:48:24 +03:00
|
|
|
|
|
|
|
# Call the MessageHistoryLimiter tests
|
2024-04-29 05:11:16 +03:00
|
|
|
|
|
|
|
for messages, expected_messages_len in zip(
|
|
|
|
message_history_limiter_apply_transform_parameters["messages"],
|
|
|
|
message_history_limiter_apply_transform_parameters["expected_messages_len"],
|
|
|
|
):
|
2024-05-06 15:16:49 +01:00
|
|
|
test_message_history_limiter_apply_transform(msg_history_limiter, messages, expected_messages_len)
|
2024-04-29 05:11:16 +03:00
|
|
|
|
|
|
|
for messages, expected_logs, expected_effect in zip(
|
|
|
|
message_history_limiter_get_logs_parameters["messages"],
|
|
|
|
message_history_limiter_get_logs_parameters["expected_logs"],
|
|
|
|
message_history_limiter_get_logs_parameters["expected_effect"],
|
|
|
|
):
|
2024-05-06 15:16:49 +01:00
|
|
|
test_message_history_limiter_get_logs(msg_history_limiter, messages, expected_logs, expected_effect)
|
2024-04-14 03:48:24 +03:00
|
|
|
|
|
|
|
# Call the MessageTokenLimiter tests
|
2024-04-29 05:11:16 +03:00
|
|
|
|
|
|
|
for messages, expected_token_count, expected_messages_len in zip(
|
|
|
|
message_token_limiter_apply_transform_parameters["messages"],
|
|
|
|
message_token_limiter_apply_transform_parameters["expected_token_count"],
|
|
|
|
message_token_limiter_apply_transform_parameters["expected_messages_len"],
|
|
|
|
):
|
|
|
|
test_message_token_limiter_apply_transform(
|
2024-05-06 15:16:49 +01:00
|
|
|
msg_token_limiter, messages, expected_token_count, expected_messages_len
|
2024-04-29 05:11:16 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
for messages, expected_token_count, expected_messages_len in zip(
|
|
|
|
message_token_limiter_with_threshold_apply_transform_parameters["messages"],
|
|
|
|
message_token_limiter_with_threshold_apply_transform_parameters["expected_token_count"],
|
|
|
|
message_token_limiter_with_threshold_apply_transform_parameters["expected_messages_len"],
|
|
|
|
):
|
|
|
|
test_message_token_limiter_with_threshold_apply_transform(
|
2024-05-06 15:16:49 +01:00
|
|
|
msg_token_limiter_with_threshold, messages, expected_token_count, expected_messages_len
|
2024-04-29 05:11:16 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
for messages, expected_logs, expected_effect in zip(
|
|
|
|
message_token_limiter_get_logs_parameters["messages"],
|
|
|
|
message_token_limiter_get_logs_parameters["expected_logs"],
|
|
|
|
message_token_limiter_get_logs_parameters["expected_effect"],
|
|
|
|
):
|
2024-05-06 15:16:49 +01:00
|
|
|
test_message_token_limiter_get_logs(msg_token_limiter, messages, expected_logs, expected_effect)
|