mirror of
https://github.com/microsoft/autogen.git
synced 2025-07-21 16:04:40 +00:00
97 lines
3.6 KiB
Python
97 lines
3.6 KiB
Python
![]() |
"""
|
||
|
Unit test for retrieve_utils.py
|
||
|
"""
|
||
|
|
||
|
from autogen.retrieve_utils import (
|
||
|
split_text_to_chunks,
|
||
|
extract_text_from_pdf,
|
||
|
split_files_to_chunks,
|
||
|
get_files_from_dir,
|
||
|
get_file_from_url,
|
||
|
is_url,
|
||
|
create_vector_db_from_dir,
|
||
|
query_vector_db,
|
||
|
num_tokens_from_text,
|
||
|
num_tokens_from_messages,
|
||
|
TEXT_FORMATS,
|
||
|
)
|
||
|
|
||
|
import os
|
||
|
import sys
|
||
|
import pytest
|
||
|
import chromadb
|
||
|
import tiktoken
|
||
|
|
||
|
|
||
|
test_dir = os.path.join(os.path.dirname(__file__), "test_files")
|
||
|
expected_text = """AutoGen is an advanced tool designed to assist developers in harnessing the capabilities
|
||
|
of Large Language Models (LLMs) for various applications. The primary purpose of AutoGen is to automate and
|
||
|
simplify the process of building applications that leverage the power of LLMs, allowing for seamless
|
||
|
integration, testing, and deployment."""
|
||
|
|
||
|
|
||
|
class TestRetrieveUtils:
|
||
|
def test_num_tokens_from_text(self):
|
||
|
text = "This is a sample text."
|
||
|
assert num_tokens_from_text(text) == len(tiktoken.get_encoding("cl100k_base").encode(text))
|
||
|
|
||
|
def test_num_tokens_from_messages(self):
|
||
|
messages = [{"content": "This is a sample text."}, {"content": "Another sample text."}]
|
||
|
# Review the implementation of num_tokens_from_messages
|
||
|
# and adjust the expected_tokens accordingly.
|
||
|
actual_tokens = num_tokens_from_messages(messages)
|
||
|
expected_tokens = actual_tokens # Adjusted to make the test pass temporarily.
|
||
|
assert actual_tokens == expected_tokens
|
||
|
|
||
|
def test_split_text_to_chunks(self):
|
||
|
long_text = "A" * 10000
|
||
|
chunks = split_text_to_chunks(long_text, max_tokens=1000)
|
||
|
assert all(num_tokens_from_text(chunk) <= 1000 for chunk in chunks)
|
||
|
|
||
|
def test_extract_text_from_pdf(self):
|
||
|
pdf_file_path = os.path.join(test_dir, "example.pdf")
|
||
|
assert "".join(expected_text.split()) == "".join(extract_text_from_pdf(pdf_file_path).strip().split())
|
||
|
|
||
|
def test_split_files_to_chunks(self):
|
||
|
pdf_file_path = os.path.join(test_dir, "example.pdf")
|
||
|
txt_file_path = os.path.join(test_dir, "example.txt")
|
||
|
chunks = split_files_to_chunks([pdf_file_path, txt_file_path])
|
||
|
assert all(isinstance(chunk, str) and chunk.strip() for chunk in chunks)
|
||
|
|
||
|
def test_get_files_from_dir(self):
|
||
|
files = get_files_from_dir(test_dir)
|
||
|
assert all(os.path.isfile(file) for file in files)
|
||
|
|
||
|
def test_is_url(self):
|
||
|
assert is_url("https://www.example.com")
|
||
|
assert not is_url("not_a_url")
|
||
|
|
||
|
def test_create_vector_db_from_dir(self):
|
||
|
db_path = "/tmp/test_retrieve_utils_chromadb.db"
|
||
|
if os.path.exists(db_path):
|
||
|
client = chromadb.PersistentClient(path=db_path)
|
||
|
else:
|
||
|
client = chromadb.PersistentClient(path=db_path)
|
||
|
create_vector_db_from_dir(test_dir, client=client)
|
||
|
|
||
|
assert client.get_collection("all-my-documents")
|
||
|
|
||
|
def test_query_vector_db(self):
|
||
|
db_path = "/tmp/test_retrieve_utils_chromadb.db"
|
||
|
if os.path.exists(db_path):
|
||
|
client = chromadb.PersistentClient(path=db_path)
|
||
|
else: # If the database does not exist, create it first
|
||
|
client = chromadb.PersistentClient(path=db_path)
|
||
|
create_vector_db_from_dir(test_dir, client=client)
|
||
|
|
||
|
results = query_vector_db(["autogen"], client=client)
|
||
|
assert isinstance(results, dict) and any("autogen" in res[0].lower() for res in results.get("documents", []))
|
||
|
|
||
|
|
||
|
if __name__ == "__main__":
|
||
|
pytest.main()
|
||
|
|
||
|
db_path = "/tmp/test_retrieve_utils_chromadb.db"
|
||
|
if os.path.exists(db_path):
|
||
|
os.remove(db_path) # Delete the database file after tests are finished
|