mirror of
https://github.com/microsoft/autogen.git
synced 2025-07-19 23:11:22 +00:00

* update orm branch + accesibility tweaks * general file location refactor * add support for LocalCommandLineCodeExecutor and DockerCommandLineCodeExecutor * update code execution config laoding * version update * bump version rc1 * add model type selection (openai , gemini, azure) * add ability to test workflow * psycopg3 support * add close logic to build tab pop ups, enable testing of workflows in build view * updates to dbmanager, version bump * add max_tokens default value * ensure sessions are used correctly in dbmanager * initial support for migrations * update sessions/workflow api routing for clarity. * general refactor, + add support for initial sample workflows * orm branch updates * Removed incorrect Git LFS files * update git lfs tracking --------- Co-authored-by: Audel Rouhi <knucklessg1@gmail.com>
324 lines
15 KiB
Python
324 lines
15 KiB
Python
# from .util import get_app_root
|
|
import os
|
|
import time
|
|
from datetime import datetime
|
|
from pathlib import Path
|
|
from typing import Any
|
|
|
|
from alembic import command, util
|
|
from alembic.config import Config
|
|
from loguru import logger
|
|
|
|
# from ..utils.db_utils import get_db_uri
|
|
from sqlmodel import Session, create_engine, text
|
|
|
|
from autogen.agentchat import AssistantAgent
|
|
|
|
from ..datamodel import (
|
|
Agent,
|
|
AgentConfig,
|
|
AgentType,
|
|
CodeExecutionConfigTypes,
|
|
Model,
|
|
Skill,
|
|
Workflow,
|
|
WorkflowAgentLink,
|
|
)
|
|
|
|
|
|
def workflow_from_id(workflow_id: int, dbmanager: Any):
|
|
workflow = dbmanager.get(Workflow, filters={"id": workflow_id}).data
|
|
if not workflow or len(workflow) == 0:
|
|
raise ValueError("The specified workflow does not exist.")
|
|
workflow = workflow[0].model_dump(mode="json")
|
|
workflow_agent_links = dbmanager.get(WorkflowAgentLink, filters={"workflow_id": workflow_id}).data
|
|
|
|
def dump_agent(agent: Agent):
|
|
exclude = []
|
|
if agent.type != AgentType.groupchat:
|
|
exclude = [
|
|
"admin_name",
|
|
"messages",
|
|
"max_round",
|
|
"admin_name",
|
|
"speaker_selection_method",
|
|
"allow_repeat_speaker",
|
|
]
|
|
return agent.model_dump(warnings=False, mode="json", exclude=exclude)
|
|
|
|
def get_agent(agent_id):
|
|
with Session(dbmanager.engine) as session:
|
|
agent: Agent = dbmanager.get_items(Agent, filters={"id": agent_id}, session=session).data[0]
|
|
agent_dict = dump_agent(agent)
|
|
agent_dict["skills"] = [Skill.model_validate(skill.model_dump(mode="json")) for skill in agent.skills]
|
|
model_exclude = [
|
|
"id",
|
|
"agent_id",
|
|
"created_at",
|
|
"updated_at",
|
|
"user_id",
|
|
"description",
|
|
]
|
|
models = [model.model_dump(mode="json", exclude=model_exclude) for model in agent.models]
|
|
agent_dict["models"] = [model.model_dump(mode="json") for model in agent.models]
|
|
|
|
if len(models) > 0:
|
|
agent_dict["config"]["llm_config"] = agent_dict.get("config", {}).get("llm_config", {})
|
|
llm_config = agent_dict["config"]["llm_config"]
|
|
if llm_config:
|
|
llm_config["config_list"] = models
|
|
agent_dict["config"]["llm_config"] = llm_config
|
|
agent_dict["agents"] = [get_agent(agent.id) for agent in agent.agents]
|
|
return agent_dict
|
|
|
|
for link in workflow_agent_links:
|
|
agent_dict = get_agent(link.agent_id)
|
|
workflow[str(link.agent_type.value)] = agent_dict
|
|
return workflow
|
|
|
|
|
|
def run_migration(engine_uri: str):
|
|
database_dir = Path(__file__).parent
|
|
script_location = database_dir / "migrations"
|
|
|
|
engine = create_engine(engine_uri)
|
|
buffer = open(script_location / "alembic.log", "w")
|
|
alembic_cfg = Config(stdout=buffer)
|
|
alembic_cfg.set_main_option("script_location", str(script_location))
|
|
alembic_cfg.set_main_option("sqlalchemy.url", engine_uri)
|
|
|
|
print(f"Running migrations with engine_uri: {engine_uri}")
|
|
|
|
should_initialize_alembic = False
|
|
with Session(engine) as session:
|
|
try:
|
|
session.exec(text("SELECT * FROM alembic_version"))
|
|
except Exception:
|
|
logger.info("Alembic not initialized")
|
|
should_initialize_alembic = True
|
|
else:
|
|
logger.info("Alembic already initialized")
|
|
|
|
if should_initialize_alembic:
|
|
try:
|
|
logger.info("Initializing alembic")
|
|
command.ensure_version(alembic_cfg)
|
|
command.upgrade(alembic_cfg, "head")
|
|
logger.info("Alembic initialized")
|
|
except Exception as exc:
|
|
logger.error(f"Error initializing alembic: {exc}")
|
|
raise RuntimeError("Error initializing alembic") from exc
|
|
|
|
logger.info(f"Running DB migrations in {script_location}")
|
|
|
|
try:
|
|
buffer.write(f"{datetime.now().isoformat()}: Checking migrations\n")
|
|
command.check(alembic_cfg)
|
|
except Exception as exc:
|
|
if isinstance(exc, (util.exc.CommandError, util.exc.AutogenerateDiffsDetected)):
|
|
try:
|
|
command.upgrade(alembic_cfg, "head")
|
|
time.sleep(3)
|
|
except Exception as exc:
|
|
logger.error(f"Error running migrations: {exc}")
|
|
|
|
try:
|
|
buffer.write(f"{datetime.now().isoformat()}: Checking migrations\n")
|
|
command.check(alembic_cfg)
|
|
except util.exc.AutogenerateDiffsDetected as exc:
|
|
logger.info(f"AutogenerateDiffsDetected: {exc}")
|
|
# raise RuntimeError(
|
|
# f"There's a mismatch between the models and the database.\n{exc}")
|
|
except util.exc.CommandError as exc:
|
|
logger.error(f"CommandError: {exc}")
|
|
# raise RuntimeError(f"Error running migrations: {exc}")
|
|
|
|
|
|
def init_db_samples(dbmanager: Any):
|
|
workflows = dbmanager.get(Workflow).data
|
|
workflow_names = [w.name for w in workflows]
|
|
if "Default Workflow" in workflow_names and "Travel Planning Workflow" in workflow_names:
|
|
logger.info("Database already initialized with Default and Travel Planning Workflows")
|
|
return
|
|
logger.info("Initializing database with Default and Travel Planning Workflows")
|
|
# models
|
|
gpt_4_model = Model(
|
|
model="gpt-4-1106-preview", description="OpenAI GPT-4 model", user_id="guestuser@gmail.com", api_type="open_ai"
|
|
)
|
|
azure_model = Model(
|
|
model="gpt4-turbo",
|
|
description="Azure OpenAI model",
|
|
user_id="guestuser@gmail.com",
|
|
api_type="azure",
|
|
base_url="https://api.your azureendpoint.com/v1",
|
|
)
|
|
zephyr_model = Model(
|
|
model="zephyr",
|
|
description="Local Huggingface Zephyr model via vLLM, LMStudio or Ollama",
|
|
base_url="http://localhost:1234/v1",
|
|
user_id="guestuser@gmail.com",
|
|
api_type="open_ai",
|
|
)
|
|
|
|
google_gemini_model = Model(
|
|
model="gemini-1.5-pro-latest",
|
|
description="Google's Gemini model",
|
|
user_id="guestuser@gmail.com",
|
|
api_type="google",
|
|
)
|
|
|
|
# skills
|
|
|
|
generate_image_skill = Skill(
|
|
name="generate_images",
|
|
description="Generate and save images based on a user's query.",
|
|
content='\nfrom typing import List\nimport uuid\nimport requests # to perform HTTP requests\nfrom pathlib import Path\n\nfrom openai import OpenAI\n\n\ndef generate_and_save_images(query: str, image_size: str = "1024x1024") -> List[str]:\n """\n Function to paint, draw or illustrate images based on the users query or request. Generates images from a given query using OpenAI\'s DALL-E model and saves them to disk. Use the code below anytime there is a request to create an image.\n\n :param query: A natural language description of the image to be generated.\n :param image_size: The size of the image to be generated. (default is "1024x1024")\n :return: A list of filenames for the saved images.\n """\n\n client = OpenAI() # Initialize the OpenAI client\n response = client.images.generate(model="dall-e-3", prompt=query, n=1, size=image_size) # Generate images\n\n # List to store the file names of saved images\n saved_files = []\n\n # Check if the response is successful\n if response.data:\n for image_data in response.data:\n # Generate a random UUID as the file name\n file_name = str(uuid.uuid4()) + ".png" # Assuming the image is a PNG\n file_path = Path(file_name)\n\n img_url = image_data.url\n img_response = requests.get(img_url)\n if img_response.status_code == 200:\n # Write the binary content to a file\n with open(file_path, "wb") as img_file:\n img_file.write(img_response.content)\n print(f"Image saved to {file_path}")\n saved_files.append(str(file_path))\n else:\n print(f"Failed to download the image from {img_url}")\n else:\n print("No image data found in the response!")\n\n # Return the list of saved files\n return saved_files\n\n\n# Example usage of the function:\n# generate_and_save_images("A cute baby sea otter")\n',
|
|
user_id="guestuser@gmail.com",
|
|
)
|
|
|
|
# agents
|
|
user_proxy_config = AgentConfig(
|
|
name="user_proxy",
|
|
description="User Proxy Agent Configuration",
|
|
human_input_mode="NEVER",
|
|
max_consecutive_auto_reply=25,
|
|
system_message="You are a helpful assistant",
|
|
code_execution_config=CodeExecutionConfigTypes.local,
|
|
default_auto_reply="TERMINATE",
|
|
llm_config=False,
|
|
)
|
|
user_proxy = Agent(
|
|
user_id="guestuser@gmail.com", type=AgentType.userproxy, config=user_proxy_config.model_dump(mode="json")
|
|
)
|
|
|
|
painter_assistant_config = AgentConfig(
|
|
name="default_assistant",
|
|
description="Assistant Agent",
|
|
human_input_mode="NEVER",
|
|
max_consecutive_auto_reply=25,
|
|
system_message=AssistantAgent.DEFAULT_SYSTEM_MESSAGE,
|
|
code_execution_config=CodeExecutionConfigTypes.none,
|
|
llm_config={},
|
|
)
|
|
painter_assistant = Agent(
|
|
user_id="guestuser@gmail.com", type=AgentType.assistant, config=painter_assistant_config.model_dump(mode="json")
|
|
)
|
|
|
|
planner_assistant_config = AgentConfig(
|
|
name="planner_assistant",
|
|
description="Assistant Agent",
|
|
human_input_mode="NEVER",
|
|
max_consecutive_auto_reply=25,
|
|
system_message="You are a helpful assistant that can suggest a travel plan for a user. You are the primary cordinator who will receive suggestions or advice from other agents (local_assistant, language_assistant). You must ensure that the finally plan integrates the suggestions from other agents or team members. YOUR FINAL RESPONSE MUST BE THE COMPLETE PLAN. When the plan is complete and all perspectives are integrated, you can respond with TERMINATE.",
|
|
code_execution_config=CodeExecutionConfigTypes.none,
|
|
llm_config={},
|
|
)
|
|
planner_assistant = Agent(
|
|
user_id="guestuser@gmail.com", type=AgentType.assistant, config=planner_assistant_config.model_dump(mode="json")
|
|
)
|
|
|
|
local_assistant_config = AgentConfig(
|
|
name="local_assistant",
|
|
description="Local Assistant Agent",
|
|
human_input_mode="NEVER",
|
|
max_consecutive_auto_reply=25,
|
|
system_message="You are a local assistant that can suggest local activities or places to visit for a user. You can suggest local activities, places to visit, restaurants to eat at, etc. You can also provide information about the weather, local events, etc. You can provide information about the local area, but you cannot suggest a complete travel plan. You can only provide information about the local area.",
|
|
code_execution_config=CodeExecutionConfigTypes.none,
|
|
llm_config={},
|
|
)
|
|
local_assistant = Agent(
|
|
user_id="guestuser@gmail.com", type=AgentType.assistant, config=local_assistant_config.model_dump(mode="json")
|
|
)
|
|
|
|
language_assistant_config = AgentConfig(
|
|
name="language_assistant",
|
|
description="Language Assistant Agent",
|
|
human_input_mode="NEVER",
|
|
max_consecutive_auto_reply=25,
|
|
system_message="You are a helpful assistant that can review travel plans, providing feedback on important/critical tips about how best to address language or communication challenges for the given destination. If the plan already includes language tips, you can mention that the plan is satisfactory, with rationale.",
|
|
code_execution_config=CodeExecutionConfigTypes.none,
|
|
llm_config={},
|
|
)
|
|
language_assistant = Agent(
|
|
user_id="guestuser@gmail.com",
|
|
type=AgentType.assistant,
|
|
config=language_assistant_config.model_dump(mode="json"),
|
|
)
|
|
|
|
# group chat
|
|
travel_groupchat_config = AgentConfig(
|
|
name="travel_groupchat",
|
|
admin_name="groupchat",
|
|
description="Group Chat Agent Configuration",
|
|
human_input_mode="NEVER",
|
|
max_consecutive_auto_reply=25,
|
|
system_message="You are a group chat manager",
|
|
code_execution_config=CodeExecutionConfigTypes.none,
|
|
default_auto_reply="TERMINATE",
|
|
llm_config={},
|
|
speaker_selection_method="auto",
|
|
)
|
|
travel_groupchat_agent = Agent(
|
|
user_id="guestuser@gmail.com", type=AgentType.groupchat, config=travel_groupchat_config.model_dump(mode="json")
|
|
)
|
|
|
|
# workflows
|
|
default_workflow = Workflow(name="Default Workflow", description="Default workflow", user_id="guestuser@gmail.com")
|
|
|
|
travel_workflow = Workflow(
|
|
name="Travel Planning Workflow", description="Travel workflow", user_id="guestuser@gmail.com"
|
|
)
|
|
|
|
with Session(dbmanager.engine) as session:
|
|
session.add(zephyr_model)
|
|
session.add(google_gemini_model)
|
|
session.add(azure_model)
|
|
session.add(gpt_4_model)
|
|
session.add(generate_image_skill)
|
|
session.add(user_proxy)
|
|
session.add(painter_assistant)
|
|
session.add(travel_groupchat_agent)
|
|
session.add(planner_assistant)
|
|
session.add(local_assistant)
|
|
session.add(language_assistant)
|
|
|
|
session.add(default_workflow)
|
|
session.add(travel_workflow)
|
|
session.commit()
|
|
|
|
dbmanager.link(link_type="agent_model", primary_id=painter_assistant.id, secondary_id=gpt_4_model.id)
|
|
dbmanager.link(link_type="agent_skill", primary_id=painter_assistant.id, secondary_id=generate_image_skill.id)
|
|
dbmanager.link(
|
|
link_type="workflow_agent", primary_id=default_workflow.id, secondary_id=user_proxy.id, agent_type="sender"
|
|
)
|
|
dbmanager.link(
|
|
link_type="workflow_agent",
|
|
primary_id=default_workflow.id,
|
|
secondary_id=painter_assistant.id,
|
|
agent_type="receiver",
|
|
)
|
|
|
|
# link agents to travel groupchat agent
|
|
|
|
dbmanager.link(link_type="agent_agent", primary_id=travel_groupchat_agent.id, secondary_id=planner_assistant.id)
|
|
dbmanager.link(link_type="agent_agent", primary_id=travel_groupchat_agent.id, secondary_id=local_assistant.id)
|
|
dbmanager.link(
|
|
link_type="agent_agent", primary_id=travel_groupchat_agent.id, secondary_id=language_assistant.id
|
|
)
|
|
dbmanager.link(link_type="agent_agent", primary_id=travel_groupchat_agent.id, secondary_id=user_proxy.id)
|
|
dbmanager.link(link_type="agent_model", primary_id=travel_groupchat_agent.id, secondary_id=gpt_4_model.id)
|
|
dbmanager.link(link_type="agent_model", primary_id=planner_assistant.id, secondary_id=gpt_4_model.id)
|
|
dbmanager.link(link_type="agent_model", primary_id=local_assistant.id, secondary_id=gpt_4_model.id)
|
|
dbmanager.link(link_type="agent_model", primary_id=language_assistant.id, secondary_id=gpt_4_model.id)
|
|
|
|
dbmanager.link(
|
|
link_type="workflow_agent", primary_id=travel_workflow.id, secondary_id=user_proxy.id, agent_type="sender"
|
|
)
|
|
dbmanager.link(
|
|
link_type="workflow_agent",
|
|
primary_id=travel_workflow.id,
|
|
secondary_id=travel_groupchat_agent.id,
|
|
agent_type="receiver",
|
|
)
|
|
logger.info("Successfully initialized database with Default and Travel Planning Workflows")
|