* First pass together.ai client class
* Config handling, models and cost
* Added tests, moved param management to create function
* Tests, parameter, validation, logging updates
* Added use of client_utils PR 2949
* Updated to return OAI response
* Notebook example
* Improved function calling, updated tests, updated notebook with Chess example
* Tidied up together client class, better parameter handling, simpler exception capture, warning for no cost, reuse in tests, cleaner tests
* Update of documentation notebook, replacement of old version
* Fix of messages parameter for hide_tools function call
* Update autogen/oai/together.py
Co-authored-by: Qingyun Wu <qingyun0327@gmail.com>
* Update together.py to fix text
---------
Co-authored-by: Qingyun Wu <qingyun0327@gmail.com>
Co-authored-by: Yiran Wu <32823396+yiranwu0@users.noreply.github.com>
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* add claude 3.5 sonnet to pricing
* Fix import error for client_utils
* fix import order for ruff formatter
* name key is not supported in anthropic message so let's remove it
* Improved tool use message conversion, changed create to return standard response
* Converted tools to messages for speaker selection, moved message conversion to function, corrected bugs
* Minor bracket typo.
* Renaming function
* add groupchat and run notebook
---------
Co-authored-by: Mark Sze <mark@sze.family>
Co-authored-by: Qingyun Wu <qingyun0327@gmail.com>
Co-authored-by: HRUSHIKESH DOKALA <96101829+Hk669@users.noreply.github.com>
* Initial commit of Mistral client class
* Updated to manage final system message for reflection_with_llm
* Add Mistral support to client class
* Add Mistral support across the board (based on Gemini changes)
* Test file for Mistral client
* Updated handling of config, added notebook for documentation
* Added support for additional API parameters
* Remove unneeded code, updated exception raising
* Updated handling of keywords, including type checks, defaults, warnings. Updated notebook example to remove logging warnings.
* Added class description.
* Updated tests to support new config handling.
* Moved parameter parsing to create function, minimised init, added parameter tests
* Refined parameter validation
* Correct spacing
* Fixed string concat in parameter validation
* Corrected upper/lower bound warning
* Use client_tools, tidy up Mistral create, better handle tool call response, tidy tests
* Update of documentation notebook, replacement of old version
* Update to handle multiple tool_call recommendations in a message
* Updated tests to accommodate multiple tool_calls as well as content in message
* Update autogen/oai/mistral.py comment
Co-authored-by: Qingyun Wu <qingyun0327@gmail.com>
* cleanup, rewrite mock
* update
---------
Co-authored-by: Qingyun Wu <qingyun0327@gmail.com>
Co-authored-by: kevin666aa <yrwu000627@gmail.com>
* intial setup for the anthropic client with cost config
* update: other methods added
* fix: formatting
* fix: config unused
* update: changes made in the client
* update: test added to the workflow
* update: added tests to the anthropic client
* fix: errors in workflows and client
* fix
* fix: anthropic tools type
* update: notebook anthropic
* Nonetype fixed
* fix-tests config
* update: tests and client issues
* logger support
* remove sys path
* updated the functioning of the client
* update: type hints and stream
* skip tests- importerror
* fix: anthropic client and tests
* none fix
* Alternating roles, parameter keywords, cost on response,
* update: anthropic notebook
* update: notebook with more details
* devcontainer
* update: added validate_params from the client_utils
* fix: formatting
* fix: minor comment
---------
Co-authored-by: Mark Sze <mark@sze.family>
* google default auth and svc keyfile for Gemini
* [.Net] Release note for 0.0.14 (#2815)
* update release note
* update trigger
* [.Net] Update website for AutoGen.SemanticKernel and AutoGen.Ollama (#2814)
support vertex ai compute region
* [CAP] User supplied threads for agents (#2812)
* First pass: message loop in main thread
* pypi version bump
* Fix readme
* Better example
* Fixed docs
* pre-commit fixes
* refactoring, minor fixes, update gemini demo ipynb
* add new deps again and reset line endings
* Docstring for the init function. Use private methods
* improve docstring
---------
Co-authored-by: Xiaoyun Zhang <bigmiao.zhang@gmail.com>
Co-authored-by: Rajan <rajan.chari@yahoo.com>
Co-authored-by: Zoltan Lux <z.lux@campus.tu-berlin.de>
* white spaces
* add llamaindex agent wrapper for autogen
* formatting
* formatting fixes
* add support for llamaindex agents
* fix style
* fix style
* delete file
* re-add file
* fixes pre-commit errors
* feat: Add agentchat_group_chat_with_llamaindex_agents notebook
This commit adds the notebook "agentchat_group_chat_with_llamaindex_agents.ipynb" which demonstrates how to integrate Llamaindex agents into Autogen. The notebook includes code for setting up the API endpoint, creating Llamaindex agents, and setting up a group chat.
* Refactor code
* feat: Add test for LLamaIndexConversableAgent
This commit adds a new test file `test_llamaindex_conversable_agent.py` that contains a test case for the `LLamaIndexConversableAgent` class. The test verifies the functionality of group chat with two MultimodalConversable Agents, limiting the chat by the `max_round` parameter. It also checks if the number of rounds does not exceed the maximum specified rounds.
The purpose of this change is to ensure that the `LLamaIndexConversableAgent` behaves as expected and correctly handles group chats with limited rounds.
Note: This commit includes import statements and setup code necessary for running the test case.
* fix formatting
* feat: Add LlamaIndexAgent job to GitHub Actions workflow
This commit adds a new job called "LlamaIndexAgent" to the GitHub Actions workflow. The job runs on multiple operating systems (ubuntu-latest, macos-latest, windows-2019) and uses Python version 3.11. It sets up the Python environment, installs necessary packages and dependencies for LMM, performs coverage testing using pytest, and uploads the coverage report to Codecov.
The commit also includes changes to the test_llamaindex_conversable_agent.py file. It imports os and sys modules, appends a path to sys.path, and adds skip conditions for tests based on certain conditions.
These changes improve the CI/CD pipeline by adding a new job for LlamaIndexAgent and enhancing test conditions in test_llamaindex_conversable_agent.py.
* fix test yaml
* cleanup tests
* fix test run
* formatting
* add test
* fix yaml
* pr feedback
* add documentation to website
* fixed style
* edit to document page
* newline
* make skip reason easier to see
* compose skip reasons
* fix env variable name
* refactor: Update package installation in contrib workflows
- Replaced specific package installations with more general ones
- Updated the installation of llama-index packages and dependencies
- Added new packages for llama-index and Wikipedia tools
* Update dependencies and add new agents for group chat
- Update dependencies to specific versions
- Add new agent `entertainent_specialist` for discovering entertainment opportunities in a location
- Modify the `test_group_chat_with_llama_index_conversable_agent` function to include the new agent in the group chat
* Update pydantic version requirement in setup.py
- Update pydantic version requirement from "pydantic>=1.10,<3,!=2.6.0" to "pydantic>=1.10,<3"
- Remove comment about issue with pydantic 2.6.0
* Refactor ChatMessage instantiation in _extract_message_and_history()
- Refactored the instantiation of ChatMessage in the _extract_message_and_history() function.
- Added an empty dictionary as additional_kwargs to the ChatMessage constructor.
* Refactor test_llamaindex_conversable_agent.py
- Removed unused import and variable
- Updated OpenAI model to gpt-4
- Reduced max_iterations for location_specialist and entertainment_specialist from 30 to 5
- Changed human_input_mode for user_proxy to "NEVER"
- Added assertion for max_rounds in entertainent_assistant
These changes improve code efficiency and ensure proper functionality.
* Remove entertainment_specialist agent and update user_proxy settings in test_llamaindex_conversable_agent.py
- Remove the creation of entertainent_specialist agent
- Update the max_consecutive_auto_reply setting for user_proxy to 10
- Update the default_auto_reply setting for user_proxy to "Thank you. TERMINATE"
* Refactor installation of LlamaIndex packages and dependencies
- Simplify installation commands for LlamaIndex packages
- Remove specific version numbers from pip install commands
* Update test_llamaindex_conversable_agent.py to include verbose output during pytest.
- Add the -v flag to the pytest command in contrib-openai.yml.
- Print a message when skipping the test due to missing dependencies or key.
* Refactor OpenAI workflow, remove LlamaIndexAgent
This commit removes the LlamaIndexAgent from the OpenAI workflow in order to streamline and simplify the code. The LlamaIndexAgent was no longer necessary and its removal improves overall code organization and maintainability.
* feat: Add test for group chat functionality with LLamaIndexConversableAgent
use mock reactagent in test
* Update Dockerfile for devcontainer
- Updated the Dockerfile for the devcontainer environment.
- Added installation of build-essential, npm, git-lfs, and other packages.
- Upgraded pip and installed pydoc-markdown, pyyaml, and colored libraries.
* Update devcontainer.json with new VS Code extensions and terminal settings
- Updated the list of VS Code extensions in devcontainer.json to include "GitHub.copilot"
- Added a new terminal profile for Linux with the path set to "/bin/bash"
- Set the default Linux terminal profile to "bash"
* removeall
* feat: Add Dockerfiles and devcontainer configurations
This commit adds Dockerfiles and devcontainer configurations for different use cases in the `.devcontainer` directory. The following changes were made:
- Added `Dockerfile` for basic setups (`base`)
- Added `Dockerfile` for advanced features (`full`)
- Added `Dockerfile` for AutoGen project developers (`dev`)
- Added `Dockerfile` for AutoGen project developers using Studio (`studio`)
- Updated existing files with necessary dependencies and configurations
- Modified README.md to provide instructions on customizing Dockerfiles and managing the Docker environment
These changes allow users to easily set up their AutoGen development environment using Docker containers.
* delete
* Add authors.yml file with author information
This commit adds the authors.yml file, which contains information about various authors contributing to the project. Each author entry includes their name, title, URL, and image URL. This file will be used to display author information on the website.
* delete
* Add test cases for agent chat functionality
This commit adds new test cases for the agent chat functionality. The test cases include scenarios such as auto feedback from code execution, function calls, currency calculator, async function calls, group chat finite state machine, cost token tracking, and group chat state flow. These test cases cover different versions of Python (3.10, 3.11, and 3.12) and are skipped if OpenAI is not installed or the Python version does not match.
* delete
* feat: Add LLM configuration documentation
This commit adds documentation for configuring an agent's access to LLMs. It includes information on the `llm_config` argument, `config_list`, and other configuration parameters. The commit also provides examples of filtering the `config_list` based on model names and tags. Additionally, it demonstrates how to add an HTTP client in `llm_config` for proxy usage. Finally, it mentions helper functions for loading a config list from API keys, environment variables, files, or `.env` files.
Closes#1234
* delete
* feat: Add LLM configuration documentation
This commit adds documentation for configuring an agent's access to LLMs. It includes information on the `llm_config` argument, `config_list`, and other configuration parameters. The commit also provides examples of filtering the `config_list` based on model names and tags. Additionally, it demonstrates how to add an HTTP client in `llm_config` for proxy usage. Finally, it mentions helper functions for loading a config list from various sources.
Closes#1234
* delete
* adding back notebooks
* reset
* feat: Add setup.py for package installation
This commit adds a new file, `setup.py`, which is used for installing the package. The `setup.py` file includes information such as the author, description, and dependencies of the package. This allows users to easily install and use the package in their projects.
The `setup.py` file also includes different extra requirements for specific functionalities, such as retrieving chat data or running Jupyter notebooks. These extra requirements are installed when specified during installation.
Overall, this addition improves the usability and installation process of the package.
* Resolving test failures locally
* Resolving test failures locally
* Updates to website resources and docs, author
* Adding image
* Fixes to precommit and doc files for lfd
* Fixing ruff exclusion of new notebook
* Updates to support notebook rendering
* Updates to support notebook rendering
* Removing some results to try to fix docs render issue
* pre-commit to standardize formatting
---------
Co-authored-by: Eric Zhu <ekzhu@users.noreply.github.com>
* Add ability to use virtual environments in local executor
* Copy environment variables from parent environment
* Fix mypy errors and formatting
* Account for venv on Windows
* Use a virtual environment context object instead of path
* Add utility method to create a virtual environment
* Remove assertion using `_venv_path`
* Add tests for `create_virtual_env`
* Modify test code and add output assertion
* Modify test code and assertion
* Execute activation script before actual command on windows
* Add docs for using a virtual env
* Initial resume
* Cleaned up resume function
* Further updating resuming group chat
* Added async resume_chat and documentation
* Added test cases, refined group chat function parameters
* compiled documentation
* Added tests to main
* Removed mdx file.
* Revert "Merge remote-tracking branch 'origin/main' into groupchatresume"
This reverts commit 8f709308f049d6f07f32a93db6900ef05a130751, reversing
changes made to 8bfcb2bff13384877ec6734f6d5ddcb1a028246d.
* Refactored resume to remove initiate_chat
* fix git history
* fix history
* Added clean-up of objects, _groupchat references, and messages_to_string updated
* Added termination-based resumption in notebook, added test cases and improved robustness on resuming messages parameter
---------
Co-authored-by: Eric Zhu <ekzhu@users.noreply.github.com>
* Correct link to Jupyter Code Executor in code-executors.ipynb
The link to the code executor was referencing the wrong folder.
* Update website/docs/tutorial/code-executors.ipynb
Co-authored-by: Eric Zhu <ekzhu@users.noreply.github.com>
* Update website/docs/tutorial/code-executors.ipynb
Co-authored-by: Eric Zhu <ekzhu@users.noreply.github.com>
---------
Co-authored-by: Eric Zhu <ekzhu@users.noreply.github.com>
* Fix for http client
* fixed constructor to only ignore the http_client while copying
* fixed comment formating
* removed check for http_client and added error message with docs
* fix formatting
* fix formatting
* added test for http-fix
* changed title and content of docs
* changed test func name
* Function Calling with GPTAssistantAgent
* Add Link to Notebook in Website
* Add metadata to the notebook
* formatting of H2 and H3 text
* updated to new method of function calling
* Run Pre-commit
* utilize get_function_schema
* Add minimum token threshold in MessageHistoryLimiter
* Update transforms tests for the threshold
* Move min_threshold_tokens from Message to Token Limiter
* Optimize _check_tokens_threshold
Co-authored-by: Wael Karkoub <wael.karkoub96@gmail.com>
* Apply requested changes (renaming, phrasing, validations)
* Fix format
* Fix _check_tokens_threshold logic
* Update docs and notebook
* Improve phrasing
* Add min_tokens example in notebook
* Add min_tokens example in website docs
* Add min_tokens example in notebook
* Update website docs to be in sync with get_logs change
---------
Co-authored-by: Wael Karkoub <wael.karkoub96@gmail.com>
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* PGVector Contrib Initial Commit - KnucklesTeam:autogen:pgvector_contrib fork
* Update website/docs/ecosystem/pgvector.md
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* Updated qdrant installation instructions.
* Fixed openai version.
* Added dependencies to install for qdrant and pgvector in contrib tests.
* Added dependencies to install for qdrant and pgvector in contrib tests.
* Cleaned up dependencies.
* Removed flaml out of setup.py. Used only for notebook example.
* Added PGVector notebook link
---------
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* Use code executors in notebook.
* Add custom code executor guide
* Add to gitignore
* Update
* fix links
* link
* typo
---------
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* Start Gemini integration: works ok with Text now
* Gemini notebook lint
* try catch "import" for Gemini
* Debug: id issue for chat completion in Gemini
* Add RAG example
* Update docs for RAG
* Fix missing pydash
* Remove temp folder
* Fix test error in runs/7206014032/job/19630042864
* Fix tqdm warning
* Fix notebook output
* Gemini's vision model is supported now
* Install instructions for the Gemini branch
* Catch and retry when see Interval Server Error 500
* Allow gemini to take more flexible messages
i.e., it can take messages where "user" is not the last role.
* Use int time for Gemini client
* Handle other exceptions in gemini call
* rename to "create" function for gemini
* GeminiClient compatible with ModelClient now
* Lint
* Update instructions in Gemini notebook
* Lint
* Remove empty blocks from Gemini notebook
* Add gemini into example page
* self.create instead of call
* Add py and Py into python execution
* Remove error code from merging
* Remove pydash dependency for gemini
* Add cloud-gemini doc
* Remove temp file
* cache import update
* Add test case for summary with mm input
* Lint: warnings instead of print
* Add test cases for gemini
* Gemini test config
* Disable default model for gemini
* Typo fix in gemini workflow
* Correct grammar in example notebook
* Raise if "model" is not provided in create(...)
* Move TODOs into a roadmap
* Update .github/workflows/contrib-tests.yml
Co-authored-by: Davor Runje <davor@airt.ai>
* Gemini test config update
* Update setup.py
Co-authored-by: Davor Runje <davor@airt.ai>
* Update test/oai/test_gemini.py
Co-authored-by: Davor Runje <davor@airt.ai>
* Update test/oai/test_gemini.py
Co-authored-by: Davor Runje <davor@airt.ai>
* Remove python 3.8 from gemini
No google's generativeai for Windows with Python 3.8
* Update import error handling for gemini
* Count tokens and cost for gemini
---------
Co-authored-by: Li Jiang <bnujli@gmail.com>
Co-authored-by: Davor Runje <davor@airt.ai>