model capabilities FAQ (#4471)

* model capabilities faq

* Add links to FAQ in more places

---------

Co-authored-by: Eric Zhu <ekzhu@users.noreply.github.com>
This commit is contained in:
peterychang 2024-12-06 02:53:22 -05:00 committed by GitHub
parent 5f61ba0c2f
commit b286c99c1f
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
3 changed files with 30 additions and 1 deletions

View File

@ -157,7 +157,9 @@
"source": [
"## Local Models\n",
"\n",
"We are working on it. Stay tuned!"
"See [this guide](../../core-user-guide/faqs.md#what-are-model-capabilities-and-how-do-i-specify-them) for how to override a model's default capabilities definitions in autogen.\n",
"\n",
"More to come. Stay tuned!"
]
}
],

View File

@ -34,3 +34,27 @@ worker1 = GrpcWorkerAgentRuntime(host_address=host_address, extra_grpc_config=ex
```
**Note**: When `GrpcWorkerAgentRuntime` creates a host connection for the clients, it uses `DEFAULT_GRPC_CONFIG` from `HostConnection` class as default set of values which will can be overriden if you pass parameters with the same name using `extra_grpc_config`.
## What are model capabilities and how do I specify them?
Model capabilites are additional capabilities an LLM may have beyond the standard natural language features. There are currently 3 additional capabilities that can be specified within Autogen
- vision: The model is capable of processing and interpreting image data.
- function_calling: The model has the capacity to accept function descriptions; such as the function name, purpose, input parameters, etc; and can respond with an appropriate function to call including any necessary parameters.
- json_output: The model is capable of outputting responses to conform with a specified json format.
Model capabilities can be passed into a model, which will override the default definitions. These capabilities will not affect what the underlying model is actually capable of, but will allow or disallow behaviors associated with them. This is particularly useful when [using local LLMs](cookbook/local-llms-ollama-litellm.ipynb).
```python
from autogen_ext.models import OpenAIChatCompletionClient
client = OpenAIChatCompletionClient(
model="gpt-4o",
api_key="YourApiKey",
model_capabilities={
"vision": True,
"function_calling": False,
"json_output": False,
}
)
```

View File

@ -96,6 +96,9 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"Default [Model Capabilities](../faqs.md#what-are-model-capabilities-and-how-do-i-specify-them) may be overridden should the need arise.\n",
"\n",
"\n",
"### Streaming Response\n",
"\n",
"You can use the {py:meth}`~autogen_ext.models.OpenAIChatCompletionClient.create_streaming` method to create a\n",