docs/decisions/0070-declarative-agent-schema.md
This ADR describes a schema which can be used to define an Agent which can be loaded and executed using the Semantic Kernel Agent Framework.
Currently the Agent Framework uses a code first approach to allow Agents to be defined and executed. Using the schema defined by this ADR developers will be able to declaratively define an Agent and have the Semantic Kernel instantiate and execute the Agent.
Here is some pseudo code to illustrate what we need to be able to do:
Kernel kernel = Kernel
.CreateBuilder()
.AddAzureAIClientProvider(...)
.Build();
var text =
"""
type: azureai_agent
name: AzureAIAgent
description: AzureAIAgent Description
instructions: AzureAIAgent Instructions
model:
id: gpt-4o-mini
tools:
- name: tool1
type: code_interpreter
""";
AzureAIAgentFactory factory = new();
var agent = await KernelAgentYaml.FromAgentYamlAsync(kernel, text, factory);
The above code represents the simplest case would work as follows:
Kernel instance has the appropriate services e.g. an instance of AzureAIClientProvider when creating AzureAI agents.KernelAgentYaml.FromAgentYamlAsync will create one of the built-in Agent instances i.e., one of ChatCompletionAgent, OpenAIAssistantsAgent, AzureAIAgent.Kernel instance configured the services and tools it requires and a default initial state.Note: Consider creating just plain Agent instances and extending the Agent abstraction to contain a method which allows the Agent instance to be invoked with user input.
Kernel kernel = ...
string text = EmbeddedResource.Read("MyAgent.yaml");
AgentFactory agentFactory = new AggregatorAgentFactory(
new ChatCompletionAgentFactory(),
new OpenAIAssistantAgentFactory(),
new AzureAIAgentFactory());
var agent = KernelAgentYaml.FromAgentYamlAsync(kernel, text, factory);;
The above example shows how different Agent types are supported.
Note:
Currently Semantic Kernel supports three Agent types and these have the following properties:
ChatCompletionAgent:
Arguments: Optional arguments for the agent. (Inherited from ChatHistoryKernelAgent)Description: The description of the agent (optional). (Inherited from Agent)HistoryReducer: (Inherited from ChatHistoryKernelAgent)Id: The identifier of the agent (optional). (Inherited from Agent)Instructions: The instructions of the agent (optional). (Inherited from KernelAgent)Kernel: The Kernel containing services, plugins, and filters for use throughout the agent lifetime. (Inherited from KernelAgent)Logger: The ILogger associated with this Agent. (Inherited from Agent)LoggerFactory: A ILoggerFactory for this Agent. (Inherited from Agent)Name: The name of the agent (optional). (Inherited from Agent)OpenAIAssistantAgent:
Arguments: Optional arguments for the agent.Definition: The assistant definition.Description: The description of the agent (optional). (Inherited from Agent)Id: The identifier of the agent (optional). (Inherited from Agent)Instructions: The instructions of the agent (optional). (Inherited from KernelAgent)IsDeleted: Set when the assistant has been deleted via DeleteAsync(CancellationToken). An assistant removed by other means will result in an exception when invoked.Kernel: The Kernel containing services, plugins, and filters for use throughout the agent lifetime. (Inherited from KernelAgent)Logger: The ILogger associated with this Agent. (Inherited from Agent)LoggerFactory: A ILoggerFactory for this Agent. (Inherited from Agent)Name: The name of the agent (optional). (Inherited from Agent)PollingOptions: Defines polling behaviorAzureAIAgent
Definition: The assistant definition.PollingOptions: Defines polling behavior for run processing.Description: The description of the agent (optional). (Inherited from Agent)Id: The identifier of the agent (optional). (Inherited from Agent)Instructions: The instructions of the agent (optional). (Inherited from KernelAgent)IsDeleted: Set when the assistant has been deleted via DeleteAsync(CancellationToken). An assistant removed by other means will result in an exception when invoked.Kernel: The Kernel containing services, plugins, and filters for use throughout the agent lifetime. (Inherited from KernelAgent)Logger: The ILogger associated with this Agent. (Inherited from Agent)LoggerFactory: A ILoggerFactory for this Agent. (Inherited from Agent)Name: The name of the agent (optional). (Inherited from Agent)When executing an Agent that was defined declaratively some of the properties will be determined by the runtime:
Kernel: The runtime will be responsible for create the Kernel instance to be used by the Agent. This Kernel instance must be configured with the models and tools that the Agent requires.Logger or LoggerFactory: The runtime will be responsible for providing a correctly configured Logger or LoggerFactory.KernelFunctions defined in the current project. See later in the ADR for an example of this.For Agent properties that define behaviors e.g. HistoryReducer the Semantic Kernel SHOULD:
Kernel e.g., as required services or possibly KernelFunction's.The document will describe the following use cases:
Semantic Kernel already has support this, see the declarative Agent concept sample.
capabilities which includes code interpreter and actions which specifies an API plugin manifest.actions property is focussed on calling REST API's and cater for native and semantic functions.Some of the possible extensions include:
Chosen option: "{title of option 1}", because {justification. e.g., only option, which meets k.o. criterion decision driver | which resolves force {force} | … | comes out best (see below)}.
<!-- This is an optional element. Feel free to remove. -->{describe how the implementation of/compliance with the ADR is validated. E.g., by a review or an ArchUnit test}
<!-- This is an optional element. Feel free to remove. -->Below are examples showing the code first and equivalent declarative syntax for creating different types of Agents.
Consider the following use cases:
ChatCompletionAgentChatCompletionAgent using Prompt TemplateChatCompletionAgent with Function CallingOpenAIAssistantAgent with Function CallingOpenAIAssistantAgent with ToolsChatCompletionAgentCode first approach:
ChatCompletionAgent agent =
new()
{
Name = "Parrot",
Instructions = "Repeat the user message in the voice of a pirate and then end with a parrot sound.",
Kernel = kernel,
};
Declarative Semantic Kernel schema:
type: chat_completion_agent
name: Parrot
instructions: Repeat the user message in the voice of a pirate and then end with a parrot sound.
Note:
ChatCompletionAgent could be the default agent type hence no explicit type property is required.ChatCompletionAgent using Prompt TemplateCode first approach:
string generateStoryYaml = EmbeddedResource.Read("GenerateStory.yaml");
PromptTemplateConfig templateConfig = KernelFunctionYaml.ToPromptTemplateConfig(generateStoryYaml);
ChatCompletionAgent agent =
new(templateConfig, new KernelPromptTemplateFactory())
{
Kernel = this.CreateKernelWithChatCompletion(),
Arguments = new KernelArguments()
{
{ "topic", "Dog" },
{ "length", "3" },
}
};
Agent YAML points to another file, the Declarative Agent implementation in Semantic Kernel already uses this technique to load a separate instructions file.
Prompt template which is used to define the instructions.
---
name: GenerateStory
description: A function that generates a story about a topic.
template:
format: semantic-kernel
parser: semantic-kernel
inputs:
- name: topic
description: The topic of the story.
is_required: true
default: dog
- name: length
description: The number of sentences in the story.
is_required: true
default: 3
---
Tell a story about {{$topic}} that is {{$length}} sentences long.
Note: Semantic Kernel could load this file directly.
ChatCompletionAgent with Function CallingCode first approach:
ChatCompletionAgent agent =
new()
{
Instructions = "Answer questions about the menu.",
Name = "RestaurantHost",
Description = "This agent answers questions about the menu.",
Kernel = kernel,
Arguments = new KernelArguments(new OpenAIPromptExecutionSettings() { Temperature = 0.4, FunctionChoiceBehavior = FunctionChoiceBehavior.Auto() }),
};
KernelPlugin plugin = KernelPluginFactory.CreateFromType<MenuPlugin>();
agent.Kernel.Plugins.Add(plugin);
Declarative using Semantic Kernel schema:
---
name: RestaurantHost
name: RestaurantHost
description: This agent answers questions about the menu.
model:
id: gpt-4o-mini
options:
temperature: 0.4
function_choice_behavior:
type: auto
functions:
- MenuPlugin.GetSpecials
- MenuPlugin.GetItemPrice
---
Answer questions about the menu.
OpenAIAssistantAgent with Function CallingCode first approach:
OpenAIAssistantAgent agent =
await OpenAIAssistantAgent.CreateAsync(
clientProvider: this.GetClientProvider(),
definition: new OpenAIAssistantDefinition("gpt_4o")
{
Instructions = "Answer questions about the menu.",
Name = "RestaurantHost",
Metadata = new Dictionary<string, string> { { AssistantSampleMetadataKey, bool.TrueString } },
},
kernel: new Kernel());
KernelPlugin plugin = KernelPluginFactory.CreateFromType<MenuPlugin>();
agent.Kernel.Plugins.Add(plugin);
Declarative using Semantic Kernel schema:
Using the syntax below the assistant does not have the functions included in it's definition.
The functions must be added to the Kernel instance associated with the Agent and will be passed when the Agent is invoked.
---
name: RestaurantHost
type: openai_assistant
description: This agent answers questions about the menu.
model:
id: gpt-4o-mini
options:
temperature: 0.4
function_choice_behavior:
type: auto
functions:
- MenuPlugin.GetSpecials
- MenuPlugin.GetItemPrice
metadata:
sksample: true
---
Answer questions about the menu.
``
or
```yml
---
name: RestaurantHost
type: openai_assistant
description: This agent answers questions about the menu.
execution_settings:
default:
temperature: 0.4
tools:
- type: function
name: MenuPlugin-GetSpecials
description: Provides a list of specials from the menu.
- type: function
name: MenuPlugin-GetItemPrice
description: Provides the price of the requested menu item.
parameters: '{"type":"object","properties":{"menuItem":{"type":"string","description":"The name of the menu item."}},"required":["menuItem"]}'
---
Answer questions about the menu.
Note: The Kernel instance used to create the Agent must have an instance of OpenAIClientProvider registered as a service.
OpenAIAssistantAgent with ToolsCode first approach:
OpenAIAssistantAgent agent =
await OpenAIAssistantAgent.CreateAsync(
clientProvider: this.GetClientProvider(),
definition: new(this.Model)
{
Instructions = "You are an Agent that can write and execute code to answer questions.",
Name = "Coder",
EnableCodeInterpreter = true,
EnableFileSearch = true,
Metadata = new Dictionary<string, string> { { AssistantSampleMetadataKey, bool.TrueString } },
},
kernel: new Kernel());
Declarative using Semantic Kernel:
---
name: Coder
type: openai_assistant
tools:
- type: code_interpreter
- type: file_search
---
You are an Agent that can write and execute code to answer questions.
name: RestaurantHost
type: azureai_agent
description: This agent answers questions about the menu.
version: 0.0.1