-
Notifications
You must be signed in to change notification settings - Fork 33
Expand file tree
/
Copy pathagent.py
More file actions
147 lines (125 loc) · 4.88 KB
/
agent.py
File metadata and controls
147 lines (125 loc) · 4.88 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
from typing import Callable, Sequence, Type, TypeVar
from langchain_core.language_models import BaseChatModel
from langchain_core.messages import HumanMessage, SystemMessage
from langchain_core.tools import BaseTool
from langgraph.constants import END, START
from langgraph.graph import StateGraph
from pydantic import BaseModel
from uipath.platform.guardrails import BaseGuardrail
from ..guardrails.actions import GuardrailAction
from .guardrails.guardrails_subgraph import (
create_agent_init_guardrails_subgraph,
create_agent_terminate_guardrails_subgraph,
create_llm_guardrails_subgraph,
create_tools_guardrails_subgraph,
)
from .init_node import (
create_init_node,
)
from .llm_node import (
create_llm_node,
)
from .router import (
create_route_agent,
)
from .router_conversational import create_route_agent_conversational
from .terminate_node import (
create_terminate_node,
)
from .tools import create_flow_control_tools
from .types import (
AgentGraphConfig,
AgentGraphNode,
AgentGraphState,
)
from .utils import create_state_with_input
InputT = TypeVar("InputT", bound=BaseModel)
OutputT = TypeVar("OutputT", bound=BaseModel)
def create_agent(
model: BaseChatModel,
tools: Sequence[BaseTool],
messages: Sequence[SystemMessage | HumanMessage]
| Callable[[InputT], Sequence[SystemMessage | HumanMessage]],
*,
input_schema: Type[InputT] | None = None,
output_schema: Type[OutputT] | None = None,
config: AgentGraphConfig | None = None,
guardrails: Sequence[tuple[BaseGuardrail, GuardrailAction]] | None = None,
) -> StateGraph[AgentGraphState, None, InputT, OutputT]:
"""Build agent graph with INIT -> AGENT (subgraph) <-> TOOLS loop, terminated by control flow tools.
The AGENT node is a subgraph that runs:
- before-agent guardrail middlewares
- the LLM tool-executing node
- after-agent guardrail middlewares
Control flow tools (end_execution, raise_error) are auto-injected alongside regular tools.
"""
from ..tools import create_tool_node
if config is None:
config = AgentGraphConfig()
agent_tools = list(tools)
flow_control_tools: list[BaseTool] = (
[] if config.is_conversational else create_flow_control_tools(output_schema)
)
llm_tools: list[BaseTool] = [*agent_tools, *flow_control_tools]
init_node = create_init_node(messages, input_schema, config.is_conversational)
tool_nodes = create_tool_node(agent_tools)
tool_nodes_with_guardrails = create_tools_guardrails_subgraph(
tool_nodes, guardrails, input_schema=input_schema
)
terminate_node = create_terminate_node(output_schema, config.is_conversational)
CompleteAgentGraphState = create_state_with_input(
input_schema if input_schema is not None else BaseModel
)
builder: StateGraph[AgentGraphState, None, InputT, OutputT] = StateGraph(
CompleteAgentGraphState, input_schema=input_schema, output_schema=output_schema
)
init_with_guardrails_subgraph = create_agent_init_guardrails_subgraph(
(AgentGraphNode.GUARDED_INIT, init_node),
guardrails,
input_schema=input_schema,
)
builder.add_node(AgentGraphNode.INIT, init_with_guardrails_subgraph)
for tool_name, tool_node in tool_nodes_with_guardrails.items():
builder.add_node(tool_name, tool_node)
terminate_with_guardrails_subgraph = create_agent_terminate_guardrails_subgraph(
(AgentGraphNode.GUARDED_TERMINATE, terminate_node),
guardrails,
input_schema=input_schema,
)
builder.add_node(AgentGraphNode.TERMINATE, terminate_with_guardrails_subgraph)
builder.add_edge(START, AgentGraphNode.INIT)
llm_node = create_llm_node(
model,
llm_tools,
config.is_conversational,
config.llm_messages_limit,
config.thinking_messages_limit,
)
llm_with_guardrails_subgraph = create_llm_guardrails_subgraph(
(AgentGraphNode.LLM, llm_node), guardrails, input_schema=input_schema
)
builder.add_node(AgentGraphNode.AGENT, llm_with_guardrails_subgraph)
builder.add_edge(AgentGraphNode.INIT, AgentGraphNode.AGENT)
tool_node_names = list(tool_nodes_with_guardrails.keys())
if config.is_conversational:
route_agent = create_route_agent_conversational()
target_node_names = [
*tool_node_names,
AgentGraphNode.TERMINATE,
]
else:
route_agent = create_route_agent(config.thinking_messages_limit)
target_node_names = [
AgentGraphNode.AGENT,
*tool_node_names,
AgentGraphNode.TERMINATE,
]
builder.add_conditional_edges(
AgentGraphNode.AGENT,
route_agent,
target_node_names,
)
for tool_name in tool_node_names:
builder.add_conditional_edges(tool_name, route_agent, target_node_names)
builder.add_edge(AgentGraphNode.TERMINATE, END)
return builder