From 9ab9f0c36e2449359c9529c37ccce26e23c9420a Mon Sep 17 00:00:00 2001 From: goulustis Date: Thu, 12 Feb 2026 15:12:42 +0800 Subject: [PATCH] update tool_node config --- lang_agent/graphs/tool_nodes.py | 16 +++++----------- 1 file changed, 5 insertions(+), 11 deletions(-) diff --git a/lang_agent/graphs/tool_nodes.py b/lang_agent/graphs/tool_nodes.py index 1905605..d12586c 100644 --- a/lang_agent/graphs/tool_nodes.py +++ b/lang_agent/graphs/tool_nodes.py @@ -23,17 +23,11 @@ from langgraph.graph import StateGraph, START, END @dataclass -class ToolNodeConfig(InstantiateConfig): +class ToolNodeConfig(LLMNodeConfig): _target: Type = field(default_factory=lambda: ToolNode) tool_prompt_f:str = osp.join(osp.dirname(osp.dirname(osp.dirname(__file__))), "configs", "route_sys_prompts", "tool_prompt.txt") - pipeline_id: Optional[str] = None - """If set, load prompts from database (with file fallback)""" - - prompt_set_id: Optional[str] = None - """If set, load from this specific prompt set instead of the active one""" - class ToolNode(ToolNodeBase): def __init__(self, config: ToolNodeConfig, @@ -46,7 +40,9 @@ class ToolNode(ToolNodeBase): self.populate_modules() def populate_modules(self): - self.llm = make_llm(tags=["tool_llm"]) + self.llm = make_llm(model=self.config.llm_name, + api_key=self.config.api_key, + tags=["tool_llm"]) self.tool_agent = create_agent(self.llm, self.tool_manager.get_langchain_tools(), checkpointer=self.mem) self.prompt_store = build_prompt_store( @@ -85,14 +81,12 @@ class ToolNode(ToolNodeBase): @dataclass -class ChattyToolNodeConfig(LLMNodeConfig, ToolNodeConfig): +class ChattyToolNodeConfig(LLMNodeConfig): _target: Type = field(default_factory=lambda: ChattyToolNode) chatty_sys_prompt_f:str = osp.join(osp.dirname(osp.dirname(osp.dirname(__file__))), "configs", "route_sys_prompts", "chatty_prompt.txt") """path to chatty system prompt""" - # pipeline_id and prompt_set_id are inherited from ToolNodeConfig - tool_node_conf:ToolNodeConfig = field(default_factory=ToolNodeConfig)