Files
lang-agent/lang_agent/graphs/react.py
2025-10-22 12:30:06 +08:00

54 lines
1.7 KiB
Python

from dataclasses import dataclass, field, is_dataclass
from typing import Type, List, Callable, Any
import tyro
from lang_agent.config import KeyConfig
from lang_agent.tool_manager import ToolManager, ToolManagerConfig
from lang_agent.base import GraphBase
from langchain.chat_models import init_chat_model
from langchain_core.messages import SystemMessage, HumanMessage
from langchain.agents import create_agent
from langgraph.checkpoint.memory import MemorySaver
@tyro.conf.configure(tyro.conf.SuppressFixed)
@dataclass
class ReactGraphConfig(KeyConfig):
_target: Type = field(default_factory=lambda: ReactGraph)
llm_name: str = "qwen-turbo"
"""name of llm"""
llm_provider:str = "openai"
"""provider of the llm"""
base_url:str = "https://dashscope.aliyuncs.com/compatible-mode/v1"
"""base url; could be used to overwrite the baseurl in llm provider"""
tool_manager_config: ToolManagerConfig = field(default_factory=ToolManagerConfig)
class ReactGraph(GraphBase):
def __init__(self, config: ReactGraphConfig):
self.config = config
self.populate_modules()
def populate_modules(self):
self.llm = init_chat_model(model=self.config.llm_name,
model_provider=self.config.llm_provider,
api_key=self.config.api_key,
base_url=self.config.base_url)
# NOTE: placeholder for now, add graph later
self.tool_manager:ToolManager = self.config.tool_manager_config.setup()
memory = MemorySaver()
tools = self.tool_manager.get_langchain_tools()
self.agent = create_agent(self.llm, tools, checkpointer=memory)
def get_graph(self):
return self.agent