Compare commits
10 Commits
9b3db40b94
...
e049e4e104
| Author | SHA1 | Date | |
|---|---|---|---|
| e049e4e104 | |||
| 8d0cd7861b | |||
| 7920a1bab3 | |||
| 30e93ed0a7 | |||
| 0087ac41e6 | |||
| 5358b46113 | |||
| 3306da038d | |||
| ac99dfd56b | |||
| 3932d695bf | |||
| 9ed141ba42 |
@@ -39,6 +39,7 @@ type EditableAgent = {
|
||||
prompts: Record<string, string>;
|
||||
apiKey: string;
|
||||
llmName: string;
|
||||
actBackend: DeepAgentActBackend;
|
||||
};
|
||||
|
||||
type AgentChatMessage = {
|
||||
@@ -48,6 +49,7 @@ type AgentChatMessage = {
|
||||
};
|
||||
|
||||
type ActiveTab = "agents" | "discussions" | "mcp";
|
||||
type DeepAgentActBackend = "state_bk" | "local_shell" | "daytona_sandbox";
|
||||
type McpTransport = "streamable_http" | "sse" | "stdio";
|
||||
type McpEntry = {
|
||||
id: string;
|
||||
@@ -62,6 +64,15 @@ type McpEntry = {
|
||||
|
||||
const DEFAULT_LLM_NAME = "qwen-plus";
|
||||
const DEFAULT_API_KEY = "";
|
||||
const DEFAULT_DEEPAGENT_ACT_BACKEND: DeepAgentActBackend = "state_bk";
|
||||
const DEEPAGENT_BACKEND_OPTIONS: Array<{
|
||||
value: DeepAgentActBackend;
|
||||
label: string;
|
||||
}> = [
|
||||
{ value: "state_bk", label: "state_bk" },
|
||||
{ value: "local_shell", label: "local_shell" },
|
||||
{ value: "daytona_sandbox", label: "daytona_sandbox" },
|
||||
];
|
||||
const LOCAL_DASHSCOPE_BASE = "http://127.0.0.1:8500/v1/apps";
|
||||
const MCP_TRANSPORT_OPTIONS: McpTransport[] = ["streamable_http", "sse", "stdio"];
|
||||
const GRAPH_ARCH_IMAGE_MODULES = import.meta.glob(
|
||||
@@ -74,6 +85,9 @@ const FALLBACK_PROMPTS_BY_GRAPH: Record<string, Record<string, string>> = {
|
||||
chat_prompt: "",
|
||||
tool_prompt: "",
|
||||
},
|
||||
deepagent: {
|
||||
sys_prompt: "",
|
||||
},
|
||||
react: {
|
||||
sys_prompt: "",
|
||||
},
|
||||
@@ -412,6 +426,26 @@ function createConversationId(): string {
|
||||
return `conv-${Date.now()}-${Math.random().toString(36).slice(2, 8)}`;
|
||||
}
|
||||
|
||||
function normalizeDeepAgentActBackend(value: unknown): DeepAgentActBackend {
|
||||
if (value === "local_shell" || value === "localshell") {
|
||||
return "local_shell";
|
||||
}
|
||||
if (value === "daytona_sandbox" || value === "daytonasandbox") {
|
||||
return "daytona_sandbox";
|
||||
}
|
||||
if (value === "state_bk" || value === "statebk") {
|
||||
return "state_bk";
|
||||
}
|
||||
return DEFAULT_DEEPAGENT_ACT_BACKEND;
|
||||
}
|
||||
|
||||
function buildGraphParams(editor: EditableAgent): Record<string, unknown> {
|
||||
if (editor.graphId === "deepagent") {
|
||||
return { act_bkend: editor.actBackend };
|
||||
}
|
||||
return {};
|
||||
}
|
||||
|
||||
function toEditable(
|
||||
config: GraphConfigReadResponse,
|
||||
draft: boolean
|
||||
@@ -428,6 +462,7 @@ function toEditable(
|
||||
prompts: config.prompt_dict || {},
|
||||
apiKey: config.api_key || DEFAULT_API_KEY,
|
||||
llmName: DEFAULT_LLM_NAME,
|
||||
actBackend: DEFAULT_DEEPAGENT_ACT_BACKEND,
|
||||
};
|
||||
}
|
||||
|
||||
@@ -455,6 +490,7 @@ export default function App() {
|
||||
const [chatInput, setChatInput] = useState<string>("");
|
||||
const [chatMessages, setChatMessages] = useState<AgentChatMessage[]>([]);
|
||||
const [chatSending, setChatSending] = useState(false);
|
||||
const [chatAbortController, setChatAbortController] = useState<AbortController | null>(null);
|
||||
const [busy, setBusy] = useState(false);
|
||||
|
||||
const configKeySet = useMemo(
|
||||
@@ -714,6 +750,10 @@ export default function App() {
|
||||
graphId,
|
||||
prompts: { ...defaults.prompt_dict },
|
||||
toolKeys: defaults.tool_keys || [],
|
||||
actBackend:
|
||||
graphId === "deepagent"
|
||||
? prev.actBackend || DEFAULT_DEEPAGENT_ACT_BACKEND
|
||||
: DEFAULT_DEEPAGENT_ACT_BACKEND,
|
||||
};
|
||||
if (next.isDraft) {
|
||||
setDraftAgents((drafts) => drafts.map((draft) => (draft.id === next.id ? next : draft)));
|
||||
@@ -905,9 +945,8 @@ export default function App() {
|
||||
const active = await getPipelineDefaultConfig(editor.pipelineId.trim());
|
||||
targetPromptSetId = active.prompt_set_id;
|
||||
} catch {
|
||||
throw new Error(
|
||||
"No active prompt set for this pipeline. Create/activate one via backend first."
|
||||
);
|
||||
// For a brand-new pipeline, let backend create a new prompt set.
|
||||
targetPromptSetId = undefined;
|
||||
}
|
||||
}
|
||||
const upsertResp = await upsertGraphConfig({
|
||||
@@ -929,6 +968,7 @@ export default function App() {
|
||||
api_key: editor.apiKey.trim(),
|
||||
llm_name: editor.llmName || DEFAULT_LLM_NAME,
|
||||
enabled: isEditorRunning,
|
||||
graph_params: buildGraphParams(editor),
|
||||
});
|
||||
await refreshRunning();
|
||||
} catch (error) {
|
||||
@@ -1012,6 +1052,7 @@ export default function App() {
|
||||
api_key: editor.apiKey.trim(),
|
||||
llm_name: editor.llmName,
|
||||
enabled: true,
|
||||
graph_params: buildGraphParams(editor),
|
||||
});
|
||||
await refreshRunning();
|
||||
if (resp.reload_required) {
|
||||
@@ -1081,6 +1122,9 @@ export default function App() {
|
||||
}
|
||||
|
||||
const authKey = runtimeFastApiKey.trim() || "dev-key";
|
||||
const controller = new AbortController();
|
||||
setChatAbortController(controller);
|
||||
|
||||
const userMessage: AgentChatMessage = {
|
||||
id: `user-${Date.now()}`,
|
||||
role: "user",
|
||||
@@ -1104,6 +1148,7 @@ export default function App() {
|
||||
sessionId: chatConversationId,
|
||||
apiKey: authKey,
|
||||
message,
|
||||
signal: controller.signal,
|
||||
onText: (text) => {
|
||||
setChatMessages((prev) =>
|
||||
prev.map((item) =>
|
||||
@@ -1118,6 +1163,19 @@ export default function App() {
|
||||
},
|
||||
});
|
||||
} catch (error) {
|
||||
if ((error as Error).message === "Request cancelled") {
|
||||
setChatMessages((prev) =>
|
||||
prev.map((item) =>
|
||||
item.id === assistantMessageId
|
||||
? {
|
||||
...item,
|
||||
content: item.content + "\n\n*[Stopped]*",
|
||||
}
|
||||
: item
|
||||
)
|
||||
);
|
||||
return;
|
||||
}
|
||||
const messageText = (error as Error).message || "Chat request failed.";
|
||||
setStatusMessage(messageText);
|
||||
setChatMessages((prev) =>
|
||||
@@ -1132,6 +1190,7 @@ export default function App() {
|
||||
);
|
||||
} finally {
|
||||
setChatSending(false);
|
||||
setChatAbortController(null);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1337,6 +1396,28 @@ export default function App() {
|
||||
/>
|
||||
</label>
|
||||
|
||||
{editor.graphId === "deepagent" ? (
|
||||
<label>
|
||||
act_bkend
|
||||
<select
|
||||
value={editor.actBackend}
|
||||
onChange={(e) =>
|
||||
updateEditor(
|
||||
"actBackend",
|
||||
normalizeDeepAgentActBackend(e.target.value)
|
||||
)
|
||||
}
|
||||
disabled={busy}
|
||||
>
|
||||
{DEEPAGENT_BACKEND_OPTIONS.map((option) => (
|
||||
<option key={option.value} value={option.value}>
|
||||
{option.label}
|
||||
</option>
|
||||
))}
|
||||
</select>
|
||||
</label>
|
||||
) : null}
|
||||
|
||||
<div className="prompt-section">
|
||||
<h3>Prompts</h3>
|
||||
{Object.keys(editor.prompts).length === 0 ? (
|
||||
@@ -1670,7 +1751,11 @@ export default function App() {
|
||||
className={`chat-modal-message ${message.role === "assistant" ? "assistant" : "user"}`}
|
||||
>
|
||||
<strong>{message.role === "assistant" ? "Agent" : "You"}</strong>
|
||||
<p>{message.content || (chatSending && message.role === "assistant" ? "..." : "")}</p>
|
||||
<div className="chat-message-content">
|
||||
<ReactMarkdown remarkPlugins={[remarkGfm]}>
|
||||
{message.content || (chatSending && message.role === "assistant" ? "..." : "")}
|
||||
</ReactMarkdown>
|
||||
</div>
|
||||
</article>
|
||||
))
|
||||
)}
|
||||
@@ -1679,19 +1764,41 @@ export default function App() {
|
||||
<textarea
|
||||
value={chatInput}
|
||||
onChange={(event) => setChatInput(event.target.value)}
|
||||
placeholder="Type your message..."
|
||||
onKeyDown={(event) => {
|
||||
if (event.key === "Enter" && !event.shiftKey) {
|
||||
event.preventDefault();
|
||||
if (chatInput.trim() && !chatSending) {
|
||||
sendAgentChatMessage().catch(() => undefined);
|
||||
}
|
||||
}
|
||||
}}
|
||||
placeholder="Type your message... (Enter to send, Shift+Enter for new line)"
|
||||
rows={3}
|
||||
disabled={chatSending}
|
||||
/>
|
||||
<button
|
||||
type="button"
|
||||
onClick={() => {
|
||||
sendAgentChatMessage().catch(() => undefined);
|
||||
}}
|
||||
disabled={chatSending || !chatInput.trim()}
|
||||
>
|
||||
Send (Stream)
|
||||
</button>
|
||||
<div className="chat-modal-actions">
|
||||
{chatSending ? (
|
||||
<button
|
||||
type="button"
|
||||
className="chat-stop-button"
|
||||
onClick={() => {
|
||||
chatAbortController?.abort();
|
||||
}}
|
||||
>
|
||||
Stop
|
||||
</button>
|
||||
) : (
|
||||
<button
|
||||
type="button"
|
||||
onClick={() => {
|
||||
sendAgentChatMessage().catch(() => undefined);
|
||||
}}
|
||||
disabled={chatSending || !chatInput.trim()}
|
||||
>
|
||||
Send (Stream)
|
||||
</button>
|
||||
)}
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
</div>
|
||||
|
||||
@@ -173,6 +173,7 @@ type StreamAgentChatOptions = {
|
||||
apiKey: string;
|
||||
message: string;
|
||||
onText: (text: string) => void;
|
||||
signal?: AbortSignal;
|
||||
};
|
||||
|
||||
function parseErrorDetail(payload: unknown): string | null {
|
||||
@@ -186,7 +187,7 @@ function parseErrorDetail(payload: unknown): string | null {
|
||||
export async function streamAgentChatResponse(
|
||||
options: StreamAgentChatOptions
|
||||
): Promise<string> {
|
||||
const { appId, sessionId, apiKey, message, onText } = options;
|
||||
const { appId, sessionId, apiKey, message, onText, signal } = options;
|
||||
const response = await fetch(
|
||||
`${API_BASE_URL}/v1/apps/${encodeURIComponent(appId)}/sessions/${encodeURIComponent(sessionId)}/responses`,
|
||||
{
|
||||
@@ -199,6 +200,7 @@ export async function streamAgentChatResponse(
|
||||
messages: [{ role: "user", content: message }],
|
||||
stream: true,
|
||||
}),
|
||||
signal,
|
||||
}
|
||||
);
|
||||
|
||||
@@ -226,6 +228,10 @@ export async function streamAgentChatResponse(
|
||||
let latestText = "";
|
||||
|
||||
while (true) {
|
||||
if (signal?.aborted) {
|
||||
reader.cancel();
|
||||
throw new Error("Request cancelled");
|
||||
}
|
||||
const { value, done } = await reader.read();
|
||||
if (done) {
|
||||
break;
|
||||
|
||||
@@ -618,6 +618,7 @@ button:disabled {
|
||||
display: grid;
|
||||
gap: 8px;
|
||||
grid-template-columns: 1fr auto;
|
||||
align-items: start;
|
||||
}
|
||||
|
||||
.chat-modal-input textarea {
|
||||
@@ -628,3 +629,23 @@ button:disabled {
|
||||
resize: vertical;
|
||||
}
|
||||
|
||||
.chat-modal-actions {
|
||||
display: flex;
|
||||
gap: 8px;
|
||||
height: 100%;
|
||||
}
|
||||
|
||||
.chat-modal-actions button {
|
||||
height: auto;
|
||||
white-space: nowrap;
|
||||
}
|
||||
|
||||
.chat-stop-button {
|
||||
background-color: #dc3545;
|
||||
color: white;
|
||||
}
|
||||
|
||||
.chat-stop-button:hover {
|
||||
background-color: #bb2d3b;
|
||||
}
|
||||
|
||||
|
||||
@@ -55,6 +55,7 @@ export type PipelineCreateRequest = {
|
||||
api_key?: string;
|
||||
llm_name: string;
|
||||
enabled?: boolean;
|
||||
graph_params?: Record<string, unknown>;
|
||||
};
|
||||
|
||||
export type PipelineSpec = {
|
||||
@@ -63,7 +64,6 @@ export type PipelineSpec = {
|
||||
enabled: boolean;
|
||||
config_file: string;
|
||||
llm_name: string;
|
||||
overrides: Record<string, unknown>;
|
||||
};
|
||||
|
||||
export type PipelineCreateResponse = {
|
||||
|
||||
@@ -55,7 +55,7 @@ class ServerPipelineManager:
|
||||
parsed_specs[pipeline_id] = {
|
||||
"enabled": bool(spec.get("enabled", True)),
|
||||
"config_file": spec.get("config_file"),
|
||||
"overrides": spec.get("overrides", {}),
|
||||
"llm_name": spec.get("llm_name"),
|
||||
}
|
||||
if not parsed_specs:
|
||||
raise ValueError("pipeline registry must define at least one pipeline.")
|
||||
@@ -152,7 +152,7 @@ class ServerPipelineManager:
|
||||
)
|
||||
|
||||
config_file = spec.get("config_file")
|
||||
overrides = spec.get("overrides", {})
|
||||
registry_llm_name = spec.get("llm_name")
|
||||
if config_file:
|
||||
loaded_cfg = load_tyro_conf(self._resolve_config_path(config_file))
|
||||
if hasattr(loaded_cfg, "setup"):
|
||||
@@ -166,19 +166,11 @@ class ServerPipelineManager:
|
||||
)
|
||||
else:
|
||||
cfg = copy.deepcopy(self.default_config)
|
||||
if not isinstance(overrides, dict):
|
||||
raise ValueError(
|
||||
f"pipeline `overrides` for `{pipeline_id}` must be an object."
|
||||
)
|
||||
for key, value in overrides.items():
|
||||
if not hasattr(cfg, key):
|
||||
raise ValueError(
|
||||
f"unknown override field `{key}` for pipeline `{pipeline_id}`"
|
||||
)
|
||||
setattr(cfg, key, value)
|
||||
if registry_llm_name is not None and hasattr(cfg, "llm_name"):
|
||||
setattr(cfg, "llm_name", registry_llm_name)
|
||||
|
||||
p = cfg.setup()
|
||||
llm_name = getattr(cfg, "llm_name", "unknown-model")
|
||||
llm_name = str(getattr(cfg, "llm_name", registry_llm_name or "unknown-model"))
|
||||
return p, llm_name
|
||||
|
||||
def _authorize(self, api_key: str, pipeline_id: str) -> None:
|
||||
|
||||
@@ -85,6 +85,8 @@ class PipelineCreateRequest(BaseModel):
|
||||
api_key: Optional[str] = Field(default=None)
|
||||
llm_name: str = Field(default="qwen-plus")
|
||||
enabled: bool = Field(default=True)
|
||||
# Arbitrary per-graph options forwarded to the graph build function.
|
||||
graph_params: Dict[str, Any] = Field(default_factory=dict)
|
||||
|
||||
|
||||
class PipelineSpec(BaseModel):
|
||||
@@ -93,7 +95,6 @@ class PipelineSpec(BaseModel):
|
||||
enabled: bool
|
||||
config_file: str
|
||||
llm_name: str
|
||||
overrides: Dict[str, Any] = Field(default_factory=dict)
|
||||
|
||||
|
||||
class PipelineCreateResponse(BaseModel):
|
||||
@@ -314,19 +315,13 @@ def _resolve_runtime_fast_api_key() -> RuntimeAuthInfoResponse:
|
||||
def _normalize_pipeline_spec(pipeline_id: str, spec: Dict[str, Any]) -> PipelineSpec:
|
||||
if not isinstance(spec, dict):
|
||||
raise ValueError(f"pipeline spec for '{pipeline_id}' must be an object")
|
||||
overrides = spec.get("overrides", {})
|
||||
if overrides is None:
|
||||
overrides = {}
|
||||
if not isinstance(overrides, dict):
|
||||
raise ValueError(f"`overrides` for pipeline '{pipeline_id}' must be an object")
|
||||
llm_name = str(overrides.get("llm_name") or "unknown")
|
||||
llm_name = str(spec.get("llm_name") or "unknown")
|
||||
return PipelineSpec(
|
||||
pipeline_id=pipeline_id,
|
||||
graph_id=str(spec.get("graph_id") or pipeline_id),
|
||||
enabled=bool(spec.get("enabled", True)),
|
||||
config_file=str(spec.get("config_file") or ""),
|
||||
llm_name=llm_name,
|
||||
overrides=overrides,
|
||||
)
|
||||
|
||||
|
||||
@@ -616,6 +611,7 @@ async def create_pipeline(body: PipelineCreateRequest):
|
||||
|
||||
config_file = f"configs/pipelines/{pipeline_id}.yaml"
|
||||
config_abs_dir = osp.join(_PROJECT_ROOT, "configs", "pipelines")
|
||||
extra_params = dict(body.graph_params or {})
|
||||
try:
|
||||
build_fn(
|
||||
pipeline_id=pipeline_id,
|
||||
@@ -624,6 +620,7 @@ async def create_pipeline(body: PipelineCreateRequest):
|
||||
api_key=resolved_api_key,
|
||||
llm_name=body.llm_name,
|
||||
pipeline_config_dir=config_abs_dir,
|
||||
**extra_params,
|
||||
)
|
||||
|
||||
update_pipeline_registry(
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
from typing import Any, Dict, List
|
||||
from typing import Any, Dict, List, Literal
|
||||
import os
|
||||
import os.path as osp
|
||||
import subprocess
|
||||
@@ -7,6 +7,15 @@ import json
|
||||
from lang_agent.config.core_config import load_tyro_conf
|
||||
from lang_agent.config.constants import TY_BUILD_SCRIPT, _PROJECT_ROOT
|
||||
|
||||
_DEEP_AGENT_BACKEND_ALIASES = {
|
||||
"state_bk": "statebk",
|
||||
"statebk": "statebk",
|
||||
"local_shell": "localshell",
|
||||
"localshell": "localshell",
|
||||
"daytona_sandbox": "daytonasandbox",
|
||||
"daytonasandbox": "daytonasandbox",
|
||||
}
|
||||
|
||||
|
||||
def opt_to_config(save_path: str, *nargs):
|
||||
os.makedirs(osp.dirname(save_path), exist_ok=True)
|
||||
@@ -50,7 +59,7 @@ def update_pipeline_registry(
|
||||
pipeline["enabled"] = bool(enabled)
|
||||
pipeline["config_file"] = config_file
|
||||
pipeline["graph_id"] = graph_id
|
||||
pipeline["overrides"] = {"llm_name": llm_name}
|
||||
pipeline["llm_name"] = llm_name
|
||||
|
||||
with open(registry_f, "w", encoding="utf-8") as f:
|
||||
json.dump(registry, f, indent=4)
|
||||
@@ -62,7 +71,8 @@ def build_route(
|
||||
tool_keys: List[str],
|
||||
api_key: str,
|
||||
llm_name: str = "qwen-plus",
|
||||
pipeline_config_dir="configs/pipelines",
|
||||
pipeline_config_dir: str = "configs/pipelines",
|
||||
**_: Any,
|
||||
):
|
||||
cmd_opt = [
|
||||
"--pipeline.pipeline-id", pipeline_id,
|
||||
@@ -92,7 +102,8 @@ def build_react(
|
||||
tool_keys: List[str],
|
||||
api_key: str,
|
||||
llm_name: str = "qwen-plus",
|
||||
pipeline_config_dir="configs/pipelines",
|
||||
pipeline_config_dir: str = "configs/pipelines",
|
||||
**_: Any,
|
||||
):
|
||||
cmd_opt = [
|
||||
"--pipeline.pipeline-id", pipeline_id,
|
||||
@@ -110,8 +121,50 @@ def build_react(
|
||||
return _build_and_load_pipeline_config(pipeline_id, pipeline_config_dir, cmd_opt)
|
||||
|
||||
|
||||
def build_deep_agent(
|
||||
pipeline_id: str,
|
||||
prompt_set: str,
|
||||
tool_keys: List[str],
|
||||
api_key: str,
|
||||
llm_name: str = "qwen-plus",
|
||||
pipeline_config_dir: str = "configs/pipelines",
|
||||
act_bkend: Literal[
|
||||
"local_shell",
|
||||
"localshell",
|
||||
"state_bk",
|
||||
"statebk",
|
||||
"daytona_sandbox",
|
||||
"daytonasandbox",
|
||||
] = "state_bk",
|
||||
**_: Any,
|
||||
):
|
||||
backend_subcommand = _DEEP_AGENT_BACKEND_ALIASES.get(act_bkend)
|
||||
if backend_subcommand is None:
|
||||
raise ValueError(
|
||||
"Unsupported deepagent backend "
|
||||
f"'{act_bkend}'. Expected one of {sorted(_DEEP_AGENT_BACKEND_ALIASES.keys())}"
|
||||
)
|
||||
|
||||
cmd_opt = [
|
||||
"--pipeline.pipeline-id", pipeline_id,
|
||||
"deepagent",
|
||||
"--llm-name", llm_name,
|
||||
"--api-key", api_key,
|
||||
"--pipeline-id", pipeline_id,
|
||||
"--prompt-set-id", prompt_set,
|
||||
backend_subcommand,
|
||||
]
|
||||
|
||||
if tool_keys:
|
||||
cmd_opt.extend(
|
||||
["--tool-manager-config.client-tool-manager.tool-keys", *tool_keys]
|
||||
)
|
||||
|
||||
return _build_and_load_pipeline_config(pipeline_id, pipeline_config_dir, cmd_opt)
|
||||
|
||||
# {pipeline_id: build_function}
|
||||
GRAPH_BUILD_FNCS = {
|
||||
"routing": build_route,
|
||||
"react": build_react,
|
||||
"deepagent": build_deep_agent,
|
||||
}
|
||||
|
||||
@@ -38,4 +38,24 @@ class LocalShell(BaseFilesystemBackend):
|
||||
self.backend = LocalShellBackend(root_dir=self.config.workspace_dir,
|
||||
virtual_mode=True,
|
||||
# env={"PATH": "/usr/bin:/bin"}
|
||||
inherit_env=True)
|
||||
inherit_env=True)
|
||||
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
import sys
|
||||
|
||||
# Instantiate a LocalShell instance with the default config
|
||||
config = LocalShellConfig()
|
||||
shell = LocalShell(config)
|
||||
|
||||
# Try checking access to 'npx'
|
||||
try:
|
||||
result = shell.backend.execute("npx --version")
|
||||
if result.exit_code == 0:
|
||||
print("npx is available, version:", result.output.strip())
|
||||
else:
|
||||
print("npx returned non-zero exit code:", result.exit_code, file=sys.stderr)
|
||||
print("output:", result.output, file=sys.stderr)
|
||||
except Exception as e:
|
||||
print("Could not access 'npx':", str(e), file=sys.stderr)
|
||||
@@ -61,7 +61,12 @@ class DeepAgent(GraphBase):
|
||||
checkpointer=self.mem,
|
||||
**bkend_agent_params)
|
||||
|
||||
self.prompt_store = build_prompt_store(file_path=self.config.sys_prompt_f, default_key="sys_prompt")
|
||||
self.prompt_store = build_prompt_store(
|
||||
pipeline_id=self.config.pipeline_id,
|
||||
prompt_set_id=self.config.prompt_set_id,
|
||||
file_path=self.config.sys_prompt_f,
|
||||
default_key="sys_prompt",
|
||||
)
|
||||
self.sys_prompt = self.prompt_store.get("sys_prompt")
|
||||
|
||||
def _agent_call(self, state:State):
|
||||
|
||||
@@ -160,3 +160,30 @@ ON CONFLICT (prompt_set_id, prompt_key)
|
||||
DO UPDATE SET
|
||||
content = EXCLUDED.content,
|
||||
updated_at = now();
|
||||
|
||||
-- Seed: initial prompt set for lang_agent/graphs/deepagents_qt.py
|
||||
-- DeepAgent uses prompt key "sys_prompt" with DB-first, file-fallback loading.
|
||||
INSERT INTO prompt_sets (pipeline_id, graph_id, name, description, is_active, list)
|
||||
SELECT
|
||||
'deepagent',
|
||||
'deepagent',
|
||||
'default',
|
||||
'Initial prompt set for DeepAgent',
|
||||
true,
|
||||
''
|
||||
WHERE NOT EXISTS (
|
||||
SELECT 1
|
||||
FROM prompt_sets
|
||||
WHERE pipeline_id = 'deepagent'
|
||||
AND name = 'default'
|
||||
);
|
||||
|
||||
INSERT INTO prompt_templates (prompt_set_id, prompt_key, content)
|
||||
SELECT ps.id, 'sys_prompt', '你是一个擅长调用工具和处理文件任务的深度代理。'
|
||||
FROM prompt_sets ps
|
||||
WHERE ps.pipeline_id = 'deepagent'
|
||||
AND ps.name = 'default'
|
||||
ON CONFLICT (prompt_set_id, prompt_key)
|
||||
DO UPDATE SET
|
||||
content = EXCLUDED.content,
|
||||
updated_at = now();
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
Simple chat loop to interact with the blueberry pipeline via DashScope-compatible API.
|
||||
|
||||
Usage:
|
||||
python chat_blueberry.py
|
||||
python scripts/py_scripts/chat_dashcope.py
|
||||
|
||||
The script connects to the server running on http://localhost:8500
|
||||
and uses the API key from the pipeline registry.
|
||||
|
||||
@@ -325,3 +325,27 @@ def test_pipeline_conversation_messages_404(monkeypatch):
|
||||
resp = client.get("/v1/pipelines/agent-a/conversations/agent-b:conv-9/messages")
|
||||
assert resp.status_code == 404, resp.text
|
||||
assert "not found for pipeline 'agent-a'" in resp.json()["detail"]
|
||||
|
||||
|
||||
def test_runtime_auth_info_prefers_registry_then_env(monkeypatch, tmp_path):
|
||||
registry_path = tmp_path / "pipeline_registry.json"
|
||||
registry_path.write_text(
|
||||
json.dumps(
|
||||
{
|
||||
"pipelines": {},
|
||||
"api_keys": {
|
||||
"sk-from-registry": {"default_pipeline_id": "blueberry"},
|
||||
},
|
||||
}
|
||||
),
|
||||
encoding="utf-8",
|
||||
)
|
||||
monkeypatch.setattr(front_apis, "PIPELINE_REGISTRY_PATH", str(registry_path))
|
||||
monkeypatch.setenv("FAST_AUTH_KEYS", "sk-from-env,other")
|
||||
|
||||
client = TestClient(front_apis.app)
|
||||
resp = client.get("/v1/runtime-auth")
|
||||
assert resp.status_code == 200, resp.text
|
||||
data = resp.json()
|
||||
assert data["fast_api_key"] == "sk-from-registry"
|
||||
assert data["source"] == "pipeline_registry"
|
||||
|
||||
@@ -35,7 +35,7 @@ def test_refresh_registry_picks_up_new_pipeline(tmp_path):
|
||||
"default": {
|
||||
"enabled": True,
|
||||
"config_file": None,
|
||||
"overrides": {"llm_name": "qwen-plus"},
|
||||
"llm_name": "qwen-plus",
|
||||
}
|
||||
},
|
||||
)
|
||||
@@ -57,12 +57,12 @@ def test_refresh_registry_picks_up_new_pipeline(tmp_path):
|
||||
"default": {
|
||||
"enabled": True,
|
||||
"config_file": None,
|
||||
"overrides": {"llm_name": "qwen-plus"},
|
||||
"llm_name": "qwen-plus",
|
||||
},
|
||||
"blueberry": {
|
||||
"enabled": True,
|
||||
"config_file": None,
|
||||
"overrides": {"llm_name": "qwen-max"},
|
||||
"llm_name": "qwen-max",
|
||||
},
|
||||
},
|
||||
)
|
||||
@@ -83,7 +83,7 @@ def test_refresh_registry_invalidates_cache_for_changed_pipeline(tmp_path):
|
||||
"blueberry": {
|
||||
"enabled": True,
|
||||
"config_file": None,
|
||||
"overrides": {"llm_name": "qwen-plus"},
|
||||
"llm_name": "qwen-plus",
|
||||
}
|
||||
},
|
||||
)
|
||||
@@ -102,7 +102,7 @@ def test_refresh_registry_invalidates_cache_for_changed_pipeline(tmp_path):
|
||||
"blueberry": {
|
||||
"enabled": True,
|
||||
"config_file": None,
|
||||
"overrides": {"llm_name": "qwen-max"},
|
||||
"llm_name": "qwen-max",
|
||||
}
|
||||
},
|
||||
)
|
||||
@@ -122,7 +122,7 @@ def test_refresh_registry_applies_disabled_state_immediately(tmp_path):
|
||||
"blueberry": {
|
||||
"enabled": True,
|
||||
"config_file": None,
|
||||
"overrides": {"llm_name": "qwen-plus"},
|
||||
"llm_name": "qwen-plus",
|
||||
}
|
||||
},
|
||||
)
|
||||
@@ -139,7 +139,7 @@ def test_refresh_registry_applies_disabled_state_immediately(tmp_path):
|
||||
"blueberry": {
|
||||
"enabled": False,
|
||||
"config_file": None,
|
||||
"overrides": {"llm_name": "qwen-plus"},
|
||||
"llm_name": "qwen-plus",
|
||||
}
|
||||
},
|
||||
)
|
||||
@@ -152,3 +152,5 @@ def test_refresh_registry_applies_disabled_state_immediately(tmp_path):
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
Reference in New Issue
Block a user