dynamic updates of pipelines from updating pipeline_registry

This commit is contained in:
2026-03-05 11:23:37 +08:00
parent f58d80ac1b
commit 8c6dd3344f

View File

@@ -4,6 +4,7 @@ from pathlib import Path as FsPath
import os.path as osp import os.path as osp
import json import json
import copy import copy
from threading import RLock
from loguru import logger from loguru import logger
from lang_agent.pipeline import Pipeline, PipelineConfig from lang_agent.pipeline import Pipeline, PipelineConfig
@@ -20,6 +21,9 @@ class ServerPipelineManager:
self._api_key_policy: Dict[str, Dict[str, Any]] = {} self._api_key_policy: Dict[str, Dict[str, Any]] = {}
self._pipelines: Dict[str, Pipeline] = {} self._pipelines: Dict[str, Pipeline] = {}
self._pipeline_llm: Dict[str, str] = {} self._pipeline_llm: Dict[str, str] = {}
self._registry_path: Optional[str] = None
self._registry_mtime_ns: Optional[int] = None
self._lock = RLock()
def _resolve_registry_path(self, registry_path: str) -> str: def _resolve_registry_path(self, registry_path: str) -> str:
path = FsPath(registry_path) path = FsPath(registry_path)
@@ -30,39 +34,102 @@ class ServerPipelineManager:
root = FsPath(__file__).resolve().parents[2] root = FsPath(__file__).resolve().parents[2]
return str((root / path).resolve()) return str((root / path).resolve())
def load_registry(self, registry_path: str) -> None: def _stat_registry_mtime_ns(self, abs_path: str) -> int:
abs_path = self._resolve_registry_path(registry_path) return FsPath(abs_path).stat().st_mtime_ns
if not osp.exists(abs_path):
raise ValueError(f"pipeline registry file not found: {abs_path}")
def _read_registry(self, abs_path: str) -> Dict[str, Any]:
with open(abs_path, "r", encoding="utf-8") as f: with open(abs_path, "r", encoding="utf-8") as f:
registry: dict = json.load(f) return json.load(f)
def _apply_registry(self, abs_path: str, registry: Dict[str, Any], mtime_ns: int) -> bool:
pipelines = registry.get("pipelines") pipelines = registry.get("pipelines")
if pipelines is None: if pipelines is None or not isinstance(pipelines, dict):
raise ValueError("`pipelines` in pipeline registry must be an object.") raise ValueError("`pipelines` in pipeline registry must be an object.")
self._pipeline_specs = {} parsed_specs: Dict[str, Dict[str, Any]] = {}
for pipeline_id, spec in pipelines.items(): for pipeline_id, spec in pipelines.items():
if not isinstance(spec, dict): if not isinstance(spec, dict):
raise ValueError( raise ValueError(
f"pipeline spec for `{pipeline_id}` must be an object." f"pipeline spec for `{pipeline_id}` must be an object."
) )
self._pipeline_specs[pipeline_id] = { parsed_specs[pipeline_id] = {
"enabled": bool(spec.get("enabled", True)), "enabled": bool(spec.get("enabled", True)),
"config_file": spec.get("config_file"), "config_file": spec.get("config_file"),
"overrides": spec.get("overrides", {}), "overrides": spec.get("overrides", {}),
} }
if not self._pipeline_specs: if not parsed_specs:
raise ValueError("pipeline registry must define at least one pipeline.") raise ValueError("pipeline registry must define at least one pipeline.")
api_key_policy = registry.get("api_keys", {}) api_key_policy = registry.get("api_keys", {})
if api_key_policy and not isinstance(api_key_policy, dict): if api_key_policy and not isinstance(api_key_policy, dict):
raise ValueError("`api_keys` in pipeline registry must be an object.") raise ValueError("`api_keys` in pipeline registry must be an object.")
with self._lock:
old_specs = self._pipeline_specs
old_policy = self._api_key_policy
old_mtime = self._registry_mtime_ns
removed = set(old_specs.keys()) - set(parsed_specs.keys())
added = set(parsed_specs.keys()) - set(old_specs.keys())
modified = {
pipeline_id
for pipeline_id in (set(old_specs.keys()) & set(parsed_specs.keys()))
if old_specs[pipeline_id] != parsed_specs[pipeline_id]
}
changed = bool(added or removed or modified or old_policy != api_key_policy)
# Drop stale cache entries for deleted/changed pipelines so future requests
# lazily rebuild from the refreshed registry spec.
for pipeline_id in (removed | modified):
self._pipelines.pop(pipeline_id, None)
self._pipeline_llm.pop(pipeline_id, None)
self._pipeline_specs = parsed_specs
self._api_key_policy = api_key_policy self._api_key_policy = api_key_policy
self._registry_path = abs_path
self._registry_mtime_ns = mtime_ns
if changed:
logger.info( logger.info(
f"loaded pipeline registry: {abs_path}, pipelines={list(self._pipeline_specs.keys())}" "refreshed pipeline registry: {} | added={} modified={} removed={} mtime={}",
abs_path,
sorted(added),
sorted(modified),
sorted(removed),
mtime_ns,
) )
elif old_mtime != mtime_ns:
logger.debug("pipeline registry mtime changed but specs were unchanged: {}", abs_path)
return changed
def load_registry(self, registry_path: str) -> None:
abs_path = self._resolve_registry_path(registry_path)
if not osp.exists(abs_path):
raise ValueError(f"pipeline registry file not found: {abs_path}")
registry = self._read_registry(abs_path)
mtime_ns = self._stat_registry_mtime_ns(abs_path)
self._apply_registry(abs_path=abs_path, registry=registry, mtime_ns=mtime_ns)
def refresh_registry_if_needed(
self, registry_path: Optional[str] = None, force: bool = False
) -> bool:
abs_path = (
self._resolve_registry_path(registry_path)
if registry_path
else self._registry_path
)
if not abs_path:
raise ValueError("registry path is not initialized")
if not osp.exists(abs_path):
raise ValueError(f"pipeline registry file not found: {abs_path}")
mtime_ns = self._stat_registry_mtime_ns(abs_path)
with self._lock:
if not force and self._registry_mtime_ns == mtime_ns:
return False
registry = self._read_registry(abs_path)
return self._apply_registry(abs_path=abs_path, registry=registry, mtime_ns=mtime_ns)
def _resolve_config_path(self, config_file: str) -> str: def _resolve_config_path(self, config_file: str) -> str:
path = FsPath(config_file) path = FsPath(config_file)
@@ -138,6 +205,7 @@ class ServerPipelineManager:
or app_id or app_id
) )
with self._lock:
if not pipeline_id: if not pipeline_id:
key_policy = ( key_policy = (
self._api_key_policy.get(api_key, {}) if self._api_key_policy else {} self._api_key_policy.get(api_key, {}) if self._api_key_policy else {}
@@ -155,10 +223,13 @@ class ServerPipelineManager:
return pipeline_id return pipeline_id
def get_pipeline(self, pipeline_id: str) -> Tuple[Pipeline, str]: def get_pipeline(self, pipeline_id: str) -> Tuple[Pipeline, str]:
with self._lock:
cached = self._pipelines.get(pipeline_id) cached = self._pipelines.get(pipeline_id)
if cached is not None: if cached is not None:
return cached, self._pipeline_llm[pipeline_id] return cached, self._pipeline_llm[pipeline_id]
# Build while holding the lock to avoid duplicate construction for
# the same pipeline on concurrent first requests.
pipeline_obj, llm_name = self._build_pipeline(pipeline_id) pipeline_obj, llm_name = self._build_pipeline(pipeline_id)
self._pipelines[pipeline_id] = pipeline_obj self._pipelines[pipeline_id] = pipeline_obj
self._pipeline_llm[pipeline_id] = llm_name self._pipeline_llm[pipeline_id] = llm_name