From c9eb9353495072d3389e159328e41e642763cf55 Mon Sep 17 00:00:00 2001 From: goulustis Date: Fri, 10 Oct 2025 21:48:05 +0800 Subject: [PATCH] fixes --- lang_agent/rag/emb.py | 9 +++++---- 1 file changed, 5 insertions(+), 4 deletions(-) diff --git a/lang_agent/rag/emb.py b/lang_agent/rag/emb.py index 43d884b..675ced3 100644 --- a/lang_agent/rag/emb.py +++ b/lang_agent/rag/emb.py @@ -31,11 +31,12 @@ class QwenEmbeddings(Embeddings): dashscope.api_key = api_key if api_key is None: logger.warning("no api_key provided!!") - + + self.MAX_BATCH_SIZE = 10 self.model = model self.max_workers = max_workers self.embedding_dimension = embedding_dimension - self.batch_size = min(batch_size, 10) # DashScope limit + self.batch_size = min(batch_size, self.MAX_BATCH_SIZE) # DashScope limit self.rate_limit_delay = rate_limit_delay def _get_batch_embeddings(self, texts: List[str]) -> List[List[float]]: @@ -166,7 +167,7 @@ class QwenEmbeddings(Embeddings): if batch_size is not None and batch_size != self.batch_size: # Temporarily override batch size original_batch_size = self.batch_size - self.batch_size = min(batch_size, 10) + self.batch_size = min(batch_size, self.MAX_BATCH_SIZE) try: return self.embed_documents(texts) finally: @@ -185,7 +186,7 @@ class QwenEmbeddings(Embeddings): if batch_size is not None and batch_size != self.batch_size: # Temporarily override batch size original_batch_size = self.batch_size - self.batch_size = min(batch_size, 10) + self.batch_size = min(batch_size, self.MAX_BATCH_SIZE) try: return await self.aembed_documents(texts) finally: