From 5d2f7136dd3fede33545f8f97c3ffca79ea51379 Mon Sep 17 00:00:00 2001 From: KevinHuSh Date: Fri, 31 May 2024 15:45:11 +0800 Subject: [PATCH] fix chunk modification bug (#1011) ### What problem does this PR solve? As title. ### Type of change - [x] Bug Fix (non-breaking change which fixes an issue) --- api/db/services/llm_service.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/api/db/services/llm_service.py b/api/db/services/llm_service.py index c484afcc3..5906927b7 100644 --- a/api/db/services/llm_service.py +++ b/api/db/services/llm_service.py @@ -82,9 +82,9 @@ class TenantLLMService(CommonService): if model_config: model_config = model_config.to_dict() if not model_config: if llm_type in [LLMType.EMBEDDING, LLMType.RERANK]: - llm = LLMService.query(llm_name=llm_name) + llm = LLMService.query(llm_name=llm_name if llm_name else mdlnm) if llm and llm[0].fid in ["Youdao", "FastEmbed", "BAAI"]: - model_config = {"llm_factory": llm[0].fid, "api_key":"", "llm_name": llm_name, "api_base": ""} + model_config = {"llm_factory": llm[0].fid, "api_key":"", "llm_name": llm_name if llm_name else mdlnm, "api_base": ""} if not model_config: if llm_name == "flag-embedding": model_config = {"llm_factory": "Tongyi-Qianwen", "api_key": "",