diff --git a/api/core/workflow/nodes/knowledge_retrieval/knowledge_retrieval_node.py b/api/core/workflow/nodes/knowledge_retrieval/knowledge_retrieval_node.py index c08dc143a6..0caf99a963 100644 --- a/api/core/workflow/nodes/knowledge_retrieval/knowledge_retrieval_node.py +++ b/api/core/workflow/nodes/knowledge_retrieval/knowledge_retrieval_node.py @@ -229,7 +229,9 @@ class KnowledgeRetrievalNode(BaseNode): source["content"] = segment.get_sign_content() retrieval_resource_list.append(source) if retrieval_resource_list: - retrieval_resource_list = sorted(retrieval_resource_list, key=lambda x: x.get("score"), reverse=True) + retrieval_resource_list = sorted( + retrieval_resource_list, key=lambda x: x.get("metadata").get("score"), reverse=True + ) position = 1 for item in retrieval_resource_list: item["metadata"]["position"] = position diff --git a/api/services/website_service.py b/api/services/website_service.py index be01815720..13cc9c679a 100644 --- a/api/services/website_service.py +++ b/api/services/website_service.py @@ -166,6 +166,8 @@ class WebsiteService: @classmethod def get_crawl_url_data(cls, job_id: str, provider: str, url: str, tenant_id: str) -> dict | None: credentials = ApiKeyAuthService.get_auth_credentials(tenant_id, "website", provider) + # decrypt api_key + api_key = encrypter.decrypt_token(tenant_id=tenant_id, token=credentials.get("config").get("api_key")) if provider == "firecrawl": file_key = "website_files/" + job_id + ".txt" if storage.exists(file_key): @@ -173,8 +175,6 @@ class WebsiteService: if data: data = json.loads(data.decode("utf-8")) else: - # decrypt api_key - api_key = encrypter.decrypt_token(tenant_id=tenant_id, token=credentials.get("config").get("api_key")) firecrawl_app = FirecrawlApp(api_key=api_key, base_url=credentials.get("config").get("base_url", None)) result = firecrawl_app.check_crawl_status(job_id) if result.get("status") != "completed":