From 2f7b234cc5486f941b38395c46741014b58e08f6 Mon Sep 17 00:00:00 2001 From: takatost Date: Thu, 17 Aug 2023 16:33:32 +0800 Subject: [PATCH] fix: max token not exist in generate summary when calc rest tokens (#891) --- api/core/generator/llm_generator.py | 1 + 1 file changed, 1 insertion(+) diff --git a/api/core/generator/llm_generator.py b/api/core/generator/llm_generator.py index 91b324c631..034483ebd5 100644 --- a/api/core/generator/llm_generator.py +++ b/api/core/generator/llm_generator.py @@ -51,6 +51,7 @@ class LLMGenerator: prompt_with_empty_context = prompt.format(context='') prompt_tokens = model_instance.get_num_tokens([PromptMessage(content=prompt_with_empty_context)]) max_context_token_length = model_instance.model_rules.max_tokens.max + max_context_token_length = max_context_token_length if max_context_token_length else 1500 rest_tokens = max_context_token_length - prompt_tokens - max_tokens - 1 context = ''