From 3b23d6764fba9c83c867e6e5baa08ebb86eca0c0 Mon Sep 17 00:00:00 2001 From: Masashi Tomooka Date: Sat, 6 Jul 2024 17:53:32 +0900 Subject: [PATCH] fix: token count includes base64 string of input images (#5868) --- api/core/model_runtime/model_providers/bedrock/llm/llm.py | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/api/core/model_runtime/model_providers/bedrock/llm/llm.py b/api/core/model_runtime/model_providers/bedrock/llm/llm.py index f3ea705e19..efb8c395fa 100644 --- a/api/core/model_runtime/model_providers/bedrock/llm/llm.py +++ b/api/core/model_runtime/model_providers/bedrock/llm/llm.py @@ -554,7 +554,10 @@ class BedrockLargeLanguageModel(LargeLanguageModel): content = message.content if isinstance(message, UserPromptMessage): - message_text = f"{human_prompt_prefix} {content} {human_prompt_postfix}" + body = content + if (isinstance(content, list)): + body = "".join([c.data for c in content if c.type == PromptMessageContentType.TEXT]) + message_text = f"{human_prompt_prefix} {body} {human_prompt_postfix}" elif isinstance(message, AssistantPromptMessage): message_text = f"{ai_prompt} {content}" elif isinstance(message, SystemPromptMessage):