From 7898937eae301c8f416f93d45cda5712ce9fb87e Mon Sep 17 00:00:00 2001 From: takatost Date: Sun, 13 Aug 2023 13:51:12 +0800 Subject: [PATCH] feat: optimize message return (#822) --- api/core/model_providers/models/llm/base.py | 9 ++++++--- 1 file changed, 6 insertions(+), 3 deletions(-) diff --git a/api/core/model_providers/models/llm/base.py b/api/core/model_providers/models/llm/base.py index 31573dd580..d9216f6b26 100644 --- a/api/core/model_providers/models/llm/base.py +++ b/api/core/model_providers/models/llm/base.py @@ -218,15 +218,18 @@ class BaseLLM(BaseProviderModel): def _get_prompt_from_messages(self, messages: List[PromptMessage], model_mode: Optional[ModelMode] = None) -> Union[str | List[BaseMessage]]: - if len(messages) == 0: - raise ValueError("prompt must not be empty.") - if not model_mode: model_mode = self.model_mode if model_mode == ModelMode.COMPLETION: + if len(messages) == 0: + return '' + return messages[0].content else: + if len(messages) == 0: + return [] + chat_messages = [] for message in messages: if message.type == MessageType.HUMAN: