fix: ensure proper conversation role alternation for vLLM (#18837)

This commit is contained in:
Contrail 2025-06-03 13:47:39 +09:00 committed by GitHub
parent c29cb503be
commit a145c2a8fe
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -79,9 +79,13 @@ class QuestionClassifierNode(LLMNode):
memory=memory, memory=memory,
max_token_limit=rest_token, max_token_limit=rest_token,
) )
# Some models (e.g. Gemma, Mistral) force roles alternation (user/assistant/user/assistant...).
# If both self._get_prompt_template and self._fetch_prompt_messages append a user prompt,
# two consecutive user prompts will be generated, causing model's error.
# To avoid this, set sys_query to an empty string so that only one user prompt is appended at the end.
prompt_messages, stop = self._fetch_prompt_messages( prompt_messages, stop = self._fetch_prompt_messages(
prompt_template=prompt_template, prompt_template=prompt_template,
sys_query=query, sys_query="",
memory=memory, memory=memory,
model_config=model_config, model_config=model_config,
sys_files=files, sys_files=files,