mirror of
https://git.mirrors.martin98.com/https://github.com/langgenius/dify.git
synced 2025-08-15 20:06:03 +08:00
fix: ensure proper conversation role alternation for vLLM (#18837)
This commit is contained in:
parent
c29cb503be
commit
a145c2a8fe
@ -79,9 +79,13 @@ class QuestionClassifierNode(LLMNode):
|
|||||||
memory=memory,
|
memory=memory,
|
||||||
max_token_limit=rest_token,
|
max_token_limit=rest_token,
|
||||||
)
|
)
|
||||||
|
# Some models (e.g. Gemma, Mistral) force roles alternation (user/assistant/user/assistant...).
|
||||||
|
# If both self._get_prompt_template and self._fetch_prompt_messages append a user prompt,
|
||||||
|
# two consecutive user prompts will be generated, causing model's error.
|
||||||
|
# To avoid this, set sys_query to an empty string so that only one user prompt is appended at the end.
|
||||||
prompt_messages, stop = self._fetch_prompt_messages(
|
prompt_messages, stop = self._fetch_prompt_messages(
|
||||||
prompt_template=prompt_template,
|
prompt_template=prompt_template,
|
||||||
sys_query=query,
|
sys_query="",
|
||||||
memory=memory,
|
memory=memory,
|
||||||
model_config=model_config,
|
model_config=model_config,
|
||||||
sys_files=files,
|
sys_files=files,
|
||||||
|
Loading…
x
Reference in New Issue
Block a user