mirror of
https://git.mirrors.martin98.com/https://github.com/langgenius/dify.git
synced 2025-08-19 22:09:11 +08:00
Fix code indentation errors (#9164)
This commit is contained in:
parent
fc60b554a1
commit
6b6e94da08
@ -312,39 +312,39 @@ class AzureOpenAILargeLanguageModel(_CommonAzureOpenAI, LargeLanguageModel):
|
|||||||
if user:
|
if user:
|
||||||
extra_model_kwargs["user"] = user
|
extra_model_kwargs["user"] = user
|
||||||
|
|
||||||
# clear illegal prompt messages
|
# clear illegal prompt messages
|
||||||
prompt_messages = self._clear_illegal_prompt_messages(model, prompt_messages)
|
prompt_messages = self._clear_illegal_prompt_messages(model, prompt_messages)
|
||||||
|
|
||||||
block_as_stream = False
|
|
||||||
if model.startswith("o1"):
|
|
||||||
if stream:
|
|
||||||
block_as_stream = True
|
|
||||||
stream = False
|
|
||||||
|
|
||||||
if "stream_options" in extra_model_kwargs:
|
|
||||||
del extra_model_kwargs["stream_options"]
|
|
||||||
|
|
||||||
if "stop" in extra_model_kwargs:
|
|
||||||
del extra_model_kwargs["stop"]
|
|
||||||
|
|
||||||
# chat model
|
|
||||||
response = client.chat.completions.create(
|
|
||||||
messages=[self._convert_prompt_message_to_dict(m) for m in prompt_messages],
|
|
||||||
model=model,
|
|
||||||
stream=stream,
|
|
||||||
**model_parameters,
|
|
||||||
**extra_model_kwargs,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
block_as_stream = False
|
||||||
|
if model.startswith("o1"):
|
||||||
if stream:
|
if stream:
|
||||||
return self._handle_chat_generate_stream_response(model, credentials, response, prompt_messages, tools)
|
block_as_stream = True
|
||||||
|
stream = False
|
||||||
|
|
||||||
block_result = self._handle_chat_generate_response(model, credentials, response, prompt_messages, tools)
|
if "stream_options" in extra_model_kwargs:
|
||||||
|
del extra_model_kwargs["stream_options"]
|
||||||
|
|
||||||
if block_as_stream:
|
if "stop" in extra_model_kwargs:
|
||||||
return self._handle_chat_block_as_stream_response(block_result, prompt_messages, stop)
|
del extra_model_kwargs["stop"]
|
||||||
|
|
||||||
return block_result
|
# chat model
|
||||||
|
response = client.chat.completions.create(
|
||||||
|
messages=[self._convert_prompt_message_to_dict(m) for m in prompt_messages],
|
||||||
|
model=model,
|
||||||
|
stream=stream,
|
||||||
|
**model_parameters,
|
||||||
|
**extra_model_kwargs,
|
||||||
|
)
|
||||||
|
|
||||||
|
if stream:
|
||||||
|
return self._handle_chat_generate_stream_response(model, credentials, response, prompt_messages, tools)
|
||||||
|
|
||||||
|
block_result = self._handle_chat_generate_response(model, credentials, response, prompt_messages, tools)
|
||||||
|
|
||||||
|
if block_as_stream:
|
||||||
|
return self._handle_chat_block_as_stream_response(block_result, prompt_messages, stop)
|
||||||
|
|
||||||
|
return block_result
|
||||||
|
|
||||||
def _handle_chat_block_as_stream_response(
|
def _handle_chat_block_as_stream_response(
|
||||||
self,
|
self,
|
||||||
|
Loading…
x
Reference in New Issue
Block a user