mirror of
https://git.mirrors.martin98.com/https://github.com/langgenius/dify.git
synced 2025-08-12 06:19:03 +08:00
feat: add completion mode and context size options for LLM configuration (#13325)
Signed-off-by: -LAN- <laipz8200@outlook.com>
This commit is contained in:
parent
f9515901cc
commit
413dfd5628
@ -1,4 +1,4 @@
|
|||||||
from .llm_entities import LLMResult, LLMResultChunk, LLMResultChunkDelta, LLMUsage
|
from .llm_entities import LLMMode, LLMResult, LLMResultChunk, LLMResultChunkDelta, LLMUsage
|
||||||
from .message_entities import (
|
from .message_entities import (
|
||||||
AssistantPromptMessage,
|
AssistantPromptMessage,
|
||||||
AudioPromptMessageContent,
|
AudioPromptMessageContent,
|
||||||
@ -23,6 +23,7 @@ __all__ = [
|
|||||||
"AudioPromptMessageContent",
|
"AudioPromptMessageContent",
|
||||||
"DocumentPromptMessageContent",
|
"DocumentPromptMessageContent",
|
||||||
"ImagePromptMessageContent",
|
"ImagePromptMessageContent",
|
||||||
|
"LLMMode",
|
||||||
"LLMResult",
|
"LLMResult",
|
||||||
"LLMResultChunk",
|
"LLMResultChunk",
|
||||||
"LLMResultChunkDelta",
|
"LLMResultChunkDelta",
|
||||||
|
@ -51,6 +51,40 @@ model_credential_schema:
|
|||||||
show_on:
|
show_on:
|
||||||
- variable: __model_type
|
- variable: __model_type
|
||||||
value: llm
|
value: llm
|
||||||
|
- variable: mode
|
||||||
|
show_on:
|
||||||
|
- variable: __model_type
|
||||||
|
value: llm
|
||||||
|
label:
|
||||||
|
en_US: Completion mode
|
||||||
|
type: select
|
||||||
|
required: false
|
||||||
|
default: chat
|
||||||
|
placeholder:
|
||||||
|
zh_Hans: 选择对话类型
|
||||||
|
en_US: Select completion mode
|
||||||
|
options:
|
||||||
|
- value: completion
|
||||||
|
label:
|
||||||
|
en_US: Completion
|
||||||
|
zh_Hans: 补全
|
||||||
|
- value: chat
|
||||||
|
label:
|
||||||
|
en_US: Chat
|
||||||
|
zh_Hans: 对话
|
||||||
|
- variable: context_size
|
||||||
|
label:
|
||||||
|
zh_Hans: 模型上下文长度
|
||||||
|
en_US: Model context size
|
||||||
|
required: true
|
||||||
|
show_on:
|
||||||
|
- variable: __model_type
|
||||||
|
value: llm
|
||||||
|
type: text-input
|
||||||
|
default: "4096"
|
||||||
|
placeholder:
|
||||||
|
zh_Hans: 在此输入您的模型上下文长度
|
||||||
|
en_US: Enter your Model context size
|
||||||
- variable: jwt_token
|
- variable: jwt_token
|
||||||
required: true
|
required: true
|
||||||
label:
|
label:
|
||||||
|
@ -20,7 +20,7 @@ from azure.core.exceptions import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
from core.model_runtime.callbacks.base_callback import Callback
|
from core.model_runtime.callbacks.base_callback import Callback
|
||||||
from core.model_runtime.entities.llm_entities import LLMResult, LLMResultChunk, LLMResultChunkDelta, LLMUsage
|
from core.model_runtime.entities.llm_entities import LLMMode, LLMResult, LLMResultChunk, LLMResultChunkDelta, LLMUsage
|
||||||
from core.model_runtime.entities.message_entities import (
|
from core.model_runtime.entities.message_entities import (
|
||||||
AssistantPromptMessage,
|
AssistantPromptMessage,
|
||||||
PromptMessage,
|
PromptMessage,
|
||||||
@ -30,6 +30,7 @@ from core.model_runtime.entities.model_entities import (
|
|||||||
AIModelEntity,
|
AIModelEntity,
|
||||||
FetchFrom,
|
FetchFrom,
|
||||||
I18nObject,
|
I18nObject,
|
||||||
|
ModelPropertyKey,
|
||||||
ModelType,
|
ModelType,
|
||||||
ParameterRule,
|
ParameterRule,
|
||||||
ParameterType,
|
ParameterType,
|
||||||
@ -334,7 +335,10 @@ class AzureAIStudioLargeLanguageModel(LargeLanguageModel):
|
|||||||
fetch_from=FetchFrom.CUSTOMIZABLE_MODEL,
|
fetch_from=FetchFrom.CUSTOMIZABLE_MODEL,
|
||||||
model_type=ModelType.LLM,
|
model_type=ModelType.LLM,
|
||||||
features=[],
|
features=[],
|
||||||
model_properties={},
|
model_properties={
|
||||||
|
ModelPropertyKey.CONTEXT_SIZE: int(credentials.get("context_size", "4096")),
|
||||||
|
ModelPropertyKey.MODE: credentials.get("mode", LLMMode.CHAT),
|
||||||
|
},
|
||||||
parameter_rules=rules,
|
parameter_rules=rules,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -3,8 +3,7 @@ from typing import Any, Optional
|
|||||||
|
|
||||||
from pydantic import BaseModel, Field, field_validator
|
from pydantic import BaseModel, Field, field_validator
|
||||||
|
|
||||||
from core.model_runtime.entities import ImagePromptMessageContent
|
from core.model_runtime.entities import ImagePromptMessageContent, LLMMode
|
||||||
from core.model_runtime.entities.llm_entities import LLMMode
|
|
||||||
from core.prompt.entities.advanced_prompt_entities import ChatModelMessage, CompletionModelPromptTemplate, MemoryConfig
|
from core.prompt.entities.advanced_prompt_entities import ChatModelMessage, CompletionModelPromptTemplate, MemoryConfig
|
||||||
from core.workflow.entities.variable_entities import VariableSelector
|
from core.workflow.entities.variable_entities import VariableSelector
|
||||||
from core.workflow.nodes.base import BaseNodeData
|
from core.workflow.nodes.base import BaseNodeData
|
||||||
@ -13,7 +12,7 @@ from core.workflow.nodes.base import BaseNodeData
|
|||||||
class ModelConfig(BaseModel):
|
class ModelConfig(BaseModel):
|
||||||
provider: str
|
provider: str
|
||||||
name: str
|
name: str
|
||||||
mode: LLMMode = LLMMode.COMPLETION
|
mode: LLMMode
|
||||||
completion_params: dict[str, Any] = {}
|
completion_params: dict[str, Any] = {}
|
||||||
|
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user