mirror of
https://git.mirrors.martin98.com/https://github.com/langgenius/dify.git
synced 2025-07-20 19:54:25 +08:00
add bedrock deepseek-r1 (#18908)
This commit is contained in:
parent
9f7d8a981f
commit
2e097a1ac0
@ -58,6 +58,7 @@ class BedrockLargeLanguageModel(LargeLanguageModel):
|
|||||||
# TODO There is invoke issue: context limit on Cohere Model, will add them after fixed.
|
# TODO There is invoke issue: context limit on Cohere Model, will add them after fixed.
|
||||||
CONVERSE_API_ENABLED_MODEL_INFO = [
|
CONVERSE_API_ENABLED_MODEL_INFO = [
|
||||||
{"prefix": "anthropic.claude-v2", "support_system_prompts": True, "support_tool_use": False},
|
{"prefix": "anthropic.claude-v2", "support_system_prompts": True, "support_tool_use": False},
|
||||||
|
{"prefix": "us.deepseek", "support_system_prompts": True, "support_tool_use": False},
|
||||||
{"prefix": "anthropic.claude-v1", "support_system_prompts": True, "support_tool_use": False},
|
{"prefix": "anthropic.claude-v1", "support_system_prompts": True, "support_tool_use": False},
|
||||||
{"prefix": "us.anthropic.claude-3", "support_system_prompts": True, "support_tool_use": True},
|
{"prefix": "us.anthropic.claude-3", "support_system_prompts": True, "support_tool_use": True},
|
||||||
{"prefix": "eu.anthropic.claude-3", "support_system_prompts": True, "support_tool_use": True},
|
{"prefix": "eu.anthropic.claude-3", "support_system_prompts": True, "support_tool_use": True},
|
||||||
|
@ -0,0 +1,63 @@
|
|||||||
|
model: us.deepseek.r1-v1:0
|
||||||
|
label:
|
||||||
|
en_US: DeepSeek-R1(US.Cross Region Inference)
|
||||||
|
icon: icon_s_en.svg
|
||||||
|
model_type: llm
|
||||||
|
features:
|
||||||
|
- agent-thought
|
||||||
|
- vision
|
||||||
|
- tool-call
|
||||||
|
- stream-tool-call
|
||||||
|
model_properties:
|
||||||
|
mode: chat
|
||||||
|
context_size: 32768
|
||||||
|
parameter_rules:
|
||||||
|
- name: max_tokens
|
||||||
|
use_template: max_tokens
|
||||||
|
required: true
|
||||||
|
label:
|
||||||
|
zh_Hans: 最大token数
|
||||||
|
en_US: Max Tokens
|
||||||
|
type: int
|
||||||
|
default: 8192
|
||||||
|
min: 1
|
||||||
|
max: 128000
|
||||||
|
help:
|
||||||
|
zh_Hans: 停止前生成的最大令牌数。
|
||||||
|
en_US: The maximum number of tokens to generate before stopping.
|
||||||
|
- name: temperature
|
||||||
|
use_template: temperature
|
||||||
|
required: false
|
||||||
|
label:
|
||||||
|
zh_Hans: 模型温度
|
||||||
|
en_US: Model Temperature
|
||||||
|
type: float
|
||||||
|
default: 1
|
||||||
|
min: 0.0
|
||||||
|
max: 1.0
|
||||||
|
help:
|
||||||
|
zh_Hans: 生成内容的随机性。当推理功能启用时,该值将被固定为1。
|
||||||
|
en_US: The amount of randomness injected into the response. When reasoning is enabled, this value will be fixed to 1.
|
||||||
|
- name: top_p
|
||||||
|
show_on:
|
||||||
|
- variable: reasoning_type
|
||||||
|
value: disabled
|
||||||
|
use_template: top_p
|
||||||
|
label:
|
||||||
|
zh_Hans: Top P
|
||||||
|
en_US: Top P
|
||||||
|
required: false
|
||||||
|
type: float
|
||||||
|
default: 0.999
|
||||||
|
min: 0.000
|
||||||
|
max: 1.000
|
||||||
|
help:
|
||||||
|
zh_Hans: 在核采样中的概率阈值。当推理功能启用时,该参数将被禁用。
|
||||||
|
en_US: The probability threshold in nucleus sampling. When reasoning is enabled, this parameter will be disabled.
|
||||||
|
- name: response_format
|
||||||
|
use_template: response_format
|
||||||
|
pricing:
|
||||||
|
input: '0.001'
|
||||||
|
output: '0.005'
|
||||||
|
unit: '0.001'
|
||||||
|
currency: USD
|
Loading…
x
Reference in New Issue
Block a user