mirror of
https://git.mirrors.martin98.com/https://github.com/infiniflow/ragflow.git
synced 2025-08-14 14:35:52 +08:00
fix minimax bug (#1528)
### What problem does this PR solve? #1353 ### Type of change - [x] Bug Fix (non-breaking change which fixes an issue)
This commit is contained in:
parent
2a647162a8
commit
607de74ace
@ -573,42 +573,35 @@ def init_llm_factory():
|
|||||||
# ------------------------ Minimax -----------------------
|
# ------------------------ Minimax -----------------------
|
||||||
{
|
{
|
||||||
"fid": factory_infos[13]["name"],
|
"fid": factory_infos[13]["name"],
|
||||||
"llm_name": "abab6.5-chat",
|
"llm_name": "abab6.5",
|
||||||
"tags": "LLM,CHAT,8k",
|
"tags": "LLM,CHAT,8k",
|
||||||
"max_tokens": 8192,
|
"max_tokens": 8192,
|
||||||
"model_type": LLMType.CHAT.value
|
"model_type": LLMType.CHAT.value
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"fid": factory_infos[13]["name"],
|
"fid": factory_infos[13]["name"],
|
||||||
"llm_name": "abab6.5s-chat",
|
"llm_name": "abab6.5s",
|
||||||
"tags": "LLM,CHAT,245k",
|
"tags": "LLM,CHAT,245k",
|
||||||
"max_tokens": 245760,
|
"max_tokens": 245760,
|
||||||
"model_type": LLMType.CHAT.value
|
"model_type": LLMType.CHAT.value
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"fid": factory_infos[13]["name"],
|
"fid": factory_infos[13]["name"],
|
||||||
"llm_name": "abab6.5t-chat",
|
"llm_name": "abab6.5t",
|
||||||
"tags": "LLM,CHAT,8k",
|
"tags": "LLM,CHAT,8k",
|
||||||
"max_tokens": 8192,
|
"max_tokens": 8192,
|
||||||
"model_type": LLMType.CHAT.value
|
"model_type": LLMType.CHAT.value
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"fid": factory_infos[13]["name"],
|
"fid": factory_infos[13]["name"],
|
||||||
"llm_name": "abab6.5g-chat",
|
"llm_name": "abab6.5g",
|
||||||
"tags": "LLM,CHAT,8k",
|
"tags": "LLM,CHAT,8k",
|
||||||
"max_tokens": 8192,
|
"max_tokens": 8192,
|
||||||
"model_type": LLMType.CHAT.value
|
"model_type": LLMType.CHAT.value
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"fid": factory_infos[13]["name"],
|
"fid": factory_infos[13]["name"],
|
||||||
"llm_name": "abab5.5-chat",
|
"llm_name": "abab5.5s",
|
||||||
"tags": "LLM,CHAT,16k",
|
|
||||||
"max_tokens": 16384,
|
|
||||||
"model_type": LLMType.CHAT.value
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"fid": factory_infos[13]["name"],
|
|
||||||
"llm_name": "abab5.5s-chat",
|
|
||||||
"tags": "LLM,CHAT,8k",
|
"tags": "LLM,CHAT,8k",
|
||||||
"max_tokens": 8192,
|
"max_tokens": 8192,
|
||||||
"model_type": LLMType.CHAT.value
|
"model_type": LLMType.CHAT.value
|
||||||
@ -987,6 +980,8 @@ def init_llm_factory():
|
|||||||
LLMFactoriesService.save(**info)
|
LLMFactoriesService.save(**info)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
LLMService.filter_delete([(LLM.fid == "MiniMax" or LLM.fid == "Minimax")])
|
||||||
for info in llm_infos:
|
for info in llm_infos:
|
||||||
try:
|
try:
|
||||||
LLMService.save(**info)
|
LLMService.save(**info)
|
||||||
|
@ -61,6 +61,7 @@ ChatModel = {
|
|||||||
"VolcEngine": VolcEngineChat,
|
"VolcEngine": VolcEngineChat,
|
||||||
"BaiChuan": BaiChuanChat,
|
"BaiChuan": BaiChuanChat,
|
||||||
"MiniMax": MiniMaxChat,
|
"MiniMax": MiniMaxChat,
|
||||||
|
"Minimax": MiniMaxChat,
|
||||||
"Mistral": MistralChat,
|
"Mistral": MistralChat,
|
||||||
'Gemini' : GeminiChat,
|
'Gemini' : GeminiChat,
|
||||||
"Bedrock": BedrockChat,
|
"Bedrock": BedrockChat,
|
||||||
|
Loading…
x
Reference in New Issue
Block a user