mirror of
https://git.mirrors.martin98.com/https://github.com/infiniflow/ragflow.git
synced 2025-08-20 01:59:20 +08:00
add support for Baidu yiyan (#2049)
### What problem does this PR solve? add support for Baidu yiyan ### Type of change - [x] New Feature (non-breaking change which adds functionality) --------- Co-authored-by: Zhedong Cen <cenzhedong2@126.com>
This commit is contained in:
parent
21f2c5838b
commit
733219cc3f
@ -140,7 +140,11 @@ def add_llm():
|
|||||||
api_key = req.get("api_key","xxxxxxxxxxxxxxx")
|
api_key = req.get("api_key","xxxxxxxxxxxxxxx")
|
||||||
elif factory =="XunFei Spark":
|
elif factory =="XunFei Spark":
|
||||||
llm_name = req["llm_name"]
|
llm_name = req["llm_name"]
|
||||||
api_key = req.get("spark_api_password","")
|
api_key = req.get("spark_api_password","xxxxxxxxxxxxxxx")
|
||||||
|
elif factory == "BaiduYiyan":
|
||||||
|
llm_name = req["llm_name"]
|
||||||
|
api_key = '{' + f'"yiyan_ak": "{req.get("yiyan_ak", "")}", ' \
|
||||||
|
f'"yiyan_sk": "{req.get("yiyan_sk", "")}"' + '}'
|
||||||
else:
|
else:
|
||||||
llm_name = req["llm_name"]
|
llm_name = req["llm_name"]
|
||||||
api_key = req.get("api_key","xxxxxxxxxxxxxxx")
|
api_key = req.get("api_key","xxxxxxxxxxxxxxx")
|
||||||
@ -157,7 +161,7 @@ def add_llm():
|
|||||||
msg = ""
|
msg = ""
|
||||||
if llm["model_type"] == LLMType.EMBEDDING.value:
|
if llm["model_type"] == LLMType.EMBEDDING.value:
|
||||||
mdl = EmbeddingModel[factory](
|
mdl = EmbeddingModel[factory](
|
||||||
key=llm['api_key'] if factory in ["VolcEngine", "Bedrock","OpenAI-API-Compatible","Replicate"] else None,
|
key=llm['api_key'],
|
||||||
model_name=llm["llm_name"],
|
model_name=llm["llm_name"],
|
||||||
base_url=llm["api_base"])
|
base_url=llm["api_base"])
|
||||||
try:
|
try:
|
||||||
@ -168,7 +172,7 @@ def add_llm():
|
|||||||
msg += f"\nFail to access embedding model({llm['llm_name']})." + str(e)
|
msg += f"\nFail to access embedding model({llm['llm_name']})." + str(e)
|
||||||
elif llm["model_type"] == LLMType.CHAT.value:
|
elif llm["model_type"] == LLMType.CHAT.value:
|
||||||
mdl = ChatModel[factory](
|
mdl = ChatModel[factory](
|
||||||
key=llm['api_key'] if factory in ["VolcEngine", "Bedrock","OpenAI-API-Compatible","Replicate","XunFei Spark"] else None,
|
key=llm['api_key'],
|
||||||
model_name=llm["llm_name"],
|
model_name=llm["llm_name"],
|
||||||
base_url=llm["api_base"]
|
base_url=llm["api_base"]
|
||||||
)
|
)
|
||||||
@ -182,7 +186,9 @@ def add_llm():
|
|||||||
e)
|
e)
|
||||||
elif llm["model_type"] == LLMType.RERANK:
|
elif llm["model_type"] == LLMType.RERANK:
|
||||||
mdl = RerankModel[factory](
|
mdl = RerankModel[factory](
|
||||||
key=None, model_name=llm["llm_name"], base_url=llm["api_base"]
|
key=llm["api_key"],
|
||||||
|
model_name=llm["llm_name"],
|
||||||
|
base_url=llm["api_base"]
|
||||||
)
|
)
|
||||||
try:
|
try:
|
||||||
arr, tc = mdl.similarity("Hello~ Ragflower!", ["Hi, there!"])
|
arr, tc = mdl.similarity("Hello~ Ragflower!", ["Hi, there!"])
|
||||||
@ -193,7 +199,9 @@ def add_llm():
|
|||||||
e)
|
e)
|
||||||
elif llm["model_type"] == LLMType.IMAGE2TEXT.value:
|
elif llm["model_type"] == LLMType.IMAGE2TEXT.value:
|
||||||
mdl = CvModel[factory](
|
mdl = CvModel[factory](
|
||||||
key=llm["api_key"] if factory in ["OpenAI-API-Compatible"] else None, model_name=llm["llm_name"], base_url=llm["api_base"]
|
key=llm["api_key"],
|
||||||
|
model_name=llm["llm_name"],
|
||||||
|
base_url=llm["api_base"]
|
||||||
)
|
)
|
||||||
try:
|
try:
|
||||||
img_url = (
|
img_url = (
|
||||||
|
@ -3201,6 +3201,13 @@
|
|||||||
"tags": "LLM",
|
"tags": "LLM",
|
||||||
"status": "1",
|
"status": "1",
|
||||||
"llm": []
|
"llm": []
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"name": "BaiduYiyan",
|
||||||
|
"logo": "",
|
||||||
|
"tags": "LLM",
|
||||||
|
"status": "1",
|
||||||
|
"llm": []
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
|
@ -119,7 +119,7 @@ This section provides instructions on setting up the RAGFlow server on Linux. If
|
|||||||
```
|
```
|
||||||
|
|
||||||
:::note
|
:::note
|
||||||
If the above steps does not work, consider using [this workaround](https://github.com/docker/for-mac/issues/7047#issuecomment-1791912053), which employs a container and does not require manual editing of the macOS settings.
|
If the above steps do not work, consider using [this workaround](https://github.com/docker/for-mac/issues/7047#issuecomment-1791912053), which employs a container and does not require manual editing of the macOS settings.
|
||||||
:::
|
:::
|
||||||
|
|
||||||
</TabItem>
|
</TabItem>
|
||||||
|
@ -43,7 +43,8 @@ EmbeddingModel = {
|
|||||||
"PerfXCloud": PerfXCloudEmbed,
|
"PerfXCloud": PerfXCloudEmbed,
|
||||||
"Upstage": UpstageEmbed,
|
"Upstage": UpstageEmbed,
|
||||||
"SILICONFLOW": SILICONFLOWEmbed,
|
"SILICONFLOW": SILICONFLOWEmbed,
|
||||||
"Replicate": ReplicateEmbed
|
"Replicate": ReplicateEmbed,
|
||||||
|
"BaiduYiyan": BaiduYiyanEmbed
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@ -101,7 +102,8 @@ ChatModel = {
|
|||||||
"01.AI": YiChat,
|
"01.AI": YiChat,
|
||||||
"Replicate": ReplicateChat,
|
"Replicate": ReplicateChat,
|
||||||
"Tencent Hunyuan": HunyuanChat,
|
"Tencent Hunyuan": HunyuanChat,
|
||||||
"XunFei Spark": SparkChat
|
"XunFei Spark": SparkChat,
|
||||||
|
"BaiduYiyan": BaiduYiyanChat
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@ -115,7 +117,8 @@ RerankModel = {
|
|||||||
"OpenAI-API-Compatible": OpenAI_APIRerank,
|
"OpenAI-API-Compatible": OpenAI_APIRerank,
|
||||||
"cohere": CoHereRerank,
|
"cohere": CoHereRerank,
|
||||||
"TogetherAI": TogetherAIRerank,
|
"TogetherAI": TogetherAIRerank,
|
||||||
"SILICONFLOW": SILICONFLOWRerank
|
"SILICONFLOW": SILICONFLOWRerank,
|
||||||
|
"BaiduYiyan": BaiduYiyanRerank
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@ -1185,3 +1185,69 @@ class SparkChat(Base):
|
|||||||
}
|
}
|
||||||
model_version = model2version[model_name]
|
model_version = model2version[model_name]
|
||||||
super().__init__(key, model_version, base_url)
|
super().__init__(key, model_version, base_url)
|
||||||
|
|
||||||
|
|
||||||
|
class BaiduYiyanChat(Base):
|
||||||
|
def __init__(self, key, model_name, base_url=None):
|
||||||
|
import qianfan
|
||||||
|
|
||||||
|
key = json.loads(key)
|
||||||
|
ak = key.get("yiyan_ak","")
|
||||||
|
sk = key.get("yiyan_sk","")
|
||||||
|
self.client = qianfan.ChatCompletion(ak=ak,sk=sk)
|
||||||
|
self.model_name = model_name.lower()
|
||||||
|
self.system = ""
|
||||||
|
|
||||||
|
def chat(self, system, history, gen_conf):
|
||||||
|
if system:
|
||||||
|
self.system = system
|
||||||
|
gen_conf["penalty_score"] = (
|
||||||
|
(gen_conf.get("presence_penalty", 0) + gen_conf.get("frequency_penalty", 0)) / 2
|
||||||
|
) + 1
|
||||||
|
if "max_tokens" in gen_conf:
|
||||||
|
gen_conf["max_output_tokens"] = gen_conf["max_tokens"]
|
||||||
|
ans = ""
|
||||||
|
|
||||||
|
try:
|
||||||
|
response = self.client.do(
|
||||||
|
model=self.model_name,
|
||||||
|
messages=history,
|
||||||
|
system=self.system,
|
||||||
|
**gen_conf
|
||||||
|
).body
|
||||||
|
ans = response['result']
|
||||||
|
return ans, response["usage"]["total_tokens"]
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
return ans + "\n**ERROR**: " + str(e), 0
|
||||||
|
|
||||||
|
def chat_streamly(self, system, history, gen_conf):
|
||||||
|
if system:
|
||||||
|
self.system = system
|
||||||
|
gen_conf["penalty_score"] = (
|
||||||
|
(gen_conf.get("presence_penalty", 0) + gen_conf.get("frequency_penalty", 0)) / 2
|
||||||
|
) + 1
|
||||||
|
if "max_tokens" in gen_conf:
|
||||||
|
gen_conf["max_output_tokens"] = gen_conf["max_tokens"]
|
||||||
|
ans = ""
|
||||||
|
total_tokens = 0
|
||||||
|
|
||||||
|
try:
|
||||||
|
response = self.client.do(
|
||||||
|
model=self.model_name,
|
||||||
|
messages=history,
|
||||||
|
system=self.system,
|
||||||
|
stream=True,
|
||||||
|
**gen_conf
|
||||||
|
)
|
||||||
|
for resp in response:
|
||||||
|
resp = resp.body
|
||||||
|
ans += resp['result']
|
||||||
|
total_tokens = resp["usage"]["total_tokens"]
|
||||||
|
|
||||||
|
yield ans
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
return ans + "\n**ERROR**: " + str(e), 0
|
||||||
|
|
||||||
|
yield total_tokens
|
||||||
|
@ -32,6 +32,7 @@ import asyncio
|
|||||||
from api.utils.file_utils import get_home_cache_dir
|
from api.utils.file_utils import get_home_cache_dir
|
||||||
from rag.utils import num_tokens_from_string, truncate
|
from rag.utils import num_tokens_from_string, truncate
|
||||||
import google.generativeai as genai
|
import google.generativeai as genai
|
||||||
|
import json
|
||||||
|
|
||||||
class Base(ABC):
|
class Base(ABC):
|
||||||
def __init__(self, key, model_name):
|
def __init__(self, key, model_name):
|
||||||
@ -591,11 +592,34 @@ class ReplicateEmbed(Base):
|
|||||||
self.client = Client(api_token=key)
|
self.client = Client(api_token=key)
|
||||||
|
|
||||||
def encode(self, texts: list, batch_size=32):
|
def encode(self, texts: list, batch_size=32):
|
||||||
from json import dumps
|
res = self.client.run(self.model_name, input={"texts": json.dumps(texts)})
|
||||||
|
|
||||||
res = self.client.run(self.model_name, input={"texts": dumps(texts)})
|
|
||||||
return np.array(res), sum([num_tokens_from_string(text) for text in texts])
|
return np.array(res), sum([num_tokens_from_string(text) for text in texts])
|
||||||
|
|
||||||
def encode_queries(self, text):
|
def encode_queries(self, text):
|
||||||
res = self.client.embed(self.model_name, input={"texts": [text]})
|
res = self.client.embed(self.model_name, input={"texts": [text]})
|
||||||
return np.array(res), num_tokens_from_string(text)
|
return np.array(res), num_tokens_from_string(text)
|
||||||
|
|
||||||
|
|
||||||
|
class BaiduYiyanEmbed(Base):
|
||||||
|
def __init__(self, key, model_name, base_url=None):
|
||||||
|
import qianfan
|
||||||
|
|
||||||
|
key = json.loads(key)
|
||||||
|
ak = key.get("yiyan_ak", "")
|
||||||
|
sk = key.get("yiyan_sk", "")
|
||||||
|
self.client = qianfan.Embedding(ak=ak, sk=sk)
|
||||||
|
self.model_name = model_name
|
||||||
|
|
||||||
|
def encode(self, texts: list, batch_size=32):
|
||||||
|
res = self.client.do(model=self.model_name, texts=texts).body
|
||||||
|
return (
|
||||||
|
np.array([r["embedding"] for r in res["data"]]),
|
||||||
|
res["usage"]["total_tokens"],
|
||||||
|
)
|
||||||
|
|
||||||
|
def encode_queries(self, text):
|
||||||
|
res = self.client.do(model=self.model_name, texts=[text]).body
|
||||||
|
return (
|
||||||
|
np.array([r["embedding"] for r in res["data"]]),
|
||||||
|
res["usage"]["total_tokens"],
|
||||||
|
)
|
||||||
|
@ -24,6 +24,7 @@ from abc import ABC
|
|||||||
import numpy as np
|
import numpy as np
|
||||||
from api.utils.file_utils import get_home_cache_dir
|
from api.utils.file_utils import get_home_cache_dir
|
||||||
from rag.utils import num_tokens_from_string, truncate
|
from rag.utils import num_tokens_from_string, truncate
|
||||||
|
import json
|
||||||
|
|
||||||
def sigmoid(x):
|
def sigmoid(x):
|
||||||
return 1 / (1 + np.exp(-x))
|
return 1 / (1 + np.exp(-x))
|
||||||
@ -288,3 +289,25 @@ class SILICONFLOWRerank(Base):
|
|||||||
rank[indexs],
|
rank[indexs],
|
||||||
response["meta"]["tokens"]["input_tokens"] + response["meta"]["tokens"]["output_tokens"],
|
response["meta"]["tokens"]["input_tokens"] + response["meta"]["tokens"]["output_tokens"],
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class BaiduYiyanRerank(Base):
|
||||||
|
def __init__(self, key, model_name, base_url=None):
|
||||||
|
from qianfan.resources import Reranker
|
||||||
|
|
||||||
|
key = json.loads(key)
|
||||||
|
ak = key.get("yiyan_ak", "")
|
||||||
|
sk = key.get("yiyan_sk", "")
|
||||||
|
self.client = Reranker(ak=ak, sk=sk)
|
||||||
|
self.model_name = model_name
|
||||||
|
|
||||||
|
def similarity(self, query: str, texts: list):
|
||||||
|
res = self.client.do(
|
||||||
|
model=self.model_name,
|
||||||
|
query=query,
|
||||||
|
documents=texts,
|
||||||
|
top_n=len(texts),
|
||||||
|
).body
|
||||||
|
rank = np.array([d["relevance_score"] for d in res["results"]])
|
||||||
|
indexs = [d["index"] for d in res["results"]]
|
||||||
|
return rank[indexs], res["usage"]["total_tokens"]
|
||||||
|
@ -62,6 +62,7 @@ pytest==8.2.2
|
|||||||
python-dotenv==1.0.1
|
python-dotenv==1.0.1
|
||||||
python_dateutil==2.8.2
|
python_dateutil==2.8.2
|
||||||
python_pptx==0.6.23
|
python_pptx==0.6.23
|
||||||
|
qianfan==0.4.6
|
||||||
readability_lxml==0.8.1
|
readability_lxml==0.8.1
|
||||||
redis==5.0.3
|
redis==5.0.3
|
||||||
Requests==2.32.2
|
Requests==2.32.2
|
||||||
|
@ -100,6 +100,7 @@ python-docx==1.1.0
|
|||||||
python-dotenv==1.0.1
|
python-dotenv==1.0.1
|
||||||
python-pptx==0.6.23
|
python-pptx==0.6.23
|
||||||
PyYAML==6.0.1
|
PyYAML==6.0.1
|
||||||
|
qianfan==0.4.6
|
||||||
redis==5.0.3
|
redis==5.0.3
|
||||||
regex==2023.12.25
|
regex==2023.12.25
|
||||||
replicate==0.31.0
|
replicate==0.31.0
|
||||||
|
1
web/src/assets/svg/llm/yiyan.svg
Normal file
1
web/src/assets/svg/llm/yiyan.svg
Normal file
@ -0,0 +1 @@
|
|||||||
|
<?xml version="1.0" standalone="no"?><!DOCTYPE svg PUBLIC "-//W3C//DTD SVG 1.1//EN" "http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd"><svg t="1724295709885" class="icon" viewBox="0 0 1024 1024" version="1.1" xmlns="http://www.w3.org/2000/svg" p-id="4238" xmlns:xlink="http://www.w3.org/1999/xlink" width="200" height="200"><path d="M96.20121136 636.3124965c-0.1472897-113.41305959-0.29457937-226.8261192-0.29457937-340.23917879 0-14.87625845 7.65906378-26.51214381 20.4732666-34.02391789 45.51251353-26.65943349 91.02502705-53.31886698 136.83211997-79.53643141 71.1409192-40.94653321 142.42912809-81.59848704 213.71733698-122.39773055 7.36448439-4.12411126 14.58167909-8.3955122 21.50429441-13.2560719 19.44223878-13.40336159 39.03176725-16.05457598 60.09419263-3.53495252 27.39588193 16.34915535 54.93905355 32.25644163 82.48222516 48.16372793 88.0792333 50.96223197 176.30575629 101.77717426 264.38498958 152.59211653 9.86840908 5.74429781 19.88410785 11.19401627 29.60522725 17.0856038 14.13981003 8.54280189 21.50429441 21.06242535 21.50429443 37.70616007 0 147.73155685 0.29457937 295.46311371-0.1472897 443.19467057 0 15.46541722-7.2171947 28.57419943-21.7988738 36.96971163-34.7603663 20.17868721-70.55176044 38.88447758-104.57567833 59.94690293-48.90017634 30.19438599-100.00969801 56.11737105-148.76258466 86.60633642-29.01606849 18.11663161-59.50503387 34.02391789-89.11026112 50.96223197-13.10878221 7.51177407-26.07027474 15.17083783-39.03176726 22.9771913-13.84523065 8.3955122-27.83775099 8.83738127-41.97756102 0.73644843-56.41195043-32.55102101-112.82390085-65.10204201-169.38314098-97.653063-61.86166887-35.64410444-123.72333775-71.1409192-185.4377169-106.78502365-11.19401627-6.48074626-22.24074286-12.81420285-32.99289009-19.88410785-11.48859565-7.65906378-17.08560379-19.14765941-17.08560378-32.69831069-0.1472897-34.7603663 0.1472897-69.52073264 0.29457938-104.28109895 1.62018657-0.58915875 1.62018657-1.62018657-0.29457938-2.65121438z m356.58833414-225.500512c2.20934532-1.76747625 4.41869063-3.68224221 6.77532565-5.15513907 68.93157389-39.62092601 137.86314777-79.24185204 206.94201135-118.86277807 2.79850407-1.62018657 6.48074626-1.62018657 6.62803594-6.18616688 0.1472897-4.8605597-4.12411126-4.71327001-6.77532564-6.18616688-40.65195383-23.56635005-81.59848704-46.83812071-122.10315117-70.84633984-16.79102442-10.01569877-32.84560039-8.54280189-48.45830728 0.58915876-45.9543826 26.51214381-91.46689612 53.61344636-137.27398903 80.42016953-31.96186226 18.70579035-64.21830387 37.11700133-96.32745581 55.67550198-18.41121097 10.60485751-27.54317163 25.33382629-27.24859225 47.72185885 0.88373813 89.55213018 0.58915875 179.10426036 0.14728969 268.65639053-0.1472897 20.17868721 9.27925033 33.58204881 25.33382629 43.15587853 31.3727035 18.70579035 63.18727606 37.11700133 95.14913832 54.93905355 10.89943689 6.03887719 21.06242535 13.99252034 35.79139414 18.41121096V505.51925374c6.48074626 19.58952848 18.55850066 34.02391789 36.67513226 44.6287754 27.83775099 16.20186565 63.18727606 12.51962347 86.31175705-10.45756784 26.95401286-26.65943349 28.72148912-62.89269668 12.81420282-90.14128893-16.34915535-28.42690974-43.59774757-37.55887038-74.38129233-38.73718787z m82.48222517 429.64401928c14.28709972-3.82953187 25.92298506-13.99252034 38.88447758-21.35700473 40.94653321-23.27177067 81.30390766-47.72185885 122.54502023-70.55176046 26.95401286-15.02354815 52.87699792-31.66728287 80.71474891-45.21793415 16.79102442-8.10093283 29.60522723-22.53532223 29.60522726-43.4504579 0.1472897-92.939793 0.29457937-185.73229631 0.14728969-278.6720893 0-11.19401627-5.15513907-13.99252034-13.84523067-7.06990501-26.51214381 20.76784598-57.29568854 34.46578693-86.16446735 51.25681135-54.49718448 31.81457257-109.14165865 63.33456576-163.78613282 95.00184862-8.54280189 4.8605597-11.78317502 10.45756784-11.63588535 20.47326662 0.29457937 96.18016613 0.1472897 192.50762194 0.1472897 288.68778806-0.29457937 3.5349525-1.47289687 7.65906378 3.38766282 10.8994369z" fill="#066AF3" p-id="4239"></path><path d="M96.20121136 636.3124965c1.91476594 1.03102783 1.91476594 2.06205563 0 3.09308345v-3.09308345z" fill="#4372E0" p-id="4240"></path><path d="M391.3697457 505.37196405c-5.44971845-44.33419602 13.84523065-74.08671296 61.4197998-94.55997955 30.93083443 1.17831749 58.03213699 10.31027814 74.38129233 38.5898982 15.75999659 27.39588193 14.13981003 63.48185543-12.81420282 90.14128893-23.27177067 22.97719129-58.47400606 26.65943349-86.31175705 10.45756783-18.11663161-10.60485751-30.34167568-25.03924691-36.67513226-44.62877541z" fill="#002A9A" p-id="4241"></path></svg>
|
After Width: | Height: | Size: 4.5 KiB |
@ -528,6 +528,11 @@ The above is the content you need to summarize.`,
|
|||||||
SparkModelNameMessage: 'Please select Spark model',
|
SparkModelNameMessage: 'Please select Spark model',
|
||||||
addSparkAPIPassword: 'Spark APIPassword',
|
addSparkAPIPassword: 'Spark APIPassword',
|
||||||
SparkAPIPasswordMessage: 'please input your APIPassword',
|
SparkAPIPasswordMessage: 'please input your APIPassword',
|
||||||
|
yiyanModelNameMessage: 'Please input model name',
|
||||||
|
addyiyanAK: 'yiyan API KEY',
|
||||||
|
yiyanAKMessage: 'Please input your API KEY',
|
||||||
|
addyiyanSK: 'yiyan Secret KEY',
|
||||||
|
yiyanSKMessage: 'Please input your Secret KEY',
|
||||||
},
|
},
|
||||||
message: {
|
message: {
|
||||||
registered: 'Registered!',
|
registered: 'Registered!',
|
||||||
|
@ -491,6 +491,11 @@ export default {
|
|||||||
SparkModelNameMessage: '請選擇星火模型!',
|
SparkModelNameMessage: '請選擇星火模型!',
|
||||||
addSparkAPIPassword: '星火 APIPassword',
|
addSparkAPIPassword: '星火 APIPassword',
|
||||||
SparkAPIPasswordMessage: '請輸入 APIPassword',
|
SparkAPIPasswordMessage: '請輸入 APIPassword',
|
||||||
|
yiyanModelNameMessage: '輸入模型名稱',
|
||||||
|
addyiyanAK: '一言 API KEY',
|
||||||
|
yiyanAKMessage: '請輸入 API KEY',
|
||||||
|
addyiyanSK: '一言 Secret KEY',
|
||||||
|
yiyanSKMessage: '請輸入 Secret KEY',
|
||||||
},
|
},
|
||||||
message: {
|
message: {
|
||||||
registered: '註冊成功',
|
registered: '註冊成功',
|
||||||
|
@ -508,6 +508,11 @@ export default {
|
|||||||
SparkModelNameMessage: '请选择星火模型!',
|
SparkModelNameMessage: '请选择星火模型!',
|
||||||
addSparkAPIPassword: '星火 APIPassword',
|
addSparkAPIPassword: '星火 APIPassword',
|
||||||
SparkAPIPasswordMessage: '请输入 APIPassword',
|
SparkAPIPasswordMessage: '请输入 APIPassword',
|
||||||
|
yiyanModelNameMessage: '请输入模型名称',
|
||||||
|
addyiyanAK: '一言 API KEY',
|
||||||
|
yiyanAKMessage: '请输入 API KEY',
|
||||||
|
addyiyanSK: '一言 Secret KEY',
|
||||||
|
yiyanSKMessage: '请输入 Secret KEY',
|
||||||
},
|
},
|
||||||
message: {
|
message: {
|
||||||
registered: '注册成功',
|
registered: '注册成功',
|
||||||
|
@ -34,6 +34,7 @@ export const IconMap = {
|
|||||||
Replicate: 'replicate',
|
Replicate: 'replicate',
|
||||||
'Tencent Hunyuan': 'hunyuan',
|
'Tencent Hunyuan': 'hunyuan',
|
||||||
'XunFei Spark': 'spark',
|
'XunFei Spark': 'spark',
|
||||||
|
BaiduYiyan: 'yiyan',
|
||||||
};
|
};
|
||||||
|
|
||||||
export const BedrockRegionList = [
|
export const BedrockRegionList = [
|
||||||
|
@ -217,6 +217,33 @@ export const useSubmitSpark = () => {
|
|||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
|
export const useSubmityiyan = () => {
|
||||||
|
const { addLlm, loading } = useAddLlm();
|
||||||
|
const {
|
||||||
|
visible: yiyanAddingVisible,
|
||||||
|
hideModal: hideyiyanAddingModal,
|
||||||
|
showModal: showyiyanAddingModal,
|
||||||
|
} = useSetModalState();
|
||||||
|
|
||||||
|
const onyiyanAddingOk = useCallback(
|
||||||
|
async (payload: IAddLlmRequestBody) => {
|
||||||
|
const ret = await addLlm(payload);
|
||||||
|
if (ret === 0) {
|
||||||
|
hideyiyanAddingModal();
|
||||||
|
}
|
||||||
|
},
|
||||||
|
[hideyiyanAddingModal, addLlm],
|
||||||
|
);
|
||||||
|
|
||||||
|
return {
|
||||||
|
yiyanAddingLoading: loading,
|
||||||
|
onyiyanAddingOk,
|
||||||
|
yiyanAddingVisible,
|
||||||
|
hideyiyanAddingModal,
|
||||||
|
showyiyanAddingModal,
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|
||||||
export const useSubmitBedrock = () => {
|
export const useSubmitBedrock = () => {
|
||||||
const { addLlm, loading } = useAddLlm();
|
const { addLlm, loading } = useAddLlm();
|
||||||
const {
|
const {
|
||||||
|
@ -39,6 +39,7 @@ import {
|
|||||||
useSubmitSpark,
|
useSubmitSpark,
|
||||||
useSubmitSystemModelSetting,
|
useSubmitSystemModelSetting,
|
||||||
useSubmitVolcEngine,
|
useSubmitVolcEngine,
|
||||||
|
useSubmityiyan,
|
||||||
} from './hooks';
|
} from './hooks';
|
||||||
import HunyuanModal from './hunyuan-modal';
|
import HunyuanModal from './hunyuan-modal';
|
||||||
import styles from './index.less';
|
import styles from './index.less';
|
||||||
@ -46,6 +47,7 @@ import OllamaModal from './ollama-modal';
|
|||||||
import SparkModal from './spark-modal';
|
import SparkModal from './spark-modal';
|
||||||
import SystemModelSettingModal from './system-model-setting-modal';
|
import SystemModelSettingModal from './system-model-setting-modal';
|
||||||
import VolcEngineModal from './volcengine-modal';
|
import VolcEngineModal from './volcengine-modal';
|
||||||
|
import YiyanModal from './yiyan-modal';
|
||||||
|
|
||||||
const LlmIcon = ({ name }: { name: string }) => {
|
const LlmIcon = ({ name }: { name: string }) => {
|
||||||
const icon = IconMap[name as keyof typeof IconMap];
|
const icon = IconMap[name as keyof typeof IconMap];
|
||||||
@ -95,7 +97,8 @@ const ModelCard = ({ item, clickApiKey }: IModelCardProps) => {
|
|||||||
{isLocalLlmFactory(item.name) ||
|
{isLocalLlmFactory(item.name) ||
|
||||||
item.name === 'VolcEngine' ||
|
item.name === 'VolcEngine' ||
|
||||||
item.name === 'Tencent Hunyuan' ||
|
item.name === 'Tencent Hunyuan' ||
|
||||||
item.name === 'XunFei Spark'
|
item.name === 'XunFei Spark' ||
|
||||||
|
item.name === 'BaiduYiyan'
|
||||||
? t('addTheModel')
|
? t('addTheModel')
|
||||||
: 'API-Key'}
|
: 'API-Key'}
|
||||||
<SettingOutlined />
|
<SettingOutlined />
|
||||||
@ -185,6 +188,14 @@ const UserSettingModel = () => {
|
|||||||
SparkAddingLoading,
|
SparkAddingLoading,
|
||||||
} = useSubmitSpark();
|
} = useSubmitSpark();
|
||||||
|
|
||||||
|
const {
|
||||||
|
yiyanAddingVisible,
|
||||||
|
hideyiyanAddingModal,
|
||||||
|
showyiyanAddingModal,
|
||||||
|
onyiyanAddingOk,
|
||||||
|
yiyanAddingLoading,
|
||||||
|
} = useSubmityiyan();
|
||||||
|
|
||||||
const {
|
const {
|
||||||
bedrockAddingLoading,
|
bedrockAddingLoading,
|
||||||
onBedrockAddingOk,
|
onBedrockAddingOk,
|
||||||
@ -199,12 +210,14 @@ const UserSettingModel = () => {
|
|||||||
VolcEngine: showVolcAddingModal,
|
VolcEngine: showVolcAddingModal,
|
||||||
'Tencent Hunyuan': showHunyuanAddingModal,
|
'Tencent Hunyuan': showHunyuanAddingModal,
|
||||||
'XunFei Spark': showSparkAddingModal,
|
'XunFei Spark': showSparkAddingModal,
|
||||||
|
BaiduYiyan: showyiyanAddingModal,
|
||||||
}),
|
}),
|
||||||
[
|
[
|
||||||
showBedrockAddingModal,
|
showBedrockAddingModal,
|
||||||
showVolcAddingModal,
|
showVolcAddingModal,
|
||||||
showHunyuanAddingModal,
|
showHunyuanAddingModal,
|
||||||
showSparkAddingModal,
|
showSparkAddingModal,
|
||||||
|
showyiyanAddingModal,
|
||||||
],
|
],
|
||||||
);
|
);
|
||||||
|
|
||||||
@ -330,6 +343,13 @@ const UserSettingModel = () => {
|
|||||||
loading={SparkAddingLoading}
|
loading={SparkAddingLoading}
|
||||||
llmFactory={'XunFei Spark'}
|
llmFactory={'XunFei Spark'}
|
||||||
></SparkModal>
|
></SparkModal>
|
||||||
|
<YiyanModal
|
||||||
|
visible={yiyanAddingVisible}
|
||||||
|
hideModal={hideyiyanAddingModal}
|
||||||
|
onOk={onyiyanAddingOk}
|
||||||
|
loading={yiyanAddingLoading}
|
||||||
|
llmFactory={'BaiduYiyan'}
|
||||||
|
></YiyanModal>
|
||||||
<BedrockModal
|
<BedrockModal
|
||||||
visible={bedrockAddingVisible}
|
visible={bedrockAddingVisible}
|
||||||
hideModal={hideBedrockAddingModal}
|
hideModal={hideBedrockAddingModal}
|
||||||
|
@ -0,0 +1,97 @@
|
|||||||
|
import { useTranslate } from '@/hooks/common-hooks';
|
||||||
|
import { IModalProps } from '@/interfaces/common';
|
||||||
|
import { IAddLlmRequestBody } from '@/interfaces/request/llm';
|
||||||
|
import { Form, Input, Modal, Select } from 'antd';
|
||||||
|
import omit from 'lodash/omit';
|
||||||
|
|
||||||
|
type FieldType = IAddLlmRequestBody & {
|
||||||
|
vision: boolean;
|
||||||
|
yiyan_ak: string;
|
||||||
|
yiyan_sk: string;
|
||||||
|
};
|
||||||
|
|
||||||
|
const { Option } = Select;
|
||||||
|
|
||||||
|
const YiyanModal = ({
|
||||||
|
visible,
|
||||||
|
hideModal,
|
||||||
|
onOk,
|
||||||
|
loading,
|
||||||
|
llmFactory,
|
||||||
|
}: IModalProps<IAddLlmRequestBody> & { llmFactory: string }) => {
|
||||||
|
const [form] = Form.useForm<FieldType>();
|
||||||
|
|
||||||
|
const { t } = useTranslate('setting');
|
||||||
|
|
||||||
|
const handleOk = async () => {
|
||||||
|
const values = await form.validateFields();
|
||||||
|
const modelType =
|
||||||
|
values.model_type === 'chat' && values.vision
|
||||||
|
? 'image2text'
|
||||||
|
: values.model_type;
|
||||||
|
|
||||||
|
const data = {
|
||||||
|
...omit(values, ['vision']),
|
||||||
|
model_type: modelType,
|
||||||
|
llm_factory: llmFactory,
|
||||||
|
};
|
||||||
|
console.info(data);
|
||||||
|
|
||||||
|
onOk?.(data);
|
||||||
|
};
|
||||||
|
|
||||||
|
return (
|
||||||
|
<Modal
|
||||||
|
title={t('addLlmTitle', { name: llmFactory })}
|
||||||
|
open={visible}
|
||||||
|
onOk={handleOk}
|
||||||
|
onCancel={hideModal}
|
||||||
|
okButtonProps={{ loading }}
|
||||||
|
confirmLoading={loading}
|
||||||
|
>
|
||||||
|
<Form
|
||||||
|
name="basic"
|
||||||
|
style={{ maxWidth: 600 }}
|
||||||
|
autoComplete="off"
|
||||||
|
layout={'vertical'}
|
||||||
|
form={form}
|
||||||
|
>
|
||||||
|
<Form.Item<FieldType>
|
||||||
|
label={t('modelType')}
|
||||||
|
name="model_type"
|
||||||
|
initialValue={'chat'}
|
||||||
|
rules={[{ required: true, message: t('modelTypeMessage') }]}
|
||||||
|
>
|
||||||
|
<Select placeholder={t('modelTypeMessage')}>
|
||||||
|
<Option value="chat">chat</Option>
|
||||||
|
<Option value="embedding">embedding</Option>
|
||||||
|
<Option value="rerank">rerank</Option>
|
||||||
|
</Select>
|
||||||
|
</Form.Item>
|
||||||
|
<Form.Item<FieldType>
|
||||||
|
label={t('modelName')}
|
||||||
|
name="llm_name"
|
||||||
|
rules={[{ required: true, message: t('yiyanModelNameMessage') }]}
|
||||||
|
>
|
||||||
|
<Input placeholder={t('yiyanModelNameMessage')} />
|
||||||
|
</Form.Item>
|
||||||
|
<Form.Item<FieldType>
|
||||||
|
label={t('addyiyanAK')}
|
||||||
|
name="yiyan_ak"
|
||||||
|
rules={[{ required: true, message: t('yiyanAKMessage') }]}
|
||||||
|
>
|
||||||
|
<Input placeholder={t('yiyanAKMessage')} />
|
||||||
|
</Form.Item>
|
||||||
|
<Form.Item<FieldType>
|
||||||
|
label={t('addyiyanSK')}
|
||||||
|
name="yiyan_sk"
|
||||||
|
rules={[{ required: true, message: t('yiyanSKMessage') }]}
|
||||||
|
>
|
||||||
|
<Input placeholder={t('yiyanSKMessage')} />
|
||||||
|
</Form.Item>
|
||||||
|
</Form>
|
||||||
|
</Modal>
|
||||||
|
);
|
||||||
|
};
|
||||||
|
|
||||||
|
export default YiyanModal;
|
Loading…
x
Reference in New Issue
Block a user