mirror of
https://git.mirrors.martin98.com/https://github.com/infiniflow/ragflow.git
synced 2025-07-25 13:14:27 +08:00
add support for Google Cloud (#2175)
### What problem does this PR solve? #1853 add support for Google Cloud ### Type of change - [x] New Feature (non-breaking change which adds functionality) --------- Co-authored-by: Zhedong Cen <cenzhedong2@126.com>
This commit is contained in:
parent
def18308d0
commit
5decdde182
@ -150,6 +150,14 @@ def add_llm():
|
|||||||
llm_name = req["llm_name"]
|
llm_name = req["llm_name"]
|
||||||
api_key = '{' + f'"fish_audio_ak": "{req.get("fish_audio_ak", "")}", ' \
|
api_key = '{' + f'"fish_audio_ak": "{req.get("fish_audio_ak", "")}", ' \
|
||||||
f'"fish_audio_refid": "{req.get("fish_audio_refid", "59cb5986671546eaa6ca8ae6f29f6d22")}"' + '}'
|
f'"fish_audio_refid": "{req.get("fish_audio_refid", "59cb5986671546eaa6ca8ae6f29f6d22")}"' + '}'
|
||||||
|
elif factory == "Google Cloud":
|
||||||
|
llm_name = req["llm_name"]
|
||||||
|
api_key = (
|
||||||
|
"{" + f'"google_project_id": "{req.get("google_project_id", "")}", '
|
||||||
|
f'"google_region": "{req.get("google_region", "")}", '
|
||||||
|
f'"google_service_account_key": "{req.get("google_service_account_key", "")}"'
|
||||||
|
+ "}"
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
llm_name = req["llm_name"]
|
llm_name = req["llm_name"]
|
||||||
api_key = req.get("api_key","xxxxxxxxxxxxxxx")
|
api_key = req.get("api_key","xxxxxxxxxxxxxxx")
|
||||||
|
@ -3352,6 +3352,13 @@
|
|||||||
"model_type": "rerank"
|
"model_type": "rerank"
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"name": "Google Cloud",
|
||||||
|
"logo": "",
|
||||||
|
"tags": "LLM",
|
||||||
|
"status": "1",
|
||||||
|
"llm": []
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
|
@ -107,6 +107,7 @@ ChatModel = {
|
|||||||
"XunFei Spark": SparkChat,
|
"XunFei Spark": SparkChat,
|
||||||
"BaiduYiyan": BaiduYiyanChat,
|
"BaiduYiyan": BaiduYiyanChat,
|
||||||
"Anthropic": AnthropicChat,
|
"Anthropic": AnthropicChat,
|
||||||
|
"Google Cloud": GoogleChat,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@ -701,9 +701,13 @@ class GeminiChat(Base):
|
|||||||
self.model = GenerativeModel(model_name=self.model_name)
|
self.model = GenerativeModel(model_name=self.model_name)
|
||||||
self.model._client = _client
|
self.model._client = _client
|
||||||
|
|
||||||
|
|
||||||
def chat(self,system,history,gen_conf):
|
def chat(self,system,history,gen_conf):
|
||||||
|
from google.generativeai.types import content_types
|
||||||
|
|
||||||
if system:
|
if system:
|
||||||
history.insert(0, {"role": "user", "parts": system})
|
self.model._system_instruction = content_types.to_content(system)
|
||||||
|
|
||||||
if 'max_tokens' in gen_conf:
|
if 'max_tokens' in gen_conf:
|
||||||
gen_conf['max_output_tokens'] = gen_conf['max_tokens']
|
gen_conf['max_output_tokens'] = gen_conf['max_tokens']
|
||||||
for k in list(gen_conf.keys()):
|
for k in list(gen_conf.keys()):
|
||||||
@ -725,8 +729,10 @@ class GeminiChat(Base):
|
|||||||
return "**ERROR**: " + str(e), 0
|
return "**ERROR**: " + str(e), 0
|
||||||
|
|
||||||
def chat_streamly(self, system, history, gen_conf):
|
def chat_streamly(self, system, history, gen_conf):
|
||||||
|
from google.generativeai.types import content_types
|
||||||
|
|
||||||
if system:
|
if system:
|
||||||
history.insert(0, {"role": "user", "parts": system})
|
self.model._system_instruction = content_types.to_content(system)
|
||||||
if 'max_tokens' in gen_conf:
|
if 'max_tokens' in gen_conf:
|
||||||
gen_conf['max_output_tokens'] = gen_conf['max_tokens']
|
gen_conf['max_output_tokens'] = gen_conf['max_tokens']
|
||||||
for k in list(gen_conf.keys()):
|
for k in list(gen_conf.keys()):
|
||||||
@ -1257,3 +1263,154 @@ class AnthropicChat(Base):
|
|||||||
yield ans + "\n**ERROR**: " + str(e)
|
yield ans + "\n**ERROR**: " + str(e)
|
||||||
|
|
||||||
yield total_tokens
|
yield total_tokens
|
||||||
|
|
||||||
|
|
||||||
|
class GoogleChat(Base):
|
||||||
|
def __init__(self, key, model_name, base_url=None):
|
||||||
|
from google.oauth2 import service_account
|
||||||
|
import base64
|
||||||
|
|
||||||
|
key = json.load(key)
|
||||||
|
access_token = json.loads(
|
||||||
|
base64.b64decode(key.get("google_service_account_key", ""))
|
||||||
|
)
|
||||||
|
project_id = key.get("google_project_id", "")
|
||||||
|
region = key.get("google_region", "")
|
||||||
|
|
||||||
|
scopes = ["https://www.googleapis.com/auth/cloud-platform"]
|
||||||
|
self.model_name = model_name
|
||||||
|
self.system = ""
|
||||||
|
|
||||||
|
if "claude" in self.model_name:
|
||||||
|
from anthropic import AnthropicVertex
|
||||||
|
from google.auth.transport.requests import Request
|
||||||
|
|
||||||
|
if access_token:
|
||||||
|
credits = service_account.Credentials.from_service_account_info(
|
||||||
|
access_token, scopes=scopes
|
||||||
|
)
|
||||||
|
request = Request()
|
||||||
|
credits.refresh(request)
|
||||||
|
token = credits.token
|
||||||
|
self.client = AnthropicVertex(
|
||||||
|
region=region, project_id=project_id, access_token=token
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
self.client = AnthropicVertex(region=region, project_id=project_id)
|
||||||
|
else:
|
||||||
|
from google.cloud import aiplatform
|
||||||
|
import vertexai.generative_models as glm
|
||||||
|
|
||||||
|
if access_token:
|
||||||
|
credits = service_account.Credentials.from_service_account_info(
|
||||||
|
access_token
|
||||||
|
)
|
||||||
|
aiplatform.init(
|
||||||
|
credentials=credits, project=project_id, location=region
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
aiplatform.init(project=project_id, location=region)
|
||||||
|
self.client = glm.GenerativeModel(model_name=self.model_name)
|
||||||
|
|
||||||
|
def chat(self, system, history, gen_conf):
|
||||||
|
if system:
|
||||||
|
self.system = system
|
||||||
|
|
||||||
|
if "claude" in self.model_name:
|
||||||
|
if "max_tokens" not in gen_conf:
|
||||||
|
gen_conf["max_tokens"] = 4096
|
||||||
|
try:
|
||||||
|
response = self.client.messages.create(
|
||||||
|
model=self.model_name,
|
||||||
|
messages=history,
|
||||||
|
system=self.system,
|
||||||
|
stream=False,
|
||||||
|
**gen_conf,
|
||||||
|
).json()
|
||||||
|
ans = response["content"][0]["text"]
|
||||||
|
if response["stop_reason"] == "max_tokens":
|
||||||
|
ans += (
|
||||||
|
"...\nFor the content length reason, it stopped, continue?"
|
||||||
|
if is_english([ans])
|
||||||
|
else "······\n由于长度的原因,回答被截断了,要继续吗?"
|
||||||
|
)
|
||||||
|
return (
|
||||||
|
ans,
|
||||||
|
response["usage"]["input_tokens"]
|
||||||
|
+ response["usage"]["output_tokens"],
|
||||||
|
)
|
||||||
|
except Exception as e:
|
||||||
|
return ans + "\n**ERROR**: " + str(e), 0
|
||||||
|
else:
|
||||||
|
self.client._system_instruction = self.system
|
||||||
|
if "max_tokens" in gen_conf:
|
||||||
|
gen_conf["max_output_tokens"] = gen_conf["max_tokens"]
|
||||||
|
for k in list(gen_conf.keys()):
|
||||||
|
if k not in ["temperature", "top_p", "max_output_tokens"]:
|
||||||
|
del gen_conf[k]
|
||||||
|
for item in history:
|
||||||
|
if "role" in item and item["role"] == "assistant":
|
||||||
|
item["role"] = "model"
|
||||||
|
if "content" in item:
|
||||||
|
item["parts"] = item.pop("content")
|
||||||
|
try:
|
||||||
|
response = self.client.generate_content(
|
||||||
|
history, generation_config=gen_conf
|
||||||
|
)
|
||||||
|
ans = response.text
|
||||||
|
return ans, response.usage_metadata.total_token_count
|
||||||
|
except Exception as e:
|
||||||
|
return "**ERROR**: " + str(e), 0
|
||||||
|
|
||||||
|
def chat_streamly(self, system, history, gen_conf):
|
||||||
|
if system:
|
||||||
|
self.system = system
|
||||||
|
|
||||||
|
if "claude" in self.model_name:
|
||||||
|
if "max_tokens" not in gen_conf:
|
||||||
|
gen_conf["max_tokens"] = 4096
|
||||||
|
ans = ""
|
||||||
|
total_tokens = 0
|
||||||
|
try:
|
||||||
|
response = self.client.messages.create(
|
||||||
|
model=self.model_name,
|
||||||
|
messages=history,
|
||||||
|
system=self.system,
|
||||||
|
stream=True,
|
||||||
|
**gen_conf,
|
||||||
|
)
|
||||||
|
for res in response.iter_lines():
|
||||||
|
res = res.decode("utf-8")
|
||||||
|
if "content_block_delta" in res and "data" in res:
|
||||||
|
text = json.loads(res[6:])["delta"]["text"]
|
||||||
|
ans += text
|
||||||
|
total_tokens += num_tokens_from_string(text)
|
||||||
|
except Exception as e:
|
||||||
|
yield ans + "\n**ERROR**: " + str(e)
|
||||||
|
|
||||||
|
yield total_tokens
|
||||||
|
else:
|
||||||
|
self.client._system_instruction = self.system
|
||||||
|
if "max_tokens" in gen_conf:
|
||||||
|
gen_conf["max_output_tokens"] = gen_conf["max_tokens"]
|
||||||
|
for k in list(gen_conf.keys()):
|
||||||
|
if k not in ["temperature", "top_p", "max_output_tokens"]:
|
||||||
|
del gen_conf[k]
|
||||||
|
for item in history:
|
||||||
|
if "role" in item and item["role"] == "assistant":
|
||||||
|
item["role"] = "model"
|
||||||
|
if "content" in item:
|
||||||
|
item["parts"] = item.pop("content")
|
||||||
|
ans = ""
|
||||||
|
try:
|
||||||
|
response = self.model.generate_content(
|
||||||
|
history, generation_config=gen_conf, stream=True
|
||||||
|
)
|
||||||
|
for resp in response:
|
||||||
|
ans += resp.text
|
||||||
|
yield ans
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
yield ans + "\n**ERROR**: " + str(e)
|
||||||
|
|
||||||
|
yield response._chunks[-1].usage_metadata.total_token_count
|
||||||
|
@ -85,6 +85,7 @@ tiktoken==0.6.0
|
|||||||
torch==2.3.0
|
torch==2.3.0
|
||||||
transformers==4.38.1
|
transformers==4.38.1
|
||||||
umap==0.1.1
|
umap==0.1.1
|
||||||
|
vertexai==1.64.0
|
||||||
volcengine==1.0.146
|
volcengine==1.0.146
|
||||||
voyageai==0.2.3
|
voyageai==0.2.3
|
||||||
webdriver_manager==4.0.1
|
webdriver_manager==4.0.1
|
||||||
|
@ -167,3 +167,4 @@ scholarly==1.7.11
|
|||||||
deepl==1.18.0
|
deepl==1.18.0
|
||||||
psycopg2-binary==2.9.9
|
psycopg2-binary==2.9.9
|
||||||
tabulate==0.9.0
|
tabulate==0.9.0
|
||||||
|
vertexai==1.64.0
|
1
web/src/assets/svg/llm/google-cloud.svg
Normal file
1
web/src/assets/svg/llm/google-cloud.svg
Normal file
File diff suppressed because one or more lines are too long
After Width: | Height: | Size: 11 KiB |
@ -499,6 +499,7 @@ The above is the content you need to summarize.`,
|
|||||||
upgrade: 'Upgrade',
|
upgrade: 'Upgrade',
|
||||||
addLlmTitle: 'Add LLM',
|
addLlmTitle: 'Add LLM',
|
||||||
modelName: 'Model name',
|
modelName: 'Model name',
|
||||||
|
modelID: 'Model ID',
|
||||||
modelUid: 'Model UID',
|
modelUid: 'Model UID',
|
||||||
modelNameMessage: 'Please input your model name!',
|
modelNameMessage: 'Please input your model name!',
|
||||||
modelType: 'Model type',
|
modelType: 'Model type',
|
||||||
@ -551,6 +552,15 @@ The above is the content you need to summarize.`,
|
|||||||
addFishAudioRefID: 'FishAudio Refrence ID',
|
addFishAudioRefID: 'FishAudio Refrence ID',
|
||||||
addFishAudioRefIDMessage:
|
addFishAudioRefIDMessage:
|
||||||
'Please input the Reference ID (leave blank to use the default model).',
|
'Please input the Reference ID (leave blank to use the default model).',
|
||||||
|
GoogleModelIDMessage: 'Please input your model ID!',
|
||||||
|
addGoogleProjectID: 'Project ID',
|
||||||
|
GoogleProjectIDMessage: 'Please input your Project ID',
|
||||||
|
addGoogleServiceAccountKey:
|
||||||
|
'Service Account Key(Leave blank if you use Application Default Credentials)',
|
||||||
|
GoogleServiceAccountKeyMessage:
|
||||||
|
'Please input Google Cloud Service Account Key in base64 format',
|
||||||
|
addGoogleRegion: 'Google Cloud Region',
|
||||||
|
GoogleRegionMessage: 'Please input Google Cloud Region',
|
||||||
},
|
},
|
||||||
message: {
|
message: {
|
||||||
registered: 'Registered!',
|
registered: 'Registered!',
|
||||||
|
@ -461,6 +461,7 @@ export default {
|
|||||||
upgrade: '升級',
|
upgrade: '升級',
|
||||||
addLlmTitle: '添加Llm',
|
addLlmTitle: '添加Llm',
|
||||||
modelName: '模型名稱',
|
modelName: '模型名稱',
|
||||||
|
modelID: '模型ID',
|
||||||
modelUid: '模型uid',
|
modelUid: '模型uid',
|
||||||
modelType: '模型類型',
|
modelType: '模型類型',
|
||||||
addLlmBaseUrl: '基礎 Url',
|
addLlmBaseUrl: '基礎 Url',
|
||||||
@ -511,6 +512,15 @@ export default {
|
|||||||
addFishAudioAKMessage: '請輸入 API KEY',
|
addFishAudioAKMessage: '請輸入 API KEY',
|
||||||
addFishAudioRefID: 'FishAudio Refrence ID',
|
addFishAudioRefID: 'FishAudio Refrence ID',
|
||||||
addFishAudioRefIDMessage: '請輸入引用模型的ID(留空表示使用默認模型)',
|
addFishAudioRefIDMessage: '請輸入引用模型的ID(留空表示使用默認模型)',
|
||||||
|
GoogleModelIDMessage: '請輸入 model ID!',
|
||||||
|
addGoogleProjectID: 'Project ID',
|
||||||
|
GoogleProjectIDMessage: '請輸入 Project ID',
|
||||||
|
addGoogleServiceAccountKey:
|
||||||
|
'Service Account Key(Leave blank if you use Application Default Credentials)',
|
||||||
|
GoogleServiceAccountKeyMessage:
|
||||||
|
'請輸入 Google Cloud Service Account Key in base64 format',
|
||||||
|
addGoogleRegion: 'Google Cloud 區域',
|
||||||
|
GoogleRegionMessage: '請輸入 Google Cloud 區域',
|
||||||
},
|
},
|
||||||
message: {
|
message: {
|
||||||
registered: '註冊成功',
|
registered: '註冊成功',
|
||||||
|
@ -478,6 +478,7 @@ export default {
|
|||||||
upgrade: '升级',
|
upgrade: '升级',
|
||||||
addLlmTitle: '添加 LLM',
|
addLlmTitle: '添加 LLM',
|
||||||
modelName: '模型名称',
|
modelName: '模型名称',
|
||||||
|
modelID: '模型ID',
|
||||||
modelUid: '模型UID',
|
modelUid: '模型UID',
|
||||||
modelType: '模型类型',
|
modelType: '模型类型',
|
||||||
addLlmBaseUrl: '基础 Url',
|
addLlmBaseUrl: '基础 Url',
|
||||||
@ -528,6 +529,15 @@ export default {
|
|||||||
FishAudioAKMessage: '请输入 API KEY',
|
FishAudioAKMessage: '请输入 API KEY',
|
||||||
addFishAudioRefID: 'FishAudio Refrence ID',
|
addFishAudioRefID: 'FishAudio Refrence ID',
|
||||||
FishAudioRefIDMessage: '请输入引用模型的ID(留空表示使用默认模型)',
|
FishAudioRefIDMessage: '请输入引用模型的ID(留空表示使用默认模型)',
|
||||||
|
GoogleModelIDMessage: '请输入 model ID!',
|
||||||
|
addGoogleProjectID: 'Project ID',
|
||||||
|
GoogleProjectIDMessage: '请输入 Project ID',
|
||||||
|
addGoogleServiceAccountKey:
|
||||||
|
'Service Account Key(Leave blank if you use Application Default Credentials)',
|
||||||
|
GoogleServiceAccountKeyMessage:
|
||||||
|
'请输入 Google Cloud Service Account Key in base64 format',
|
||||||
|
addGoogleRegion: 'Google Cloud 区域',
|
||||||
|
GoogleRegionMessage: '请输入 Google Cloud 区域',
|
||||||
},
|
},
|
||||||
message: {
|
message: {
|
||||||
registered: '注册成功',
|
registered: '注册成功',
|
||||||
|
@ -39,6 +39,7 @@ export const IconMap = {
|
|||||||
'Tencent Cloud': 'tencent-cloud',
|
'Tencent Cloud': 'tencent-cloud',
|
||||||
Anthropic: 'anthropic',
|
Anthropic: 'anthropic',
|
||||||
'Voyage AI': 'voyage',
|
'Voyage AI': 'voyage',
|
||||||
|
'Google Cloud': 'google-cloud',
|
||||||
};
|
};
|
||||||
|
|
||||||
export const BedrockRegionList = [
|
export const BedrockRegionList = [
|
||||||
|
@ -0,0 +1,95 @@
|
|||||||
|
import { useTranslate } from '@/hooks/common-hooks';
|
||||||
|
import { IModalProps } from '@/interfaces/common';
|
||||||
|
import { IAddLlmRequestBody } from '@/interfaces/request/llm';
|
||||||
|
import { Form, Input, Modal, Select } from 'antd';
|
||||||
|
|
||||||
|
type FieldType = IAddLlmRequestBody & {
|
||||||
|
google_project_id: string;
|
||||||
|
google_region: string;
|
||||||
|
google_service_account_key: string;
|
||||||
|
};
|
||||||
|
|
||||||
|
const { Option } = Select;
|
||||||
|
|
||||||
|
const GoogleModal = ({
|
||||||
|
visible,
|
||||||
|
hideModal,
|
||||||
|
onOk,
|
||||||
|
loading,
|
||||||
|
llmFactory,
|
||||||
|
}: IModalProps<IAddLlmRequestBody> & { llmFactory: string }) => {
|
||||||
|
const [form] = Form.useForm<FieldType>();
|
||||||
|
|
||||||
|
const { t } = useTranslate('setting');
|
||||||
|
const handleOk = async () => {
|
||||||
|
const values = await form.validateFields();
|
||||||
|
|
||||||
|
const data = {
|
||||||
|
...values,
|
||||||
|
llm_factory: llmFactory,
|
||||||
|
};
|
||||||
|
|
||||||
|
onOk?.(data);
|
||||||
|
};
|
||||||
|
|
||||||
|
return (
|
||||||
|
<Modal
|
||||||
|
title={t('addLlmTitle', { name: llmFactory })}
|
||||||
|
open={visible}
|
||||||
|
onOk={handleOk}
|
||||||
|
onCancel={hideModal}
|
||||||
|
okButtonProps={{ loading }}
|
||||||
|
>
|
||||||
|
<Form
|
||||||
|
name="basic"
|
||||||
|
style={{ maxWidth: 600 }}
|
||||||
|
autoComplete="off"
|
||||||
|
layout={'vertical'}
|
||||||
|
form={form}
|
||||||
|
>
|
||||||
|
<Form.Item<FieldType>
|
||||||
|
label={t('modelType')}
|
||||||
|
name="model_type"
|
||||||
|
initialValue={'chat'}
|
||||||
|
rules={[{ required: true, message: t('modelTypeMessage') }]}
|
||||||
|
>
|
||||||
|
<Select placeholder={t('modelTypeMessage')}>
|
||||||
|
<Option value="chat">chat</Option>
|
||||||
|
</Select>
|
||||||
|
</Form.Item>
|
||||||
|
<Form.Item<FieldType>
|
||||||
|
label={t('modelID')}
|
||||||
|
name="llm_name"
|
||||||
|
rules={[{ required: true, message: t('GoogleModelIDMessage') }]}
|
||||||
|
>
|
||||||
|
<Input placeholder={t('GoogleModelIDMessage')} />
|
||||||
|
</Form.Item>
|
||||||
|
<Form.Item<FieldType>
|
||||||
|
label={t('addGoogleProjectID')}
|
||||||
|
name="google_project_id"
|
||||||
|
rules={[{ required: true, message: t('GoogleProjectIDMessage') }]}
|
||||||
|
>
|
||||||
|
<Input placeholder={t('GoogleProjectIDMessage')} />
|
||||||
|
</Form.Item>
|
||||||
|
<Form.Item<FieldType>
|
||||||
|
label={t('addGoogleRegion')}
|
||||||
|
name="google_region"
|
||||||
|
rules={[{ required: true, message: t('GoogleRegionMessage') }]}
|
||||||
|
>
|
||||||
|
<Input placeholder={t('GoogleRegionMessage')} />
|
||||||
|
</Form.Item>
|
||||||
|
<Form.Item<FieldType>
|
||||||
|
label={t('addGoogleServiceAccountKey')}
|
||||||
|
name="google_service_account_key"
|
||||||
|
rules={[
|
||||||
|
{ required: true, message: t('GoogleServiceAccountKeyMessage') },
|
||||||
|
]}
|
||||||
|
>
|
||||||
|
<Input placeholder={t('GoogleServiceAccountKeyMessage')} />
|
||||||
|
</Form.Item>
|
||||||
|
</Form>
|
||||||
|
</Modal>
|
||||||
|
);
|
||||||
|
};
|
||||||
|
|
||||||
|
export default GoogleModal;
|
@ -298,6 +298,33 @@ export const useSubmitFishAudio = () => {
|
|||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
|
export const useSubmitGoogle = () => {
|
||||||
|
const { addLlm, loading } = useAddLlm();
|
||||||
|
const {
|
||||||
|
visible: GoogleAddingVisible,
|
||||||
|
hideModal: hideGoogleAddingModal,
|
||||||
|
showModal: showGoogleAddingModal,
|
||||||
|
} = useSetModalState();
|
||||||
|
|
||||||
|
const onGoogleAddingOk = useCallback(
|
||||||
|
async (payload: IAddLlmRequestBody) => {
|
||||||
|
const ret = await addLlm(payload);
|
||||||
|
if (ret === 0) {
|
||||||
|
hideGoogleAddingModal();
|
||||||
|
}
|
||||||
|
},
|
||||||
|
[hideGoogleAddingModal, addLlm],
|
||||||
|
);
|
||||||
|
|
||||||
|
return {
|
||||||
|
GoogleAddingLoading: loading,
|
||||||
|
onGoogleAddingOk,
|
||||||
|
GoogleAddingVisible,
|
||||||
|
hideGoogleAddingModal,
|
||||||
|
showGoogleAddingModal,
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|
||||||
export const useSubmitBedrock = () => {
|
export const useSubmitBedrock = () => {
|
||||||
const { addLlm, loading } = useAddLlm();
|
const { addLlm, loading } = useAddLlm();
|
||||||
const {
|
const {
|
||||||
|
@ -32,11 +32,13 @@ import ApiKeyModal from './api-key-modal';
|
|||||||
import BedrockModal from './bedrock-modal';
|
import BedrockModal from './bedrock-modal';
|
||||||
import { IconMap } from './constant';
|
import { IconMap } from './constant';
|
||||||
import FishAudioModal from './fish-audio-modal';
|
import FishAudioModal from './fish-audio-modal';
|
||||||
|
import GoogleModal from './google-modal';
|
||||||
import {
|
import {
|
||||||
useHandleDeleteLlm,
|
useHandleDeleteLlm,
|
||||||
useSubmitApiKey,
|
useSubmitApiKey,
|
||||||
useSubmitBedrock,
|
useSubmitBedrock,
|
||||||
useSubmitFishAudio,
|
useSubmitFishAudio,
|
||||||
|
useSubmitGoogle,
|
||||||
useSubmitHunyuan,
|
useSubmitHunyuan,
|
||||||
useSubmitOllama,
|
useSubmitOllama,
|
||||||
useSubmitSpark,
|
useSubmitSpark,
|
||||||
@ -104,7 +106,8 @@ const ModelCard = ({ item, clickApiKey }: IModelCardProps) => {
|
|||||||
item.name === 'XunFei Spark' ||
|
item.name === 'XunFei Spark' ||
|
||||||
item.name === 'BaiduYiyan' ||
|
item.name === 'BaiduYiyan' ||
|
||||||
item.name === 'Fish Audio' ||
|
item.name === 'Fish Audio' ||
|
||||||
item.name === 'Tencent Cloud'
|
item.name === 'Tencent Cloud' ||
|
||||||
|
item.name === 'Google Cloud'
|
||||||
? t('addTheModel')
|
? t('addTheModel')
|
||||||
: 'API-Key'}
|
: 'API-Key'}
|
||||||
<SettingOutlined />
|
<SettingOutlined />
|
||||||
@ -186,6 +189,14 @@ const UserSettingModel = () => {
|
|||||||
HunyuanAddingLoading,
|
HunyuanAddingLoading,
|
||||||
} = useSubmitHunyuan();
|
} = useSubmitHunyuan();
|
||||||
|
|
||||||
|
const {
|
||||||
|
GoogleAddingVisible,
|
||||||
|
hideGoogleAddingModal,
|
||||||
|
showGoogleAddingModal,
|
||||||
|
onGoogleAddingOk,
|
||||||
|
GoogleAddingLoading,
|
||||||
|
} = useSubmitGoogle();
|
||||||
|
|
||||||
const {
|
const {
|
||||||
TencentCloudAddingVisible,
|
TencentCloudAddingVisible,
|
||||||
hideTencentCloudAddingModal,
|
hideTencentCloudAddingModal,
|
||||||
@ -235,6 +246,7 @@ const UserSettingModel = () => {
|
|||||||
BaiduYiyan: showyiyanAddingModal,
|
BaiduYiyan: showyiyanAddingModal,
|
||||||
'Fish Audio': showFishAudioAddingModal,
|
'Fish Audio': showFishAudioAddingModal,
|
||||||
'Tencent Cloud': showTencentCloudAddingModal,
|
'Tencent Cloud': showTencentCloudAddingModal,
|
||||||
|
'Google Cloud': showGoogleAddingModal,
|
||||||
}),
|
}),
|
||||||
[
|
[
|
||||||
showBedrockAddingModal,
|
showBedrockAddingModal,
|
||||||
@ -244,6 +256,7 @@ const UserSettingModel = () => {
|
|||||||
showSparkAddingModal,
|
showSparkAddingModal,
|
||||||
showyiyanAddingModal,
|
showyiyanAddingModal,
|
||||||
showFishAudioAddingModal,
|
showFishAudioAddingModal,
|
||||||
|
showGoogleAddingModal,
|
||||||
],
|
],
|
||||||
);
|
);
|
||||||
|
|
||||||
@ -364,6 +377,13 @@ const UserSettingModel = () => {
|
|||||||
loading={HunyuanAddingLoading}
|
loading={HunyuanAddingLoading}
|
||||||
llmFactory={'Tencent Hunyuan'}
|
llmFactory={'Tencent Hunyuan'}
|
||||||
></HunyuanModal>
|
></HunyuanModal>
|
||||||
|
<GoogleModal
|
||||||
|
visible={GoogleAddingVisible}
|
||||||
|
hideModal={hideGoogleAddingModal}
|
||||||
|
onOk={onGoogleAddingOk}
|
||||||
|
loading={GoogleAddingLoading}
|
||||||
|
llmFactory={'Google Cloud'}
|
||||||
|
></GoogleModal>
|
||||||
<TencentCloudModal
|
<TencentCloudModal
|
||||||
visible={TencentCloudAddingVisible}
|
visible={TencentCloudAddingVisible}
|
||||||
hideModal={hideTencentCloudAddingModal}
|
hideModal={hideTencentCloudAddingModal}
|
||||||
|
Loading…
x
Reference in New Issue
Block a user