From 202acbd628c5eb5e8409446449914fcaa1fc8501 Mon Sep 17 00:00:00 2001 From: Debug Doctor <109025558+debugdoctor@users.noreply.github.com> Date: Tue, 4 Mar 2025 11:35:25 +0800 Subject: [PATCH] Perf: update novita.ai LLM library (#5574) ### What problem does this PR solve? LLM library update ### Type of change - [x] Other : config update --- conf/llm_factories.json | 214 +++++++++++++++++++++++++++------------- 1 file changed, 146 insertions(+), 68 deletions(-) diff --git a/conf/llm_factories.json b/conf/llm_factories.json index db7864a76..39d780304 100644 --- a/conf/llm_factories.json +++ b/conf/llm_factories.json @@ -2292,11 +2292,83 @@ { "name": "novita.ai", "logo": "", - "tags": "LLM", + "tags": "LLM,IMAGE2TEXT", "status": "1", "llm": [ { - "llm_name": "meta-llama/llama-3-8b-instruct", + "llm_name": "deepseek/deepseek-r1", + "tags": "LLM,CHAT,64k", + "max_tokens": 64000, + "model_type": "chat" + }, + { + "llm_name": "deepseek/deepseek_v3", + "tags": "LLM,CHAT,64k", + "max_tokens": 64000, + "model_type": "chat" + }, + { + "llm_name": "deepseek/deepseek-r1-distill-llama-70b", + "tags": "LLM,CHAT,64k", + "max_tokens": 64000, + "model_type": "chat" + }, + { + "llm_name": "deepseek/deepseek-r1-distill-qwen-32b", + "tags": "LLM,CHAT,64k", + "max_tokens": 64000, + "model_type": "chat" + }, + { + "llm_name": "deepseek/deepseek-r1-distill-qwen-14b", + "tags": "LLM,CHAT,64k", + "max_tokens": 64000, + "model_type": "chat" + }, + { + "llm_name": "deepseek/deepseek-r1-distill-llama-8b", + "tags": "LLM,CHAT,32k", + "max_tokens": 32000, + "model_type": "chat" + }, + { + "llm_name": "meta-llama/llama-3.3-70b-instruct", + "tags": "LLM,CHAT,128k", + "max_tokens": 131072, + "model_type": "chat" + }, + { + "llm_name": "meta-llama/llama-3.2-11b-vision-instruct", + "tags": "LLM,CHAT,32k", + "max_tokens": 32768, + "model_type": "chat" + }, + { + "llm_name": "meta-llama/llama-3.2-3b-instruct", + "tags": "LLM,CHAT,32k", + "max_tokens": 32768, + "model_type": "chat" + }, + { + "llm_name": "meta-llama/llama-3.2-1b-instruct", + "tags": "LLM,CHAT,32k", + "max_tokens": 32768, + "model_type": "chat" + }, + { + "llm_name": "meta-llama/llama-3.1-70b-instruct", + "tags": "LLM,CHAT,32k", + "max_tokens": 32768, + "model_type": "chat" + }, + { + "llm_name": "meta-llama/llama-3.1-8b-instruct", + "tags": "LLM,CHAT,16k", + "max_tokens": 16384, + "model_type": "chat" + }, + { + "llm_name": "meta-llama/llama-3.1-8b-instruct-bf16", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" @@ -2307,58 +2379,34 @@ "max_tokens": 8192, "model_type": "chat" }, + { + "llm_name": "meta-llama/llama-3-8b-instruct", + "tags": "LLM,CHAT,8k", + "max_tokens": 8192, + "model_type": "chat" + }, + { + "llm_name": "qwen/qwen-2.5-72b-instruct", + "tags": "LLM,CHAT,32k", + "max_tokens": 32000, + "model_type": "chat" + }, + { + "llm_name": "qwen/qwen-2-vl-72b-instruct", + "tags": "LLM,IMAGE2TEXT,32k", + "max_tokens": 32768, + "model_type": "image2text" + }, + { + "llm_name": "qwen/qwen-2-7b-instruct", + "tags": "LLM,CHAT,32k", + "max_tokens": 32768, + "model_type": "chat" + }, { "llm_name": "mistralai/mistral-nemo", - "tags": "LLM,CHAT,32k", - "max_tokens": 32768, - "model_type": "chat" - }, - { - "llm_name": "microsoft/wizardlm-2-7b", - "tags": "LLM,CHAT,32k", - "max_tokens": 32768, - "model_type": "chat" - }, - { - "llm_name": "openchat/openchat-7b", - "tags": "LLM,CHAT,4k", - "max_tokens": 4096, - "model_type": "chat" - }, - { - "llm_name": "meta-llama/llama-3.1-8b-instruct", - "tags": "LLM,CHAT,8k", - "max_tokens": 8192, - "model_type": "chat" - }, - { - "llm_name": "meta-llama/llama-3.1-70b-instruct", - "tags": "LLM,CHAT,8k", - "max_tokens": 8192, - "model_type": "chat" - }, - { - "llm_name": "meta-llama/llama-3.1-405b-instruct", - "tags": "LLM,CHAT,32k", - "max_tokens": 32768, - "model_type": "chat" - }, - { - "llm_name": "google/gemma-2-9b-it", - "tags": "LLM,CHAT,8k", - "max_tokens": 8192, - "model_type": "chat" - }, - { - "llm_name": "jondurbin/airoboros-l2-70b", - "tags": "LLM,CHAT,4k", - "max_tokens": 4096, - "model_type": "chat" - }, - { - "llm_name": "nousresearch/hermes-2-pro-llama-3-8b", - "tags": "LLM,CHAT,8k", - "max_tokens": 8192, + "tags": "LLM,CHAT,128k", + "max_tokens": 131072, "model_type": "chat" }, { @@ -2368,19 +2416,43 @@ "model_type": "chat" }, { - "llm_name": "cognitivecomputations/dolphin-mixtral-8x22b", - "tags": "LLM,CHAT,15k", - "max_tokens": 16000, + "llm_name": "Sao10K/L3-8B-Stheno-v3.2", + "tags": "LLM,CHAT,8k", + "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "sao10k/l3-70b-euryale-v2.1", - "tags": "LLM,CHAT,15k", - "max_tokens": 16000, + "tags": "LLM,CHAT,8k", + "max_tokens": 8192, "model_type": "chat" }, { - "llm_name": "sophosympatheia/midnight-rose-70b", + "llm_name": "sao10k/l3-8b-lunaris", + "tags": "LLM,CHAT,8k", + "max_tokens": 8192, + "model_type": "chat" + }, + { + "llm_name": "sao10k/l31-70b-euryale-v2.2", + "tags": "LLM,CHAT,8k", + "max_tokens": 8192, + "model_type": "chat" + }, + { + "llm_name": "nousresearch/hermes-2-pro-llama-3-8b", + "tags": "LLM,CHAT,8k", + "max_tokens": 8192, + "model_type": "chat" + }, + { + "llm_name": "nousresearch/nous-hermes-llama2-13b", + "tags": "LLM,CHAT,4k", + "max_tokens": 4096, + "model_type": "chat" + }, + { + "llm_name": "openchat/openchat-7b", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" @@ -2392,19 +2464,25 @@ "model_type": "chat" }, { - "llm_name": "nousresearch/nous-hermes-llama2-13b", - "tags": "LLM,CHAT,4k", - "max_tokens": 4096, + "llm_name": "microsoft/wizardlm-2-8x22b", + "tags": "LLM,CHAT,65k", + "max_tokens": 65535, "model_type": "chat" }, { - "llm_name": "Nous-Hermes-2-Mixtral-8x7B-DPO", - "tags": "LLM,CHAT,32k", - "max_tokens": 32768, + "llm_name": "google/gemma-2-9b-it", + "tags": "LLM,CHAT,8k", + "max_tokens": 8192, "model_type": "chat" }, { - "llm_name": "lzlv_70b", + "llm_name": "cognitivecomputations/dolphin-mixtral-8x22b", + "tags": "LLM,CHAT,16k", + "max_tokens": 16000, + "model_type": "chat" + }, + { + "llm_name": "jondurbin/airoboros-l2-70b", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" @@ -2416,9 +2494,9 @@ "model_type": "chat" }, { - "llm_name": "microsoft/wizardlm-2-8x22b", - "tags": "LLM,CHAT,64k", - "max_tokens": 65535, + "llm_name": "sophosympatheia/midnight-rose-70b", + "tags": "LLM,CHAT,4k", + "max_tokens": 4096, "model_type": "chat" } ]