From 934dbc2e2bd9c11c6e546903308ceaf92521e92d Mon Sep 17 00:00:00 2001 From: Kevin Hu Date: Tue, 3 Dec 2024 15:18:38 +0800 Subject: [PATCH] Add more mistral models. (#3826) ### What problem does this PR solve? #3647 ### Type of change - [x] New Feature (non-breaking change which adds functionality) --- agent/canvas.py | 1 + conf/llm_factories.json | 48 ++++++++++++++++++++++++++++++----------- 2 files changed, 37 insertions(+), 12 deletions(-) diff --git a/agent/canvas.py b/agent/canvas.py index dc84a23aa..72ab3c103 100644 --- a/agent/canvas.py +++ b/agent/canvas.py @@ -234,6 +234,7 @@ class Canvas(ABC): except Exception as e: yield {"content": "*Exception*: {}".format(e), "running_status": True} logging.exception("Canvas.run got exception") + ran += 1 if ran >= len(self.path[-1]) and waiting: without_dependent_checking = waiting diff --git a/conf/llm_factories.json b/conf/llm_factories.json index 0250b8913..3caa831c7 100644 --- a/conf/llm_factories.json +++ b/conf/llm_factories.json @@ -618,26 +618,32 @@ }, { "llm_name": "open-mistral-7b", - "tags": "LLM,CHAT,32k", - "max_tokens": 32000, + "tags": "LLM,CHAT,128k", + "max_tokens": 128000, + "model_type": "chat" + }, + { + "llm_name": "ministral-8b-latest", + "tags": "LLM,CHAT,128k", + "max_tokens": 128000, + "model_type": "chat" + }, + { + "llm_name": "ministral-3b-latest", + "tags": "LLM,CHAT,128k", + "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "mistral-large-latest", - "tags": "LLM,CHAT,32k", - "max_tokens": 32000, + "tags": "LLM,CHAT,128k", + "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "mistral-small-latest", - "tags": "LLM,CHAT,32k", - "max_tokens": 32000, - "model_type": "chat" - }, - { - "llm_name": "mistral-medium-latest", - "tags": "LLM,CHAT,32k", - "max_tokens": 32000, + "tags": "LLM,CHAT,128k", + "max_tokens": 128000, "model_type": "chat" }, { @@ -646,11 +652,29 @@ "max_tokens": 32000, "model_type": "chat" }, + { + "llm_name": "mistral-nemo", + "tags": "LLM,CHAT,128k", + "max_tokens": 128000, + "model_type": "chat" + }, { "llm_name": "mistral-embed", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "embedding" + }, + { + "llm_name": "pixtral-large-latest", + "tags": "LLM,CHAT,32k", + "max_tokens": 32000, + "model_type": "image2text" + }, + { + "llm_name": "pixtral-12b", + "tags": "LLM,CHAT,32k", + "max_tokens": 32000, + "model_type": "image2text" } ] },