From e786f596e292cd4a037e3eff6dc5ff417811d8f9 Mon Sep 17 00:00:00 2001 From: writinwaters <93570324+writinwaters@users.noreply.github.com> Date: Thu, 6 Feb 2025 17:14:13 +0800 Subject: [PATCH] Updated template description (#4744) ### What problem does this PR solve? ### Type of change - [x] Documentation Update --- agent/templates/research_report.json | 4 ++-- docs/guides/accelerate_question_answering.mdx | 2 +- docs/guides/develop/build_docker_image.mdx | 4 ++-- docs/quickstart.mdx | 2 +- docs/release_notes.md | 2 +- 5 files changed, 7 insertions(+), 7 deletions(-) diff --git a/agent/templates/research_report.json b/agent/templates/research_report.json index 18437725a..126967b73 100644 --- a/agent/templates/research_report.json +++ b/agent/templates/research_report.json @@ -1,7 +1,7 @@ { "id": 10, - "title": "Research Report Generator", - "description": "This generator can produce a research report based on the given title and language. It decomposes into sub-titles and queries to search engine from different angles, and generates sections based on search engine results and comprehension of the sub-titles.", + "title": "Research report generator", + "description": "A report generator that creates a research report from a given title, in the specified target language. It generates queries from the input title, then uses these to create subtitles and sections, compiling everything into a comprehensive report.", "canvas_type": "chatbot", "dsl": { "answer": [], diff --git a/docs/guides/accelerate_question_answering.mdx b/docs/guides/accelerate_question_answering.mdx index 9a3ed0fa4..42ef664bf 100644 --- a/docs/guides/accelerate_question_answering.mdx +++ b/docs/guides/accelerate_question_answering.mdx @@ -23,7 +23,7 @@ Please note that some of your settings may consume a significant amount of time. - In the **Prompt Engine** tab of your **Chat Configuration** dialogue, disabling **Multi-turn optimization** will reduce the time required to get an answer from the LLM. - In the **Prompt Engine** tab of your **Chat Configuration** dialogue, leaving the **Rerank model** field empty will significantly decrease retrieval time. -- In the **Assistant Setting** tab of your **Chat Configuration** dialogue, disabling **Keyword analysis** will reduce the time to get get an answer from the LLM. +- In the **Assistant Setting** tab of your **Chat Configuration** dialogue, disabling **Keyword analysis** will reduce the time to receive an answer from the LLM. - When chatting with your chat assistant, click the light bulb icon above the *current* dialogue and scroll down the popup window to view the time taken for each task: ![enlighten](https://github.com/user-attachments/assets/fedfa2ee-21a7-451b-be66-20125619923c) diff --git a/docs/guides/develop/build_docker_image.mdx b/docs/guides/develop/build_docker_image.mdx index 72d519035..bea9f4688 100644 --- a/docs/guides/develop/build_docker_image.mdx +++ b/docs/guides/develop/build_docker_image.mdx @@ -35,7 +35,7 @@ A guide explaining how to build a RAGFlow Docker image from its source code. By This image is approximately 2 GB in size and relies on external LLM and embedding services. :::tip NOTE -While we also test RAGFlow on ARM64 platforms, we do not plan to maintain RAGFlow Docker images for ARM. However, you can build an image yourself on a `linux/arm64` or `darwin/arm64` host machine as well. +While we also test RAGFlow on ARM64 platforms, we do not maintain RAGFlow Docker images for ARM. However, you can build an image yourself on a `linux/arm64` or `darwin/arm64` host machine as well. ::: ```bash @@ -51,7 +51,7 @@ docker build --build-arg LIGHTEN=1 -f Dockerfile -t infiniflow/ragflow:nightly-s This image is approximately 9 GB in size. As it includes embedding models, it relies on external LLM services only. :::tip NOTE -While we also test RAGFlow on ARM64 platforms, we do not plan to maintain RAGFlow Docker images for ARM. However, you can build an image yourself on a `linux/arm64` or `darwin/arm64` host machine. +While we also test RAGFlow on ARM64 platforms, we do not maintain RAGFlow Docker images for ARM. However, you can build an image yourself on a `linux/arm64` or `darwin/arm64` host machine. ::: ```bash diff --git a/docs/quickstart.mdx b/docs/quickstart.mdx index 372de68b2..e95d8b7ab 100644 --- a/docs/quickstart.mdx +++ b/docs/quickstart.mdx @@ -18,7 +18,7 @@ This quick start guide describes a general process from: - Establishing an AI chat based on your datasets. :::danger IMPORTANT -We officially support x86 CPU and Nvidia GPU, and this document offers instructions on deploying RAGFlow using Docker on x86 platforms. While we also test RAGFlow on ARM64 platforms, we do not plan to maintain RAGFlow Docker images for ARM. +We officially support x86 CPU and Nvidia GPU, and this document offers instructions on deploying RAGFlow using Docker on x86 platforms. While we also test RAGFlow on ARM64 platforms, we do not maintain RAGFlow Docker images for ARM. If you are on an ARM platform, follow [this guide](https://ragflow.io/docs/dev/build_docker_image) to build a RAGFlow Docker image. ::: diff --git a/docs/release_notes.md b/docs/release_notes.md index 32840c559..f2f4d54de 100644 --- a/docs/release_notes.md +++ b/docs/release_notes.md @@ -274,7 +274,7 @@ Released on May 31, 2024. - Supports ARM64 platforms. :::danger IMPORTANT -While we also test RAGFlow on ARM64 platforms, we do not plan to maintain RAGFlow Docker images for ARM. +While we also test RAGFlow on ARM64 platforms, we do not maintain RAGFlow Docker images for ARM. If you are on an ARM platform, follow [this guide](https://ragflow.io/docs/dev/build_docker_image) to build a RAGFlow Docker image. :::