diff --git a/docs/guides/accelerate_question_answering.mdx b/docs/guides/accelerate_question_answering.mdx new file mode 100644 index 000000000..e1c258155 --- /dev/null +++ b/docs/guides/accelerate_question_answering.mdx @@ -0,0 +1,50 @@ +--- +sidebar_position: 2 +slug: /accelerate_doc_indexing_and_question_answering +--- + +# Accelerate document indexing and question answering +import APITable from '@site/src/components/APITable'; + +A checklist to speed up document parsing and question answering. + +--- + +Please note that some of your settings may consume a significant amount of time. If you often find that document parsing and question answering are time-consuming, here is a checklist to consider: + +## 1. Accelerate document indexing + +- Use GPU to reduce embedding time. +- On the configuration page of your knowledge base, toggle off **Use RAPTOR to enhance retrieval**. +- The **Knowledge Graph** chunk method (GraphRAG) is time-consuming. +- Disable **Auto-keyword** and **Auto-question** on the configuration page of yor knowledge base, as both depend on the LLM. + +## 2. Accelerate question answering + +- In the **Prompt Engine** tab of your **Chat Configuration** dialogue, disabling **Multi-turn optimization** will reduce the time required to get an answer from the LLM. +- In the **Prompt Engine** tab of your **Chat Configuration** dialogue, leaving the **Rerank model** field empty will significantly decrease retrieval time. +- In the **Assistant Setting** tab of your **Chat Configuration** dialogue, disabling **Keyword analysis** will reduce the time to get get an answer from the LLM. +- When chatting with your chat assistant, click the light bulb icon above the *current* dialogue and scroll down the popup window to view the time taken for each task: + ![enlighten](https://github.com/user-attachments/assets/fedfa2ee-21a7-451b-be66-20125619923c) + + +```mdx-code-block + +``` + +| Item name | Description | +| ----------------- | ------------------------------------------------------------ | +| Total | Total time spent on this conversation round, including chunk retrieval and answer generation. | +| Check LLM | Time to validate the specified LLM. | +| Create retriever | Time to create a chunk retriever. | +| Bind embedding | Time to initialize an embedding model instance. | +| Bind LLM | Time to intialize an LLM instance. | +| Tune question | Time to optimize the user query using the context of the mult-turn conversation. | +| Bind reranker | Time to initialize an reranker model instance for chunk retrieval. | +| Generate keywords | Time to extract keywords from the user query. | +| Retrieval | Time to retrieve the chunks. | +| Generate answer | Time to generate the answer. | + +```mdx-code-block + +``` \ No newline at end of file diff --git a/docs/guides/configure_knowledge_base.md b/docs/guides/configure_knowledge_base.md index 4a3f5b6cf..43378365c 100644 --- a/docs/guides/configure_knowledge_base.md +++ b/docs/guides/configure_knowledge_base.md @@ -1,5 +1,5 @@ --- -sidebar_position: 1 +sidebar_position: 0 slug: /configure_knowledge_base --- diff --git a/docs/guides/develop/build_docker_image.mdx b/docs/guides/develop/build_docker_image.mdx index 48c60bdfa..91f06fcf3 100644 --- a/docs/guides/develop/build_docker_image.mdx +++ b/docs/guides/develop/build_docker_image.mdx @@ -12,6 +12,7 @@ A guide explaining how to build a RAGFlow Docker image from its source code. By ## Target Audience - Developers who have added new features or modified the existing code and require a Docker image to view and debug their changes. +- Developers looking to build a RAGFlow Docker image for an ARM64 platform. - Testers looking to explore the latest features of RAGFlow in a Docker image. ## Prerequisites diff --git a/docs/guides/start_chat.md b/docs/guides/start_chat.md index 0eaf0d8f7..7bc234cde 100644 --- a/docs/guides/start_chat.md +++ b/docs/guides/start_chat.md @@ -1,5 +1,5 @@ --- -sidebar_position: 2 +sidebar_position: 1 slug: /start_chat --- @@ -57,11 +57,11 @@ You start an AI conversation by creating an assistant. :::tip NOTE -1. Click the light bubble icon above the answer to view the expanded system prompt: +1. Click the light bulb icon above the answer to view the expanded system prompt: ![](https://github.com/user-attachments/assets/515ab187-94e8-412a-82f2-aba52cd79e09) - *The light bubble icon is available only for the current dialogue.* + *The light bulb icon is available only for the current dialogue.* 2. Scroll down the expanded prompt to view the time consumed for each task: