Skip to content

Commit 2e8da10

Browse files
infinityCuratorRAGFlow Curator
andauthored
Synchronize documentation. (#435)
Co-authored-by: RAGFlow Curator <infinitydocs.curator@users.noreply.github.com>
1 parent 6fb34ae commit 2e8da10

File tree

1 file changed

+29
-2
lines changed

1 file changed

+29
-2
lines changed

website/docs/faq.mdx

Lines changed: 29 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -540,11 +540,38 @@ uv pip install -U "mineru[core]" -i https://mirrors.aliyun.com/pypi/simple
540540

541541
### How to configure MinerU-specific settings?
542542

543-
1. Set `MINERU_EXECUTABLE` (default: `mineru`) to the path of the MinerU executable.
543+
1. Set `MINERU_EXECUTABLE` (default: `mineru`) to the path to the MinerU executable.
544544
2. Set `MINERU_DELETE_OUTPUT` to `0` to keep MinerU's output. (Default: `1`, which deletes temporary output)
545545
3. Set `MINERU_OUTPUT_DIR` to specify the output directory for MinerU.
546-
4. Set `MINERU_BACKEND` to `"pipeline"`. (Options: `"pipeline"` (default) | `"vlm-transformers"`)
546+
4. Set `MINERU_BACKEND` to specify a parsing backend:
547+
- `"pipeline"` (default): The traditional multimodel pipeline.
548+
- `"vlm-transformers"`: A vision-language model using HuggingFace Transformers.
549+
- `"vlm-vllm-engine"`: A vision-language model using local vLLM engine (requires a local GPU).
550+
- `"vlm-http-client"`: A vision-language model via HTTP client to remote vLLM server (RAGFlow only requires CPU).
551+
5. If using the `"vlm-http-client"` backend, you must also set `MINERU_SERVER_URL` to the URL of your vLLM server.
547552

548553
:::tip NOTE
549554
For information about other environment variables natively supported by MinerU, see [here](https://opendatalab.github.io/MinerU/usage/cli_tools/#environment-variables-description).
550555
:::
556+
557+
---
558+
559+
### How to use MinerU with a vLLM server for document parsing?
560+
561+
RAGFlow supports MinerU's `vlm-http-client` backend, enabling you to delegate document parsing tasks to a remote vLLM server. With this configuration, RAGFlow will connect to your remote vLLM server as a client and use its powerful GPU resources for document parsing. This significantly improves performance for parsing complex documents while reducing the resources required on your RAGFlow server. To configure MinerU with a vLLM server:
562+
563+
1. Set up a vLLM server running MinerU:
564+
```bash
565+
mineru-vllm-server --port 30000
566+
```
567+
568+
2. Configure the following environment variables in your **docker/.env** file:
569+
- `MINERU_EXECUTABLE=/ragflow/uv_tools/.venv/bin/mineru` (or the path to your MinerU executable)
570+
- `MINERU_BACKEND="vlm-http-client"`
571+
- `MINERU_SERVER_URL="http://your-vllm-server-ip:30000"`
572+
573+
3. Complete the rest standard MinerU setup steps as described [here](#how-to-configure-mineru-specific-settings).
574+
575+
:::tip NOTE
576+
When using the `vlm-http-client` backend, the RAGFlow server requires no GPU, only network connectivity. This enables cost-effective distributed deployment with multiple RAGFlow instances sharing one remote vLLM server.
577+
:::

0 commit comments

Comments
 (0)