RAGFlow

Deploy RAGFlow deep document understanding RAG engine on Clore.ai GPUs

RAGFlow is an open-source Retrieval-Augmented Generation (RAG) engine with deep document understanding capabilities. With over 50,000 GitHub stars, it's one of the most comprehensive RAG platforms available — designed to extract, chunk, and reason over complex documents including PDFs, Word files, spreadsheets, images, and more.

Unlike basic RAG systems that naively split documents into chunks, RAGFlow uses layout-aware parsing to understand document structure, tables, figures, and multi-column layouts. This results in dramatically higher retrieval precision and answer quality.

Key features:

  • 📄 Deep document understanding — OCR, table extraction, figure recognition

  • 🔍 Multiple chunking strategies — semantic, layout-aware, fixed-size, Q&A style

  • 🤖 LLM integration — works with OpenAI, Ollama, Anthropic, local models

  • 🌐 Full-featured WebUI — drag-and-drop document management

  • 🔌 REST API — integrate RAGFlow into any application

  • 📊 Citation tracking — answers include source document references

  • 🏗️ Multi-tenant — team workspaces with permission control

circle-check

Server Requirements

Parameter
Minimum
Recommended

GPU

NVIDIA RTX 3080 (10 GB)

NVIDIA RTX 4090 (24 GB)

VRAM

8 GB

16–24 GB

RAM

16 GB

32–64 GB

CPU

8 cores

16+ cores

Disk

50 GB

100–500 GB

OS

Ubuntu 20.04+

Ubuntu 22.04

CUDA

11.8+

12.1+

Ports

22, 9380, 80

22, 9380, 80

Docker

Required

Docker + Docker Compose

circle-exclamation

Quick Deploy on CLORE.AI

1. Find a suitable server

Go to CLORE.AI Marketplacearrow-up-right and filter by:

  • VRAM: ≥ 8 GB

  • RAM: ≥ 16 GB

  • Disk: ≥ 50 GB

  • GPU: RTX 3090, 4090, A100, H100

2. Configure your deployment

Docker Image:

Port Mappings:

Startup Command:

3. Access the WebUI

Default credentials: [email protected] / admin


Step-by-Step Setup

Step 1: SSH into your server

Step 2: Install Docker Compose

Step 3: Clone the RAGFlow repository

Step 4: Configure environment

Key settings to configure:

Step 5: Choose the right image variant

Step 6: Start all services

Wait for:

Step 7: Create admin account

Open http://<server-ip>:80 and register the first admin account.

Step 8: Configure LLM model

  1. Go to Settings → Model Providers

  2. Add your LLM (OpenAI, Ollama, etc.)

  3. Set the default chat model and embedding model


Usage Examples

Example 1: Upload and Query Documents via WebUI

  1. Log in to http://<server-ip>:80

  2. Click "Knowledge Base""Create Knowledge Base"

  3. Name it: "Clore.ai Documentation"

  4. Upload PDF/Word/TXT files using drag-and-drop

  5. Wait for parsing (progress shown in UI)

  6. Go to "Chat" → Create a new assistant linked to your knowledge base

  7. Ask questions about your documents


Example 2: API — Create Knowledge Base and Upload Documents


Example 3: Query Documents via API


Example 4: Batch Document Processing Pipeline


Example 5: RAGFlow with Local Ollama LLM


Configuration

docker-compose.yml Key Services

Chunking Strategies

Method
Best For
Description

naive

General documents

Fixed-size chunks with overlap

qa

FAQ/Q&A documents

Splits on question-answer pairs

table

Spreadsheets, tables

Preserves table structure

paper

Academic papers

Sections, abstract, references

book

Long books, manuals

Chapter-aware chunking

laws

Legal documents

Article-based chunking

manual

Technical manuals

Section hierarchy preservation


Performance Tips

1. Scale Elasticsearch Memory

2. GPU-Accelerated Embedding

Configure RAGFlow to use a GPU-based embedding model:

  • In Settings → Model Providers, use a local GPU model via Ollama

  • Or point to a dedicated embedding service running on the Clore.ai GPU

3. Parallel Document Processing

RAGFlow processes documents in parallel by default. Configure worker count:

4. MinIO for Large Document Sets

For deployments with thousands of documents, configure dedicated MinIO storage with larger disk allocation in your CLORE.AI order.


Troubleshooting

Issue: Services fail to start (memory)

Issue: Cannot access WebUI on port 80

Issue: Document parsing stuck

Issue: Elasticsearch heap out of memory

Issue: Embedding model not found



Clore.ai GPU Recommendations

Use Case
Recommended GPU
Est. Cost on Clore.ai

Development/Testing

RTX 3090 (24GB)

~$0.12/gpu/hr

Production RAG

RTX 3090 (24GB)

~$0.12/gpu/hr

High-throughput Embedding

RTX 4090 (24GB)

~$0.70/gpu/hr

💡 All examples in this guide can be deployed on Clore.aiarrow-up-right GPU servers. Browse available GPUs and rent by the hour — no commitments, full root access.

Last updated

Was this helpful?