LibreChat Multi-Provider

Deploy LibreChat on Clore.ai — a powerful, self-hosted ChatGPT alternative with multi-provider AI support, conversation branching, and plugin ecosystem on affordable GPU cloud infrastructure.

Overview

LibreChatarrow-up-right is an enhanced, open-source ChatGPT-like interface with 22K+ GitHub stars. It faithfully reimagines the ChatGPT experience while adding features the original lacks — multi-provider switching within the same conversation, conversation branching/forking, a rich plugin system, file uploads with vision, and a full code interpreter sandbox.

Why run LibreChat on Clore.ai?

  • True multi-provider in one UI — Switch between GPT-4, Claude 3.5, Gemini Pro, Mistral, and local Ollama models mid-session.

  • No GPU needed for the app — LibreChat is a Node.js application; it only needs compute for inference if you attach a local LLM backend.

  • Cost-effective self-hosting — Clore.ai pricing starts at fractions of a cent per minute, ideal for running a personal AI hub.

  • Persistent conversations — MongoDB stores your full chat history server-side, unlike browser-local solutions.

  • Team-friendly — Multi-user support with individual API key management.

Key Features

Feature
Description

Multi-provider

OpenAI, Anthropic, Google, Azure, Mistral, Ollama, OpenRouter

Conversation branching

Fork and explore alternative responses

Plugins

Bing search, Zapier, WolframAlpha, custom tools

File uploads

Images, PDFs, documents with vision analysis

Code interpreter

Execute Python in an isolated sandbox

Artifacts

Render HTML, React, and Markdown outputs

Presets

Save and share custom model configurations


Requirements

Server Specifications

Component
Minimum
Recommended
Notes

GPU

None required

RTX 3090 (if adding Ollama)

Only for local LLM inference

VRAM

24 GB

For local models via Ollama

CPU

2 vCPU

4 vCPU

Node.js + MongoDB

RAM

4 GB

8 GB

MongoDB benefits from more RAM

Storage

20 GB

50+ GB

File uploads, model cache if local

Clore.ai Pricing Reference

Server Type
Approx. Cost
Use Case

CPU-focused (4 vCPU, 8 GB RAM)

~$0.05–0.10/hr

LibreChat + external API providers

RTX 3090 (24 GB VRAM)

~$0.20/hr

LibreChat + Ollama local inference

RTX 4090 (24 GB VRAM)

~$0.35/hr

LibreChat + faster Ollama/vLLM

A100 80 GB

~$1.10/hr

LibreChat + large 70B+ models

💡 Cost tip: If you only use LibreChat to route API calls to OpenAI/Anthropic/Google, you only pay for the Clore.ai server compute (cheap), not the inference hardware. Budget ~$0.05–0.15/hr for a reliable LibreChat host.

Prerequisites

  • Clore.ai server with SSH access

  • Docker + Docker Compose (pre-installed on Clore.ai)

  • Git (pre-installed on Clore.ai)

  • At least one LLM API key or a local Ollama/vLLM backend


Quick Start

LibreChat's official deployment uses Docker Compose with MongoDB and MeiliSearch for full functionality.

Step 1: Connect to your Clore.ai server

Step 2: Clone the repository

Step 3: Configure environment

Set at minimum:

Generate secrets quickly:

Step 4: Start the stack

This starts:

  • LibreChat — main application on port 3080

  • MongoDB — conversation and user storage

  • MeiliSearch — fast conversation search

Step 5: Verify and access

Open in browser:

Register a new account on the login page.


Method 2: Pre-built Docker Image (Fastest)

If you want to skip building from source:


Method 3: Single-Container Quick Test

For a rapid proof-of-concept without MongoDB (limited functionality):

⚠️ This method requires a separate MongoDB instance. Use Method 1 for a complete setup.


Configuration

Adding AI Providers

Edit librechat.yaml (create it in the project root) for advanced provider configuration:

Mount this file in your docker-compose.yml:

Environment Variables Reference

Variable
Description
Example

MONGO_URI

MongoDB connection string

mongodb://mongodb:27017/LibreChat

JWT_SECRET

JWT signing secret (64+ chars)

Random hex string

OPENAI_API_KEY

OpenAI key

sk-...

ANTHROPIC_API_KEY

Anthropic key

sk-ant-...

GOOGLE_KEY

Google Gemini key

AI...

ALLOW_REGISTRATION

Enable public signup

true / false

ALLOW_EMAIL_LOGIN

Enable email/password login

true

DEBUG_LOGGING

Verbose logs

true

SEARCH

Enable MeiliSearch

true

MEILI_MASTER_KEY

MeiliSearch API key

Random string

Restricting Registration

For private use, disable public registration after creating your account:

Then restart: docker compose restart LibreChat

Enabling Code Interpreter

The code interpreter runs Python in an isolated Docker container. Ensure Docker socket is accessible.

File Upload Configuration


GPU Acceleration

LibreChat does not use GPU directly — it's a routing layer. GPU acceleration applies to any local inference backend you connect to it.

Connecting to Ollama (Same Server)

If running Ollama on the same Clore.ai server (see Ollama Guide):

Connecting to vLLM (High Throughput)

For high-concurrency deployments (see vLLM Guide):

In librechat.yaml:

GPU Sizing for Local Models

Model Size
Min VRAM
Recommended Clore GPU
Approx. Cost

7–8B (Q4)

6 GB

RTX 3090

~$0.20/hr

13B (Q4)

10 GB

RTX 3090

~$0.20/hr

34B (Q4)

24 GB

RTX 4090

~$0.35/hr

70B (Q4)

48 GB

2× RTX 3090

~$0.40/hr

70B (FP16)

80 GB

A100 80GB

~$1.10/hr


Tips & Best Practices

Cost Management on Clore.ai

Backup Strategy

Restoring from Backup

Securing LibreChat

  • Always set strong, unique values for JWT_SECRET and CREDS_KEY

  • Disable registration after initial user creation: ALLOW_REGISTRATION=false

  • Use a reverse proxy (nginx/Caddy) with HTTPS for production

  • Regularly update the Docker image: docker compose pull && docker compose up -d

Nginx Reverse Proxy (Optional)


Troubleshooting

Port 3080 not accessible

MongoDB connection refused

JWT / Authentication errors

Ollama models not appearing

Out of disk space

Update to latest version


Further Reading

Last updated

Was this helpful?