RUNLOCALAIv38
→WILL IT RUNBEST GPUCOMPARETROUBLESHOOTSTARTPULSEMODELSHARDWARETOOLSBENCH
RUNLOCALAI

Operator-grade instrument for local-AI hardware intelligence. Hand-written verdicts. Real benchmarks. Reproducible commands.

OP·Fredoline Eruo
DIR
  • Models
  • Hardware
  • Tools
  • Benchmarks
  • Will it run?
GUIDES
  • Best GPU
  • Best laptop
  • Best Mac
  • Best used GPU
  • Best budget GPU
  • Best GPU for Ollama
  • Best GPU for SD
  • AI PC build $2K
  • CUDA vs ROCm
  • 16 vs 24 GB
  • Compare hardware
  • Custom compare
REF
  • Systems
  • Ecosystem maps
  • Pillar guides
  • Methodology
  • Glossary
  • Errors KB
  • Troubleshooting
  • Resources
  • Public API
EDITOR
  • About
  • About the author
  • Changelog
  • Latest
  • Updates
  • Submit benchmark
  • Send feedback
  • Trust
  • Editorial policy
  • How we make money
  • Contact
LEGAL
  • Privacy
  • Terms
  • Sitemap
MAIL · MONTHLY DIGEST
Get monthly local AI changes
Monthly recap. No spam.
DISCLOSURE

Some links on this site are affiliate links (Amazon Associates and other first-class retailers). When you buy through them, we earn a small commission at no extra cost to you. Affiliate links do not influence our verdicts — there are cards we rate highly that we don't have affiliate relationships with, and cards that sell well that we refuse to recommend. Read more →

SYS · ONLINEUPTIME · 100%2026 · operator-owned
RUNLOCALAI · v38
  1. >
  2. Home
  3. /Tools
  4. /LibreChat
gui
Open source
free + open-source

LibreChat

Open-source ChatGPT clone with multi-provider support (OpenAI, Anthropic, local LLMs via OpenAI-compatible APIs). The most popular self-hosted ChatGPT-shaped frontend. Strong multi-user + RAG + plugin support; pairs well with vLLM / Ollama / LM Studio backends.

By Fredoline Eruo·Last verified May 7, 2026·24,000 GitHub stars

Overview

What LibreChat actually is

LibreChat is the most production-grade self-hosted ChatGPT-shaped frontend in 2026. It started as a faithful clone of the ChatGPT UI but has evolved into something distinctly more capable: multi-provider routing (OpenAI, Anthropic, Google, AWS Bedrock, OpenRouter, plus any OAI-compatible local backend), first-class multi-user authentication, per-user usage tracking, plugin / tool support, RAG over uploaded files, image generation, and presets that let users bookmark prompt + model + parameter combinations.

If Open WebUI is the homelab default and AnythingLLM is the document-first default, LibreChat is the team-first multi-provider default. It's what you reach for when "we have ten people, some prefer Claude, some prefer Llama 3.1 70B local, and we want one auth-gated UI in front of all of them."

Where it fits in the stack

LibreChat is a frontend with serious operational machinery:

  • Frontend: React web UI (Mongo + Meilisearch backed)
  • Backend services: Node.js API server + Mongo (chats, users, presets) + Meilisearch (search) + optional Redis
  • LLM providers (configurable): Ollama, vLLM, LM Studio, OpenAI, Anthropic, AWS Bedrock, Google Vertex, OpenRouter, Groq, custom OAI-shape
  • Auth: local accounts, OAuth (Google, GitHub, Discord, etc.), LDAP, OpenID Connect
  • Plugins / tools: native plugin system + MCP support landed in 2025

The Mongo dependency is the biggest difference from Open WebUI's SQLite-default — LibreChat is architected for "real" multi-tenant deployments, not laptop convenience.

Best use cases

  • Internal team AI gateway. Auth + per-user usage tracking + multi-provider routing in one deploy.
  • Mixed local + cloud teams. Some queries hit local Llama 3.1 70B (privacy / cost), some hit Claude (frontier capability).
  • Prompt-engineering workbench. The "presets" system saves model + system prompt + parameter combinations, makes prompt iteration tractable.
  • Coding / docs assistants behind SSO. OIDC integration means LibreChat fits a corp SSO setup without bolt-on glue.
  • Cost-controlled OpenAI replacement with audit trail. Mongo retains all chat history; per-user spend reports are a built-in feature.

OS support

OS Quality Notes
Linux (Docker) excellent the production path
Linux (native Node) excellent works for dev
macOS excellent dev or single-host
Windows excellent dev or single-host
Windows (WSL2) excellent matches Linux

LibreChat is genuinely cross-platform for development; for production almost everyone runs Docker Compose with Mongo + Meilisearch + the LibreChat container.

Hardware / runtime support

Like AnythingLLM, LibreChat doesn't do inference. Hardware support is whatever the backend offers:

  • NVIDIA CUDA via Ollama / vLLM / LM Studio
  • Apple Silicon Metal via Ollama / LM Studio / MLX-LM
  • AMD ROCm via Ollama / llama.cpp
  • Intel Arc / NPU laptops via Ollama / OpenVINO
  • Cloud APIs if local isn't the goal

The frontend itself runs comfortably on a 2-CPU / 2 GB Linux VM; the heavy resource consumer is the backend Mongo if you have many users.

Model / quant format support

Whatever the backend supports. LibreChat passes the configured model name through the OpenAI-compatible API. Format details are someone else's problem.

For the cross-runtime view see /systems/quantization-formats.

Setup path

Docker Compose is the production path:

git clone https://github.com/danny-avila/LibreChat.git
cd LibreChat
cp .env.example .env
# edit .env to point at backends
docker compose up -d

That brings up Mongo, Meilisearch, the API server, and the React frontend. Browse to http://localhost:3080, register the first user (which becomes admin).

To wire in a local Ollama:

ENDPOINTS=openAI
OPENAI_REVERSE_PROXY=http://host.docker.internal:11434/v1
OPENAI_MODELS=llama3.1:8b,qwen2.5:32b

What breaks first

  1. Mongo connection drift. A Mongo upgrade or auth-change kills the whole stack; back up the Mongo volume before any version bump.
  2. Backend URL inside Docker. localhost from inside the LibreChat container is the container, not your host — you need host.docker.internal (Windows / Mac) or the host network IP (Linux).
  3. OAuth callback URLs. Setting up Google / GitHub OAuth requires the exact callback URL match; misconfigured ones produce silent auth failures.
  4. Plugin / MCP wiring. Plugins with their own auth (browser, code interpreter) need separate config and break on minor version bumps more often than the core does.
  5. Meilisearch index drift on big chat history. Search-related crashes after schema bumps; usually a re-index fixes it but it's a real ops chore.

Alternatives by intent

If you want… Reach for
Lighter, single-host, no Mongo Open WebUI
Document-first, workspace UX AnythingLLM
In-IDE assistant Continue.dev
Pure local desktop app LM Studio
Customizable React app fork LibreChat itself — it's MIT-licensed

Best pairings

  • Ollama for local + OpenAI/Anthropic for cloud — the canonical hybrid setup
  • vLLM + 32B AWQ-INT4 + LibreChat — the team-grade local-only deployment
  • OIDC SSO + LibreChat + local backend — the compliance-friendly internal-AI gateway
  • Apple M3 Ultra + MLX-LM 70B + LibreChat — the high-VRAM Mac team backend

Who should avoid LibreChat

  • Single-user laptops with no team need. Mongo overhead isn't worth it; use Open WebUI.
  • Document-RAG-first operators. Use AnythingLLM; LibreChat's RAG is fine but not the focus.
  • Teams that want a managed product. Self-hosting LibreChat is real ops work; if that's not viable, a hosted ChatGPT Team / Claude Team plan is honest about the trade.
  • Operators on hostile networks where Mongo can't run. Architectural mismatch.

Related

  • Stacks: /stacks/local-coding-agent, /stacks/private-rag-laptop
  • System guides: /setup, /compatibility
  • Tools: Ollama, vLLM, Open WebUI
  • Hardware: RTX 4090, Apple M3 Ultra, RTX 3090

Pros

  • Multi-provider out of the box — local + cloud APIs in one UI
  • First-class multi-user with auth + per-user usage tracking
  • Active development + large community

Cons

  • Heavier deployment than Open WebUI (Mongo dependency)
  • RAG path narrower than AnythingLLM for document-grounding workflows
  • Plugin ecosystem trails ChatGPT and OpenAI's mainline

Compatibility

Operating systems
Windows
macOS
Linux
Docker
GPU backends
n/a — frontend
LicenseOpen source · free + open-source

Runtime health

Operator-grade signals on how actively LibreChat is being maintained, how fresh its measurements are, and what failure classes operators have flagged. Every label below is anchored to a real date or count — we never infer maintainer activity we can't show.

Release cadence

Derived from the most recent editorial signal on this row.

Active
Updated May 7, 2026

6 days since last refresh · source: lastUpdated

Benchmark freshness

How recent the editorial measurements on this runtime are.

0editorial benchmarks

No editorial benchmarks for this runtime yet.

Community reproduction

Submissions that match an editorial measurement on similar hardware.

0reproduced reports

No community reproductions on file yet.

Get LibreChat

Official site
https://www.librechat.ai
GitHub
https://github.com/danny-avila/LibreChat

Frequently asked

Is LibreChat free?

LibreChat has a paid tier (free + open-source). Check the pricing page for current terms.

What operating systems does LibreChat support?

LibreChat supports Windows, macOS, Linux, Docker.

Which GPUs work with LibreChat?

LibreChat supports n/a — frontend. CPU-only inference is also possible but slow.
See something off?Report outdated·Suggest a correctionWe read every submission. Editorial review takes 1-7 days.

Reviewed by RunLocalAI Editorial. See our editorial policy for how we evaluate tools.

Related — keep moving

Compare hardware
  • RTX 3090 vs RTX 4090 →
  • RTX 4090 vs RTX 5090 →
Buyer guides
  • Best GPU for Ollama →
  • Best GPU for local AI (pillar) →
When it doesn't work
  • Ollama running slow →
  • Ollama port 11434 conflict →
  • Ollama model not found →
  • CUDA out of memory →
Recommended hardware
  • RTX 3090 (24 GB used) →
  • RTX 4060 Ti 16 GB (entry) →
Alternatives
Text Generation WebUI (oobabooga)JanStable Diffusion WebUI (AUTOMATIC1111)GPT4AllMstyOpen WebUIComfyUIKoboldCPP
Before you buy

Verify LibreChat runs on your specific hardware before committing money.

Will it run on my hardware? →Custom hardware comparison →GPU recommender (4 questions) →