AnythingLLM

AnythingLLM is an open-source, all-in-one AI desktop app that bundles RAG, AI agents, document chat, and multi-model support into a single privacy-first package.

What is AnythingLLM?

AnythingLLM is a full-stack AI application built by Mintplex Labs (YCombinator S22, founded by Timothy Carambat). It lets you chat with documents, run AI agents, and use retrieval-augmented generation (RAG) — all locally and offline if you want. Think of it as a self-hostable, model-agnostic ChatGPT alternative where you bring your own LLM.

The project is MIT-licensed and has 55k+ stars on GitHub.


AnythingLLM logo


GitHub logo

Deployment Options

Mode Users Local LLM built-in MCP Support Notes
Desktop Single-user Yes (one-click) No macOS, Windows, Linux. No account needed.
Docker Multi-user No (connect remote) Yes Self-hosted, admin controls, whitelabel.
Cloud Multi-user No (bring your own) No Hosted by Mintplex Labs. Isolated instances.

Key Features

  • Any LLM — Use built-in local models (downloaded in-app) or connect to OpenAI, Anthropic, Azure, AWS Bedrock, Ollama, LM Studio, Groq, OpenRouter, LiteLLM, and more.
  • Document ingestion — PDFs, Word docs, CSVs, codebases, and web content. Documents are chunked, embedded, and stored in your chosen vector database.
  • Workspaces — Documents are scoped into workspaces (like threads with containerised context). Recent versions also support full-document context when the model's context window allows it, falling back to RAG when it doesn't.
  • AI Agents — Tool-calling agents with native function calling support. Agents can do web search/scraping, deep research, and cross-app actions. Skills are available from the AnythingLLM Hub.
  • No-code agent builder — Build agentic flows without writing code.
  • MCP compatibility — Docker deployments support Model Context Protocol tools for interoperability with external toolchains.
  • Multi-modal — Works with text-only and multi-modal LLMs (images, audio) in a single interface.
  • Developer API — Built-in REST API for integration into existing products.
  • Privacy by default — Everything stored and run locally. Nothing shared unless you allow it. Telemetry can be disabled.

Architecture

AnythingLLM is a full-stack JavaScript/TypeScript application with three main components:

  • Server — Backend API (Node.js)
  • Frontend — React-based UI
  • Collector — Document ingestion and processing pipeline

Storage uses SQLite for app data and LanceDB (or your choice of vector DB) for embeddings. The whole thing can be persisted via a single Docker volume mount at /app/server/storage.

Who It's For

  • Individual users who want a private, local AI assistant without cloud subscriptions or complex setup.
  • Teams and enterprises who need multi-user access, admin controls, role-based permissions, and whitelabel branding.
  • Developers who want an extensible platform with API access, plugin ecosystem, and MCP support.

Relevance

Worth watching as a self-hosted alternative to cloud AI platforms. The combination of zero-setup desktop experience + Docker multi-user deployment + open-source MIT license hits a sweet spot for privacy-conscious users and organisations. The MCP compatibility and native tool calling in recent releases make it increasingly viable as an agentic platform, not just a document Q&A tool.

links

social