Open WebUI

A self-hosted, ChatGPT-like web interface for local LLMs. Connects to Ollama’s local API and renders streaming chat with a polished UI — file uploads, markdown rendering, edit/copy actions, multi-user accounts, image-gen integration via Automatic1111.

  • GitHub: open-webui/open-webui
  • Install: Docker container; one command from the project README
  • License: Open source

Why It’s Used

Ollama exposes local models as a REST API, but the bare CLI is awkward for sustained work. Open WebUI sits on top and makes local LLMs feel like commercial chat — without sending anything off the machine. Common stack: Ollama + Docker + Open WebUI for fully-offline ChatGPT.

Backend Options

Open WebUI sits on top of an OpenAI-API-compatible local model runtime. Two practical choices:

  • Ollama — the canonical pairing; standalone install, simplest setup
  • Docker Model Runner — newer alternative built into Docker Desktop; community compose file by Bret Fisher wires the two together via OPENAI_API_BASE_URL

Use Cases Covered in Wiki Sources

  • Cybersecurity communication — translating technical findings into plain English without leaking data to cloud LLMs (Joshua Clarke video)
  • Private chat over local models — the canonical use case

See Also