Open WebUI
Recent Changes
Update ollama to 0.6.6
New, faster model downloading:
OLLAMA_EXPERIMENT=client2 ollama serve
will run Ollama using a new downloader with improved performance and reliability when runningollama pull
. Please share feedback here!Fixed memory leak issues when running Gemma 3, Mistral Small 3.1 and other models on Ollama
Improved performance of
ollama create
when importing models from SafetensorsOllama will now allow tool function parameters with either a single type or an array of types by @rozgo
Fixed certain out of memory issues from not reserving enough memory at startup
Fix nondeterministic model unload order by @IreGaddr
Include the
items
and$defs
fields to properly handlearray
types in the API by @shefflerOpenAI-Beta
headers are now included in the CORS safelist by @drifkinFixed issue where model tensor data would be corrupted when importing models from Safetensors
About
Open WebUI is an extensible, feature-rich, and user-friendly self-hosted WebUI for various LLM runners, supported LLM runners include Ollama and OpenAI-compatible APIs.
Features
- Intuitive Interface: Our chat interface takes inspiration from ChatGPT, ensuring a user-friendly experience.
- Responsive Design: Enjoy a seamless experience on both desktop and mobile devices.
- Swift Responsiveness: Enjoy fast and responsive performance.
- Effortless Setup: Install seamlessly using Docker or Kubernetes (kubectl, kustomize or helm) for a hassle-free experience.
- Code Syntax Highlighting: Enjoy enhanced code readability with our syntax highlighting feature.
- Full Markdown and LaTeX Support: Elevate your LLM experience with comprehensive Markdown and LaTeX capabilities for enriched interaction.
- Local RAG Integration
- Web Browsing Capability
- Prompt Preset Support
- RLHF Annotation
- Conversation Tagging
- Download/Delete Models
- GGUF File Model Creation
Install Open WebUI in a few minutes on your server with Cloudron. To install Cloudron first, follow our setup steps.