The only local AI app that combines uncensored chat, image generation, and video creation in one interface. No cloud, no subscriptions, no data collection.
Download v1.9.0
macOS: Build from source
Your AI uses tools autonomously — web search, page fetching, file I/O, code execution. Multi-step chains with live tool-call blocks. Powered by Hermes 3.
Powered by Ollama. Run abliterated models like Llama 3.1, Qwen 3.5, or DeepSeek R1 locally — uncensored, private, with streaming responses and thinking display.
Text-to-image via ComfyUI. Supports Stable Diffusion XL, FLUX.1 Schnell, Pony Diffusion, and Juggernaut XL checkpoints. Full parameter control, no content filter.
Wan 2.1/2.2 and AnimateDiff support. Generate video clips from text on your own GPU. No cloud API, no watermarks, no usage limits.
Pre-built characters from Helpful Assistant to Roast Master. Switch personalities without prompt engineering. Works with any uncensored Ollama model.
Browse, install, and switch AI models from within the app. Auto-detects text, image, and video models across Ollama and ComfyUI backends.
Everything runs on localhost. No telemetry, no cloud, no accounts, no Docker required. Your conversations, images, and videos never leave your machine.
Upload PDFs, DOCX, or TXT files and chat with your documents. Hybrid search with confidence scores and source citations.
Talk to your AI with push-to-talk and hear responses with sentence-level text-to-speech streaming.
Give your AI a goal — it plans, searches the web, reads/writes files, and executes Python code autonomously.
Two-phase search: find URLs, then fetch and read actual page content. Real answers, not hallucinations.
Write and run Python code locally. Data analysis, file processing, automation — your AI does the work.
Read and write files on your system. Summarize documents, create reports, process data.
Generate images mid-conversation via ComfyUI. The only local agent that can create visuals on command.
Auto-saves tool results. Search past results by keyword, filter by category, export as markdown.
Safe tools auto-execute. File writes and code execution ask permission first. You stay in control.
Works with abliterated models. LU auto-fixes tool-calling templates via Ollama Modelfile. Models that other apps refuse to load work out of the box — with full agent capabilities.
| Feature | Locally Uncensored | Open WebUI | LM Studio | SillyTavern |
|---|---|---|---|---|
| AI Chat | Yes | Yes | Yes | Yes |
| Image Generation | Yes | No | No | No |
| Video Generation | Yes | No | No | No |
| Uncensored by Default | Yes | No | No | Partial |
| One-Click Setup | Yes | Docker | Yes | Node.js |
| Built-in Personas | 25+ | No | No | Manual |
| Open Source | MIT | MIT | No | AGPL |
| No Docker Required | Yes | No | Yes | Yes |
| Document Chat (RAG) | Yes | Yes | No | No |
| Voice (STT + TTS) | Yes | Partial | No | No |
| AI Agents + Tool Calling | 6 Tools | No | No | No |
| Agent Memory System | Yes | No | No | No |
| Abliterated Model Support | Auto-fix | No | No | Partial |
Run git clone and setup.bat (or setup.sh on Linux/macOS). The script checks for Node.js and Ollama, installs them if missing.
The setup script downloads a recommended uncensored/abliterated model (~5.7 GB). The app launches in your browser at localhost:5173.
Start chatting immediately. For image and video generation, click "Install ComfyUI" in the Create tab — one click, fully automated.
Latest & smartest open model. 6-18 GB VRAM. Exceptional reasoning, coding, and multilingual support.
Fastest all-rounder. 6 GB VRAM. Uncensored, reliable, perfect entry point for any hardware.
Chain-of-thought reasoning. Shows its thinking process. 6-48 GB VRAM. Scales to your hardware.
Best text-to-image. 8-10 GB VRAM. Incredible prompt following, detail, and coherence.
Next-gen image model. 8 GB VRAM. Fastest FLUX architecture with stunning quality.
Top photorealistic SDXL checkpoint. 6 GB VRAM. Perfect for portraits and realistic scenes.
Best text-to-video. 8-12+ GB VRAM. Lightweight 1.3B for speed, 14B for cinema quality.
Tencent's video model. 12+ GB VRAM. Excellent temporal consistency and visual quality.
Lightricks' latest. 12+ GB VRAM. Fast inference, high quality text-to-video generation.
Locally Uncensored is a free, open-source desktop app that lets you run uncensored AI locally on your own machine. It combines AI chat (via Ollama), image generation (via ComfyUI with FLUX, SDXL, and more), and video generation (via Wan 2.1, HunyuanVideo, LTX Video) in one interface. No cloud, no subscriptions, no data collection. Everything is MIT licensed.
Yes. After the initial setup and model download, no internet connection is needed. Your conversations, generated images, and videos never leave your machine. There are no accounts, no telemetry, and no usage limits. The MIT license means you can use, modify, and distribute it freely.
Those tools only handle text chat. Locally Uncensored is the only local AI app that combines chat, image generation, AND video generation in one interface. It ships with 25+ built-in personas, uses uncensored/abliterated models by default, and wraps ComfyUI's complexity behind a simple UI — no node graphs required.
For text chat: any modern computer with 8 GB RAM. For image generation: NVIDIA GPU with 8+ GB VRAM (GTX 1080 or better). For video generation: 10-12 GB VRAM recommended. The app auto-detects your hardware and recommends appropriate models. Works on Windows, macOS, and Linux.
Locally Uncensored uses abliterated AI models — models where artificial content restrictions have been removed. This means the AI responds honestly to any question without refusing or adding disclaimers. Combined with running locally, your conversations are completely private and unrestricted.
Complete comparison of GPT4All, Open WebUI, LM Studio, Jan, Kobold.cpp, SillyTavern, text-generation-webui, and Locally Uncensored.
A complete guide to running AI locally without restrictions. Setup, models, and why local beats cloud.
All-in-one AI creative suite vs the most popular local chatbot with document RAG.
Lightweight Tauri app with image/video gen vs polished Electron chat client with cloud API support.
Both are MIT-licensed Ollama frontends. Only one combines chat, image gen, and video generation.
Open source all-in-one vs polished closed-source chat client.
Both run uncensored AI locally. One is built for roleplay, the other for everything else.
Free, open source, and yours to keep. No sign-up required.