From download to your first AI chat, image, and video in under 10 minutes.
| Operating System | Windows 10 or later (64-bit) |
| RAM | 8 GB minimum, 16 GB recommended |
| GPU (for chat) | Optional. CPU inference works, GPU is faster. |
| GPU (for images) | NVIDIA GPU with 6+ GB VRAM recommended |
| GPU (for video) | NVIDIA GPU with 8-12+ GB VRAM recommended |
| Disk Space | 2 GB for the app. 5-50 GB for models depending on selection. |
Download the latest release from the Releases page or from locallyuncensored.com. The file is called Locally Uncensored_x.x.x_x64-setup.exe.
Double-click the downloaded file. Windows may show a SmartScreen warning because the app is not code-signed. Click "More info", then "Run anyway". The installer completes in a few seconds.
After installation, the app launches automatically. You will see the onboarding wizard that guides you through the initial setup.
On first launch, a setup wizard walks you through five steps. Each step takes less than a minute.
Click "Get Started" to begin.
Choose Light or Dark mode. You can change this later in Settings.
The app scans your system for local AI backends. If Ollama is installed and running, it detects it automatically. If no backend is found, click "Install Ollama" to download and install it directly from the app with a progress bar.
If you want image and video generation, the app detects or installs ComfyUI automatically. If ComfyUI is already installed on your system, the app finds it and shows the path. If not, click "Install ComfyUI" for a one-click installation with step-by-step progress.
Choose which AI models to download. Models are split into "Uncensored" and "Mainstream" tabs. Each card shows the model name, a short description, file size, and VRAM requirement. Select the ones you want and click "Download Selected". You can always add more models later from the Model Manager.
After setup, you see "You're all set!" and the app takes you to the main interface.
Click "+ New Chat" in the bottom left corner of the sidebar. The main area shows an empty chat with a message input at the bottom.
Type your message in the input field and press Enter. The AI responds in real time with streaming text.
Click the model name in the top header bar (e.g. "phi4") to open the model dropdown. Select a different model to switch. Each conversation remembers which model it uses.
Click the "Think" toggle in the message input to enable chain-of-thought reasoning. The AI shows its reasoning process before giving the final answer. Works with compatible models like QwQ, DeepSeek-R1, and Qwen 3.
Toggle "Agent" in the top right to enable tool-calling capabilities. The AI can search the web, read and write files, execute code, and take screenshots. Requires a compatible model (Hermes 3, Qwen 3, Gemma 4).
Click the paintbrush icon in the top header bar or navigate to Create in the sidebar.
If you have not installed an image model yet, go to the Model Manager (grid icon in the header), switch to the "Discover" tab, and look under "Image" models. Click "Install" on a model bundle. Recommended starting models:
Back in the Create view, select your installed model from the dropdown, type a prompt describing what you want, and click "Generate". The first generation takes longer because the model needs to load into VRAM. Subsequent generations are faster.
Go to the Model Manager, switch to "Discover", and look under "Video" models. The VRAM tier tabs (All / Lightweight / Mid-Range / High-End) help you find models that match your GPU. Recommended:
In the Create view, switch to the "Video" tab. Select your installed model, type a prompt, and click "Generate". Video generation takes 2-10 minutes depending on the model, resolution, and your hardware.
Click the gear icon in the top right corner to open Settings.
Manage your AI backends. Ollama is enabled by default. You can add cloud providers (OpenAI, Anthropic, Groq) or additional local backends (LM Studio, vLLM, KoboldCpp). Click "+ Add Provider" to see all 20+ presets.
View ComfyUI status, start/stop/restart it, and configure the path and port. If you use the ComfyUI Desktop App instead of the browser-based version, you may need to change the port here to match your setup.
Adjust temperature, top-p, top-k, and max tokens for chat responses.
ollama serve.