LocalForge AILocalForge AI
BlogFAQ
← Back to Blog

How to Run Flux Uncensored Locally in 2026 (CHROMA Setup Guide)

Step-by-step guide to running Flux uncensored on your own PC using CHROMA. Covers Forge and ComfyUI setup, model downloads, VRAM requirements, and LoRA compatibility — completely offline and private.

DIY setup time: 2-4 hours (Python, Git, model downloads, troubleshooting)
LocalForge AI: under 10 minutes

Why CHROMA, Not Base Flux?

Base Flux (dev/schnell) is "lightly" censored. It was trained on filtered data, so it struggles with anatomy, explicit content, and certain body compositions. You can force it with LoRAs, but results are inconsistent.

CHROMA is a community fork that solves this:

  • Censorship training removed
  • Anatomical accuracy improved
  • Same Flux architecture = same LoRA compatibility
  • Near-identical quality to base Flux
  • Open weights, free to download

If you want to run "Flux uncensored" locally — CHROMA is what you're looking for.

Requirements

Component Minimum Recommended
GPU VRAM 8 GB (quantized only) 12-16 GB+
RAM 16 GB 32 GB
Storage 25 GB free 50 GB+
GPU Brand NVIDIA recommended. AMD guide here.

Best cards: RTX 3060 12GB (budget), RTX 4070 Ti Super (mid-range), RTX 4090 (best). Full GPU comparison here.

Option A: Forge Setup (Easiest)

Forge is a fork of AUTOMATIC1111 with native Flux/CHROMA support. Simplest path for most users.

Step 1: Install Forge

  • Install Python 3.10.x and Git
  • Clone Forge: git clone https://github.com/lllyasviel/stable-diffusion-webui-forge
  • Run webui.bat (Windows) or webui.sh (Linux/Mac)
  • Wait for first-time setup to complete (installs dependencies)

Step 2: Download CHROMA

  • Download CHROMA model from HuggingFace (search "CHROMA flux")
  • Place the model file in models/Stable-diffusion/
  • Also download the CLIP and VAE files if not bundled

Step 3: Configure

  • Restart Forge
  • Select CHROMA from the model dropdown
  • Set resolution to 1024×1024 (Flux native resolution)
  • Use 20-30 steps, CFG 1.0 (Flux models use low CFG)
  • Generate

Common issues: CUDA out of memory → enable --medvram flag. Black images → wrong VAE. Slow first generation → model loading into VRAM (normal).

Option B: ComfyUI Setup (More Control)

ComfyUI gives you more flexibility but requires learning the node-based interface.

Step 1: Install ComfyUI

  • Download the standalone package from ComfyUI releases (no Python required)
  • Or clone from GitHub and set up manually with Python 3.10+

Step 2: Add CHROMA Model

  • Place CHROMA model in ComfyUI/models/diffusion_models/
  • Place CLIP models in ComfyUI/models/text_encoders/
  • Place VAE in ComfyUI/models/vae/

Step 3: Load Workflow

  • Find a CHROMA workflow JSON on CivitAI or the ComfyUI community
  • Drag-and-drop the JSON into ComfyUI
  • Update file paths if needed
  • Click "Queue Prompt" to generate

Not sure which UI? See our ComfyUI vs Forge comparison.

Using LoRAs With CHROMA

CHROMA is architecturally compatible with Flux, so existing Flux LoRAs work. This gives you access to a growing library of style, character, and concept LoRAs:

  • Download Flux LoRAs from CivitAI or HuggingFace
  • Place in your models/Lora/ folder
  • Apply in Forge with the LoRA tab, or wire into ComfyUI workflow
  • Start with LoRA weight 0.7-1.0 and adjust

For more on LoRAs: Complete LoRA, Checkpoint & Embeddings Guide

Low VRAM? Quantized Options

If you're on 8 GB VRAM, you can run quantized versions of CHROMA:

  • FP8: ~50% smaller, minimal quality loss. Best compromise.
  • NF4: ~75% smaller, noticeable quality reduction. For 6-8 GB cards.
  • Alternative: Use SDXL or Pony V6 instead — better experience on low VRAM. Low VRAM optimization guide

Privacy Note

When you run CHROMA locally, your prompts never leave your computer. No telemetry, no cloud calls, no logging. Unlike cloud services that store every prompt, local generation means absolute privacy.

This is especially important for uncensored content — you don't want your prompts on someone else's server. Learn more about AI privacy.

Bottom Line

Running Flux uncensored locally is absolutely possible in 2026 — CHROMA is the model you want. The manual setup takes 2-4 hours if you're comfortable with Python and Git.

If you'd rather skip the setup entirely: LocalForge AI gives you a private offline setup with everything already configured for $50, one-time. Under 10 minutes from download to generating.