LocalForge AILocalForge AI
BlogFAQ

How to Run Stable Diffusion Locally (NSFW)

Stable Diffusion WebUI Forge is the easiest “first PC install” for most beginners: clone the repo, run webui-user.bat, wait for dependencies, then drop .safetensors files into the right folders. GPU: NVIDIA + current drivers is the path with the most tutorials. NSFW: nothing is “unlocked by magic” — you’re downloading community weights and running them locally, so keep it legal and age-appropriate.

The Models

1. 1) Prerequisites

Top Pick

Do drivers + PATH first — it prevents spooky errors later.

Architecture: Windows + NVIDIA + Python 3.10 + Git · VRAM: GPU-dependent · Best for: Smooth tutorial path before you touch models

View on CivitAI →

2. 2) Install Forge

Clone repo → run webui batch → wait for first-run downloads.

Architecture: WebUI fork · VRAM: 6–12 GB+ typical (depends on model and res) · Best for: First local install with a normal UI

View on CivitAI →

3. 3) Model folders

Wrong folder = invisible model — follow Civitai’s file guide.

Architecture: Filesystem layout · VRAM: N/A · Best for: Checkpoints vs LoRA vs VAE hygiene

View on CivitAI →

4. 4) First generation

Start conservative — prove stability, then expand.

Architecture: SD 1.5 / SDXL / Flux-class (later) · VRAM: Resolution × model size · Best for: Smoke test before NSFW downloads

View on CivitAI →

Why This Matters

SD means Stable Diffusion. VRAM is your GPU’s fast memory — it’s the usual bottleneck for image size and model size. CUDA is NVIDIA’s GPU compute stack; most beginner guides assume an NVIDIA card because the install path is smoothest. This page walks you from zero to a first local image, with adult-capable workflows in mind — you’re responsible for lawful use.

The Steps

1. Prerequisites (do this before you “install SD”)

Check you’re on Windows 10/11, 64-bit, with an NVIDIA GPU for the standard path.

Architecture VRAM Best For
PC + NVIDIA GPU 6 GB+ (tight) to 12 GB+ (more comfortable) Following WebUI tutorials

Install Python 3.10.x (add to PATH), Git for Windows, and the latest NVIDIA driver from NVIDIA’s site. Put the WebUI on an SSD if you can — model loads hurt on a slow disk.


2. Download Forge (WebUI)

Clone Stable Diffusion WebUI Forge — it’s the beginner-friendly WebUI we recommend.

| Architecture | VRAM | Best For | |---|---|--/| | Gradio WebUI | Same as your GPU | txt2img + extensions |

Open a terminal in the folder where you want the project, then clone the repo from GitHub (lllyasviel/stable-diffusion-webui-forge). Run webui-user.bat. First launch can take 20–30+ minutes while it downloads dependencies — that’s normal.

Forge on GitHub


3. Put models in the right folders

Think “checkpoint = base brain,” LoRA = small add-on.

File type Where it goes (Forge/A1111-style layout)
.safetensors checkpoints models/Stable-diffusion
LoRA models/Lora
VAE (optional) models/VAE
Textual inversion embeddings

After copying files, refresh the checkpoint list in the UI (or restart). CivitAI’s community article “Where do I put the files?” is the classic reference if you forget.


4. Make your first image (SFW test first)

Prove the pipeline works before you chase exotic models.

Checkpoint VRAM Best For
Any common SD 1.5 or SDXL base you can hold Match file to VRAM Smoke test

Pick a checkpoint in the UI, set resolution modestly (especially on 8 GB), then generate. If you OOM (out of memory), reduce resolution or enable --medvram / --lowvram in your launch args (trade: speed).


If you want a local-first bundle without assembling Python, Git, and CUDA by hand, LocalForge AI is built to cut setup friction — you still pick models and stay in control.


5. Troubleshooting (the boring stuff that fixes 80%)

VRAM errors are usually not “broken AI” — they’re a math problem.

  • OOM: lower resolution, close other GPU apps, try --medvram or --lowvram (A1111 wiki explains the tradeoffs).
  • Slow generation: expect that on small GPUs — seconds-per-image is not universal.
  • Wrong folder: 90% of “it doesn’t show up” is the file path.

Quick Comparison

Path What it is Good for
Forge WebUI fork Beginners + extensions
ComfyUI portable Node UI + batch files You want workflows later
Cloud-only Not local Privacy tradeoff

What to Do Next

Verdict

Install Forge, place models in the standard folders, and prove a small image before you download 20 checkpoints. If you’re stuck on environment setup, LocalForge AI is the “less time wrestling installs” lane — still local, still your responsibility.


What to Do Next

FAQ

What does SD, VRAM, and CUDA mean? +
SD is Stable Diffusion. VRAM is GPU memory — it limits how big your images and models can be. CUDA is NVIDIA’s GPU compute layer; most beginner guides assume an NVIDIA GPU because installs are easiest.
Is Forge safe to install? +
Get it from the official GitHub repository. You’re running a local app — still practice normal hygiene: antivirus, updates, and don’t download mystery checkpoints from random links.
Where do I put CivitAI downloads? +
Checkpoints go in `models/Stable-diffusion`, LoRAs in `models/Lora`, and optional VAE files in `models/VAE` for typical WebUI layouts. If a model doesn’t appear, it’s almost always a folder mistake.
Why did my GPU run out of memory? +
OOM means your resolution or model is too large for the VRAM you have. Lower resolution first, then try `--medvram` or `--lowvram` flags (trade speed).
Can I run NSFW locally without a cloud filter? +
Local generation doesn’t use a cloud “block” button — you’re choosing models on disk. You’re still responsible for lawful, age-appropriate use.
What is LocalForge AI here for? +
It’s an optional offline-first stack to reduce install time — same local concept, less manual plumbing.