LocalForge AILocalForge AI
BlogFAQ

How to Run Pony Diffusion Locally (Full Guide)

Install Stable Diffusion WebUI Forge, download Pony Diffusion V6 XL (~6.94 GB) into models/Stable-diffusion/, set CLIP skip to 2, and use the full score_9, score_8_up, … quality string. You'll need an NVIDIA GPU with 8 GB+ VRAM for comfortable 1024² generation. The whole process takes about 30–45 minutes on a decent connection.


The Models

1. 1. Check Your Hardware

Top Pick

NVIDIA + Python 3.10 + Git + latest drivers. Miss any and Forge won't start.

Architecture: NVIDIA GPU + Windows · VRAM: 8 GB+ recommended · Best for: Confirming you can run SDXL-class models

View on CivitAI →

2. 2. Install Forge

One-click package or git clone. First run downloads dependencies — wait 20–30 min.

Architecture: WebUI (Gradio) · VRAM: 6–12 GB+ (model-dependent) · Best for: Easiest first local UI with Pony support

View on CivitAI →

3. 3. Download Pony V6 XL

~6.94 GB safetensors → models/Stable-diffusion/. Refresh the dropdown.

Architecture: SDXL fine-tune (Pony) · VRAM: ~8 GB+ at 1024² · Best for: The checkpoint with 840k+ downloads and the full LoRA ecosystem

View on CivitAI →

4. 4. Set CLIP Skip to 2

Non-negotiable. Author says skip this and you get 'low quality blobs.'

Architecture: Model setting · VRAM: N/A · Best for: Preventing blurry/blobby output on Pony V6

View on CivitAI →

5. 5. Write Your First Prompt

Full score string + rating tag + source tag + subject. Don't use masterpiece/HD.

Architecture: Pony V6 tag system · VRAM: N/A · Best for: Getting a real image out — not debugging prompt syntax

View on CivitAI →

6. 6. Fix Common Problems

Wrong CLIP skip, wrong tags, or not enough VRAM — that's 80% of Pony issues.

Architecture: Troubleshooting · VRAM: 6 GB: --medvram flag · Best for: OOM errors, blobby output, missing models

View on CivitAI →

7. 7. V7 Footnote

Different stack, weaker tags, V7.1 planned. Start with V6 first.

Architecture: AuraFlow (not SDXL) · VRAM: Verify locally · Best for: Future experiment after V6 works

View on CivitAI →

Why Pony Needs Its Own Guide

Pony Diffusion V6 XL is an SDXL fine-tune, so the install path is the same as any SDXL model — Forge, ComfyUI, or A1111. But the settings are not. If you skip CLIP skip 2, you'll get blobs. If you type masterpiece, best quality instead of the score_9 string, you're prompting a model that doesn't understand those words. This guide covers the Pony-specific settings that generic SD install tutorials miss.

The Steps

1. Check Your Hardware

You need an NVIDIA GPU, 8 GB+ VRAM, and ~20 GB of free disk space.

Requirement Minimum Recommended
GPU NVIDIA, 6 GB VRAM (tight) NVIDIA, 8–12 GB VRAM
RAM 16 GB 32 GB
Disk 20 GB free (SSD preferred) 50 GB free
OS Windows 10/11 64-bit Same

Install the latest NVIDIA drivers from nvidia.com. You also need Python 3.10.x (check "Add to PATH" during install) and Git for Windows. If any of these are missing, Forge won't launch.


2. Install Stable Diffusion WebUI Forge

Forge is the recommended UI — it's faster than A1111 on 6 GB GPUs and actively maintained.

Method Time Best For
One-click package ~10–15 min Beginners who don't want to touch Git
Git clone ~20–30 min Devs who want easier updates

One-click: Download the latest release from Forge on GitHub. Extract, run update.bat, then run.bat. The package includes Git and Python — no separate install needed.

Git clone: Open a terminal where you want the project, run git clone https://github.com/lllyasviel/stable-diffusion-webui-forge.git, then run webui-user.bat. First launch takes 20–30+ minutes while it downloads PyTorch and other dependencies. That's normal.

ComfyUI alternative: If you want node-based control, grab the portable build from ComfyUI's GitHub. Steeper learning curve, more power. Everything below still applies — just put files in ComfyUI's folder structure instead.


3. Download Pony V6 XL

Grab the "V6 (start with this one)" build — it's the checkpoint with 840,997 downloads and the full LoRA ecosystem.

File Size Where to Put It
Pony V6 XL safetensors ~6.94 GB models/Stable-diffusion/
VAE (optional) ~335 MB models/VAE/

Download from Civitai or the Hugging Face mirror. Drop the .safetensors file into your Forge models/Stable-diffusion/ folder. If colors look washed out later, add vae-ft-mse-840000-ema-pruned.safetensors to models/VAE/ — but try without it first.

After copying, click Refresh in the Forge checkpoint dropdown. Select the Pony file.

If you want Pony V6 running without assembling Python, Git, and CUDA by hand, LocalForge AI ships with the folder layout pre-wired — you still pick models and run offline.


4. Set CLIP Skip to 2

This is non-negotiable. Without CLIP skip 2, Pony V6 produces low-quality output.

In Forge: go to Settings → search for Clip skip → set to 2 → click Apply settings. In ComfyUI, use a CLIP loader node with stop_at_clip_layer: -2. The model author states this explicitly on the Civitai card — skip this and you'll wonder why everything looks wrong.


5. Write Your First Prompt

Use the full score string — score_9 alone is much weaker than the full chain.

Here's the template:

score_9, score_8_up, score_7_up, score_6_up, source_anime, rating_safe, 1girl, smile, blue eyes, park background
  • Quality tags: score_9, score_8_up, score_7_up, score_6_up (add score_5_up, score_4_up if you want)
  • Source tag: source_anime, source_pony, source_furry, or source_cartoon — picks the style
  • Rating tag: rating_safe for SFW, rating_explicit for NSFW, rating_questionable for middle ground
  • Subject tags: Describe what you want after the system tags

Set Euler a, 25 steps, CFG 7, and a resolution like 1024×1024 or 832×1216. Hit Generate.

Don't use masterpiece, best quality, or HD — those are for non-Pony models. Don't paste negative prompt walls from Discord — Pony V6 is designed to work without them.


6. Fix Common Problems

Most Pony issues come from three mistakes: wrong CLIP skip, wrong prompt tags, or not enough VRAM.

  • OOM (out of memory) on 6 GB: Add --medvram to your launch args in webui-user.bat. Costs ~20–30% speed but cuts VRAM usage by 40–50%. Also try lowering resolution to 832×832.
  • OOM on 8 GB with hi-res fix: Drop hi-res fix first. Add --xformers to launch args.
  • Blurry/blobby output: Check CLIP skip — it must be 2.
  • Tags don't seem to work: Make sure you're using the full score_9, score_8_up, … string, not just score_9.
  • Model doesn't show up: Wrong folder. It belongs in models/Stable-diffusion/, not models/checkpoints/ or anywhere else. Hit Refresh.
  • Washed-out colors: Try adding a VAE file to models/VAE/ and selecting it in the UI.

7. What About Pony V7?

V7 is a different model family (AuraFlow, not SDXL). Don't mix V6 settings with V7.

Pony V7 base exists on Civitai and Hugging Face with GGUF quantized weights for lower VRAM. But the tooling is newer, quality tags are weaker, and V7.1 is still planned. Start with V6, prove your workflow, then experiment with V7 as a separate install. See Best Pony Diffusion Models for the full V6 vs V7 comparison.


Quick Comparison

UI Install Time Best For Pony-Specific Notes
Forge 10–30 min Beginners + extensions CLIP skip in Settings → easy
ComfyUI 15–30 min Power users + workflows CLIP skip via node parameter
A1111 20–40 min Legacy — Forge is better Skip this; Forge replaces it

What to Do Next

  • Pick NSFW modelsPony Diffusion NSFW models covers rating_explicit workflows, NSFW LoRAs, and which merges work best for explicit content.
  • Compare all Pony modelsBest Pony Diffusion models ranks V6, V7, AutismMix, and Illustrious-lane picks with download stats.
  • Broaden to all SD NSFWBest NSFW Stable Diffusion models if you want SD 1.5 photoreal or non-Pony SDXL options.
  • Try a managed stackLocalForge AI handles the install plumbing so you start generating instead of debugging Python paths.

What to Do Next

FAQ

What VRAM do I need for Pony Diffusion V6 XL? +
Plan for 8 GB+ VRAM for comfortable 1024² generation. The safetensors file is about 6.94 GB, and FP16 inference adds overhead. 6 GB works with the --medvram flag but you'll lose 20–30% speed.
Why does my Pony V6 output look blurry or blobby? +
CLIP skip is probably wrong. Pony V6 requires CLIP skip 2 — the author says this explicitly on the model card. Check Settings in Forge and set it to 2.
Should I use masterpiece or best quality tags with Pony? +
No. Those tags are for non-Pony models like DAMN or general SD checkpoints. Pony V6 uses its own quality system: score_9, score_8_up, score_7_up, and so on. Using the wrong tags won't crash anything but you'll get weaker results.
Can I run Pony on ComfyUI instead of Forge? +
Yes. Drop the safetensors into ComfyUI's checkpoints folder, set stop_at_clip_layer to -2 in the CLIP loader node, and use the same prompt tags. The model doesn't care which UI runs it.
Do I need to disable a safety filter for NSFW? +
No. Pony V6 was trained on explicit content natively. Add rating_explicit to your prompt and the model generates NSFW output. There's no built-in censor to bypass when running locally.
Is Pony V7 better than V6? +
V7 runs on AuraFlow (not SDXL), supports up to 1536×1536, and has stronger backgrounds. But quality tags are weaker, the LoRA ecosystem is tiny, and V7.1 is still coming. Start with V6 — it has 840k+ downloads and years of community support.