How to Run Pony Diffusion Locally (Full Guide)
Install Stable Diffusion WebUI Forge, download Pony Diffusion V6 XL (~6.94 GB) into models/Stable-diffusion/, set CLIP skip to 2, and use the full score_9, score_8_up, … quality string. You'll need an NVIDIA GPU with 8 GB+ VRAM for comfortable 1024² generation. The whole process takes about 30–45 minutes on a decent connection.
The Models
1. 1. Check Your Hardware
Top PickNVIDIA + Python 3.10 + Git + latest drivers. Miss any and Forge won't start.
Architecture: NVIDIA GPU + Windows · VRAM: 8 GB+ recommended · Best for: Confirming you can run SDXL-class models
View on CivitAI →2. 2. Install Forge
One-click package or git clone. First run downloads dependencies — wait 20–30 min.
Architecture: WebUI (Gradio) · VRAM: 6–12 GB+ (model-dependent) · Best for: Easiest first local UI with Pony support
View on CivitAI →3. 3. Download Pony V6 XL
~6.94 GB safetensors → models/Stable-diffusion/. Refresh the dropdown.
Architecture: SDXL fine-tune (Pony) · VRAM: ~8 GB+ at 1024² · Best for: The checkpoint with 840k+ downloads and the full LoRA ecosystem
View on CivitAI →4. 4. Set CLIP Skip to 2
Non-negotiable. Author says skip this and you get 'low quality blobs.'
Architecture: Model setting · VRAM: N/A · Best for: Preventing blurry/blobby output on Pony V6
View on CivitAI →5. 5. Write Your First Prompt
Full score string + rating tag + source tag + subject. Don't use masterpiece/HD.
Architecture: Pony V6 tag system · VRAM: N/A · Best for: Getting a real image out — not debugging prompt syntax
View on CivitAI →6. 6. Fix Common Problems
Wrong CLIP skip, wrong tags, or not enough VRAM — that's 80% of Pony issues.
Architecture: Troubleshooting · VRAM: 6 GB: --medvram flag · Best for: OOM errors, blobby output, missing models
View on CivitAI →7. 7. V7 Footnote
Different stack, weaker tags, V7.1 planned. Start with V6 first.
Architecture: AuraFlow (not SDXL) · VRAM: Verify locally · Best for: Future experiment after V6 works
View on CivitAI →Why Pony Needs Its Own Guide
Pony Diffusion V6 XL is an SDXL fine-tune, so the install path is the same as any SDXL model — Forge, ComfyUI, or A1111. But the settings are not. If you skip CLIP skip 2, you'll get blobs. If you type masterpiece, best quality instead of the score_9 string, you're prompting a model that doesn't understand those words. This guide covers the Pony-specific settings that generic SD install tutorials miss.
The Steps
1. Check Your Hardware
You need an NVIDIA GPU, 8 GB+ VRAM, and ~20 GB of free disk space.
| Requirement | Minimum | Recommended |
|---|---|---|
| GPU | NVIDIA, 6 GB VRAM (tight) | NVIDIA, 8–12 GB VRAM |
| RAM | 16 GB | 32 GB |
| Disk | 20 GB free (SSD preferred) | 50 GB free |
| OS | Windows 10/11 64-bit | Same |
Install the latest NVIDIA drivers from nvidia.com. You also need Python 3.10.x (check "Add to PATH" during install) and Git for Windows. If any of these are missing, Forge won't launch.
2. Install Stable Diffusion WebUI Forge
Forge is the recommended UI — it's faster than A1111 on 6 GB GPUs and actively maintained.
| Method | Time | Best For |
|---|---|---|
| One-click package | ~10–15 min | Beginners who don't want to touch Git |
| Git clone | ~20–30 min | Devs who want easier updates |
One-click: Download the latest release from Forge on GitHub. Extract, run update.bat, then run.bat. The package includes Git and Python — no separate install needed.
Git clone: Open a terminal where you want the project, run git clone https://github.com/lllyasviel/stable-diffusion-webui-forge.git, then run webui-user.bat. First launch takes 20–30+ minutes while it downloads PyTorch and other dependencies. That's normal.
ComfyUI alternative: If you want node-based control, grab the portable build from ComfyUI's GitHub. Steeper learning curve, more power. Everything below still applies — just put files in ComfyUI's folder structure instead.
3. Download Pony V6 XL
Grab the "V6 (start with this one)" build — it's the checkpoint with 840,997 downloads and the full LoRA ecosystem.
| File | Size | Where to Put It |
|---|---|---|
| Pony V6 XL safetensors | ~6.94 GB | models/Stable-diffusion/ |
| VAE (optional) | ~335 MB | models/VAE/ |
Download from Civitai or the Hugging Face mirror. Drop the .safetensors file into your Forge models/Stable-diffusion/ folder. If colors look washed out later, add vae-ft-mse-840000-ema-pruned.safetensors to models/VAE/ — but try without it first.
After copying, click Refresh in the Forge checkpoint dropdown. Select the Pony file.
If you want Pony V6 running without assembling Python, Git, and CUDA by hand, LocalForge AI ships with the folder layout pre-wired — you still pick models and run offline.
4. Set CLIP Skip to 2
This is non-negotiable. Without CLIP skip 2, Pony V6 produces low-quality output.
In Forge: go to Settings → search for Clip skip → set to 2 → click Apply settings. In ComfyUI, use a CLIP loader node with stop_at_clip_layer: -2. The model author states this explicitly on the Civitai card — skip this and you'll wonder why everything looks wrong.
5. Write Your First Prompt
Use the full score string — score_9 alone is much weaker than the full chain.
Here's the template:
score_9, score_8_up, score_7_up, score_6_up, source_anime, rating_safe, 1girl, smile, blue eyes, park background
- Quality tags:
score_9, score_8_up, score_7_up, score_6_up(addscore_5_up, score_4_upif you want) - Source tag:
source_anime,source_pony,source_furry, orsource_cartoon— picks the style - Rating tag:
rating_safefor SFW,rating_explicitfor NSFW,rating_questionablefor middle ground - Subject tags: Describe what you want after the system tags
Set Euler a, 25 steps, CFG 7, and a resolution like 1024×1024 or 832×1216. Hit Generate.
Don't use masterpiece, best quality, or HD — those are for non-Pony models. Don't paste negative prompt walls from Discord — Pony V6 is designed to work without them.
6. Fix Common Problems
Most Pony issues come from three mistakes: wrong CLIP skip, wrong prompt tags, or not enough VRAM.
- OOM (out of memory) on 6 GB: Add
--medvramto your launch args inwebui-user.bat. Costs ~20–30% speed but cuts VRAM usage by 40–50%. Also try lowering resolution to 832×832. - OOM on 8 GB with hi-res fix: Drop hi-res fix first. Add
--xformersto launch args. - Blurry/blobby output: Check CLIP skip — it must be 2.
- Tags don't seem to work: Make sure you're using the full
score_9, score_8_up, …string, not justscore_9. - Model doesn't show up: Wrong folder. It belongs in
models/Stable-diffusion/, notmodels/checkpoints/or anywhere else. Hit Refresh. - Washed-out colors: Try adding a VAE file to
models/VAE/and selecting it in the UI.
7. What About Pony V7?
V7 is a different model family (AuraFlow, not SDXL). Don't mix V6 settings with V7.
Pony V7 base exists on Civitai and Hugging Face with GGUF quantized weights for lower VRAM. But the tooling is newer, quality tags are weaker, and V7.1 is still planned. Start with V6, prove your workflow, then experiment with V7 as a separate install. See Best Pony Diffusion Models for the full V6 vs V7 comparison.
Quick Comparison
| UI | Install Time | Best For | Pony-Specific Notes |
|---|---|---|---|
| Forge | 10–30 min | Beginners + extensions | CLIP skip in Settings → easy |
| ComfyUI | 15–30 min | Power users + workflows | CLIP skip via node parameter |
| A1111 | 20–40 min | Legacy — Forge is better | Skip this; Forge replaces it |
What to Do Next
- Pick NSFW models — Pony Diffusion NSFW models covers
rating_explicitworkflows, NSFW LoRAs, and which merges work best for explicit content. - Compare all Pony models — Best Pony Diffusion models ranks V6, V7, AutismMix, and Illustrious-lane picks with download stats.
- Broaden to all SD NSFW — Best NSFW Stable Diffusion models if you want SD 1.5 photoreal or non-Pony SDXL options.
- Try a managed stack — LocalForge AI handles the install plumbing so you start generating instead of debugging Python paths.
What to Do Next
Pick NSFW models
rating_explicit workflows, NSFW LoRAs, and uncensored merge picks.
Compare all Pony models
V6 vs V7, AutismMix, Gehenna, DAMN — ranked with Civitai stats.
Broader SD NSFW list
SD 1.5 photoreal + non-Pony SDXL picks.
Managed local stack
LocalForge AI — skip Python/Git setup, start generating.
