LocalForge AILocalForge AI
BlogFAQ
← Back to Blog

ControlNet for Stable Diffusion: Precise Image Control on Your Local Machine

Complete guide to setting up and using ControlNet with local Stable Diffusion. Control poses, edges, depth, and composition in your uncensored AI image generation — all offline.

Why ControlNet Changes Everything

Text prompts are powerful, but they're terrible at precise composition. Try getting a specific pose, a hand position, or a room layout from text alone — it's a slot machine.

ControlNet solves this. Feed it a reference image — a pose skeleton, an edge map, a depth map — and Stable Diffusion follows the structure while applying your prompt's style and subject. Locally, with zero restrictions.

ControlNet Types Explained

Type Input Best For
OpenPoseStick figure poseControlling body pose, hand positions
CannyEdge detection mapPreserving outlines and shapes
DepthDepth mapScene layout, foreground/background separation
LineartClean line drawingAnime/illustration coloring, sketch-to-image
ScribbleRough sketchQuick concept art from rough drawings
IP-AdapterReference photoStyle transfer, face consistency
TileLow-res imageUpscaling with detail regeneration

Setup in Forge UI

  1. ControlNet is built into Forge — no extension install needed
  2. Download ControlNet models from Hugging Face — get the ones matching your base model (SDXL or SD 1.5)
  3. Place models in models/ControlNet/
  4. In Forge UI, expand the ControlNet panel below the prompt box
  5. Upload your reference image, select the preprocessor (e.g., "openpose"), select the model, hit Generate

Key Settings

  • Control Weight: How strongly the reference image influences output. Start at 0.7–1.0. Lower values = more creative freedom.
  • Starting/Ending Step: When ControlNet activates during generation. Default 0–1 (full range). Setting end to 0.5 lets the model add its own details in later steps.
  • Preprocessor: Extracts the control signal from your image. Use "none" if you're providing a pre-processed map (e.g., a hand-drawn pose skeleton).
  • Multiple ControlNets: You can stack them — e.g., OpenPose for body + Depth for scene layout. Forge supports up to 3 simultaneously.

Honest warning: ControlNet setup is where most beginners quit. Downloading the right models for the right architecture, matching preprocessors, resolving version conflicts — it takes 1–2 hours even for experienced users. LocalForge AI ships with ControlNet and Forge UI pre-configured. Upload a reference image and go.

Practical Workflows

Consistent Character Poses

Use OpenPose with a reference photo of the pose you want. Combine with a character LoRA for a consistent face across multiple poses. This is how people create consistent character series locally.

Sketch to Finished Art

Draw a rough sketch on paper or in MS Paint. Use the Scribble preprocessor. ControlNet turns your rough lines into a fully rendered image matching your prompt's style.

Photo-to-Photo Style Transfer

Use Canny or Depth on a real photo, then prompt for a different style: "oil painting", "cyberpunk", "anime". The composition stays the same, but the aesthetic transforms.

FAQ

Does ControlNet work with Flux models?

As of early 2026, ControlNet support for Flux is experimental and limited to a few types (Canny, Depth) via ComfyUI custom nodes. SDXL has the most mature ControlNet support.

Can I use ControlNet for fixing hands?

Yes — use OpenPose or Depth focused on just the hand area with inpainting. Or use ADetailer which automates hand/face fixes. Both run locally in Forge.