Browse Models
The simplest way to self-host ControlNet SD 1.5 Open Pose. Launch a dedicated cloud GPU server running Lab Station OS to download and serve the model using any compatible app or framework.
Download model weights for local inference. Must be used with a compatible app, notebook, or codebase. May run slowly, or not work at all, depending on your system resources, particularly GPU(s) and available VRAM.
ControlNet SD 1.5 Open Pose enables pose-guided image generation by detecting skeletal keypoints in input images. Available in basic (body only) and full (body, hands, face) variants, it uses Low-Rank Adaptation to reduce model size while maintaining functionality. Features smart resampling for consistent results across resolutions.
ControlNet SD 1.5 Open Pose is a neural network designed to add conditional control to Stable Diffusion 1.5's text-to-image generation capabilities. Based on the architecture detailed in Adding Conditional Control to Text-to-Image Diffusion Models, the model specializes in replicating subject poses from input images into newly generated images while allowing for style and color transfer.
The model is part of the ControlNet 1.1 release, which maintains architectural consistency with ControlNet 1.0 while offering significant improvements in robustness and result quality. The technology leverages preprocessors to extract pose information, creating a "skeleton" of key body points that guides the Stable Diffusion generation process.
Multiple variants of the OpenPose model exist, each offering different levels of detail in pose detection:
openpose
: Focuses on basic body pose detectionopenpose_face
: Includes facial feature detectionopenpose_faceonly
: Specializes in facial feature detectionopenpose_full
: Combines body pose, hand, and facial feature detection for comprehensive pose informationThe ControlNet 1.1 developers recommend using primarily two options for optimal results:
A notable advancement in version 1.1 is the improved implementation of the OpenPose algorithm, resulting in more accurate keypoint detection, particularly for hand poses. The model has also benefited from cleaned training data, addressing issues present in version 1.0 such as duplicate images and low-quality data.
The model can be implemented through various interfaces, with official support for combining multiple ControlNets limited to the Automatic1111 Stable Diffusion web UI. The extension includes a smart resampling algorithm that maintains pixel-perfect control images regardless of resolution changes, distinguishing it from other implementations.
For systems with varying VRAM capabilities:
--medvram-sdxl
command-line flag--lowvram
flagThe model can be deployed in two directory locations:
stable-diffusion-webui\extensions\sd-webui-controlnet\models
stable-diffusion-webui\models\ControlNet
A more efficient Low-Rank Adaptation (LoRA) version has been developed by Stability AI, significantly reducing the model size: