Browse Models
The simplest way to self-host ControlNet SD 1.5 Segmentation. Launch a dedicated cloud GPU server running Lab Station OS to download and serve the model using any compatible app or framework.
Download model weights for local inference. Must be used with a compatible app, notebook, or codebase. May run slowly, or not work at all, depending on your system resources, particularly GPU(s) and available VRAM.
ControlNet SD 1.5 Segmentation enables controlled image generation through semantic segmentation maps, allowing users to define object boundaries and placement. It features pixel-perfect resampling for detail preservation and supports both COCO and ADE20K annotation protocols for flexible composition control.
ControlNet SD 1.5 Segmentation is a neural network model that adds conditional control to Stable Diffusion 1.5 text-to-image generation. It's part of the larger ControlNet 1.1 family, which includes fourteen different control models designed to guide image generation through various input sources. The model was developed based on research detailed in "Adding Conditional Control to Text-to-Image Diffusion Models".
The model leverages Stable Diffusion 1.5 as its base architecture and was trained on a combination of COCO and ADE20K datasets. Version 1.1 introduced significant improvements over its predecessor, notably adding support for both ADE20K and COCO annotation protocols, which enhanced the model's overall performance and versatility. The model accepts semantic segmentation maps as input, allowing precise control over object boundaries and classifications in the generated images.
A notable advancement in the ControlNet family is the Control-LoRA variant, which offers a parameter-efficient implementation. While the original ControlNet models are approximately 4.7GB in size, Control-LoRA versions reduce this significantly - Rank 256 variants are about 738MB, while Rank 128 versions are roughly 377MB, making them more accessible for consumer GPUs.
The model's primary feature is its ability to process various types of segmentation maps through multiple preprocessors:
ControlNet SD 1.5 Segmentation can be used in conjunction with other ControlNet models through the multi-ControlNet feature, allowing for complex control over image generation. The "Pixel Perfect" resampling algorithm ensures that input control images maintain their detail and resolution during the upscaling process, regardless of the target image resolution.
The model is primarily implemented through two main interfaces:
Performance may vary depending on available VRAM, and users can optimize performance using command-line flags such as --medvram-sdxl
and --lowvram
. The model can be used alongside other ControlNet variants, though compatibility testing is recommended, particularly with extensions like Deforum.