Launch a dedicated cloud GPU server running Laboratory OS to download and run ControlNet SDXL Recolor using any compatible app or framework.
Direct Download
Must be used with a compatible app, notebook, or codebase. May run slowly, or not work at all, depending on local system resources, particularly GPU(s) and available VRAM.
Forge is a platform built on top of Stable Diffusion WebUI to make development easier, optimize resource management, speed up inference, and study experimental features.
Train your own LoRAs and finetunes for Stable Diffusion and Flux using this popular GUI for the Kohya trainers.
Model Report
stabilityai / ControlNet SDXL Recolor
ControlNet SDXL Recolor is a Control-LoRA extension for Stable Diffusion XL that specializes in adding color to grayscale photographs and hand-drawn sketches. Built on the ControlNet 1.1 architecture, it uses low-rank adaptation for parameter-efficient fine-tuning, reducing model size while maintaining performance. The model supports both photo restoration and sketch colorization modes, integrating with SDXL workflows and compatible with multi-ControlNet compositions for complex image manipulation tasks.
Explore the Future of AI
Your server, your data, under your control
ControlNet SDXL Recolor is a generative AI model designed for image colorization, functioning as a Control-LoRA extension for the Stable Diffusion XL (SDXL) architecture. This model specializes in adding color to grayscale photographs and hand-drawn sketches, supporting production applications within the broader ControlNet 1.1 model suite. Developed with a focus on efficiency and versatility, ControlNet SDXL Recolor enables users to transform monochrome media into colored outputs. The model implements parameter-efficient fine-tuning strategies, making it suitable for various computational environments and integration with other ControlNet models.
Sample outputs of ControlNet SDXL Recolor: colorized versions of both black and white photographs and sketches demonstrate the model's dual functionality for restoration and creative coloring tasks.
ControlNet SDXL Recolor is built upon the ControlNet 1.1 architecture, inheriting the modular framework designed for guided image synthesis. Control-LoRA models utilize low-rank adaptation (LoRA), which introduces parameter-efficient fine-tuning by injecting small trainable matrices into the original architecture. This approach reduces the file size compared to full ControlNet checkpoints, decreasing typical model storage from 4.7GB to as little as 738MB or 377MB, depending on the selected rank, while maintaining performance characteristics consistent with the original model.
At its core, ControlNet overlays conditioning networks onto the image generation pipeline of SDXL, allowing for fine-grained control based on input hints such as edges, depth maps, or, in the case of Recolor, grayscale tonal distribution and sketch outlines. A crucial implementation detail involves the use of global average pooling to aggregate features before merging them into the SDXL UNet layers. This enables efficient modulation of the generation process, ensuring colorization is applied only on the conditional side of the classifier-free guidance (Cfg) scale. These technical improvements foster the model's integration with SDXL, resulting in context-aware colorization from varied forms of monochrome input. Details of the architecture and LoRA methodology are available in the ControlNet 1.1 documentation and Hugging Face release notes.
Colorization Capabilities and Use Cases
ControlNet SDXL Recolor is configured for two principal colorization tasks. First, in “Recolor” mode, it restores and enhances black and white photographs, simulating naturalistic or creatively stylized color palettes. Second, its “Sketch” mode targets white-on-black drawings—either hand-drawn or generated with edge detection methods such as the pidi edge model—providing a framework for artists to apply color to line art.
Both modes employ learned mappings from tonal cues to plausible color distributions, leveraging the SDXL backbone’s expressivity. The model can be deployed for photo restoration, archival enhancement, creative illustration, and preprocessing in digital workflows where coloring is desired. Its compatibility with other ControlNets further expands its utility, facilitating combined operations such as joint colorization and inpainting, or mixing with depth or posture controls for complex image manipulations. General model usage and multi-ControlNet workflows are supported as described in the ControlNet project documentation and the Automatic1111 plugin integration.
The ComfyUI interface demonstrates ControlNet SDXL Recolor in action: a grayscale portrait is transformed into a realistic, colorized image through a visual processing workflow. The prompt used is a grayscale headshot of a woman, illustrating both the input and the resulting output.
Details about the precise datasets used to train ControlNet SDXL Recolor have not been fully disclosed. However, the broader ControlNet 1.1 family emphasizes enhanced data quality and augmentation techniques. Enhancements over earlier versions include the removal of duplicate grayscale human images, reduction of low-quality or compressed images, and correction of prompts for better paired training. Augmentation strategies—such as random image flipping—help generalize model performance across diverse photographic and sketch styles. These refinements contribute to reduced artifacts in colorization outcomes, as discussed in the official ControlNet 1.1 updates.
The overall training process involves conditioning the model to learn plausible mappings between input grayscale or line art images and their corresponding colorized forms. The Control-LoRA training procedure focuses on minimizing additional parameter footprint while maintaining fidelity in generated results.
Integration and Compatibility
ControlNet SDXL Recolor is engineered for integration with SDXL workflows, benefiting from compatibility with various user interfaces such as ComfyUI and StableSwarmUI. The model also participates in the “Multi-ControlNet” ecosystem, wherein additive controls from multiple ControlNets (including community models and custom LoRAs) can be arbitrarily combined, especially in production pipelines managed through the Automatic1111 plugin.
Model deployment typically involves referencing the appropriate Control-LoRA checkpoint (by rank and mode), ensuring that associated configuration files activate recommended features such as global average pooling. Implementation instructions, compatibility details, and advanced setup recommendations are regularly maintained in the ControlNet project repositories.
Model Family and Related Models
ControlNet SDXL Recolor belongs to a suite of ControlNet 1.1 models, each designed to condition SDXL-based image generation on different input modalities. Notable models in the family include ControlNet SDXL Canny for edge-based guidance, ControlNet SDXL Depth for spatial awareness, ControlNet SDXL Openpose for pose estimation, and ControlNet SDXL IP-Adapter for image prompt adaptation. This modular approach enables coverage of image transformation tasks, with each Control-LoRA providing efficient, focused parameter updates for its target function.
The full set of ControlNet 1.1 models, including both production-ready and experimental controls, is described in the project documentation. The suite’s heterogeneous architecture and shared integration strategy facilitate advanced, multi-stage workflows in digital art, photography, and research contexts.
Limitations and Known Issues
While ControlNet SDXL Recolor and its companion models offer broad utility, they inherit several known limitations from the ControlNet 1.1 framework. Some models in the family are classified as experimental and may require manual selection of optimal outputs. Functionality such as multi-ControlNet composition or tiled upscaling is primarily supported via specific plugins and user interfaces, with standalone deployments often lacking full feature parity. Dependency on the Automatic1111 plugin is recommended for utilizing advanced orchestration features. Additionally, some related models—such as Anime Lineart—may require supplementary files not provided directly by the project.
External Resources
For further information and technical details, the following resources are available: