Browse Models
The simplest way to self-host Stable Diffusion 1.5. Launch a dedicated cloud GPU server running Lab Station OS to download and serve the model using any compatible app or framework.
Download model weights for local inference. Must be used with a compatible app, notebook, or codebase. May run slowly, or not work at all, depending on your system resources, particularly GPU(s) and available VRAM.
Stable Diffusion 1.5 is a text-to-image model with 983M parameters, trained on LAION-aesthetics v2 5+ dataset. It uses a CLIP text encoder, U-Net denoiser, and VAE architecture. Notable for image generation, modification, and inpainting at 512x512 resolution, with optimal results at 7-8.5 guidance scale.
Stable Diffusion 1.5, released in October 2022 by RunwayML, represents a significant advancement in text-to-image generation technology. Building upon the foundational research in Latent Diffusion Models, it introduces improvements in image quality and generation capabilities while maintaining accessibility through open-source availability.
At its core, Stable Diffusion 1.5 is a latent diffusion model with 983 million parameters. The architecture consists of three main components:
The model operates in a compressed latent space (reducing 512x512x3 images to 64x64x4 latent representations), significantly reducing computational requirements compared to pixel-space diffusion models. This efficiency allows the model to run on consumer hardware with as little as 4GB VRAM.
The model underwent extensive training on the LAION-aesthetics v2 5+ dataset, with 595,000 fine-tuning steps at 512x512 resolution. The training process involved:
Performance evaluations show improvements over previous versions (v1.1-v1.4) in terms of image quality and prompt adherence. The model performs particularly well with classifier-free guidance scales between 7.0 and 8.5, typically using 50 inference steps for optimal quality.
The model supports multiple generation modes:
Key parameters affecting generation include:
While powerful, Stable Diffusion 1.5 has several notable limitations:
The model operates under the CreativeML OpenRAIL-M license, which permits commercial use of generated images while prohibiting harmful applications. Users must comply with content restrictions and include appropriate license documentation when redistributing the model.