by segmind
Open source · 7k downloads · 87 likes
The tiny-sd model is an optimized and lightweight version of the Stable Diffusion model, designed to generate images from text more quickly and efficiently. Distilled from Realistic Vision V4.0, it leverages an enhanced subset of artistic data to produce realistic and detailed results. Its key capabilities include generating high-resolution images (512x512) with reduced latency, up to 80% faster than base models, while maintaining comparable visual quality. This model is particularly well-suited for applications requiring real-time or large-scale image generation, such as creative tools, prototypes, or automated pipelines. What sets it apart is its balance between performance and accessibility, offering a lightweight alternative without sacrificing result fidelity.
license: creativeml-openrail-m base_model: SG161222/Realistic_Vision_V4.0 datasets:
This pipeline was distilled from SG161222/Realistic_Vision_V4.0 on a Subset of recastai/LAION-art-EN-improved-captions dataset. Below are some example images generated with the tiny-sd model.

This Pipeline is based upon the paper. Training Code can be found here.
You can use the pipeline like so:
from diffusers import DiffusionPipeline
import torch
pipeline = DiffusionPipeline.from_pretrained("segmind/tiny-sd", torch_dtype=torch.float16)
prompt = "Portrait of a pretty girl"
image = pipeline(prompt).images[0]
image.save("my_image.png")
These are the key hyperparameters used during training:
We have observed that the distilled models are upto 80% faster than the Base SD1.5 Models. Below is a comparision on an A100 80GB.

Here is the code for benchmarking the speeds.