Stable Diffusion
  • 👋Welcome to Stable Diffusion
  • Stable Diffusion Overview
    • 💡Technology
      • Architecture
      • Training data
      • Training procedures
      • Limitations
      • End-user fine tuning
    • âš™ī¸Capabilities
      • Text to image generation
      • Image modification
    • đŸ•šī¸Usage and controversy
    • ✨License
    • 🔗External links
  • Stable Diffusion Chain
    • đŸ“ĒToken
    • 👑Tokenomics
    • đŸ’ģIntegrated Systems
    • ⌚Stable Diffusion Chain Roadmap
    • 🐧Build 3D NFT
Powered by GitBook
On this page
  1. Stable Diffusion Overview
  2. Technology

Training data

Stable Diffusion was trained on pairs of images and captions taken from LAION-5B, a publicly available dataset derived from Common Crawl data scraped from the web, where 5 billion image-text pairs were classified based on language and filtered into separate datasets by resolution, a predicted likelihood of containing a watermark, and predicted "aesthetic" score (e.g. subjective visual quality).[15]

The dataset was created by LAION, a German non-profit which receives funding from Stability AI.[15][16] The Stable Diffusion model was trained on three subsets of LAION-5B: laion2B-en, laion-high-resolution, and laion-aesthetics v2 5+.[15] A third-party analysis of the model's training data identified that out of a smaller subset of 12 million images taken from the original wider dataset used, approximately 47% of the sample size of images came from 100 different domains, with Pinterest taking up 8.5% of the subset, followed by websites such as WordPress, Blogspot, Flickr, DeviantArt and Wikimedia Commons.[17][15]

PreviousArchitectureNextTraining procedures

Last updated 2 years ago

💡