Stable Diffusion web UI
Go to file
Travis Fletcher f6394a89ca
Fixed Nvidia Docker dependencies (#1327)
1. The update to Streamlit 1.13 has bee mirrored
2. Streamlit needs to be in Headless mode so it doesn't pester about
wanting an email on startup, which crashes in docker since there's no
stdin
3. Fixed an issue where opencv-python from the base image was not
allowing us to override it with opencv-python-headless, which is
required because of its lack of runtime DLLs (+ it's smaller)
2022-09-27 00:42:02 +03:00
.github License header 2022-09-26 14:02:48 +01:00
.idea Added gitignore 2022-08-29 00:41:17 +01:00
.streamlit Fixed Nvidia Docker dependencies (#1327) 2022-09-27 00:42:02 +03:00
configs Merge branch 'dev' into dev 2022-09-26 07:03:59 -07:00
data Repo merge (#712) 2022-09-06 23:50:14 +01:00
docs License header 2022-09-26 14:02:48 +01:00
frontend License header 2022-09-26 14:02:48 +01:00
images Add NSFW filter to avoid unexpected (#955) 2022-09-11 12:32:36 +03:00
ldm move scale multiplication (#1232) 2022-09-20 02:10:46 +03:00
models Repo merge (#712) 2022-09-06 23:50:14 +01:00
optimizedSD Repo merge (#712) 2022-09-06 23:50:14 +01:00
scripts Merge branch 'dev' into dev 2022-09-26 07:03:59 -07:00
_config.yml dark theme 2022-09-19 01:05:46 +03:00
.dockerignore Audit dependencies and revise Docker environment specification 2022-09-20 00:21:54 -07:00
.env_docker.example License header 2022-09-26 14:02:48 +01:00
.gitattributes Repo merge (#712) 2022-09-06 23:50:14 +01:00
.gitignore Fixed missing frontend/dist folder. 2022-09-20 21:45:16 -07:00
build_docker.sh License header 2022-09-26 14:02:48 +01:00
CONTRIBUTING.md add contribution guide 2022-09-19 01:05:46 +03:00
daisi_app.py Added config.toml file again. 2022-09-26 06:53:14 -07:00
docker-compose.yml License header 2022-09-26 14:02:48 +01:00
docker-reset.sh License header 2022-09-26 14:02:48 +01:00
Dockerfile Fixed Nvidia Docker dependencies (#1327) 2022-09-27 00:42:02 +03:00
entrypoint.sh License header 2022-09-26 14:02:48 +01:00
environment.yaml License header 2022-09-26 14:02:48 +01:00
LICENSE Create LICENSE 2022-08-29 12:16:23 +01:00
README.md Update README.md 2022-09-22 10:44:11 +01:00
requirements.txt Fixed Nvidia Docker dependencies (#1327) 2022-09-27 00:42:02 +03:00
setup.py refactor: sort out dependencies 2022-09-11 04:33:43 +03:00
Stable_Diffusion_v1_Model_Card.md Repo merge (#712) 2022-09-06 23:50:14 +01:00
webui-streamlit.cmd License header 2022-09-26 14:02:48 +01:00
webui.cmd License header 2022-09-26 14:02:48 +01:00
webui.sh License header 2022-09-26 14:02:48 +01:00

Web based UI for Stable Diffusion by sd-webui

Visit sd-webui's Discord Server Discord Server

Installation instructions for Windows, Linux

Want to ask a question or request a feature?

Come to our Discord Server or use Discussions.

Documentation

Documentaion is located here

Want to contribute?

Check the Contribution Guide

sd-webui is

Gradio

Features

Screenshots

Streamlit

Features

Screenshots


Stable Diffusion was made possible thanks to a collaboration with Stability AI and Runway and builds upon our previous work:

High-Resolution Image Synthesis with Latent Diffusion Models
Robin Rombach*, Andreas Blattmann*, Dominik Lorenz, Patrick Esser, Björn Ommer

CVPR '22 Oral

which is available on GitHub. PDF at arXiv. Please also visit our Project page.

Stable Diffusion is a latent text-to-image diffusion model. Thanks to a generous compute donation from Stability AI and support from LAION, we were able to train a Latent Diffusion Model on 512x512 images from a subset of the LAION-5B database. Similar to Google's Imagen, this model uses a frozen CLIP ViT-L/14 text encoder to condition the model on text prompts. With its 860M UNet and 123M text encoder, the model is relatively lightweight and runs on a GPU with at least 10GB VRAM. See this section below and the model card.

Stable Diffusion v1 refers to a specific configuration of the model architecture that uses a downsampling-factor 8 autoencoder with an 860M UNet and CLIP ViT-L/14 text encoder for the diffusion model. The model was pretrained on 256x256 images and then finetuned on 512x512 images.

*Note: Stable Diffusion v1 is a general text-to-image diffusion model and therefore mirrors biases and (mis-)conceptions that are present in its training data. Details on the training procedure and data, as well as the intended use of the model can be found in the corresponding model card.

BibTeX

@misc{rombach2021highresolution,
      title={High-Resolution Image Synthesis with Latent Diffusion Models}, 
      author={Robin Rombach and Andreas Blattmann and Dominik Lorenz and Patrick Esser and Björn Ommer},
      year={2021},
      eprint={2112.10752},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}