1e3b4f9969
Merge from main |
||
---|---|---|
.github | ||
media | ||
scripts | ||
ui | ||
.gitignore | ||
build.bat | ||
build.sh | ||
CONTRIBUTING.md | ||
CreativeML Open RAIL-M License | ||
environment.yaml | ||
How to install and run.txt | ||
LICENSE | ||
README BEFORE YOU RUN THIS.txt | ||
README.md | ||
Troubleshooting.md |
Stable Diffusion UI v2
A simple 1-click way to install and use Stable Diffusion on your own computer. No dependencies or technical knowledge required.
(for support, and development discussion) | Troubleshooting guide for common problems
️🔥🎉 New! Use Custom Weights, Task Queue, Negative Prompt, Live Preview, More Samplers, In-Painting, Face Correction (GFPGAN) and Upscaling (RealESRGAN) have been added!
This distribution currently uses Stable Diffusion 1.4. Once the model for 1.5 becomes publicly available, the model in this distribution will be updated.
Features in the new v2 Version:
- No Dependencies or Technical Knowledge Required: 1-click install for Windows 10/11 and Linux. No dependencies, no need for WSL or Docker or Conda or technical setup. Just download and run!
- Face Correction (GFPGAN) and Upscaling (RealESRGAN)
- In-Painting
- Live Preview: See the image as the AI is drawing it
- Task Queue: Queue up all your ideas, without waiting for the current task to finish
- Custom Weights: Use your own
.ckpt
file, by placing it inside thestable-diffusion
folder (rename it tocustom-model.ckpt
) - Negative Prompt: Specify aspects of the image to remove.
- Lots of Samplers: ddim, plms, heun, euler, euler_a, dpm2, dpm2_a, lms
- Image Modifiers: A library of modifier tags like "Realistic", "Pencil Sketch", "ArtStation" etc. Experiment with various styles quickly.
- New UI: with cleaner design
- Waifu Model Support: Just replace the
stable-diffusion\sd-v1-4.ckpt
file after installation with the Waifu model - Supports "Text to Image" and "Image to Image"
- NSFW Setting: A setting in the UI to control NSFW content
- Use CPU setting: If you don't have a compatible graphics card, but still want to run it on your CPU.
- Auto-updater: Gets you the latest improvements and bug-fixes to a rapidly evolving project.
- Low Memory Usage: Creates 512x512 images with less than 4GB of VRAM!
Live Preview
System Requirements
- Windows 10/11, or Linux. Experimental support for Mac is coming soon.
- An NVIDIA graphics card, preferably with 4GB or more of VRAM. But if you don't have a compatible graphics card, you can still use it with a "Use CPU" setting. It'll be very slow, but it should still work.
You do not need anything else. You do not need WSL, Docker or Conda. The installer will take care of it.
Installation
-
Download for Windows or for Linux.
-
Extract:
- For Windows: After unzipping the file, please move the
stable-diffusion-ui
folder to yourC:
(or any drive like D:, at the top root level), e.g.C:\stable-diffusion-ui
. This will avoid a common problem with Windows (file path length limits). - For Linux: After extracting the .tar.xz file, please open a terminal, and go to the
stable-diffusion-ui
directory.
- Run:
- For Windows:
Start Stable Diffusion UI.cmd
by double-clicking it. - For Linux: In the terminal, run
./start.sh
(orbash start.sh
)
This will automatically install Stable Diffusion, set it up, and start the interface. No additional steps are needed.
To Uninstall: Just delete the stable-diffusion-ui
folder to uninstall all the downloaded packages.
Usage
Open http://localhost:9000 in your browser (after running step 3 previously). It may take a few moments for the back-end to be ready.
With a text description
- Enter a text prompt, like
a photograph of an astronaut riding a horse
in the textbox. - Press
Make Image
. This will take some time, depending on your system's processing power. - See the image generated using your prompt.
With an image
- Click
Browse..
next toInitial Image
. Select your desired image. - An optional text prompt can help you further describe the kind of image you want to generate.
- Press
Make Image
. See the image generated using your prompt.
You can use Face Correction or Upscaling to improve the image further.
Pro tip: You can also click Use as Input
on a generated image, to use it as the input image for your next generation. This can be useful for sequentially refining the generated image with a single click.
Another tip: Images with the same aspect ratio of your generated image work best. E.g. 1:1 if you're generating images sized 512x512.
Problems? Troubleshooting
Please try the common troubleshooting steps. If that doesn't fix it, please ask on the discord server, or file an issue.
Image Settings
You can also set the configuration like seed
, width
, height
, num_outputs
, num_inference_steps
and guidance_scale
using the 'show' button next to 'Image settings'.
Use the same seed
number to get the same image for a certain prompt. This is useful for refining a prompt without losing the basic image design. Enable the random images
checkbox to get random images.
System Settings
The system settings are reachable via the cogwheel symbol on the top right. It can be used to configure whether all generated images should saved be automically, or to tune the Stable Diffusion image generation.
Image Modifiers
Bugs reports and code contributions welcome
If there are any problems or suggestions, please feel free to ask on the discord server or file an issue.
Also, please feel free to submit a pull request, if you have any code contributions in mind. Join the discord server for development-related discussions, and for helping other users.
Disclaimer
The authors of this project are not responsible for any content generated using this interface.
The license of this software forbids you from sharing any content that violates any laws, produce any harm to a person, disseminate any personal information that would be meant for harm, spread misinformation, or target vulnerable groups. For the full list of restrictions please read the license. You agree to these terms by using this software.