mirror of
https://github.com/AUTOMATIC1111/stable-diffusion-webui.git
synced 2025-08-10 18:59:49 +00:00
silence the warning from transformers
add feature demonstrations to readme
This commit is contained in:
48
README.md
48
README.md
@@ -4,8 +4,9 @@ A browser interface based on Gradio library for Stable Diffusion.
|
||||
Original script with Gradio UI was written by a kind anonymopus user. This is a modification.
|
||||
|
||||

|
||||
## Installing and running
|
||||
|
||||
## Stable Diffusion
|
||||
### Stable Diffusion
|
||||
|
||||
This script assumes that you already have main Stable Diffusion sutff installed, assumed to be in directory `/sd`.
|
||||
If you don't have it installed, follow the guide:
|
||||
@@ -21,7 +22,7 @@ Particularly, following files must exist:
|
||||
- `/sd/ldm/util.py`
|
||||
- `/sd/k_diffusion/__init__.py`
|
||||
|
||||
## GFPGAN
|
||||
### GFPGAN
|
||||
|
||||
If you want to use GFPGAN to improve generated faces, you need to install it separately.
|
||||
Follow instructions from https://github.com/TencentARC/GFPGAN, but when cloning it, do so into Stable Diffusion main directory, `/sd`.
|
||||
@@ -37,7 +38,7 @@ The following files must exist:
|
||||
If the GFPGAN directory does not exist, you will not get the option to use GFPGAN in the UI. If it does exist, you will either be able
|
||||
to use it, or there will be a message in console with an error related to GFPGAN.
|
||||
|
||||
## Web UI
|
||||
### Web UI
|
||||
|
||||
Run the script as:
|
||||
|
||||
@@ -56,3 +57,44 @@ Running on local URL: http://127.0.0.1:7860/
|
||||
```
|
||||
|
||||
Open the URL in browser, and you are good to go.
|
||||
|
||||
## Features
|
||||
The script creates a web UI for Stable Diffusion's txt2img and img2img scripts. Following are features added
|
||||
that are not in original script.
|
||||
|
||||
### GFPGAN
|
||||
Lets you improve faces in pictures using the GFPGAN model. There is a checkbox in every tab to use GFPGAN at 100%, and
|
||||
also a separate tab that just allows you to use GFPGAN on any picture, with a slider that controls how strongthe effect is.
|
||||
|
||||

|
||||
|
||||
### Sampling method selection
|
||||
Pick out of three sampling methods for txt2img: DDIM, PLMS, k-diffusion:
|
||||
|
||||

|
||||
|
||||
### Prompt matrix
|
||||
Separate multiple prompts using the `|` character, and the system will produce an image for every combination of them.
|
||||
For example, if you use `a house in a field of grass|at dawn|illustration` prompt, there are four combinations possible (first part of prompt is always kept):
|
||||
|
||||
- `a house in a field of grass`
|
||||
- `a house in a field of grass, at dawn`
|
||||
- `a house in a field of grass, illustration`
|
||||
- `a house in a field of grass, at dawn, illustration`
|
||||
|
||||
Four images will be produced, in this order, all with same seed and each with corresponding prompt:
|
||||
|
||||

|
||||
|
||||
### Flagging
|
||||
Click the Flag button under the output section, and generated images will be saved to `log/images` directory, and generation parameters
|
||||
will be appended to a csv file `log/log.csv` in the `/sd` directory.
|
||||
|
||||
### Copy-paste generation parameters
|
||||
A text output provides generation parameters in an easy to copy-paste form for easy sharing.
|
||||
|
||||

|
||||
|
||||
### Correct seeds for batches
|
||||
If you use a seed of 1000 to generate two batches of two images each, four generated images will have seeds: `1000, 1001, 1002, 1003`.
|
||||
Previous versions of the UI would produce `1000, x, 1001, x`, where x is an iamge that can't be generated by any seed.
|
||||
|
Reference in New Issue
Block a user