change description of sdp, cross attention optimization under settings

w-e-w 2023-06-21 23:03:31 +09:00
parent 329554be04
commit 92ed873b26

@ -2,10 +2,10 @@ A number of optimization can be enabled by [commandline arguments](Command-Line-
| commandline argument | explanation |
|--------------------------------|--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| `--opt-sdp-attention` | Faster speeds than using xformers. (non-deterministic)
| `--opt-sdp-no-mem-attention` | Faster speeds than using xformers. (deterministic, slightly slower than `--opt-sdp-attention` and uses more VRAM)
| `--xformers` | Use [xformers](https://github.com/facebookresearch/xformers) library. Great improvement to memory consumption and speed. Nvidia GPUs only. |
| `--force-enable-xformers` | Enables xformers above regardless of whether the program thinks you can run it or not. Do not report bugs you get running this. |
| `--opt-sdp-attention` | May results in faster speeds than using xformers on some systems but requires more VRAM. (non-deterministic)
| `--opt-sdp-no-mem-attention` | May results in faster speeds than using xformers on some systems but requires more VRAM. (deterministic, slightly slower than `--opt-sdp-attention` and uses more VRAM)
| `--xformers` | Use [xformers](https://github.com/facebookresearch/xformers) library. Great improvement to memory consumption and speed. Nvidia GPUs only. (non-deterministic) |
| `--force-enable-xformers` | Enables xformers regardless of whether the program thinks you can run it or not. Do not report bugs you get running this. |
| `--opt-split-attention` | Cross attention layer optimization significantly reducing memory use for almost no cost (some report improved performance with it). Black magic. <br/>On by default for `torch.cuda`, which includes both NVidia and AMD cards. |
| `--disable-opt-split-attention` | Disables the optimization above. |
| `--opt-sub-quad-attention` | Sub-quadratic attention, a memory efficient Cross Attention layer optimization that can significantly reduce required memory, sometimes at a slight performance cost. Recommended if getting poor performance or failed generations with a hardware/software configuration that xformers doesn't work for. On macOS, this will also allow for generation of larger images. |
@ -15,7 +15,10 @@ A number of optimization can be enabled by [commandline arguments](Command-Line-
| `*do-not-batch-cond-uncond` | Prevents batching of positive and negative prompts during sampling, which essentially lets you run at 0.5 batch size, saving a lot of memory. Decreases performance. Not a command line option, but an optimization implicitly enabled by using `--medvram` or `--lowvram`. |
| `--always-batch-cond-uncond` | Disables the optimization above. Only makes sense together with `--medvram` or `--lowvram` |
| `--opt-channelslast` | Changes torch memory type for stable diffusion to channels last. Effects not closely studied. |
| `--upcast-sampling` | For Nvidia and AMD cards normally forced to run with `--no-half`, [should improve generation speed](https://github.com/AUTOMATIC1111/stable-diffusion-webui/pull/8782). |
| `--upcast-sampling` | For Nvidia and AMD cards normally forced to run with `--no-half`, [should improve generation speed](https://github.com/AUTOMATIC1111/stable-diffusion-webui/pull/8782).
since webui 1.3.0 `Cross attention optimization` can be selected under settings (xformers still needs to enabled via COMMANDLINE_ARGS)
![2023-06-21 22_53_54_877 chrome](https://github.com/AUTOMATIC1111/stable-diffusion-webui/assets/40751091/c72576e1-0f51-4643-ad91-e9aaec4fc125)
Extra tips (Windows):