Sign In

[35+ Resources] Must-have Extensions for Stable Diffusion

[35+ Resources] Must-have Extensions for Stable Diffusion

Below is a list of extensions for Stable Diffusion (mainly for Automatic1111 WebUI). I have included ones that efficiently enhanced my workflow, as well as other highly-rated extensions even if I haven't fully tested them yet. The aim is to build a handy list for everyone. I'll continue learning and expanding this list as I go.

Edit (Aug. 9, 2024): I've included new discoveries of handy extensions for a precise control in Stable Diffusion. Happy summer days, folks.

Get any insights to add? Please share them in the comment section, or you can vote or specify other tools in the Poll below (powered by StrawPoll).

How to Install the Extensions?

Using the Github link of the text2prompt extension as an example:

• Install from the extension tab on A1111 WebUI

Copy https://github.com/toshiaki1729/stable-diffusion-webui-text2prompt.git into "Install from URL" tab and hit "Install".

• Install Manually

Clone the repository into the extensions folder and restart the web UI.

git clone https://github.com/toshiaki1729/stable-diffusion-webui-text2prompt.git extensions/text2prompt

You can also download the zip file and unzip it to the extensions folder: \extensions\[extension name]

Part 1. Best Stable Diffusion Extensions for Image Upscaling and Enhancement

1. Tiled Diffusion & VAE

4.6k stars | 330 forks

https://github.com/pkuliyi2015/multidiffusion-upscaler-for-automatic1111

Tiled Diffusion allows you to upscale images larger than 2K at lower VRAM usage. It can be integrated into the Stable Diffusion upscaling workflow, after highres.fix 1.5x or 2x hits the CUDA memory constraints.

It supports integrating Regional Prompt Control, ControlNet, and DemoFusion into the workflow, making it flexible and convenient for the production pipeline.

Here is a detailed tutorial (PDF) by PotatCat: https://civitai.com/models/34726

2. Ultimate SD Upscale

1.5k stars | 153 forks

https://github.com/Coyote-A/ultimate-upscale-for-automatic1111

Ultimate SD Upscale extension helps you to achieve higher resolutions, especially when highres.fix has cuda of out memory errors or cannot reach a larger size. It is better to combine this extension with ControlNet to keep the composition of the image, without extra heads or other image distortions.

It works by dividing the images into grids (tiles) and inpaints it. The extension will fix the seams along the edges and redraw them. You can tweak the mask blur and padding parameters to naturalize the seams. As for upscaling algorithms, the no-brainer choice is ESRGAN for photorealistic images, and R-ESRGAN 4x+ for the rest. For detailed tutorials and explanations on each parameter, read the Wiki page of this extension.

3. Standalone Software (License Giveaway)

If you have CUDA issues (low Vram), or you prefer batch upscaling tons of images with details kept, you can also try using a standalone AI image upscaler -  Aiarty Image Enhancer. See how it handles 512px to 8K/16K upscaling in the video below. Here's also a chance to get a 1 Year License for free.

Part 2. Best Extensions for Stable Diffusion Prompts and Tags

1. Promptgen

448 stars | 85 forks

https://github.com/AUTOMATIC1111/stable-diffusion-webui-promptgen

Promptgen lets you generate prompts right inside the Stable Diffusion WebUI. The author has GPT2 finetunes for prompts scraped from lexica.art and majinai.art (with/without NSFW filters). Users can add more models in the setting.

2. Tokenizer

139 stars | 23 forks

https://github.com/AUTOMATIC1111/stable-diffusion-webui-tokenizer

Tokenizer is an extension for stable-diffusion-webui that adds a tab that lets you preview how CLIP model would tokenize your text.

3. Model-keyword

233 stars | 13 forks

https://github.com/mix1009/model-keyword

 

Model-keyword is an extension for A1111 WebUI that can auto fill keyword for custom stable diffusion models and LoRA models. It works for txt2img and img2img. When prompting, make sure to enable this extension. It will detect the custom checkpoint model or LoRA you are using, and then insert the matching keyword.

4. Prompt Generator

228 stars | 19 forks

https://github.com/imrayya/stable-diffusion-webui-Prompt_Generator

Prompt Generator works in AUTOMATIC1111 WebUI. It helps you to extend the base prompt into detailed descriptions and styles. The script is based on distilgpt2-stable-diffusion-v2 by FredZhang7 and MagicPrompt-Stable-Diffusion by Gustavosta and it runs locally without internet access.

Fred's script can generate descriptive safebooru and danbooru tags, making it a handy extension for txt2img models focusing on anime styles. MagicPrompt series are GPT-2 models to generate prompts for AI images.

For more models, you can edit the json file at \extensions\stable-diffusion-webui-Prompt_Generator\models.json as long as the model is hosted on huggingface.

5. Text2prompt

159 stars | 15 forks

https://github.com/toshiaki1729/stable-diffusion-webui-text2prompt

It is another prompt generator for A1111 WebUI, mainly used for anime. You can increase the K and P value for the prompts to be more creative, or enable the Cutoff and Power option while decreasing Power value. For a more strict result, do the opposite.

Note: only prompts that contain some danbooru tags can be generated.

6. Novelai-2-local-prompt

69 stars | 14 forks

https://github.com/animerl/novelai-2-local-prompt

This script helps you to convert the NovelAI prompt format to Stable Diffusion compliable prompts. It also has a button to recall the previously used prompts.

7. Ranbooru

61 stars| 7 forks

https://github.com/Inzaniak/sd-webui-ranbooru

This extension can pick a random sets of tags from boorus pictures (incl. Gelbooru, Rule34, Safebooru, yande.re, konachan, aibooru, danbooru and xbooru) and add into your prompts. Recommended by ivragi.

1. Stable Diffusion Themes – Lobe Theme

2.3K stars | 215 forks

https://github.com/lobehub/sd-webui-lobe-theme

If you don't like the default Gradio-powered WebUI interface, you can use this extension that comes with a modern sleek UI. Its Github page has detailed instructions on installation and customizing guidance.

Note: before installing the theme, please read carefully about the compatibility issues. For instance, Lobe Theme v3 only works for SD WebUI v1.6 and above. For older versions, you need to use the legacy themes.

2. A1111 WebUI Image Browser

596 stars | 109 forks

https://github.com/AlUlkesh/stable-diffusion-webui-images-browser

The extension helps you to review previously created images, examine the metadata associated with their generation, and utilize that data for further image generation processes such as text-to-image or image-to-image prompts, directly inside the A1111 WebUI.

It also allows you to curate a collection of your preferred images by adding them to a "favorites" album, and provides the functionality to remove images that are no longer of interest to you.

3. Config-Presets

252 stars | 16 forks

https://github.com/Zyin055/Config-Presets

It allows you to configure user-friendly, adjustable dropdown menus to facilitate the modification of parameters, both for the txt2img and the img2img tabs.

You can adjust values to your personal preferences to enhance your workflow efficiency.

4. Inspiration (Artist Styles and Genre)

114 stars | 22 forks

https://github.com/yfszzx/stable-diffusion-webui-inspiration

The Stable Diffusion WebUI Inspiration extension can display random images with signature style of a particular artist or artistic genre. Upon selection, more images from that artist or genre will be presented, making it easy for you to visualize the wanted styles.

It has a vast collection of approximately 6000 artists and styles at your disposal, you can easily refine your search by using keywords to filter through artists or genres. You also have the option to save preferred styles to your favorites or to block those that do not appeal to you.

5. Infinity Grid Generator Script

174 stars | 24 forks

https://github.com/mcmonkeyprojects/sd-infinity-grid-generator-script

This extension is designed to generate infinite-dimensional grids. If you are familiar with the X/Y plot grid, this is the "infinite axis grid" that has more axes on it. Utilizing this tool, you can swiftly analyze outcomes from a myriad of parameter combinations.

Note: Please be aware that the time required to produce a grid is exponential. For example, with four variables, each offering five different options, the result would be a staggering 625 images, calculated as 5 raised to the power of 4.

Part 4. Best Stable Diffusion Extensions for Animations and Videos

1. Animate Diff

2.9k stars | 246 forks

https://github.com/continue-revolution/sd-webui-animatediff

Animate Diff is one of the most popular ways to generate videos with Stable Diffusion. As an easy-to-use video toolkit, you can also generate GIFs with this script.

The author of Animate Diff also created Segment Anything to enhance inpainting and ControlNet semantic segmentation.

2. Deforum

2.6k stars | 383 forks

https://github.com/deforum-art/sd-webui-deforum

It is the official porting of Deforum for Stable Diffusion Automatic1111 WebUI. Using Deforum, you can create animation videos with a list of text prompts and camera movement settings.

The tool offers choices between 2D and 3D motion styles, allowing you to tailor the visual appearance of your animations or videos to your specific preferences. You can also use the interpolation feature.

3. Seed Travel

302 stars | 25 forks

https://github.com/yownas/seed_travel

Seed Travel is an extension for the Stable Diffusion WebUI that helps you generate a series of images, based on the range of seeds you designate.

To illustrate, it compares all possible noise patterns as a map, with each seed represented as a point on that map. What seed traveling does is it picks two "points" on this map and then "travels" between them, delivering a series of images that you can use to create animations from these still frames.

4. Prompt Fusion

257 stars | 16 forks

https://github.com/ljleb/prompt-fusion-extension

Prompt Fusion allows you to "travel" during the sampling stage of a single image, thus creating consistent and slightly iterated visuals. The prompt interpolation feature allows you to create videos by progressively exploring the latent space to generate a sequence of images iteratively. It also has an attention interpolation feature for generating multiple images with subtle variations.

Other similar extensions:

Prompt Travel: https://github.com/Kahsolt/stable-diffusion-webui-prompt-travel

Part 5. Extensions for Controls, Editing and Fine-tuning

1. ControlNet

16.3k stars | 1.9k forks

https://github.com/Mikubill/sd-webui-controlnet

ControlNet is an advanced feature integrated into the Stable Diffusion model. It introduces a level of control over the image generation process that was previously challenging to achieve. The extension allows users to guide the diffusion process by providing specific constraints or conditions, using utilities such as OpenPose, Canny, HED, Scribble, Mlsd, Seg, Depth and Normal Map for various scenarios.

2. DWPose

2k stars | 133 forks

https://github.com/IDEA-Research/DWPose

DWPose claims to output better results than Openpose for ControlNet. It supports consistent and controllable image-to-video synthesis for character animation. Note: DWPose is implemented as a preprocessor (dw_openpose_full) for sd-webui-controlnet v1.1237 and higher.

3. ADetailer

3.9k stars | 303 forks

https://github.com/Bing-su/adetailer

ADetailer is one of the most popular auto-masking and inpainting tools for Stable Diffusion WebUI. Once enabled, it can automatically detect, mask and enhance the faces with more details.

Depending on your image output, sometimes the enhanced image can be overfitting. You can experiment with the parameter settings for better results.

4. Detection Detailer

336 stars | 74 forks

https://github.com/dustysys/ddetailer

This extension is designed for object detection and auto-mask. It can segment people and faces from the background, and automatically improve the faces for details.

5. Reactor (Face Swap)

2.3k stars | 245 forks

https://github.com/Gourieff/sd-webui-reactor

Reactor is a face swap extension for Stable Diffusion WebUI. Before using this script, you need to install dependencies on your system. For instance, for A1111, you need to install Visual Studio or VS C++ Build Tools only (if you don't need the entire Visual Studio). The author's GitHub page has detailed tutorials on this.

You can index faces or specify the gender. If the generated image has blurry faces, you can enable the "Restore face" option.

You can also use Roop: https://github.com/s0md3v/sd-webui-roop

Disclaimer: It is imperative to exercise caution and respect local regulations when using face swap features. It is advised to seek permission from the concerned individual if the software is being employed to alter the visage of an actual person. It is crucial to explicitly state that the resulting image or video is a deepfake.

6. Depth Map Script

1.6k stars | 152 forks

https://github.com/thygate/stable-diffusion-webui-depthmap-script

This extension allows you to generate depth maps, 3D meshes, normalmaps and create 3D stereo image pairs.

For stereo images, you can either go with the "Generate Stereo SBS" or "Generator Anaglyph" options. "Generate 3D inpainted mesh" will deliver mesh needed to generate videos. Besides generation, you can also use the Depth Map script on existing images.

7. Dynamic Thresholding

1.1k stars | 100 forks

https://github.com/mcmonkeyprojects/sd-dynamic-thresholding

Dynamic thresholding (CFG Scale Fix) is an extension for several Stable Diffusion WebUI and ComfyUI. It allows you to use higher CFG scales without color issues. In this way, it gives you more flexible choices when generating images in Stable Diffusion, without worrying about possible distortions introduced by high CFG scales.

8. Multi-subject Render

365 stars | 27 forks

https://github.com/Extraltodeus/multi-subject-render

This add-on allows you to generate multiple subjects for a complex scene at once, powered by depth analysis. The usual prompt UI will be used to generate the background only. Therefore, you just need to describe the background/scenes/settings, without adding any token for the foreground subjects.

For the subjects, there is a "Foreground prompt" section. You can include all the subjects in one line, or use multiple lines, each line describing one character. There's a final blend section to configure. The blend width and height will be the final resolution of the image.

9. Composable LoRA

470 stars | 72 forks

https://github.com/opparco/stable-diffusion-webui-composable-lora

This extension can confine the impact of a LoRA to a designated subprompt. The benefit of this approach is that it can effectively mitigate the adverse impacts introduced by the built-in settings, where negative prompts are always affected by LoRA.

10. Regional Prompter

1.5k stars | 125 forks

https://github.com/hako-mikan/sd-webui-regional-prompter

Regional Prompter allows you to specify prompts for different regions of an image, thus achieving a precise control of content and colors. With a dedicated syntax, you can split the regions by columns and rows, or use masks for a more flexible regional control.

11. Cutoff for SD WebUI

1.2k stars | 85 forks

https://github.com/hnmr293/sd-webui-cutoff

This extension can limit the tokens' influence scope. What does that mean? Let's take a look at the following scenario.

We may find Stable Diffusion mess up with the image when there are lots of colors for different parts of the image. For instance, when we prefer an anime girl to have blue hair and wear a pink skit, the generated image may mess up the skirt and change the hair color to pink.

After enabling the Cutoff extension, we can put the words that we want to limit the scope in "Target tokens".

As you can see, the second rows are images generated with the Cutoff extension turned on. The hair colors (blue) and skirt colors (pink) are not messed up.

12. Latent Couple (Two Shot Diffusion Port)

704 stars| 105 forks

https://github.com/opparco/stable-diffusion-webui-two-shot

This extension allows you to determine the region of the latent space that reflects your subprompts. In plain language, you can mask the canvas, and control what subject will appear on each region. For instance, prompt for mask 1: mountains, for mask 2: trees, for mask 3: moon. The result is more controlled with the desired composition and subject matters.

Part 6. Other Handy Extensions for Stable Diffusion

1. A1111 Aspect Ratio Selector

209 stars | 35 forks

https://github.com/alemelis/sd-webui-ar

Aspect Ratio Selector for Stable Diffusion Automatic1111 WebUI saves you from manually doing the math for different aspect ratios.

Install it by clicking the Extensions tab in WebUI > Install from URL > paste the url https://github.com/alemelis/sd-webui-ar > Install.

There will be aspect ratios appearing under width and height. For aspect ratios larger than 1 (such as 3:2, 4:3, 16:9), the script allows you to change the height while fixing the width. For aspect ratio smaller than 1 (such as 2:3, 3:4, 9:16), the width changes while the height is determined by you.

You can edit /sd-webui-ar/aspect_ratios.txt to personal preferences. The format is:

button-label, width, height, # optional comment.

Example:

1, 512, 512 # 1:1 square

# marks a comment, so lines starting with # will be ignored.

2. OpenOutpaint

395 stars | 24 forks

https://github.com/zero01101/openOutpaint-webUI-extension

It integrates the openOutpaint feature seamlessly into the AUTOMATIC1111 web user interface natively. You can also send the output from A1111 webUI txt2img and img2img tools directly to openOutpaint.

Please note that to activate this extension, the --api flag must be included in your webui-user launch script.

3. 3D Model Loader

228 stars | 22 forks

https://github.com/jtydhr88/sd-3dmodel-loader

Although Gradio has Gradio.Model3D component, but it only supports ojb, glb and gltf, so the author built this extension. You can load your 3D model or anime, edit model pose, and send it to txt2img/img2img.

4. Save Intermediate Images

109 stars | 16 forks

https://github.com/AlUlkesh/sd_save_intermediate_images

As its name indicates, this extension will save the intermediate images upon generation. You can also create videos (webm/mp4) and GIFs out of those images to showcase the "drawing process".

5. Conditioning-highres-fix

46 stars | 8 forks

https://github.com/klimaleksus/stable-diffusion-webui-conditioning-highres-fix

The script can improve the highres.fix feature for sd v1.5 inpainting model. Upon generating, it can rewrite the "Inpainting conditioning mask strength" value by making it relative to "Denoising strength" – set to half of the latter.

Here is a grid comparison provided by the author of this script. As you can see, the lower the Conditioning strength is, the better the image will be, with sharper quality. The AI image will become over saturated and simplified with high denoising strength.

6. SD Scale Calculator

https://preyx.github.io/sd-scale-calc/

This is not an extension for Stable Diffusion but an online tool, but it truly enhances my workflow, so I also bookmarked it here.

As its name suggests, SD scale calculator helps you to calculate the first-pass width and height by entering the target width and height. It is quite handy for highres. fix and img2img resolution settings.

Thanks for reading and participating in the ongoing vote. From the current results, it seems that many of you are interested in Stable Diffusion extensions for precise controls, editing, fine-tuning, and AI upscaling. Should I focus more on exploring extensions in these areas? Let me know your thoughts in the comment below.

205

Comments