Skip to main content
Milloz.com
Rejuvanated Web Tech Tracker

Main navigation

  • Home
User account menu
  • Log in

Breadcrumb

  1. Home

AUTOMATIC1111 Stable Diffusion WebUI Guide — Features, Extensions, Installation & Hardware Requirements 2025 🎨

🎨 The All Rounder of All Stable Diffusion UIs

If you've been anywhere near AI image generation in the last three years, you've probably heard of AUTOMATIC1111's Stable Diffusion WebUI — or simply AUTOMATIC1111 as most people call it. With a staggering 162,000+ stars on GitHub and over 30,000 forks, it isn't just the most popular Stable Diffusion interface — it's one of the most-starred AI projects in existence.

Think of it as the Swiss Army knife of AI image generation. Whether you want to generate images from text prompts (txt2img), edit existing photos (img2img), upscale low-res images, remove backgrounds, or even train your own AI models, AUTOMATIC1111 does it all from a clean web browser interface. It's free, open-source, and runs entirely on your own hardware — no subscriptions, no cloud fees, no data leaving your computer.

In this guide, we'll walk you through everything: what AUTOMATIC1111 is, why it became the undisputed king of Stable Diffusion UIs, its key features, how to install it on every platform, the best extensions that make it even more powerful, hardware requirements, and where things stand now with newer forks like SD WebUI Forge.


🧠 What Is AUTOMATIC1111?

AUTOMATIC1111 is the username of the developer who created the Stable Diffusion WebUI — and over time, the name stuck. The actual project is AUTOMATIC1111/stable-diffusion-webui on GitHub, but everyone just says "AUTOMATIC1111" or "A1111" for short.

Launched in August 2022 — just days after Stable Diffusion 1.4 was released — it quickly became the go-to interface because it was:

  • 🧩 Feature-packed from day one — txt2img, img2img, inpainting, upscaling, all in one place
  • 🔌 Highly extensible — a thriving ecosystem of community extensions
  • 🆓 Free and open-source — AGPL-3.0 license, no paywalls
  • 🖥️ Runs locally — your data never leaves your PC
  • 💪 Works on modest hardware — 4GB VRAM minimum, even reports of 2GB working

Before AUTOMATIC1111, running Stable Diffusion meant wrestling with Python scripts, Jupyter notebooks, or command-line tools. AUTOMATIC1111 wrapped everything into a beautiful Gradio web interface with one-click installers. Suddenly, anyone with a half-decent GPU could generate stunning AI art in minutes.


🌟 Core Features Deep Dive

🎯 txt2img — Text to Image

The bread and butter. Type a prompt, get an image. But AUTOMATIC1111 takes it much further:

  • Negative prompts — tell the AI what you don't want in the image
  • Prompt weighting — make certain words more important: (beautiful:1.5) or ((stunning))
  • Prompt matrix — combine multiple prompts and generate all combinations
  • Attention editing — press Ctrl+Up/Down on selected text to fine-tune emphasis
  • Composable Diffusion — use AND to mix separate prompts with individual weights
  • No token limit — unlike original SD which caps at 75 tokens

🖼️ img2img — Image to Image

Start with an existing image and transform it:

  • Change artistic style while preserving composition
  • Inpaint specific areas (edit parts of an image)
  • Outpaint beyond the canvas boundaries
  • Loopback — run img2img multiple times for iterative refinement
  • Batch processing — process entire folders of images automatically
  • Color sketch — draw rough colors and let AI fill in the details

🔧 Highres Fix

Generate at one resolution, then upscale with added detail in a single click. This avoids the usual distortion of generating directly at high resolutions. A magic button that every AI artist uses daily.

📈 Upscaling Suite

AUTOMATIC1111 bundles multiple upscalers:

  • ESRGAN — classic neural upscaler with many community models
  • RealESRGAN — enhanced version for realistic photos
  • SwinIR / Swin2SR — transformer-based upscalers
  • LDSR — latent diffusion super-resolution
  • Stable Diffusion Upscale — AI-native upscaling using SD itself

🎨 Training Tab

Yes, AUTOMATIC1111 can even train models:

  • Textual Inversion embeddings — teach the AI new concepts using just 3-5 images
  • Hypernetworks — lightweight model fine-tuning
  • LoRA training — the most popular method, gives you portable style/character packs
  • Image preprocessing — auto-crop, mirror, tag using BLIP or DeepDanbooru

Training Textual Inversions works on as little as 8GB VRAM (some report 6GB).

🔬 Advanced Features

  • X/Y/Z Plot — generate a grid of images varying 2-3 parameters (sampler, seed, CFG scale) to find the perfect combination
  • CLIP Interrogator — upload any image, and AUTOMATIC1111 will try to guess the prompt that generated it
  • PNG Info — drag an image to see exactly what prompt/settings/seed produced it
  • Seed resizing — generate the same image at different resolutions
  • Variations — generate slight variations of the same image
  • Prompt editing — change the prompt mid-generation (start with "watermelon" and switch to "anime girl" halfway)
  • Checkpoint Merger — merge up to 3 models into one
  • Reload checkpoints on the fly — no restart needed to switch models
  • Tiling support — create seamless textures and patterns
  • API mode — use AUTOMATIC1111 as a backend for other apps

🔌 Top 10 Must-Have Extensions

AUTOMATIC1111's superpower is its extension ecosystem. Here are the most popular ones:

1. 🎛️ ControlNet (via sd-webui-controlnet)

The single most important extension ever made for Stable Diffusion. ControlNet lets you control image generation using reference images — pose detection (OpenPose), depth maps, Canny edges, normal maps, scribbles, and more. Want a character in a specific pose? Draw a stick figure. Want your building to match a real photo's structure? Use depth ControlNet. This extension single-handedly made Stable Diffusion usable for professional workflows.

2. 🔄 Deforum (deforum/sd-webui-deforum)

⭐ 2,854 stars. The animation powerhouse. Deforum turns AUTOMATIC1111 into a video generation engine — creating stunning AI animations, morphing sequences, and music-reactive videos. It's what most "AI animation" clips on social media use behind the scenes.

3. 🧩 ADetailer (Bing-su/adetailer)

⭐ 4,741 stars. Auto-detecting, masking, and inpainting with detection models. Fix faces, hands, and other problem areas automatically. Since Stable Diffusion famously struggles with hands and faces, ADetailer is essentially mandatory for anyone generating photorealistic humans.

4. 📐 MultiDiffusion Upscaler (pkuliyi2015/multidiffusion-upscaler-for-automatic1111)

⭐ 5,001 stars. Tiled VAE and diffusion for generating ultra-high-resolution images (4K, 8K+) without running out of VRAM. Instead of processing the whole image at once, it tiles it — making large outputs possible even on consumer GPUs.

5. ⚡ NVIDIA TensorRT (NVIDIA/Stable-Diffusion-WebUI-TensorRT)

⭐ 1,992 stars. Official NVIDIA extension that optimizes Stable Diffusion models using TensorRT. Up to 2x faster inference on NVIDIA GPUs. If you have an RTX card, this is the single biggest performance boost you can get.

6. 🎨 Lobe Theme (lobehub/sd-webui-lobe-theme)

⭐ 2,683 stars. A modern, beautiful theme for AUTOMATIC1111. Highly customizable, with dark mode, compact layouts, and a cleaner interface. If the default UI feels dated, this transforms it completely.

7. 🏷️ Civitai Helper (butaixianran/Stable-Diffusion-Webui-Civitai-Helper)

⭐ 2,523 stars. Manage models from Civitai directly within AUTOMATIC1111. Download, preview, and organize community models without ever leaving the UI. Civitai is the largest repository of community-trained Stable Diffusion models, and this extension makes it effortless to browse and install them.

8. 🏷️ WD14 Tagger (kawalain/stable-diffusion-webui-wd14-tagger)

⭐ 1,417 stars. Automatic tagging for training datasets. Uses Waifu Diffusion 14 models to generate caption tags for images — essential for anyone training LoRAs or hypernetworks.

9. 🖼️ Infinite Image Browsing (zanllp/infinite-image-browsing)

⭐ 1,291 stars. A full-featured image management system built into AUTOMATIC1111. Search, filter, organize, and delete your generated images with semantic search and AI-powered tagging. Essential once you've generated thousands of images.

10. 🎭 Segment Anything (continue-revolution/sd-webui-segment-anything)

⭐ 3,515 stars. Meta's Segment Anything Model integrated into AUTOMATIC1111. Click on any object in an image to automatically create a perfect mask — incredibly useful for advanced inpainting and compositing workflows.


💻 Installation Guide (All Platforms)

🪪 Windows (Easiest Method)

  1. Install Python 3.10.6 (critical — newer versions may break PyTorch compatibility)
  2. Install Git from git-scm.com
  3. Download the one-click release package from the v1.0.0-pre release page on GitHub, or clone the repo
  4. Run update.bat to pull dependencies
  5. Run run.bat to start the web interface

That's it. The first launch downloads PyTorch and all dependencies automatically. Your browser opens to http://127.0.0.1:7860.

🐧 Linux

Install dependencies, then one command to get started:

# Debian/Ubuntu
sudo apt install wget git python3 python3-venv libgl1 libglib2.0-0

# Download and run
wget -q https://raw.githubusercontent.com/AUTOMATIC1111/stable-diffusion-webui/master/webui.sh
bash webui.sh

On first run, it creates a Python venv, downloads PyTorch, installs dependencies, and launches the web server — all automatically.

🍎 macOS (Apple Silicon)

M1/M2/M3 Macs are supported natively. Follow the Installation on Apple Silicon guide on the project wiki. Apple Silicon users benefit from the unified memory architecture — Macs with 16GB+ unified memory can run Stable Diffusion quite well.

🐳 Docker

For those who prefer containers, AbdBarho/stable-diffusion-webui-docker (⭐ 7,321) provides a battle-tested Docker setup:

git clone https://github.com/AbdBarho/stable-diffusion-webui-docker
cd stable-diffusion-webui-docker
docker compose --profile download up --build
docker compose --profile auto up --build

This approach keeps your system clean and handles all dependencies in an isolated container.

☁️ Google Colab (No GPU Required)

If you don't have a powerful GPU, camenduru/stable-diffusion-webui-colab (⭐ 15,944) provides a ready-to-use Colab notebook that runs AUTOMATIC1111 on Google's free cloud GPUs.


💾 Hardware Requirements

🥇 Tier 1 — Minimum (Stable Diffusion 1.5 & Light SDXL)

GPU VRAM: 4GB (2GB reports for SD 1.5) | RAM: 8GB | Storage: 10GB+
Example GPUs: GTX 1060 6GB, RTX 3050, GTX 1660 Super
At this tier, you can run SD 1.5 models (512×768) at modest speeds. Textual Inversion training works. SDXL is very slow but possible with tiling. Use --medvram or --lowvram flags.

🥈 Tier 2 — Recommended (Best Experience)

GPU VRAM: 8-12GB | RAM: 16GB | Storage: 50GB+
Example GPUs: RTX 3060 12GB, RTX 3070, RTX 4060 Ti
Smooth SDXL at 1024×1024, decent LoRA training, ControlNet works well, upscaling to 2K. This is the sweet spot for most users.

🥉 Tier 3 — Enthusiast (4K & Heavy Workflows)

GPU VRAM: 16-24GB | RAM: 32GB | Storage: 100GB+
Example GPUs: RTX 4080 Super, RTX 4090, RTX 5090
Ultra-high-res generation (4K+), batch processing, heavy LoRA training, multi-model merging, video generation via Deforum and AnimateDiff. The full experience.


🔄 AUTOMATIC1111 vs SD WebUI Forge vs ComfyUI

AUTOMATIC1111 (162K ⭐)

Best for: Beginners, all-in-one workflow, vast extension ecosystem.
Strengths: Most features built-in, one-click install, largest community, works out of the box.
Weaknesses: Slower than Forge for SDXL and Flux, occasional VRAM bloat, development has slowed.

SD WebUI Forge (12.5K ⭐)

Best for: Users wanting faster performance, Flux model support.
Strengths: Optimized memory management, faster SDXL/Flux inference, compatible with most A1111 extensions.
Weaknesses: Smaller community, some extensions not yet compatible, fewer tutorials.
Created by lllyasviel (the same developer behind ControlNet), Forge is essentially AUTOMATIC1111 with heavily optimized backend code. It achieves 30-45% faster generation on the same hardware by improving memory scheduling and attention optimization.

ComfyUI (55K+ ⭐)

Best for: Advanced users, complex workflows, production pipelines.
Strengths: Node-based visual programming, most memory-efficient, best for complex multi-model workflows, first to support new models.
Weaknesses: Steeper learning curve, not as beginner-friendly, fewer built-in features.

Bottom line: If you're new, start with AUTOMATIC1111. It's the simplest path to generating great images. If you need speed and Flux support, try Forge. For complex production workflows, graduate to ComfyUI.


📦 What Models Can AUTOMATIC1111 Run?

  • ✅ Stable Diffusion 1.4 / 1.5 — the classics, full support
  • ✅ Stable Diffusion 2.0 / 2.1 — supported
  • ✅ SDXL 0.9 / 1.0 — fully supported (may need --medvram on 8GB cards)
  • ✅ SD 3.5 Medium — supported with optimizations
  • ✅ FLUX.1 — limited support (checkpoint merge method); Forge recommended
  • ✅ Segmind SSD-1B — supported (lightweight SDXL alternative)
  • ✅ Alt-Diffusion — supported for multilingual generation
  • ✅ Inpainting models — dedicated support
  • ✅ VAEs — hot-swappable from settings
  • ✅ LoRAs / Hypernetworks / Textual Inversions — full support with preview browser
  • ✅ Checkpoints in .safetensors — modern, secure format supported

🔧 Performance Optimization Tips

  1. Use --xformers — the single biggest speed boost (40%+ faster on NVIDIA). Add it to your COMMANDLINE_ARGS in webui-user.bat or webui-user.sh
  2. Use --medvram on 6-8GB cards — enables SDXL without crashes
  3. Use --lowvram on 4GB cards — sacrifices speed for compatibility
  4. Enable TAESD — a tiny separate neural network for fast preview generation during rendering
  5. Use TensorRT (NVIDIA only) — compile models for 2x faster inference with negligible quality loss
  6. Don't max out batch size — larger batches consume VRAM linearly. Use 1-2 for 8GB cards.
  7. Use half-precision (fp16) — default in AUTOMATIC1111, but verify in settings

🏁 Final Verdict

AUTOMATIC1111 is the gold standard of Stable Diffusion interfaces. It democratized AI image generation the way WordPress democratized web publishing — by making a powerful technology accessible to everyone with a decent computer.

Its 162K GitHub stars, 30K forks, and massive extension ecosystem speak for themselves. No other Stable Diffusion interface comes close in terms of community, documentation, or breadth of features. Even as newer alternatives like Forge and ComfyUI emerge, AUTOMATIC1111 remains the best starting point and the most versatile all-rounder.

Whether you're a complete beginner wanting to generate your first AI image, a digital artist building production workflows, or a developer integrating Stable Diffusion into apps via its API, AUTOMATIC1111 has you covered.

Ready to begin? Visit the GitHub repository, pick your platform from the installation guide above, and start creating. Your first AI-generated image is just a prompt away. 🚀

📷 Header image: AUTOMATIC1111 Stable Diffusion WebUI screenshot — Wikimedia Commons (CC)

Recent content

  • Privacy Policy — Milloz.com
    13 hours ago
  • AUTOMATIC1111 Stable Diffusion WebUI Guide — Features, Extensions, Installation & Hardware Requirements 2025 🎨
    13 hours ago
  • Top 10 ComfyUI Models (2025) - Node-Based AI Image and Video Generation Guide
    13 hours 45 minutes ago
  • Pinokio - One-click AI app store for your PC - Top 10 most downloaded Pinokio apps 2025 — Pinokio Install Instructions
    15 hours ago
  • Here are the top 10 video understanding models on Ollama + the real reason video gen isn't available.
    15 hours 55 minutes ago
  • Ollama's Best Vision models and Image Generation ranked — from FLUX.2 Klein to Llama 3.2 Vision and Gemma 4. 🎨
    16 hours 4 minutes ago
  • Ollama's most popular local LLM models ranked by pulls — Ollama Install guide, Alternatives 🚀
    16 hours 20 minutes ago