New release, primarily focused around three major new features: full LCM support, completely new Model Merge functionality and Stable-fast compile support
Also included are several other improvements and large number of hotfixes - see full changelog for details
- Diffusers
- LCM support for any SD 1.5 or SD-XL model!
- download lcm-lora-sd15 and/or lcm-lora-sdxl
- load for favorite SD 1.5 or SD-XL model (original LCM was SD 1.5 only, this is both)
- load lcm lora (note: lcm lora is processed differently than any other lora)
- set sampler to LCM
- set number of steps to some low number, for SD-XL 6-7 steps is normally sufficient
note: LCM scheduler does not support steps higher than 50 - set CFG to between 1 and 2
- Add
cli/lcm-convert.py
script to convert any SD 1.5 or SD-XL model to LCM model
by baking in LORA and uploading to Huggingface, thanks @Disty0 - Support for Stable Fast model compile on Windows/Linux/WSL2 with CUDA
See Wiki:Benchmark for details and comparisment
of different backends, precision modes, advanced settings and compile modes
Hint: 70+ it/s is possible on RTX4090 with no special tweaks - Add additional pipeline types for manual model loads when loading from
safetensors
- Updated logic for calculating steps when using base/hires/refiner workflows
- Improve model offloading for both model and sequential cpu offload when dealing with meta tensors
- Safe model offloading for non-standard models
- Fix DPM SDE scheduler
- Better support for SD 1.5 inpainting models
- Add support for OpenAI Consistency decoder VAE
- Enhance prompt parsing with long prompts and support for BREAK keyword
Change-in-behavior: new line in prompt now means BREAK - Add alternative Lora loading algorithm, triggered if
SD_LORA_DIFFUSERS
is set
- LCM support for any SD 1.5 or SD-XL model!
- Models
- Model merge
- completely redesigned, now based on best-of-class
meh
by @s1dlx
and heavily modified for additional functionality and fully integrated by @AI-Casanova (thanks!) - merge SD or SD-XL models using simple merge (12 methods),
using one of presets (20 built-in presets) or custom block merge values - merge with ReBasin permuatations and/or clipping protection
- fully multithreaded for fastest merge possible
- completely redesigned, now based on best-of-class
- Model update
- under UI -> Models - Update
- scan existing models for updated metadata on CivitAI and
provide download functionality for models with available
- Model merge
- Extra networks
- Use multi-threading for 5x load speedup
- Better Lora trigger words support
- Auto refresh styles on change
- General
- Many mobile UI optimizations, thanks @iDeNoh
- Support for Torch 2.1.1 with CUDA 12.1 or CUDA 11.8
- Configurable location for HF cache folder
Default is standard~/.cache/huggingface/hub
- Reworked parser when pasting previously generated images/prompts
includes alltxt2img
,img2img
andoverride
params - Reworked model compile
- Support custom upscalers in subfolders
- Add additional image info when loading image in process tab
- Better file locking when sharing config and/or models between multiple instances
- Handle custom API endpoints when using auth
- Show logged in user in log when accessing via UI and/or API
- Support
--ckpt none
to skip loading a model
- XYZ grid
- Add refiner options to XYZ Grid
- Add option to create only subgrids in XYZ grid, thanks @midcoastal
- Allow custom font, background and text color in settings
- Fixes
- Fix
params.txt
saved before actual image - Fix inpaint
- Fix manual grid image save
- Fix img2img init image save
- Fix upscale in txt2img for batch counts when no hires is used
- More uniform models paths
- Safe scripts callback execution
- Improved extension compatibility
- Improved BF16 support
- Match previews for reference models with downloaded models
- Fix
Another pretty big release, this time with focus on new models (3 new model types), new backends and optimizations Plus quite a few fixes
Also, Wiki has been updated with new content, so check it out!
Some highlights: OpenVINO, IntelArc, DirectML, ONNX/Olive
- Diffusers
- since now SD.Next supports 12 different model types, we've added reference model for each type in
Extra networks -> Reference for easier select & auto-download
Models can still be downloaded manually, this is just a convenience feature & a showcase for supported models - new model type: Segmind SSD-1B
its a distilled model trained at 1024px, this time 50% smaller and faster version of SD-XL!
(and quality does not suffer, its just more optimized)
test shows batch-size:4 with 1k images at full quality used less than 6.5GB of VRAM
and for further optimization, you can use built-in TAESD decoder,
which results in batch-size:16 with 1k images using 7.9GB of VRAM select from extra networks -> reference or download using built-in Huggingface downloader:segmind/SSD-1B
- new model type: Pixart-α XL 2
in medium/512px and large/1024px variations
comparable in quality to SD 1.5 and SD-XL, but with better text encoder and highly optimized training pipeline
so finetunes can be done in as little as 10% compared to SD/SD-XL (note that due to much larger text encodeder, it is a large model)
select from extra networks -> reference or download using built-in Huggingface downloader:PixArt-alpha/PixArt-XL-2-1024-MS
- new model type: LCM: Latent Consistency Models
trained at 512px, but with near-instant generate in a as little as 3 steps!
combined with OpenVINO, generate on CPU takes less than 5-10 seconds: https://www.youtube.com/watch?v=b90ESUTLsRo
and absolute beast when combined with HyperTile and TAESD decoder resulting in 28 FPS
(on RTX4090 for batch 16x16 at 512px)
note: set sampler to Default before loading model as LCM comes with its own LCMScheduler sampler
select from extra networks -> reference or download using built-in Huggingface downloader:SimianLuo/LCM_Dreamshaper_v7
- support for Custom pipelines, thanks @disty0
download using built-in Huggingface downloader
think of them as plugins for diffusers not unlike original extensions that modify behavior ofldm
backend
list of community pipelines: https://github.com/huggingface/diffusers/blob/main/examples/community/README.md - new custom pipeline:
Disty0/zero123plus-pipeline
, thanks @disty0
generate 4 output images with different camera positions: front, side, top, back!
for more details, see vladmandic#2421 - new backend: ONNX/Olive (experimental), thanks @lshqqytiger
for details, see WiKi - extend support for Free-U
improve generations quality at no cost (other than finding params that work for you)
- since now SD.Next supports 12 different model types, we've added reference model for each type in
- General
- attempt to auto-fix invalid samples which occure due to math errors in lower precision
example:RuntimeWarning: invalid value encountered in cast: sample = sample.astype(np.uint8)
begone black images (note: if it proves as working, this solution will need to be expanded to cover all scenarios) - add Lora OFT support, thanks @antis0007 and @ai-casanova
- Upscalers
- compile option, thanks @disty0
- chaiNNer add high quality models from Helaman
- redesigned Progress bar with full details on current operation
- new option: settings -> images -> keep incomplete
can be used to skip vae decode on aborted/skipped/interrupted image generations - new option: settings -> system paths -> models
can be used to set custom base path for all models (previously only as cli option) - remove external clone of items in
/repositories
- Interrogator module has been removed from
extensions-builtin
and fully implemented (and improved) natively
- attempt to auto-fix invalid samples which occure due to math errors in lower precision
- UI
- UI tweaks for default themes
- UI switch core font in default theme to noto-sans
previously default font was simply system-ui, but it lead to too much variations between browsers and platforms - UI tweaks for mobile devices, thanks @iDeNoh
- updated Context menu
right-click on any button in action menu (e.g. generate button)
- Extra networks
- sort by name, size, date, etc.
- switch between gallery and list views
- add tags from user metadata (in addition to tags in model metadata) for lora
- added Reference models for diffusers backend
- faster enumeration of all networks on server startup
- Packages
- updated
diffusers
to 0.22.0,transformers
to 4.34.1 - update openvino, thanks @disty0
- update directml, @lshqqytiger
- updated
- Compute
- OpenVINO:
- updated to mainstream
torch
2.1.0 - support for ESRGAN upscalers
- updated to mainstream
- OpenVINO:
- Fixes
- fix freeu for backend original and add it to xyz grid
- fix loading diffuser models in huggingface format from non-standard location
- fix default styles looking in wrong location
- fix missing upscaler folder on initial startup
- fix handling of relative path for models
- fix simple live preview device mismatch
- fix batch img2img
- fix diffusers samplers: dpm++ 2m, dpm++ 1s, deis
- fix new style filename template
- fix image name template using model name
- fix image name sequence
- fix model path using relative path
- fix safari/webkit layour, thanks @eadnams22
- fix
torch-rocm
andtensorflow-rocm
version detection, thanks @xangelix - fix chainner upscalers color clipping
- fix for base+refiner workflow in diffusers mode: number of steps, diffuser pipe mode
- fix for prompt encoder with refiner in diffusers mode
- fix prompts-from-file saving incorrect metadata
- fix add/remove extra networks to prompt
- fix before-hires step
- fix diffusers switch from invalid model
- force second requirements check on startup
- remove lyco, multiple_tqdm
- enhance extension compatibility for exensions directly importing codeformers
- enhance extension compatibility for exensions directly accessing processing params
- css fixes
- clearly mark external themes in ui
- update
typing-extensions
This is a major release, with many changes and new functionality...
Changelog is massive, but do read through or you'll be missing on some very cool new functionality
or even free speedups and quality improvements (regardless of which workflows you're using)!
Note that for this release its recommended to perform a clean install (e.g. fresh git clone
)
Upgrades are still possible and supported, but clean install is recommended for best experience
- UI
- added change log to UI
see System -> Changelog - converted submenus from checkboxes to accordion elements
any ui state including state of open/closed menus can be saved as default!
see System -> User interface -> Set menu states - new built-in theme invoked
thanks @BinaryQuantumSoul - add compact view option in settings -> user interface
- small visual indicator bottom right of page showing internal server job state
- added change log to UI
- Extra networks:
- Details
- new details interface to view and save data about extra networks
main ui now has a single button on each en to trigger details view - details view includes model/lora metadata parser!
- details view includes civitai model metadata!
- new details interface to view and save data about extra networks
- Metadata:
- you can scan civitai
for missing metadata and previews directly from extra networks
simply click on button in top-right corner of extra networks page
- you can scan civitai
- Styles
- save/apply icons moved to extra networks
- can be edited in details view
- support for single or multiple styles per json
- support for embedded previews
- large database of art styles included by default
can be disabled in settings -> extra networks -> show built-in - styles can also be used in a prompt directly:
<style:style_name>
if style if an exact match, it will be used
otherwise it will rotate between styles that match the start of the name
that way you can use different styles as wildcards when processing batches - styles can have extra fields, not just prompt and negative prompt
for example: "Extra: sampler: Euler a, width: 480, height: 640, steps: 30, cfg scale: 10, clip skip: 2"
- VAE
- VAEs are now also listed as part of extra networks
- Image preview methods have been redesigned: simple, approximate, taesd, full
please set desired preview method in settings - both original and diffusers backend now support "full quality" setting
if you desired model or platform does not support FP16 and/or you have a low-end hardware and cannot use FP32
you can disable "full quality" in advanced params and it will likely reduce decode errors (infamous black images)
- LoRA
- LoRAs are now automatically filtered based on compatibility with currently loaded model
note that if lora type cannot be auto-determined, it will be left in the list
- LoRAs are now automatically filtered based on compatibility with currently loaded model
- Refiner
- you can load model from extra networks as base model or as refiner
simply select button in top-right of models page
- you can load model from extra networks as base model or as refiner
- General
- faster search, ability to show/hide/sort networks
- refactored subfolder handling
note: this will trigger model hash recaclulation on first model use
- Details
- Diffusers:
- better pipeline auto-detect when loading from safetensors
- SDXL Inpaint
- although any model can be used for inpainiting, there is a case to be made for
dedicated inpainting models as they are tuned to inpaint and not generate - model can be used as base model for img2img or refiner model for txt2img
To download go to Models -> Huggingface:diffusers/stable-diffusion-xl-1.0-inpainting-0.1
(6.7GB)
- although any model can be used for inpainiting, there is a case to be made for
- SDXL Instruct-Pix2Pix
- model can be used as base model for img2img or refiner model for txt2img
this model is massive and requires a lot of resources!
to download go to Models -> Huggingface:diffusers/sdxl-instructpix2pix-768
(11.9GB)
- model can be used as base model for img2img or refiner model for txt2img
- SD Latent Upscale
- you can use SD Latent Upscale models as refiner models
this is a bit experimental, but it works quite well!
to download go to Models -> Huggingface:stabilityai/sd-x2-latent-upscaler
(2.2GB)stabilityai/stable-diffusion-x4-upscaler
(1.7GB)
- you can use SD Latent Upscale models as refiner models
- better Prompt attention
should better handle more complex prompts
for sdxl, choose which part of prompt goes to second text encoder - just addTE2:
separator in the prompt
for hires and refiner, second pass prompt is used if present, otherwise primary prompt is used
new option in settings -> diffusers -> sdxl pooled embeds
thanks @AI-Casanova - better Hires support for SD and SDXL
- better TI embeddings support for SD and SDXL
faster loading, wider compatibility and support for embeddings with multiple vectors
information about used embedding is now also added to image metadata
thanks @AI-Casanova - better Lora handling
thanks @AI-Casanova - better SDXL preview quality (approx method)
thanks @BlueAmulet - new setting: settings -> diffusers -> force inpaint
as some models behave better when in inpaint mode even for normal img2img tasks
- Upscalers:
- pretty much a rewrite and tons of new upscalers - built-in list is now at 42
- fix long outstanding memory leak in legacy code, amazing this went undetected for so long
- more high quality upscalers available by default
SwinIR (2), ESRGAN (12), RealESRGAN (6), SCUNet (2) - if that is not enough, there is new chaiNNer integration:
adds 15 more upscalers from different families out-of-the-box:
HAT (6), RealHAT (2), DAT (1), RRDBNet (1), SPSRNet (1), SRFormer (2), SwiftSR (2)
and yes, you can download and add your own, just place them inmodels/chaiNNer
- two additional latent upscalers based on SD upscale models when using Diffusers backend
SD Upscale 2x, SD Upscale 4x*
note: Recommended usage for SD Upscale is by using second pass instead of upscaler
as it allows for tuning of prompt, seed, sampler settings which are used to guide upscaler - upscalers are available in xyz grid
- simplified settings->postprocessing->upscalers
e.g. all upsamplers share same settings for tiling - allow upscale-only as part of txt2img and img2img workflows
simply set denoising strength to 0 so hires does not get triggered - unified init/download/execute/progress code
- easier installation
- Samplers:
- moved ui options to submenu
- default list for new installs is now all samplers, list can be modified in settings
- simplified samplers configuration in settings
plus added few new ones like sigma min/max which can highly impact sampler behavior - note that list of samplers is now different since keeping a flat-list of all possible
combinations results in 50+ samplers which is not practical
items such as algorithm (e.g. karras) is actually a sampler option, not a sampler itself
- CivitAI:
- civitai model download is now multithreaded and resumable
meaning that you can download multiple models in parallel
as well as resume aborted/incomplete downloads - civitai integration in models -> civitai can now find most
previews AND metadata for most models (checkpoints, loras, embeddings)
metadata is now parsed and saved in [model].json
typical hit rate is >95% for models, loras and embeddings - description from parsed model metadata is used as model description if there is no manual
description file present in format of [model].txt - to enable search for models, make sure all models have set hash values
Models -> Valida -> Calculate hashes
- civitai model download is now multithreaded and resumable
- LoRA
- new unified LoRA handler for all LoRA types (lora, lyco, loha, lokr, locon, ia3, etc.)
applies to both original and diffusers backend
thanks @AI-Casanova for diffusers port - for backend:original, separate lyco handler has been removed
- new unified LoRA handler for all LoRA types (lora, lyco, loha, lokr, locon, ia3, etc.)
- Compute
- CUDA:
- default updated to
torch
2.1.0 with cuda 12.1 - testing moved to
torch
2.2.0-dev/cu122 - check out generate context menu -> show nvml for live gpu stats (memory, power, temp, clock, etc.)
- default updated to
- Intel Arc/IPEX:
- tons of optimizations, built-in binary wheels for Windows
i have to say, intel arc/ipex is getting to be quite a player, especially with openvino
thanks @Disty0 @Nuullll
- tons of optimizations, built-in binary wheels for Windows
- AMD ROCm:
- updated installer to support detect
ROCm
5.4/5.5/5.6/5.7 - support for
torch-rocm-5.7
- updated installer to support detect
- xFormers:
- default updated to 0.0.23
- note that latest xformers are still not compatible with cuda 12.1
recommended to use torch 2.1.0 with cuda 11.8
if you attempt to use xformers with cuda 12.1, it will force a full xformers rebuild on install
which can take a very long time and may/may-not work - added cmd param
--use-xformers
to force usage of exformers
- GC:
- custom garbage collect threshold to reduce vram memory usage, thanks @Disty0
see settings -> compute -> gc
- custom garbage collect threshold to reduce vram memory usage, thanks @Disty0
- CUDA:
- Inference
- new section in settings
- HyperTile: new!
available for diffusers and original backends
massive (up to 2x) speed-up your generations for free :)
note: hypertile is not compatible with any extension that modifies processing parameters such as resolution
thanks @tfernd - Free-U: new!
available for diffusers and original backends
improve generations quality at no cost (other than finding params that work for you)
note: temporarily disabled for diffusers pending release of diffusers==0.22
thanks @ljleb - Token Merging: not new, but updated
available for diffusers and original backends
speed-up your generations by merging redundant tokens
speed up will depend on how aggressive you want to be with token merging - Batch mode
new option settings -> inference -> batch mode
when using img2img process batch, optionally process multiple images in batch in parallel
thanks @Symbiomatrix
- HyperTile: new!
- new section in settings
- NSFW Detection/Censor
- install extension: NudeNet
body part detection, image metadata, advanced censoring, etc...
works for text, image and process workflows
more in the extension notes
- install extension: NudeNet
- Extensions
- automatic discovery of new extensions on github
no more waiting for them to appear in index! - new framework for extension validation
extensions ui now shows actual status of extensions for reviewed extensions
if you want to contribute/flag/update extension status, reach out on github or discord - better overall compatibility with A1111 extensions (up to a point)
- MultiDiffusion
has been removed from list of built-in extensions
you can still install it manually if desired - [LyCORIS]https://github.com/KohakuBlueleaf/a1111-sd-webui-lycoris
has been removed from list of built-in extensions
it is considered obsolete given that all functionality is now built-in
- automatic discovery of new extensions on github
- General
- Startup
- all main CLI parameters can now be set as environment variable as well
for example--data-dir <path>
can be specified asSD_DATADIR=<path>
before starting SD.Next
- all main CLI parameters can now be set as environment variable as well
- XYZ Grid
- more flexibility to use selection or strings
- Logging
- get browser session info in server log
- allow custom log file destination
seewebui --log
- when running with
--debug
flag, log is force-rotated
so eachsdnext.log.*
represents exactly one server run - internal server job state tracking
- Launcher
- new
webui.ps1
powershell launcher for windows (oldwebui.bat
is still valid)
thanks @em411
- new
- API
- add end-to-end example how to use API:
cli/simple-txt2img.js
covers txt2img, upscale, hires, refiner
- add end-to-end example how to use API:
- train.py
- wrapper script around built-in kohya's lora training script
seecli/train.py --help
new support for sd and sdxl, thanks @evshiron
new support for full offline mode (without sdnext server running)
- wrapper script around built-in kohya's lora training script
- Startup
- Themes
- all built-in themes are fully supported:
- black-teal (default), light-teal, black-orange, invoked, amethyst-nightfall, midnight-barbie
- if you're using any gradio default themes or a 3rd party theme or that are not optimized for SD.Next, you may experience issues
default minimal style has been updated for compatibility, but actual styling is completely outside of SD.Next control
- all built-in themes are fully supported:
Started as a mostly a service release with quite a few fixes, but then...
Major changes how hires works as well as support for a very interesting new model Wuerstchen
- tons of fixes
- changes to hires
- enable non-latent upscale modes (standard upscalers)
- when using latent upscale, hires pass is run automatically
- when using non-latent upscalers, hires pass is skipped by default
enabled using force hires option in ui
hires was not designed to work with standard upscalers, but i understand this is a common workflow - when using refiner, upscale/hires runs before refiner pass
- second pass can now also utilize full/quick vae quality
- note that when combining non-latent upscale, hires and refiner output quality is maximum,
but operations are really resource intensive as it includes: base->decode->upscale->encode->hires->refine - all combinations of: decode full/quick + upscale none/latent/non-latent + hires on/off + refiner on/off
should be supported, but given the number of combinations, issues are possible - all operations are captured in image medata
- diffusers:
- allow loading of sd/sdxl models from safetensors without online connectivity
- support for new model: wuerstchen
its a high-resolution model (1024px+) thats ~40% faster than sd-xl with a bit lower resource requirements
go to models -> huggingface -> search "warp-ai/wuerstchen" -> download
its nearly 12gb in size, so be patient :)
- minor re-layout of the main ui
- updated ui hints
- updated models -> civitai
- search and download loras
- find previews for already downloaded models or loras
- new option inference mode
- default is standard
torch.no_grad
new option istorch.inference_only
which is slightly faster and uses less vram, but only works on some gpus
- default is standard
- new cmdline param
--no-metadata
skips reading metadata from models that are not already cached - updated gradio
- styles support for subfolders
- css optimizations
- clean-up logging
- capture system info in startup log
- better diagnostic output
- capture extension output
- capture ldm output
- cleaner server restart
- custom exception handling
One week later, another large update!
- system:
- full python 3.11 support
note that changing python version does require reinstall
and if you're already on python 3.10, really no need to upgrade
- full python 3.11 support
- themes:
- new default theme: black-teal
- new light theme: light-teal
- new additional theme: midnight-barbie, thanks @nyxia
- extra networks:
- support for tags
show tags on hover, search by tag, list tags, add to prompt, etc. - styles are now also listed as part of extra networks
existingstyles.csv
is converted upon startup to individual styles insidemodels/style
this is stage one of new styles functionality
old styles interface is still available, but will be removed in future - cache file lists for much faster startup
speedups are 50+% for large number of extra networks - ui refresh button now refreshes selected page, not all pages
- simplified handling of descriptions
now shows on-mouse-over without the need for user interaction - metadata and info buttons only show if there is actual content
- support for tags
- diffusers:
- add full support for textual inversions (embeddings)
this applies to both sd15 and sdxl
thanks @ai-casanova for porting compel/sdxl code - mix&match base and refiner models (experimental):
most of those are "because why not" and can result in corrupt images, but some are actually useful
also note that if you're not using actual refiner model, you need to bump refiner steps
as normal models are not designed to work with low step count
and if you're having issues, try setting prompt parser to "fixed attention" as majority of problems
are due to token mismatches when using prompt attention- any sd15 + any sd15
- any sd15 + sdxl-refiner
- any sdxl-base + sdxl-refiner
- any sdxl-base + any sd15
- any sdxl-base + any sdxl-base
- ability to interrupt (stop/skip) model generate
- added aesthetics score setting (for sdxl)
used to automatically guide unet towards higher pleasing images
highly recommended for simple prompts - added force zeros setting
create zero-tensor for prompt if prompt is empty (positive or negative)
- add full support for textual inversions (embeddings)
- general:
rembg
remove backgrounds support for is-net model- settings now show markers for all items set to non-default values
- metadata refactored how/what/when metadata is added to images
should result in much cleaner and more complete metadata - pre-create all system folders on startup
- handle model load errors gracefully
- improved vram reporting in ui
- improved script profiling (when running in debug mode)
Time for a quite a large update that has been leaking bit-by-bit over the past week or so...
Note: due to large changes, it is recommended to reset (delete) your ui-config.json
- diffusers:
- support for distilled sd models
just go to models/huggingface and download a model, for example:
segmind/tiny-sd
,segmind/small-sd
,segmind/portrait-finetuned
those are lower quality, but extremely small and fast
up to 50% faster than sd 1.5 and execute in as little as 2.1gb of vram
- support for distilled sd models
- general:
- redesigned settings
- new layout with separated sections:
settings, ui config, licenses, system info, benchmark, models - system info tab is now part of settings
when running outside of sdnext, system info is shown in main ui - all system and image paths are now relative by default
- add settings validation when performing load/save
- settings tab in ui now shows settings that are changed from default values
- settings tab switch to compact view
- new layout with separated sections:
- update gradio major version
this may result in some smaller layout changes since its a major version change
however, browser page load is now much faster - optimizations:
- optimize model hashing
- add cli param
--skip-all
that skips all installer checks
use at personal discretion, but it can be useful for bulk deployments - add model precompile option (when model compile is enabled)
- extra network folder info caching
results in much faster startup when you have large number of extra networks - faster xyz grid switching
especially when using different checkpoints
- update second pass options for clarity
- models:
- civitai download missing model previews
- add openvino (experimental) cpu optimized model compile and inference
enable with--use-openvino
thanks @disty0 - enable batch img2img scale-by workflows
now you can batch process with rescaling based on each individual original image size - fixes:
- fix extra networks previews
- css fixes
- improved extensions compatibility (e.g. sd-cn-animation)
- allow changing vae on-the-fly for both original and diffusers backend
- redesigned settings
Another release thats been baking in dev branch for a while...
- general:
- caching of extra network information to enable much faster create/refresh operations
thanks @midcoastal
- caching of extra network information to enable much faster create/refresh operations
- diffusers:
- add hires support (experimental)
applies to all model types that support img2img, including sd and sd-xl
also supports all hires upscaler types as well as standard params like steps and denoising strength
when used with sd-xl, it can be used with or without refiner loaded
how to enable - there are no explicit checkboxes other than second pass itself:- hires: upscaler is set and target resolution is not at default
- refiner: if refiner model is loaded
- images save options: before hires, before refiner
- redo
move model to cpu
logic in settings -> diffusers to be more reliable
note that system defaults have also changed, so you may need to tweak to your liking - update dependencies
- add hires support (experimental)
Smaller update, but with some breaking changes (to prepare for future larger functionality)...
- general:
- update all metadata saved with images
see https://github.com/vladmandic/automatic/wiki/Metadata for details - improved amd installer with support for navi 2x & 3x and rocm 5.4/5.5/5.6
thanks @evshiron - fix img2img resizing (applies to original, diffusers, hires)
- config change: main
config.json
no longer contains entire configuration
but only differences from defaults (simmilar to recent change performed toui-config.json
)
- update all metadata saved with images
- diffusers:
- enable batch img2img workflows
- original:
- new samplers: dpm++ 3M sde (standard and karras variations)
enable in settings -> samplers -> show samplers - expose always/never discard penultimage sigma
enable in settings -> samplers
- new samplers: dpm++ 3M sde (standard and karras variations)
This is a big one that's been cooking in dev
for a while now, but finally ready for release...
- diffusers:
- pipeline autodetect
if pipeline is set to autodetect (default for new installs), app will try to autodetect pipeline based on selected model
this should reduce user errors such as loading sd-xl model when sd pipeline is selected - quick vae decode as alternative to full vae decode which is very resource intensive
quick decode is based ontaesd
and produces lower quality, but its great for tests or grids as it runs much faster and uses far less vram
disabled by default, selectable in txt2img/img2img -> advanced -> full quality - prompt attention for sd and sd-xl
supports bothfull parser
and nativecompel
thanks @ai-casanova - advanced lora load/apply methods
in addition to standard lora loading that was recently added to sd-xl using diffusers, now we have- sequential apply (load & apply multiple loras in sequential manner) and
- merge and apply (load multiple loras and merge before applying to model)
see settings -> diffusers -> lora methods
thanks @hameerabbasi and @ai-casanova
- sd-xl vae from safetensors now applies correct config
result is that 3rd party vaes can be used without washed out colors - options for optimized memory handling for lower memory usage
see settings -> diffusers
- pipeline autodetect
if pipeline is set to autodetect (default for new installs), app will try to autodetect pipeline based on selected model
- general:
- new civitai model search and download
native support for civitai, integrated into ui as models -> civitai - updated requirements
this time its a bigger change so upgrade may take longer to install new requirements - improved extra networks performance with large number of networks
- new civitai model search and download
Another minor update, but it unlocks some cool new items...
- diffusers:
- vaesd live preview (sd and sd-xl)
- fix inpainting (sd and sd-xl)
- general:
- new torch 2.0 with ipex (intel arc)
- additional callbacks for extensions
enables latest comfyui extension
Smaller release, but IMO worth a post...
- diffusers:
- sd-xl lora's are now supported!
- memory optimizations: Enhanced sequential CPU offloading, model CPU offload, FP16 VAE
- significant impact if running SD-XL (for example, but applies to any model) with only 8GB VRAM
- update packages
- minor bugfixes
This is a big one, new models, new diffusers, new features and updated UI...
First, SD-XL 1.0 is released and yes, SD.Next supports it out of the box!
Also fresh is new Kandinsky 2.2 model that does look quite nice:
Actual changelog is:
-
general:
- new loading screens and artwork
- major ui simplification for both txt2img and img2img
nothing is removed, but you can show/hide individual sections
default is very simple interface, but you can enable any sections and save it as default in settings - themes: add additional built-in theme,
amethyst-nightfall
- extra networks: add add/remove tags to prompt (e.g. lora activation keywords)
- extensions: fix couple of compatibility items
- firefox compatibility improvements
- minor image viewer improvements
- add backend and operation info to metadata
-
diffusers:
- we're out of experimental phase and diffusers backend is considered stable
- sd-xl: support for sd-xl 1.0 official model
- sd-xl: loading vae now applies to both base and refiner and saves a bit of vram
- sd-xl: denoising_start/denoising_end
- sd-xl: enable dual prompts
dual prompt is used if set regardless if refiner is enabled/loaded
if refiner is loaded & enabled, refiner prompt will also be used for refiner pass- primary prompt goes to OpenAI CLIP-ViT/L-14
- refiner prompt goes to OpenCLIP-ViT/bigG-14
- kandinsky 2.2 support
note: kandinsky model must be downloaded using model downloader, not as safetensors due to specific model format - refiner: fix batch processing
- vae: enable loading of pure-safetensors vae files without config
also enable automatic selection to work with diffusers - sd-xl: initial lora support
right now this applies to official lora released by stability-ai, support for kohya's lora is expected soon - implement img2img and inpainting (experimental)
actual support and quality depends on model
it works as expected for sd 1.5, but not so much for sd-xl for now - implement limited stop/interrupt for diffusers works between stages, not within steps
- add option to save image before refiner pass
- option to set vae upcast in settings
- enable fp16 vae decode when using optimized vae
this pretty much doubles performance of decode step (delay after generate is done)
-
original
- fix hires secondary sampler
this now fully obsoletesfallback_sampler
andforce_latent_sampler
- fix hires secondary sampler
While we're waiting for official SD-XL release, here's another update with some fixes and enhancements...
- global
- image save: option to add invisible image watermark to all your generated images
disabled by default, can be enabled in settings -> image options
watermark information will be shown when loading image such as in process image tab
also additional cli utility/cli/image-watermark.py
to read/write/strip watermarks from images - batch processing: fix metadata saving, also allow to drag&drop images for batch processing
- ui configuration: you can modify all ui default values from settings as usual,
but only values that are non-default will be written toui-config.json
- startup: add cmd flag to skip all
torch
checks - startup: force requirements check on each server start
there are too many misbehaving extensions that change system requirements - internal: safe handling of all config file read/write operations
this allows sdnext to run in fully shared environments and prevents any possible configuration corruptions
- image save: option to add invisible image watermark to all your generated images
- diffusers:
- sd-xl: remove image watermarks autocreated by 0.9 model
- vae: enable loading of external vae, documented in diffusers wiki
and mix&match continues, you can even use sd-xl vae with sd 1.5 models! - samplers: add concept of default sampler to avoid needing to tweak settings for primary or second pass
note that sampler details will be printed in log when running in debug level - samplers: allow overriding of sampler beta values in settings
- refiner: fix refiner applying only to first image in batch
- refiner: allow using direct latents or processed output in refiner
- model: basic support for one more model: UniDiffuser
download using model downloader:thu-ml/unidiffuser-v1
and set resolution to 512x512
Trying to unify settings for both original and diffusers backend without introducing duplicates...
- renamed hires fix to second pass
as that is what it actually is, name hires fix is misleading to start with - actual hires fix and refiner are now options inside second pass section
- obsoleted settings -> sampler -> force_latent_sampler
it is now part of second pass options and it works the same for both original and diffusers backend
which means you can use different scheduler settings for txt2img and hires if you want - sd-xl refiner will run if its loaded and if second pass is enabled
so you can quickly enable/disable refiner by simply enabling/disabling second pass - you can mix&match model and refiner
for example, you can generate image using sd 1.5 and still use sd-xl refiner as second pass - reorganized settings -> samplers to show which section refers to which backend
- added diffusers lmsd sampler
Another big one, but now improvements to both diffusers and original backends as well plus ability to dynamically switch between them!
- swich backend between diffusers and original on-the-fly
- you can still use
--backend <backend>
and now that only means in which mode app will start, but you can change it anytime in ui settings - for example, you can even do things like generate image using sd-xl,
then switch to original backend and perform inpaint using a different model
- you can still use
- diffusers backend:
- separate ui settings for refiner pass with sd-xl
you can specify: prompt, negative prompt, steps, denoise start - fix loading from pure safetensors files
now you can load sd-xl from safetensors file or from huggingface folder format - fix kandinsky model (2.1 working, 2.2 was just released and will be soon)
- separate ui settings for refiner pass with sd-xl
- original backend:
- improvements to vae/unet handling as well as cross-optimization heads
in non-technical terms, this means lower memory usage and higher performance
and you should be able to generate higher resolution images without any other changes
- improvements to vae/unet handling as well as cross-optimization heads
- other:
- major refactoring of the javascript code
includes fixes for text selections and navigation - system info tab now reports on nvidia driver version as well
- minor fixes in extra-networks
- installer handles origin changes for submodules
- major refactoring of the javascript code
big thanks to @huggingface team for great communication, support and fixing all the reported issues asap!
Service release with some fixes and enhancements:
- diffusers:
- option to move base and/or refiner model to cpu to free up vram
- model downloader options to specify model variant / revision / mirror
- now you can download
fp16
variant directly for reduced memory footprint - basic img2img workflow (sketch and inpaint are not supported yet)
note that sd-xl img2img workflows are architecturaly different so it will take longer to implement - updated hints for settings
- extra networks:
- fix corrupt display on refesh when new extra network type found
- additional ui tweaks
- generate thumbnails from previews only if preview resolution is above 1k
- image viewer:
- fixes for non-chromium browsers and mobile users and add option to download image
- option to download image directly from image viewer
- general
- fix startup issue with incorrect config
- installer should always check requirements on upgrades
This is a massive update which has been baking in a dev
branch for a while now
- merge experimental diffusers support
TL;DR: Yes, you can run SD-XL model in SD.Next now
For details, see Wiki page: Diffusers
Note this is still experimental, so please follow Wiki
Additional enhancements and fixes will be provided over the next few days
Thanks to @huggingface team for making this possible and our internal @team for all the early testing
Release also contains number of smaller updates:
- add pan & zoom controls (touch and mouse) to image viewer (lightbox)
- cache extra networks between tabs
this should result in neat 2x speedup on building extra networks - add settings -> extra networks -> do not automatically build extra network pages
speeds up app start if you have a lot of extra networks and you want to build them manually when needed - extra network ui tweaks
Small quality-of-life updates and bugfixes:
- add option to disallow usage of ckpt checkpoints
- change lora and lyco dir without server restart
- additional filename template fields:
uuid
,seq
,image_hash
- image toolbar is now shown only when image is present
- image
Zip
button gone and its not optional setting that applies to standardSave
button - folder
Show
button is present only when working on localhost,
otherwise its replaced withCopy
that places image URLs on clipboard so they can be used in other apps
A bit bigger update this time, but contained to specific areas...
- change in behavior
extensions no longer auto-update on startup
using--upgrade
flag upgrades core app as well as all submodules and extensions - live server log monitoring in ui
configurable via settings -> live preview - new extra networks interface
note: if you're using a 3rd party ui extension for extra networks, it will likely need to be updated to work with new interface- display in front of main ui, inline with main ui or as a sidebar
- lazy load thumbnails
drastically reduces load times for large number of extra networks - auto-create thumbnails from preview images in extra networks in a background thread
significant load time saving on subsequent restarts - support for info files in addition to description files
- support for variable aspect-ratio thumbnails
- new folder view
- extensions sort by trending
- add requirements check for training
- new training tab interface
- redesigned preprocess, train embedding, train hypernetwork
- new models tab interface
- new model convert functionality, thanks @akegarasu
- new model verify functionality
- lot of ipex specific fixes/optimizations, thanks @disty0
This one is less relevant for standard users, but pretty major if you're running an actual server
But even if not, it still includes bunch of cumulative fixes since last release - and going by number of new issues, this is probably the most stable release so far...
(next one is not going to be as stable, but it will be fun :) )
- minor improvements to extra networks ui
- more hints/tooltips integrated into ui
- new dedicated api server
- but highly promising for high throughput server
- improve server logging and monitoring with
- server log file rotation
- ring buffer with api endpoint
/sdapi/v1/log
- real-time status and load endpoint
/sdapi/v1/system-info/status
Second stage of a jumbo merge from upstream plus few minor changes...
- simplify token merging
- reorganize some settings
- all updates from upstream: A1111 v1.3.2 [df004be] (latest release)
pretty much nothing major that i haven't released in previous versions, but its still a long list of tiny changes- skipped/did-not-port:
add separate hires prompt: unnecessarily complicated and spread over large number of commits due to many regressions
allow external scripts to add cross-optimization methods: dangerous and i don't see a use case for it so far
load extension info in threads: unnecessary as other optimizations i've already put place perform equally good - broken/reverted:
sub-quadratic optimization changes
- skipped/did-not-port:
Just a day later and one bigger update... Both some new functionality as well as massive merges from upstream
- new cache for models/lora/lyco metadata:
metadata.json
drastically reduces disk access on app startup - allow saving/resetting of ui default values
settings -> ui defaults - ability to run server without loaded model
default is to auto-load model on startup, can be changed in settings -> stable diffusion
if disabled, model will be loaded on first request, e.g. when you click generate
useful when you want to start server to perform other tasks like upscaling which do not rely on model - updated
accelerate
andxformers
- huge nubmer of changes ported from A1111 upstream
this was a massive merge, hopefully this does not cause any regressions
and still a bit more pending...
- updated ui labels and hints to improve clarity and provide some extra info
this is 1st stage of the process, more to come...
if you want to join the effort, see vladmandic#1246 - new localization and hints engine
how hints are displayed can be selected in settings -> ui - reworked installer sequence
as some extensions are loading packages directly from their preload sequence
which was preventing some optimizations to take effect - updated settings tab functionality, thanks @gegell
with real-time monitor for all new and/or updated settings - launcher will now warn if application owned files are modified
you are free to add any user files, but do not modify app files unless you're sure in what you're doing - add more profiling for scripts/extensions so you can see what takes time
this applies both to initial load as well as execution - experimental
sd_model_dict
setting which allows you to load model dictionary
from one model and apply weights from another model specified insd_model_checkpoint
results? who am i to judge :)
Few new features and extra handling for broken extensions
that caused my phone to go crazy with notifications over the weekend...
- added extra networks to xyz grid options
now you can have more fun with all your embeddings and loras :) - new vae decode method to help with larger batch sizes, thanks @bigdog
- new setting -> lora -> use lycoris to handle all lora types
this is still experimental, but the goal is to obsolete old built-in lora module
as it doesn't understand many new loras and built-in lyco module can handle it all - somewhat optimize browser page loading
still slower than i'd want, but gradio is pretty bad at this - profiling of scripts/extensions callbacks
you can now see how much or pre/post processing is done, not just how long generate takes - additional exception handling so bad exception does not crash main app
- additional background removal models
- some work on bfloat16 which nobody really should be using, but why not 🙂
Some quality-of-life improvements while working on larger stuff in the background...
- redesign action box to be uniform across all themes
- add pause option next to stop/skip
- redesigned progress bar
- add new built-in extension: agent-scheduler
very elegant way to getting full queueing capabilities, thank @artventurdev - enable more image formats
note: not all are understood by browser so previews and images may appear as blank
unless you have some browser extensions that can handle them
but they are saved correctly. and cant beat raw quality of 32-bittiff
orpsd
:) - change in behavior:
xformers
will be uninstalled on startup if they are not active
if you do havexformers
selected as your desired cross-optimization method, then they will be used
reason is that a lot of libaries try to blindly import xformers even if they are not selected or not functional
Another bigger one...And more to come in the next few days...
- new live preview mode: taesd
i really like this one, so its enabled as default for new installs - settings search feature
- new sampler: dpm++ 2m sde
- fully common save/zip/delete (new) options in all tabs
which (again) meant rework of process image tab - system info tab: live gpu utilization/memory graphs for nvidia gpus
- updated controlnet interface
- minor style changes
- updated lora, swinir, scunet and ldsr code from upstream
- start of merge from a1111 v1.3
Some quality-of-life improvements...
- updated README
- created CHANGELOG
this will be the source for all info about new things moving forward
and cross-posted to Discussions#99 as well as discord announcements - optimize model loading on startup
this should reduce startup time significantly - set default cross-optimization method for each platform backend
applicable for new installs onlycuda
=> Scaled-Dot-Productrocm
=> Sub-quadraticdirectml
=> Sub-quadraticipex
=> InvokeAI'smps
=> Doggettx'scpu
=> Doggettx's
- optimize logging
- optimize profiling
now includes startup profiling as well ascuda
profiling during generate - minor lightbox improvements
- bugfixes...i don't recall when was a release with at least several of those
other than that - first stage of Diffusers integration is now in master branch
i don't recommend anyone to try it (and dont even think reporting issues for it)
but if anyone wants to contribute, take a look at project page
Major internal work with perhaps not that much user-facing to show for it ;)
- update core repos: stability-ai, taming-transformers, k-diffusion, blip, codeformer
note: to avoid disruptions, this is applicable for new installs only - tested with torch 2.1, cuda 12.1, cudnn 8.9
(production remains on torch2.0.1+cuda11.8+cudnn8.8) - fully extend support of
--data-dir
allows multiple installations to share pretty much everything, not just models
especially useful if you want to run in a stateless container or cloud instance - redo api authentication
now api authentication will use same user/pwd (if specified) for ui and strictly enforce it using httpbasicauth
new authentication is also fully supported in combination with ssl for both sync and async calls
if you want to use api programatically, see examples incli/sdapi.py
- add dark/light theme mode toggle
- redo some
clip-skip
functionality - better matching for vae vs model
- update to
xyz grid
to allow creation of large number of images without creating grid itself - update
gradio
(again) - more prompt parser optimizations
- better error handling when importing image settings which are not compatible with current install
for example, when upscaler or sampler originally used is not available - fixes...amazing how many issues were introduced by porting a1111 v1.20 code without adding almost no new functionality
next one is v1.30 (still in dev) which does bring a lot of new features
This is a massive one due to huge number of changes,
but hopefully it will go ok...
- new prompt parsers
select in UI -> Settings -> Stable Diffusion- Full: my new implementation
- A1111: for backward compatibility
- Compel: as used in ComfyUI and InvokeAI (a.k.a Temporal Weighting)
- Fixed: for really old backward compatibility
- monitor extensions install/startup and
log if they modify any packages/requirements
this is a deep-experimental python hack, but i think its worth it as extensions modifying requirements
is one of most common causes of issues - added
--safe
command line flag mode which skips loading user extensions
please try to use it before opening new issue - reintroduce
--api-only
mode to start server without ui - port all upstream changes from A1111
up to today - commit hash89f9faa
- major work on prompt parsing
this can cause some differences in results compared to what you're used to, but its all about fixes & improvements
- prompt parser was adding commas and spaces as separate words and tokens and/or prefixes
- negative prompt weight using
[word:weight]
was ignored, it was always0.909
- bracket matching was anything but correct. complex nested attention brackets are now working.
- btw, if you run with
--debug
flag, you'll now actually see parsed prompt & schedule
- updated all scripts in
/cli
- add option in settings to force different latent sampler instead of using primary only
- add interrupt/skip capabilities to process images
This is mostly about optimizations...
- improved
torch-directml
support
especially interesting for amd users on windows where torch+rocm is not yet available
dont forget to run using--use-directml
or default is cpu - improved compatibility with nvidia rtx 1xxx/2xxx series gpus
- fully working
torch.compile
with torch 2.0.1
usinginductor
compile takes a while on first run, but does result in 5-10% performance increase - improved memory handling
for highest performance, you can also disable aggressive gc in settings - improved performance
especially after generate as image handling has been moved to separate thread - allow per-extension updates in extension manager
- option to reset configuration in settings
- brand new extension manager
this is pretty much a complete rewrite, so new issues are possible - support for
torch
2.0.1
note that if you are experiencing frequent hangs, this may be a worth a try - updated
gradio
to 3.29.0 - added
--reinstall
flag to force reinstall of all packages - auto-recover & re-attempt when
--upgrade
is requested but fails - check for duplicate extensions
Back online with few updates:
- bugfixes. yup, quite a lot of those
- auto-detect some cpu/gpu capabilities on startup
this should reduce need to tweak and tune settings like no-half, no-half-vae, fp16 vs fp32, etc - configurable order of top level tabs
- configurable order of scripts in txt2img and img2img
for both, see sections in ui-> settings -> user interface
Again, few days later...
- reviewed/ported all commits from A1111 upstream
some a few are not applicable as i already have alternative implementations
and very few i choose not to implement (save/restore last-known-good-config is a bad hack)
otherwise, we're fully up to date (its doesn't show on fork status as code merges were mostly manual due to conflicts)
but...due to sheer size of the updates, this may introduce some temporary issues - redesigned server restart function
now available and working in ui
actually, since server restart is now a true restart and not ui restart, it can be used much more flexibly - faster model load
plus support for slower devices via stream-load function (in ui settings) - better logging
this includes new--debug
flag for more verbose logging when troubleshooting
Been a bit quieter for last few days as changes were quite significant, but finally here we are...
- Updated core libraries: Gradio, Diffusers, Transformers
- Added support for Intel ARC GPUs via Intel OneAPI IPEX (auto-detected)
- Added support for TorchML (set by default when running on non-compatible GPU or on CPU)
- Enhanced support for AMD GPUs with ROCm
- Enhanced support for Apple M1/M2
- Redesigned command params: run
webui --help
for details - Redesigned API and script processing
- Experimental support for multiple Torch compile options
- Improved sampler support
- Google Colab: https://colab.research.google.com/drive/126cDNwHfifCyUpCCQF9IHpEdiXRfHrLN
Maintained by https://github.com/Linaqruf/sd-notebook-collection - Fixes, fixes, fixes...
To take advantage of new out-of-the-box tunings, its recommended to delete your config.json
so new defaults are applied. Its not necessary, but otherwise you may need to play with UI Settings to get the best of Intel ARC, TorchML, ROCm or Apple M1/M2.
a bit shorter list as:
- i've been busy with bugfixing
there are a lot of them, not going to list each here.
but seems like critical issues backlog is quieting down and soon i can focus on new features development. - i've started collaboration with couple of major projects, hopefully this will accelerate future development.
what's new:
- ability to view/add/edit model description shown in extra networks cards
- add option to specify fallback sampler if primary sampler is not compatible with desired operation
- make clip skip a local parameter
- remove obsolete items from UI settings
- set defaults for AMD ROCm
if you have issues, you may want to start with a fresh install so configuration can be created from scratch - set defaults for Apple M1/M2
if you have issues, you may want to start with a fresh install so configuration can be created from scratch
- update process image -> info
- add VAE info to metadata
- update GPU utility search paths for better GPU type detection
- update git flags for wider compatibility
- update environment tuning
- update ti training defaults
- update VAE search paths
- add compatibility opts for some old extensions
- validate script args for always-on scripts
fixes: deforum with controlnet
- identify race condition where generate locks up while fetching preview
- add pulldowns to x/y/z script
- add VAE rollback feature in case of NaNs
- use samples format for live preview
- add token merging
- use Approx NN for live preview
- create default
styles.csv
- fix setup not installing
tensorflow
dependencies - update default git flags to reduce number of warnings
- fix VAE dtype
should fix most issues with NaN or black images - add built-in Gradio themes
- reduce requirements
- more AMD specific work
- initial work on Apple platform support
- additional PR merges
- handle torch cuda crashing in setup
- fix setup race conditions
- fix ui lightbox
- mark tensorflow as optional
- add additional image name templates
- autodetect which system libs should be installed
this is a first pass of autoconfig for nVidia vs AMD environments - fix parse cmd line args from extensions
- only install
xformers
if actually selected as desired cross-attention method - do not attempt to use
xformers
orsdp
if running on cpu - merge tomesd token merging
- merge 23 PRs pending from a1111 backlog (!!)
expect shorter updates for the next few days as i'll be partially ooo
- full CUDA tuning section in UI Settings
- improve exif/pnginfo metadata parsing
it can now handle 3rd party images or images edited in external software - optimized setup performance and logging
- improve compatibility with some 3rd party extensions for example handle extensions that install packages directly from github urls
- fix initial model download if no models found
- fix vae not found issues
- fix multiple git issues
note: if you previously had command line optimizations such as --no-half, those are now ignored and moved to ui settings
- fix live preview
- fix model merge
- fix handling of user-defined temp folders
- fix submit benchmark
- option to override
torch
andxformers
installer - separate benchmark data for system-info extension
- minor css fixes
- created initial merge backlog from pending prs on a1111 repo
see #258 for details
- reconnect ui to active session on browser restart
this is one of most frequently asked for items, finally figured it out
works for text and image generation, but not for process as there is no progress bar reported there to start with - force unload
xformers
when not used
improves compatibility with AMD/M1 platforms - add
styles.csv
to UI settings to allow customizing path - add
--skip-git
to cmd flags for power users that want
to skip all git checks and operations and perform manual updates - add
--disable-queue
to cmd flags that disables Gradio queues (experimental) this forces it to use HTTP instead of WebSockets and can help on unreliable network connections - set scripts & extensions loading priority and allow custom priorities
fixes random extension issues:
ScuNet
upscaler disappearing,Additional Networks
not showing up on XYZ axis, etc. - improve html loading order
- remove some
asserts
causing runtime errors and replace with user-friendly messages - update README.md
- update TODO.md
- themes are now dynamic and discovered from list of available gradio themes on huggingface
its quite a list of 30+ supported themes so far - added option to see theme preview without the need to apply it or restart server
- integrated image info functionality into process image tab and removed separate image info tab
- more installer improvements
- fix urls
- updated github integration
- make model download as optional if no models found
- support for ui themes! to to settings -> user interface -> "ui theme* includes 12 predefined themes
- ability to restart server from ui
- updated requirements
- removed
styles.csv
from repo, its now fully under user control - removed model-keyword extension as overly aggressive
- rewrite of the fastapi middleware handlers
- install bugfixes, hopefully new installer is now ok
i really want to focus on features and not troubleshooting installer
- update default values
- remove
ui-config.json
from repo, its not fully under user control - updated extensions manager
- updated locon/lycoris plugin
- enable quick launch by default
- add multidiffusion upscaler extensions
- add model keyword extension
- enable strong linting
- fix circular imports
- fix extensions updated
- fix git update issues
- update github templates
- handle duplicate extensions
- redo exception handler
- fix generate forever
- enable cmdflags compatibility
- change default css font
- fix ti previews on initial start
- enhance tracebacks
- pin transformers version to last known good version
- fix extension loader
This has been pending for a while, but finally uploaded some massive changes
- New launcher
webui.bat
andwebui.sh
:
Platform specific wrapper scripts that startslaunch.py
in Python virtual environment
Note: Server can run without virtual environment, but it is recommended to use it
This is carry-over from original repo
If you're unsure which launcher to use, this is the one you wantlaunch.py
:
Main startup script
Can be used directly to start server in manually activatedvenv
or to run it withoutvenv
installer.py
:
Main installer, used bylaunch.py
webui.py
:
Main server script
- New logger
- New exception handler
- Built-in performance profiler
- New requirements handling
- Move of most of command line flags into UI Settings