Releases: Teriks/dgenerate
v4.2.3 (Stable Diffusion 3 Inpainting, ROCm (AMD) support)
v4.2.3 release with Windows installer.
Due to the size of the packaged python environment, the installer is within a multi-part zip file.
The multipart zip can be extracted using 7-Zip: https://www.7-zip.org/
Download both dgenerate_installer.zip.001
and dgenerate_installer.zip.002
to a folder.
Unzip dgenerate_installer.zip.001
to a directory (Right click, 7-Zip -> Extract to "dgenerate_installer") and then run dgenerate_installer\dgenerate.msi
to install.
dgenerate will be installed under C:\Program Files\dgenerate
by default with an isolated python environment provided.
The install directory will be added to PATH, and dgenerate will be available from the command line.
Portable Install
A portable install is provided via dgenerate_portable.zip.001
and dgenerate_portable.zip.002
, these contain
nothing but the dgenerate executable and a frozen python environment which can be placed anywhere.
4.2.3 Patch Fix
Fix missing upscaler-ncnn
processor in Console UI recipe forms.
4.2.0 Features
-
Basic inpainting is now supported with Stable Diffusion 3
-
Config directive
\setp
can be used to evaluate python ternary statements during template variable assignment. i.e.\setp my_var "Yes" if True else "No"
-
Improved config error reporting. 1 indexed line number in the interpreter to align with Console UI line numbering. Better pinpointing of the error line number inside "here template" control structures involving Jinja. Patches for issues with
asteval
affecting syntax error output. -
PyPI Packaging now supports installing dgenerate with the ROCm torch backend on Linux, allowing for use with AMD Cards / Accelerators. See installation documentation: Linux with ROCm (AMD Cards)
v4.2.2 (Stable Diffusion 3 Inpainting, ROCm (AMD) support)
v4.2.2 release with Windows installer.
Due to the size of the packaged python environment, the installer is within a multi-part zip file.
The multipart zip can be extracted using 7-Zip: https://www.7-zip.org/
Download both dgenerate_installer.zip.001
and dgenerate_installer.zip.002
to a folder.
Unzip dgenerate_installer.zip.001
to a directory (Right click, 7-Zip -> Extract to "dgenerate_installer") and then run dgenerate_installer\dgenerate.msi
to install.
dgenerate will be installed under C:\Program Files\dgenerate
by default with an isolated python environment provided.
The install directory will be added to PATH, and dgenerate will be available from the command line.
Portable Install
A portable install is provided via dgenerate_portable.zip.001
and dgenerate_portable.zip.002
, these contain
nothing but the dgenerate executable and a frozen python environment which can be placed anywhere.
4.2.2 Patch Fix
Backport fixes from: https://github.com/Teriks/dgenerate/tree/flux_newpipes
Fix optional arguments: --sdxl-refiner-inference-steps
, --sdxl-refiner-guidance-scales
, and --sdxl-refiner-guidance-rescales
not being considered when calculating the number of generation steps displayed to the user.
Minor fix for example runner script, inform user when running from an environment where dgenerate is not accessible as an importable package. In this case, dgenerate will run in a subprocess by invoking the command in a shell if it is available.
4.2.0 Features
-
Basic inpainting is now supported with Stable Diffusion 3
-
Config directive
\setp
can be used to evaluate python ternary statements during template variable assignment. i.e.\setp my_var "Yes" if True else "No"
-
Improved config error reporting. 1 indexed line number in the interpreter to align with Console UI line numbering. Better pinpointing of the error line number inside "here template" control structures involving Jinja. Patches for issues with
asteval
affecting syntax error output. -
PyPI Packaging now supports installing dgenerate with the ROCm torch backend on Linux, allowing for use with AMD Cards / Accelerators. See installation documentation: Linux with ROCm (AMD Cards)
v4.2.1 (Stable Diffusion 3 Inpainting, ROCm (AMD) support)
v4.2.1 release with Windows installer.
Due to the size of the packaged python environment, the installer is within a multi-part zip file.
The multipart zip can be extracted using 7-Zip: https://www.7-zip.org/
Download both dgenerate_installer.zip.001
and dgenerate_installer.zip.002
to a folder.
Unzip dgenerate_installer.zip.001
to a directory (Right click, 7-Zip -> Extract to "dgenerate_installer") and then run dgenerate_installer\dgenerate.msi
to install.
dgenerate will be installed under C:\Program Files\dgenerate
by default with an isolated python environment provided.
The install directory will be added to PATH, and dgenerate will be available from the command line.
Portable Install
A portable install is provided via dgenerate_portable.zip.001
and dgenerate_portable.zip.002
, these contain
nothing but the dgenerate executable and a frozen python environment which can be placed anywhere.
4.2.1 Patch Fix
Upgrade asteval
to include error handling fixes from new release, instead of monkey patching it inside dgenerate
.
4.2.0 Features
-
Basic inpainting is now supported with Stable Diffusion 3
-
Config directive
\setp
can be used to evaluate python ternary statements during template variable assignment. i.e.\setp my_var "Yes" if True else "No"
-
Improved config error reporting. 1 indexed line number in the interpreter to align with Console UI line numbering. Better pinpointing of the error line number inside "here template" control structures involving Jinja. Patches for issues with
asteval
affecting syntax error output. -
PyPI Packaging now supports installing dgenerate with the ROCm torch backend on Linux, allowing for use with AMD Cards / Accelerators. See installation documentation: Linux with ROCm (AMD Cards)
v4.2.0 (Stable Diffusion 3 Inpainting, ROCm (AMD) support)
v4.2.0 release with Windows installer.
Due to the size of the packaged python environment, the installer is within a multi-part zip file.
The multipart zip can be extracted using 7-Zip: https://www.7-zip.org/
Download both dgenerate_installer.zip.001
and dgenerate_installer.zip.002
to a folder.
Unzip dgenerate_installer.zip.001
to a directory (Right click, 7-Zip -> Extract to "dgenerate_installer") and then run dgenerate_installer\dgenerate.msi
to install.
dgenerate will be installed under C:\Program Files\dgenerate
by default with an isolated python environment provided.
The install directory will be added to PATH, and dgenerate will be available from the command line.
Portable Install
A portable install is provided via dgenerate_portable.zip.001
and dgenerate_portable.zip.002
, these contain
nothing but the dgenerate executable and a frozen python environment which can be placed anywhere.
4.2.0 Features
-
Basic inpainting is now supported with Stable Diffusion 3
-
Config directive
\setp
can be used to evaluate python ternary statements during template variable assignment. i.e.\setp my_var "Yes" if True else "No"
-
Improved config error reporting. 1 indexed line number in the interpreter to align with Console UI line numbering. Better pinpointing of the error line number inside "here template" control structures involving Jinja. Patches for issues with
asteval
affecting syntax error output. -
PyPI Packaging now supports installing dgenerate with the ROCm torch backend on Linux, allowing for use with AMD Cards / Accelerators. See installation documentation: Linux with ROCm (AMD Cards)
v4.1.2 (Stable Cascade prompt weighting, MacOS arm64 support)
v4.1.2 release with Windows installer.
Due to the size of the packaged python environment, the installer is within a multi-part zip file.
The multipart zip can be extracted using 7-Zip: https://www.7-zip.org/
Download both dgenerate_installer.zip.001
and dgenerate_installer.zip.002
to a folder.
Unzip dgenerate_installer.zip.001
to a directory (Right click, 7-Zip -> Extract to "dgenerate_installer") and then run dgenerate_installer\dgenerate.msi
to install.
dgenerate will be installed under C:\Program Files\dgenerate
by default with an isolated python environment provided.
The install directory will be added to PATH, and dgenerate will be available from the command line.
Portable Install
A portable install is provided via dgenerate_portable.zip.001
and dgenerate_portable.zip.002
, these contain
nothing but the dgenerate executable and a frozen python environment which can be placed anywhere.
4.1.2 Patch
HuggingFace repository file link for transformer involved in Flux (Schnell quantized)
console recipe had moved. Updated link.
Better error message when a HuggingFace repository is not accessible due to lack of authentication.
Console output window Clear
right click context menu functionality bug fix. Output window started in editable state, it should be disabled for text input and only enabled temporarily for writes internally. Improper state management resulted in inability to write new output to the window after using Clear
from the UI menu, as the text box was left in a disabled UI state.
spandrel 0.4.0
4.1.0 Features
-
Stable Cascade can now use long prompts with weighting via the
compel
andsd-embed
prompt weighters, simply specify one of these two values to--prompt-weighter
to enable the feature when using Stable Cascade. -
Built and published wheels for MacOS on Apple Silicon, dgenerate can now be used with
--device mps
on MacOS (arm64). See install instructions here: MacOS Install (Apple Silicon Only). This is very experimental. -
Console UI hotkeys dynamically adjust for useability when the UI is running on MacOS, I am sure that the hotkeys are very abnormal when compared to typical MacOS convention for text editors, but they all function and do not overlap with important system hotkeys.
Insert
(for multiline input) changes toCommand+i
, andCtrl+Space
(for run) changes toCommand+r
. All other hotkeys remain the same, including copy, paste, and cut. -
Added template function
frange
which is justrange
but for float ranges. This is useful for stepping through guidance scale values or LoRA scale values for example. It uses identical logic torange
, but the step value may be a float, e.g.frange(0.1, 1.0, 0.2) -> [0.1, 0.3, 0.5, 0.7, 0.9]
. The default step value is:0.1
-
torch 2.4.1
v4.1.1 (Stable Cascade prompt weighting, MacOS arm64 support)
v4.1.1 release with Windows installer.
Due to the size of the packaged python environment, the installer is within a multi-part zip file.
The multipart zip can be extracted using 7-Zip: https://www.7-zip.org/
Download both dgenerate_installer.zip.001
and dgenerate_installer.zip.002
to a folder.
Unzip dgenerate_installer.zip.001
to a directory (Right click, 7-Zip -> Extract to "dgenerate_installer") and then run dgenerate_installer\dgenerate.msi
to install.
dgenerate will be installed under C:\Program Files\dgenerate
by default with an isolated python environment provided.
The install directory will be added to PATH, and dgenerate will be available from the command line.
Portable Install
A portable install is provided via dgenerate_portable.zip.001
and dgenerate_portable.zip.002
, these contain
nothing but the dgenerate executable and a frozen python environment which can be placed anywhere.
4.1.1 Patch Fix
The Stable Cascade Decoder and SDXL Refiner now get placed on the correct device, they were defaulting to cuda
instead of honoring --device
. This became apparent on MacOS where they were attempting to use CUDA instead of MPS causing an exception.
4.1.0 Features
-
Stable Cascade can now use long prompts with weighting via the
compel
andsd-embed
prompt weighters, simply specify one of these two values to--prompt-weighter
to enable the feature when using Stable Cascade. -
Built and published wheels for MacOS on Apple Silicon, dgenerate can now be used with
--device mps
on MacOS (arm64). See install instructions here: MacOS Install (Apple Silicon Only). This is very experimental. -
Console UI hotkeys dynamically adjust for useability when the UI is running on MacOS, I am sure that the hotkeys are very abnormal when compared to typical MacOS convention for text editors, but they all function and do not overlap with important system hotkeys.
Insert
(for multiline input) changes toCommand+i
, andCtrl+Space
(for run) changes toCommand+r
. All other hotkeys remain the same, including copy, paste, and cut. -
Added template function
frange
which is justrange
but for float ranges. This is useful for stepping through guidance scale values or LoRA scale values for example. It uses identical logic torange
, but the step value may be a float, e.g.frange(0.1, 1.0, 0.2) -> [0.1, 0.3, 0.5, 0.7, 0.9]
. The default step value is:0.1
-
torch 2.4.1
v4.1.0 (Stable Cascade prompt weighting, MacOS arm64 support)
v4.1.0 release with Windows installer.
Due to the size of the packaged python environment, the installer is within a multi-part zip file.
The multipart zip can be extracted using 7-Zip: https://www.7-zip.org/
Download both dgenerate_installer.zip.001
and dgenerate_installer.zip.002
to a folder.
Unzip dgenerate_installer.zip.001
to a directory (Right click, 7-Zip -> Extract to "dgenerate_installer") and then run dgenerate_installer\dgenerate.msi
to install.
dgenerate will be installed under C:\Program Files\dgenerate
by default with an isolated python environment provided.
The install directory will be added to PATH, and dgenerate will be available from the command line.
Portable Install
A portable install is provided via dgenerate_portable.zip.001
and dgenerate_portable.zip.002
, these contain
nothing but the dgenerate executable and a frozen python environment which can be placed anywhere.
4.1.0 Features
-
Stable Cascade can now use long prompts with weighting via the
compel
andsd-embed
prompt weighters, simply specify one of these two values to--prompt-weighter
to enable the feature when using Stable Cascade. -
Built and published wheels for MacOS on Apple Silicon, dgenerate can now be used with
--device mps
on MacOS (arm64). See install instructions here: MacOS Install (Apple Silicon Only). This is very experimental. -
Console UI hotkeys dynamically adjust for useability when the UI is running on MacOS, I am sure that the hotkeys are very abnormal when compared to typical MacOS convention for text editors, but they all function and do not overlap with important system hotkeys.
Insert
(for multiline input) changes toCommand+i
, andCtrl+Space
(for run) changes toCommand+r
. All other hotkeys remain the same, including copy, paste, and cut. -
Added template function
frange
which is justrange
but for float ranges. This is useful for stepping through guidance scale values or LoRA scale values for example. It uses identical logic torange
, but the step value may be a float, e.g.frange(0.1, 1.0, 0.2) -> [0.1, 0.3, 0.5, 0.7, 0.9]
. The default step value is:0.1
-
torch 2.4.1
v4.0.3 (Flux, IP Adapters, T2I Adapters, Input Batching)
v4.0.3 release with Windows installer.
Due to the size of the packaged python environment, the installer is within a multi-part zip file.
The multipart zip can be extracted using 7-Zip: https://www.7-zip.org/
Download both dgenerate_installer.zip.001
and dgenerate_installer.zip.002
to a folder.
Unzip dgenerate_installer.zip.001
to a directory (Right click, 7-Zip -> Extract to "dgenerate_installer") and then run dgenerate_installer\dgenerate.msi
to install.
dgenerate will be installed under C:\Program Files\dgenerate
by default with an isolated python environment provided.
The install directory will be added to PATH, and dgenerate will be available from the command line.
Portable Install
A portable install is provided via dgenerate_portable.zip.001
and dgenerate_portable.zip.002
, these contain
nothing but the dgenerate executable and a frozen python environment which can be placed anywhere.
4.0.3 Patch Fix
runwayml has removed their stable diffusion 1.5 repository from Hugging Face, breaking single file loads in the diffusers library and anything that uses it.
This release upgrades diffusers to 0.30.2 where the issue has been worked around by instead referencing Lykon/dreamshaper-8
All references to the runwayml model on Hugging Face have been removed from this project, including in documentation.
4.0.0 Features
The command line interface remains backwards compatible for all supported model types, the major version change is mostly due to breaking library API changes and the complete removal of Flax/Jax support.
All support for Flax/Jax has been removed because there is only minimal support or adoption of it anywhere currently, and seemingly no work being done on it in the diffusers library. Also this alleviates quite a bit of dependency hell and possible issues caused by having torch and Flax/Jax in the same package as dependencies.
-
Support for Flux in
txt2img
mode, add associated options:--flux-second-prompts
,--flux-max-sequence-length
. Single file loads from a.safetensor
file is partially supported for Flux, you must specify the VAE and Text Encoders manually. -
Add Flux support to
sd-embed
prompt weighter internal plugin, prompt weighting syntax can be used with Flux when--prompt-weighter sd-embed
is specified. -
New Console UI recipes for Flux that are optimized by default to run (slowly) on even old hardware or GPUs with less than 12gb of VRAM.
-
Stable Diffusion 3 now supports per LoRA scale values.
-
The option
--lora-fuse-scale
has been added to control the LoRA scale once all LoRA models have been fused together at their individual scale values. This controls at what scale the merged LoRA weights are fused into the main model, these weights possibly being merged from multiple LoRA models at different individual scales. -
Support loading a transformer model for new diffusion architectures based on transformer architecture (SD3 and Flux) via the argument
--transformer
, this new option is nearly identical to--unet/--unet2
. Transformers can be quantized via thequantize
URI argument usingoptimum.quanto
. The quantize argument accepts a string value which is the datatype for quantization, for exampleqint8
,qfloat8
, etc. Acceptable values are listed in the--help
text of the--transformer
argument. -
Text encoders specified with
--text-encoders/--text-encoders2
can now be quantized via thequantize
URI argument usingoptimum.quanto
. Acceptable values are listed in the--help
text of the--text-encoders
argument. -
Support loading multiple IP Adapter models with
--ip-adapters
. -
Support loading an image encoder with
--image-encoder
for use with IP Adapters and Stable Cascade. -
New
--image-seeds
syntax for specifying IP Adapter images, see: Specifying IP Adapters -
Support loading multiple T2I Adapter models with
--t2i-adapters
, T2I Adapters function identically to Control Nets in terms of image input specification with--image-seeds
. They are essentially another form of control guidance model, and support Open Pose / Depth inputs etc. similar to Control Nets. T2I Adapters are a bit less resource intensive, but do not currently supportimg2img
orinpainting
, onlytxt2img
. -
Support image input batching for
img2img
andinpainting
mask images, allowing multiple input images to be processed on the GPU at once. New--image-seeds
syntax feature added to support this, see: Batching Input Images and Inpaint Masks -
--seed-image-processors
and--mask-image-processors
now support the + syntax previously supported by--control-image-processors
in order to allow for selective processing of input images when multipleimg2img
orinpaint
masks are specified. This is covered in the Image Processors section: Multiple control net images, and input image batching -
Added the
resize
image processor which implements basic image resizing using the usual resampling algorithms in Pillow. This can be useful for resizing input images to Stable Cascade when multiple are provided, as there is not a mechanism to do so from--image-seeds
-
Stable Cascade can now use the new input batching syntax of
--image-seeds
to process multiple input images as a style reference, Stable Cascade does not use multiple images for batching like other pipelines, but as image prompts. -
Improved
--help
output formatting and readability fordgenerate
anddgenerate --sub-command image-process
-
torch 2.4.0 and diffusers 0.30.0
v4.0.2 (Flux, IP Adapters, T2I Adapters, Input Batching)
v4.0.2 release with Windows installer.
Due to the size of the packaged python environment, the installer is within a multi-part zip file.
The multipart zip can be extracted using 7-Zip: https://www.7-zip.org/
Download both dgenerate_installer.zip.001
and dgenerate_installer.zip.002
to a folder.
Unzip dgenerate_installer.zip.001
to a directory (Right click, 7-Zip -> Extract to "dgenerate_installer") and then run dgenerate_installer\dgenerate.msi
to install.
dgenerate will be installed under C:\Program Files\dgenerate
by default with an isolated python environment provided.
The install directory will be added to PATH, and dgenerate will be available from the command line.
Portable Install
A portable install is provided via dgenerate_portable.zip.001
and dgenerate_portable.zip.002
, these contain
nothing but the dgenerate executable and a frozen python environment which can be placed anywhere.
4.0.2 Patch Fix
Add the arguments --flux-second-prompts
, --flux-max-sequence-length
, and --lora-fuse-scale
to the argument
reconstruction used by --output-configs
and --output-metadata
. These arguments were not being replicated in config output.
4.0.0 Features
The command line interface remains backwards compatible for all supported model types, the major version change is mostly due to breaking library API changes and the complete removal of Flax/Jax support.
All support for Flax/Jax has been removed because there is only minimal support or adoption of it anywhere currently, and seemingly no work being done on it in the diffusers library. Also this alleviates quite a bit of dependency hell and possible issues caused by having torch and Flax/Jax in the same package as dependencies.
-
Support for Flux in
txt2img
mode, add associated options:--flux-second-prompts
,--flux-max-sequence-length
. Single file loads from a.safetensor
file is partially supported for Flux, you must specify the VAE and Text Encoders manually. -
Add Flux support to
sd-embed
prompt weighter internal plugin, prompt weighting syntax can be used with Flux when--prompt-weighter sd-embed
is specified. -
New Console UI recipes for Flux that are optimized by default to run (slowly) on even old hardware or GPUs with less than 12gb of VRAM.
-
Stable Diffusion 3 now supports per LoRA scale values.
-
The option
--lora-fuse-scale
has been added to control the LoRA scale once all LoRA models have been fused together at their individual scale values. This controls at what scale the merged LoRA weights are fused into the main model, these weights possibly being merged from multiple LoRA models at different individual scales. -
Support loading a transformer model for new diffusion architectures based on transformer architecture (SD3 and Flux) via the argument
--transformer
, this new option is nearly identical to--unet/--unet2
. Transformers can be quantized via thequantize
URI argument usingoptimum.quanto
. The quantize argument accepts a string value which is the datatype for quantization, for exampleqint8
,qfloat8
, etc. Acceptable values are listed in the--help
text of the--transformer
argument. -
Text encoders specified with
--text-encoders/--text-encoders2
can now be quantized via thequantize
URI argument usingoptimum.quanto
. Acceptable values are listed in the--help
text of the--text-encoders
argument. -
Support loading multiple IP Adapter models with
--ip-adapters
. -
Support loading an image encoder with
--image-encoder
for use with IP Adapters and Stable Cascade. -
New
--image-seeds
syntax for specifying IP Adapter images, see: Specifying IP Adapters -
Support loading multiple T2I Adapter models with
--t2i-adapters
, T2I Adapters function identically to Control Nets in terms of image input specification with--image-seeds
. They are essentially another form of control guidance model, and support Open Pose / Depth inputs etc. similar to Control Nets. T2I Adapters are a bit less resource intensive, but do not currently supportimg2img
orinpainting
, onlytxt2img
. -
Support image input batching for
img2img
andinpainting
mask images, allowing multiple input images to be processed on the GPU at once. New--image-seeds
syntax feature added to support this, see: Batching Input Images and Inpaint Masks -
--seed-image-processors
and--mask-image-processors
now support the + syntax previously supported by--control-image-processors
in order to allow for selective processing of input images when multipleimg2img
orinpaint
masks are specified. This is covered in the Image Processors section: Multiple control net images, and input image batching -
Added the
resize
image processor which implements basic image resizing using the usual resampling algorithms in Pillow. This can be useful for resizing input images to Stable Cascade when multiple are provided, as there is not a mechanism to do so from--image-seeds
-
Stable Cascade can now use the new input batching syntax of
--image-seeds
to process multiple input images as a style reference, Stable Cascade does not use multiple images for batching like other pipelines, but as image prompts. -
Improved
--help
output formatting and readability fordgenerate
anddgenerate --sub-command image-process
-
torch 2.4.0 and diffusers 0.30.0
v4.0.1 (Flux, IP Adapters, T2I Adapters, Input Batching)
v4.0.1 release with Windows installer.
Due to the size of the packaged python environment, the installer is within a multi-part zip file.
The multipart zip can be extracted using 7-Zip: https://www.7-zip.org/
Download both dgenerate_installer.zip.001
and dgenerate_installer.zip.002
to a folder.
Unzip dgenerate_installer.zip.001
to a directory (Right click, 7-Zip -> Extract to "dgenerate_installer") and then run dgenerate_installer\dgenerate.msi
to install.
dgenerate will be installed under C:\Program Files\dgenerate
by default with an isolated python environment provided.
The install directory will be added to PATH, and dgenerate will be available from the command line.
Portable Install
A portable install is provided via dgenerate_portable.zip.001
and dgenerate_portable.zip.002
, these contain
nothing but the dgenerate executable and a frozen python environment which can be placed anywhere.
4.0.1 Patch Fix
Add missing VAE tiling checkbox for Flux recipes in Console UI recipe form.
4.0.0 Features
The command line interface remains backwards compatible for all supported model types, the major version change is mostly due to breaking library API changes and the complete removal of Flax/Jax support.
All support for Flax/Jax has been removed because there is only minimal support or adoption of it anywhere currently, and seemingly no work being done on it in the diffusers library. Also this alleviates quite a bit of dependency hell and possible issues caused by having torch and Flax/Jax in the same package as dependencies.
-
Support for Flux in
txt2img
mode, add associated options:--flux-second-prompts
,--flux-max-sequence-length
. Single file loads from a.safetensor
file is partially supported for Flux, you must specify the VAE and Text Encoders manually. -
Add Flux support to
sd-embed
prompt weighter internal plugin, prompt weighting syntax can be used with Flux when--prompt-weighter sd-embed
is specified. -
New Console UI recipes for Flux that are optimized by default to run (slowly) on even old hardware or GPUs with less than 12gb of VRAM.
-
Stable Diffusion 3 now supports per LoRA scale values.
-
The option
--lora-fuse-scale
has been added to control the LoRA scale once all LoRA models have been fused together at their individual scale values. This controls at what scale the merged LoRA weights are fused into the main model, these weights possibly being merged from multiple LoRA models at different individual scales. -
Support loading a transformer model for new diffusion architectures based on transformer architecture (SD3 and Flux) via the argument
--transformer
, this new option is nearly identical to--unet/--unet2
. Transformers can be quantized via thequantize
URI argument usingoptimum.quanto
. The quantize argument accepts a string value which is the datatype for quantization, for exampleqint8
,qfloat8
, etc. Acceptable values are listed in the--help
text of the--transformer
argument. -
Text encoders specified with
--text-encoders/--text-encoders2
can now be quantized via thequantize
URI argument usingoptimum.quanto
. Acceptable values are listed in the--help
text of the--text-encoders
argument. -
Support loading multiple IP Adapter models with
--ip-adapters
. -
Support loading an image encoder with
--image-encoder
for use with IP Adapters and Stable Cascade. -
New
--image-seeds
syntax for specifying IP Adapter images, see: Specifying IP Adapters -
Support loading multiple T2I Adapter models with
--t2i-adapters
, T2I Adapters function identically to Control Nets in terms of image input specification with--image-seeds
. They are essentially another form of control guidance model, and support Open Pose / Depth inputs etc. similar to Control Nets. T2I Adapters are a bit less resource intensive, but do not currently supportimg2img
orinpainting
, onlytxt2img
. -
Support image input batching for
img2img
andinpainting
mask images, allowing multiple input images to be processed on the GPU at once. New--image-seeds
syntax feature added to support this, see: Batching Input Images and Inpaint Masks -
--seed-image-processors
and--mask-image-processors
now support the + syntax previously supported by--control-image-processors
in order to allow for selective processing of input images when multipleimg2img
orinpaint
masks are specified. This is covered in the Image Processors section: Multiple control net images, and input image batching -
Added the
resize
image processor which implements basic image resizing using the usual resampling algorithms in Pillow. This can be useful for resizing input images to Stable Cascade when multiple are provided, as there is not a mechanism to do so from--image-seeds
-
Stable Cascade can now use the new input batching syntax of
--image-seeds
to process multiple input images as a style reference, Stable Cascade does not use multiple images for batching like other pipelines, but as image prompts. -
Improved
--help
output formatting and readability fordgenerate
anddgenerate --sub-command image-process
-
torch 2.4.0 and diffusers 0.30.0