hunyuan-output.mp4
This release took a while, but it has many exciting updates. It contains several new pipelines for image and video generation, new quantization backends, and more.
Going forward, to provide more transparency to the community about ongoing developments and releases in Diffusers, we will be making use of a roadmap tracker.
New Video Generation Pipelines 📹
Open video generation models are on the rise, and we’re pleased to provide comprehensive integration support for all of them. The following video pipelines are bundled in this release:
Check out this section to learn more about the fine-tuning options available for these new video models.
New Image Generation Pipelines
- SANA
- Flux Control (including Control LoRA)
- Flux Redux
- Flux Fill Inpainting / Outpainting
- Flux RF-Inversion
- SD3.5 ControlNet
- ControlNet Union XL
- SD3.5 IP Adapter
- Flux IP adapter
Important Note about the new Flux Models
We can combine the regular Flux.1 Dev LoRAs with Flux Control LoRAs, Flux Control, and Flux Fill. For example, you can enable few-steps inference with Flux Fill using:
from diffusers import FluxFillPipeline
from diffusers.utils import load_image
import torch
pipe = FluxFillPipeline.from_pretrained(
"black-forest-labs/FLUX.1-Fill-dev", torch_dtype=torch.bfloat16
).to("cuda")
adapter_id = "alimama-creative/FLUX.1-Turbo-Alpha"
pipe.load_lora_weights(adapter_id)
image = load_image("https://huggingface.co/datasets/diffusers/diffusers-images-docs/resolve/main/cup.png")
mask = load_image("https://huggingface.co/datasets/diffusers/diffusers-images-docs/resolve/main/cup_mask.png")
image = pipe(
prompt="a white paper cup",
image=image,
mask_image=mask,
height=1632,
width=1232,
guidance_scale=30,
num_inference_steps=8,
max_sequence_length=512,
generator=torch.Generator("cpu").manual_seed(0)
).images[0]
image.save("flux-fill-dev.png")
To learn more, check out the documentation.
Note
SANA is a small model compared to other models like Flux and Sana-0.6B can be deployed on a 16GB laptop GPU, taking less than 1 second to generate a 1024×1024 resolution image. We support LoRA fine-tuning of SANA. Check out this section for more details.
Acknowledgements
- Shoutout to @lawrence-cj and @chenjy2003 for contributing SANA in this PR. SANA also features a Deep Compression Autoencoder, which was contributed by @lawrence-cj in this PR.
- Shoutout to @guiyrt for contributing SD3.5 IP Adapter in this PR.
New Quantization Backends
Please be aware of the following caveats:
- TorchAO quantized checkpoints cannot be serialized in
safetensors
currently. This may change in the future. - GGUF currently only supports loading pre-quantized checkpoints into models in this release. Support for saving models with GGUF quantization will be added in the future.
New training scripts
This release features many new training scripts for the community to play:
All commits
- post-release 0.31.0 by @sayakpaul in #9742
- fix bug in
require_accelerate_version_greater
by @faaany in #9746 - [Official callbacks] SDXL Controlnet CFG Cutoff by @asomoza in #9311
- [SD3-5 dreambooth lora] update model cards by @linoytsaban in #9749
- config attribute not foud error for FluxImagetoImage Pipeline for multi controlnet solved by @rshah240 in #9586
- Some minor updates to the nightly and push workflows by @sayakpaul in #9759
- [Docs] fix docstring typo in SD3 pipeline by @shenzhiy21 in #9765
- [bugfix] bugfix for npu free memory by @leisuzz in #9640
- [research_projects] add flux training script with quantization by @sayakpaul in #9754
- Add a doc for AWS Neuron in Diffusers by @JingyaHuang in #9766
- [refactor] enhance readability of flux related pipelines by @Luciennnnnnn in #9711
- Added Support of Xlabs controlnet to FluxControlNetInpaintPipeline by @SahilCarterr in #9770
- [research_projects] Update README.md to include a note about NF5 T5-xxl by @sayakpaul in #9775
- [Fix] train_dreambooth_lora_flux_advanced ValueError: unexpected save model: <class 'transformers.models.t5.modeling_t5.T5EncoderModel'> by @rootonchair in #9777
- [Fix] remove setting lr for T5 text encoder when using prodigy in flux dreambooth lora script by @biswaroop1547 in #9473
- [SD 3.5 Dreambooth LoRA] support configurable training block & layers by @linoytsaban in #9762
- [flux dreambooth lora training] make LoRA target modules configurable + small bug fix by @linoytsaban in #9646
- adds the pipeline for pixart alpha controlnet by @raulc0399 in #8857
- [core] Allegro T2V by @a-r-r-o-w in #9736
- Allegro VAE fix by @a-r-r-o-w in #9811
- [CI] add new runner for testing by @sayakpaul in #9699
- [training] fixes to the quantization training script and add AdEMAMix optimizer as an option by @sayakpaul in #9806
- [training] use the lr when using 8bit adam. by @sayakpaul in #9796
- [Tests] clean up and refactor gradient checkpointing tests by @sayakpaul in #9494
- [CI] add a big GPU marker to run memory-intensive tests separately on CI by @sayakpaul in #9691
- [LoRA] fix: lora loading when using with a device_mapped model. by @sayakpaul in #9449
- Revert "[LoRA] fix: lora loading when using with a device_mapped mode… by @yiyixuxu in #9823
- [Model Card] standardize advanced diffusion training sd15 lora by @chiral-carbon in #7613
- NPU Adaption for FLUX by @leisuzz in #9751
- Fixes EMAModel "from_pretrained" method by @SahilCarterr in #9779
- Update train_controlnet_flux.py,Fix size mismatch issue in validation by @ScilenceForest in #9679
- Handling mixed precision for dreambooth flux lora training by @icsl-Jeon in #9565
- Reduce Memory Cost in Flux Training by @leisuzz in #9829
- Add Diffusion Policy for Reinforcement Learning by @DorsaRoh in #9824
- [feat] add
load_lora_adapter()
for compatible models by @sayakpaul in #9712 - Refac training utils.py by @RogerSinghChugh in #9815
- [core] Mochi T2V by @a-r-r-o-w in #9769
- [Fix] Test of sd3 lora by @SahilCarterr in #9843
- Fix: Remove duplicated comma in distributed_inference.md by @vahidaskari in #9868
- Add new community pipeline for 'Adaptive Mask Inpainting', introduced in [ECCV2024] ComA by @jellyheadandrew in #9228
- Updated _encode_prompt_with_clip and encode_prompt in train_dreamboth_sd3 by @SahilCarterr in #9800
- [Core] introduce
controlnet
module by @sayakpaul in #8768 - [Flux] reduce explicit device transfers and typecasting in flux. by @sayakpaul in #9817
- Improve downloads of sharded variants by @DN6 in #9869
- [fix] Replaced shutil.copy with shutil.copyfile by @SahilCarterr in #9885
- Enabling gradient checkpointing in eval() mode by @MikeTkachuk in #9878
- [FIX] Fix TypeError in DreamBooth SDXL when use_dora is False by @SahilCarterr in #9879
- [Advanced LoRA v1.5] fix: gradient unscaling problem by @sayakpaul in #7018
- Revert "[Flux] reduce explicit device transfers and typecasting in flux." by @sayakpaul in #9896
- Feature IP Adapter Xformers Attention Processor by @elismasilva in #9881
- Notebooks for Community Scripts Examples by @ParagEkbote in #9905
- Fix Progress Bar Updates in SD 1.5 PAG Img2Img pipeline by @painebenjamin in #9925
- Update pipeline_flux_img2img.py by @example-git in #9928
- add depth controlnet sd3 pre-trained checkpoints to docs by @pureexe in #9937
- Move Wuerstchen Dreambooth to research_projects by @ParagEkbote in #9935
- Update ip_adapter.py by @mkknightr in #8882
- Modify apply_overlay for inpainting with padding_mask_crop (Inpainting area: "Only Masked") by @clarkkent0618 in #8793
- Correct pipeline_output.py to the type Mochi by @twobob in #9945
- Add all AttnProcessor classes in
AttentionProcessor
type by @Prgckwb in #9909 - Fixed Nits in Docs and Example Script by @ParagEkbote in #9940
- Add server example by @thealmightygrant in #9918
- CogVideoX 1.5 by @zRzRzRzRzRzRzR in #9877
- Notebooks for Community Scripts-2 by @ParagEkbote in #9952
- [advanced flux training] bug fix + reduce memory cost as in #9829 by @linoytsaban in #9838
- [LoRA] feat:
save_lora_adapter()
by @sayakpaul in #9862 - Make CogVideoX RoPE implementation consistent by @a-r-r-o-w in #9963
- [CI] Unpin torch<2.5 in CI by @DN6 in #9961
- Move IP Adapter Scripts to research project by @ParagEkbote in #9960
- add skip_layers argument to SD3 transformer model class by @bghira in #9880
- Fix beta and exponential sigmas + add tests by @hlky in #9954
- Flux latents fix by @DN6 in #9929
- [LoRA] enable LoRA for Mochi-1 by @sayakpaul in #9943
- Improve control net block index for sd3 by @linjiapro in #9758
- Update handle single blocks on _convert_xlabs_flux_lora_to_diffusers by @raulmosa in #9915
- fix controlnet module refactor by @yiyixuxu in #9968
- Fix prepare latent image ids and vae sample generators for flux by @a-r-r-o-w in #9981
- [Tests] skip nan lora tests on PyTorch 2.5.1 CPU. by @sayakpaul in #9975
- make
pipelines
tests device-agnostic (part1) by @faaany in #9399 - ControlNet from_single_file when already converted by @hlky in #9978
- Flux Fill, Canny, Depth, Redux by @a-r-r-o-w in #9985
- [SD3 dreambooth lora] smol fix to checkpoint saving by @linoytsaban in #9993
- [Docs] add: missing pipelines from the spec. by @sayakpaul in #10005
- Add prompt about wandb in examples/dreambooth/readme. by @SkyCol in #10014
- [docs] Fix CogVideoX table by @a-r-r-o-w in #10008
- Notebooks for Community Scripts-3 by @ParagEkbote in #10032
- Sd35 controlnet by @yiyixuxu in #10020
- Add
beta
,exponential
andkarras
sigmas toFlowMatchEulerDiscreteScheduler
by @hlky in #10001 - Update sdxl reference pipeline to latest sdxl pipeline by @dimitribarbot in #9938
- [Community Pipeline] Add some feature for regional prompting pipeline by @cjkangme in #9874
- Add sdxl controlnet reference community pipeline by @dimitribarbot in #9893
- Change image_gen_aux repository URL by @asomoza in #10048
- make
pipelines
tests device-agnostic (part2) by @faaany in #9400 - [Mochi-1] ensuring to compute the fourier features in FP32 in Mochi encoder by @sayakpaul in #10031
- [Fix] Syntax error by @SahilCarterr in #10068
- [CI] Add quantization by @sayakpaul in #9832
- Add
sigmas
to Flux pipelines by @hlky in #10081 - Fixed Nits in Evaluation Docs by @ParagEkbote in #10063
- fix link in the docs by @coding-famer in #10058
- fix offloading for sd3.5 controlnets by @yiyixuxu in #10072
- [Single File] Fix SD3.5 single file loading by @DN6 in #10077
- Fix
num_images_per_prompt>1
with Skip Guidance Layers inStableDiffusion3Pipeline
by @hlky in #10086 - [Single File] Pass token when fetching interpreted config by @DN6 in #10082
- Interpolate fix on cuda for large output tensors by @pcuenca in #10067
- Convert
sigmas
tonp.array
in FlowMatch set_timesteps by @hlky in #10088 - fix: missing AutoencoderKL lora adapter by @beniz in #9807
- Let server decide default repo visibility by @Wauplin in #10047
- Fix some documentation in ./src/diffusers/models/embeddings.py for demo by @DTG2005 in #9579
- Don't stale close-to-merge by @pcuenca in #10096
- Add StableDiffusion3PAGImg2Img Pipeline + Fix SD3 Unconditional PAG by @painebenjamin in #9932
- Notebooks for Community Scripts-4 by @ParagEkbote in #10094
- Fix Broken Link in Optimization Docs by @ParagEkbote in #10105
- DPM++ third order fixes by @StAlKeR7779 in #9104
- update by @aihao2000 in #7067
- Avoid compiling a progress bar. by @lsb in #10098
- [Bug fix] "previous_timestep()" in DDPM scheduling compatible with "trailing" and "linspace" options by @AnandK27 in #9384
- Fix multi-prompt inference by @hlky in #10103
- Test
skip_guidance_layers
in SD3 pipeline by @hlky in #10102 - Use parameters + buffers when deciding upscale_dtype by @universome in #9882
- [tests] refactor vae tests by @sayakpaul in #9808
- add torch_xla support in pipeline_stable_audio.py by @ in #10109
- Fix
pipeline_stable_audio
formating by @hlky in #10114 - [bitsandbytes] allow directly CUDA placements of pipelines loaded with bnb components by @sayakpaul in #9840
- Fix Broken Links in ReadMe by @ParagEkbote in #10117
- Add
sigmas
to pipelines using FlowMatch by @hlky in #10116 - [Flux Redux] add prompt & multiple image input by @linoytsaban in #10056
- Fix a bug in the state dict judgment in ip_adapter.py. by @zhangp365 in #10095
- Fix a bug for SD35 control net training and improve control net block index by @linjiapro in #10065
- pass attn mask arg for flux by @yiyixuxu in #10122
- [docs] load_lora_adapter by @stevhliu in #10119
- Use torch.device instead of current device index for BnB quantizer by @a-r-r-o-w in #10069
- [Tests] fix condition argument in xfail. by @sayakpaul in #10099
- [Tests] xfail incompatible SD configs. by @sayakpaul in #10127
- [FIX] Bug in FluxPosEmbed by @SahilCarterr in #10115
- [Guide] Quantize your Diffusion Models with
bnb
by @ariG23498 in #10012 - Remove duplicate checks for len(generator) != batch_size when generator is a list by @a-r-r-o-w in #10134
- [community] Load Models from Sources like
Civitai
into Existing Pipelines by @suzukimain in #9986 - [DC-AE] Add the official Deep Compression Autoencoder code(32x,64x,128x compression ratio); by @lawrence-cj in #9708
- fixed a dtype bfloat16 bug in torch_utils.py by @zhangp365 in #10125
- [LoRA] depcrecate save_attn_procs(). by @sayakpaul in #10126
- Update ptxla training by @entrpn in #9864
- support sd3.5 for controlnet example by @DavyMorgan in #9860
- [Single file] Support
revision
argument when loading single file config by @a-r-r-o-w in #10168 - [community pipeline] Add RF-inversion Flux pipeline by @linoytsaban in #9816
- Improve post-processing performance by @soof-golan in #10170
- Use
torch
inget_3d_rotary_pos_embed
/_allegro
by @hlky in #10161 - Flux Control LoRA by @a-r-r-o-w in #9999
- Add PAG Support for Stable Diffusion Inpaint Pipeline by @darshil0805 in #9386
- [community pipeline rf-inversion] - fix example in doc by @linoytsaban in #10179
- Fix Nonetype attribute error when loading multiple Flux loras by @jonathanyin12 in #10182
- Added Error when len(gligen_images ) is not equal to len(gligen_phrases) in StableDiffusionGLIGENTextImagePipeline by @SahilCarterr in #10176
- [Single File] Add single file support for AutoencoderDC by @DN6 in #10183
- Add ControlNetUnion by @hlky in #10131
- fix min-snr implementation by @ethansmith2000 in #8466
- Add support for XFormers in SD3 by @CanvaChen in #8583
- [LoRA] add a test to ensure
set_adapters()
and attn kwargs outs match by @sayakpaul in #10110 - [CI] merge peft pr workflow into the main pr workflow. by @sayakpaul in #10042
- [WIP][Training] Flux Control LoRA training script by @sayakpaul in #10130
- [core] LTX Video by @a-r-r-o-w in #10021
- Ci update tpu by @paulinebm in #10197
- Remove
negative_*
from SDXL callback by @hlky in #10203 - refactor StableDiffusionXLControlNetUnion by @hlky in #10200
- update StableDiffusion3Img2ImgPipeline.add image size validation by @ZHJ19970917 in #10166
- Remove mps workaround for fp16 GELU, which is now supported natively by @skotapati in #10133
- [RF inversion community pipeline] add eta_decay by @linoytsaban in #10199
- Allow image resolutions multiple of 8 instead of 64 in SVD pipeline by @mlfarinha in #6646
- Use
torch
inget_2d_sincos_pos_embed
andget_3d_sincos_pos_embed
by @hlky in #10156 - add reshape to fix use_memory_efficient_attention in flax by @entrpn in #7918
- Add offload option in flux-control training by @Adenialzz in #10225
- Test error raised when loading normal and expanding loras together in Flux by @a-r-r-o-w in #10188
- [Sana] Add Sana, including
SanaPipeline
,SanaPAGPipeline
,LinearAttentionProcessor
,Flow-based DPM-sovler
and so on. by @lawrence-cj in #9982 - [Tests] update always test pipelines list. by @sayakpaul in #10143
- Update sana.md with minor corrections by @sayakpaul in #10232
- [docs] minor stuff to ltx video docs. by @sayakpaul in #10229
- Fix format issue in push_test yml by @DN6 in #10235
- [core] Hunyuan Video by @a-r-r-o-w in #10136
- Update pipeline_controlnet.py add support for pytorch_xla by @ in #10222
- [Docs] add rest of the lora loader mixins to the docs. by @sayakpaul in #10230
- Use
t
instead oftimestep
in_apply_perturbed_attention_guidance
by @hlky in #10243 - Add
dynamic_shifting
to SD3 by @hlky in #10236 - Fix
use_flow_sigmas
by @hlky in #10242 - Fix ControlNetUnion _callback_tensor_inputs by @hlky in #10218
- Use non-human subject in StableDiffusion3ControlNetPipeline example by @hlky in #10214
- Add enable_vae_tiling to AllegroPipeline, fix example by @hlky in #10212
- Fix checkpoint in CogView3PlusPipeline example by @hlky in #10211
- Fix RePaint Scheduler by @hlky in #10185
- Add ControlNetUnion to AutoPipeline from_pretrained by @hlky in #10219
- fix downsample bug in MidResTemporalBlock1D by @holmosaint in #10250
- [core] TorchAO Quantizer by @a-r-r-o-w in #10009
- [docs] Add missing AttnProcessors by @stevhliu in #10246
- [chore] add contribution note for lawrence. by @sayakpaul in #10253
- Fix copied from comment in Mochi lora loader by @a-r-r-o-w in #10255
- [LoRA] Support LTX Video by @a-r-r-o-w in #10228
- [docs] Clarify dtypes for Sana by @a-r-r-o-w in #10248
- [Single File] Add GGUF support by @DN6 in #9964
- Fix Mochi Quality Issues by @DN6 in #10033
- [tests] Remove/rename unsupported quantization torchao type by @a-r-r-o-w in #10263
- [docs] delete_adapters() by @stevhliu in #10245
- [Community Pipeline] Fix typo that cause error on regional prompting pipeline by @cjkangme in #10251
- Add
set_shift
to FlowMatchEulerDiscreteScheduler by @hlky in #10269 - [LoRA] feat: lora support for SANA. by @sayakpaul in #10234
- [chore] fix: licensing headers in mochi and ltx by @sayakpaul in #10275
- Use
torch
inget_2d_rotary_pos_embed
by @hlky in #10155 - [chore] fix: reamde -> readme by @sayakpaul in #10276
- Make
time_embed_dim
ofUNet2DModel
changeable by @Bichidian in #10262 - Support pass kwargs to sd3 custom attention processor by @Matrix53 in #9818
- Flux Control(Depth/Canny) + Inpaint by @affromero in #10192
- Fix sigma_last with use_flow_sigmas by @hlky in #10267
- Fix Doc links in GGUF and Quantization overview docs by @DN6 in #10279
- Make zeroing prompt embeds for Mochi Pipeline configurable by @DN6 in #10284
- [Single File] Add single file support for Flux Canny, Depth and Fill by @DN6 in #10288
- [tests] Fix broken cuda, nightly and lora tests on main for CogVideoX by @a-r-r-o-w in #10270
- Rename Mochi integration test correctly by @a-r-r-o-w in #10220
- [tests] remove nullop import checks from lora tests by @a-r-r-o-w in #10273
- [chore] Update README_sana.md to update the default model by @sayakpaul in #10285
- Hunyuan VAE tiling fixes and transformer docs by @a-r-r-o-w in #10295
- Add Flux Control to AutoPipeline by @hlky in #10292
- Update lora_conversion_utils.py by @zhaowendao30 in #9980
- Check correct model type is passed to
from_pretrained
by @hlky in #10189 - [LoRA] Support HunyuanVideo by @SHYuanBest in #10254
- [Single File] Add single file support for Mochi Transformer by @DN6 in #10268
- Allow Mochi Transformer to be split across multiple GPUs by @DN6 in #10300
- Fix
local_files_only
for checkpoints with shards by @hlky in #10294 - Fix failing lora tests after HunyuanVideo lora by @a-r-r-o-w in #10307
- unet's
sample_size
attribute is to accept tuple(h, w) inStableDiffusionPipeline
by @Foundsheep in #10181 - Enable Gradient Checkpointing for UNet2DModel (New) by @dg845 in #7201
- [WIP] SD3.5 IP-Adapter Pipeline Integration by @guiyrt in #9987
- Add support for sharded models when TorchAO quantization is enabled by @a-r-r-o-w in #10256
- Make tensors in ResNet contiguous for Hunyuan VAE by @a-r-r-o-w in #10309
- [Single File] Add GGUF support for LTX by @DN6 in #10298
- [LoRA] feat: support loading regular Flux LoRAs into Flux Control, and Fill by @sayakpaul in #10259
- [Tests] add integration tests for lora expansion stuff in Flux. by @sayakpaul in #10318
- Mochi docs by @DN6 in #9934
- [Docs] Update ltx_video.md to remove generator from
from_pretrained()
by @sayakpaul in #10316 - docs: fix a mistake in docstring by @Leojc in #10319
- [BUG FIX] [Stable Audio Pipeline] Resolve torch.Tensor.new_zeros() TypeError in function prepare_latents caused by audio_vae_length by @syntaxticsugr in #10306
- [docs] Fix quantization links by @stevhliu in #10323
- [Sana]add 2K related model for Sana by @lawrence-cj in #10322
- [Docs] Update gguf.md to remove generator from the pipeline from_pretrained by @sayakpaul in #10299
- Fix push_tests_mps.yml by @hlky in #10326
- Fix EMAModel test_from_pretrained by @hlky in #10325
- Support Flux IP Adapter by @hlky in #10261
- flux controlnet inpaint config bug by @yigitozgenc in #10291
- Community hosted weights for diffusers format HunyuanVideo weights by @a-r-r-o-w in #10344
- Fix enable_sequential_cpu_offload in test_kandinsky_combined by @hlky in #10324
- update
get_parameter_dtype
by @yiyixuxu in #10342 - [Single File] Add Single File support for HunYuan video by @DN6 in #10320
- [Sana bug] bug fix for 2K model config by @lawrence-cj in #10340
.from_single_file()
- Add missing.shape
by @gau-nernst in #10332- Bump minimum TorchAO version to 0.7.0 by @a-r-r-o-w in #10293
- [docs] fix: torchao example. by @sayakpaul in #10278
- [tests] Refactor TorchAO serialization fast tests by @a-r-r-o-w in #10271
- [SANA LoRA] sana lora training tests and misc. by @sayakpaul in #10296
- [Single File] Fix loading by @DN6 in #10349
- [Tests] QoL improvements to the LoRA test suite by @sayakpaul in #10304
- Fix FluxIPAdapterTesterMixin by @hlky in #10354
- Fix failing CogVideoX LoRA fuse test by @a-r-r-o-w in #10352
- Rename LTX blocks and docs title by @a-r-r-o-w in #10213
- [LoRA] test fix by @sayakpaul in #10351
- [Tests] Fix more tests sayak by @sayakpaul in #10359
- [core] LTX Video 0.9.1 by @a-r-r-o-w in #10330
- Release: v0.32.0 by @sayakpaul (direct commit on v0.32.0-release)
Significant community contributions
The following contributors have made significant changes to the library over the last release:
- @faaany
- @linoytsaban
- [SD3-5 dreambooth lora] update model cards (#9749)
- [SD 3.5 Dreambooth LoRA] support configurable training block & layers (#9762)
- [flux dreambooth lora training] make LoRA target modules configurable + small bug fix (#9646)
- [advanced flux training] bug fix + reduce memory cost as in #9829 (#9838)
- [SD3 dreambooth lora] smol fix to checkpoint saving (#9993)
- [Flux Redux] add prompt & multiple image input (#10056)
- [community pipeline] Add RF-inversion Flux pipeline (#9816)
- [community pipeline rf-inversion] - fix example in doc (#10179)
- [RF inversion community pipeline] add eta_decay (#10199)
- @raulc0399
- adds the pipeline for pixart alpha controlnet (#8857)
- @yiyixuxu
- @jellyheadandrew
- Add new community pipeline for 'Adaptive Mask Inpainting', introduced in [ECCV2024] ComA (#9228)
- @DN6
- Improve downloads of sharded variants (#9869)
- [CI] Unpin torch<2.5 in CI (#9961)
- Flux latents fix (#9929)
- [Single File] Fix SD3.5 single file loading (#10077)
- [Single File] Pass token when fetching interpreted config (#10082)
- [Single File] Add single file support for AutoencoderDC (#10183)
- Fix format issue in push_test yml (#10235)
- [Single File] Add GGUF support (#9964)
- Fix Mochi Quality Issues (#10033)
- Fix Doc links in GGUF and Quantization overview docs (#10279)
- Make zeroing prompt embeds for Mochi Pipeline configurable (#10284)
- [Single File] Add single file support for Flux Canny, Depth and Fill (#10288)
- [Single File] Add single file support for Mochi Transformer (#10268)
- Allow Mochi Transformer to be split across multiple GPUs (#10300)
- [Single File] Add GGUF support for LTX (#10298)
- Mochi docs (#9934)
- [Single File] Add Single File support for HunYuan video (#10320)
- [Single File] Fix loading (#10349)
- @ParagEkbote
- Notebooks for Community Scripts Examples (#9905)
- Move Wuerstchen Dreambooth to research_projects (#9935)
- Fixed Nits in Docs and Example Script (#9940)
- Notebooks for Community Scripts-2 (#9952)
- Move IP Adapter Scripts to research project (#9960)
- Notebooks for Community Scripts-3 (#10032)
- Fixed Nits in Evaluation Docs (#10063)
- Notebooks for Community Scripts-4 (#10094)
- Fix Broken Link in Optimization Docs (#10105)
- Fix Broken Links in ReadMe (#10117)
- @painebenjamin
- @hlky
- Fix beta and exponential sigmas + add tests (#9954)
- ControlNet from_single_file when already converted (#9978)
- Add
beta
,exponential
andkarras
sigmas toFlowMatchEulerDiscreteScheduler
(#10001) - Add
sigmas
to Flux pipelines (#10081) - Fix
num_images_per_prompt>1
with Skip Guidance Layers inStableDiffusion3Pipeline
(#10086) - Convert
sigmas
tonp.array
in FlowMatch set_timesteps (#10088) - Fix multi-prompt inference (#10103)
- Test
skip_guidance_layers
in SD3 pipeline (#10102) - Fix
pipeline_stable_audio
formating (#10114) - Add
sigmas
to pipelines using FlowMatch (#10116) - Use
torch
inget_3d_rotary_pos_embed
/_allegro
(#10161) - Add ControlNetUnion (#10131)
- Remove
negative_*
from SDXL callback (#10203) - refactor StableDiffusionXLControlNetUnion (#10200)
- Use
torch
inget_2d_sincos_pos_embed
andget_3d_sincos_pos_embed
(#10156) - Use
t
instead oftimestep
in_apply_perturbed_attention_guidance
(#10243) - Add
dynamic_shifting
to SD3 (#10236) - Fix
use_flow_sigmas
(#10242) - Fix ControlNetUnion _callback_tensor_inputs (#10218)
- Use non-human subject in StableDiffusion3ControlNetPipeline example (#10214)
- Add enable_vae_tiling to AllegroPipeline, fix example (#10212)
- Fix checkpoint in CogView3PlusPipeline example (#10211)
- Fix RePaint Scheduler (#10185)
- Add ControlNetUnion to AutoPipeline from_pretrained (#10219)
- Add
set_shift
to FlowMatchEulerDiscreteScheduler (#10269) - Use
torch
inget_2d_rotary_pos_embed
(#10155) - Fix sigma_last with use_flow_sigmas (#10267)
- Add Flux Control to AutoPipeline (#10292)
- Check correct model type is passed to
from_pretrained
(#10189) - Fix
local_files_only
for checkpoints with shards (#10294) - Fix push_tests_mps.yml (#10326)
- Fix EMAModel test_from_pretrained (#10325)
- Support Flux IP Adapter (#10261)
- Fix enable_sequential_cpu_offload in test_kandinsky_combined (#10324)
- Fix FluxIPAdapterTesterMixin (#10354)
- @dimitribarbot
- @suzukimain
- [community] Load Models from Sources like
Civitai
into Existing Pipelines (#9986)
- [community] Load Models from Sources like
- @lawrence-cj
- [DC-AE] Add the official Deep Compression Autoencoder code(32x,64x,128x compression ratio); (#9708)
- [Sana] Add Sana, including
SanaPipeline
,SanaPAGPipeline
,LinearAttentionProcessor
,Flow-based DPM-sovler
and so on. (#9982) - [Sana]add 2K related model for Sana (#10322)
- [Sana bug] bug fix for 2K model config (#10340)
- @darshil0805
- Add PAG Support for Stable Diffusion Inpaint Pipeline (#9386)
- @affromero
- Flux Control(Depth/Canny) + Inpaint (#10192)
- @SHYuanBest
- [LoRA] Support HunyuanVideo (#10254)
- @guiyrt
- [WIP] SD3.5 IP-Adapter Pipeline Integration (#9987)