Merged
Conversation
* pag_initial * pag_docs * edit_docs * custom * typo * delete_docs * whitespace * make style --------- Co-authored-by: Sayak Paul <[email protected]>
* Add `final_sigma_zero` to UniPCMultistep Effectively the same trick as DDIM's `set_alpha_to_one` and DPM's `final_sigma_type='zero'`. Currently False by default but maybe this should be True? * `final_sigma_zero: bool` -> `final_sigmas_type: str` Should 1:1 match DPM Multistep now. * Set `final_sigmas_type='sigma_min'` in UniPC UTs
* fix ip adapter support * Update sag pipelines tests, adjust sag pipeline to pass tests --------- Co-authored-by: YiYi Xu <[email protected]>
…arigold v0.1.5 (huggingface#7524) * add resample option; check denoise_step; update ckpt path * Add seeding in pipeline to increase reproducibility * fix typo * fix typo
* Fix SVD bug (shape of `time_context`) * Formatting code * Formatting src/diffusers/models/transformers/transformer_temporal.py by `make style && make quality` --------- Co-authored-by: kevinkhwu <[email protected]> Co-authored-by: Sayak Paul <[email protected]> Co-authored-by: Dhruv Nair <[email protected]>
* add HD-Painter pipeline * style fixing * refactor, change doc, fix ruff * fix docs * used correct ruff version --------- Co-authored-by: Hayk Manukyan <[email protected]> Co-authored-by: Sayak Paul <[email protected]>
* add from_pipe --------- Co-authored-by: yiyixuxu <yixu310@gmail,com> Co-authored-by: Sayak Paul <[email protected]> Co-authored-by: Dhruv Nair <[email protected]>
…ggingface#7550) * initial-commit pipeline created * updated README.md
* make nightly workflow dispatchable. * add a note about running the release tests to setup.py
* remove class assignments for linear and conv. * fix: self.nn
* start printing the tensors. * print full throttle * set static slices for 7 tests. * remove printing. * flatten * disable test for controlnet * what happens when things are seeded properly? * set the right value * style./ * make pia test fail to check things * print. * fix pia. * checking for animatediff. * fix: animatediff. * video synthesis * final piece. * style. * print guess. * fix: assertion for control guess. --------- Co-authored-by: Dhruv Nair <[email protected]>
* 7529 do not disable autocast for cuda devices * Remove typecasting error check for non-mps platforms, as a correct autocast implementation makes it a non-issue * add autocast fix to other training examples * disable native_amp for dreambooth (sdxl) * disable native_amp for pix2pix (sdxl) * remove tests from remaining files * disable native_amp on huggingface accelerator for every training example that uses it * convert more usages of autocast to nullcontext, make style fixes * make style fixes * style. * Empty-Commit --------- Co-authored-by: bghira <[email protected]> Co-authored-by: Sayak Paul <[email protected]>
* add: utility to format our docs too 📜 * debugging saga * fix: message * checking * should be fixed. * revert pipeline_fixture * remove empty line * make style * fix: setup.py * style.
* UniPC UTs iterate solvers on FP16 It wasn't catching errs on order==3. Might be excessive? * UniPC Multistep fix tensor dtype/device on order=3 * UniPC UTs Add v_pred to fp16 test iter For completions sake. Probably overkill?
* UniPC Multistep add `rescale_betas_zero_snr` Same patch as DPM and Euler with the patched final alpha cumprod BF16 doesn't seem to break down, I think cause UniPC upcasts during some phases already? We could still force an upcast since it only loses ≈ 0.005 it/s for me but the difference in output is very small. A better endeavor might upcasting in step() and removing all the other upcasts elsewhere? * UniPC ZSNR UT * Re-add `rescale_betas_zsnr` doc oops
…face#7491) * refactor transformers 2d into multiple legacy variants. * fix: init. * fix recursive init. * add inits. * make transformer block creation more modular. * complete refactor. * remove forward * debug * remove legacy blocks and refactor within the module itself. * remove print * guard caption projection * remove fetcher. * reduce the number of args. * fix: norm_type * group variables that are shared. * remove _get_transformer_blocks * harmonize the init function signatures. * transformer_blocks to common * repeat .
* increase number of workers for the tests. * move to beefier runner. * improve the fast push tests too. * use a beefy machine for pytorch pipeline tests * up the number of workers further.
* Update pipeline_animatediff_video2video.py * commit with test for whether latent input can be passed into animatediffvid2vid
* Skip `test_freeu_enabled ` on MPS * Small fixes - import skip_mps correctly - disable all instances of test_freeu_enabled * Empty commit to trigger tests * Empty commit to trigger CI
* reduce block sizes for unet1d. * reduce blocks for unet_2d. * reduce block size for unet_motion * increase channels. * correctly increase channels. * reduce number of layers in unet2dconditionmodel tests. * reduce block sizes for unet2dconditionmodel tests * reduce block sizes for unet3dconditionmodel. * fix: test_feed_forward_chunking * fix: test_forward_with_norm_groups * skip spatiotemporal tests on MPS. * reduce block size in AutoencoderKL. * reduce block sizes for vqmodel. * further reduce block size. * make style. * Empty-Commit * reduce sizes for ConsistencyDecoderVAETests * further reduction. * further block reductions in AutoencoderKL and AssymetricAutoencoderKL. * massively reduce the block size in unet2dcontionmodel. * reduce sizes for unet3d * fix tests in unet3d. * reduce blocks further in motion unet. * fix: output shape * add attention_head_dim to the test configuration. * remove unexpected keyword arg * up a bit. * groups. * up again * fix
add set_begin_index for all if pipelines
* add audioldm2 tts * change gpt2 max new tokens * remove unnecessary pipeline and class * add TTS to AudioLDM2Pipeline * add TTS docs * delete unnecessary file * remove unnecessary import * add audioldm2 slow testcase * fix code quality * remove AudioLDMLearnablePositionalEmbedding * add variable check vits encoder * add use_learned_position_embedding --------- Co-authored-by: Dhruv Nair <[email protected]>
) Allow safety and feature extractor arguments to be passed to convert_from_ckpt Allows management of safety checker and feature extractor from outside of the convert ckpt class. Co-authored-by: Sayak Paul <[email protected]>
* Restore unet params back to normal from EMA when validation call is finished * empty commit --------- Co-authored-by: Sayak Paul <[email protected]>
* disable test * update --------- Co-authored-by: yiyixuxu <yixu310@gmail,com>
* Support multiimage masking --------- Co-authored-by: Sayak Paul <[email protected]> Co-authored-by: YiYi Xu <[email protected]>
* Hunyuan Team: add HunyuanDiT related updates --------- Co-authored-by: XCLiu <[email protected]> Co-authored-by: yiyixuxu <[email protected]>
* add hunyuandit doc * update hunyuandit doc * update hunyuandit 2d model * update toctree.yml for hunyuandit
* update * update * update * update
* Update transformer2d.md title For the other classes (e.g., UNet2DModel) the title of the documentation coincides with the name of the class, but that was not the case for Transformer2DModel. * Update model docs titles for consistency with class names
minor docs changes in hunyuandit
…ngface#8370) * handle norm_type of transformer2d_model safely. * log an info when old model class is being returned. * Apply suggestions from code review Co-authored-by: Dhruv Nair <[email protected]> * remove extra stuff --------- Co-authored-by: Dhruv Nair <[email protected]>
…STRING (huggingface#8401) Update code example in pipeline_stable_unclip_img2img.py Previous code caused an error when run
huggingface#8396) * feat: introduce qkv fusion for Hunyuan * fix copies
feat: support chunked ff.
…e#8399) * allow hunyuan dit to run under 6GB for GPU VRAM * add section in the docs/
…uggingface#8385) * fix: euledm when using the exp sigma schedule. * fix-copies * remove print. * reduce friction * yiyi's suggestioms
* add training code of gligen * fix code quality tests. --------- Co-authored-by: Sayak Paul <[email protected]>
* Fix typos * Trim trailing whitespaces * Remove a trailing whitespace * chore: Update MarigoldDepthPipeline checkpoint to prs-eth/marigold-lcm-v1-0 * Revert "chore: Update MarigoldDepthPipeline checkpoint to prs-eth/marigold-lcm-v1-0" This reverts commit fd742b3. * pokemon -> naruto * `DPMSolverMultistep` -> `DPMSolverMultistepScheduler` * Improve Markdown stylization * Improve style * Improve style * Refactor pipeline variable names for consistency * up style
…uggingface#8402) * optimizations to the hunyuan dit docs. * Apply suggestions from code review Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/api/pipelines/hunyuandit.md Co-authored-by: Steven Liu <[email protected]> --------- Co-authored-by: Steven Liu <[email protected]>
* fix: legacy model mapping * remove print
* single file usage * edit
* Refactor code to remove unnecessary calls to `to(torch_device)`
* Refactor code to remove unnecessary calls to `to("cuda")`
* Update pipeline_stable_diffusion_diffedit.py
* first draft * secret * tiktok * capital matters * dataset matter * don't be a prick * refact * only on main or tag * document with an example * Update destination dataset * link * allow manual trigger * better * lin --------- Co-authored-by: Sayak Paul <[email protected]>
…#7830) * feat: support saving a model in sharded checkpoints. * feat: make loading of sharded checkpoints work. * add tests * cleanse the loading logic a bit more. * more resilience while loading from the Hub. * parallelize shard downloads by using snapshot_download()/ * default to a shard size. * more fix * Empty-Commit * debug * fix * uality * more debugging * fix more * initial comments from Benjamin * move certain methods to loading_utils * add test to check if the correct number of shards are present. * add a test to check if loading of sharded checkpoints from the Hub is okay * clarify the unit when passed as an int. * use hf_hub for sharding. * remove unnecessary code * remove unnecessary function * lucain's comments. * fixes * address high-level comments. * fix test * subfolder shenanigans./ * Update src/diffusers/utils/hub_utils.py Co-authored-by: Lucain <[email protected]> * Apply suggestions from code review Co-authored-by: Lucain <[email protected]> * remove _huggingface_hub_version as not needed. * address more feedback. * add a test for local_files_only=True/ * need hf hub to be at least 0.23.2 * style * final comment. * clean up subfolder. * deal with suffixes in code. * _add_variant default. * use weights_name_pattern * remove add_suffix_keyword * clean up downloading of sharded ckpts. * don't return something special when using index.json * fix more * don't use bare except * remove comments and catch the errors better * fix a couple of things when using is_file() * empty --------- Co-authored-by: Lucain <[email protected]>
* Move away from * unused constant * Add custom error
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
What does this PR do?
clipdrop-mainbranch withdiffusersmainbranch