- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - Documentation path like ".*/docs/.*" 312 files: docs/README.md docs/TRANSLATING.md docs/source/_config.py docs/source/pt/installation.md docs/source/pt/index.md docs/source/pt/quicktour.md docs/source/ja/stable_diffusion.md docs/source/ja/installation.md docs/source/ja/index.md docs/source/ja/quicktour.md docs/source/ja/tutorials/autopipeline.md docs/source/ja/tutorials/tutorial_overview.md docs/source/en/stable_diffusion.md docs/source/en/hybrid_inference/vae_encode.md docs/source/en/hybrid_inference/overview.md docs/source/en/hybrid_inference/vae_decode.md docs/source/en/hybrid_inference/api_reference.md docs/source/en/api/normalization.md docs/source/en/api/cache.md docs/source/en/api/logging.md docs/source/en/api/outputs.md docs/source/en/api/pipelines/flux.md docs/source/en/api/pipelines/dance_diffusion.md docs/source/en/api/pipelines/hunyuandit.md docs/source/en/api/pipelines/controlnetxs_sdxl.md docs/source/en/api/pipelines/shap_e.md docs/source/en/api/pipelines/cogview4.md docs/source/en/api/pipelines/controlnet_sdxl.md docs/source/en/api/pipelines/kandinsky_v22.md docs/source/en/api/pipelines/dit.md docs/source/en/api/pipelines/stable_cascade.md docs/source/en/api/pipelines/control_flux_inpaint.md docs/source/en/api/pipelines/hunyuan_video.md docs/source/en/api/pipelines/amused.md docs/source/en/api/pipelines/stable_diffusion/inpaint.md docs/source/en/api/pipelines/stable_diffusion/k_diffusion.md docs/source/en/api/pipelines/stable_diffusion/stable_diffusion_2.md docs/source/en/api/pipelines/stable_diffusion/depth2img.md docs/source/en/api/pipelines/stable_diffusion/sdxl_turbo.md docs/source/en/api/pipelines/stable_diffusion/image_variation.md docs/source/en/api/pipelines/stable_diffusion/stable_diffusion_xl.md docs/source/en/api/pipelines/stable_diffusion/gligen.md docs/source/en/api/pipelines/stable_diffusion/ldm3d_diffusion.md docs/source/en/api/pipelines/stable_diffusion/latent_upscale.md docs/source/en/api/pipelines/stable_diffusion/overview.md docs/source/en/api/pipelines/stable_diffusion/upscale.md docs/source/en/api/pipelines/mochi.md docs/source/en/api/pipelines/controlnetxs.md docs/source/en/api/pipelines/controlnet.md docs/source/en/api/pipelines/controlnet_sana.md docs/source/en/api/pipelines/ltx_video.md docs/source/en/api/pipelines/kandinsky.md docs/source/en/api/pipelines/chroma.md docs/source/en/api/pipelines/kandinsky3.md docs/source/en/api/pipelines/paint_by_example.md docs/source/en/api/pipelines/controlnet_hunyuandit.md docs/source/en/api/pipelines/ddpm.md docs/source/en/api/pipelines/text_to_video.md docs/source/en/api/pipelines/controlnet_union.md docs/source/en/api/pipelines/deepfloyd_if.md docs/source/en/api/pipelines/marigold.md docs/source/en/api/pipelines/aura_flow.md docs/source/en/api/pipelines/text_to_video_zero.md docs/source/en/api/pipelines/framepack.md docs/source/en/api/pipelines/cogvideox.md docs/source/en/api/pipelines/overview.md docs/source/en/api/pipelines/controlnet_sd3.md docs/source/en/api/pipelines/stable_audio.md docs/source/en/api/pipelines/auto_pipeline.md docs/source/en/api/pipelines/ddim.md docs/source/en/api/pipelines/pix2pix.md docs/source/en/api/pipelines/controlnet_flux.md docs/source/en/api/pipelines/easyanimate.md docs/source/en/api/pipelines/hidream.md docs/source/en/api/pipelines/cosmos.md docs/source/en/api/models/cosmos_transformer3d.md docs/source/en/api/models/sd3_transformer2d.md docs/source/en/api/models/pixart_transformer2d.md docs/source/en/api/models/aura_flow_transformer2d.md docs/source/en/api/models/transformer2d.md docs/source/en/api/models/autoencoderkl_mochi.md docs/source/en/api/models/cogview4_transformer2d.md docs/source/en/api/models/cogview3plus_transformer2d.md docs/source/en/api/models/autoencoder_oobleck.md docs/source/en/api/models/allegro_transformer3d.md docs/source/en/api/models/flux_transformer.md docs/source/en/api/models/autoencoderkl_cosmos.md docs/source/en/api/models/autoencoder_kl_hunyuan_video.md docs/source/en/api/models/autoencoder_tiny.md docs/source/en/api/models/lumina2_transformer2d.md docs/source/en/api/models/mochi_transformer3d.md docs/source/en/api/models/autoencoderkl_magvit.md docs/source/en/api/models/controlnet.md docs/source/en/api/models/wan_transformer_3d.md docs/source/en/api/models/autoencoderkl_allegro.md docs/source/en/api/models/hidream_image_transformer.md docs/source/en/api/models/controlnet_sana.md docs/source/en/api/models/consistency_decoder_vae.md docs/source/en/api/models/ltx_video_transformer3d.md docs/source/en/api/models/hunyuan_transformer2d.md docs/source/en/api/models/easyanimate_transformer3d.md docs/source/en/api/models/controlnet_hunyuandit.md docs/source/en/api/models/autoencoderkl.md docs/source/en/api/models/controlnet_union.md docs/source/en/api/models/hunyuan_video_transformer_3d.md docs/source/en/api/models/dit_transformer2d.md docs/source/en/api/models/latte_transformer3d.md docs/source/en/api/models/consisid_transformer3d.md docs/source/en/api/models/cogvideox_transformer3d.md docs/source/en/api/models/chroma_transformer.md docs/source/en/api/models/autoencoder_kl_wan.md docs/source/en/api/models/overview.md docs/source/en/api/models/controlnet_sd3.md docs/source/en/api/models/transformer_temporal.md docs/source/en/api/models/stable_audio_transformer.md docs/source/en/api/models/stable_cascade_unet.md docs/source/en/api/models/controlnet_flux.md docs/source/en/api/models/autoencoderkl_ltx_video.md docs/source/en/api/models/auto_model.md docs/source/en/api/models/lumina_nextdit2d.md docs/source/en/api/models/autoencoderkl_cogvideox.md docs/source/en/api/video_processor.md docs/source/en/api/quantization.md docs/source/en/api/loaders/transformer_sd3.md docs/source/en/api/loaders/ip_adapter.md docs/source/en/api/loaders/textual_inversion.md docs/source/en/api/loaders/unet.md docs/source/en/api/loaders/peft.md docs/source/en/api/loaders/lora.md docs/source/en/api/loaders/single_file.md docs/source/en/api/schedulers/heun.md docs/source/en/api/schedulers/flow_match_euler_discrete.md docs/source/en/api/schedulers/pndm.md docs/source/en/api/schedulers/euler.md docs/source/en/api/schedulers/multistep_dpm_solver_inverse.md docs/source/en/api/schedulers/consistency_decoder.md docs/source/en/api/schedulers/multistep_dpm_solver_cogvideox.md docs/source/en/api/schedulers/ipndm.md docs/source/en/api/schedulers/ddim_inverse.md docs/source/en/api/schedulers/flow_match_heun_discrete.md docs/source/en/api/schedulers/singlestep_dpm_solver.md docs/source/en/api/schedulers/lms_discrete.md docs/source/en/api/schedulers/dpm_discrete.md docs/source/en/api/schedulers/ddpm.md docs/source/en/api/schedulers/edm_euler.md docs/source/en/api/schedulers/dpm_sde.md docs/source/en/api/schedulers/edm_multistep_dpm_solver.md docs/source/en/api/schedulers/euler_ancestral.md docs/source/en/api/schedulers/overview.md docs/source/en/api/schedulers/dpm_discrete_ancestral.md docs/source/en/api/schedulers/ddim.md docs/source/en/api/schedulers/multistep_dpm_solver.md docs/source/en/api/schedulers/lcm.md docs/source/en/api/schedulers/cosine_dpm.md docs/source/en/api/schedulers/ddim_cogvideox.md docs/source/en/api/schedulers/stochastic_karras_ve.md docs/source/en/api/attnprocessor.md docs/source/en/api/internal_classes_overview.md docs/source/en/api/configuration.md docs/source/en/api/image_processor.md docs/source/en/api/activations.md docs/source/en/api/utilities.md docs/source/en/advanced_inference/outpaint.md docs/source/en/installation.md docs/source/en/community_projects.md docs/source/en/training/t2i_adapters.md docs/source/en/training/lcm_distill.md docs/source/en/training/wuerstchen.md docs/source/en/training/create_dataset.md docs/source/en/training/sdxl.md docs/source/en/training/controlnet.md docs/source/en/training/text2image.md docs/source/en/training/dreambooth.md docs/source/en/training/kandinsky.md docs/source/en/training/unconditional_training.md docs/source/en/training/lora.md docs/source/en/training/ddpo.md docs/source/en/training/cogvideox.md docs/source/en/training/adapt_a_model.md docs/source/en/training/text_inversion.md docs/source/en/training/overview.md docs/source/en/training/custom_diffusion.md docs/source/en/training/instructpix2pix.md docs/source/en/training/distributed_inference.md docs/source/en/index.md docs/source/en/optimization/pruna.md docs/source/en/optimization/onnx.md docs/source/en/optimization/habana.md docs/source/en/optimization/fp16.md docs/source/en/optimization/tgate.md docs/source/en/optimization/cache.md docs/source/en/optimization/deepcache.md docs/source/en/optimization/xformers.md docs/source/en/optimization/para_attn.md docs/source/en/optimization/xdit.md docs/source/en/optimization/tome.md docs/source/en/optimization/open_vino.md docs/source/en/optimization/neuron.md docs/source/en/optimization/mps.md docs/source/en/optimization/coreml.md docs/source/en/optimization/speed-memory-optims.md docs/source/en/quantization/torchao.md docs/source/en/quantization/bitsandbytes.md docs/source/en/quantization/quanto.md docs/source/en/quantization/overview.md docs/source/en/quantization/gguf.md docs/source/en/using-diffusers/t2i_adapter.md docs/source/en/using-diffusers/inference_with_tcd_lora.md docs/source/en/using-diffusers/stable_diffusion_jax_how_to.md docs/source/en/using-diffusers/inpaint.md docs/source/en/using-diffusers/ip_adapter.md docs/source/en/using-diffusers/custom_pipeline_overview.md docs/source/en/using-diffusers/marigold_usage.md docs/source/en/using-diffusers/img2img.md docs/source/en/using-diffusers/unconditional_image_generation.md docs/source/en/using-diffusers/schedulers.md docs/source/en/using-diffusers/conditional_image_generation.md docs/source/en/using-diffusers/textual_inversion_inference.md docs/source/en/using-diffusers/depth2img.md docs/source/en/using-diffusers/write_own_pipeline.md docs/source/en/using-diffusers/callback.md docs/source/en/using-diffusers/overview_techniques.md docs/source/en/using-diffusers/diffedit.md docs/source/en/using-diffusers/sdxl.md docs/source/en/using-diffusers/push_to_hub.md docs/source/en/using-diffusers/pag.md docs/source/en/using-diffusers/reusing_seeds.md docs/source/en/using-diffusers/controlnet.md docs/source/en/using-diffusers/consisid.md docs/source/en/using-diffusers/dreambooth.md docs/source/en/using-diffusers/scheduler_features.md docs/source/en/using-diffusers/omnigen.md docs/source/en/using-diffusers/shap-e.md docs/source/en/using-diffusers/kandinsky.md docs/source/en/using-diffusers/controlling_generation.md docs/source/en/using-diffusers/sdxl_turbo.md docs/source/en/using-diffusers/loading.md docs/source/en/using-diffusers/inference_with_lcm.md docs/source/en/using-diffusers/create_a_server.md docs/source/en/using-diffusers/svd.md docs/source/en/using-diffusers/image_quality.md docs/source/en/using-diffusers/text-img2vid.md docs/source/en/using-diffusers/other-formats.md docs/source/en/using-diffusers/weighted_prompts.md docs/source/en/conceptual/evaluation.md docs/source/en/conceptual/philosophy.md docs/source/en/conceptual/contribution.md docs/source/en/conceptual/ethical_guidelines.md docs/source/en/quicktour.md docs/source/en/tutorials/autopipeline.md docs/source/en/tutorials/tutorial_overview.md docs/source/en/tutorials/using_peft_for_inference.md docs/source/en/tutorials/basic_training.md docs/source/ko/stable_diffusion.md docs/source/ko/api/pipelines/stable_diffusion/stable_diffusion_xl.md docs/source/ko/installation.md docs/source/ko/training/create_dataset.md docs/source/ko/training/controlnet.md docs/source/ko/training/text2image.md docs/source/ko/training/dreambooth.md docs/source/ko/training/unconditional_training.md docs/source/ko/training/lora.md docs/source/ko/training/adapt_a_model.md docs/source/ko/training/text_inversion.md docs/source/ko/training/overview.md docs/source/ko/training/custom_diffusion.md docs/source/ko/training/instructpix2pix.md docs/source/ko/training/distributed_inference.md docs/source/ko/index.md docs/source/ko/optimization/torch2.0.md docs/source/ko/optimization/onnx.md docs/source/ko/optimization/habana.md docs/source/ko/optimization/fp16.md docs/source/ko/optimization/xformers.md docs/source/ko/optimization/tome.md docs/source/ko/optimization/open_vino.md docs/source/ko/optimization/mps.md docs/source/ko/optimization/coreml.md docs/source/ko/using-diffusers/stable_diffusion_jax_how_to.md docs/source/ko/using-diffusers/inpaint.md docs/source/ko/using-diffusers/custom_pipeline_overview.md docs/source/ko/using-diffusers/img2img.md docs/source/ko/using-diffusers/unconditional_image_generation.md docs/source/ko/using-diffusers/schedulers.md docs/source/ko/using-diffusers/conditional_image_generation.md docs/source/ko/using-diffusers/loading_adapters.md docs/source/ko/using-diffusers/textual_inversion_inference.md docs/source/ko/using-diffusers/depth2img.md docs/source/ko/using-diffusers/write_own_pipeline.md docs/source/ko/using-diffusers/diffedit.md docs/source/ko/using-diffusers/push_to_hub.md docs/source/ko/using-diffusers/shap-e.md docs/source/ko/using-diffusers/kandinsky.md docs/source/ko/using-diffusers/controlling_generation.md docs/source/ko/using-diffusers/sdxl_turbo.md docs/source/ko/using-diffusers/loading.md docs/source/ko/using-diffusers/svd.md docs/source/ko/using-diffusers/other-formats.md docs/source/ko/using-diffusers/weighted_prompts.md docs/source/ko/conceptual/evaluation.md docs/source/ko/conceptual/philosophy.md docs/source/ko/conceptual/contribution.md docs/source/ko/conceptual/ethical_guidelines.md docs/source/ko/in_translation.md docs/source/ko/quicktour.md docs/source/ko/tutorials/tutorial_overview.md docs/source/ko/tutorials/basic_training.md docs/source/zh/stable_diffusion.md docs/source/zh/installation.md docs/source/zh/index.md docs/source/zh/consisid.md docs/source/zh/quicktour.md - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - Too long lines (1000+ characters) 67 files: examples/research_projects/geodiff/geodiff_molecule_conformation.ipynb examples/community/README.md src/diffusers/models/autoencoders/autoencoder_kl_cosmos.py docs/source/en/api/pipelines/latte.md docs/source/en/api/pipelines/value_guided_sampling.md docs/source/en/api/pipelines/sana.md docs/source/en/api/pipelines/consistency_models.md docs/source/en/api/pipelines/blip_diffusion.md docs/source/en/api/pipelines/kolors.md docs/source/en/api/pipelines/wuerstchen.md docs/source/en/api/pipelines/stable_unclip.md docs/source/en/api/pipelines/visualcloze.md docs/source/en/api/pipelines/stable_diffusion/stable_diffusion_3.md docs/source/en/api/pipelines/stable_diffusion/img2img.md docs/source/en/api/pipelines/stable_diffusion/text2img.md docs/source/en/api/pipelines/stable_diffusion/svd.md docs/source/en/api/pipelines/stable_diffusion/adapter.md docs/source/en/api/pipelines/stable_diffusion/stable_diffusion_safe.md docs/source/en/api/pipelines/diffedit.md docs/source/en/api/pipelines/pag.md docs/source/en/api/pipelines/consisid.md docs/source/en/api/pipelines/pixart_sigma.md docs/source/en/api/pipelines/audioldm.md docs/source/en/api/pipelines/lumina.md docs/source/en/api/pipelines/omnigen.md docs/source/en/api/pipelines/sana_sprint.md docs/source/en/api/pipelines/unidiffuser.md docs/source/en/api/pipelines/ledits_pp.md docs/source/en/api/pipelines/allegro.md docs/source/en/api/pipelines/musicldm.md docs/source/en/api/pipelines/wan.md docs/source/en/api/pipelines/animatediff.md docs/source/en/api/pipelines/cogview3.md docs/source/en/api/pipelines/semantic_stable_diffusion.md docs/source/en/api/pipelines/pixart.md docs/source/en/api/pipelines/self_attention_guidance.md docs/source/en/api/pipelines/pia.md docs/source/en/api/pipelines/latent_diffusion.md docs/source/en/api/pipelines/lumina2.md docs/source/en/api/pipelines/latent_consistency_models.md docs/source/en/api/pipelines/panorama.md docs/source/en/api/pipelines/attend_and_excite.md docs/source/en/api/pipelines/audioldm2.md docs/source/en/api/pipelines/unclip.md docs/source/en/api/pipelines/i2vgenxl.md docs/source/en/api/models/uvit2d.md docs/source/en/api/models/vq.md docs/source/en/api/models/unet-motion.md docs/source/en/api/models/sana_transformer2d.md docs/source/en/api/models/unet.md docs/source/en/api/models/unet2d-cond.md docs/source/en/api/models/autoencoder_dc.md docs/source/en/api/models/asymmetricautoencoderkl.md docs/source/en/api/models/controlnet_sparsectrl.md docs/source/en/api/models/prior_transformer.md docs/source/en/api/models/omnigen_transformer.md docs/source/en/api/models/unet3d-cond.md docs/source/en/api/models/unet2d.md docs/source/en/api/schedulers/repaint.md docs/source/en/api/schedulers/tcd.md docs/source/en/api/schedulers/deis.md docs/source/en/api/schedulers/score_sde_vp.md docs/source/en/api/schedulers/unipc.md docs/source/en/api/schedulers/score_sde_ve.md docs/source/en/api/schedulers/vq_diffusion.md docs/source/en/api/schedulers/cm_stochastic_iterative.md docs/source/en/optimization/memory.md - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - Hidden files and folders path like ".*/[.][a-zA-Z0-9_]+.*" 7 files: examples/research_projects/pixart/.gitignore .github/workflows/pypi_publish.yaml .github/ISSUE_TEMPLATE/feedback.md .github/ISSUE_TEMPLATE/feature_request.md .github/ISSUE_TEMPLATE/translate.md .github/PULL_REQUEST_TEMPLATE.md .gitignore - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - Too long file (1000000+ bytes) 1 files: git-history.txt - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -