SYMBOL INDEX (6563 symbols across 386 files) FILE: run_stable_flow.py class StableFlow (line 9) | class StableFlow: method __init__ (line 13) | def __init__(self): method _parse_args (line 17) | def _parse_args(self): method _load_pipeline (line 32) | def _load_pipeline(self): method infer_and_save (line 46) | def infer_and_save(self, prompts): method image2latent (line 70) | def image2latent(self, image, latent_nudging_scalar = 1.15): method invert_and_save (line 87) | def invert_and_save(self, prompts): FILE: setup.py function deps_list (line 165) | def deps_list(*pkgs): class DepsTableUpdateCommand (line 169) | class DepsTableUpdateCommand(Command): method initialize_options (line 185) | def initialize_options(self): method finalize_options (line 188) | def finalize_options(self): method run (line 191) | def run(self): FILE: src/diffusers/callbacks.py class PipelineCallback (line 7) | class PipelineCallback(ConfigMixin): method __init__ (line 22) | def __init__(self, cutoff_step_ratio=1.0, cutoff_step_index=None): method tensor_inputs (line 36) | def tensor_inputs(self) -> List[str]: method callback_fn (line 39) | def callback_fn(self, pipeline, step_index, timesteps, callback_kwargs... method __call__ (line 42) | def __call__(self, pipeline, step_index, timestep, callback_kwargs) ->... class MultiPipelineCallbacks (line 46) | class MultiPipelineCallbacks: method __init__ (line 52) | def __init__(self, callbacks: List[PipelineCallback]): method tensor_inputs (line 56) | def tensor_inputs(self) -> List[str]: method __call__ (line 59) | def __call__(self, pipeline, step_index, timestep, callback_kwargs) ->... class SDCFGCutoffCallback (line 69) | class SDCFGCutoffCallback(PipelineCallback): method callback_fn (line 79) | def callback_fn(self, pipeline, step_index, timestep, callback_kwargs)... class SDXLCFGCutoffCallback (line 98) | class SDXLCFGCutoffCallback(PipelineCallback): method callback_fn (line 108) | def callback_fn(self, pipeline, step_index, timestep, callback_kwargs)... class IPAdapterScaleCutoffCallback (line 135) | class IPAdapterScaleCutoffCallback(PipelineCallback): method callback_fn (line 145) | def callback_fn(self, pipeline, step_index, timestep, callback_kwargs)... FILE: src/diffusers/commands/__init__.py class BaseDiffusersCLICommand (line 19) | class BaseDiffusersCLICommand(ABC): method register_subcommand (line 22) | def register_subcommand(parser: ArgumentParser): method run (line 26) | def run(self): FILE: src/diffusers/commands/diffusers_cli.py function main (line 22) | def main(): FILE: src/diffusers/commands/env.py function info_command_factory (line 36) | def info_command_factory(_): class EnvironmentCommand (line 40) | class EnvironmentCommand(BaseDiffusersCLICommand): method register_subcommand (line 42) | def register_subcommand(parser: ArgumentParser) -> None: method run (line 46) | def run(self) -> dict: method format_dict (line 179) | def format_dict(d: dict) -> str: FILE: src/diffusers/commands/fp16_safetensors.py function conversion_command_factory (line 35) | def conversion_command_factory(args: Namespace): class FP16SafetensorsCommand (line 44) | class FP16SafetensorsCommand(BaseDiffusersCLICommand): method register_subcommand (line 46) | def register_subcommand(parser: ArgumentParser): method __init__ (line 66) | def __init__(self, ckpt_id: str, fp16: bool, use_safetensors: bool): method run (line 79) | def run(self): FILE: src/diffusers/configuration_utils.py class FrozenDict (line 55) | class FrozenDict(OrderedDict): method __init__ (line 56) | def __init__(self, *args, **kwargs): method __delitem__ (line 64) | def __delitem__(self, *args, **kwargs): method setdefault (line 67) | def setdefault(self, *args, **kwargs): method pop (line 70) | def pop(self, *args, **kwargs): method update (line 73) | def update(self, *args, **kwargs): method __setattr__ (line 76) | def __setattr__(self, name, value): method __setitem__ (line 81) | def __setitem__(self, name, value): class ConfigMixin (line 87) | class ConfigMixin: method register_to_config (line 110) | def register_to_config(self, **kwargs): method __getattr__ (line 127) | def __getattr__(self, name: str) -> Any: method save_config (line 145) | def save_config(self, save_directory: Union[str, os.PathLike], push_to... method from_config (line 188) | def from_config(cls, config: Union[FrozenDict, Dict[str, Any]] = None,... method get_config_dict (line 278) | def get_config_dict(cls, *args, **kwargs): method load_config (line 288) | def load_config( method _get_init_keys (line 452) | def _get_init_keys(input_class): method extract_init_dict (line 456) | def extract_init_dict(cls, config_dict, **kwargs): method _dict_from_json_file (line 552) | def _dict_from_json_file(cls, json_file: Union[str, os.PathLike]): method __repr__ (line 557) | def __repr__(self): method config (line 561) | def config(self) -> Dict[str, Any]: method to_json_string (line 570) | def to_json_string(self) -> str: method to_json_file (line 596) | def to_json_file(self, json_file_path: Union[str, os.PathLike]): function register_to_config (line 608) | def register_to_config(init): function flax_register_to_config (line 658) | def flax_register_to_config(cls): class LegacyConfigMixin (line 706) | class LegacyConfigMixin(ConfigMixin): method from_config (line 713) | def from_config(cls, config: Union[FrozenDict, Dict[str, Any]] = None,... FILE: src/diffusers/dependency_versions_check.py function dep_version_check (line 33) | def dep_version_check(pkg, hint=None): FILE: src/diffusers/experimental/rl/value_guided_sampling.py class ValueGuidedRLPipeline (line 25) | class ValueGuidedRLPipeline(DiffusionPipeline): method __init__ (line 44) | def __init__( method normalize (line 71) | def normalize(self, x_in, key): method de_normalize (line 74) | def de_normalize(self, x_in, key): method to_torch (line 77) | def to_torch(self, x_in): method reset_x0 (line 84) | def reset_x0(self, x_in, cond, act_dim): method run_diffusion (line 89) | def run_diffusion(self, x, conditions, n_guide_steps, scale): method __call__ (line 122) | def __call__(self, obs, batch_size=64, planning_horizon=32, n_guide_st... FILE: src/diffusers/image_processor.py function is_valid_image (line 41) | def is_valid_image(image): function is_valid_image_imagelist (line 45) | def is_valid_image_imagelist(images): class VaeImageProcessor (line 60) | class VaeImageProcessor(ConfigMixin): method __init__ (line 85) | def __init__( method numpy_to_pil (line 105) | def numpy_to_pil(images: np.ndarray) -> List[PIL.Image.Image]: method pil_to_numpy (line 121) | def pil_to_numpy(images: Union[List[PIL.Image.Image], PIL.Image.Image]... method numpy_to_pt (line 133) | def numpy_to_pt(images: np.ndarray) -> torch.Tensor: method pt_to_numpy (line 144) | def pt_to_numpy(images: torch.Tensor) -> np.ndarray: method normalize (line 152) | def normalize(images: Union[np.ndarray, torch.Tensor]) -> Union[np.nda... method denormalize (line 159) | def denormalize(images: Union[np.ndarray, torch.Tensor]) -> Union[np.n... method convert_to_rgb (line 166) | def convert_to_rgb(image: PIL.Image.Image) -> PIL.Image.Image: method convert_to_grayscale (line 175) | def convert_to_grayscale(image: PIL.Image.Image) -> PIL.Image.Image: method blur (line 184) | def blur(image: PIL.Image.Image, blur_factor: int = 4) -> PIL.Image.Im... method get_crop_region (line 193) | def get_crop_region(mask_image: PIL.Image.Image, width: int, height: i... method _resize_and_fill (line 282) | def _resize_and_fill( method _resize_and_crop (line 327) | def _resize_and_crop( method resize (line 353) | def resize( method binarize (line 409) | def binarize(self, image: PIL.Image.Image) -> PIL.Image.Image: method get_default_height_width (line 426) | def get_default_height_width( method preprocess (line 469) | def preprocess( method postprocess (line 596) | def postprocess( method apply_overlay (line 651) | def apply_overlay( class VaeImageProcessorLDM3D (line 687) | class VaeImageProcessorLDM3D(VaeImageProcessor): method __init__ (line 705) | def __init__( method numpy_to_pil (line 715) | def numpy_to_pil(images: np.ndarray) -> List[PIL.Image.Image]: method depth_pil_to_numpy (line 731) | def depth_pil_to_numpy(images: Union[List[PIL.Image.Image], PIL.Image.... method rgblike_to_depthmap (line 743) | def rgblike_to_depthmap(image: Union[np.ndarray, torch.Tensor]) -> Uni... method numpy_to_depth (line 753) | def numpy_to_depth(self, images: np.ndarray) -> List[PIL.Image.Image]: method postprocess (line 773) | def postprocess( method preprocess (line 828) | def preprocess( class IPAdapterMaskProcessor (line 928) | class IPAdapterMaskProcessor(VaeImageProcessor): method __init__ (line 951) | def __init__( method downsample (line 970) | def downsample(mask: torch.Tensor, batch_size: int, num_queries: int, ... class PixArtImageProcessor (line 1032) | class PixArtImageProcessor(VaeImageProcessor): method __init__ (line 1055) | def __init__( method classify_height_width_bin (line 1074) | def classify_height_width_bin(height: int, width: int, ratios: dict) -... method resize_and_crop_tensor (line 1082) | def resize_and_crop_tensor(samples: torch.Tensor, new_width: int, new_... FILE: src/diffusers/loaders/__init__.py function text_encoder_lora_state_dict (line 7) | def text_encoder_lora_state_dict(text_encoder): function text_encoder_attn_modules (line 33) | def text_encoder_attn_modules(text_encoder): FILE: src/diffusers/loaders/ip_adapter.py class IPAdapterMixin (line 51) | class IPAdapterMixin: method load_ip_adapter (line 55) | def load_ip_adapter( method set_ip_adapter_scale (line 245) | def set_ip_adapter_scale(self, scale): method unload_ip_adapter (line 297) | def unload_ip_adapter(self): FILE: src/diffusers/loaders/lora_base.py function fuse_text_encoder_lora (line 55) | def fuse_text_encoder_lora(text_encoder, lora_scale=1.0, safe_fusing=Fal... function unfuse_text_encoder_lora (line 91) | def unfuse_text_encoder_lora(text_encoder): function set_adapters_for_text_encoder (line 105) | def set_adapters_for_text_encoder( function disable_lora_for_text_encoder (line 149) | def disable_lora_for_text_encoder(text_encoder: Optional["PreTrainedMode... function enable_lora_for_text_encoder (line 163) | def enable_lora_for_text_encoder(text_encoder: Optional["PreTrainedModel... function _remove_text_encoder_monkey_patch (line 177) | def _remove_text_encoder_monkey_patch(text_encoder): class LoraBaseMixin (line 184) | class LoraBaseMixin: method load_lora_weights (line 190) | def load_lora_weights(self, **kwargs): method save_lora_weights (line 194) | def save_lora_weights(cls, **kwargs): method lora_state_dict (line 198) | def lora_state_dict(cls, **kwargs): method _optionally_disable_offloading (line 202) | def _optionally_disable_offloading(cls, _pipeline): method _fetch_state_dict (line 237) | def _fetch_state_dict( method _best_guess_weight_name (line 314) | def _best_guess_weight_name( method unload_lora_weights (line 356) | def unload_lora_weights(self): method fuse_lora (line 379) | def fuse_lora( method unfuse_lora (line 460) | def unfuse_lora(self, components: List[str] = [], **kwargs): method set_adapters (line 516) | def set_adapters( method disable_lora (line 577) | def disable_lora(self): method enable_lora (line 589) | def enable_lora(self): method delete_adapters (line 601) | def delete_adapters(self, adapter_names: Union[List[str], str]): method get_active_adapters (line 623) | def get_active_adapters(self) -> List[str]: method get_list_adapters (line 656) | def get_list_adapters(self) -> Dict[str, List[str]]: method set_lora_device (line 678) | def set_lora_device(self, adapter_names: List[str], device: Union[torc... method pack_weights (line 707) | def pack_weights(layers, prefix): method write_lora_layers (line 713) | def write_lora_layers( method lora_scale (line 749) | def lora_scale(self) -> float: FILE: src/diffusers/loaders/lora_conversion_utils.py function _maybe_map_sgm_blocks_to_diffusers (line 23) | def _maybe_map_sgm_blocks_to_diffusers(state_dict, unet_config, delimite... function _convert_non_diffusers_lora_to_diffusers (line 126) | def _convert_non_diffusers_lora_to_diffusers(state_dict, unet_name="unet... function _convert_unet_lora_key (line 234) | def _convert_unet_lora_key(key): function _convert_text_encoder_lora_key (line 290) | def _convert_text_encoder_lora_key(key, lora_name): function _get_alpha_name (line 317) | def _get_alpha_name(lora_name_alpha, diffusers_name, alpha): FILE: src/diffusers/loaders/lora_pipeline.py class StableDiffusionLoraLoaderMixin (line 50) | class StableDiffusionLoraLoaderMixin(LoraBaseMixin): method load_lora_weights (line 60) | def load_lora_weights( method lora_state_dict (line 121) | def lora_state_dict( method load_lora_into_unet (line 230) | def load_lora_into_unet(cls, state_dict, network_alphas, unet, adapter... method load_lora_into_text_encoder (line 265) | def load_lora_into_text_encoder( method save_lora_weights (line 383) | def save_lora_weights( method fuse_lora (line 436) | def fuse_lora( method unfuse_lora (line 479) | def unfuse_lora(self, components: List[str] = ["unet", "text_encoder"]... class StableDiffusionXLLoraLoaderMixin (line 500) | class StableDiffusionXLLoraLoaderMixin(LoraBaseMixin): method load_lora_weights (line 511) | def load_lora_weights( method lora_state_dict (line 592) | def lora_state_dict( method load_lora_into_unet (line 702) | def load_lora_into_unet(cls, state_dict, network_alphas, unet, adapter... method load_lora_into_text_encoder (line 738) | def load_lora_into_text_encoder( method save_lora_weights (line 856) | def save_lora_weights( method fuse_lora (line 917) | def fuse_lora( method unfuse_lora (line 960) | def unfuse_lora(self, components: List[str] = ["unet", "text_encoder",... class SD3LoraLoaderMixin (line 981) | class SD3LoraLoaderMixin(LoraBaseMixin): method lora_state_dict (line 996) | def lora_state_dict( method load_lora_weights (line 1085) | def load_lora_weights( method load_lora_into_transformer (line 1155) | def load_lora_into_transformer(cls, state_dict, transformer, adapter_n... method load_lora_into_text_encoder (line 1234) | def load_lora_into_text_encoder( method save_lora_weights (line 1352) | def save_lora_weights( method fuse_lora (line 1414) | def fuse_lora( method unfuse_lora (line 1457) | def unfuse_lora(self, components: List[str] = ["transformer", "text_en... class FluxLoraLoaderMixin (line 1478) | class FluxLoraLoaderMixin(LoraBaseMixin): method lora_state_dict (line 1493) | def lora_state_dict( method load_lora_weights (line 1582) | def load_lora_weights( method load_lora_into_transformer (line 1641) | def load_lora_into_transformer(cls, state_dict, transformer, adapter_n... method load_lora_into_text_encoder (line 1720) | def load_lora_into_text_encoder( method save_lora_weights (line 1839) | def save_lora_weights( method fuse_lora (line 1893) | def fuse_lora( method unfuse_lora (line 1936) | def unfuse_lora(self, components: List[str] = ["transformer", "text_en... class AmusedLoraLoaderMixin (line 1955) | class AmusedLoraLoaderMixin(StableDiffusionLoraLoaderMixin): method load_lora_into_transformer (line 1961) | def load_lora_into_transformer(cls, state_dict, network_alphas, transf... method load_lora_into_text_encoder (line 2046) | def load_lora_into_text_encoder( method save_lora_weights (line 2164) | def save_lora_weights( class LoraLoaderMixin (line 2218) | class LoraLoaderMixin(StableDiffusionLoraLoaderMixin): method __init__ (line 2219) | def __init__(self, *args, **kwargs): FILE: src/diffusers/loaders/peft.py class PeftAdapterMixin (line 39) | class PeftAdapterMixin: method set_adapters (line 55) | def set_adapters( method add_adapter (line 111) | def add_adapter(self, adapter_config, adapter_name: str = "default") -... method set_adapter (line 149) | def set_adapter(self, adapter_name: Union[str, List[str]]) -> None: method disable_adapters (line 198) | def disable_adapters(self) -> None: method enable_adapters (line 220) | def enable_adapters(self) -> None: method active_adapters (line 243) | def active_adapters(self) -> List[str]: method fuse_lora (line 264) | def fuse_lora(self, lora_scale=1.0, safe_fusing=False, adapter_names=N... method _fuse_lora_apply (line 272) | def _fuse_lora_apply(self, module, adapter_names=None): method unfuse_lora (line 294) | def unfuse_lora(self): method _unfuse_lora_apply (line 299) | def _unfuse_lora_apply(self, module): method unload_lora (line 305) | def unload_lora(self): method disable_lora (line 315) | def disable_lora(self): method enable_lora (line 338) | def enable_lora(self): method delete_adapters (line 361) | def delete_adapters(self, adapter_names: Union[List[str], str]): FILE: src/diffusers/loaders/single_file.py function load_single_file_sub_model (line 51) | def load_single_file_sub_model( function _map_component_types_to_config_dict (line 178) | def _map_component_types_to_config_dict(component_types): function _infer_pipeline_config_dict (line 227) | def _infer_pipeline_config_dict(pipeline_class): function _download_diffusers_model_config_from_hub (line 239) | def _download_diffusers_model_config_from_hub( class FromSingleFileMixin (line 263) | class FromSingleFileMixin: method from_single_file (line 270) | def from_single_file(cls, pretrained_model_link_or_path, **kwargs): FILE: src/diffusers/loaders/single_file_model.py function _get_single_file_loadable_mapping_class (line 88) | def _get_single_file_loadable_mapping_class(cls): function _get_mapping_function_kwargs (line 99) | def _get_mapping_function_kwargs(mapping_fn, **kwargs): class FromOriginalModelMixin (line 110) | class FromOriginalModelMixin: method from_single_file (line 117) | def from_single_file(cls, pretrained_model_link_or_path_or_dict: Optio... FILE: src/diffusers/loaders/single_file_utils.py class SingleFileComponentError (line 276) | class SingleFileComponentError(Exception): method __init__ (line 277) | def __init__(self, message=None): function is_valid_url (line 282) | def is_valid_url(url): function _extract_repo_id_and_weights_name (line 290) | def _extract_repo_id_and_weights_name(pretrained_model_name_or_path): function _is_model_weights_in_cached_folder (line 310) | def _is_model_weights_in_cached_folder(cached_folder, name): function load_single_file_checkpoint (line 321) | def load_single_file_checkpoint( function fetch_original_config (line 355) | def fetch_original_config(original_config_file, local_files_only=False): function is_clip_model (line 377) | def is_clip_model(checkpoint): function is_clip_sdxl_model (line 384) | def is_clip_sdxl_model(checkpoint): function is_clip_sd3_model (line 391) | def is_clip_sd3_model(checkpoint): function is_open_clip_model (line 398) | def is_open_clip_model(checkpoint): function is_open_clip_sdxl_model (line 405) | def is_open_clip_sdxl_model(checkpoint): function is_open_clip_sd3_model (line 412) | def is_open_clip_sd3_model(checkpoint): function is_open_clip_sdxl_refiner_model (line 419) | def is_open_clip_sdxl_refiner_model(checkpoint): function is_clip_model_in_single_file (line 426) | def is_clip_model_in_single_file(class_obj, checkpoint): function infer_diffusers_model_type (line 445) | def infer_diffusers_model_type(checkpoint): function fetch_diffusers_config (line 530) | def fetch_diffusers_config(checkpoint): function set_image_size (line 537) | def set_image_size(checkpoint, image_size=None): function conv_attn_to_linear (line 548) | def conv_attn_to_linear(checkpoint): function create_unet_diffusers_config_from_ldm (line 560) | def create_unet_diffusers_config_from_ldm( function create_controlnet_diffusers_config_from_ldm (line 691) | def create_controlnet_diffusers_config_from_ldm(original_config, checkpo... function create_vae_diffusers_config_from_ldm (line 723) | def create_vae_diffusers_config_from_ldm(original_config, checkpoint, im... function update_unet_resnet_ldm_to_diffusers (line 774) | def update_unet_resnet_ldm_to_diffusers(ldm_keys, new_checkpoint, checkp... function update_unet_attention_ldm_to_diffusers (line 789) | def update_unet_attention_ldm_to_diffusers(ldm_keys, new_checkpoint, che... function update_vae_resnet_ldm_to_diffusers (line 795) | def update_vae_resnet_ldm_to_diffusers(keys, new_checkpoint, checkpoint,... function update_vae_attentions_ldm_to_diffusers (line 801) | def update_vae_attentions_ldm_to_diffusers(keys, new_checkpoint, checkpo... function convert_stable_cascade_unet_single_file_to_diffusers (line 827) | def convert_stable_cascade_unet_single_file_to_diffusers(checkpoint, **k... function convert_ldm_unet_checkpoint (line 883) | def convert_ldm_unet_checkpoint(checkpoint, config, extract_ema=False, *... function convert_controlnet_checkpoint (line 1050) | def convert_controlnet_checkpoint( function convert_ldm_vae_checkpoint (line 1174) | def convert_ldm_vae_checkpoint(checkpoint, config): function convert_ldm_clip_checkpoint (line 1274) | def convert_ldm_clip_checkpoint(checkpoint, remove_prefix=None): function convert_open_clip_checkpoint (line 1292) | def convert_open_clip_checkpoint( function create_diffusers_clip_model_from_ldm (line 1358) | def create_diffusers_clip_model_from_ldm( function _legacy_load_scheduler (line 1468) | def _legacy_load_scheduler( function _legacy_load_clip_tokenizer (line 1598) | def _legacy_load_clip_tokenizer(cls, checkpoint, config=None, local_file... function _legacy_load_safety_checker (line 1624) | def _legacy_load_safety_checker(local_files_only, torch_dtype): function swap_scale_shift (line 1642) | def swap_scale_shift(weight, dim): function convert_sd3_transformer_checkpoint_to_diffusers (line 1648) | def convert_sd3_transformer_checkpoint_to_diffusers(checkpoint, **kwargs): function is_t5_in_single_file (line 1792) | def is_t5_in_single_file(checkpoint): function convert_sd3_t5_checkpoint_to_diffusers (line 1799) | def convert_sd3_t5_checkpoint_to_diffusers(checkpoint): function create_diffusers_t5_model_from_checkpoint (line 1814) | def create_diffusers_t5_model_from_checkpoint( function convert_animatediff_checkpoint_to_diffusers (line 1863) | def convert_animatediff_checkpoint_to_diffusers(checkpoint, **kwargs): function convert_flux_transformer_checkpoint_to_diffusers (line 1882) | def convert_flux_transformer_checkpoint_to_diffusers(checkpoint, **kwargs): FILE: src/diffusers/loaders/textual_inversion.py function load_textual_inversion_state_dicts (line 38) | def load_textual_inversion_state_dicts(pretrained_model_name_or_paths, *... class TextualInversionLoaderMixin (line 110) | class TextualInversionLoaderMixin: method maybe_convert_prompt (line 115) | def maybe_convert_prompt(self, prompt: Union[str, List[str]], tokenize... method _maybe_convert_prompt (line 142) | def _maybe_convert_prompt(self, prompt: str, tokenizer: "PreTrainedTok... method _check_text_inv_inputs (line 172) | def _check_text_inv_inputs(self, tokenizer, text_encoder, pretrained_m... method _retrieve_tokens_and_embeddings (line 196) | def _retrieve_tokens_and_embeddings(tokens, state_dicts, tokenizer): method _extend_tokens_and_embeddings (line 237) | def _extend_tokens_and_embeddings(tokens, embeddings, tokenizer): method load_textual_inversion (line 264) | def load_textual_inversion( method unload_textual_inversion (line 458) | def unload_textual_inversion( FILE: src/diffusers/loaders/unet.py class UNet2DConditionLoadersMixin (line 60) | class UNet2DConditionLoadersMixin: method load_attn_procs (line 69) | def load_attn_procs(self, pretrained_model_name_or_path_or_dict: Union... method _process_custom_diffusion (line 237) | def _process_custom_diffusion(self, state_dict): method _process_lora (line 271) | def _process_lora(self, state_dict, unet_identifier_key, network_alpha... method _optionally_disable_offloading (line 355) | def _optionally_disable_offloading(cls, _pipeline): method save_attn_procs (line 389) | def save_attn_procs( method _get_custom_diffusion_state_dict (line 488) | def _get_custom_diffusion_state_dict(self): method _convert_ip_adapter_image_proj_to_diffusers (line 516) | def _convert_ip_adapter_image_proj_to_diffusers(self, state_dict, low_... method _convert_ip_adapter_attn_to_diffusers (line 733) | def _convert_ip_adapter_attn_to_diffusers(self, state_dicts, low_cpu_m... method _load_ip_adapter_weights (line 823) | def _load_ip_adapter_weights(self, state_dicts, low_cpu_mem_usage=False): method _load_ip_adapter_loras (line 855) | def _load_ip_adapter_loras(self, state_dicts): FILE: src/diffusers/loaders/unet_loader_utils.py function _translate_into_actual_layer_name (line 27) | def _translate_into_actual_layer_name(name): function _maybe_expand_lora_scales (line 41) | def _maybe_expand_lora_scales( function _maybe_expand_lora_scales_for_one_adapter (line 64) | def _maybe_expand_lora_scales_for_one_adapter( FILE: src/diffusers/loaders/utils.py class AttnProcsLayers (line 20) | class AttnProcsLayers(torch.nn.Module): method __init__ (line 21) | def __init__(self, state_dict: Dict[str, torch.Tensor]): FILE: src/diffusers/models/activations.py function get_activation (line 36) | def get_activation(act_fn: str) -> nn.Module: class FP32SiLU (line 53) | class FP32SiLU(nn.Module): method __init__ (line 58) | def __init__(self): method forward (line 61) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: class GELU (line 65) | class GELU(nn.Module): method __init__ (line 76) | def __init__(self, dim_in: int, dim_out: int, approximate: str = "none... method gelu (line 81) | def gelu(self, gate: torch.Tensor) -> torch.Tensor: method forward (line 87) | def forward(self, hidden_states): class GEGLU (line 93) | class GEGLU(nn.Module): method __init__ (line 103) | def __init__(self, dim_in: int, dim_out: int, bias: bool = True): method gelu (line 107) | def gelu(self, gate: torch.Tensor) -> torch.Tensor: method forward (line 113) | def forward(self, hidden_states, *args, **kwargs): class SwiGLU (line 126) | class SwiGLU(nn.Module): method __init__ (line 137) | def __init__(self, dim_in: int, dim_out: int, bias: bool = True): method forward (line 142) | def forward(self, hidden_states): class ApproximateGELU (line 148) | class ApproximateGELU(nn.Module): method __init__ (line 159) | def __init__(self, dim_in: int, dim_out: int, bias: bool = True): method forward (line 163) | def forward(self, x: torch.Tensor) -> torch.Tensor: FILE: src/diffusers/models/adapter.py class MultiAdapter (line 28) | class MultiAdapter(ModelMixin): method __init__ (line 41) | def __init__(self, adapters: List["T2IAdapter"]): method forward (line 76) | def forward(self, xs: torch.Tensor, adapter_weights: Optional[List[flo... method save_pretrained (line 103) | def save_pretrained( method from_pretrained (line 146) | def from_pretrained(cls, pretrained_model_path: Optional[Union[str, os... class T2IAdapter (line 217) | class T2IAdapter(ModelMixin, ConfigMixin): method __init__ (line 245) | def __init__( method forward (line 267) | def forward(self, x: torch.Tensor) -> List[torch.Tensor]: method total_downscale_factor (line 277) | def total_downscale_factor(self): method downscale_factor (line 281) | def downscale_factor(self): class FullAdapter (line 291) | class FullAdapter(nn.Module): method __init__ (line 296) | def __init__( method forward (line 322) | def forward(self, x: torch.Tensor) -> List[torch.Tensor]: class FullAdapterXL (line 341) | class FullAdapterXL(nn.Module): method __init__ (line 346) | def __init__( method forward (line 374) | def forward(self, x: torch.Tensor) -> List[torch.Tensor]: class AdapterBlock (line 391) | class AdapterBlock(nn.Module): method __init__ (line 407) | def __init__(self, in_channels: int, out_channels: int, num_res_blocks... method forward (line 422) | def forward(self, x: torch.Tensor) -> torch.Tensor: class AdapterResnetBlock (line 439) | class AdapterResnetBlock(nn.Module): method __init__ (line 448) | def __init__(self, channels: int): method forward (line 454) | def forward(self, x: torch.Tensor) -> torch.Tensor: class LightAdapter (line 469) | class LightAdapter(nn.Module): method __init__ (line 474) | def __init__( method forward (line 500) | def forward(self, x: torch.Tensor) -> List[torch.Tensor]: class LightAdapterBlock (line 516) | class LightAdapterBlock(nn.Module): method __init__ (line 532) | def __init__(self, in_channels: int, out_channels: int, num_res_blocks... method forward (line 544) | def forward(self, x: torch.Tensor) -> torch.Tensor: class LightAdapterResnetBlock (line 559) | class LightAdapterResnetBlock(nn.Module): method __init__ (line 569) | def __init__(self, channels: int): method forward (line 575) | def forward(self, x: torch.Tensor) -> torch.Tensor: FILE: src/diffusers/models/attention.py function _chunked_feed_forward (line 31) | def _chunked_feed_forward(ff: nn.Module, hidden_states: torch.Tensor, ch... class GatedSelfAttentionDense (line 47) | class GatedSelfAttentionDense(nn.Module): method __init__ (line 58) | def __init__(self, query_dim: int, context_dim: int, n_heads: int, d_h... method forward (line 75) | def forward(self, x: torch.Tensor, objs: torch.Tensor) -> torch.Tensor: class JointTransformerBlock (line 89) | class JointTransformerBlock(nn.Module): method __init__ (line 103) | def __init__(self, dim, num_attention_heads, attention_head_dim, conte... method set_chunk_feed_forward (line 154) | def set_chunk_feed_forward(self, chunk_size: Optional[int], dim: int =... method forward (line 159) | def forward( class BasicTransformerBlock (line 213) | class BasicTransformerBlock(nn.Module): method __init__ (line 248) | def __init__( method set_chunk_feed_forward (line 416) | def set_chunk_feed_forward(self, chunk_size: Optional[int], dim: int =... method forward (line 421) | def forward( class LuminaFeedForward (line 544) | class LuminaFeedForward(nn.Module): method __init__ (line 559) | def __init__( method forward (line 590) | def forward(self, x): class TemporalBasicTransformerBlock (line 595) | class TemporalBasicTransformerBlock(nn.Module): method __init__ (line 607) | def __init__( method set_chunk_feed_forward (line 660) | def set_chunk_feed_forward(self, chunk_size: Optional[int], **kwargs): method forward (line 666) | def forward( class SkipFFTransformerBlock (line 724) | class SkipFFTransformerBlock(nn.Module): method __init__ (line 725) | def __init__( method forward (line 767) | def forward(self, hidden_states, encoder_hidden_states, cross_attentio... class FreeNoiseTransformerBlock (line 797) | class FreeNoiseTransformerBlock(nn.Module): method __init__ (line 852) | def __init__( method _get_frame_indices (line 966) | def _get_frame_indices(self, num_frames: int) -> List[Tuple[int, int]]: method _get_frame_weights (line 974) | def _get_frame_weights(self, num_frames: int, weighting_scheme: str = ... method set_free_noise_properties (line 989) | def set_free_noise_properties( method set_chunk_feed_forward (line 996) | def set_chunk_feed_forward(self, chunk_size: Optional[int], dim: int =... method forward (line 1001) | def forward( class FeedForward (line 1109) | class FeedForward(nn.Module): method __init__ (line 1123) | def __init__( method forward (line 1161) | def forward(self, hidden_states: torch.Tensor, *args, **kwargs) -> tor... FILE: src/diffusers/models/attention_flax.py function _query_chunk_attention (line 23) | def _query_chunk_attention(query, key, value, precision, key_chunk_size:... function jax_memory_efficient_attention (line 74) | def jax_memory_efficient_attention( class FlaxAttention (line 122) | class FlaxAttention(nn.Module): method setup (line 153) | def setup(self): method reshape_heads_to_batch_dim (line 165) | def reshape_heads_to_batch_dim(self, tensor): method reshape_batch_dim_to_heads (line 173) | def reshape_batch_dim_to_heads(self, tensor): method __call__ (line 181) | def __call__(self, hidden_states, context=None, deterministic=True): class FlaxBasicTransformerBlock (line 244) | class FlaxBasicTransformerBlock(nn.Module): method setup (line 279) | def setup(self): method __call__ (line 306) | def __call__(self, hidden_states, context, deterministic=True): class FlaxTransformer2DModel (line 328) | class FlaxTransformer2DModel(nn.Module): method setup (line 367) | def setup(self): method __call__ (line 409) | def __call__(self, hidden_states, context, deterministic=True): class FlaxFeedForward (line 434) | class FlaxFeedForward(nn.Module): method setup (line 456) | def setup(self): method __call__ (line 462) | def __call__(self, hidden_states, deterministic=True): class FlaxGEGLU (line 468) | class FlaxGEGLU(nn.Module): method setup (line 486) | def setup(self): method __call__ (line 491) | def __call__(self, hidden_states, deterministic=True): FILE: src/diffusers/models/attention_processor.py class Attention (line 41) | class Attention(nn.Module): method __init__ (line 96) | def __init__( method set_use_npu_flash_attention (line 266) | def set_use_npu_flash_attention(self, use_npu_flash_attention: bool) -... method set_use_memory_efficient_attention_xformers (line 283) | def set_use_memory_efficient_attention_xformers( method set_attention_slice (line 390) | def set_attention_slice(self, slice_size: int) -> None: method set_processor (line 418) | def set_processor(self, processor: "AttnProcessor") -> None: method get_processor (line 438) | def get_processor(self, return_deprecated_lora: bool = False) -> "Atte... method forward (line 452) | def forward( method batch_to_head_dim (line 498) | def batch_to_head_dim(self, tensor: torch.Tensor) -> torch.Tensor: method head_to_batch_dim (line 515) | def head_to_batch_dim(self, tensor: torch.Tensor, out_dim: int = 3) ->... method get_attention_scores (line 542) | def get_attention_scores( method prepare_attention_mask (line 589) | def prepare_attention_mask( method norm_encoder_hidden_states (line 636) | def norm_encoder_hidden_states(self, encoder_hidden_states: torch.Tens... method fuse_projections (line 666) | def fuse_projections(self, fuse=True): class AttnProcessor (line 715) | class AttnProcessor: method __call__ (line 720) | def __call__( class CustomDiffusionAttnProcessor (line 787) | class CustomDiffusionAttnProcessor(nn.Module): method __init__ (line 806) | def __init__( method __call__ (line 832) | def __call__( class AttnAddedKVProcessor (line 891) | class AttnAddedKVProcessor: method __call__ (line 897) | def __call__( class AttnAddedKVProcessor2_0 (line 958) | class AttnAddedKVProcessor2_0: method __init__ (line 964) | def __init__(self): method __call__ (line 970) | def __call__( class JointAttnProcessor2_0 (line 1034) | class JointAttnProcessor2_0: method __init__ (line 1037) | def __init__(self): method __call__ (line 1041) | def __call__( class PAGJointAttnProcessor2_0 (line 1117) | class PAGJointAttnProcessor2_0: method __init__ (line 1120) | def __init__(self): method __call__ (line 1126) | def __call__( class PAGCFGJointAttnProcessor2_0 (line 1272) | class PAGCFGJointAttnProcessor2_0: method __init__ (line 1275) | def __init__(self): method __call__ (line 1281) | def __call__( class FusedJointAttnProcessor2_0 (line 1437) | class FusedJointAttnProcessor2_0: method __init__ (line 1440) | def __init__(self): method __call__ (line 1444) | def __call__( class AuraFlowAttnProcessor2_0 (line 1516) | class AuraFlowAttnProcessor2_0: method __init__ (line 1519) | def __init__(self): method __call__ (line 1525) | def __call__( class FusedAuraFlowAttnProcessor2_0 (line 1609) | class FusedAuraFlowAttnProcessor2_0: method __init__ (line 1612) | def __init__(self): method __call__ (line 1618) | def __call__( function apply_rope (line 1707) | def apply_rope(xq, xk, freqs_cis): class FluxSingleAttnProcessor2_0 (line 1715) | class FluxSingleAttnProcessor2_0: method __init__ (line 1720) | def __init__(self): method __call__ (line 1724) | def __call__( class FluxAttnProcessor2_0 (line 1790) | class FluxAttnProcessor2_0: method __init__ (line 1793) | def __init__(self): method __call__ (line 1797) | def __call__( class XFormersAttnAddedKVProcessor (line 1901) | class XFormersAttnAddedKVProcessor: method __init__ (line 1913) | def __init__(self, attention_op: Optional[Callable] = None): method __call__ (line 1916) | def __call__( class XFormersAttnProcessor (line 1972) | class XFormersAttnProcessor: method __init__ (line 1984) | def __init__(self, attention_op: Optional[Callable] = None): method __call__ (line 1987) | def __call__( class AttnProcessorNPU (line 2066) | class AttnProcessorNPU: method __init__ (line 2074) | def __init__(self): method __call__ (line 2078) | def __call__( class AttnProcessor2_0 (line 2175) | class AttnProcessor2_0: method __init__ (line 2180) | def __init__(self): method __call__ (line 2184) | def __call__( class StableAudioAttnProcessor2_0 (line 2269) | class StableAudioAttnProcessor2_0: method __init__ (line 2275) | def __init__(self): method apply_partial_rotary_emb (line 2281) | def apply_partial_rotary_emb( method __call__ (line 2296) | def __call__( class HunyuanAttnProcessor2_0 (line 2400) | class HunyuanAttnProcessor2_0: method __init__ (line 2406) | def __init__(self): method __call__ (line 2410) | def __call__( class FusedHunyuanAttnProcessor2_0 (line 2498) | class FusedHunyuanAttnProcessor2_0: method __init__ (line 2505) | def __init__(self): method __call__ (line 2511) | def __call__( class PAGHunyuanAttnProcessor2_0 (line 2601) | class PAGHunyuanAttnProcessor2_0: method __init__ (line 2608) | def __init__(self): method __call__ (line 2614) | def __call__( class PAGCFGHunyuanAttnProcessor2_0 (line 2724) | class PAGCFGHunyuanAttnProcessor2_0: method __init__ (line 2731) | def __init__(self): method __call__ (line 2737) | def __call__( class LuminaAttnProcessor2_0 (line 2848) | class LuminaAttnProcessor2_0: method __init__ (line 2854) | def __init__(self): method __call__ (line 2858) | def __call__( class FusedAttnProcessor2_0 (line 2944) | class FusedAttnProcessor2_0: method __init__ (line 2957) | def __init__(self): method __call__ (line 2963) | def __call__( class CustomDiffusionXFormersAttnProcessor (line 3050) | class CustomDiffusionXFormersAttnProcessor(nn.Module): method __init__ (line 3073) | def __init__( method __call__ (line 3101) | def __call__( class CustomDiffusionAttnProcessor2_0 (line 3166) | class CustomDiffusionAttnProcessor2_0(nn.Module): method __init__ (line 3186) | def __init__( method __call__ (line 3212) | def __call__( class SlicedAttnProcessor (line 3280) | class SlicedAttnProcessor: method __init__ (line 3290) | def __init__(self, slice_size: int): method __call__ (line 3293) | def __call__( class SlicedAttnAddedKVProcessor (line 3367) | class SlicedAttnAddedKVProcessor: method __init__ (line 3377) | def __init__(self, slice_size): method __call__ (line 3380) | def __call__( class SpatialNorm (line 3459) | class SpatialNorm(nn.Module): method __init__ (line 3470) | def __init__( method forward (line 3480) | def forward(self, f: torch.Tensor, zq: torch.Tensor) -> torch.Tensor: class IPAdapterAttnProcessor (line 3488) | class IPAdapterAttnProcessor(nn.Module): method __init__ (line 3503) | def __init__(self, hidden_size, cross_attention_dim=None, num_tokens=(... method __call__ (line 3526) | def __call__( class IPAdapterAttnProcessor2_0 (line 3686) | class IPAdapterAttnProcessor2_0(torch.nn.Module): method __init__ (line 3701) | def __init__(self, hidden_size, cross_attention_dim=None, num_tokens=(... method __call__ (line 3729) | def __call__( class PAGIdentitySelfAttnProcessor2_0 (line 3916) | class PAGIdentitySelfAttnProcessor2_0: method __init__ (line 3922) | def __init__(self): method __call__ (line 3928) | def __call__( class PAGCFGIdentitySelfAttnProcessor2_0 (line 4015) | class PAGCFGIdentitySelfAttnProcessor2_0: method __init__ (line 4021) | def __init__(self): method __call__ (line 4027) | def __call__( class LoRAAttnProcessor (line 4118) | class LoRAAttnProcessor: method __init__ (line 4119) | def __init__(self): class LoRAAttnProcessor2_0 (line 4123) | class LoRAAttnProcessor2_0: method __init__ (line 4124) | def __init__(self): class LoRAXFormersAttnProcessor (line 4128) | class LoRAXFormersAttnProcessor: method __init__ (line 4129) | def __init__(self): class LoRAAttnAddedKVProcessor (line 4133) | class LoRAAttnAddedKVProcessor: method __init__ (line 4134) | def __init__(self): FILE: src/diffusers/models/autoencoders/autoencoder_asym_kl.py class AsymmetricAutoencoderKL (line 26) | class AsymmetricAutoencoderKL(ModelMixin, ConfigMixin): method __init__ (line 64) | def __init__( method encode (line 115) | def encode(self, x: torch.Tensor, return_dict: bool = True) -> Union[A... method _decode (line 125) | def _decode( method decode (line 141) | def decode( method forward (line 156) | def forward( FILE: src/diffusers/models/autoencoders/autoencoder_kl.py class AutoencoderKL (line 36) | class AutoencoderKL(ModelMixin, ConfigMixin, FromOriginalModelMixin): method __init__ (line 75) | def __init__( method _set_gradient_checkpointing (line 139) | def _set_gradient_checkpointing(self, module, value=False): method enable_tiling (line 143) | def enable_tiling(self, use_tiling: bool = True): method disable_tiling (line 151) | def disable_tiling(self): method enable_slicing (line 158) | def enable_slicing(self): method disable_slicing (line 165) | def disable_slicing(self): method attn_processors (line 174) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 198) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 233) | def set_default_attn_processor(self): method encode (line 249) | def encode( method _decode (line 285) | def _decode(self, z: torch.Tensor, return_dict: bool = True) -> Union[... method decode (line 300) | def decode( method blend_v (line 328) | def blend_v(self, a: torch.Tensor, b: torch.Tensor, blend_extent: int)... method blend_h (line 334) | def blend_h(self, a: torch.Tensor, b: torch.Tensor, blend_extent: int)... method tiled_encode (line 340) | def tiled_encode(self, x: torch.Tensor, return_dict: bool = True) -> A... method tiled_decode (line 395) | def tiled_decode(self, z: torch.Tensor, return_dict: bool = True) -> U... method forward (line 444) | def forward( method fuse_qkv_projections (line 473) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 499) | def unfuse_qkv_projections(self): FILE: src/diffusers/models/autoencoders/autoencoder_kl_cogvideox.py class CogVideoXSafeConv3d (line 38) | class CogVideoXSafeConv3d(nn.Conv3d): method forward (line 43) | def forward(self, input: torch.Tensor) -> torch.Tensor: class CogVideoXCausalConv3d (line 67) | class CogVideoXCausalConv3d(nn.Module): method __init__ (line 79) | def __init__( method fake_context_parallel_forward (line 120) | def fake_context_parallel_forward(self, inputs: torch.Tensor) -> torch... method _clear_fake_context_parallel_cache (line 136) | def _clear_fake_context_parallel_cache(self): method forward (line 140) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: class CogVideoXSpatialNorm3D (line 154) | class CogVideoXSpatialNorm3D(nn.Module): method __init__ (line 168) | def __init__( method forward (line 179) | def forward(self, f: torch.Tensor, zq: torch.Tensor) -> torch.Tensor: class CogVideoXResnetBlock3D (line 195) | class CogVideoXResnetBlock3D(nn.Module): method __init__ (line 213) | def __init__( method forward (line 272) | def forward( class CogVideoXDownBlock3D (line 307) | class CogVideoXDownBlock3D(nn.Module): method __init__ (line 328) | def __init__( method forward (line 375) | def forward( class CogVideoXMidBlock3D (line 403) | class CogVideoXMidBlock3D(nn.Module): method __init__ (line 421) | def __init__( method forward (line 454) | def forward( class CogVideoXUpBlock3D (line 478) | class CogVideoXUpBlock3D(nn.Module): method __init__ (line 498) | def __init__( method forward (line 547) | def forward( class CogVideoXEncoder3D (line 576) | class CogVideoXEncoder3D(nn.Module): method __init__ (line 602) | def __init__( method forward (line 675) | def forward(self, sample: torch.Tensor, temb: Optional[torch.Tensor] =... class CogVideoXDecoder3D (line 712) | class CogVideoXDecoder3D(nn.Module): method __init__ (line 738) | def __init__( method forward (line 818) | def forward(self, sample: torch.Tensor, temb: Optional[torch.Tensor] =... class AutoencoderKLCogVideoX (line 855) | class AutoencoderKLCogVideoX(ModelMixin, ConfigMixin, FromOriginalModelM... method __init__ (line 891) | def __init__( method _set_gradient_checkpointing (line 962) | def _set_gradient_checkpointing(self, module, value=False): method clear_fake_context_parallel_cache (line 966) | def clear_fake_context_parallel_cache(self): method encode (line 973) | def encode( method decode (line 997) | def decode(self, z: torch.FloatTensor, return_dict: bool = True) -> Un... method forward (line 1019) | def forward( FILE: src/diffusers/models/autoencoders/autoencoder_kl_temporal_decoder.py class TemporalDecoder (line 29) | class TemporalDecoder(nn.Module): method __init__ (line 30) | def __init__( method forward (line 87) | def forward( class AutoencoderKLTemporalDecoder (line 164) | class AutoencoderKLTemporalDecoder(ModelMixin, ConfigMixin): method __init__ (line 197) | def __init__( method _set_gradient_checkpointing (line 239) | def _set_gradient_checkpointing(self, module, value=False): method attn_processors (line 245) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 269) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 303) | def set_default_attn_processor(self): method encode (line 317) | def encode( method decode (line 344) | def decode( method forward (line 373) | def forward( FILE: src/diffusers/models/autoencoders/autoencoder_oobleck.py class Snake1d (line 30) | class Snake1d(nn.Module): method __init__ (line 35) | def __init__(self, hidden_dim, logscale=True): method forward (line 44) | def forward(self, hidden_states): class OobleckResidualUnit (line 56) | class OobleckResidualUnit(nn.Module): method __init__ (line 61) | def __init__(self, dimension: int = 16, dilation: int = 1): method forward (line 70) | def forward(self, hidden_state): class OobleckEncoderBlock (line 93) | class OobleckEncoderBlock(nn.Module): method __init__ (line 96) | def __init__(self, input_dim, output_dim, stride: int = 1): method forward (line 107) | def forward(self, hidden_state): class OobleckDecoderBlock (line 116) | class OobleckDecoderBlock(nn.Module): method __init__ (line 119) | def __init__(self, input_dim, output_dim, stride: int = 1): method forward (line 136) | def forward(self, hidden_state): class OobleckDiagonalGaussianDistribution (line 146) | class OobleckDiagonalGaussianDistribution(object): method __init__ (line 147) | def __init__(self, parameters: torch.Tensor, deterministic: bool = Fal... method sample (line 155) | def sample(self, generator: Optional[torch.Generator] = None) -> torch... method kl (line 166) | def kl(self, other: "OobleckDiagonalGaussianDistribution" = None) -> t... method mode (line 182) | def mode(self) -> torch.Tensor: class AutoencoderOobleckOutput (line 187) | class AutoencoderOobleckOutput(BaseOutput): class OobleckDecoderOutput (line 202) | class OobleckDecoderOutput(BaseOutput): class OobleckEncoder (line 214) | class OobleckEncoder(nn.Module): method __init__ (line 217) | def __init__(self, encoder_hidden_size, audio_channels, downsampling_r... method forward (line 242) | def forward(self, hidden_state): class OobleckDecoder (line 254) | class OobleckDecoder(nn.Module): method __init__ (line 257) | def __init__(self, channels, input_channels, audio_channels, upsamplin... method forward (line 282) | def forward(self, hidden_state): class AutoencoderOobleck (line 294) | class AutoencoderOobleck(ModelMixin, ConfigMixin): method __init__ (line 322) | def __init__( method enable_slicing (line 358) | def enable_slicing(self): method disable_slicing (line 365) | def disable_slicing(self): method encode (line 373) | def encode( method _decode (line 401) | def _decode(self, z: torch.Tensor, return_dict: bool = True) -> Union[... method decode (line 410) | def decode( method forward (line 438) | def forward( FILE: src/diffusers/models/autoencoders/autoencoder_tiny.py class AutoencoderTinyOutput (line 29) | class AutoencoderTinyOutput(BaseOutput): class AutoencoderTiny (line 41) | class AutoencoderTiny(ModelMixin, ConfigMixin): method __init__ (line 98) | def __init__( method _set_gradient_checkpointing (line 157) | def _set_gradient_checkpointing(self, module, value: bool = False) -> ... method scale_latents (line 161) | def scale_latents(self, x: torch.Tensor) -> torch.Tensor: method unscale_latents (line 165) | def unscale_latents(self, x: torch.Tensor) -> torch.Tensor: method enable_slicing (line 169) | def enable_slicing(self) -> None: method disable_slicing (line 176) | def disable_slicing(self) -> None: method enable_tiling (line 183) | def enable_tiling(self, use_tiling: bool = True) -> None: method disable_tiling (line 191) | def disable_tiling(self) -> None: method _tiled_encode (line 198) | def _tiled_encode(self, x: torch.Tensor) -> torch.Tensor: method _tiled_decode (line 246) | def _tiled_decode(self, x: torch.Tensor) -> torch.Tensor: method encode (line 294) | def encode(self, x: torch.Tensor, return_dict: bool = True) -> Union[A... method decode (line 309) | def decode( method forward (line 323) | def forward( FILE: src/diffusers/models/autoencoders/consistency_decoder_vae.py class ConsistencyDecoderVAEOutput (line 39) | class ConsistencyDecoderVAEOutput(BaseOutput): class ConsistencyDecoderVAE (line 52) | class ConsistencyDecoderVAE(ModelMixin, ConfigMixin): method __init__ (line 72) | def __init__( method enable_tiling (line 169) | def enable_tiling(self, use_tiling: bool = True): method disable_tiling (line 178) | def disable_tiling(self): method enable_slicing (line 186) | def enable_slicing(self): method disable_slicing (line 194) | def disable_slicing(self): method attn_processors (line 203) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 227) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 262) | def set_default_attn_processor(self): method encode (line 278) | def encode( method decode (line 313) | def decode( method blend_v (line 360) | def blend_v(self, a: torch.Tensor, b: torch.Tensor, blend_extent: int)... method blend_h (line 367) | def blend_h(self, a: torch.Tensor, b: torch.Tensor, blend_extent: int)... method tiled_encode (line 373) | def tiled_encode(self, x: torch.Tensor, return_dict: bool = True) -> U... method forward (line 428) | def forward( FILE: src/diffusers/models/autoencoders/vae.py class DecoderOutput (line 34) | class DecoderOutput(BaseOutput): class Encoder (line 47) | class Encoder(nn.Module): method __init__ (line 71) | def __init__( method forward (line 140) | def forward(self, sample: torch.Tensor) -> torch.Tensor: class Decoder (line 185) | class Decoder(nn.Module): method __init__ (line 208) | def __init__( method forward (line 284) | def forward( class UpSample (line 350) | class UpSample(nn.Module): method __init__ (line 361) | def __init__( method forward (line 371) | def forward(self, x: torch.Tensor) -> torch.Tensor: class MaskConditionEncoder (line 378) | class MaskConditionEncoder(nn.Module): method __init__ (line 383) | def __init__( method forward (line 420) | def forward(self, x: torch.Tensor, mask=None) -> torch.Tensor: class MaskConditionDecoder (line 431) | class MaskConditionDecoder(nn.Module): method __init__ (line 454) | def __init__( method forward (line 535) | def forward( class VectorQuantizer (line 645) | class VectorQuantizer(nn.Module): method __init__ (line 654) | def __init__( method remap_to_used (line 691) | def remap_to_used(self, inds: torch.LongTensor) -> torch.LongTensor: method unmap_to_all (line 705) | def unmap_to_all(self, inds: torch.LongTensor) -> torch.LongTensor: method forward (line 715) | def forward(self, z: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor... method get_codebook_entry (line 749) | def get_codebook_entry(self, indices: torch.LongTensor, shape: Tuple[i... class DiagonalGaussianDistribution (line 767) | class DiagonalGaussianDistribution(object): method __init__ (line 768) | def __init__(self, parameters: torch.Tensor, deterministic: bool = Fal... method sample (line 780) | def sample(self, generator: Optional[torch.Generator] = None) -> torch... method kl (line 791) | def kl(self, other: "DiagonalGaussianDistribution" = None) -> torch.Te... method nll (line 810) | def nll(self, sample: torch.Tensor, dims: Tuple[int, ...] = [1, 2, 3])... method mode (line 819) | def mode(self) -> torch.Tensor: class EncoderTiny (line 823) | class EncoderTiny(nn.Module): method __init__ (line 841) | def __init__( method forward (line 877) | def forward(self, x: torch.Tensor) -> torch.Tensor: class DecoderTiny (line 899) | class DecoderTiny(nn.Module): method __init__ (line 919) | def __init__( method forward (line 960) | def forward(self, x: torch.Tensor) -> torch.Tensor: FILE: src/diffusers/models/autoencoders/vq_model.py class VQEncoderOutput (line 28) | class VQEncoderOutput(BaseOutput): class VQModel (line 40) | class VQModel(ModelMixin, ConfigMixin): method __init__ (line 75) | def __init__( method encode (line 130) | def encode(self, x: torch.Tensor, return_dict: bool = True) -> VQEncod... method decode (line 140) | def decode( method forward (line 160) | def forward( FILE: src/diffusers/models/controlnet.py class ControlNetOutput (line 47) | class ControlNetOutput(BaseOutput): class ControlNetConditioningEmbedding (line 66) | class ControlNetConditioningEmbedding(nn.Module): method __init__ (line 76) | def __init__( method forward (line 98) | def forward(self, conditioning): class ControlNetModel (line 111) | class ControlNetModel(ModelMixin, ConfigMixin, FromOriginalModelMixin): method __init__ (line 183) | def __init__( method from_unet (line 445) | def from_unet( method attn_processors (line 522) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 546) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 581) | def set_default_attn_processor(self): method set_attention_slice (line 597) | def set_attention_slice(self, slice_size: Union[str, int, List[int]]) ... method _set_gradient_checkpointing (line 662) | def _set_gradient_checkpointing(self, module, value: bool = False) -> ... method forward (line 666) | def forward( function zero_module (line 867) | def zero_module(module): FILE: src/diffusers/models/controlnet_flax.py class FlaxControlNetOutput (line 34) | class FlaxControlNetOutput(BaseOutput): class FlaxControlNetConditioningEmbedding (line 47) | class FlaxControlNetConditioningEmbedding(nn.Module): method setup (line 52) | def setup(self) -> None: method __call__ (line 90) | def __call__(self, conditioning: jnp.ndarray) -> jnp.ndarray: class FlaxControlNetModel (line 104) | class FlaxControlNetModel(nn.Module, FlaxModelMixin, ConfigMixin): method init_weights (line 172) | def init_weights(self, rng: jax.Array) -> FrozenDict: method setup (line 186) | def setup(self) -> None: method __call__ (line 314) | def __call__( FILE: src/diffusers/models/controlnet_hunyuan.py class HunyuanControlNetOutput (line 37) | class HunyuanControlNetOutput(BaseOutput): class HunyuanDiT2DControlNetModel (line 41) | class HunyuanDiT2DControlNetModel(ModelMixin, ConfigMixin): method __init__ (line 43) | def __init__( method attn_processors (line 119) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 142) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method from_transformer (line 175) | def from_transformer( method forward (line 216) | def forward( class HunyuanDiT2DMultiControlNetModel (line 312) | class HunyuanDiT2DMultiControlNetModel(ModelMixin): method __init__ (line 325) | def __init__(self, controlnets): method forward (line 329) | def forward( FILE: src/diffusers/models/controlnet_sd3.py class SD3ControlNetOutput (line 37) | class SD3ControlNetOutput(BaseOutput): class SD3ControlNetModel (line 41) | class SD3ControlNetModel(ModelMixin, ConfigMixin, PeftAdapterMixin, From... method __init__ (line 45) | def __init__( method enable_forward_chunking (line 110) | def enable_forward_chunking(self, chunk_size: Optional[int] = None, di... method attn_processors (line 141) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 165) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method fuse_qkv_projections (line 200) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 226) | def unfuse_qkv_projections(self): method _set_gradient_checkpointing (line 239) | def _set_gradient_checkpointing(self, module, value=False): method from_transformer (line 244) | def from_transformer(cls, transformer, num_layers=12, load_weights_fro... method forward (line 259) | def forward( class SD3MultiControlNetModel (line 368) | class SD3MultiControlNetModel(ModelMixin): method __init__ (line 381) | def __init__(self, controlnets): method forward (line 385) | def forward( FILE: src/diffusers/models/controlnet_sparsectrl.py class SparseControlNetOutput (line 43) | class SparseControlNetOutput(BaseOutput): class SparseControlNetConditioningEmbedding (line 62) | class SparseControlNetConditioningEmbedding(nn.Module): method __init__ (line 63) | def __init__( method forward (line 84) | def forward(self, conditioning: torch.Tensor) -> torch.Tensor: class SparseControlNetModel (line 96) | class SparseControlNetModel(ModelMixin, ConfigMixin, FromOriginalModelMi... method __init__ (line 161) | def __init__( method from_unet (line 388) | def from_unet( method attn_processors (line 453) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 477) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 512) | def set_default_attn_processor(self): method set_attention_slice (line 528) | def set_attention_slice(self, slice_size: Union[str, int, List[int]]) ... method _set_gradient_checkpointing (line 593) | def _set_gradient_checkpointing(self, module, value: bool = False) -> ... method forward (line 597) | def forward( function zero_module (line 786) | def zero_module(module: nn.Module) -> nn.Module: FILE: src/diffusers/models/controlnet_xs.py class ControlNetXSOutput (line 53) | class ControlNetXSOutput(BaseOutput): class DownBlockControlNetXSAdapter (line 66) | class DownBlockControlNetXSAdapter(nn.Module): method __init__ (line 70) | def __init__( class MidBlockControlNetXSAdapter (line 86) | class MidBlockControlNetXSAdapter(nn.Module): method __init__ (line 90) | def __init__(self, midblock: UNetMidBlock2DCrossAttn, base_to_ctrl: nn... class UpBlockControlNetXSAdapter (line 97) | class UpBlockControlNetXSAdapter(nn.Module): method __init__ (line 100) | def __init__(self, ctrl_to_base: nn.ModuleList): function get_down_block_adapter (line 105) | def get_down_block_adapter( function get_mid_block_adapter (line 196) | def get_mid_block_adapter( function get_up_block_adapter (line 231) | def get_up_block_adapter( class ControlNetXSAdapter (line 245) | class ControlNetXSAdapter(ModelMixin, ConfigMixin): method __init__ (line 293) | def __init__( method from_unet (line 430) | def from_unet( method forward (line 506) | def forward(self, *args, **kwargs): class UNetControlNetXSModel (line 512) | class UNetControlNetXSModel(ModelMixin, ConfigMixin): method __init__ (line 529) | def __init__( method from_unet (line 720) | def from_unet( method freeze_unet_params (line 838) | def freeze_unet_params(self) -> None: method _set_gradient_checkpointing (line 867) | def _set_gradient_checkpointing(self, module, value=False): method attn_processors (line 873) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 897) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 932) | def set_default_attn_processor(self): method enable_freeu (line 948) | def enable_freeu(self, s1: float, s2: float, b1: float, b2: float): method disable_freeu (line 973) | def disable_freeu(self): method fuse_qkv_projections (line 982) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 1008) | def unfuse_qkv_projections(self): method forward (line 1021) | def forward( class ControlNetXSCrossAttnDownBlock2D (line 1224) | class ControlNetXSCrossAttnDownBlock2D(nn.Module): method __init__ (line 1225) | def __init__( method from_modules (line 1344) | def from_modules(cls, base_downblock: CrossAttnDownBlock2D, ctrl_downb... method freeze_base_params (line 1409) | def freeze_base_params(self) -> None: method forward (line 1426) | def forward( class ControlNetXSCrossAttnMidBlock2D (line 1540) | class ControlNetXSCrossAttnMidBlock2D(nn.Module): method __init__ (line 1541) | def __init__( method from_modules (line 1594) | def from_modules( method freeze_base_params (line 1642) | def freeze_base_params(self) -> None: method forward (line 1653) | def forward( class ControlNetXSCrossAttnUpBlock2D (line 1690) | class ControlNetXSCrossAttnUpBlock2D(nn.Module): method __init__ (line 1691) | def __init__( method from_modules (line 1763) | def from_modules(cls, base_upblock: CrossAttnUpBlock2D, ctrl_upblock: ... method freeze_base_params (line 1821) | def freeze_base_params(self) -> None: method forward (line 1838) | def forward( function make_zero_conv (line 1927) | def make_zero_conv(in_channels, out_channels=None): function zero_module (line 1931) | def zero_module(module): function find_largest_factor (line 1937) | def find_largest_factor(number, max_factor): FILE: src/diffusers/models/downsampling.py class Downsample1D (line 26) | class Downsample1D(nn.Module): method __init__ (line 42) | def __init__( method forward (line 64) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: class Downsample2D (line 69) | class Downsample2D(nn.Module): method __init__ (line 85) | def __init__( method forward (line 132) | def forward(self, hidden_states: torch.Tensor, *args, **kwargs) -> tor... class FirDownsample2D (line 152) | class FirDownsample2D(nn.Module): method __init__ (line 166) | def __init__( method _downsample_2d (line 181) | def _downsample_2d( method forward (line 247) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class KDownsample2D (line 258) | class KDownsample2D(nn.Module): method __init__ (line 265) | def __init__(self, pad_mode: str = "reflect"): method forward (line 272) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: class CogVideoXDownsample3D (line 288) | class CogVideoXDownsample3D(nn.Module): method __init__ (line 308) | def __init__( method forward (line 322) | def forward(self, x: torch.Tensor) -> torch.Tensor: function downsample_2d (line 356) | def downsample_2d( FILE: src/diffusers/models/embeddings.py function get_timestep_embedding (line 27) | def get_timestep_embedding( function get_3d_sincos_pos_embed (line 81) | def get_3d_sincos_pos_embed( function get_2d_sincos_pos_embed (line 128) | def get_2d_sincos_pos_embed( function get_2d_sincos_pos_embed_from_grid (line 150) | def get_2d_sincos_pos_embed_from_grid(embed_dim, grid): function get_1d_sincos_pos_embed_from_grid (line 162) | def get_1d_sincos_pos_embed_from_grid(embed_dim, pos): class PatchEmbed (line 183) | class PatchEmbed(nn.Module): method __init__ (line 186) | def __init__( method cropped_pos_embed (line 237) | def cropped_pos_embed(self, height, width): method forward (line 260) | def forward(self, latent): class LuminaPatchEmbed (line 291) | class LuminaPatchEmbed(nn.Module): method __init__ (line 294) | def __init__(self, patch_size=2, in_channels=4, embed_dim=768, bias=Tr... method forward (line 303) | def forward(self, x, freqs_cis): class CogVideoXPatchEmbed (line 337) | class CogVideoXPatchEmbed(nn.Module): method __init__ (line 338) | def __init__( method forward (line 354) | def forward(self, text_embeds: torch.Tensor, image_embeds: torch.Tensor): function get_2d_rotary_pos_embed (line 377) | def get_2d_rotary_pos_embed(embed_dim, crops_coords, grid_size, use_real... function get_2d_rotary_pos_embed_from_grid (line 405) | def get_2d_rotary_pos_embed_from_grid(embed_dim, grid, use_real=False): function get_2d_rotary_pos_embed_lumina (line 425) | def get_2d_rotary_pos_embed_lumina(embed_dim, len_h, len_w, linear_facto... function get_1d_rotary_pos_embed (line 441) | def get_1d_rotary_pos_embed( function apply_rotary_emb (line 495) | def apply_rotary_emb( class TimestepEmbedding (line 543) | class TimestepEmbedding(nn.Module): method __init__ (line 544) | def __init__( method forward (line 576) | def forward(self, sample, condition=None): class Timesteps (line 591) | class Timesteps(nn.Module): method __init__ (line 592) | def __init__(self, num_channels: int, flip_sin_to_cos: bool, downscale... method forward (line 599) | def forward(self, timesteps): class GaussianFourierProjection (line 610) | class GaussianFourierProjection(nn.Module): method __init__ (line 613) | def __init__( method forward (line 628) | def forward(self, x): class SinusoidalPositionalEmbedding (line 641) | class SinusoidalPositionalEmbedding(nn.Module): method __init__ (line 653) | def __init__(self, embed_dim: int, max_seq_length: int = 32): method forward (line 662) | def forward(self, x): class ImagePositionalEmbeddings (line 668) | class ImagePositionalEmbeddings(nn.Module): method __init__ (line 692) | def __init__( method forward (line 710) | def forward(self, index): class LabelEmbedding (line 733) | class LabelEmbedding(nn.Module): method __init__ (line 743) | def __init__(self, num_classes, hidden_size, dropout_prob): method token_drop (line 750) | def token_drop(self, labels, force_drop_ids=None): method forward (line 761) | def forward(self, labels: torch.LongTensor, force_drop_ids=None): class TextImageProjection (line 769) | class TextImageProjection(nn.Module): method __init__ (line 770) | def __init__( method forward (line 783) | def forward(self, text_embeds: torch.Tensor, image_embeds: torch.Tensor): class ImageProjection (line 796) | class ImageProjection(nn.Module): method __init__ (line 797) | def __init__( method forward (line 809) | def forward(self, image_embeds: torch.Tensor): class IPAdapterFullImageProjection (line 819) | class IPAdapterFullImageProjection(nn.Module): method __init__ (line 820) | def __init__(self, image_embed_dim=1024, cross_attention_dim=1024): method forward (line 827) | def forward(self, image_embeds: torch.Tensor): class IPAdapterFaceIDImageProjection (line 831) | class IPAdapterFaceIDImageProjection(nn.Module): method __init__ (line 832) | def __init__(self, image_embed_dim=1024, cross_attention_dim=1024, mul... method forward (line 841) | def forward(self, image_embeds: torch.Tensor): class CombinedTimestepLabelEmbeddings (line 847) | class CombinedTimestepLabelEmbeddings(nn.Module): method __init__ (line 848) | def __init__(self, num_classes, embedding_dim, class_dropout_prob=0.1): method forward (line 855) | def forward(self, timestep, class_labels, hidden_dtype=None): class CombinedTimestepTextProjEmbeddings (line 866) | class CombinedTimestepTextProjEmbeddings(nn.Module): method __init__ (line 867) | def __init__(self, embedding_dim, pooled_projection_dim): method forward (line 874) | def forward(self, timestep, pooled_projection): class CombinedTimestepGuidanceTextProjEmbeddings (line 885) | class CombinedTimestepGuidanceTextProjEmbeddings(nn.Module): method __init__ (line 886) | def __init__(self, embedding_dim, pooled_projection_dim): method forward (line 894) | def forward(self, timestep, guidance, pooled_projection): class HunyuanDiTAttentionPool (line 909) | class HunyuanDiTAttentionPool(nn.Module): method __init__ (line 912) | def __init__(self, spacial_dim: int, embed_dim: int, num_heads: int, o... method forward (line 921) | def forward(self, x): class HunyuanCombinedTimestepTextSizeStyleEmbedding (line 949) | class HunyuanCombinedTimestepTextSizeStyleEmbedding(nn.Module): method __init__ (line 950) | def __init__( method forward (line 984) | def forward(self, timestep, encoder_hidden_states, image_meta_size, st... class LuminaCombinedTimestepCaptionEmbedding (line 1010) | class LuminaCombinedTimestepCaptionEmbedding(nn.Module): method __init__ (line 1011) | def __init__(self, hidden_size=4096, cross_attention_dim=2048, frequen... method forward (line 1028) | def forward(self, timestep, caption_feat, caption_mask): class TextTimeEmbedding (line 1044) | class TextTimeEmbedding(nn.Module): method __init__ (line 1045) | def __init__(self, encoder_dim: int, time_embed_dim: int, num_heads: i... method forward (line 1052) | def forward(self, hidden_states): class TextImageTimeEmbedding (line 1060) | class TextImageTimeEmbedding(nn.Module): method __init__ (line 1061) | def __init__(self, text_embed_dim: int = 768, image_embed_dim: int = 7... method forward (line 1067) | def forward(self, text_embeds: torch.Tensor, image_embeds: torch.Tensor): class ImageTimeEmbedding (line 1078) | class ImageTimeEmbedding(nn.Module): method __init__ (line 1079) | def __init__(self, image_embed_dim: int = 768, time_embed_dim: int = 1... method forward (line 1084) | def forward(self, image_embeds: torch.Tensor): class ImageHintTimeEmbedding (line 1091) | class ImageHintTimeEmbedding(nn.Module): method __init__ (line 1092) | def __init__(self, image_embed_dim: int = 768, time_embed_dim: int = 1... method forward (line 1114) | def forward(self, image_embeds: torch.Tensor, hint: torch.Tensor): class AttentionPooling (line 1122) | class AttentionPooling(nn.Module): method __init__ (line 1125) | def __init__(self, num_heads, embed_dim, dtype=None): method forward (line 1135) | def forward(self, x): function get_fourier_embeds_from_boundingbox (line 1172) | def get_fourier_embeds_from_boundingbox(embed_dim, box): class GLIGENTextBoundingboxProjection (line 1193) | class GLIGENTextBoundingboxProjection(nn.Module): method __init__ (line 1194) | def __init__(self, positive_len, out_dim, feature_type="text-only", fo... method forward (line 1235) | def forward( class PixArtAlphaCombinedTimestepSizeEmbeddings (line 1286) | class PixArtAlphaCombinedTimestepSizeEmbeddings(nn.Module): method __init__ (line 1294) | def __init__(self, embedding_dim, size_emb_dim, use_additional_conditi... method forward (line 1307) | def forward(self, timestep, resolution, aspect_ratio, batch_size, hidd... class PixArtAlphaTextProjection (line 1323) | class PixArtAlphaTextProjection(nn.Module): method __init__ (line 1330) | def __init__(self, in_features, hidden_size, out_features=None, act_fn... method forward (line 1345) | def forward(self, caption): class IPAdapterPlusImageProjectionBlock (line 1352) | class IPAdapterPlusImageProjectionBlock(nn.Module): method __init__ (line 1353) | def __init__( method forward (line 1376) | def forward(self, x, latents, residual): class IPAdapterPlusImageProjection (line 1385) | class IPAdapterPlusImageProjection(nn.Module): method __init__ (line 1401) | def __init__( method forward (line 1424) | def forward(self, x: torch.Tensor) -> torch.Tensor: class IPAdapterFaceIDPlusImageProjection (line 1444) | class IPAdapterFaceIDPlusImageProjection(nn.Module): method __init__ (line 1461) | def __init__( method forward (line 1496) | def forward(self, id_embeds: torch.Tensor) -> torch.Tensor: class MultiIPAdapterImageProjection (line 1524) | class MultiIPAdapterImageProjection(nn.Module): method __init__ (line 1525) | def __init__(self, IPAdapterImageProjectionLayers: Union[List[nn.Modul... method forward (line 1529) | def forward(self, image_embeds: List[torch.Tensor]): FILE: src/diffusers/models/embeddings_flax.py function get_sinusoidal_embeddings (line 20) | def get_sinusoidal_embeddings( class FlaxTimestepEmbedding (line 58) | class FlaxTimestepEmbedding(nn.Module): method __call__ (line 73) | def __call__(self, temb): class FlaxTimesteps (line 80) | class FlaxTimesteps(nn.Module): method __call__ (line 94) | def __call__(self, timesteps): FILE: src/diffusers/models/lora.py function text_encoder_attn_modules (line 41) | def text_encoder_attn_modules(text_encoder): function text_encoder_mlp_modules (line 55) | def text_encoder_mlp_modules(text_encoder): function adjust_lora_scale_text_encoder (line 69) | def adjust_lora_scale_text_encoder(text_encoder, lora_scale: float = 1.0): class PatchedLoraProjection (line 83) | class PatchedLoraProjection(torch.nn.Module): method __init__ (line 84) | def __init__(self, regular_linear_layer, lora_scale=1, network_alpha=N... method state_dict (line 111) | def state_dict(self, *args, destination=None, prefix="", keep_vars=Fal... method _fuse_lora (line 119) | def _fuse_lora(self, lora_scale=1.0, safe_fusing=False): method _unfuse_lora (line 151) | def _unfuse_lora(self): method forward (line 167) | def forward(self, input): class LoRALinearLayer (line 175) | class LoRALinearLayer(nn.Module): method __init__ (line 196) | def __init__( method forward (line 222) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class LoRAConv2dLayer (line 235) | class LoRAConv2dLayer(nn.Module): method __init__ (line 258) | def __init__( method forward (line 286) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class LoRACompatibleConv (line 299) | class LoRACompatibleConv(nn.Conv2d): method __init__ (line 304) | def __init__(self, *args, lora_layer: Optional[LoRAConv2dLayer] = None... method set_lora_layer (line 311) | def set_lora_layer(self, lora_layer: Optional[LoRAConv2dLayer]): method _fuse_lora (line 317) | def _fuse_lora(self, lora_scale: float = 1.0, safe_fusing: bool = False): method _unfuse_lora (line 351) | def _unfuse_lora(self): method forward (line 369) | def forward(self, hidden_states: torch.Tensor, scale: float = 1.0) -> ... class LoRACompatibleLinear (line 386) | class LoRACompatibleLinear(nn.Linear): method __init__ (line 391) | def __init__(self, *args, lora_layer: Optional[LoRALinearLayer] = None... method set_lora_layer (line 398) | def set_lora_layer(self, lora_layer: Optional[LoRALinearLayer]): method _fuse_lora (line 403) | def _fuse_lora(self, lora_scale: float = 1.0, safe_fusing: bool = False): method _unfuse_lora (line 435) | def _unfuse_lora(self): method forward (line 451) | def forward(self, hidden_states: torch.Tensor, scale: float = 1.0) -> ... FILE: src/diffusers/models/model_loading_utils.py function _determine_device_map (line 56) | def _determine_device_map(model: torch.nn.Module, device_map, max_memory... function _fetch_remapped_cls_from_config (line 78) | def _fetch_remapped_cls_from_config(config, old_class): function load_state_dict (line 98) | def load_state_dict(checkpoint_file: Union[str, os.PathLike], variant: O... function load_model_dict_into_meta (line 133) | def load_model_dict_into_meta( function _load_state_dict_into_model (line 165) | def _load_state_dict_into_model(model_to_load, state_dict: OrderedDict) ... function _fetch_index_file (line 186) | def _fetch_index_file( FILE: src/diffusers/models/modeling_flax_pytorch_utils.py function rename_key (line 29) | def rename_key(key): function rename_key_and_reshape_tensor (line 44) | def rename_key_and_reshape_tensor(pt_tuple_key, pt_tensor, random_flax_s... function convert_pytorch_state_dict_to_flax (line 107) | def convert_pytorch_state_dict_to_flax(pt_state_dict, flax_model, init_k... FILE: src/diffusers/models/modeling_flax_utils.py class FlaxModelMixin (line 50) | class FlaxModelMixin(PushToHubMixin): method _from_config (line 65) | def _from_config(cls, config, **kwargs): method _cast_floating_to (line 71) | def _cast_floating_to(self, params: Union[Dict, FrozenDict], dtype: jn... method to_bf16 (line 95) | def to_bf16(self, params: Union[Dict, FrozenDict], mask: Any = None): method to_fp32 (line 134) | def to_fp32(self, params: Union[Dict, FrozenDict], mask: Any = None): method to_fp16 (line 161) | def to_fp16(self, params: Union[Dict, FrozenDict], mask: Any = None): method init_weights (line 200) | def init_weights(self, rng: jax.Array) -> Dict: method from_pretrained (line 205) | def from_pretrained( method save_pretrained (line 497) | def save_pretrained( FILE: src/diffusers/models/modeling_outputs.py class AutoencoderKLOutput (line 7) | class AutoencoderKLOutput(BaseOutput): class Transformer2DModelOutput (line 21) | class Transformer2DModelOutput(BaseOutput): FILE: src/diffusers/models/modeling_pytorch_flax_utils.py function load_flax_checkpoint_in_pytorch_model (line 37) | def load_flax_checkpoint_in_pytorch_model(pt_model, model_file): function load_flax_weights_in_pytorch_model (line 58) | def load_flax_weights_in_pytorch_model(pt_model, flax_state): FILE: src/diffusers/models/modeling_utils.py function get_parameter_device (line 78) | def get_parameter_device(parameter: torch.nn.Module) -> torch.device: function get_parameter_dtype (line 94) | def get_parameter_dtype(parameter: torch.nn.Module) -> torch.dtype: class ModelMixin (line 116) | class ModelMixin(torch.nn.Module, PushToHubMixin): method __init__ (line 132) | def __init__(self): method __getattr__ (line 135) | def __getattr__(self, name: str) -> Any: method is_gradient_checkpointing (line 154) | def is_gradient_checkpointing(self) -> bool: method enable_gradient_checkpointing (line 160) | def enable_gradient_checkpointing(self) -> None: method disable_gradient_checkpointing (line 169) | def disable_gradient_checkpointing(self) -> None: method set_use_npu_flash_attention (line 177) | def set_use_npu_flash_attention(self, valid: bool) -> None: method enable_npu_flash_attention (line 193) | def enable_npu_flash_attention(self) -> None: method disable_npu_flash_attention (line 200) | def disable_npu_flash_attention(self) -> None: method set_use_memory_efficient_attention_xformers (line 207) | def set_use_memory_efficient_attention_xformers( method enable_xformers_memory_efficient_attention (line 224) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method disable_xformers_memory_efficient_attention (line 260) | def disable_xformers_memory_efficient_attention(self) -> None: method save_pretrained (line 266) | def save_pretrained( method from_pretrained (line 412) | def from_pretrained(cls, pretrained_model_name_or_path: Optional[Union... method _load_pretrained_model (line 863) | def _load_pretrained_model( method _get_signature_keys (line 967) | def _get_signature_keys(cls, obj): method _get_no_split_modules (line 976) | def _get_no_split_modules(self, device_map: str): method device (line 1006) | def device(self) -> torch.device: method dtype (line 1014) | def dtype(self) -> torch.dtype: method num_parameters (line 1020) | def num_parameters(self, only_trainable: bool = False, exclude_embeddi... method _convert_deprecated_attention_blocks (line 1058) | def _convert_deprecated_attention_blocks(self, state_dict: OrderedDict... method _temp_convert_self_to_deprecated_attention_blocks (line 1102) | def _temp_convert_self_to_deprecated_attention_blocks(self) -> None: method _undo_temp_convert_self_to_deprecated_attention_blocks (line 1129) | def _undo_temp_convert_self_to_deprecated_attention_blocks(self) -> None: class LegacyModelMixin (line 1153) | class LegacyModelMixin(ModelMixin): method from_pretrained (line 1161) | def from_pretrained(cls, pretrained_model_name_or_path: Optional[Union... FILE: src/diffusers/models/normalization.py class AdaLayerNorm (line 31) | class AdaLayerNorm(nn.Module): method __init__ (line 44) | def __init__( method forward (line 67) | def forward( class FP32LayerNorm (line 88) | class FP32LayerNorm(nn.LayerNorm): method forward (line 89) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: class AdaLayerNormZero (line 100) | class AdaLayerNormZero(nn.Module): method __init__ (line 109) | def __init__(self, embedding_dim: int, num_embeddings: Optional[int] =... method forward (line 127) | def forward( class AdaLayerNormZeroSingle (line 143) | class AdaLayerNormZeroSingle(nn.Module): method __init__ (line 152) | def __init__(self, embedding_dim: int, norm_type="layer_norm", bias=Tr... method forward (line 164) | def forward( class LuminaRMSNormZero (line 175) | class LuminaRMSNormZero(nn.Module): method __init__ (line 183) | def __init__(self, embedding_dim: int, norm_eps: float, norm_elementwi... method forward (line 193) | def forward( class AdaLayerNormSingle (line 206) | class AdaLayerNormSingle(nn.Module): method __init__ (line 217) | def __init__(self, embedding_dim: int, use_additional_conditions: bool... method forward (line 227) | def forward( class AdaGroupNorm (line 239) | class AdaGroupNorm(nn.Module): method __init__ (line 251) | def __init__( method forward (line 265) | def forward(self, x: torch.Tensor, emb: torch.Tensor) -> torch.Tensor: class AdaLayerNormContinuous (line 277) | class AdaLayerNormContinuous(nn.Module): method __init__ (line 278) | def __init__( method forward (line 302) | def forward(self, x: torch.Tensor, conditioning_embedding: torch.Tenso... class LuminaLayerNormContinuous (line 310) | class LuminaLayerNormContinuous(nn.Module): method __init__ (line 311) | def __init__( method forward (line 342) | def forward( class CogVideoXLayerNormZero (line 358) | class CogVideoXLayerNormZero(nn.Module): method __init__ (line 359) | def __init__( method forward (line 373) | def forward( class LayerNorm (line 387) | class LayerNorm(nn.Module): method __init__ (line 388) | def __init__(self, dim, eps: float = 1e-5, elementwise_affine: bool = ... method forward (line 405) | def forward(self, input): class RMSNorm (line 409) | class RMSNorm(nn.Module): method __init__ (line 410) | def __init__(self, dim, eps: float, elementwise_affine: bool = True): method forward (line 425) | def forward(self, hidden_states): class GlobalResponseNorm (line 441) | class GlobalResponseNorm(nn.Module): method __init__ (line 443) | def __init__(self, dim): method forward (line 448) | def forward(self, x): FILE: src/diffusers/models/resnet.py class ResnetBlockCondNorm2D (line 44) | class ResnetBlockCondNorm2D(nn.Module): method __init__ (line 74) | def __init__( method forward (line 149) | def forward(self, input_tensor: torch.Tensor, temb: torch.Tensor, *arg... class ResnetBlock2D (line 189) | class ResnetBlock2D(nn.Module): method __init__ (line 220) | def __init__( method forward (line 320) | def forward(self, input_tensor: torch.Tensor, temb: torch.Tensor, *arg... function rearrange_dims (line 377) | def rearrange_dims(tensor: torch.Tensor) -> torch.Tensor: class Conv1dBlock (line 388) | class Conv1dBlock(nn.Module): method __init__ (line 400) | def __init__( method forward (line 414) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: class ResidualTemporalBlock1D (line 424) | class ResidualTemporalBlock1D(nn.Module): method __init__ (line 436) | def __init__( method forward (line 455) | def forward(self, inputs: torch.Tensor, t: torch.Tensor) -> torch.Tensor: class TemporalConvLayer (line 471) | class TemporalConvLayer(nn.Module): method __init__ (line 482) | def __init__( method forward (line 523) | def forward(self, hidden_states: torch.Tensor, num_frames: int = 1) ->... class TemporalResnetBlock (line 542) | class TemporalResnetBlock(nn.Module): method __init__ (line 554) | def __init__( method forward (line 608) | def forward(self, input_tensor: torch.Tensor, temb: torch.Tensor) -> t... class SpatioTemporalResBlock (line 635) | class SpatioTemporalResBlock(nn.Module): method __init__ (line 653) | def __init__( method forward (line 686) | def forward( class AlphaBlender (line 719) | class AlphaBlender(nn.Module): method __init__ (line 733) | def __init__( method get_alpha (line 753) | def get_alpha(self, image_only_indicator: torch.Tensor, ndims: int) ->... method forward (line 784) | def forward( FILE: src/diffusers/models/resnet_flax.py class FlaxUpsample2D (line 19) | class FlaxUpsample2D(nn.Module): method setup (line 23) | def setup(self): method __call__ (line 32) | def __call__(self, hidden_states): class FlaxDownsample2D (line 43) | class FlaxDownsample2D(nn.Module): method setup (line 47) | def setup(self): method __call__ (line 56) | def __call__(self, hidden_states): class FlaxResnetBlock2D (line 63) | class FlaxResnetBlock2D(nn.Module): method setup (line 70) | def setup(self): method __call__ (line 106) | def __call__(self, hidden_states, temb, deterministic=True): FILE: src/diffusers/models/transformers/auraflow_transformer_2d.py function find_multiple (line 41) | def find_multiple(n: int, k: int) -> int: class AuraFlowPatchEmbed (line 49) | class AuraFlowPatchEmbed(nn.Module): method __init__ (line 50) | def __init__( method forward (line 71) | def forward(self, latent): class AuraFlowFeedForward (line 88) | class AuraFlowFeedForward(nn.Module): method __init__ (line 89) | def __init__(self, dim, hidden_dim=None) -> None: method forward (line 101) | def forward(self, x: torch.Tensor) -> torch.Tensor: class AuraFlowPreFinalBlock (line 107) | class AuraFlowPreFinalBlock(nn.Module): method __init__ (line 108) | def __init__(self, embedding_dim: int, conditioning_embedding_dim: int): method forward (line 114) | def forward(self, x: torch.Tensor, conditioning_embedding: torch.Tenso... class AuraFlowSingleTransformerBlock (line 122) | class AuraFlowSingleTransformerBlock(nn.Module): method __init__ (line 125) | def __init__(self, dim, num_attention_heads, attention_head_dim): method forward (line 146) | def forward(self, hidden_states: torch.FloatTensor, temb: torch.FloatT... class AuraFlowJointTransformerBlock (line 166) | class AuraFlowJointTransformerBlock(nn.Module): method __init__ (line 181) | def __init__(self, dim, num_attention_heads, attention_head_dim): method forward (line 208) | def forward( class AuraFlowTransformer2DModel (line 240) | class AuraFlowTransformer2DModel(ModelMixin, ConfigMixin): method __init__ (line 264) | def __init__( method attn_processors (line 330) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 354) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method fuse_qkv_projections (line 389) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 415) | def unfuse_qkv_projections(self): method _set_gradient_checkpointing (line 428) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 432) | def forward( FILE: src/diffusers/models/transformers/cogvideox_transformer_3d.py class CogVideoXBlock (line 35) | class CogVideoXBlock(nn.Module): method __init__ (line 63) | def __init__( method forward (line 107) | def forward( class CogVideoXTransformer3DModel (line 145) | class CogVideoXTransformer3DModel(ModelMixin, ConfigMixin): method __init__ (line 185) | def __init__( method _set_gradient_checkpointing (line 271) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 274) | def forward( FILE: src/diffusers/models/transformers/dit_transformer_2d.py class DiTTransformer2DModel (line 31) | class DiTTransformer2DModel(ModelMixin, ConfigMixin): method __init__ (line 70) | def __init__( method _set_gradient_checkpointing (line 146) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 150) | def forward( FILE: src/diffusers/models/transformers/dual_transformer_2d.py class DualTransformer2DModel (line 22) | class DualTransformer2DModel(nn.Module): method __init__ (line 49) | def __init__( method forward (line 98) | def forward( FILE: src/diffusers/models/transformers/hunyuan_transformer_2d.py class AdaLayerNormShift (line 37) | class AdaLayerNormShift(nn.Module): method __init__ (line 46) | def __init__(self, embedding_dim: int, elementwise_affine=True, eps=1e... method forward (line 52) | def forward(self, x: torch.Tensor, emb: torch.Tensor) -> torch.Tensor: class HunyuanDiTBlock (line 59) | class HunyuanDiTBlock(nn.Module): method __init__ (line 91) | def __init__( method set_chunk_feed_forward (line 161) | def set_chunk_feed_forward(self, chunk_size: Optional[int], dim: int =... method forward (line 166) | def forward( class HunyuanDiT2DModel (line 203) | class HunyuanDiT2DModel(ModelMixin, ConfigMixin): method __init__ (line 248) | def __init__( method fuse_qkv_projections (line 321) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 347) | def unfuse_qkv_projections(self): method attn_processors (line 362) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 386) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 420) | def set_default_attn_processor(self): method forward (line 426) | def forward( method enable_forward_chunking (line 539) | def enable_forward_chunking(self, chunk_size: Optional[int] = None, di... method disable_forward_chunking (line 569) | def disable_forward_chunking(self): FILE: src/diffusers/models/transformers/latte_transformer_3d.py class LatteTransformer3DModel (line 28) | class LatteTransformer3DModel(ModelMixin, ConfigMixin): method __init__ (line 69) | def __init__( method _set_gradient_checkpointing (line 165) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 168) | def forward( FILE: src/diffusers/models/transformers/lumina_nextdit2d.py class LuminaNextDiTBlock (line 36) | class LuminaNextDiTBlock(nn.Module): method __init__ (line 53) | def __init__( method forward (line 118) | def forward( class LuminaNextDiT2DModel (line 178) | class LuminaNextDiT2DModel(ModelMixin, ConfigMixin): method __init__ (line 225) | def __init__( method forward (line 289) | def forward( FILE: src/diffusers/models/transformers/pixart_transformer_2d.py class PixArtTransformer2DModel (line 32) | class PixArtTransformer2DModel(ModelMixin, ConfigMixin): method __init__ (line 84) | def __init__( method _set_gradient_checkpointing (line 186) | def _set_gradient_checkpointing(self, module, value=False): method attn_processors (line 192) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 216) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method fuse_qkv_projections (line 251) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 277) | def unfuse_qkv_projections(self): method forward (line 290) | def forward( FILE: src/diffusers/models/transformers/prior_transformer.py class PriorTransformerOutput (line 24) | class PriorTransformerOutput(BaseOutput): class PriorTransformer (line 36) | class PriorTransformer(ModelMixin, ConfigMixin, UNet2DConditionLoadersMi... method __init__ (line 74) | def __init__( method attn_processors (line 171) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 195) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 230) | def set_default_attn_processor(self): method forward (line 245) | def forward( method post_process_latents (line 378) | def post_process_latents(self, prior_latents): FILE: src/diffusers/models/transformers/stable_audio_transformer.py class StableAudioGaussianFourierProjection (line 39) | class StableAudioGaussianFourierProjection(nn.Module): method __init__ (line 43) | def __init__( method forward (line 58) | def forward(self, x): class StableAudioDiTBlock (line 72) | class StableAudioDiTBlock(nn.Module): method __init__ (line 88) | def __init__( method set_chunk_feed_forward (line 146) | def set_chunk_feed_forward(self, chunk_size: Optional[int], dim: int =... method forward (line 151) | def forward( class StableAudioDiTModel (line 190) | class StableAudioDiTModel(ModelMixin, ConfigMixin): method __init__ (line 216) | def __init__( method attn_processors (line 283) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 307) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 342) | def set_default_attn_processor(self): method _set_gradient_checkpointing (line 348) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 352) | def forward( FILE: src/diffusers/models/transformers/t5_film_transformer.py class T5FilmDecoder (line 26) | class T5FilmDecoder(ModelMixin, ConfigMixin): method __init__ (line 50) | def __init__( method encoder_decoder_mask (line 89) | def encoder_decoder_mask(self, query_input: torch.Tensor, key_input: t... method forward (line 93) | def forward(self, encodings_and_masks, decoder_input_tokens, decoder_n... class DecoderLayer (line 150) | class DecoderLayer(nn.Module): method __init__ (line 169) | def __init__( method forward (line 196) | def forward( class T5LayerSelfAttentionCond (line 228) | class T5LayerSelfAttentionCond(nn.Module): method __init__ (line 243) | def __init__(self, d_model: int, d_kv: int, num_heads: int, dropout_ra... method forward (line 250) | def forward( class T5LayerCrossAttention (line 270) | class T5LayerCrossAttention(nn.Module): method __init__ (line 287) | def __init__(self, d_model: int, d_kv: int, num_heads: int, dropout_ra... method forward (line 293) | def forward( class T5LayerFFCond (line 309) | class T5LayerFFCond(nn.Module): method __init__ (line 324) | def __init__(self, d_model: int, d_ff: int, dropout_rate: float, layer... method forward (line 331) | def forward(self, hidden_states: torch.Tensor, conditioning_emb: Optio... class T5DenseGatedActDense (line 341) | class T5DenseGatedActDense(nn.Module): method __init__ (line 354) | def __init__(self, d_model: int, d_ff: int, dropout_rate: float): method forward (line 362) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class T5LayerNorm (line 372) | class T5LayerNorm(nn.Module): method __init__ (line 383) | def __init__(self, hidden_size: int, eps: float = 1e-6): method forward (line 391) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class NewGELUActivation (line 407) | class NewGELUActivation(nn.Module): method forward (line 413) | def forward(self, input: torch.Tensor) -> torch.Tensor: class T5FiLMLayer (line 417) | class T5FiLMLayer(nn.Module): method __init__ (line 428) | def __init__(self, in_features: int, out_features: int): method forward (line 432) | def forward(self, x: torch.Tensor, conditioning_emb: torch.Tensor) -> ... FILE: src/diffusers/models/transformers/transformer_2d.py class Transformer2DModelOutput (line 32) | class Transformer2DModelOutput(Transformer2DModelOutput): method __init__ (line 33) | def __init__(self, *args, **kwargs): class Transformer2DModel (line 39) | class Transformer2DModel(LegacyModelMixin, LegacyConfigMixin): method __init__ (line 71) | def __init__( method _init_continuous_input (line 174) | def _init_continuous_input(self, norm_type): method _init_vectorized_inputs (line 211) | def _init_vectorized_inputs(self, norm_type): method _init_patched_inputs (line 251) | def _init_patched_inputs(self, norm_type): method _set_gradient_checkpointing (line 323) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 327) | def forward( method _operate_on_continuous_inputs (line 479) | def _operate_on_continuous_inputs(self, hidden_states): method _operate_on_patched_inputs (line 494) | def _operate_on_patched_inputs(self, hidden_states, encoder_hidden_sta... method _get_output_for_continuous_inputs (line 514) | def _get_output_for_continuous_inputs(self, hidden_states, residual, b... method _get_output_for_vectorized_inputs (line 529) | def _get_output_for_vectorized_inputs(self, hidden_states): method _get_output_for_patched_inputs (line 538) | def _get_output_for_patched_inputs( FILE: src/diffusers/models/transformers/transformer_flux.py function rope (line 38) | def rope(pos: torch.Tensor, dim: int, theta: int) -> torch.Tensor: class EmbedND (line 55) | class EmbedND(nn.Module): method __init__ (line 56) | def __init__(self, dim: int, theta: int, axes_dim: List[int]): method forward (line 62) | def forward(self, ids: torch.Tensor) -> torch.Tensor: class FluxSingleTransformerBlock (line 72) | class FluxSingleTransformerBlock(nn.Module): method __init__ (line 86) | def __init__(self, dim, num_attention_heads, attention_head_dim, mlp_r... method forward (line 109) | def forward( class FluxTransformerBlock (line 143) | class FluxTransformerBlock(nn.Module): method __init__ (line 157) | def __init__(self, dim, num_attention_heads, attention_head_dim, qk_no... method forward (line 194) | def forward( class FluxTransformer2DModel (line 250) | class FluxTransformer2DModel(ModelMixin, ConfigMixin, PeftAdapterMixin, ... method __init__ (line 271) | def __init__( method _set_gradient_checkpointing (line 326) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 330) | def forward( FILE: src/diffusers/models/transformers/transformer_sd3.py class SD3Transformer2DModel (line 35) | class SD3Transformer2DModel(ModelMixin, ConfigMixin, PeftAdapterMixin, F... method __init__ (line 59) | def __init__( method enable_forward_chunking (line 111) | def enable_forward_chunking(self, chunk_size: Optional[int] = None, di... method disable_forward_chunking (line 141) | def disable_forward_chunking(self): method attn_processors (line 154) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 178) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method fuse_qkv_projections (line 213) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 239) | def unfuse_qkv_projections(self): method _set_gradient_checkpointing (line 252) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 256) | def forward( FILE: src/diffusers/models/transformers/transformer_temporal.py class TransformerTemporalModelOutput (line 29) | class TransformerTemporalModelOutput(BaseOutput): class TransformerTemporalModel (line 41) | class TransformerTemporalModel(ModelMixin, ConfigMixin): method __init__ (line 71) | def __init__( method forward (line 121) | def forward( class TransformerSpatioTemporalModel (line 203) | class TransformerSpatioTemporalModel(nn.Module): method __init__ (line 218) | def __init__( method forward (line 278) | def forward( FILE: src/diffusers/models/unets/unet_1d.py class UNet1DOutput (line 29) | class UNet1DOutput(BaseOutput): class UNet1DModel (line 41) | class UNet1DModel(ModelMixin, ConfigMixin): method __init__ (line 75) | def __init__( method forward (line 195) | def forward( FILE: src/diffusers/models/unets/unet_1d_blocks.py class DownResnetBlock1D (line 25) | class DownResnetBlock1D(nn.Module): method __init__ (line 26) | def __init__( method forward (line 69) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class UpResnetBlock1D (line 87) | class UpResnetBlock1D(nn.Module): method __init__ (line 88) | def __init__( method forward (line 129) | def forward( class ValueFunctionMidBlock1D (line 152) | class ValueFunctionMidBlock1D(nn.Module): method __init__ (line 153) | def __init__(self, in_channels: int, out_channels: int, embed_dim: int): method forward (line 164) | def forward(self, x: torch.Tensor, temb: Optional[torch.Tensor] = None... class MidResTemporalBlock1D (line 172) | class MidResTemporalBlock1D(nn.Module): method __init__ (line 173) | def __init__( method forward (line 212) | def forward(self, hidden_states: torch.Tensor, temb: torch.Tensor) -> ... class OutConv1DBlock (line 225) | class OutConv1DBlock(nn.Module): method __init__ (line 226) | def __init__(self, num_groups_out: int, out_channels: int, embed_dim: ... method forward (line 233) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class OutValueFunctionBlock (line 243) | class OutValueFunctionBlock(nn.Module): method __init__ (line 244) | def __init__(self, fc_dim: int, embed_dim: int, act_fn: str = "mish"): method forward (line 254) | def forward(self, hidden_states: torch.Tensor, temb: torch.Tensor) -> ... class Downsample1d (line 283) | class Downsample1d(nn.Module): method __init__ (line 284) | def __init__(self, kernel: str = "linear", pad_mode: str = "reflect"): method forward (line 291) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class Upsample1d (line 300) | class Upsample1d(nn.Module): method __init__ (line 301) | def __init__(self, kernel: str = "linear", pad_mode: str = "reflect"): method forward (line 308) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class SelfAttention1d (line 317) | class SelfAttention1d(nn.Module): method __init__ (line 318) | def __init__(self, in_channels: int, n_head: int = 1, dropout_rate: fl... method transpose_for_scores (line 332) | def transpose_for_scores(self, projection: torch.Tensor) -> torch.Tensor: method forward (line 338) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class ResConvBlock (line 375) | class ResConvBlock(nn.Module): method __init__ (line 376) | def __init__(self, in_channels: int, mid_channels: int, out_channels: ... method forward (line 393) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class UNetMidBlock1D (line 409) | class UNetMidBlock1D(nn.Module): method __init__ (line 410) | def __init__(self, mid_channels: int, in_channels: int, out_channels: ... method forward (line 438) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class AttnDownBlock1D (line 449) | class AttnDownBlock1D(nn.Module): method __init__ (line 450) | def __init__(self, out_channels: int, in_channels: int, mid_channels: ... method forward (line 469) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class DownBlock1D (line 479) | class DownBlock1D(nn.Module): method __init__ (line 480) | def __init__(self, out_channels: int, in_channels: int, mid_channels: ... method forward (line 493) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class DownBlock1DNoSkip (line 502) | class DownBlock1DNoSkip(nn.Module): method __init__ (line 503) | def __init__(self, out_channels: int, in_channels: int, mid_channels: ... method forward (line 515) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class AttnUpBlock1D (line 523) | class AttnUpBlock1D(nn.Module): method __init__ (line 524) | def __init__(self, in_channels: int, out_channels: int, mid_channels: ... method forward (line 543) | def forward( class UpBlock1D (line 561) | class UpBlock1D(nn.Module): method __init__ (line 562) | def __init__(self, in_channels: int, out_channels: int, mid_channels: ... method forward (line 575) | def forward( class UpBlock1DNoSkip (line 592) | class UpBlock1DNoSkip(nn.Module): method __init__ (line 593) | def __init__(self, in_channels: int, out_channels: int, mid_channels: ... method forward (line 605) | def forward( function get_down_block (line 626) | def get_down_block( function get_up_block (line 651) | def get_up_block( function get_mid_block (line 671) | def get_mid_block( function get_out_block (line 695) | def get_out_block( FILE: src/diffusers/models/unets/unet_2d.py class UNet2DOutput (line 28) | class UNet2DOutput(BaseOutput): class UNet2DModel (line 40) | class UNet2DModel(ModelMixin, ConfigMixin): method __init__ (line 93) | def __init__( method forward (line 243) | def forward( FILE: src/diffusers/models/unets/unet_2d_blocks.py function get_down_block (line 43) | def get_down_block( function get_mid_block (line 252) | def get_mid_block( function get_up_block (line 327) | def get_up_block( class AutoencoderTinyBlock (line 552) | class AutoencoderTinyBlock(nn.Module): method __init__ (line 568) | def __init__(self, in_channels: int, out_channels: int, act_fn: str): method forward (line 585) | def forward(self, x: torch.Tensor) -> torch.Tensor: class UNetMidBlock2D (line 589) | class UNetMidBlock2D(nn.Module): method __init__ (line 620) | def __init__( method forward (line 734) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class UNetMidBlock2DCrossAttn (line 744) | class UNetMidBlock2DCrossAttn(nn.Module): method __init__ (line 745) | def __init__( method forward (line 847) | def forward( class UNetMidBlock2DSimpleCrossAttn (line 902) | class UNetMidBlock2DSimpleCrossAttn(nn.Module): method __init__ (line 903) | def __init__( method forward (line 987) | def forward( class AttnDownBlock2D (line 1027) | class AttnDownBlock2D(nn.Module): method __init__ (line 1028) | def __init__( method forward (line 1119) | def forward( class CrossAttnDownBlock2D (line 1149) | class CrossAttnDownBlock2D(nn.Module): method __init__ (line 1150) | def __init__( method forward (line 1241) | def forward( class DownBlock2D (line 1312) | class DownBlock2D(nn.Module): method __init__ (line 1313) | def __init__( method forward (line 1364) | def forward( class DownEncoderBlock2D (line 1404) | class DownEncoderBlock2D(nn.Module): method __init__ (line 1405) | def __init__( method forward (line 1468) | def forward(self, hidden_states: torch.Tensor, *args, **kwargs) -> tor... class AttnDownEncoderBlock2D (line 1483) | class AttnDownEncoderBlock2D(nn.Module): method __init__ (line 1484) | def __init__( method forward (line 1570) | def forward(self, hidden_states: torch.Tensor, *args, **kwargs) -> tor... class AttnSkipDownBlock2D (line 1586) | class AttnSkipDownBlock2D(nn.Module): method __init__ (line 1587) | def __init__( method forward (line 1667) | def forward( class SkipDownBlock2D (line 1698) | class SkipDownBlock2D(nn.Module): method __init__ (line 1699) | def __init__( method forward (line 1758) | def forward( class ResnetDownsampleBlock2D (line 1788) | class ResnetDownsampleBlock2D(nn.Module): method __init__ (line 1789) | def __init__( method forward (line 1852) | def forward( class SimpleCrossAttnDownBlock2D (line 1892) | class SimpleCrossAttnDownBlock2D(nn.Module): method __init__ (line 1893) | def __init__( method forward (line 1987) | def forward( class KDownBlock2D (line 2053) | class KDownBlock2D(nn.Module): method __init__ (line 2054) | def __init__( method forward (line 2099) | def forward( class KCrossAttnDownBlock2D (line 2137) | class KCrossAttnDownBlock2D(nn.Module): method __init__ (line 2138) | def __init__( method forward (line 2202) | def forward( class AttnUpBlock2D (line 2267) | class AttnUpBlock2D(nn.Module): method __init__ (line 2268) | def __init__( method forward (line 2359) | def forward( class CrossAttnUpBlock2D (line 2391) | class CrossAttnUpBlock2D(nn.Module): method __init__ (line 2392) | def __init__( method forward (line 2482) | def forward( class UpBlock2D (line 2567) | class UpBlock2D(nn.Module): method __init__ (line 2568) | def __init__( method forward (line 2617) | def forward( class UpDecoderBlock2D (line 2682) | class UpDecoderBlock2D(nn.Module): method __init__ (line 2683) | def __init__( method forward (line 2744) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class AttnUpDecoderBlock2D (line 2755) | class AttnUpDecoderBlock2D(nn.Module): method __init__ (line 2756) | def __init__( method forward (line 2842) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class AttnSkipUpBlock2D (line 2854) | class AttnSkipUpBlock2D(nn.Module): method __init__ (line 2855) | def __init__( method forward (line 2948) | def forward( class SkipUpBlock2D (line 2988) | class SkipUpBlock2D(nn.Module): method __init__ (line 2989) | def __init__( method forward (line 3060) | def forward( class ResnetUpsampleBlock2D (line 3098) | class ResnetUpsampleBlock2D(nn.Module): method __init__ (line 3099) | def __init__( method forward (line 3167) | def forward( class SimpleCrossAttnUpBlock2D (line 3212) | class SimpleCrossAttnUpBlock2D(nn.Module): method __init__ (line 3213) | def __init__( method forward (line 3311) | def forward( class KUpBlock2D (line 3379) | class KUpBlock2D(nn.Module): method __init__ (line 3380) | def __init__( method forward (line 3429) | def forward( class KCrossAttnUpBlock2D (line 3473) | class KCrossAttnUpBlock2D(nn.Module): method __init__ (line 3474) | def __init__( method forward (line 3559) | def forward( class KAttentionBlock (line 3620) | class KAttentionBlock(nn.Module): method __init__ (line 3644) | def __init__( method _to_3d (line 3687) | def _to_3d(self, hidden_states: torch.Tensor, height: int, weight: int... method _to_4d (line 3690) | def _to_4d(self, hidden_states: torch.Tensor, height: int, weight: int... method forward (line 3693) | def forward( FILE: src/diffusers/models/unets/unet_2d_blocks_flax.py class FlaxCrossAttnDownBlock2D (line 22) | class FlaxCrossAttnDownBlock2D(nn.Module): method setup (line 62) | def setup(self): method __call__ (line 96) | def __call__(self, hidden_states, temb, encoder_hidden_states, determi... class FlaxDownBlock2D (line 111) | class FlaxDownBlock2D(nn.Module): method setup (line 137) | def setup(self): method __call__ (line 155) | def __call__(self, hidden_states, temb, deterministic=True): class FlaxCrossAttnUpBlock2D (line 169) | class FlaxCrossAttnUpBlock2D(nn.Module): method setup (line 210) | def setup(self): method __call__ (line 245) | def __call__(self, hidden_states, res_hidden_states_tuple, temb, encod... class FlaxUpBlock2D (line 261) | class FlaxUpBlock2D(nn.Module): method setup (line 290) | def setup(self): method __call__ (line 310) | def __call__(self, hidden_states, res_hidden_states_tuple, temb, deter... class FlaxUNetMidBlock2DCrossAttn (line 325) | class FlaxUNetMidBlock2DCrossAttn(nn.Module): method setup (line 357) | def setup(self): method __call__ (line 394) | def __call__(self, hidden_states, temb, encoder_hidden_states, determi... FILE: src/diffusers/models/unets/unet_2d_condition.py class UNet2DConditionOutput (line 59) | class UNet2DConditionOutput(BaseOutput): class UNet2DConditionModel (line 71) | class UNet2DConditionModel( method __init__ (line 171) | def __init__( method _check_config (line 487) | def _check_config( method _set_time_proj (line 539) | def _set_time_proj( method _set_encoder_hid_proj (line 567) | def _set_encoder_hid_proj( method _set_class_embedding (line 607) | def _set_class_embedding( method _set_add_embedding (line 644) | def _set_add_embedding( method _set_pos_net_if_use_gligen (line 684) | def _set_pos_net_if_use_gligen(self, attention_type: str, cross_attent... method attn_processors (line 698) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 721) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 755) | def set_default_attn_processor(self): method set_attention_slice (line 770) | def set_attention_slice(self, slice_size: Union[str, int, List[int]] =... method _set_gradient_checkpointing (line 835) | def _set_gradient_checkpointing(self, module, value=False): method enable_freeu (line 839) | def enable_freeu(self, s1: float, s2: float, b1: float, b2: float): method disable_freeu (line 863) | def disable_freeu(self): method fuse_qkv_projections (line 871) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 896) | def unfuse_qkv_projections(self): method get_time_embed (line 909) | def get_time_embed( method get_class_embed (line 935) | def get_class_embed(self, sample: torch.Tensor, class_labels: Optional... method get_aug_embed (line 951) | def get_aug_embed( method process_encoder_hidden_states (line 1003) | def process_encoder_hidden_states( method forward (line 1039) | def forward( FILE: src/diffusers/models/unets/unet_2d_condition_flax.py class FlaxUNet2DConditionOutput (line 36) | class FlaxUNet2DConditionOutput(BaseOutput): class FlaxUNet2DConditionModel (line 49) | class FlaxUNet2DConditionModel(nn.Module, FlaxModelMixin, ConfigMixin): method init_weights (line 133) | def init_weights(self, rng: jax.Array) -> FrozenDict: method setup (line 165) | def setup(self) -> None: method __call__ (line 330) | def __call__( FILE: src/diffusers/models/unets/unet_3d_blocks.py class DownBlockMotion (line 47) | class DownBlockMotion(DownBlockMotion): method __init__ (line 48) | def __init__(self, *args, **kwargs): class CrossAttnDownBlockMotion (line 54) | class CrossAttnDownBlockMotion(CrossAttnDownBlockMotion): method __init__ (line 55) | def __init__(self, *args, **kwargs): class UpBlockMotion (line 61) | class UpBlockMotion(UpBlockMotion): method __init__ (line 62) | def __init__(self, *args, **kwargs): class CrossAttnUpBlockMotion (line 68) | class CrossAttnUpBlockMotion(CrossAttnUpBlockMotion): method __init__ (line 69) | def __init__(self, *args, **kwargs): class UNetMidBlockCrossAttnMotion (line 75) | class UNetMidBlockCrossAttnMotion(UNetMidBlockCrossAttnMotion): method __init__ (line 76) | def __init__(self, *args, **kwargs): function get_down_block (line 82) | def get_down_block( function get_up_block (line 174) | def get_up_block( class UNetMidBlock3DCrossAttn (line 274) | class UNetMidBlock3DCrossAttn(nn.Module): method __init__ (line 275) | def __init__( method forward (line 376) | def forward( class CrossAttnDownBlock3D (line 408) | class CrossAttnDownBlock3D(nn.Module): method __init__ (line 409) | def __init__( method forward (line 509) | def forward( class DownBlock3D (line 550) | class DownBlock3D(nn.Module): method __init__ (line 551) | def __init__( method forward (line 616) | def forward( class CrossAttnUpBlock3D (line 639) | class CrossAttnUpBlock3D(nn.Module): method __init__ (line 640) | def __init__( method forward (line 734) | def forward( class UpBlock3D (line 796) | class UpBlock3D(nn.Module): method __init__ (line 797) | def __init__( method forward (line 856) | def forward( class MidBlockTemporalDecoder (line 899) | class MidBlockTemporalDecoder(nn.Module): method __init__ (line 900) | def __init__( method forward (line 943) | def forward( class UpBlockTemporalDecoder (line 962) | class UpBlockTemporalDecoder(nn.Module): method __init__ (line 963) | def __init__( method forward (line 994) | def forward( class UNetMidBlockSpatioTemporal (line 1012) | class UNetMidBlockSpatioTemporal(nn.Module): method __init__ (line 1013) | def __init__( method forward (line 1067) | def forward( class DownBlockSpatioTemporal (line 1122) | class DownBlockSpatioTemporal(nn.Module): method __init__ (line 1123) | def __init__( method forward (line 1163) | def forward( class CrossAttnDownBlockSpatioTemporal (line 1212) | class CrossAttnDownBlockSpatioTemporal(nn.Module): method __init__ (line 1213) | def __init__( method forward (line 1273) | def forward( class UpBlockSpatioTemporal (line 1334) | class UpBlockSpatioTemporal(nn.Module): method __init__ (line 1335) | def __init__( method forward (line 1372) | def forward( class CrossAttnUpBlockSpatioTemporal (line 1423) | class CrossAttnUpBlockSpatioTemporal(nn.Module): method __init__ (line 1424) | def __init__( method forward (line 1481) | def forward( FILE: src/diffusers/models/unets/unet_3d_condition.py class UNet3DConditionOutput (line 54) | class UNet3DConditionOutput(BaseOutput): class UNet3DConditionModel (line 66) | class UNet3DConditionModel(ModelMixin, ConfigMixin, UNet2DConditionLoade... method __init__ (line 102) | def __init__( method attn_processors (line 294) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attention_slice (line 318) | def set_attention_slice(self, slice_size: Union[str, int, List[int]]) ... method set_attn_processor (line 384) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method enable_forward_chunking (line 418) | def enable_forward_chunking(self, chunk_size: Optional[int] = None, di... method disable_forward_chunking (line 447) | def disable_forward_chunking(self): method set_default_attn_processor (line 459) | def set_default_attn_processor(self): method _set_gradient_checkpointing (line 474) | def _set_gradient_checkpointing(self, module, value: bool = False) -> ... method enable_freeu (line 479) | def enable_freeu(self, s1, s2, b1, b2): method disable_freeu (line 504) | def disable_freeu(self): method fuse_qkv_projections (line 513) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 539) | def unfuse_qkv_projections(self): method forward (line 552) | def forward( FILE: src/diffusers/models/unets/unet_i2vgen_xl.py class I2VGenXLTransformerTemporalEncoder (line 52) | class I2VGenXLTransformerTemporalEncoder(nn.Module): method __init__ (line 53) | def __init__( method forward (line 83) | def forward( class I2VGenXLUNet (line 101) | class I2VGenXLUNet(ModelMixin, ConfigMixin, UNet2DConditionLoadersMixin): method __init__ (line 131) | def __init__( method attn_processors (line 323) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 347) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method enable_forward_chunking (line 382) | def enable_forward_chunking(self, chunk_size: Optional[int] = None, di... method disable_forward_chunking (line 412) | def disable_forward_chunking(self): method set_default_attn_processor (line 424) | def set_default_attn_processor(self): method _set_gradient_checkpointing (line 440) | def _set_gradient_checkpointing(self, module, value: bool = False) -> ... method enable_freeu (line 445) | def enable_freeu(self, s1, s2, b1, b2): method disable_freeu (line 470) | def disable_freeu(self): method fuse_qkv_projections (line 479) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 505) | def unfuse_qkv_projections(self): method forward (line 518) | def forward( FILE: src/diffusers/models/unets/unet_kandinsky3.py class Kandinsky3UNetOutput (line 33) | class Kandinsky3UNetOutput(BaseOutput): class Kandinsky3EncoderProj (line 37) | class Kandinsky3EncoderProj(nn.Module): method __init__ (line 38) | def __init__(self, encoder_hid_dim, cross_attention_dim): method forward (line 43) | def forward(self, x): class Kandinsky3UNet (line 49) | class Kandinsky3UNet(ModelMixin, ConfigMixin): method __init__ (line 51) | def __init__( method attn_processors (line 145) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 168) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 202) | def set_default_attn_processor(self): method _set_gradient_checkpointing (line 208) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 212) | def forward(self, sample, timestep, encoder_hidden_states=None, encode... class Kandinsky3UpSampleBlock (line 254) | class Kandinsky3UpSampleBlock(nn.Module): method __init__ (line 255) | def __init__( method forward (line 313) | def forward(self, x, time_embed, context=None, context_mask=None, imag... class Kandinsky3DownSampleBlock (line 325) | class Kandinsky3DownSampleBlock(nn.Module): method __init__ (line 326) | def __init__( method forward (line 381) | def forward(self, x, time_embed, context=None, context_mask=None, imag... class Kandinsky3ConditionalGroupNorm (line 393) | class Kandinsky3ConditionalGroupNorm(nn.Module): method __init__ (line 394) | def __init__(self, groups, normalized_shape, context_dim): method forward (line 401) | def forward(self, x, context): class Kandinsky3Block (line 412) | class Kandinsky3Block(nn.Module): method __init__ (line 413) | def __init__(self, in_channels, out_channels, time_embed_dim, kernel_s... method forward (line 430) | def forward(self, x, time_embed): class Kandinsky3ResNetBlock (line 439) | class Kandinsky3ResNetBlock(nn.Module): method __init__ (line 440) | def __init__( method forward (line 471) | def forward(self, x, time_embed): class Kandinsky3AttentionPooling (line 483) | class Kandinsky3AttentionPooling(nn.Module): method __init__ (line 484) | def __init__(self, num_channels, context_dim, head_dim=64): method forward (line 494) | def forward(self, x, context, context_mask=None): class Kandinsky3AttentionBlock (line 500) | class Kandinsky3AttentionBlock(nn.Module): method __init__ (line 501) | def __init__(self, num_channels, time_embed_dim, context_dim=None, nor... method forward (line 520) | def forward(self, x, time_embed, context=None, context_mask=None, imag... FILE: src/diffusers/models/unets/unet_motion_model.py class UNetMotionOutput (line 53) | class UNetMotionOutput(BaseOutput): class AnimateDiffTransformer3D (line 65) | class AnimateDiffTransformer3D(nn.Module): method __init__ (line 94) | def __init__( method forward (line 144) | def forward( class DownBlockMotion (line 216) | class DownBlockMotion(nn.Module): method __init__ (line 217) | def __init__( method forward (line 310) | def forward( class CrossAttnDownBlockMotion (line 362) | class CrossAttnDownBlockMotion(nn.Module): method __init__ (line 363) | def __init__( method forward (line 497) | def forward( class CrossAttnUpBlockMotion (line 573) | class CrossAttnUpBlockMotion(nn.Module): method __init__ (line 574) | def __init__( method forward (line 699) | def forward( class UpBlockMotion (line 791) | class UpBlockMotion(nn.Module): method __init__ (line 792) | def __init__( method forward (line 870) | def forward( class UNetMidBlockCrossAttnMotion (line 943) | class UNetMidBlockCrossAttnMotion(nn.Module): method __init__ (line 944) | def __init__( method forward (line 1069) | def forward( class MotionModules (line 1137) | class MotionModules(nn.Module): method __init__ (line 1138) | def __init__( class MotionAdapter (line 1178) | class MotionAdapter(ModelMixin, ConfigMixin, FromOriginalModelMixin): method __init__ (line 1180) | def __init__( method forward (line 1311) | def forward(self, sample): class UNetMotionModel (line 1315) | class UNetMotionModel(ModelMixin, ConfigMixin, UNet2DConditionLoadersMix... method __init__ (line 1327) | def __init__( method from_unet2d (line 1651) | def from_unet2d( method freeze_unet2d_params (line 1804) | def freeze_unet2d_params(self) -> None: method load_motion_modules (line 1828) | def load_motion_modules(self, motion_adapter: Optional[MotionAdapter])... method save_motion_modules (line 1838) | def save_motion_modules( method attn_processors (line 1875) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 1899) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method enable_forward_chunking (line 1933) | def enable_forward_chunking(self, chunk_size: Optional[int] = None, di... method disable_forward_chunking (line 1962) | def disable_forward_chunking(self) -> None: method set_default_attn_processor (line 1974) | def set_default_attn_processor(self) -> None: method _set_gradient_checkpointing (line 1989) | def _set_gradient_checkpointing(self, module, value: bool = False) -> ... method enable_freeu (line 1994) | def enable_freeu(self, s1: float, s2: float, b1: float, b2: float) -> ... method disable_freeu (line 2019) | def disable_freeu(self) -> None: method fuse_qkv_projections (line 2028) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 2054) | def unfuse_qkv_projections(self): method forward (line 2067) | def forward( FILE: src/diffusers/models/unets/unet_spatio_temporal_condition.py class UNetSpatioTemporalConditionOutput (line 20) | class UNetSpatioTemporalConditionOutput(BaseOutput): class UNetSpatioTemporalConditionModel (line 32) | class UNetSpatioTemporalConditionModel(ModelMixin, ConfigMixin, UNet2DCo... method __init__ (line 71) | def __init__( method attn_processors (line 249) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 276) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 310) | def set_default_attn_processor(self): method _set_gradient_checkpointing (line 323) | def _set_gradient_checkpointing(self, module, value=False): method enable_forward_chunking (line 328) | def enable_forward_chunking(self, chunk_size: Optional[int] = None, di... method forward (line 357) | def forward( FILE: src/diffusers/models/unets/unet_stable_cascade.py class SDCascadeLayerNorm (line 31) | class SDCascadeLayerNorm(nn.LayerNorm): method __init__ (line 32) | def __init__(self, *args, **kwargs): method forward (line 35) | def forward(self, x): class SDCascadeTimestepBlock (line 41) | class SDCascadeTimestepBlock(nn.Module): method __init__ (line 42) | def __init__(self, c, c_timestep, conds=[]): method forward (line 50) | def forward(self, x, t): class SDCascadeResBlock (line 59) | class SDCascadeResBlock(nn.Module): method __init__ (line 60) | def __init__(self, c, c_skip=0, kernel_size=3, dropout=0.0): method forward (line 72) | def forward(self, x, x_skip=None): class GlobalResponseNorm (line 82) | class GlobalResponseNorm(nn.Module): method __init__ (line 83) | def __init__(self, dim): method forward (line 88) | def forward(self, x): class SDCascadeAttnBlock (line 94) | class SDCascadeAttnBlock(nn.Module): method __init__ (line 95) | def __init__(self, c, c_cond, nhead, self_attn=True, dropout=0.0): method forward (line 103) | def forward(self, x, kv): class UpDownBlock2d (line 113) | class UpDownBlock2d(nn.Module): method __init__ (line 114) | def __init__(self, in_channels, out_channels, mode, enabled=True): method forward (line 126) | def forward(self, x): class StableCascadeUNetOutput (line 133) | class StableCascadeUNetOutput(BaseOutput): class StableCascadeUNet (line 137) | class StableCascadeUNet(ModelMixin, ConfigMixin, FromOriginalModelMixin): method __init__ (line 141) | def __init__( method _set_gradient_checkpointing (line 390) | def _set_gradient_checkpointing(self, value=False): method _init_weights (line 393) | def _init_weights(self, m): method get_timestep_ratio_embedding (line 422) | def get_timestep_ratio_embedding(self, timestep_ratio, max_positions=1... method get_clip_embeddings (line 436) | def get_clip_embeddings(self, clip_txt_pooled, clip_txt=None, clip_img... method _down_encode (line 454) | def _down_encode(self, x, r_embed, clip): method _up_decode (line 503) | def _up_decode(self, level_outputs, r_embed, clip): method forward (line 566) | def forward( FILE: src/diffusers/models/unets/uvit_2d.py class UVit2DModel (line 39) | class UVit2DModel(ModelMixin, ConfigMixin, PeftAdapterMixin): method __init__ (line 43) | def __init__( method _set_gradient_checkpointing (line 151) | def _set_gradient_checkpointing(self, module, value: bool = False) -> ... method forward (line 154) | def forward(self, input_ids, encoder_hidden_states, pooled_text_emb, m... method attn_processors (line 217) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 241) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 276) | def set_default_attn_processor(self): class UVit2DConvEmbed (line 292) | class UVit2DConvEmbed(nn.Module): method __init__ (line 293) | def __init__(self, in_channels, block_out_channels, vocab_size, elemen... method forward (line 299) | def forward(self, input_ids): class UVitBlock (line 307) | class UVitBlock(nn.Module): method __init__ (line 308) | def __init__( method forward (line 386) | def forward(self, x, pooled_text_emb, encoder_hidden_states, cross_att... class ConvNextBlock (line 406) | class ConvNextBlock(nn.Module): method __init__ (line 407) | def __init__( method forward (line 427) | def forward(self, x, cond_embeds): class ConvMlmLayer (line 451) | class ConvMlmLayer(nn.Module): method __init__ (line 452) | def __init__( method forward (line 466) | def forward(self, hidden_states): FILE: src/diffusers/models/upsampling.py class Upsample1D (line 25) | class Upsample1D(nn.Module): method __init__ (line 41) | def __init__( method forward (line 62) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: class Upsample2D (line 75) | class Upsample2D(nn.Module): method __init__ (line 91) | def __init__( method forward (line 141) | def forward(self, hidden_states: torch.Tensor, output_size: Optional[i... class FirUpsample2D (line 187) | class FirUpsample2D(nn.Module): method __init__ (line 201) | def __init__( method _upsample_2d (line 216) | def _upsample_2d( method forward (line 311) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class KUpsample2D (line 321) | class KUpsample2D(nn.Module): method __init__ (line 328) | def __init__(self, pad_mode: str = "reflect"): method forward (line 335) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: class CogVideoXUpsample3D (line 351) | class CogVideoXUpsample3D(nn.Module): method __init__ (line 370) | def __init__( method forward (line 384) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: function upfirdn2d_native (line 415) | def upfirdn2d_native( function upsample_2d (line 465) | def upsample_2d( FILE: src/diffusers/models/vae_flax.py class FlaxDecoderOutput (line 33) | class FlaxDecoderOutput(BaseOutput): class FlaxAutoencoderKLOutput (line 48) | class FlaxAutoencoderKLOutput(BaseOutput): class FlaxUpsample2D (line 61) | class FlaxUpsample2D(nn.Module): method setup (line 75) | def setup(self): method __call__ (line 84) | def __call__(self, hidden_states): class FlaxDownsample2D (line 95) | class FlaxDownsample2D(nn.Module): method setup (line 109) | def setup(self): method __call__ (line 118) | def __call__(self, hidden_states): class FlaxResnetBlock2D (line 125) | class FlaxResnetBlock2D(nn.Module): method setup (line 151) | def setup(self): method __call__ (line 185) | def __call__(self, hidden_states, deterministic=True): class FlaxAttentionBlock (line 202) | class FlaxAttentionBlock(nn.Module): method setup (line 223) | def setup(self): method transpose_for_scores (line 232) | def transpose_for_scores(self, projection): method __call__ (line 240) | def __call__(self, hidden_states): class FlaxDownEncoderBlock2D (line 275) | class FlaxDownEncoderBlock2D(nn.Module): method setup (line 304) | def setup(self): method __call__ (line 322) | def __call__(self, hidden_states, deterministic=True): class FlaxUpDecoderBlock2D (line 332) | class FlaxUpDecoderBlock2D(nn.Module): method setup (line 361) | def setup(self): method __call__ (line 379) | def __call__(self, hidden_states, deterministic=True): class FlaxUNetMidBlock2D (line 389) | class FlaxUNetMidBlock2D(nn.Module): method setup (line 415) | def setup(self): method __call__ (line 452) | def __call__(self, hidden_states, deterministic=True): class FlaxEncoder (line 461) | class FlaxEncoder(nn.Module): method setup (line 506) | def setup(self): method __call__ (line 555) | def __call__(self, sample, deterministic: bool = True): class FlaxDecoder (line 574) | class FlaxDecoder(nn.Module): method setup (line 618) | def setup(self): method __call__ (line 671) | def __call__(self, sample, deterministic: bool = True): class FlaxDiagonalGaussianDistribution (line 689) | class FlaxDiagonalGaussianDistribution(object): method __init__ (line 690) | def __init__(self, parameters, deterministic=False): method sample (line 700) | def sample(self, key): method kl (line 703) | def kl(self, other=None): method nll (line 715) | def nll(self, sample, axis=[1, 2, 3]): method mode (line 722) | def mode(self): class FlaxAutoencoderKL (line 727) | class FlaxAutoencoderKL(nn.Module, FlaxModelMixin, ConfigMixin): method setup (line 790) | def setup(self): method init_weights (line 827) | def init_weights(self, rng: jax.Array) -> FrozenDict: method encode (line 837) | def encode(self, sample, deterministic: bool = True, return_dict: bool... method decode (line 849) | def decode(self, latents, deterministic: bool = True, return_dict: boo... method __call__ (line 863) | def __call__(self, sample, sample_posterior=False, deterministic: bool... FILE: src/diffusers/models/vq_model.py class VQEncoderOutput (line 18) | class VQEncoderOutput(VQEncoderOutput): method __init__ (line 19) | def __init__(self, *args, **kwargs): class VQModel (line 25) | class VQModel(VQModel): method __init__ (line 26) | def __init__(self, *args, **kwargs): FILE: src/diffusers/optimization.py class SchedulerType (line 30) | class SchedulerType(Enum): function get_constant_schedule (line 40) | def get_constant_schedule(optimizer: Optimizer, last_epoch: int = -1) ->... function get_constant_schedule_with_warmup (line 56) | def get_constant_schedule_with_warmup(optimizer: Optimizer, num_warmup_s... function get_piecewise_constant_schedule (line 81) | def get_piecewise_constant_schedule(optimizer: Optimizer, step_rules: st... function get_linear_schedule_with_warmup (line 123) | def get_linear_schedule_with_warmup( function get_cosine_schedule_with_warmup (line 154) | def get_cosine_schedule_with_warmup( function get_cosine_with_hard_restarts_schedule_with_warmup (line 188) | def get_cosine_with_hard_restarts_schedule_with_warmup( function get_polynomial_decay_schedule_with_warmup (line 223) | def get_polynomial_decay_schedule_with_warmup( function get_scheduler (line 289) | def get_scheduler( FILE: src/diffusers/pipelines/amused/pipeline_amused.py class AmusedPipeline (line 42) | class AmusedPipeline(DiffusionPipeline): method __init__ (line 52) | def __init__( method __call__ (line 74) | def __call__( FILE: src/diffusers/pipelines/amused/pipeline_amused_img2img.py class AmusedImg2ImgPipeline (line 52) | class AmusedImg2ImgPipeline(DiffusionPipeline): method __init__ (line 67) | def __init__( method __call__ (line 89) | def __call__( FILE: src/diffusers/pipelines/amused/pipeline_amused_inpaint.py class AmusedInpaintPipeline (line 60) | class AmusedInpaintPipeline(DiffusionPipeline): method __init__ (line 75) | def __init__( method __call__ (line 105) | def __call__( FILE: src/diffusers/pipelines/animatediff/pipeline_animatediff.py class AnimateDiffPipeline (line 69) | class AnimateDiffPipeline( method __init__ (line 110) | def __init__( method encode_prompt (line 146) | def encode_prompt( method encode_image (line 329) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 354) | def prepare_ip_adapter_image_embeds( method decode_latents (line 399) | def decode_latents(self, latents, decode_chunk_size: int = 16): method prepare_extra_step_kwargs (line 418) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 436) | def check_inputs( method prepare_latents (line 505) | def prepare_latents( method guidance_scale (line 538) | def guidance_scale(self): method clip_skip (line 542) | def clip_skip(self): method do_classifier_free_guidance (line 549) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 553) | def cross_attention_kwargs(self): method num_timesteps (line 557) | def num_timesteps(self): method __call__ (line 562) | def __call__( FILE: src/diffusers/pipelines/animatediff/pipeline_animatediff_controlnet.py class AnimateDiffControlNetPipeline (line 106) | class AnimateDiffControlNetPipeline( method __init__ (line 147) | def __init__( method encode_prompt (line 184) | def encode_prompt( method encode_image (line 367) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 392) | def prepare_ip_adapter_image_embeds( method decode_latents (line 438) | def decode_latents(self, latents, decode_chunk_size: int = 16): method prepare_extra_step_kwargs (line 457) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 474) | def check_inputs( method prepare_latents (line 615) | def prepare_latents( method prepare_video (line 647) | def prepare_video( method guidance_scale (line 680) | def guidance_scale(self): method clip_skip (line 684) | def clip_skip(self): method do_classifier_free_guidance (line 691) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 695) | def cross_attention_kwargs(self): method num_timesteps (line 699) | def num_timesteps(self): method __call__ (line 703) | def __call__( FILE: src/diffusers/pipelines/animatediff/pipeline_animatediff_sdxl.py function rescale_noise_cfg (line 115) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 130) | def retrieve_timesteps( class AnimateDiffSDXLPipeline (line 189) | class AnimateDiffSDXLPipeline( method __init__ (line 259) | def __init__( method encode_prompt (line 304) | def encode_prompt( method encode_image (line 539) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 564) | def prepare_ip_adapter_image_embeds( method decode_latents (line 610) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 623) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 640) | def check_inputs( method prepare_latents (line 713) | def prepare_latents( method _get_add_time_ids (line 738) | def _get_add_time_ids( method upcast_vae (line 756) | def upcast_vae(self): method get_guidance_scale_embedding (line 775) | def get_guidance_scale_embedding( method guidance_scale (line 806) | def guidance_scale(self): method guidance_rescale (line 810) | def guidance_rescale(self): method clip_skip (line 814) | def clip_skip(self): method do_classifier_free_guidance (line 821) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 825) | def cross_attention_kwargs(self): method denoising_end (line 829) | def denoising_end(self): method num_timesteps (line 833) | def num_timesteps(self): method interrupt (line 837) | def interrupt(self): method __call__ (line 842) | def __call__( FILE: src/diffusers/pipelines/animatediff/pipeline_animatediff_sparsectrl.py function retrieve_latents (line 110) | def retrieve_latents( class AnimateDiffSparseControlNetPipeline (line 123) | class AnimateDiffSparseControlNetPipeline( method __init__ (line 164) | def __init__( method encode_prompt (line 198) | def encode_prompt( method encode_image (line 381) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 406) | def prepare_ip_adapter_image_embeds( method decode_latents (line 452) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 465) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 482) | def check_inputs( method check_image (line 577) | def check_image(self, image, prompt, prompt_embeds): method prepare_latents (line 615) | def prepare_latents( method prepare_image (line 640) | def prepare_image(self, image, width, height, device, dtype): method prepare_sparse_control_conditioning (line 661) | def prepare_sparse_control_conditioning( method guidance_scale (line 680) | def guidance_scale(self): method clip_skip (line 684) | def clip_skip(self): method do_classifier_free_guidance (line 691) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 695) | def cross_attention_kwargs(self): method num_timesteps (line 699) | def num_timesteps(self): method __call__ (line 704) | def __call__( FILE: src/diffusers/pipelines/animatediff/pipeline_animatediff_video2video.py function retrieve_latents (line 100) | def retrieve_latents( function retrieve_timesteps (line 114) | def retrieve_timesteps( class AnimateDiffVideoToVideoPipeline (line 173) | class AnimateDiffVideoToVideoPipeline( method __init__ (line 214) | def __init__( method encode_prompt (line 250) | def encode_prompt( method encode_image (line 433) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 458) | def prepare_ip_adapter_image_embeds( method encode_video (line 503) | def encode_video(self, video, generator, decode_chunk_size: int = 16) ... method decode_latents (line 512) | def decode_latents(self, latents, decode_chunk_size: int = 16): method prepare_extra_step_kwargs (line 531) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 548) | def check_inputs( method get_timesteps (line 620) | def get_timesteps(self, num_inference_steps, timesteps, strength, devi... method prepare_latents (line 629) | def prepare_latents( method guidance_scale (line 716) | def guidance_scale(self): method clip_skip (line 720) | def clip_skip(self): method do_classifier_free_guidance (line 727) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 731) | def cross_attention_kwargs(self): method num_timesteps (line 735) | def num_timesteps(self): method __call__ (line 739) | def __call__( FILE: src/diffusers/pipelines/animatediff/pipeline_output.py class AnimateDiffPipelineOutput (line 12) | class AnimateDiffPipelineOutput(BaseOutput): FILE: src/diffusers/pipelines/audioldm/pipeline_audioldm.py class AudioLDMPipeline (line 52) | class AudioLDMPipeline(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 78) | def __init__( method _encode_prompt (line 99) | def _encode_prompt( method decode_latents (line 238) | def decode_latents(self, latents): method mel_spectrogram_to_waveform (line 243) | def mel_spectrogram_to_waveform(self, mel_spectrogram): method prepare_extra_step_kwargs (line 253) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 270) | def check_inputs( method prepare_latents (line 329) | def prepare_latents(self, batch_size, num_channels_latents, height, dt... method __call__ (line 353) | def __call__( FILE: src/diffusers/pipelines/audioldm2/modeling_audioldm2.py function add_special_tokens (line 47) | def add_special_tokens(hidden_states, attention_mask, sos_token, eos_tok... class AudioLDM2ProjectionModelOutput (line 63) | class AudioLDM2ProjectionModelOutput(BaseOutput): class AudioLDM2ProjectionModel (line 82) | class AudioLDM2ProjectionModel(ModelMixin, ConfigMixin): method __init__ (line 98) | def __init__( method forward (line 126) | def forward( class AudioLDM2UNet2DConditionModel (line 167) | class AudioLDM2UNet2DConditionModel(ModelMixin, ConfigMixin, UNet2DCondi... method __init__ (line 242) | def __init__( method attn_processors (line 536) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 560) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 595) | def set_default_attn_processor(self): method set_attention_slice (line 611) | def set_attention_slice(self, slice_size): method _set_gradient_checkpointing (line 677) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 681) | def forward( function get_down_block (line 892) | def get_down_block( function get_up_block (line 949) | def get_up_block( class CrossAttnDownBlock2D (line 1006) | class CrossAttnDownBlock2D(nn.Module): method __init__ (line 1007) | def __init__( method forward (line 1092) | def forward( class UNetMidBlock2DCrossAttn (line 1185) | class UNetMidBlock2DCrossAttn(nn.Module): method __init__ (line 1186) | def __init__( method forward (line 1271) | def forward( class CrossAttnUpBlock2D (line 1356) | class CrossAttnUpBlock2D(nn.Module): method __init__ (line 1357) | def __init__( method forward (line 1438) | def forward( FILE: src/diffusers/pipelines/audioldm2/pipeline_audioldm2.py function prepare_inputs_for_generation (line 118) | def prepare_inputs_for_generation( class AudioLDM2Pipeline (line 136) | class AudioLDM2Pipeline(DiffusionPipeline): method __init__ (line 181) | def __init__( method enable_vae_slicing (line 213) | def enable_vae_slicing(self): method disable_vae_slicing (line 221) | def disable_vae_slicing(self): method enable_model_cpu_offload (line 228) | def enable_model_cpu_offload(self, gpu_id=0): method generate_language_model (line 265) | def generate_language_model( method encode_prompt (line 307) | def encode_prompt( method mel_spectrogram_to_waveform (line 640) | def mel_spectrogram_to_waveform(self, mel_spectrogram): method score_waveforms (line 649) | def score_waveforms(self, text, audio, num_waveforms_per_prompt, devic... method prepare_extra_step_kwargs (line 674) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 691) | def check_inputs( method prepare_latents (line 790) | def prepare_latents(self, batch_size, num_channels_latents, height, dt... method __call__ (line 814) | def __call__( FILE: src/diffusers/pipelines/aura_flow/pipeline_aura_flow.py function retrieve_timesteps (line 48) | def retrieve_timesteps( class AuraFlowPipeline (line 107) | class AuraFlowPipeline(DiffusionPipeline): method __init__ (line 128) | def __init__( method check_inputs (line 147) | def check_inputs( method encode_prompt (line 205) | def encode_prompt( method prepare_latents (line 333) | def prepare_latents( method upcast_vae (line 365) | def upcast_vae(self): method __call__ (line 385) | def __call__( FILE: src/diffusers/pipelines/auto_pipeline.py function _get_connected_pipeline (line 181) | def _get_connected_pipeline(pipeline_cls): function _get_task_class (line 195) | def _get_task_class(mapping, pipeline_class_name, throw_error_if_not_exi... class AutoPipelineForText2Image (line 213) | class AutoPipelineForText2Image(ConfigMixin): method __init__ (line 231) | def __init__(self, *args, **kwargs): method from_pretrained (line 240) | def from_pretrained(cls, pretrained_model_or_path, **kwargs): method from_pipe (line 385) | def from_pipe(cls, pipeline, **kwargs): class AutoPipelineForImage2Image (line 499) | class AutoPipelineForImage2Image(ConfigMixin): method __init__ (line 517) | def __init__(self, *args, **kwargs): method from_pretrained (line 526) | def from_pretrained(cls, pretrained_model_or_path, **kwargs): method from_pipe (line 672) | def from_pipe(cls, pipeline, **kwargs): class AutoPipelineForInpainting (line 792) | class AutoPipelineForInpainting(ConfigMixin): method __init__ (line 810) | def __init__(self, *args, **kwargs): method from_pretrained (line 819) | def from_pretrained(cls, pretrained_model_or_path, **kwargs): method from_pipe (line 964) | def from_pipe(cls, pipeline, **kwargs): FILE: src/diffusers/pipelines/blip_diffusion/blip_image_processing.py class BlipImageProcessor (line 49) | class BlipImageProcessor(BaseImageProcessor): method __init__ (line 86) | def __init__( method resize (line 116) | def resize( method preprocess (line 164) | def preprocess( method postprocess (line 301) | def postprocess(self, sample: torch.Tensor, output_type: str = "pil"): FILE: src/diffusers/pipelines/blip_diffusion/modeling_blip2.py class Blip2TextEmbeddings (line 47) | class Blip2TextEmbeddings(nn.Module): method __init__ (line 50) | def __init__(self, config): method forward (line 66) | def forward( class Blip2VisionEmbeddings (line 101) | class Blip2VisionEmbeddings(nn.Module): method __init__ (line 102) | def __init__(self, config: Blip2VisionConfig): method forward (line 120) | def forward(self, pixel_values: torch.Tensor) -> torch.Tensor: class Blip2QFormerEncoder (line 133) | class Blip2QFormerEncoder(nn.Module): method __init__ (line 134) | def __init__(self, config): method forward (line 142) | def forward( class Blip2QFormerLayer (line 236) | class Blip2QFormerLayer(nn.Module): method __init__ (line 237) | def __init__(self, config, layer_idx): method forward (line 256) | def forward( method feed_forward_chunk (line 327) | def feed_forward_chunk(self, attention_output): method feed_forward_chunk_query (line 332) | def feed_forward_chunk_query(self, attention_output): class ProjLayer (line 339) | class ProjLayer(nn.Module): method __init__ (line 340) | def __init__(self, in_dim, out_dim, hidden_dim, drop_p=0.1, eps=1e-12): method forward (line 351) | def forward(self, x): class Blip2VisionModel (line 361) | class Blip2VisionModel(Blip2PreTrainedModel): method __init__ (line 365) | def __init__(self, config: Blip2VisionConfig): method forward (line 377) | def forward( method get_input_embeddings (line 421) | def get_input_embeddings(self): class Blip2QFormerModel (line 426) | class Blip2QFormerModel(Blip2PreTrainedModel): method __init__ (line 431) | def __init__(self, config: Blip2Config): method get_input_embeddings (line 454) | def get_input_embeddings(self): method set_input_embeddings (line 457) | def set_input_embeddings(self, value): method _prune_heads (line 460) | def _prune_heads(self, heads_to_prune): method get_extended_attention_mask (line 468) | def get_extended_attention_mask( method forward (line 513) | def forward( FILE: src/diffusers/pipelines/blip_diffusion/modeling_ctx_clip.py function _expand_mask (line 25) | def _expand_mask(mask: torch.Tensor, dtype: torch.dtype, tgt_len: Option... class ContextCLIPTextModel (line 42) | class ContextCLIPTextModel(CLIPPreTrainedModel): method __init__ (line 47) | def __init__(self, config: CLIPTextConfig): method forward (line 53) | def forward( class ContextCLIPTextTransformer (line 76) | class ContextCLIPTextTransformer(nn.Module): method __init__ (line 77) | def __init__(self, config: CLIPTextConfig): method forward (line 85) | def forward( method _build_causal_attention_mask (line 162) | def _build_causal_attention_mask(self, bsz, seq_len, dtype): class ContextCLIPTextEmbeddings (line 172) | class ContextCLIPTextEmbeddings(nn.Module): method __init__ (line 173) | def __init__(self, config: CLIPTextConfig): method forward (line 183) | def forward( FILE: src/diffusers/pipelines/blip_diffusion/pipeline_blip_diffusion.py class BlipDiffusionPipeline (line 75) | class BlipDiffusionPipeline(DiffusionPipeline): method __init__ (line 103) | def __init__( method get_query_embeddings (line 129) | def get_query_embeddings(self, input_image, src_subject): method _build_prompt (line 133) | def _build_prompt(self, prompts, tgt_subjects, prompt_strength=1.0, pr... method prepare_latents (line 143) | def prepare_latents(self, batch_size, num_channels, height, width, dty... method encode_prompt (line 160) | def encode_prompt(self, query_embeds, prompt, device=None): method __call__ (line 188) | def __call__( FILE: src/diffusers/pipelines/cogvideo/pipeline_cogvideox.py function retrieve_timesteps (line 59) | def retrieve_timesteps( class CogVideoXPipelineOutput (line 119) | class CogVideoXPipelineOutput(BaseOutput): class CogVideoXPipeline (line 133) | class CogVideoXPipeline(DiffusionPipeline): method __init__ (line 165) | def __init__( method _get_t5_prompt_embeds (line 187) | def _get_t5_prompt_embeds( method encode_prompt (line 229) | def encode_prompt( method prepare_latents (line 310) | def prepare_latents( method decode_latents (line 335) | def decode_latents(self, latents: torch.Tensor, num_seconds: int): method prepare_extra_step_kwargs (line 352) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 370) | def check_inputs( method guidance_scale (line 422) | def guidance_scale(self): method num_timesteps (line 426) | def num_timesteps(self): method interrupt (line 430) | def interrupt(self): method __call__ (line 435) | def __call__( FILE: src/diffusers/pipelines/consistency_models/pipeline_consistency_models.py class ConsistencyModelPipeline (line 63) | class ConsistencyModelPipeline(DiffusionPipeline): method __init__ (line 80) | def __init__(self, unet: UNet2DModel, scheduler: CMStochasticIterative... method prepare_latents (line 90) | def prepare_latents(self, batch_size, num_channels, height, width, dty... method postprocess_image (line 108) | def postprocess_image(self, sample: torch.Tensor, output_type: str = "... method prepare_class_labels (line 128) | def prepare_class_labels(self, batch_size, device, class_labels=None): method check_inputs (line 144) | def check_inputs(self, num_inference_steps, timesteps, latents, batch_... method __call__ (line 169) | def __call__( FILE: src/diffusers/pipelines/controlnet/multicontrolnet.py class MultiControlNetModel (line 15) | class MultiControlNetModel(ModelMixin): method __init__ (line 28) | def __init__(self, controlnets: Union[List[ControlNetModel], Tuple[Con... method forward (line 32) | def forward( method save_pretrained (line 75) | def save_pretrained( method from_pretrained (line 114) | def from_pretrained(cls, pretrained_model_path: Optional[Union[str, os... FILE: src/diffusers/pipelines/controlnet/pipeline_controlnet.py function retrieve_timesteps (line 96) | def retrieve_timesteps( class StableDiffusionControlNetPipeline (line 155) | class StableDiffusionControlNetPipeline( method __init__ (line 205) | def __init__( method _encode_prompt (line 258) | def _encode_prompt( method encode_prompt (line 291) | def encode_prompt( method encode_image (line 474) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 499) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 545) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 560) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 572) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 589) | def check_inputs( method check_image (line 752) | def check_image(self, image, prompt, prompt_embeds): method prepare_image (line 789) | def prepare_image( method prepare_latents (line 820) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 843) | def get_guidance_scale_embedding( method guidance_scale (line 874) | def guidance_scale(self): method clip_skip (line 878) | def clip_skip(self): method do_classifier_free_guidance (line 885) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 889) | def cross_attention_kwargs(self): method num_timesteps (line 893) | def num_timesteps(self): method __call__ (line 898) | def __call__( FILE: src/diffusers/pipelines/controlnet/pipeline_controlnet_blip_diffusion.py class BlipDiffusionControlNetPipeline (line 82) | class BlipDiffusionControlNetPipeline(DiffusionPipeline): method __init__ (line 112) | def __init__( method get_query_embeddings (line 140) | def get_query_embeddings(self, input_image, src_subject): method _build_prompt (line 144) | def _build_prompt(self, prompts, tgt_subjects, prompt_strength=1.0, pr... method prepare_latents (line 154) | def prepare_latents(self, batch_size, num_channels, height, width, dty... method encode_prompt (line 171) | def encode_prompt(self, query_embeds, prompt, device=None): method prepare_control_image (line 198) | def prepare_control_image( method __call__ (line 236) | def __call__( FILE: src/diffusers/pipelines/controlnet/pipeline_controlnet_img2img.py function retrieve_latents (line 96) | def retrieve_latents( function prepare_image (line 109) | def prepare_image(image): class StableDiffusionControlNetImg2ImgPipeline (line 133) | class StableDiffusionControlNetImg2ImgPipeline( method __init__ (line 183) | def __init__( method _encode_prompt (line 236) | def _encode_prompt( method encode_prompt (line 269) | def encode_prompt( method encode_image (line 452) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 477) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 523) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 538) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 550) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 567) | def check_inputs( method check_image (line 725) | def check_image(self, image, prompt, prompt_embeds): method prepare_control_image (line 763) | def prepare_control_image( method get_timesteps (line 794) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 806) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method guidance_scale (line 872) | def guidance_scale(self): method clip_skip (line 876) | def clip_skip(self): method do_classifier_free_guidance (line 883) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 887) | def cross_attention_kwargs(self): method num_timesteps (line 891) | def num_timesteps(self): method __call__ (line 896) | def __call__( FILE: src/diffusers/pipelines/controlnet/pipeline_controlnet_inpaint.py function retrieve_latents (line 108) | def retrieve_latents( class StableDiffusionControlNetInpaintPipeline (line 121) | class StableDiffusionControlNetInpaintPipeline( method __init__ (line 182) | def __init__( method _encode_prompt (line 238) | def _encode_prompt( method encode_prompt (line 271) | def encode_prompt( method encode_image (line 454) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 479) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 525) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 540) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 552) | def prepare_extra_step_kwargs(self, generator, eta): method get_timesteps (line 570) | def get_timesteps(self, num_inference_steps, strength, device): method check_inputs (line 581) | def check_inputs( method check_image (line 760) | def check_image(self, image, prompt, prompt_embeds): method prepare_control_image (line 797) | def prepare_control_image( method prepare_latents (line 832) | def prepare_latents( method prepare_mask_latents (line 896) | def prepare_mask_latents( method _encode_vae_image (line 942) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method guidance_scale (line 957) | def guidance_scale(self): method clip_skip (line 961) | def clip_skip(self): method do_classifier_free_guidance (line 968) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 972) | def cross_attention_kwargs(self): method num_timesteps (line 976) | def num_timesteps(self): method __call__ (line 981) | def __call__( FILE: src/diffusers/pipelines/controlnet/pipeline_controlnet_inpaint_sd_xl.py function retrieve_latents (line 68) | def retrieve_latents( function rescale_noise_cfg (line 139) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class StableDiffusionXLControlNetInpaintPipeline (line 153) | class StableDiffusionXLControlNetInpaintPipeline( method __init__ (line 221) | def __init__( method encode_prompt (line 273) | def encode_prompt( method encode_image (line 508) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 533) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 579) | def prepare_extra_step_kwargs(self, generator, eta): method check_image (line 596) | def check_image(self, image, prompt, prompt_embeds): method check_inputs (line 633) | def check_inputs( method prepare_control_image (line 850) | def prepare_control_image( method prepare_latents (line 884) | def prepare_latents( method _encode_vae_image (line 951) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method prepare_mask_latents (line 974) | def prepare_mask_latents( method get_timesteps (line 1022) | def get_timesteps(self, num_inference_steps, strength, device, denoisi... method _get_add_time_ids (line 1058) | def _get_add_time_ids( method upcast_vae (line 1105) | def upcast_vae(self): method guidance_scale (line 1123) | def guidance_scale(self): method clip_skip (line 1127) | def clip_skip(self): method do_classifier_free_guidance (line 1134) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 1138) | def cross_attention_kwargs(self): method num_timesteps (line 1142) | def num_timesteps(self): method __call__ (line 1147) | def __call__( FILE: src/diffusers/pipelines/controlnet/pipeline_controlnet_sd_xl.py function retrieve_timesteps (line 117) | def retrieve_timesteps( class StableDiffusionXLControlNetPipeline (line 176) | class StableDiffusionXLControlNetPipeline( method __init__ (line 247) | def __init__( method encode_prompt (line 294) | def encode_prompt( method encode_image (line 529) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 554) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 600) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 617) | def check_inputs( method check_image (line 807) | def check_image(self, image, prompt, prompt_embeds): method prepare_image (line 845) | def prepare_image( method prepare_latents (line 876) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 899) | def _get_add_time_ids( method upcast_vae (line 918) | def upcast_vae(self): method get_guidance_scale_embedding (line 936) | def get_guidance_scale_embedding( method guidance_scale (line 967) | def guidance_scale(self): method clip_skip (line 971) | def clip_skip(self): method do_classifier_free_guidance (line 978) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 982) | def cross_attention_kwargs(self): method denoising_end (line 986) | def denoising_end(self): method num_timesteps (line 990) | def num_timesteps(self): method __call__ (line 995) | def __call__( FILE: src/diffusers/pipelines/controlnet/pipeline_controlnet_sd_xl_img2img.py function retrieve_latents (line 146) | def retrieve_latents( class StableDiffusionXLControlNetImg2ImgPipeline (line 159) | class StableDiffusionXLControlNetImg2ImgPipeline( method __init__ (line 239) | def __init__( method encode_prompt (line 288) | def encode_prompt( method encode_image (line 523) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 548) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 594) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 611) | def check_inputs( method check_image (line 813) | def check_image(self, image, prompt, prompt_embeds): method prepare_control_image (line 851) | def prepare_control_image( method get_timesteps (line 882) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 894) | def prepare_latents( method _get_add_time_ids (line 981) | def _get_add_time_ids( method upcast_vae (line 1033) | def upcast_vae(self): method guidance_scale (line 1051) | def guidance_scale(self): method clip_skip (line 1055) | def clip_skip(self): method do_classifier_free_guidance (line 1062) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 1066) | def cross_attention_kwargs(self): method num_timesteps (line 1070) | def num_timesteps(self): method __call__ (line 1075) | def __call__( FILE: src/diffusers/pipelines/controlnet/pipeline_flax_controlnet.py class FlaxStableDiffusionControlNetPipeline (line 111) | class FlaxStableDiffusionControlNetPipeline(FlaxDiffusionPipeline): method __init__ (line 141) | def __init__( method prepare_text_inputs (line 180) | def prepare_text_inputs(self, prompt: Union[str, List[str]]): method prepare_image_inputs (line 194) | def prepare_image_inputs(self, image: Union[Image.Image, List[Image.Im... method _get_has_nsfw_concepts (line 205) | def _get_has_nsfw_concepts(self, features, params): method _run_safety_checker (line 209) | def _run_safety_checker(self, images, safety_model_params, jit=False): method _generate (line 239) | def _generate( method __call__ (line 349) | def __call__( function _p_generate (line 488) | def _p_generate( function _p_get_has_nsfw_concepts (line 514) | def _p_get_has_nsfw_concepts(pipe, features, params): function unshard (line 518) | def unshard(x: jnp.ndarray): function preprocess (line 525) | def preprocess(image, dtype): FILE: src/diffusers/pipelines/controlnet_hunyuandit/pipeline_hunyuandit_controlnet.py function map_to_standard_shapes (line 114) | def map_to_standard_shapes(target_width, target_height): function get_resize_crop_region_for_grid (line 122) | def get_resize_crop_region_for_grid(src, tgt_size): function rescale_noise_cfg (line 143) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class HunyuanDiTControlNetPipeline (line 157) | class HunyuanDiTControlNetPipeline(DiffusionPipeline): method __init__ (line 208) | def __init__( method encode_prompt (line 270) | def encode_prompt( method run_safety_checker (line 436) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 451) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 469) | def check_inputs( method prepare_latents (line 546) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method prepare_image (line 569) | def prepare_image( method guidance_scale (line 604) | def guidance_scale(self): method guidance_rescale (line 608) | def guidance_rescale(self): method do_classifier_free_guidance (line 615) | def do_classifier_free_guidance(self): method num_timesteps (line 619) | def num_timesteps(self): method interrupt (line 623) | def interrupt(self): method __call__ (line 628) | def __call__( FILE: src/diffusers/pipelines/controlnet_sd3/pipeline_stable_diffusion_3_controlnet.py function retrieve_timesteps (line 78) | def retrieve_timesteps( class StableDiffusion3ControlNetPipeline (line 137) | class StableDiffusion3ControlNetPipeline(DiffusionPipeline, SD3LoraLoade... method __init__ (line 179) | def __init__( method _get_t5_prompt_embeds (line 222) | def _get_t5_prompt_embeds( method _get_clip_prompt_embeds (line 279) | def _get_clip_prompt_embeds( method encode_prompt (line 335) | def encode_prompt( method check_inputs (line 528) | def check_inputs( method prepare_latents (line 619) | def prepare_latents( method prepare_image (line 650) | def prepare_image( method guidance_scale (line 685) | def guidance_scale(self): method clip_skip (line 689) | def clip_skip(self): method do_classifier_free_guidance (line 696) | def do_classifier_free_guidance(self): method joint_attention_kwargs (line 700) | def joint_attention_kwargs(self): method num_timesteps (line 704) | def num_timesteps(self): method interrupt (line 708) | def interrupt(self): method __call__ (line 713) | def __call__( FILE: src/diffusers/pipelines/controlnet_xs/pipeline_controlnet_xs.py class StableDiffusionControlNetXSPipeline (line 92) | class StableDiffusionControlNetXSPipeline( method __init__ (line 138) | def __init__( method _encode_prompt (line 189) | def _encode_prompt( method encode_prompt (line 222) | def encode_prompt( method run_safety_checker (line 405) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 420) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 432) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 449) | def check_inputs( method check_image (line 519) | def check_image(self, image, prompt, prompt_embeds): method prepare_image (line 556) | def prepare_image( method prepare_latents (line 586) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method guidance_scale (line 610) | def guidance_scale(self): method clip_skip (line 615) | def clip_skip(self): method do_classifier_free_guidance (line 620) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 625) | def cross_attention_kwargs(self): method num_timesteps (line 630) | def num_timesteps(self): method __call__ (line 635) | def __call__( FILE: src/diffusers/pipelines/controlnet_xs/pipeline_controlnet_xs_sd_xl.py class StableDiffusionXLControlNetXSPipeline (line 106) | class StableDiffusionXLControlNetXSPipeline( method __init__ (line 169) | def __init__( method encode_prompt (line 214) | def encode_prompt( method prepare_extra_step_kwargs (line 449) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 466) | def check_inputs( method check_image (line 563) | def check_image(self, image, prompt, prompt_embeds): method prepare_image (line 600) | def prepare_image( method prepare_latents (line 630) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 652) | def _get_add_time_ids( method upcast_vae (line 671) | def upcast_vae(self): method guidance_scale (line 690) | def guidance_scale(self): method clip_skip (line 695) | def clip_skip(self): method do_classifier_free_guidance (line 700) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 705) | def cross_attention_kwargs(self): method num_timesteps (line 710) | def num_timesteps(self): method __call__ (line 715) | def __call__( FILE: src/diffusers/pipelines/dance_diffusion/pipeline_dance_diffusion.py class DanceDiffusionPipeline (line 28) | class DanceDiffusionPipeline(DiffusionPipeline): method __init__ (line 45) | def __init__(self, unet, scheduler): method __call__ (line 50) | def __call__( FILE: src/diffusers/pipelines/ddim/pipeline_ddim.py class DDIMPipeline (line 24) | class DDIMPipeline(DiffusionPipeline): method __init__ (line 41) | def __init__(self, unet, scheduler): method __call__ (line 50) | def __call__( FILE: src/diffusers/pipelines/ddpm/pipeline_ddpm.py class DDPMPipeline (line 24) | class DDPMPipeline(DiffusionPipeline): method __init__ (line 41) | def __init__(self, unet, scheduler): method __call__ (line 46) | def __call__( FILE: src/diffusers/pipelines/deepfloyd_if/pipeline_if.py class IFPipeline (line 87) | class IFPipeline(DiffusionPipeline, StableDiffusionLoraLoaderMixin): method __init__ (line 119) | def __init__( method encode_prompt (line 160) | def encode_prompt( method run_safety_checker (line 313) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 327) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 344) | def check_inputs( method prepare_intermediate_images (line 386) | def prepare_intermediate_images(self, batch_size, num_channels, height... method _text_preprocessing (line 400) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 424) | def _clean_caption(self, caption): method __call__ (line 540) | def __call__( FILE: src/diffusers/pipelines/deepfloyd_if/pipeline_if_img2img.py function resize (line 39) | def resize(images: PIL.Image.Image, img_size: int) -> PIL.Image.Image: class IFImg2ImgPipeline (line 111) | class IFImg2ImgPipeline(DiffusionPipeline, StableDiffusionLoraLoaderMixin): method __init__ (line 143) | def __init__( method encode_prompt (line 184) | def encode_prompt( method run_safety_checker (line 338) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 352) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 369) | def check_inputs( method _text_preprocessing (line 443) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 468) | def _clean_caption(self, caption): method preprocess_image (line 582) | def preprocess_image(self, image: PIL.Image.Image) -> torch.Tensor: method get_timesteps (line 619) | def get_timesteps(self, num_inference_steps, strength): method prepare_intermediate_images (line 630) | def prepare_intermediate_images( method __call__ (line 654) | def __call__( FILE: src/diffusers/pipelines/deepfloyd_if/pipeline_if_img2img_superresolution.py function resize (line 42) | def resize(images: PIL.Image.Image, img_size: int) -> PIL.Image.Image: class IFImg2ImgSuperResolutionPipeline (line 114) | class IFImg2ImgSuperResolutionPipeline(DiffusionPipeline, StableDiffusio... method __init__ (line 147) | def __init__( method _text_preprocessing (line 195) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 220) | def _clean_caption(self, caption): method encode_prompt (line 336) | def encode_prompt( method run_safety_checker (line 490) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 504) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 521) | def check_inputs( method preprocess_original_image (line 631) | def preprocess_original_image(self, image: PIL.Image.Image) -> torch.T... method preprocess_image (line 668) | def preprocess_image(self, image: PIL.Image.Image, num_images_per_prom... method get_timesteps (line 700) | def get_timesteps(self, num_inference_steps, strength): method prepare_intermediate_images (line 712) | def prepare_intermediate_images( method __call__ (line 736) | def __call__( FILE: src/diffusers/pipelines/deepfloyd_if/pipeline_if_inpainting.py function resize (line 40) | def resize(images: PIL.Image.Image, img_size: int) -> PIL.Image.Image: class IFInpaintingPipeline (line 114) | class IFInpaintingPipeline(DiffusionPipeline, StableDiffusionLoraLoaderM... method __init__ (line 146) | def __init__( method encode_prompt (line 188) | def encode_prompt( method run_safety_checker (line 342) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 356) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 373) | def check_inputs( method _text_preprocessing (line 483) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 508) | def _clean_caption(self, caption): method preprocess_image (line 623) | def preprocess_image(self, image: PIL.Image.Image) -> torch.Tensor: method preprocess_mask_image (line 659) | def preprocess_mask_image(self, mask_image) -> torch.Tensor: method get_timesteps (line 709) | def get_timesteps(self, num_inference_steps, strength): method prepare_intermediate_images (line 720) | def prepare_intermediate_images( method __call__ (line 746) | def __call__( FILE: src/diffusers/pipelines/deepfloyd_if/pipeline_if_inpainting_superresolution.py function resize (line 42) | def resize(images: PIL.Image.Image, img_size: int) -> PIL.Image.Image: class IFInpaintingSuperResolutionPipeline (line 116) | class IFInpaintingSuperResolutionPipeline(DiffusionPipeline, StableDiffu... method __init__ (line 149) | def __init__( method _text_preprocessing (line 197) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 222) | def _clean_caption(self, caption): method encode_prompt (line 338) | def encode_prompt( method run_safety_checker (line 492) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 506) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 523) | def check_inputs( method preprocess_original_image (line 667) | def preprocess_original_image(self, image: PIL.Image.Image) -> torch.T... method preprocess_image (line 704) | def preprocess_image(self, image: PIL.Image.Image, num_images_per_prom... method preprocess_mask_image (line 736) | def preprocess_mask_image(self, mask_image) -> torch.Tensor: method get_timesteps (line 786) | def get_timesteps(self, num_inference_steps, strength): method prepare_intermediate_images (line 798) | def prepare_intermediate_images( method __call__ (line 824) | def __call__( FILE: src/diffusers/pipelines/deepfloyd_if/pipeline_if_superresolution.py class IFSuperResolutionPipeline (line 72) | class IFSuperResolutionPipeline(DiffusionPipeline, StableDiffusionLoraLo... method __init__ (line 105) | def __init__( method _text_preprocessing (line 153) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 178) | def _clean_caption(self, caption): method encode_prompt (line 294) | def encode_prompt( method run_safety_checker (line 448) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 462) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 479) | def check_inputs( method prepare_intermediate_images (line 559) | def prepare_intermediate_images(self, batch_size, num_channels, height... method preprocess_image (line 573) | def preprocess_image(self, image, num_images_per_prompt, device): method __call__ (line 606) | def __call__( FILE: src/diffusers/pipelines/deepfloyd_if/pipeline_output.py class IFPipelineOutput (line 11) | class IFPipelineOutput(BaseOutput): FILE: src/diffusers/pipelines/deepfloyd_if/safety_checker.py class IFSafetyChecker (line 12) | class IFSafetyChecker(PreTrainedModel): method __init__ (line 17) | def __init__(self, config: CLIPConfig): method forward (line 26) | def forward(self, clip_input, images, p_threshold=0.5, w_threshold=0.5): FILE: src/diffusers/pipelines/deepfloyd_if/watermark.py class IFWatermarker (line 12) | class IFWatermarker(ModelMixin, ConfigMixin): method __init__ (line 13) | def __init__(self): method apply_watermark (line 19) | def apply_watermark(self, images: List[PIL.Image.Image], sample_size=N... FILE: src/diffusers/pipelines/deprecated/alt_diffusion/modeling_roberta_series.py class TransformationModelOutput (line 11) | class TransformationModelOutput(ModelOutput): class RobertaSeriesConfig (line 39) | class RobertaSeriesConfig(XLMRobertaConfig): method __init__ (line 40) | def __init__( class RobertaSeriesModelWithTransformation (line 58) | class RobertaSeriesModelWithTransformation(RobertaPreTrainedModel): method __init__ (line 64) | def __init__(self, config): method forward (line 74) | def forward( FILE: src/diffusers/pipelines/deprecated/alt_diffusion/pipeline_alt_diffusion.py function rescale_noise_cfg (line 67) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 82) | def retrieve_timesteps( class AltDiffusionPipeline (line 141) | class AltDiffusionPipeline( method __init__ (line 187) | def __init__( method _encode_prompt (line 279) | def _encode_prompt( method encode_prompt (line 311) | def encode_prompt( method encode_image (line 492) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method run_safety_checker (line 516) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 530) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 541) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 558) | def check_inputs( method prepare_latents (line 610) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 632) | def get_guidance_scale_embedding(self, w, embedding_dim=512, dtype=tor... method guidance_scale (line 661) | def guidance_scale(self): method guidance_rescale (line 665) | def guidance_rescale(self): method clip_skip (line 669) | def clip_skip(self): method do_classifier_free_guidance (line 676) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 680) | def cross_attention_kwargs(self): method num_timesteps (line 684) | def num_timesteps(self): method __call__ (line 689) | def __call__( FILE: src/diffusers/pipelines/deprecated/alt_diffusion/pipeline_alt_diffusion_img2img.py function retrieve_latents (line 84) | def retrieve_latents( function preprocess (line 98) | def preprocess(image): function retrieve_timesteps (line 122) | def retrieve_timesteps( class AltDiffusionImg2ImgPipeline (line 181) | class AltDiffusionImg2ImgPipeline( method __init__ (line 227) | def __init__( method _encode_prompt (line 319) | def _encode_prompt( method encode_prompt (line 351) | def encode_prompt( method encode_image (line 532) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method run_safety_checker (line 556) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 570) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 581) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 598) | def check_inputs( method get_timesteps (line 649) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 658) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method get_guidance_scale_embedding (line 716) | def get_guidance_scale_embedding(self, w, embedding_dim=512, dtype=tor... method guidance_scale (line 745) | def guidance_scale(self): method clip_skip (line 749) | def clip_skip(self): method do_classifier_free_guidance (line 756) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 760) | def cross_attention_kwargs(self): method num_timesteps (line 764) | def num_timesteps(self): method __call__ (line 769) | def __call__( FILE: src/diffusers/pipelines/deprecated/alt_diffusion/pipeline_output.py class AltDiffusionPipelineOutput (line 14) | class AltDiffusionPipelineOutput(BaseOutput): FILE: src/diffusers/pipelines/deprecated/audio_diffusion/mel.py class Mel (line 37) | class Mel(ConfigMixin, SchedulerMixin): method __init__ (line 59) | def __init__( method set_resolution (line 80) | def set_resolution(self, x_res: int, y_res: int): method load_audio (line 94) | def load_audio(self, audio_file: str = None, raw_audio: np.ndarray = N... method get_number_of_slices (line 112) | def get_number_of_slices(self) -> int: method get_audio_slice (line 121) | def get_audio_slice(self, slice: int = 0) -> np.ndarray: method get_sample_rate (line 134) | def get_sample_rate(self) -> int: method audio_slice_to_image (line 143) | def audio_slice_to_image(self, slice: int) -> Image.Image: method image_to_audio (line 162) | def image_to_audio(self, image: Image.Image) -> np.ndarray: FILE: src/diffusers/pipelines/deprecated/audio_diffusion/pipeline_audio_diffusion.py class AudioDiffusionPipeline (line 30) | class AudioDiffusionPipeline(DiffusionPipeline): method __init__ (line 51) | def __init__( method get_default_steps (line 61) | def get_default_steps(self) -> int: method __call__ (line 71) | def __call__( method encode (line 271) | def encode(self, images: List[Image.Image], steps: int = 50) -> np.nda... method slerp (line 312) | def slerp(x0: torch.Tensor, x1: torch.Tensor, alpha: float) -> torch.T... FILE: src/diffusers/pipelines/deprecated/latent_diffusion_uncond/pipeline_latent_diffusion_uncond.py class LDMPipeline (line 26) | class LDMPipeline(DiffusionPipeline): method __init__ (line 42) | def __init__(self, vqvae: VQModel, unet: UNet2DModel, scheduler: DDIMS... method __call__ (line 47) | def __call__( FILE: src/diffusers/pipelines/deprecated/pndm/pipeline_pndm.py class PNDMPipeline (line 26) | class PNDMPipeline(DiffusionPipeline): method __init__ (line 43) | def __init__(self, unet: UNet2DModel, scheduler: PNDMScheduler): method __call__ (line 51) | def __call__( FILE: src/diffusers/pipelines/deprecated/repaint/pipeline_repaint.py function _preprocess_image (line 33) | def _preprocess_image(image: Union[List, PIL.Image.Image, torch.Tensor]): function _preprocess_mask (line 56) | def _preprocess_mask(mask: Union[List, PIL.Image.Image, torch.Tensor]): class RePaintPipeline (line 76) | class RePaintPipeline(DiffusionPipeline): method __init__ (line 94) | def __init__(self, unet, scheduler): method __call__ (line 99) | def __call__( FILE: src/diffusers/pipelines/deprecated/score_sde_ve/pipeline_score_sde_ve.py class ScoreSdeVePipeline (line 25) | class ScoreSdeVePipeline(DiffusionPipeline): method __init__ (line 42) | def __init__(self, unet: UNet2DModel, scheduler: ScoreSdeVeScheduler): method __call__ (line 47) | def __call__( FILE: src/diffusers/pipelines/deprecated/spectrogram_diffusion/continuous_encoder.py class SpectrogramContEncoder (line 29) | class SpectrogramContEncoder(ModelMixin, ConfigMixin, ModuleUtilsMixin): method __init__ (line 31) | def __init__( method forward (line 71) | def forward(self, encoder_inputs, encoder_inputs_mask): FILE: src/diffusers/pipelines/deprecated/spectrogram_diffusion/midi_utils.py class NoteRepresentationConfig (line 84) | class NoteRepresentationConfig: class NoteEventData (line 92) | class NoteEventData: class NoteEncodingState (line 101) | class NoteEncodingState: class EventRange (line 109) | class EventRange: class Event (line 116) | class Event: class Tokenizer (line 121) | class Tokenizer: method __init__ (line 122) | def __init__(self, regular_ids: int): method encode (line 127) | def encode(self, token_ids): class Codec (line 145) | class Codec: method __init__ (line 156) | def __init__(self, max_shift_steps: int, steps_per_second: float, even... method num_classes (line 172) | def num_classes(self) -> int: method is_shift_event_index (line 178) | def is_shift_event_index(self, index: int) -> bool: method max_shift_steps (line 182) | def max_shift_steps(self) -> int: method encode_event (line 185) | def encode_event(self, event: Event) -> int: method event_type_range (line 200) | def event_type_range(self, event_type: str) -> Tuple[int, int]: method decode_event_index (line 210) | def decode_event_index(self, index: int) -> Event: class ProgramGranularity (line 222) | class ProgramGranularity: function drop_programs (line 228) | def drop_programs(tokens, codec: Codec): function programs_to_midi_classes (line 234) | def programs_to_midi_classes(tokens, codec): function frame (line 254) | def frame(signal, frame_length, frame_step, pad_end=False, pad_value=0, ... function program_to_slakh_program (line 272) | def program_to_slakh_program(program): function audio_to_frames (line 279) | def audio_to_frames( function note_sequence_to_onsets_and_offsets_and_programs (line 302) | def note_sequence_to_onsets_and_offsets_and_programs( function num_velocity_bins_from_codec (line 332) | def num_velocity_bins_from_codec(codec: Codec): function segment (line 339) | def segment(a, n): function velocity_to_bin (line 343) | def velocity_to_bin(velocity, num_velocity_bins): function note_event_data_to_events (line 350) | def note_event_data_to_events( function note_encoding_state_to_events (line 382) | def note_encoding_state_to_events(state: NoteEncodingState) -> Sequence[... function encode_and_index_events (line 392) | def encode_and_index_events( function extract_sequence_with_indices (line 498) | def extract_sequence_with_indices(features, state_events_end_token=None,... function map_midi_programs (line 524) | def map_midi_programs( function run_length_encode_shifts_fn (line 534) | def run_length_encode_shifts_fn( function note_representation_processor_chain (line 604) | def note_representation_processor_chain(features, codec: Codec, note_rep... class MidiProcessor (line 619) | class MidiProcessor: method __init__ (line 620) | def __init__(self): method __call__ (line 635) | def __call__(self, midi: Union[bytes, os.PathLike, str]): FILE: src/diffusers/pipelines/deprecated/spectrogram_diffusion/notes_encoder.py class SpectrogramNotesEncoder (line 25) | class SpectrogramNotesEncoder(ModelMixin, ConfigMixin, ModuleUtilsMixin): method __init__ (line 27) | def __init__( method forward (line 69) | def forward(self, encoder_input_tokens, encoder_inputs_mask): FILE: src/diffusers/pipelines/deprecated/spectrogram_diffusion/pipeline_spectrogram_diffusion.py class SpectrogramDiffusionPipeline (line 41) | class SpectrogramDiffusionPipeline(DiffusionPipeline): method __init__ (line 60) | def __init__( method scale_features (line 83) | def scale_features(self, features, output_range=(-1.0, 1.0), clip=False): method scale_to_features (line 93) | def scale_to_features(self, outputs, input_range=(-1.0, 1.0), clip=Fal... method encode (line 102) | def encode(self, input_tokens, continuous_inputs, continuous_mask): method decode (line 114) | def decode(self, encodings_and_masks, input_tokens, noise_time): method __call__ (line 130) | def __call__( FILE: src/diffusers/pipelines/deprecated/stable_diffusion_variants/pipeline_cycle_diffusion.py function preprocess (line 41) | def preprocess(image): function retrieve_latents (line 65) | def retrieve_latents( function posterior_sample (line 78) | def posterior_sample(scheduler, latents, timestep, clean_latents, genera... function compute_noise (line 105) | def compute_noise(scheduler, prev_latents, latents, timestep, noise_pred... class CycleDiffusionPipeline (line 139) | class CycleDiffusionPipeline(DiffusionPipeline, TextualInversionLoaderMi... method __init__ (line 174) | def __init__( method _encode_prompt (line 251) | def _encode_prompt( method encode_prompt (line 284) | def encode_prompt( method check_inputs (line 466) | def check_inputs( method prepare_extra_step_kwargs (line 518) | def prepare_extra_step_kwargs(self, generator, eta): method run_safety_checker (line 536) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 551) | def decode_latents(self, latents): method get_timesteps (line 563) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 574) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method __call__ (line 630) | def __call__( FILE: src/diffusers/pipelines/deprecated/stable_diffusion_variants/pipeline_onnx_stable_diffusion_inpaint_legacy.py function preprocess (line 20) | def preprocess(image): function preprocess_mask (line 29) | def preprocess_mask(mask, scale_factor=8): class OnnxStableDiffusionInpaintPipelineLegacy (line 41) | class OnnxStableDiffusionInpaintPipelineLegacy(DiffusionPipeline): method __init__ (line 82) | def __init__( method _encode_prompt (line 152) | def _encode_prompt( method check_inputs (line 255) | def check_inputs( method __call__ (line 297) | def __call__( FILE: src/diffusers/pipelines/deprecated/stable_diffusion_variants/pipeline_stable_diffusion_inpaint_legacy.py function preprocess_image (line 40) | def preprocess_image(image, batch_size): function preprocess_mask (line 50) | def preprocess_mask(mask, batch_size, scale_factor=8): class StableDiffusionInpaintPipelineLegacy (line 81) | class StableDiffusionInpaintPipelineLegacy( method __init__ (line 123) | def __init__( method _encode_prompt (line 221) | def _encode_prompt( method encode_prompt (line 254) | def encode_prompt( method run_safety_checker (line 437) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 452) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 464) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 481) | def check_inputs( method get_timesteps (line 532) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 541) | def prepare_latents(self, image, timestep, num_images_per_prompt, dtyp... method __call__ (line 558) | def __call__( FILE: src/diffusers/pipelines/deprecated/stable_diffusion_variants/pipeline_stable_diffusion_model_editing.py class StableDiffusionModelEditingPipeline (line 39) | class StableDiffusionModelEditingPipeline( method __init__ (line 81) | def __init__( method _encode_prompt (line 159) | def _encode_prompt( method encode_prompt (line 192) | def encode_prompt( method run_safety_checker (line 375) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 390) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 402) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 419) | def check_inputs( method prepare_latents (line 472) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method edit_model (line 495) | def edit_model( method __call__ (line 613) | def __call__( FILE: src/diffusers/pipelines/deprecated/stable_diffusion_variants/pipeline_stable_diffusion_paradigms.py class StableDiffusionParadigmsPipeline (line 65) | class StableDiffusionParadigmsPipeline( method __init__ (line 108) | def __init__( method _encode_prompt (line 154) | def _encode_prompt( method encode_prompt (line 187) | def encode_prompt( method run_safety_checker (line 370) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 385) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 402) | def check_inputs( method prepare_latents (line 455) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _cumsum (line 477) | def _cumsum(self, input, dim, debug=False): method __call__ (line 487) | def __call__( FILE: src/diffusers/pipelines/deprecated/stable_diffusion_variants/pipeline_stable_diffusion_pix2pix_zero.py class Pix2PixInversionPipelineOutput (line 58) | class Pix2PixInversionPipelineOutput(BaseOutput, TextualInversionLoaderM... function preprocess (line 180) | def preprocess(image): function prepare_unet (line 203) | def prepare_unet(unet: UNet2DConditionModel): class Pix2PixZeroL2Loss (line 220) | class Pix2PixZeroL2Loss: method __init__ (line 221) | def __init__(self): method compute_loss (line 224) | def compute_loss(self, predictions, targets): class Pix2PixZeroAttnProcessor (line 228) | class Pix2PixZeroAttnProcessor: method __init__ (line 232) | def __init__(self, is_pix2pix_zero=False): method __call__ (line 237) | def __call__( class StableDiffusionPix2PixZeroPipeline (line 283) | class StableDiffusionPix2PixZeroPipeline(DiffusionPipeline, StableDiffus... method __init__ (line 324) | def __init__( method _encode_prompt (line 373) | def _encode_prompt( method encode_prompt (line 406) | def encode_prompt( method run_safety_checker (line 589) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 604) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 616) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 633) | def check_inputs( method prepare_latents (line 664) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method generate_caption (line 687) | def generate_caption(self, images): method construct_direction (line 706) | def construct_direction(self, embs_source: torch.Tensor, embs_target: ... method get_embeds (line 711) | def get_embeds(self, prompt: List[str], batch_size: int = 16) -> torch... method prepare_image_latents (line 730) | def prepare_image_latents(self, image, batch_size, dtype, device, gene... method get_epsilon (line 779) | def get_epsilon(self, model_output: torch.Tensor, sample: torch.Tensor... method auto_corr_loss (line 796) | def auto_corr_loss(self, hidden_states, generator=None): method kl_divergence (line 811) | def kl_divergence(self, hidden_states): method __call__ (line 818) | def __call__( method invert (line 1109) | def invert( FILE: src/diffusers/pipelines/deprecated/stochastic_karras_ve/pipeline_stochastic_karras_ve.py class KarrasVePipeline (line 25) | class KarrasVePipeline(DiffusionPipeline): method __init__ (line 40) | def __init__(self, unet: UNet2DModel, scheduler: KarrasVeScheduler): method __call__ (line 45) | def __call__( FILE: src/diffusers/pipelines/deprecated/versatile_diffusion/modeling_text_unet.py function get_down_block (line 44) | def get_down_block( function get_up_block (line 109) | def get_up_block( class FourierEmbedder (line 175) | class FourierEmbedder(nn.Module): method __init__ (line 176) | def __init__(self, num_freqs=64, temperature=100): method __call__ (line 186) | def __call__(self, x): class GLIGENTextBoundingboxProjection (line 191) | class GLIGENTextBoundingboxProjection(nn.Module): method __init__ (line 192) | def __init__(self, positive_len, out_dim, feature_type, fourier_freqs=8): method forward (line 233) | def forward( class UNetFlatConditionModel (line 271) | class UNetFlatConditionModel(ModelMixin, ConfigMixin): method __init__ (line 369) | def __init__( method attn_processors (line 829) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 852) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 886) | def set_default_attn_processor(self): method set_attention_slice (line 901) | def set_attention_slice(self, slice_size): method _set_gradient_checkpointing (line 966) | def _set_gradient_checkpointing(self, module, value=False): method enable_freeu (line 970) | def enable_freeu(self, s1, s2, b1, b2): method disable_freeu (line 994) | def disable_freeu(self): method fuse_qkv_projections (line 1002) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 1025) | def unfuse_qkv_projections(self): method unload_lora (line 1038) | def unload_lora(self): method forward (line 1049) | def forward( class LinearMultiDim (line 1427) | class LinearMultiDim(nn.Linear): method __init__ (line 1428) | def __init__(self, in_features, out_features=None, second_dim=4, *args... method forward (line 1437) | def forward(self, input_tensor, *args, **kwargs): class ResnetBlockFlat (line 1446) | class ResnetBlockFlat(nn.Module): method __init__ (line 1447) | def __init__( method forward (line 1507) | def forward(self, input_tensor, temb): class DownBlockFlat (line 1540) | class DownBlockFlat(nn.Module): method __init__ (line 1541) | def __init__( method forward (line 1592) | def forward( class CrossAttnDownBlockFlat (line 1628) | class CrossAttnDownBlockFlat(nn.Module): method __init__ (line 1629) | def __init__( method forward (line 1720) | def forward( class UpBlockFlat (line 1788) | class UpBlockFlat(nn.Module): method __init__ (line 1789) | def __init__( method forward (line 1838) | def forward( class CrossAttnUpBlockFlat (line 1904) | class CrossAttnUpBlockFlat(nn.Module): method __init__ (line 1905) | def __init__( method forward (line 1995) | def forward( class UNetMidBlockFlat (line 2081) | class UNetMidBlockFlat(nn.Module): method __init__ (line 2112) | def __init__( method forward (line 2226) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class UNetMidBlockFlatCrossAttn (line 2237) | class UNetMidBlockFlatCrossAttn(nn.Module): method __init__ (line 2238) | def __init__( method forward (line 2340) | def forward( class UNetMidBlockFlatSimpleCrossAttn (line 2396) | class UNetMidBlockFlatSimpleCrossAttn(nn.Module): method __init__ (line 2397) | def __init__( method forward (line 2481) | def forward( FILE: src/diffusers/pipelines/deprecated/versatile_diffusion/pipeline_versatile_diffusion.py class VersatileDiffusionPipeline (line 20) | class VersatileDiffusionPipeline(DiffusionPipeline): method __init__ (line 56) | def __init__( method image_variation (line 82) | def image_variation( method text_to_image (line 194) | def text_to_image( method dual_guided (line 302) | def dual_guided( FILE: src/diffusers/pipelines/deprecated/versatile_diffusion/pipeline_versatile_diffusion_dual_guided.py class VersatileDiffusionDualGuidedPipeline (line 41) | class VersatileDiffusionDualGuidedPipeline(DiffusionPipeline): method __init__ (line 75) | def __init__( method remove_unused_weights (line 106) | def remove_unused_weights(self): method _convert_to_dual_attention (line 109) | def _convert_to_dual_attention(self): method _revert_dual_attention (line 143) | def _revert_dual_attention(self): method _encode_text_prompt (line 156) | def _encode_text_prompt(self, prompt, device, num_images_per_prompt, d... method _encode_image_prompt (line 247) | def _encode_image_prompt(self, prompt, device, num_images_per_prompt, ... method decode_latents (line 303) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 315) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 332) | def check_inputs(self, prompt, image, height, width, callback_steps): method prepare_latents (line 350) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method set_transformer_params (line 372) | def set_transformer_params(self, mix_ratio: float = 0.5, condition_typ... method __call__ (line 386) | def __call__( FILE: src/diffusers/pipelines/deprecated/versatile_diffusion/pipeline_versatile_diffusion_image_variation.py class VersatileDiffusionImageVariationPipeline (line 35) | class VersatileDiffusionImageVariationPipeline(DiffusionPipeline): method __init__ (line 64) | def __init__( method _encode_prompt (line 83) | def _encode_prompt(self, prompt, device, num_images_per_prompt, do_cla... method decode_latents (line 163) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 175) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 193) | def check_inputs(self, image, height, width, callback_steps): method prepare_latents (line 216) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 239) | def __call__( FILE: src/diffusers/pipelines/deprecated/versatile_diffusion/pipeline_versatile_diffusion_text_to_image.py class VersatileDiffusionTextToImagePipeline (line 34) | class VersatileDiffusionTextToImagePipeline(DiffusionPipeline): method __init__ (line 67) | def __init__( method _swap_unet_attention_blocks (line 91) | def _swap_unet_attention_blocks(self): method remove_unused_weights (line 104) | def remove_unused_weights(self): method _encode_prompt (line 107) | def _encode_prompt(self, prompt, device, num_images_per_prompt, do_cla... method decode_latents (line 220) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 232) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 249) | def check_inputs( method prepare_latents (line 302) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 325) | def __call__( FILE: src/diffusers/pipelines/deprecated/vq_diffusion/pipeline_vq_diffusion.py class LearnedClassifierFreeSamplingEmbeddings (line 30) | class LearnedClassifierFreeSamplingEmbeddings(ModelMixin, ConfigMixin): method __init__ (line 36) | def __init__(self, learnable: bool, hidden_size: Optional[int] = None,... class VQDiffusionPipeline (line 52) | class VQDiffusionPipeline(DiffusionPipeline): method __init__ (line 80) | def __init__( method _encode_prompt (line 100) | def _encode_prompt(self, prompt, num_images_per_prompt, do_classifier_... method __call__ (line 164) | def __call__( method truncate (line 304) | def truncate(self, log_p_x_0: torch.Tensor, truncation_rate: float) ->... FILE: src/diffusers/pipelines/dit/pipeline_dit.py class DiTPipeline (line 31) | class DiTPipeline(DiffusionPipeline): method __init__ (line 49) | def __init__( method get_label_ids (line 67) | def get_label_ids(self, label: Union[str, List[str]]) -> List[int]: method __call__ (line 93) | def __call__( FILE: src/diffusers/pipelines/flux/pipeline_flux.py function calculate_shift (line 67) | def calculate_shift( function retrieve_timesteps (line 81) | def retrieve_timesteps( class FluxPipeline (line 140) | class FluxPipeline(DiffusionPipeline, FluxLoraLoaderMixin): method __init__ (line 171) | def __init__( method _get_t5_prompt_embeds (line 201) | def _get_t5_prompt_embeds( method _get_clip_prompt_embeds (line 247) | def _get_clip_prompt_embeds( method encode_prompt (line 288) | def encode_prompt( method check_inputs (line 372) | def check_inputs( method _prepare_latent_image_ids (line 421) | def _prepare_latent_image_ids(batch_size, height, width, device, dtype): method _pack_latents (line 436) | def _pack_latents(latents, batch_size, num_channels_latents, height, w... method _unpack_latents (line 444) | def _unpack_latents(latents, height, width, vae_scale_factor): method prepare_latents (line 457) | def prepare_latents( method guidance_scale (line 491) | def guidance_scale(self): method joint_attention_kwargs (line 495) | def joint_attention_kwargs(self): method num_timesteps (line 499) | def num_timesteps(self): method interrupt (line 503) | def interrupt(self): method __call__ (line 508) | def __call__( FILE: src/diffusers/pipelines/flux/pipeline_output.py class FluxPipelineOutput (line 11) | class FluxPipelineOutput(BaseOutput): FILE: src/diffusers/pipelines/free_init_utils.py class FreeInitMixin (line 24) | class FreeInitMixin: method enable_free_init (line 27) | def enable_free_init( method disable_free_init (line 66) | def disable_free_init(self): method free_init_enabled (line 71) | def free_init_enabled(self): method _get_free_init_freq_filter (line 74) | def _get_free_init_freq_filter( method _apply_freq_filter (line 118) | def _apply_freq_filter(self, x: torch.Tensor, noise: torch.Tensor, low... method _apply_free_init (line 138) | def _apply_free_init( FILE: src/diffusers/pipelines/free_noise_utils.py class AnimateDiffFreeNoiseMixin (line 32) | class AnimateDiffFreeNoiseMixin: method _enable_free_noise_in_block (line 35) | def _enable_free_noise_in_block(self, block: Union[CrossAttnDownBlockM... method _disable_free_noise_in_block (line 73) | def _disable_free_noise_in_block(self, block: Union[CrossAttnDownBlock... method _prepare_latents_free_noise (line 101) | def _prepare_latents_free_noise( method enable_free_noise (line 175) | def enable_free_noise( method disable_free_noise (line 227) | def disable_free_noise(self) -> None: method free_noise_enabled (line 235) | def free_noise_enabled(self): FILE: src/diffusers/pipelines/hunyuandit/pipeline_hunyuandit.py function map_to_standard_shapes (line 98) | def map_to_standard_shapes(target_width, target_height): function get_resize_crop_region_for_grid (line 106) | def get_resize_crop_region_for_grid(src, tgt_size): function rescale_noise_cfg (line 127) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class HunyuanDiTPipeline (line 141) | class HunyuanDiTPipeline(DiffusionPipeline): method __init__ (line 188) | def __init__( method encode_prompt (line 242) | def encode_prompt( method run_safety_checker (line 408) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 423) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 440) | def check_inputs( method prepare_latents (line 517) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method guidance_scale (line 540) | def guidance_scale(self): method guidance_rescale (line 544) | def guidance_rescale(self): method do_classifier_free_guidance (line 551) | def do_classifier_free_guidance(self): method num_timesteps (line 555) | def num_timesteps(self): method interrupt (line 559) | def interrupt(self): method __call__ (line 564) | def __call__( FILE: src/diffusers/pipelines/i2vgen_xl/pipeline_i2vgen_xl.py class I2VGenXLPipelineOutput (line 75) | class I2VGenXLPipelineOutput(BaseOutput): class I2VGenXLPipeline (line 90) | class I2VGenXLPipeline( method __init__ (line 115) | def __init__( method guidance_scale (line 141) | def guidance_scale(self): method do_classifier_free_guidance (line 148) | def do_classifier_free_guidance(self): method encode_prompt (line 151) | def encode_prompt( method _encode_image (line 320) | def _encode_image(self, image, device, num_videos_per_prompt): method decode_latents (line 352) | def decode_latents(self, latents, decode_chunk_size=None): method prepare_extra_step_kwargs (line 375) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 392) | def check_inputs( method prepare_image_latents (line 441) | def prepare_image_latents( method prepare_latents (line 474) | def prepare_latents( method __call__ (line 501) | def __call__( function _convert_pt_to_pil (line 734) | def _convert_pt_to_pil(image: Union[torch.Tensor, List[torch.Tensor]]): function _resize_bilinear (line 749) | def _resize_bilinear( function _center_crop_wide (line 762) | def _center_crop_wide( FILE: src/diffusers/pipelines/kandinsky/pipeline_kandinsky.py function get_new_h_w (line 66) | def get_new_h_w(h, w, scale_factor=8): class KandinskyPipeline (line 76) | class KandinskyPipeline(DiffusionPipeline): method __init__ (line 98) | def __init__( method prepare_latents (line 118) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method _encode_prompt (line 129) | def _encode_prompt( method __call__ (line 233) | def __call__( FILE: src/diffusers/pipelines/kandinsky/pipeline_kandinsky_combined.py class KandinskyCombinedPipeline (line 113) | class KandinskyCombinedPipeline(DiffusionPipeline): method __init__ (line 148) | def __init__( method enable_xformers_memory_efficient_attention (line 193) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_sequential_cpu_offload (line 196) | def enable_sequential_cpu_offload(self, gpu_id=0): method progress_bar (line 206) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 211) | def set_progress_bar_config(self, **kwargs): method __call__ (line 217) | def __call__( class KandinskyImg2ImgCombinedPipeline (line 331) | class KandinskyImg2ImgCombinedPipeline(DiffusionPipeline): method __init__ (line 366) | def __init__( method enable_xformers_memory_efficient_attention (line 411) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_sequential_cpu_offload (line 414) | def enable_sequential_cpu_offload(self, gpu_id=0): method progress_bar (line 425) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 430) | def set_progress_bar_config(self, **kwargs): method __call__ (line 436) | def __call__( class KandinskyInpaintCombinedPipeline (line 572) | class KandinskyInpaintCombinedPipeline(DiffusionPipeline): method __init__ (line 607) | def __init__( method enable_xformers_memory_efficient_attention (line 652) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_sequential_cpu_offload (line 655) | def enable_sequential_cpu_offload(self, gpu_id=0): method progress_bar (line 666) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 671) | def set_progress_bar_config(self, **kwargs): method __call__ (line 677) | def __call__( FILE: src/diffusers/pipelines/kandinsky/pipeline_kandinsky_img2img.py function get_new_h_w (line 79) | def get_new_h_w(h, w, scale_factor=8): function prepare_image (line 89) | def prepare_image(pil_image, w=512, h=512): class KandinskyImg2ImgPipeline (line 98) | class KandinskyImg2ImgPipeline(DiffusionPipeline): method __init__ (line 120) | def __init__( method get_timesteps (line 139) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 148) | def prepare_latents(self, latents, latent_timestep, shape, dtype, devi... method _encode_prompt (line 164) | def _encode_prompt( method add_noise (line 267) | def add_noise( method __call__ (line 295) | def __call__( FILE: src/diffusers/pipelines/kandinsky/pipeline_kandinsky_inpaint.py function get_new_h_w (line 88) | def get_new_h_w(h, w, scale_factor=8): function prepare_mask (line 98) | def prepare_mask(masks): function prepare_mask_and_masked_image (line 122) | def prepare_mask_and_masked_image(image, mask, height, width): class KandinskyInpaintPipeline (line 240) | class KandinskyInpaintPipeline(DiffusionPipeline): method __init__ (line 262) | def __init__( method prepare_latents (line 283) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method _encode_prompt (line 294) | def _encode_prompt( method __call__ (line 398) | def __call__( FILE: src/diffusers/pipelines/kandinsky/pipeline_kandinsky_prior.py class KandinskyPriorPipelineOutput (line 113) | class KandinskyPriorPipelineOutput(BaseOutput): class KandinskyPriorPipeline (line 128) | class KandinskyPriorPipeline(DiffusionPipeline): method __init__ (line 152) | def __init__( method interpolate (line 174) | def interpolate( method prepare_latents (line 280) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method get_zero_embed (line 291) | def get_zero_embed(self, batch_size=1, device=None): method _encode_prompt (line 300) | def _encode_prompt( method __call__ (line 399) | def __call__( FILE: src/diffusers/pipelines/kandinsky/text_encoder.py class MCLIPConfig (line 5) | class MCLIPConfig(XLMRobertaConfig): method __init__ (line 8) | def __init__(self, transformerDimSize=1024, imageDimSize=768, **kwargs): class MultilingualCLIP (line 14) | class MultilingualCLIP(PreTrainedModel): method __init__ (line 17) | def __init__(self, config, *args, **kwargs): method forward (line 24) | def forward(self, input_ids, attention_mask): FILE: src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2.py function downscale_height_and_width (line 54) | def downscale_height_and_width(height, width, scale_factor=8): class KandinskyV22Pipeline (line 64) | class KandinskyV22Pipeline(DiffusionPipeline): method __init__ (line 83) | def __init__( method prepare_latents (line 99) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method guidance_scale (line 111) | def guidance_scale(self): method do_classifier_free_guidance (line 115) | def do_classifier_free_guidance(self): method num_timesteps (line 119) | def num_timesteps(self): method __call__ (line 124) | def __call__( FILE: src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_combined.py class KandinskyV22CombinedPipeline (line 107) | class KandinskyV22CombinedPipeline(DiffusionPipeline): method __init__ (line 140) | def __init__( method enable_xformers_memory_efficient_attention (line 179) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_sequential_cpu_offload (line 182) | def enable_sequential_cpu_offload(self, gpu_id: Optional[int] = None, ... method progress_bar (line 193) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 198) | def set_progress_bar_config(self, **kwargs): method __call__ (line 204) | def __call__( class KandinskyV22Img2ImgCombinedPipeline (line 335) | class KandinskyV22Img2ImgCombinedPipeline(DiffusionPipeline): method __init__ (line 368) | def __init__( method enable_xformers_memory_efficient_attention (line 407) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_model_cpu_offload (line 410) | def enable_model_cpu_offload(self, gpu_id: Optional[int] = None, devic... method enable_sequential_cpu_offload (line 420) | def enable_sequential_cpu_offload(self, gpu_id: Optional[int] = None, ... method progress_bar (line 431) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 436) | def set_progress_bar_config(self, **kwargs): method __call__ (line 442) | def __call__( class KandinskyV22InpaintCombinedPipeline (line 584) | class KandinskyV22InpaintCombinedPipeline(DiffusionPipeline): method __init__ (line 617) | def __init__( method enable_xformers_memory_efficient_attention (line 656) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_sequential_cpu_offload (line 659) | def enable_sequential_cpu_offload(self, gpu_id: Optional[int] = None, ... method progress_bar (line 670) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 675) | def set_progress_bar_config(self, **kwargs): method __call__ (line 681) | def __call__( FILE: src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_controlnet.py function downscale_height_and_width (line 96) | def downscale_height_and_width(height, width, scale_factor=8): class KandinskyV22ControlnetPipeline (line 106) | class KandinskyV22ControlnetPipeline(DiffusionPipeline): method __init__ (line 124) | def __init__( method prepare_latents (line 140) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method __call__ (line 152) | def __call__( FILE: src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_controlnet_img2img.py function downscale_height_and_width (line 100) | def downscale_height_and_width(height, width, scale_factor=8): function prepare_image (line 111) | def prepare_image(pil_image, w=512, h=512): class KandinskyV22ControlnetImg2ImgPipeline (line 120) | class KandinskyV22ControlnetImg2ImgPipeline(DiffusionPipeline): method __init__ (line 138) | def __init__( method get_timesteps (line 154) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 164) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method __call__ (line 207) | def __call__( FILE: src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_img2img.py function downscale_height_and_width (line 72) | def downscale_height_and_width(height, width, scale_factor=8): function prepare_image (line 83) | def prepare_image(pil_image, w=512, h=512): class KandinskyV22Img2ImgPipeline (line 92) | class KandinskyV22Img2ImgPipeline(DiffusionPipeline): method __init__ (line 111) | def __init__( method get_timesteps (line 127) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 136) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method guidance_scale (line 179) | def guidance_scale(self): method do_classifier_free_guidance (line 183) | def do_classifier_free_guidance(self): method num_timesteps (line 187) | def num_timesteps(self): method __call__ (line 191) | def __call__( FILE: src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_inpainting.py function downscale_height_and_width (line 81) | def downscale_height_and_width(height, width, scale_factor=8): function prepare_mask (line 92) | def prepare_mask(masks): function prepare_mask_and_masked_image (line 117) | def prepare_mask_and_masked_image(image, mask, height, width): class KandinskyV22InpaintPipeline (line 235) | class KandinskyV22InpaintPipeline(DiffusionPipeline): method __init__ (line 254) | def __init__( method prepare_latents (line 271) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method guidance_scale (line 283) | def guidance_scale(self): method do_classifier_free_guidance (line 287) | def do_classifier_free_guidance(self): method num_timesteps (line 291) | def num_timesteps(self): method __call__ (line 295) | def __call__( FILE: src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_prior.py class KandinskyV22PriorPipeline (line 84) | class KandinskyV22PriorPipeline(DiffusionPipeline): method __init__ (line 111) | def __init__( method interpolate (line 133) | def interpolate( method prepare_latents (line 239) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method get_zero_embed (line 251) | def get_zero_embed(self, batch_size=1, device=None): method _encode_prompt (line 261) | def _encode_prompt( method do_classifier_free_guidance (line 359) | def do_classifier_free_guidance(self): method guidance_scale (line 363) | def guidance_scale(self): method num_timesteps (line 367) | def num_timesteps(self): method __call__ (line 372) | def __call__( FILE: src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_prior_emb2emb.py class KandinskyV22PriorEmb2EmbPipeline (line 102) | class KandinskyV22PriorEmb2EmbPipeline(DiffusionPipeline): method __init__ (line 126) | def __init__( method get_timesteps (line 146) | def get_timesteps(self, num_inference_steps, strength, device): method interpolate (line 157) | def interpolate( method _encode_image (line 245) | def _encode_image( method prepare_latents (line 262) | def prepare_latents(self, emb, timestep, batch_size, num_images_per_pr... method get_zero_embed (line 289) | def get_zero_embed(self, batch_size=1, device=None): method _encode_prompt (line 299) | def _encode_prompt( method __call__ (line 398) | def __call__( FILE: src/diffusers/pipelines/kandinsky3/convert_kandinsky3_unet.py function convert_state_dict (line 36) | def convert_state_dict(unet_state_dict): function main (line 75) | def main(model_path, output_path): FILE: src/diffusers/pipelines/kandinsky3/pipeline_kandinsky3.py function downscale_height_and_width (line 40) | def downscale_height_and_width(height, width, scale_factor=8): class Kandinsky3Pipeline (line 50) | class Kandinsky3Pipeline(DiffusionPipeline, StableDiffusionLoraLoaderMix... method __init__ (line 60) | def __init__( method process_embeds (line 74) | def process_embeds(self, embeddings, attention_mask, cut_context): method encode_prompt (line 83) | def encode_prompt( method prepare_latents (line 232) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method check_inputs (line 243) | def check_inputs( method guidance_scale (line 314) | def guidance_scale(self): method do_classifier_free_guidance (line 318) | def do_classifier_free_guidance(self): method num_timesteps (line 322) | def num_timesteps(self): method __call__ (line 327) | def __call__( FILE: src/diffusers/pipelines/kandinsky3/pipeline_kandinsky3_img2img.py function downscale_height_and_width (line 47) | def downscale_height_and_width(height, width, scale_factor=8): function prepare_image (line 57) | def prepare_image(pil_image): class Kandinsky3Img2ImgPipeline (line 65) | class Kandinsky3Img2ImgPipeline(DiffusionPipeline, StableDiffusionLoraLo... method __init__ (line 75) | def __init__( method get_timesteps (line 89) | def get_timesteps(self, num_inference_steps, strength, device): method _process_embeds (line 98) | def _process_embeds(self, embeddings, attention_mask, cut_context): method encode_prompt (line 108) | def encode_prompt( method prepare_latents (line 257) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method prepare_extra_step_kwargs (line 300) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 317) | def check_inputs( method guidance_scale (line 390) | def guidance_scale(self): method do_classifier_free_guidance (line 394) | def do_classifier_free_guidance(self): method num_timesteps (line 398) | def num_timesteps(self): method __call__ (line 403) | def __call__( FILE: src/diffusers/pipelines/kolors/pipeline_kolors.py function retrieve_timesteps (line 65) | def retrieve_timesteps( class KolorsPipeline (line 124) | class KolorsPipeline(DiffusionPipeline, StableDiffusionMixin, StableDiff... method __init__ (line 168) | def __init__( method encode_prompt (line 198) | def encode_prompt( method encode_image (line 363) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 388) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 434) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 451) | def check_inputs( method prepare_latents (line 538) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 561) | def _get_add_time_ids( method upcast_vae (line 580) | def upcast_vae(self): method get_guidance_scale_embedding (line 599) | def get_guidance_scale_embedding( method guidance_scale (line 630) | def guidance_scale(self): method do_classifier_free_guidance (line 637) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 641) | def cross_attention_kwargs(self): method denoising_end (line 645) | def denoising_end(self): method num_timesteps (line 649) | def num_timesteps(self): method interrupt (line 653) | def interrupt(self): method __call__ (line 658) | def __call__( FILE: src/diffusers/pipelines/kolors/pipeline_kolors_img2img.py function retrieve_latents (line 70) | def retrieve_latents( function retrieve_timesteps (line 84) | def retrieve_timesteps( class KolorsImg2ImgPipeline (line 143) | class KolorsImg2ImgPipeline(DiffusionPipeline, StableDiffusionMixin, Sta... method __init__ (line 187) | def __init__( method encode_prompt (line 218) | def encode_prompt( method encode_image (line 383) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 408) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 454) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 471) | def check_inputs( method get_timesteps (line 562) | def get_timesteps(self, num_inference_steps, strength, device, denoisi... method prepare_latents (line 599) | def prepare_latents( method _get_add_time_ids (line 686) | def _get_add_time_ids( method upcast_vae (line 705) | def upcast_vae(self): method get_guidance_scale_embedding (line 724) | def get_guidance_scale_embedding( method guidance_scale (line 755) | def guidance_scale(self): method do_classifier_free_guidance (line 762) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 766) | def cross_attention_kwargs(self): method denoising_start (line 770) | def denoising_start(self): method denoising_end (line 774) | def denoising_end(self): method num_timesteps (line 778) | def num_timesteps(self): method interrupt (line 782) | def interrupt(self): method __call__ (line 787) | def __call__( FILE: src/diffusers/pipelines/kolors/pipeline_output.py class KolorsPipelineOutput (line 11) | class KolorsPipelineOutput(BaseOutput): FILE: src/diffusers/pipelines/kolors/text_encoder.py class ChatGLMConfig (line 32) | class ChatGLMConfig(PretrainedConfig): method __init__ (line 35) | def __init__( class RMSNorm (line 93) | class RMSNorm(torch.nn.Module): method __init__ (line 94) | def __init__(self, normalized_shape, eps=1e-5, device=None, dtype=None... method forward (line 99) | def forward(self, hidden_states: torch.Tensor): function _config_to_kwargs (line 107) | def _config_to_kwargs(args): class CoreAttention (line 114) | class CoreAttention(torch.nn.Module): method __init__ (line 115) | def __init__(self, config: ChatGLMConfig, layer_number): method forward (line 140) | def forward(self, query_layer, key_layer, value_layer, attention_mask): function split_tensor_along_last_dim (line 240) | def split_tensor_along_last_dim( function apply_rotary_pos_emb (line 269) | def apply_rotary_pos_emb(x: torch.Tensor, rope_cache: torch.Tensor) -> t... class SelfAttention (line 289) | class SelfAttention(torch.nn.Module): method __init__ (line 295) | def __init__(self, config: ChatGLMConfig, layer_number, device=None): method _allocate_memory (line 331) | def _allocate_memory(self, inference_max_sequence_len, batch_size, dev... method forward (line 345) | def forward(self, hidden_states, attention_mask, rotary_pos_emb, kv_ca... class MLP (line 434) | class MLP(torch.nn.Module): method __init__ (line 441) | def __init__(self, config: ChatGLMConfig, device=None): method forward (line 466) | def forward(self, hidden_states): class GLMBlock (line 475) | class GLMBlock(torch.nn.Module): method __init__ (line 481) | def __init__(self, config: ChatGLMConfig, layer_number, device=None): method forward (line 507) | def forward( class GLMTransformer (line 551) | class GLMTransformer(torch.nn.Module): method __init__ (line 554) | def __init__(self, config: ChatGLMConfig, device=None): method _get_layer (line 578) | def _get_layer(self, layer_number): method forward (line 581) | def forward( class ChatGLMPreTrainedModel (line 629) | class ChatGLMPreTrainedModel(PreTrainedModel): method _init_weights (line 641) | def _init_weights(self, module: nn.Module): method get_masks (line 645) | def get_masks(self, input_ids, past_key_values, padding_mask=None): method get_position_ids (line 664) | def get_position_ids(self, input_ids, device): method _set_gradient_checkpointing (line 669) | def _set_gradient_checkpointing(self, module, value=False): function default_init (line 674) | def default_init(cls, *args, **kwargs): class Embedding (line 678) | class Embedding(torch.nn.Module): method __init__ (line 681) | def __init__(self, config: ChatGLMConfig, device=None): method forward (line 691) | def forward(self, input_ids): class RotaryEmbedding (line 703) | class RotaryEmbedding(nn.Module): method __init__ (line 704) | def __init__(self, dim, original_impl=False, device=None, dtype=None): method forward_impl (line 711) | def forward_impl(self, seq_len: int, n_elem: int, dtype: torch.dtype, ... method forward (line 734) | def forward(self, max_seq_len, offset=0): class PrefixEncoder (line 738) | class PrefixEncoder(torch.nn.Module): method __init__ (line 744) | def __init__(self, config: ChatGLMConfig): method forward (line 761) | def forward(self, prefix: torch.Tensor): class ChatGLMModel (line 770) | class ChatGLMModel(ChatGLMPreTrainedModel): method __init__ (line 771) | def __init__(self, config: ChatGLMConfig, device=None, empty_init=True): method get_input_embeddings (line 812) | def get_input_embeddings(self): method get_prompt (line 815) | def get_prompt(self, batch_size, device, dtype=torch.half): method forward (line 826) | def forward( FILE: src/diffusers/pipelines/kolors/tokenizer.py class SPTokenizer (line 26) | class SPTokenizer: method __init__ (line 27) | def __init__(self, model_path: str): method tokenize (line 49) | def tokenize(self, s: str, encode_special_tokens=False): method encode (line 64) | def encode(self, s: str, bos: bool = False, eos: bool = False) -> List... method decode (line 73) | def decode(self, t: List[int]) -> str: method decode_tokens (line 87) | def decode_tokens(self, tokens: List[str]) -> str: method convert_token_to_id (line 91) | def convert_token_to_id(self, token): method convert_id_to_token (line 97) | def convert_id_to_token(self, index): class ChatGLMTokenizer (line 106) | class ChatGLMTokenizer(PreTrainedTokenizer): method __init__ (line 111) | def __init__( method get_command (line 136) | def get_command(self, token): method unk_token (line 143) | def unk_token(self) -> str: method unk_token (line 147) | def unk_token(self, value: str): method pad_token (line 151) | def pad_token(self) -> str: method pad_token (line 155) | def pad_token(self, value: str): method pad_token_id (line 159) | def pad_token_id(self): method eos_token (line 163) | def eos_token(self) -> str: method eos_token (line 167) | def eos_token(self, value: str): method eos_token_id (line 171) | def eos_token_id(self): method vocab_size (line 175) | def vocab_size(self): method get_vocab (line 178) | def get_vocab(self): method _tokenize (line 184) | def _tokenize(self, text, **kwargs): method _convert_token_to_id (line 187) | def _convert_token_to_id(self, token): method _convert_id_to_token (line 191) | def _convert_id_to_token(self, index): method convert_tokens_to_string (line 195) | def convert_tokens_to_string(self, tokens: List[str]) -> str: method save_vocabulary (line 198) | def save_vocabulary(self, save_directory, filename_prefix=None): method get_prefix_tokens (line 224) | def get_prefix_tokens(self): method build_single_message (line 228) | def build_single_message(self, role, metadata, message): method build_chat_input (line 235) | def build_chat_input(self, query, history=None, role="user"): method build_inputs_with_special_tokens (line 248) | def build_inputs_with_special_tokens( method _pad (line 273) | def _pad( FILE: src/diffusers/pipelines/latent_consistency_models/pipeline_latent_consistency_img2img.py function retrieve_latents (line 47) | def retrieve_latents( function retrieve_timesteps (line 61) | def retrieve_timesteps( class LatentConsistencyModelImg2ImgPipeline (line 146) | class LatentConsistencyModelImg2ImgPipeline( method __init__ (line 194) | def __init__( method encode_prompt (line 233) | def encode_prompt( method encode_image (line 416) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 441) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 487) | def run_safety_checker(self, image, device, dtype): method prepare_latents (line 502) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method get_guidance_scale_embedding (line 568) | def get_guidance_scale_embedding( method prepare_extra_step_kwargs (line 599) | def prepare_extra_step_kwargs(self, generator, eta): method get_timesteps (line 617) | def get_timesteps(self, num_inference_steps, strength, device): method check_inputs (line 628) | def check_inputs( method guidance_scale (line 682) | def guidance_scale(self): method cross_attention_kwargs (line 686) | def cross_attention_kwargs(self): method clip_skip (line 690) | def clip_skip(self): method do_classifier_free_guidance (line 694) | def do_classifier_free_guidance(self): method num_timesteps (line 698) | def num_timesteps(self): method __call__ (line 703) | def __call__( FILE: src/diffusers/pipelines/latent_consistency_models/pipeline_latent_consistency_text2img.py function retrieve_timesteps (line 65) | def retrieve_timesteps( class LatentConsistencyModelPipeline (line 124) | class LatentConsistencyModelPipeline( method __init__ (line 172) | def __init__( method encode_prompt (line 217) | def encode_prompt( method encode_image (line 400) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 425) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 471) | def run_safety_checker(self, image, device, dtype): method prepare_latents (line 486) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 508) | def get_guidance_scale_embedding( method prepare_extra_step_kwargs (line 539) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 557) | def check_inputs( method guidance_scale (line 612) | def guidance_scale(self): method cross_attention_kwargs (line 616) | def cross_attention_kwargs(self): method clip_skip (line 620) | def clip_skip(self): method do_classifier_free_guidance (line 624) | def do_classifier_free_guidance(self): method num_timesteps (line 628) | def num_timesteps(self): method __call__ (line 633) | def __call__( FILE: src/diffusers/pipelines/latent_diffusion/pipeline_latent_diffusion.py class LDMTextToImagePipeline (line 32) | class LDMTextToImagePipeline(DiffusionPipeline): method __init__ (line 55) | def __init__( method __call__ (line 68) | def __call__( class LDMBertConfig (line 242) | class LDMBertConfig(PretrainedConfig): method __init__ (line 247) | def __init__( function _expand_mask (line 289) | def _expand_mask(mask: torch.Tensor, dtype: torch.dtype, tgt_len: Option... class LDMBertAttention (line 304) | class LDMBertAttention(nn.Module): method __init__ (line 307) | def __init__( method _shape (line 331) | def _shape(self, tensor: torch.Tensor, seq_len: int, bsz: int): method forward (line 334) | def forward( class LDMBertEncoderLayer (line 448) | class LDMBertEncoderLayer(nn.Module): method __init__ (line 449) | def __init__(self, config: LDMBertConfig): method forward (line 466) | def forward( class LDMBertPreTrainedModel (line 518) | class LDMBertPreTrainedModel(PreTrainedModel): method _init_weights (line 524) | def _init_weights(self, module): method _set_gradient_checkpointing (line 535) | def _set_gradient_checkpointing(self, module, value=False): method dummy_inputs (line 540) | def dummy_inputs(self): class LDMBertEncoder (line 550) | class LDMBertEncoder(LDMBertPreTrainedModel): method __init__ (line 560) | def __init__(self, config: LDMBertConfig): method get_input_embeddings (line 578) | def get_input_embeddings(self): method set_input_embeddings (line 581) | def set_input_embeddings(self, value): method forward (line 584) | def forward( class LDMBertModel (line 717) | class LDMBertModel(LDMBertPreTrainedModel): method __init__ (line 720) | def __init__(self, config: LDMBertConfig): method forward (line 725) | def forward( FILE: src/diffusers/pipelines/latent_diffusion/pipeline_latent_diffusion_superresolution.py function preprocess (line 23) | def preprocess(image): class LDMSuperResolutionPipeline (line 33) | class LDMSuperResolutionPipeline(DiffusionPipeline): method __init__ (line 51) | def __init__( method __call__ (line 68) | def __call__( FILE: src/diffusers/pipelines/latte/pipeline_latte.py function retrieve_timesteps (line 71) | def retrieve_timesteps( class LattePipelineOutput (line 131) | class LattePipelineOutput(BaseOutput): class LattePipeline (line 135) | class LattePipeline(DiffusionPipeline): method __init__ (line 169) | def __init__( method mask_text_embeddings (line 187) | def mask_text_embeddings(self, emb, mask): method encode_prompt (line 196) | def encode_prompt( method prepare_extra_step_kwargs (line 346) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 363) | def check_inputs( method _text_preprocessing (line 415) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 440) | def _clean_caption(self, caption): method prepare_latents (line 555) | def prepare_latents( method guidance_scale (line 581) | def guidance_scale(self): method do_classifier_free_guidance (line 588) | def do_classifier_free_guidance(self): method num_timesteps (line 592) | def num_timesteps(self): method interrupt (line 596) | def interrupt(self): method __call__ (line 601) | def __call__( method decode_latents (line 854) | def decode_latents(self, latents: torch.Tensor, video_length: int, dec... FILE: src/diffusers/pipelines/ledits_pp/pipeline_leditspp_stable_diffusion.py class LeditsAttentionStore (line 63) | class LeditsAttentionStore: method get_empty_store (line 65) | def get_empty_store(): method __call__ (line 68) | def __call__(self, attn, is_cross: bool, place_in_unet: str, editing_p... method forward (line 77) | def forward(self, attn, is_cross: bool, place_in_unet: str): method between_steps (line 82) | def between_steps(self, store_step=True): method get_attention (line 100) | def get_attention(self, step: int): method aggregate_attention (line 110) | def aggregate_attention( method __init__ (line 133) | def __init__(self, average: bool, batch_size=1, max_resolution=16, max... class LeditsGaussianSmoothing (line 148) | class LeditsGaussianSmoothing: method __init__ (line 149) | def __init__(self, device): method __call__ (line 169) | def __call__(self, input): class LEDITSCrossAttnProcessor (line 180) | class LEDITSCrossAttnProcessor: method __init__ (line 181) | def __init__(self, attention_store, place_in_unet, pnp, editing_prompts): method __call__ (line 187) | def __call__( function rescale_noise_cfg (line 236) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class LEditsPPPipelineStableDiffusion (line 250) | class LEditsPPPipelineStableDiffusion( method __init__ (line 287) | def __init__( method run_safety_checker (line 389) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 404) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 416) | def prepare_extra_step_kwargs(self, eta, generator=None): method check_inputs (line 434) | def check_inputs( method prepare_latents (line 462) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method prepare_unet (line 474) | def prepare_unet(self, attention_store, PnP: bool = False): method encode_prompt (line 498) | def encode_prompt( method guidance_rescale (line 686) | def guidance_rescale(self): method clip_skip (line 690) | def clip_skip(self): method cross_attention_kwargs (line 694) | def cross_attention_kwargs(self): method __call__ (line 699) | def __call__( method invert (line 1199) | def invert( method encode_image (line 1347) | def encode_image(self, image, dtype=None, height=None, width=None, res... function compute_noise_ddim (line 1367) | def compute_noise_ddim(scheduler, prev_latents, latents, timestep, noise... function compute_noise_sde_dpm_pp_2nd (line 1405) | def compute_noise_sde_dpm_pp_2nd(scheduler, prev_latents, latents, times... function compute_noise (line 1489) | def compute_noise(scheduler, *args): FILE: src/diffusers/pipelines/ledits_pp/pipeline_leditspp_stable_diffusion_xl.py class LeditsAttentionStore (line 108) | class LeditsAttentionStore: method get_empty_store (line 110) | def get_empty_store(): method __call__ (line 113) | def __call__(self, attn, is_cross: bool, place_in_unet: str, editing_p... method forward (line 122) | def forward(self, attn, is_cross: bool, place_in_unet: str): method between_steps (line 127) | def between_steps(self, store_step=True): method get_attention (line 145) | def get_attention(self, step: int): method aggregate_attention (line 155) | def aggregate_attention( method __init__ (line 178) | def __init__(self, average: bool, batch_size=1, max_resolution=16, max... class LeditsGaussianSmoothing (line 193) | class LeditsGaussianSmoothing: method __init__ (line 194) | def __init__(self, device): method __call__ (line 214) | def __call__(self, input): class LEDITSCrossAttnProcessor (line 226) | class LEDITSCrossAttnProcessor: method __init__ (line 227) | def __init__(self, attention_store, place_in_unet, pnp, editing_prompts): method __call__ (line 233) | def __call__( class LEditsPPPipelineStableDiffusionXL (line 281) | class LEditsPPPipelineStableDiffusionXL( method __init__ (line 354) | def __init__( method encode_prompt (line 404) | def encode_prompt( method prepare_extra_step_kwargs (line 625) | def prepare_extra_step_kwargs(self, eta, generator=None): method check_inputs (line 642) | def check_inputs( method prepare_latents (line 666) | def prepare_latents(self, device, latents): method _get_add_time_ids (line 673) | def _get_add_time_ids( method upcast_vae (line 692) | def upcast_vae(self): method get_guidance_scale_embedding (line 710) | def get_guidance_scale_embedding( method guidance_scale (line 741) | def guidance_scale(self): method guidance_rescale (line 745) | def guidance_rescale(self): method clip_skip (line 749) | def clip_skip(self): method do_classifier_free_guidance (line 756) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 760) | def cross_attention_kwargs(self): method denoising_end (line 764) | def denoising_end(self): method num_timesteps (line 768) | def num_timesteps(self): method prepare_unet (line 772) | def prepare_unet(self, attention_store, PnP: bool = False): method __call__ (line 798) | def __call__( method encode_image (line 1400) | def encode_image(self, image, dtype=None, height=None, width=None, res... method invert (line 1429) | def invert( function rescale_noise_cfg (line 1645) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function compute_noise_ddim (line 1660) | def compute_noise_ddim(scheduler, prev_latents, latents, timestep, noise... function compute_noise_sde_dpm_pp_2nd (line 1699) | def compute_noise_sde_dpm_pp_2nd(scheduler, prev_latents, latents, times... function compute_noise (line 1784) | def compute_noise(scheduler, *args): FILE: src/diffusers/pipelines/ledits_pp/pipeline_output.py class LEditsPPDiffusionPipelineOutput (line 11) | class LEditsPPDiffusionPipelineOutput(BaseOutput): class LEditsPPInversionPipelineOutput (line 29) | class LEditsPPInversionPipelineOutput(BaseOutput): FILE: src/diffusers/pipelines/lumina/pipeline_lumina.py function retrieve_timesteps (line 68) | def retrieve_timesteps( class LuminaText2ImgPipeline (line 127) | class LuminaText2ImgPipeline(DiffusionPipeline): method __init__ (line 169) | def __init__( method _get_gemma_prompt_embeds (line 196) | def _get_gemma_prompt_embeds( method encode_prompt (line 252) | def encode_prompt( method prepare_extra_step_kwargs (line 362) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 379) | def check_inputs( method _text_preprocessing (line 438) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 463) | def _clean_caption(self, caption): method prepare_latents (line 577) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method guidance_scale (line 598) | def guidance_scale(self): method do_classifier_free_guidance (line 605) | def do_classifier_free_guidance(self): method num_timesteps (line 609) | def num_timesteps(self): method __call__ (line 614) | def __call__( FILE: src/diffusers/pipelines/marigold/marigold_image_processing.py class MarigoldImageProcessor (line 19) | class MarigoldImageProcessor(ConfigMixin): method __init__ (line 23) | def __init__( method expand_tensor_or_array (line 32) | def expand_tensor_or_array(images: Union[torch.Tensor, np.ndarray]) ->... method pt_to_numpy (line 51) | def pt_to_numpy(images: torch.Tensor) -> np.ndarray: method numpy_to_pt (line 59) | def numpy_to_pt(images: np.ndarray) -> torch.Tensor: method resize_antialias (line 74) | def resize_antialias( method resize_to_max_edge (line 90) | def resize_to_max_edge(image: torch.Tensor, max_edge_sz: int, mode: st... method pad_image (line 111) | def pad_image(image: torch.Tensor, align: int) -> Tuple[torch.Tensor, ... method unpad_image (line 127) | def unpad_image(image: torch.Tensor, padding: Tuple[int, int]) -> torc... method load_image_canonical (line 144) | def load_image_canonical( method check_image_values_range (line 190) | def check_image_values_range(image: torch.Tensor) -> None: method preprocess (line 198) | def preprocess( method colormap (line 240) | def colormap( method visualize_depth (line 370) | def visualize_depth( method export_depth_to_16bit_png (line 435) | def export_depth_to_16bit_png( method visualize_normals (line 471) | def visualize_normals( method visualize_uncertainty (line 532) | def visualize_uncertainty( FILE: src/diffusers/pipelines/marigold/pipeline_marigold_depth.py class MarigoldDepthOutput (line 75) | class MarigoldDepthOutput(BaseOutput): class MarigoldDepthPipeline (line 96) | class MarigoldDepthPipeline(DiffusionPipeline): method __init__ (line 141) | def __init__( method check_inputs (line 188) | def check_inputs( method progress_bar (line 315) | def progress_bar(self, iterable=None, total=None, desc=None, leave=True): method __call__ (line 335) | def __call__( method prepare_latents (line 608) | def prepare_latents( method decode_prediction (line 645) | def decode_prediction(self, pred_latent: torch.Tensor) -> torch.Tensor: method ensemble_depth (line 660) | def ensemble_depth( FILE: src/diffusers/pipelines/marigold/pipeline_marigold_normals.py class MarigoldNormalsOutput (line 70) | class MarigoldNormalsOutput(BaseOutput): class MarigoldNormalsPipeline (line 91) | class MarigoldNormalsPipeline(DiffusionPipeline): method __init__ (line 131) | def __init__( method check_inputs (line 174) | def check_inputs( method progress_bar (line 299) | def progress_bar(self, iterable=None, total=None, desc=None, leave=True): method __call__ (line 319) | def __call__( method prepare_latents (line 581) | def prepare_latents( method decode_prediction (line 618) | def decode_prediction(self, pred_latent: torch.Tensor) -> torch.Tensor: method normalize_normals (line 637) | def normalize_normals(normals: torch.Tensor, eps: float = 1e-6) -> tor... method ensemble_normals (line 647) | def ensemble_normals( FILE: src/diffusers/pipelines/musicldm/pipeline_musicldm.py class MusicLDMPipeline (line 67) | class MusicLDMPipeline(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 93) | def __init__( method _encode_prompt (line 116) | def _encode_prompt( method mel_spectrogram_to_waveform (line 250) | def mel_spectrogram_to_waveform(self, mel_spectrogram): method score_waveforms (line 260) | def score_waveforms(self, text, audio, num_waveforms_per_prompt, devic... method prepare_extra_step_kwargs (line 285) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 303) | def check_inputs( method prepare_latents (line 362) | def prepare_latents(self, batch_size, num_channels_latents, height, dt... method enable_model_cpu_offload (line 384) | def enable_model_cpu_offload(self, gpu_id=0): method __call__ (line 420) | def __call__( FILE: src/diffusers/pipelines/onnx_utils.py class OnnxRuntimeModel (line 52) | class OnnxRuntimeModel: method __init__ (line 53) | def __init__(self, model=None, **kwargs): method __call__ (line 59) | def __call__(self, **kwargs): method load_model (line 64) | def load_model(path: Union[str, Path], provider=None, sess_options=None): method _save_pretrained (line 80) | def _save_pretrained(self, save_directory: Union[str, Path], file_name... method save_pretrained (line 111) | def save_pretrained( method _from_pretrained (line 135) | def _from_pretrained( method from_pretrained (line 196) | def from_pretrained( FILE: src/diffusers/pipelines/pag/pag_utils.py class PAGMixin (line 33) | class PAGMixin: method _set_pag_attn_processor (line 36) | def _set_pag_attn_processor(self, pag_applied_layers, do_classifier_fr... method _get_pag_scale (line 88) | def _get_pag_scale(self, t): method _apply_perturbed_attention_guidance (line 101) | def _apply_perturbed_attention_guidance(self, noise_pred, do_classifie... method _prepare_perturbed_attention_guidance (line 127) | def _prepare_perturbed_attention_guidance(self, cond, uncond, do_class... method set_pag_applied_layers (line 146) | def set_pag_applied_layers( method pag_scale (line 191) | def pag_scale(self) -> float: method pag_adaptive_scale (line 196) | def pag_adaptive_scale(self) -> float: method do_pag_adaptive_scaling (line 201) | def do_pag_adaptive_scaling(self) -> bool: method do_perturbed_attention_guidance (line 206) | def do_perturbed_attention_guidance(self) -> bool: method pag_attn_processors (line 211) | def pag_attn_processors(self) -> Dict[str, AttentionProcessor]: FILE: src/diffusers/pipelines/pag/pipeline_pag_controlnet_sd.py function retrieve_timesteps (line 99) | def retrieve_timesteps( class StableDiffusionControlNetPAGPipeline (line 158) | class StableDiffusionControlNetPAGPipeline( method __init__ (line 209) | def __init__( method encode_prompt (line 265) | def encode_prompt( method encode_image (line 448) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 473) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 519) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 534) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 551) | def check_inputs( method check_image (line 708) | def check_image(self, image, prompt, prompt_embeds): method prepare_image (line 746) | def prepare_image( method prepare_latents (line 777) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 800) | def get_guidance_scale_embedding( method guidance_scale (line 831) | def guidance_scale(self): method clip_skip (line 835) | def clip_skip(self): method do_classifier_free_guidance (line 842) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 846) | def cross_attention_kwargs(self): method num_timesteps (line 850) | def num_timesteps(self): method __call__ (line 855) | def __call__( FILE: src/diffusers/pipelines/pag/pipeline_pag_controlnet_sd_xl.py function retrieve_timesteps (line 121) | def retrieve_timesteps( class StableDiffusionXLControlNetPAGPipeline (line 180) | class StableDiffusionXLControlNetPAGPipeline( method __init__ (line 252) | def __init__( method encode_prompt (line 301) | def encode_prompt( method encode_image (line 536) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 561) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 607) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 625) | def check_inputs( method check_image (line 815) | def check_image(self, image, prompt, prompt_embeds): method prepare_image (line 853) | def prepare_image( method prepare_latents (line 884) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 907) | def _get_add_time_ids( method upcast_vae (line 926) | def upcast_vae(self): method get_guidance_scale_embedding (line 944) | def get_guidance_scale_embedding( method guidance_scale (line 975) | def guidance_scale(self): method clip_skip (line 979) | def clip_skip(self): method do_classifier_free_guidance (line 986) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 990) | def cross_attention_kwargs(self): method denoising_end (line 994) | def denoising_end(self): method num_timesteps (line 998) | def num_timesteps(self): method __call__ (line 1003) | def __call__( FILE: src/diffusers/pipelines/pag/pipeline_pag_hunyuandit.py function map_to_standard_shapes (line 101) | def map_to_standard_shapes(target_width, target_height): function get_resize_crop_region_for_grid (line 109) | def get_resize_crop_region_for_grid(src, tgt_size): function rescale_noise_cfg (line 130) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class HunyuanDiTPAGPipeline (line 144) | class HunyuanDiTPAGPipeline(DiffusionPipeline, PAGMixin): method __init__ (line 192) | def __init__( method encode_prompt (line 252) | def encode_prompt( method run_safety_checker (line 418) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 433) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 451) | def check_inputs( method prepare_latents (line 528) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method guidance_scale (line 551) | def guidance_scale(self): method guidance_rescale (line 555) | def guidance_rescale(self): method do_classifier_free_guidance (line 562) | def do_classifier_free_guidance(self): method num_timesteps (line 566) | def num_timesteps(self): method interrupt (line 570) | def interrupt(self): method __call__ (line 575) | def __call__( FILE: src/diffusers/pipelines/pag/pipeline_pag_kolors.py function retrieve_timesteps (line 70) | def retrieve_timesteps( class KolorsPAGPipeline (line 129) | class KolorsPAGPipeline( method __init__ (line 181) | def __init__( method encode_prompt (line 215) | def encode_prompt( method encode_image (line 380) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 405) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 451) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 469) | def check_inputs( method prepare_latents (line 556) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 579) | def _get_add_time_ids( method upcast_vae (line 598) | def upcast_vae(self): method get_guidance_scale_embedding (line 617) | def get_guidance_scale_embedding( method guidance_scale (line 648) | def guidance_scale(self): method do_classifier_free_guidance (line 655) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 659) | def cross_attention_kwargs(self): method denoising_end (line 663) | def denoising_end(self): method num_timesteps (line 667) | def num_timesteps(self): method interrupt (line 671) | def interrupt(self): method __call__ (line 676) | def __call__( FILE: src/diffusers/pipelines/pag/pipeline_pag_pixart_sigma.py function retrieve_timesteps (line 76) | def retrieve_timesteps( class PixArtSigmaPAGPipeline (line 135) | class PixArtSigmaPAGPipeline(DiffusionPipeline, PAGMixin): method __init__ (line 160) | def __init__( method encode_prompt (line 181) | def encode_prompt( method prepare_extra_step_kwargs (line 324) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 342) | def check_inputs( method _text_preprocessing (line 410) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 435) | def _clean_caption(self, caption): method prepare_latents (line 550) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 574) | def __call__( FILE: src/diffusers/pipelines/pag/pipeline_pag_sd.py function rescale_noise_cfg (line 62) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 77) | def retrieve_timesteps( class StableDiffusionPAGPipeline (line 136) | class StableDiffusionPAGPipeline( method __init__ (line 183) | def __init__( method encode_prompt (line 279) | def encode_prompt( method encode_image (line 462) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 487) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 533) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 548) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 566) | def check_inputs( method prepare_latents (line 636) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 659) | def get_guidance_scale_embedding( method guidance_scale (line 690) | def guidance_scale(self): method guidance_rescale (line 694) | def guidance_rescale(self): method clip_skip (line 698) | def clip_skip(self): method do_classifier_free_guidance (line 705) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 709) | def cross_attention_kwargs(self): method num_timesteps (line 713) | def num_timesteps(self): method interrupt (line 717) | def interrupt(self): method __call__ (line 722) | def __call__( FILE: src/diffusers/pipelines/pag/pipeline_pag_sd_3.py function retrieve_timesteps (line 77) | def retrieve_timesteps( class StableDiffusion3PAGPipeline (line 136) | class StableDiffusion3PAGPipeline(DiffusionPipeline, SD3LoraLoaderMixin,... method __init__ (line 177) | def __init__( method _get_t5_prompt_embeds (line 221) | def _get_t5_prompt_embeds( method _get_clip_prompt_embeds (line 278) | def _get_clip_prompt_embeds( method encode_prompt (line 334) | def encode_prompt( method check_inputs (line 528) | def check_inputs( method prepare_latents (line 619) | def prepare_latents( method guidance_scale (line 651) | def guidance_scale(self): method clip_skip (line 655) | def clip_skip(self): method do_classifier_free_guidance (line 662) | def do_classifier_free_guidance(self): method joint_attention_kwargs (line 666) | def joint_attention_kwargs(self): method num_timesteps (line 670) | def num_timesteps(self): method interrupt (line 674) | def interrupt(self): method __call__ (line 679) | def __call__( FILE: src/diffusers/pipelines/pag/pipeline_pag_sd_animatediff.py class AnimateDiffPAGPipeline (line 80) | class AnimateDiffPAGPipeline( method __init__ (line 124) | def __init__( method encode_prompt (line 156) | def encode_prompt( method encode_image (line 339) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 364) | def prepare_ip_adapter_image_embeds( method decode_latents (line 410) | def decode_latents(self, latents, decode_chunk_size: int = 16): method prepare_extra_step_kwargs (line 429) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 447) | def check_inputs( method prepare_latents (line 511) | def prepare_latents( method guidance_scale (line 544) | def guidance_scale(self): method clip_skip (line 548) | def clip_skip(self): method do_classifier_free_guidance (line 555) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 559) | def cross_attention_kwargs(self): method num_timesteps (line 563) | def num_timesteps(self): method __call__ (line 568) | def __call__( FILE: src/diffusers/pipelines/pag/pipeline_pag_sd_xl.py function rescale_noise_cfg (line 90) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 105) | def retrieve_timesteps( class StableDiffusionXLPAGPipeline (line 164) | class StableDiffusionXLPAGPipeline( method __init__ (line 237) | def __init__( method encode_prompt (line 281) | def encode_prompt( method encode_image (line 516) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 541) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 587) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 605) | def check_inputs( method prepare_latents (line 702) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 725) | def _get_add_time_ids( method upcast_vae (line 744) | def upcast_vae(self): method get_guidance_scale_embedding (line 763) | def get_guidance_scale_embedding( method guidance_scale (line 794) | def guidance_scale(self): method guidance_rescale (line 798) | def guidance_rescale(self): method clip_skip (line 802) | def clip_skip(self): method do_classifier_free_guidance (line 809) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 813) | def cross_attention_kwargs(self): method denoising_end (line 817) | def denoising_end(self): method num_timesteps (line 821) | def num_timesteps(self): method interrupt (line 825) | def interrupt(self): method __call__ (line 830) | def __call__( FILE: src/diffusers/pipelines/pag/pipeline_pag_sd_xl_img2img.py function rescale_noise_cfg (line 94) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_latents (line 109) | def retrieve_latents( function retrieve_timesteps (line 123) | def retrieve_timesteps( class StableDiffusionXLPAGImg2ImgPipeline (line 182) | class StableDiffusionXLPAGImg2ImgPipeline( method __init__ (line 258) | def __init__( method encode_prompt (line 302) | def encode_prompt( method prepare_extra_step_kwargs (line 537) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 555) | def check_inputs( method get_timesteps (line 646) | def get_timesteps(self, num_inference_steps, strength, device, denoisi... method prepare_latents (line 683) | def prepare_latents( method encode_image (line 770) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 795) | def prepare_ip_adapter_image_embeds( method _get_add_time_ids (line 841) | def _get_add_time_ids( method upcast_vae (line 893) | def upcast_vae(self): method get_guidance_scale_embedding (line 911) | def get_guidance_scale_embedding( method guidance_scale (line 942) | def guidance_scale(self): method guidance_rescale (line 946) | def guidance_rescale(self): method clip_skip (line 950) | def clip_skip(self): method do_classifier_free_guidance (line 957) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 961) | def cross_attention_kwargs(self): method denoising_end (line 965) | def denoising_end(self): method denoising_start (line 969) | def denoising_start(self): method num_timesteps (line 973) | def num_timesteps(self): method interrupt (line 977) | def interrupt(self): method __call__ (line 982) | def __call__( FILE: src/diffusers/pipelines/pag/pipeline_pag_sd_xl_inpaint.py function rescale_noise_cfg (line 107) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_latents (line 122) | def retrieve_latents( function retrieve_timesteps (line 136) | def retrieve_timesteps( class StableDiffusionXLPAGInpaintPipeline (line 195) | class StableDiffusionXLPAGInpaintPipeline( method __init__ (line 274) | def __init__( method encode_image (line 321) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 346) | def prepare_ip_adapter_image_embeds( method encode_prompt (line 392) | def encode_prompt( method prepare_extra_step_kwargs (line 627) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 645) | def check_inputs( method prepare_latents (line 750) | def prepare_latents( method _encode_vae_image (line 817) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method prepare_mask_latents (line 841) | def prepare_mask_latents( method get_timesteps (line 895) | def get_timesteps(self, num_inference_steps, strength, device, denoisi... method _get_add_time_ids (line 932) | def _get_add_time_ids( method upcast_vae (line 984) | def upcast_vae(self): method get_guidance_scale_embedding (line 1002) | def get_guidance_scale_embedding( method guidance_scale (line 1033) | def guidance_scale(self): method guidance_rescale (line 1037) | def guidance_rescale(self): method clip_skip (line 1041) | def clip_skip(self): method do_classifier_free_guidance (line 1048) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 1052) | def cross_attention_kwargs(self): method denoising_end (line 1056) | def denoising_end(self): method denoising_start (line 1060) | def denoising_start(self): method num_timesteps (line 1064) | def num_timesteps(self): method interrupt (line 1068) | def interrupt(self): method __call__ (line 1073) | def __call__( FILE: src/diffusers/pipelines/paint_by_example/image_encoder.py class PaintByExampleImageEncoder (line 25) | class PaintByExampleImageEncoder(CLIPPreTrainedModel): method __init__ (line 26) | def __init__(self, config, proj_size=None): method forward (line 38) | def forward(self, pixel_values, return_uncond_vector=False): class PaintByExampleMapper (line 50) | class PaintByExampleMapper(nn.Module): method __init__ (line 51) | def __init__(self, config): method forward (line 63) | def forward(self, hidden_states): FILE: src/diffusers/pipelines/paint_by_example/pipeline_paint_by_example.py function retrieve_latents (line 38) | def retrieve_latents( function prepare_mask_and_masked_image (line 51) | def prepare_mask_and_masked_image(image, mask): class PaintByExamplePipeline (line 151) | class PaintByExamplePipeline(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 192) | def __init__( method run_safety_checker (line 217) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 232) | def prepare_extra_step_kwargs(self, generator, eta): method decode_latents (line 250) | def decode_latents(self, latents): method check_inputs (line 262) | def check_inputs(self, image, height, width, callback_steps): method prepare_latents (line 285) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method prepare_mask_latents (line 308) | def prepare_mask_latents( method _encode_vae_image (line 354) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method _encode_image (line 368) | def _encode_image(self, image, device, num_images_per_prompt, do_class... method __call__ (line 394) | def __call__( FILE: src/diffusers/pipelines/pia/pipeline_pia.py function prepare_mask_coef_by_statistics (line 92) | def prepare_mask_coef_by_statistics(num_frames: int, cond_frame: int, mo... class PIAPipelineOutput (line 111) | class PIAPipelineOutput(BaseOutput): class PIAPipeline (line 125) | class PIAPipeline( method __init__ (line 166) | def __init__( method encode_prompt (line 202) | def encode_prompt( method encode_image (line 385) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method decode_latents (line 410) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 423) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 440) | def check_inputs( method prepare_ip_adapter_image_embeds (line 504) | def prepare_ip_adapter_image_embeds( method prepare_latents (line 550) | def prepare_latents( method prepare_masked_condition (line 575) | def prepare_masked_condition( method get_timesteps (line 627) | def get_timesteps(self, num_inference_steps, strength, device): method guidance_scale (line 639) | def guidance_scale(self): method clip_skip (line 643) | def clip_skip(self): method do_classifier_free_guidance (line 650) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 654) | def cross_attention_kwargs(self): method num_timesteps (line 658) | def num_timesteps(self): method __call__ (line 663) | def __call__( FILE: src/diffusers/pipelines/pipeline_flax_utils.py function import_flax_or_no_model (line 74) | def import_flax_or_no_model(module, class_name): class FlaxImagePipelineOutput (line 88) | class FlaxImagePipelineOutput(BaseOutput): class FlaxDiffusionPipeline (line 101) | class FlaxDiffusionPipeline(ConfigMixin, PushToHubMixin): method register_modules (line 118) | def register_modules(self, **kwargs): method save_pretrained (line 151) | def save_pretrained( method from_pretrained (line 231) | def from_pretrained(cls, pretrained_model_name_or_path: Optional[Union... method _get_signature_keys (line 538) | def _get_signature_keys(cls, obj): method components (line 547) | def components(self) -> Dict[str, Any]: method numpy_to_pil (line 584) | def numpy_to_pil(images): method progress_bar (line 600) | def progress_bar(self, iterable): method set_progress_bar_config (line 610) | def set_progress_bar_config(self, **kwargs): FILE: src/diffusers/pipelines/pipeline_loading_utils.py function is_safetensors_compatible (line 92) | def is_safetensors_compatible(filenames, variant=None, passed_components... function variant_compatible_siblings (line 140) | def variant_compatible_siblings(filenames, variant=None) -> Union[List[o... function warn_deprecated_model_variant (line 208) | def warn_deprecated_model_variant(pretrained_model_name_or_path, token, ... function _unwrap_model (line 230) | def _unwrap_model(model): function maybe_raise_or_warn (line 244) | def maybe_raise_or_warn( function get_class_obj_and_candidates (line 275) | def get_class_obj_and_candidates( function _get_custom_pipeline_class (line 302) | def _get_custom_pipeline_class( function _get_pipeline_class (line 335) | def _get_pipeline_class( function _load_empty_model (line 386) | def _load_empty_model( function _assign_components_to_devices (line 470) | def _assign_components_to_devices( function _get_final_device_map (line 500) | def _get_final_device_map(device_map, pipeline_class, passed_class_obj, ... function load_sub_model (line 586) | def load_sub_model( function _fetch_class_library_tuple (line 724) | def _fetch_class_library_tuple(module): FILE: src/diffusers/pipelines/pipeline_utils.py class ImagePipelineOutput (line 101) | class ImagePipelineOutput(BaseOutput): class AudioPipelineOutput (line 115) | class AudioPipelineOutput(BaseOutput): class DiffusionPipeline (line 127) | class DiffusionPipeline(ConfigMixin, PushToHubMixin): method register_modules (line 153) | def register_modules(self, **kwargs): method __setattr__ (line 168) | def __setattr__(self, name: str, value: Any): method save_pretrained (line 183) | def save_pretrained( method to (line 303) | def to(self, *args, **kwargs): method device (line 449) | def device(self) -> torch.device: method dtype (line 464) | def dtype(self) -> torch.dtype: method from_pretrained (line 480) | def from_pretrained(cls, pretrained_model_name_or_path: Optional[Union... method name_or_path (line 975) | def name_or_path(self) -> str: method _execution_device (line 979) | def _execution_device(self): method remove_all_hooks (line 1000) | def remove_all_hooks(self): method enable_model_cpu_offload (line 1009) | def enable_model_cpu_offload(self, gpu_id: Optional[int] = None, devic... method maybe_free_model_hooks (line 1087) | def maybe_free_model_hooks(self): method enable_sequential_cpu_offload (line 1101) | def enable_sequential_cpu_offload(self, gpu_id: Optional[int] = None, ... method reset_device_map (line 1162) | def reset_device_map(self): method download (line 1177) | def download(cls, pretrained_model_name, **kwargs) -> Union[str, os.Pa... method _get_signature_keys (line 1543) | def _get_signature_keys(cls, obj): method _get_signature_types (line 1558) | def _get_signature_types(cls): method components (line 1570) | def components(self) -> Dict[str, Any]: method numpy_to_pil (line 1606) | def numpy_to_pil(images): method progress_bar (line 1612) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 1627) | def set_progress_bar_config(self, **kwargs): method enable_xformers_memory_efficient_attention (line 1630) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method disable_xformers_memory_efficient_attention (line 1665) | def disable_xformers_memory_efficient_attention(self): method set_use_memory_efficient_attention_xformers (line 1671) | def set_use_memory_efficient_attention_xformers( method enable_attention_slicing (line 1691) | def enable_attention_slicing(self, slice_size: Optional[Union[str, int... method disable_attention_slicing (line 1731) | def disable_attention_slicing(self): method set_attention_slice (line 1739) | def set_attention_slice(self, slice_size: Optional[int]): method from_pipe (line 1748) | def from_pipe(cls, pipeline, **kwargs): class StableDiffusionMixin (line 1871) | class StableDiffusionMixin: method enable_vae_slicing (line 1876) | def enable_vae_slicing(self): method disable_vae_slicing (line 1883) | def disable_vae_slicing(self): method enable_vae_tiling (line 1890) | def enable_vae_tiling(self): method disable_vae_tiling (line 1898) | def disable_vae_tiling(self): method enable_freeu (line 1905) | def enable_freeu(self, s1: float, s2: float, b1: float, b2: float): method disable_freeu (line 1927) | def disable_freeu(self): method fuse_qkv_projections (line 1931) | def fuse_qkv_projections(self, unet: bool = True, vae: bool = True): method unfuse_qkv_projections (line 1962) | def unfuse_qkv_projections(self, unet: bool = True, vae: bool = True): FILE: src/diffusers/pipelines/pixart_alpha/pipeline_pixart_alpha.py function retrieve_timesteps (line 173) | def retrieve_timesteps( class PixArtAlphaPipeline (line 232) | class PixArtAlphaPipeline(DiffusionPipeline): method __init__ (line 274) | def __init__( method encode_prompt (line 292) | def encode_prompt( method prepare_extra_step_kwargs (line 435) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 452) | def check_inputs( method _text_preprocessing (line 520) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 545) | def _clean_caption(self, caption): method prepare_latents (line 660) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 684) | def __call__( FILE: src/diffusers/pipelines/pixart_alpha/pipeline_pixart_sigma.py function retrieve_timesteps (line 117) | def retrieve_timesteps( class PixArtSigmaPipeline (line 176) | class PixArtSigmaPipeline(DiffusionPipeline): method __init__ (line 200) | def __init__( method encode_prompt (line 218) | def encode_prompt( method prepare_extra_step_kwargs (line 361) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 379) | def check_inputs( method _text_preprocessing (line 447) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 472) | def _clean_caption(self, caption): method prepare_latents (line 587) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 611) | def __call__( FILE: src/diffusers/pipelines/semantic_stable_diffusion/pipeline_output.py class SemanticStableDiffusionPipelineOutput (line 11) | class SemanticStableDiffusionPipelineOutput(BaseOutput): FILE: src/diffusers/pipelines/semantic_stable_diffusion/pipeline_semantic_stable_diffusion.py class SemanticStableDiffusionPipeline (line 21) | class SemanticStableDiffusionPipeline(DiffusionPipeline, StableDiffusion... method __init__ (line 52) | def __init__( method run_safety_checker (line 95) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 110) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 122) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 140) | def check_inputs( method prepare_latents (line 193) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 216) | def __call__( FILE: src/diffusers/pipelines/shap_e/camera.py class DifferentiableProjectiveCamera (line 23) | class DifferentiableProjectiveCamera: method __post_init__ (line 38) | def __post_init__(self): method resolution (line 43) | def resolution(self): method fov (line 46) | def fov(self): method get_image_coords (line 49) | def get_image_coords(self) -> torch.Tensor: method camera_rays (line 64) | def camera_rays(self): method get_camera_rays (line 76) | def get_camera_rays(self, coords: torch.Tensor) -> torch.Tensor: method resize_image (line 105) | def resize_image(self, width: int, height: int) -> "DifferentiableProj... function create_pan_cameras (line 122) | def create_pan_cameras(size: int) -> DifferentiableProjectiveCamera: FILE: src/diffusers/pipelines/shap_e/pipeline_shap_e.py class ShapEPipelineOutput (line 67) | class ShapEPipelineOutput(BaseOutput): class ShapEPipeline (line 79) | class ShapEPipeline(DiffusionPipeline): method __init__ (line 103) | def __init__( method prepare_latents (line 122) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method _encode_prompt (line 133) | def _encode_prompt( method __call__ (line 184) | def __call__( FILE: src/diffusers/pipelines/shap_e/pipeline_shap_e_img2img.py class ShapEPipelineOutput (line 68) | class ShapEPipelineOutput(BaseOutput): class ShapEImg2ImgPipeline (line 80) | class ShapEImg2ImgPipeline(DiffusionPipeline): method __init__ (line 104) | def __init__( method prepare_latents (line 123) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method _encode_image (line 134) | def _encode_image( method __call__ (line 166) | def __call__( FILE: src/diffusers/pipelines/shap_e/renderer.py function sample_pmf (line 30) | def sample_pmf(pmf: torch.Tensor, n_samples: int) -> torch.Tensor: function posenc_nerf (line 53) | def posenc_nerf(x: torch.Tensor, min_deg: int = 0, max_deg: int = 15) ->... function encode_position (line 70) | def encode_position(position): function encode_direction (line 74) | def encode_direction(position, direction=None): function _sanitize_name (line 81) | def _sanitize_name(x: str) -> str: function integrate_samples (line 85) | def integrate_samples(volume_range, ts, density, channels): function volume_query_points (line 119) | def volume_query_points(volume, grid_size): function _convert_srgb_to_linear (line 128) | def _convert_srgb_to_linear(u: torch.Tensor): function _create_flat_edge_indices (line 132) | def _create_flat_edge_indices( class VoidNeRFModel (line 214) | class VoidNeRFModel(nn.Module): method __init__ (line 219) | def __init__(self, background, channel_scale=255.0): method forward (line 225) | def forward(self, position): class VolumeRange (line 237) | class VolumeRange: method __post_init__ (line 242) | def __post_init__(self): method partition (line 245) | def partition(self, ts): class BoundingBoxVolume (line 269) | class BoundingBoxVolume(nn.Module): method __init__ (line 274) | def __init__( method intersect (line 300) | def intersect( class StratifiedRaySampler (line 353) | class StratifiedRaySampler(nn.Module): method __init__ (line 358) | def __init__(self, depth_mode: str = "linear"): method sample (line 366) | def sample( class ImportanceRaySampler (line 405) | class ImportanceRaySampler(nn.Module): method __init__ (line 410) | def __init__( method sample (line 433) | def sample(self, t0: torch.Tensor, t1: torch.Tensor, n_samples: int) -... class MeshDecoderOutput (line 467) | class MeshDecoderOutput(BaseOutput): class MeshDecoder (line 485) | class MeshDecoder(nn.Module): method __init__ (line 490) | def __init__(self): method forward (line 498) | def forward(self, field: torch.Tensor, min_point: torch.Tensor, size: ... class MLPNeRFModelOutput (line 602) | class MLPNeRFModelOutput(BaseOutput): class MLPNeRSTFModel (line 609) | class MLPNeRSTFModel(ModelMixin, ConfigMixin): method __init__ (line 611) | def __init__( method map_indices_to_keys (line 648) | def map_indices_to_keys(self, output): method forward (line 662) | def forward(self, *, position, direction, ts, nerf_level="coarse", ren... class ChannelsProj (line 708) | class ChannelsProj(nn.Module): method __init__ (line 709) | def __init__( method forward (line 723) | def forward(self, x: torch.Tensor) -> torch.Tensor: class ShapEParamsProjModel (line 734) | class ShapEParamsProjModel(ModelMixin, ConfigMixin): method __init__ (line 742) | def __init__( method forward (line 772) | def forward(self, x: torch.Tensor): class ShapERenderer (line 784) | class ShapERenderer(ModelMixin, ConfigMixin): method __init__ (line 786) | def __init__( method render_rays (line 826) | def render_rays(self, rays, sampler, n_samples, prev_model_out=None, r... method decode_to_image (line 905) | def decode_to_image( method decode_to_mesh (line 950) | def decode_to_mesh( FILE: src/diffusers/pipelines/stable_audio/modeling_stable_audio.py class StableAudioPositionalEmbedding (line 31) | class StableAudioPositionalEmbedding(nn.Module): method __init__ (line 34) | def __init__(self, dim: int): method forward (line 40) | def forward(self, times: torch.Tensor) -> torch.Tensor: class StableAudioProjectionModelOutput (line 49) | class StableAudioProjectionModelOutput(BaseOutput): class StableAudioNumberConditioner (line 66) | class StableAudioNumberConditioner(nn.Module): method __init__ (line 81) | def __init__( method forward (line 98) | def forward( class StableAudioProjectionModel (line 116) | class StableAudioProjectionModel(ModelMixin, ConfigMixin): method __init__ (line 132) | def __init__(self, text_encoder_dim, conditioning_dim, min_value, max_... method forward (line 140) | def forward( FILE: src/diffusers/pipelines/stable_audio/pipeline_stable_audio.py class StableAudioPipeline (line 74) | class StableAudioPipeline(DiffusionPipeline): method __init__ (line 102) | def __init__( method enable_vae_slicing (line 124) | def enable_vae_slicing(self): method disable_vae_slicing (line 132) | def disable_vae_slicing(self): method encode_prompt (line 139) | def encode_prompt( method encode_duration (line 259) | def encode_duration( method prepare_extra_step_kwargs (line 298) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 315) | def check_inputs( method prepare_latents (line 403) | def prepare_latents( method __call__ (line 476) | def __call__( FILE: src/diffusers/pipelines/stable_cascade/pipeline_stable_cascade.py class StableCascadeDecoderPipeline (line 50) | class StableCascadeDecoderPipeline(DiffusionPipeline): method __init__ (line 84) | def __init__( method prepare_latents (line 103) | def prepare_latents( method encode_prompt (line 124) | def encode_prompt( method check_inputs (line 231) | def check_inputs( method guidance_scale (line 273) | def guidance_scale(self): method do_classifier_free_guidance (line 277) | def do_classifier_free_guidance(self): method num_timesteps (line 281) | def num_timesteps(self): method __call__ (line 286) | def __call__( FILE: src/diffusers/pipelines/stable_cascade/pipeline_stable_cascade_combined.py class StableCascadeCombinedPipeline (line 45) | class StableCascadeCombinedPipeline(DiffusionPipeline): method __init__ (line 76) | def __init__( method enable_xformers_memory_efficient_attention (line 121) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_model_cpu_offload (line 124) | def enable_model_cpu_offload(self, gpu_id: Optional[int] = None, devic... method enable_sequential_cpu_offload (line 134) | def enable_sequential_cpu_offload(self, gpu_id: Optional[int] = None, ... method progress_bar (line 144) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 148) | def set_progress_bar_config(self, **kwargs): method __call__ (line 154) | def __call__( FILE: src/diffusers/pipelines/stable_cascade/pipeline_stable_cascade_prior.py class StableCascadePriorPipelineOutput (line 52) | class StableCascadePriorPipelineOutput(BaseOutput): class StableCascadePriorPipeline (line 72) | class StableCascadePriorPipeline(DiffusionPipeline): method __init__ (line 104) | def __init__( method prepare_latents (line 125) | def prepare_latents( method encode_prompt (line 145) | def encode_prompt( method encode_image (line 252) | def encode_image(self, images, device, dtype, batch_size, num_images_p... method check_inputs (line 266) | def check_inputs( method guidance_scale (line 344) | def guidance_scale(self): method do_classifier_free_guidance (line 348) | def do_classifier_free_guidance(self): method num_timesteps (line 352) | def num_timesteps(self): method get_timestep_ratio_conditioning (line 355) | def get_timestep_ratio_conditioning(self, t, alphas_cumprod): method __call__ (line 367) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion/clip_image_project_model.py class CLIPImageProjection (line 21) | class CLIPImageProjection(ModelMixin, ConfigMixin): method __init__ (line 23) | def __init__(self, hidden_size: int = 768): method forward (line 28) | def forward(self, x): FILE: src/diffusers/pipelines/stable_diffusion/convert_from_ckpt.py function shave_segments (line 69) | def shave_segments(path, n_shave_prefix_segments=1): function renew_resnet_paths (line 79) | def renew_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_vae_resnet_paths (line 101) | def renew_vae_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_attention_paths (line 117) | def renew_attention_paths(old_list, n_shave_prefix_segments=0): function renew_vae_attention_paths (line 138) | def renew_vae_attention_paths(old_list, n_shave_prefix_segments=0): function assign_to_checkpoint (line 168) | def assign_to_checkpoint( function conv_attn_to_linear (line 223) | def conv_attn_to_linear(checkpoint): function create_unet_diffusers_config (line 235) | def create_unet_diffusers_config(original_config, image_size: int, contr... function create_vae_diffusers_config (line 344) | def create_vae_diffusers_config(original_config, image_size: int): function create_diffusers_schedular (line 368) | def create_diffusers_schedular(original_config): function create_ldm_bert_config (line 378) | def create_ldm_bert_config(original_config): function convert_ldm_unet_checkpoint (line 388) | def convert_ldm_unet_checkpoint( function convert_ldm_vae_checkpoint (line 636) | def convert_ldm_vae_checkpoint(checkpoint, config): function convert_ldm_bert_checkpoint (line 743) | def convert_ldm_bert_checkpoint(checkpoint, config): function convert_ldm_clip_checkpoint (line 793) | def convert_ldm_clip_checkpoint(checkpoint, local_files_only=False, text... function convert_paint_by_example_checkpoint (line 857) | def convert_paint_by_example_checkpoint(checkpoint, local_files_only=Fal... function convert_open_clip_checkpoint (line 924) | def convert_open_clip_checkpoint( function stable_unclip_image_encoder (line 1006) | def stable_unclip_image_encoder(original_config, local_files_only=False): function stable_unclip_image_noising_components (line 1043) | def stable_unclip_image_noising_components( function convert_controlnet_checkpoint (line 1088) | def convert_controlnet_checkpoint( function download_from_original_stable_diffusion_ckpt (line 1138) | def download_from_original_stable_diffusion_ckpt( function download_controlnet_from_original_ckpt (line 1817) | def download_controlnet_from_original_ckpt( FILE: src/diffusers/pipelines/stable_diffusion/pipeline_flax_stable_diffusion.py class FlaxStableDiffusionPipeline (line 81) | class FlaxStableDiffusionPipeline(FlaxDiffusionPipeline): method __init__ (line 109) | def __init__( method prepare_inputs (line 167) | def prepare_inputs(self, prompt: Union[str, List[str]]): method _get_has_nsfw_concepts (line 180) | def _get_has_nsfw_concepts(self, features, params): method _run_safety_checker (line 184) | def _run_safety_checker(self, images, safety_model_params, jit=False): method _generate (line 214) | def _generate( method __call__ (line 311) | def __call__( function _p_generate (line 439) | def _p_generate( function _p_get_has_nsfw_concepts (line 465) | def _p_get_has_nsfw_concepts(pipe, features, params): function unshard (line 469) | def unshard(x: jnp.ndarray): FILE: src/diffusers/pipelines/stable_diffusion/pipeline_flax_stable_diffusion_img2img.py class FlaxStableDiffusionImg2ImgPipeline (line 105) | class FlaxStableDiffusionImg2ImgPipeline(FlaxDiffusionPipeline): method __init__ (line 133) | def __init__( method prepare_inputs (line 170) | def prepare_inputs(self, prompt: Union[str, List[str]], image: Union[I... method _get_has_nsfw_concepts (line 191) | def _get_has_nsfw_concepts(self, features, params): method _run_safety_checker (line 195) | def _run_safety_checker(self, images, safety_model_params, jit=False): method get_timestep_start (line 225) | def get_timestep_start(self, num_inference_steps, strength): method _generate (line 233) | def _generate( method __call__ (line 338) | def __call__( function _p_generate (line 485) | def _p_generate( function _p_get_has_nsfw_concepts (line 515) | def _p_get_has_nsfw_concepts(pipe, features, params): function unshard (line 519) | def unshard(x: jnp.ndarray): function preprocess (line 526) | def preprocess(image, dtype): FILE: src/diffusers/pipelines/stable_diffusion/pipeline_flax_stable_diffusion_inpaint.py class FlaxStableDiffusionInpaintPipeline (line 102) | class FlaxStableDiffusionInpaintPipeline(FlaxDiffusionPipeline): method __init__ (line 136) | def __init__( method prepare_inputs (line 194) | def prepare_inputs( method _get_has_nsfw_concepts (line 233) | def _get_has_nsfw_concepts(self, features, params): method _run_safety_checker (line 237) | def _run_safety_checker(self, images, safety_model_params, jit=False): method _generate (line 267) | def _generate( method __call__ (line 395) | def __call__( function _p_generate (line 532) | def _p_generate( function _p_get_has_nsfw_concepts (line 562) | def _p_get_has_nsfw_concepts(pipe, features, params): function unshard (line 566) | def unshard(x: jnp.ndarray): function preprocess_image (line 573) | def preprocess_image(image, dtype): function preprocess_mask (line 582) | def preprocess_mask(mask, dtype): FILE: src/diffusers/pipelines/stable_diffusion/pipeline_onnx_stable_diffusion.py class OnnxStableDiffusionPipeline (line 33) | class OnnxStableDiffusionPipeline(DiffusionPipeline): method __init__ (line 46) | def __init__( method _encode_prompt (line 115) | def _encode_prompt( method check_inputs (line 218) | def check_inputs( method __call__ (line 265) | def __call__( class StableDiffusionOnnxPipeline (line 464) | class StableDiffusionOnnxPipeline(OnnxStableDiffusionPipeline): method __init__ (line 465) | def __init__( FILE: src/diffusers/pipelines/stable_diffusion/pipeline_onnx_stable_diffusion_img2img.py function preprocess (line 35) | def preprocess(image): class OnnxStableDiffusionImg2ImgPipeline (line 58) | class OnnxStableDiffusionImg2ImgPipeline(DiffusionPipeline): method __init__ (line 98) | def __init__( method _encode_prompt (line 168) | def _encode_prompt( method check_inputs (line 271) | def check_inputs( method __call__ (line 313) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion/pipeline_onnx_stable_diffusion_inpaint.py function prepare_mask_and_masked_image (line 38) | def prepare_mask_and_masked_image(image, mask, latents_shape): class OnnxStableDiffusionInpaintPipeline (line 56) | class OnnxStableDiffusionInpaintPipeline(DiffusionPipeline): method __init__ (line 96) | def __init__( method _encode_prompt (line 167) | def _encode_prompt( method check_inputs (line 271) | def check_inputs( method __call__ (line 319) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion/pipeline_onnx_stable_diffusion_upscale.py function preprocess (line 34) | def preprocess(image): class OnnxStableDiffusionUpscalePipeline (line 56) | class OnnxStableDiffusionUpscalePipeline(DiffusionPipeline): method __init__ (line 69) | def __init__( method check_inputs (line 145) | def check_inputs( method prepare_latents (line 230) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method decode_latents (line 239) | def decode_latents(self, latents): method _encode_prompt (line 246) | def _encode_prompt( method __call__ (line 349) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion/pipeline_output.py class StableDiffusionPipelineOutput (line 11) | class StableDiffusionPipelineOutput(BaseOutput): class FlaxStableDiffusionPipelineOutput (line 32) | class FlaxStableDiffusionPipelineOutput(BaseOutput): FILE: src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion.py function rescale_noise_cfg (line 59) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 73) | def retrieve_timesteps( class StableDiffusionPipeline (line 132) | class StableDiffusionPipeline( method __init__ (line 178) | def __init__( method _encode_prompt (line 270) | def _encode_prompt( method encode_prompt (line 302) | def encode_prompt( method encode_image (line 484) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 508) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 553) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 567) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 578) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 595) | def check_inputs( method prepare_latents (line 664) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 687) | def get_guidance_scale_embedding( method guidance_scale (line 718) | def guidance_scale(self): method guidance_rescale (line 722) | def guidance_rescale(self): method clip_skip (line 726) | def clip_skip(self): method do_classifier_free_guidance (line 733) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 737) | def cross_attention_kwargs(self): method num_timesteps (line 741) | def num_timesteps(self): method interrupt (line 745) | def interrupt(self): method __call__ (line 750) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_depth2img.py function retrieve_latents (line 40) | def retrieve_latents( function preprocess (line 54) | def preprocess(image): class StableDiffusionDepth2ImgPipeline (line 77) | class StableDiffusionDepth2ImgPipeline(DiffusionPipeline, TextualInversi... method __init__ (line 106) | def __init__( method _encode_prompt (line 152) | def _encode_prompt( method encode_prompt (line 185) | def encode_prompt( method run_safety_checker (line 368) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 383) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 395) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 412) | def check_inputs( method get_timesteps (line 464) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 476) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method prepare_depth_map (line 541) | def prepare_depth_map(self, image, depth_map, batch_size, do_classifie... method guidance_scale (line 593) | def guidance_scale(self): method clip_skip (line 597) | def clip_skip(self): method do_classifier_free_guidance (line 604) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 608) | def cross_attention_kwargs(self): method num_timesteps (line 612) | def num_timesteps(self): method __call__ (line 616) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_image_variation.py class StableDiffusionImageVariationPipeline (line 37) | class StableDiffusionImageVariationPipeline(DiffusionPipeline, StableDif... method __init__ (line 72) | def __init__( method _encode_image (line 133) | def _encode_image(self, image, device, num_images_per_prompt, do_class... method run_safety_checker (line 159) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 174) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 186) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 203) | def check_inputs(self, image, height, width, callback_steps): method prepare_latents (line 226) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 249) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_img2img.py function retrieve_latents (line 77) | def retrieve_latents( function preprocess (line 90) | def preprocess(image): function retrieve_timesteps (line 114) | def retrieve_timesteps( class StableDiffusionImg2ImgPipeline (line 173) | class StableDiffusionImg2ImgPipeline( method __init__ (line 219) | def __init__( method _encode_prompt (line 312) | def _encode_prompt( method encode_prompt (line 345) | def encode_prompt( method encode_image (line 528) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 553) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 599) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 614) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 626) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 643) | def check_inputs( method get_timesteps (line 711) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 722) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method get_guidance_scale_embedding (line 788) | def get_guidance_scale_embedding( method guidance_scale (line 819) | def guidance_scale(self): method clip_skip (line 823) | def clip_skip(self): method do_classifier_free_guidance (line 830) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 834) | def cross_attention_kwargs(self): method num_timesteps (line 838) | def num_timesteps(self): method interrupt (line 842) | def interrupt(self): method __call__ (line 847) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_inpaint.py function retrieve_latents (line 41) | def retrieve_latents( function retrieve_timesteps (line 55) | def retrieve_timesteps( class StableDiffusionInpaintPipeline (line 114) | class StableDiffusionInpaintPipeline( method __init__ (line 160) | def __init__( method _encode_prompt (line 261) | def _encode_prompt( method encode_prompt (line 294) | def encode_prompt( method encode_image (line 477) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 502) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 548) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 563) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 580) | def check_inputs( method prepare_latents (line 670) | def prepare_latents( method _encode_vae_image (line 733) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method prepare_mask_latents (line 747) | def prepare_mask_latents( method get_timesteps (line 793) | def get_timesteps(self, num_inference_steps, strength, device): method get_guidance_scale_embedding (line 805) | def get_guidance_scale_embedding( method guidance_scale (line 836) | def guidance_scale(self): method clip_skip (line 840) | def clip_skip(self): method do_classifier_free_guidance (line 847) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 851) | def cross_attention_kwargs(self): method num_timesteps (line 855) | def num_timesteps(self): method interrupt (line 859) | def interrupt(self): method __call__ (line 863) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_instruct_pix2pix.py function preprocess (line 39) | def preprocess(image): function retrieve_latents (line 63) | def retrieve_latents( class StableDiffusionInstructPix2PixPipeline (line 76) | class StableDiffusionInstructPix2PixPipeline( method __init__ (line 120) | def __init__( method __call__ (line 165) | def __call__( method _encode_prompt (line 482) | def _encode_prompt( method encode_image (line 632) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 656) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 716) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 731) | def prepare_extra_step_kwargs(self, generator, eta): method decode_latents (line 749) | def decode_latents(self, latents): method check_inputs (line 760) | def check_inputs( method prepare_latents (line 826) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method prepare_image_latents (line 848) | def prepare_image_latents( method guidance_scale (line 890) | def guidance_scale(self): method image_guidance_scale (line 894) | def image_guidance_scale(self): method num_timesteps (line 898) | def num_timesteps(self): method do_classifier_free_guidance (line 905) | def do_classifier_free_guidance(self): FILE: src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_latent_upscale.py function preprocess (line 37) | def preprocess(image): class StableDiffusionLatentUpscalePipeline (line 63) | class StableDiffusionLatentUpscalePipeline(DiffusionPipeline, StableDiff... method __init__ (line 88) | def __init__( method _encode_prompt (line 108) | def _encode_prompt(self, prompt, device, do_classifier_free_guidance, ... method decode_latents (line 199) | def decode_latents(self, latents): method check_inputs (line 210) | def check_inputs(self, prompt, image, callback_steps): method prepare_latents (line 248) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 262) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_upscale.py function preprocess (line 42) | def preprocess(image): class StableDiffusionUpscalePipeline (line 68) | class StableDiffusionUpscalePipeline( method __init__ (line 108) | def __init__( method run_safety_checker (line 156) | def run_safety_checker(self, image, device, dtype): method _encode_prompt (line 174) | def _encode_prompt( method encode_prompt (line 207) | def encode_prompt( method prepare_extra_step_kwargs (line 390) | def prepare_extra_step_kwargs(self, generator, eta): method decode_latents (line 408) | def decode_latents(self, latents): method check_inputs (line 419) | def check_inputs( method prepare_latents (line 504) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method upcast_vae (line 517) | def upcast_vae(self): method __call__ (line 535) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion/pipeline_stable_unclip.py class StableUnCLIPPipeline (line 61) | class StableUnCLIPPipeline( method __init__ (line 123) | def __init__( method _encode_prior_prompt (line 161) | def _encode_prior_prompt( method _encode_prompt (line 255) | def _encode_prompt( method encode_prompt (line 288) | def encode_prompt( method decode_latents (line 471) | def decode_latents(self, latents): method prepare_prior_extra_step_kwargs (line 483) | def prepare_prior_extra_step_kwargs(self, generator, eta): method prepare_extra_step_kwargs (line 501) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 518) | def check_inputs( method prepare_latents (line 579) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method noise_image_embeddings (line 590) | def noise_image_embeddings( method __call__ (line 638) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion/pipeline_stable_unclip_img2img.py class StableUnCLIPImg2ImgPipeline (line 72) | class StableUnCLIPImg2ImgPipeline( method __init__ (line 128) | def __init__( method _encode_prompt (line 162) | def _encode_prompt( method _encode_image (line 194) | def _encode_image( method encode_prompt (line 250) | def encode_prompt( method decode_latents (line 433) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 445) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 462) | def check_inputs( method prepare_latents (line 546) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method noise_image_embeddings (line 569) | def noise_image_embeddings( method __call__ (line 617) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion/safety_checker.py function cosine_distance (line 26) | def cosine_distance(image_embeds, text_embeds): class StableDiffusionSafetyChecker (line 32) | class StableDiffusionSafetyChecker(PreTrainedModel): method __init__ (line 38) | def __init__(self, config: CLIPConfig): method forward (line 51) | def forward(self, clip_input, images): method forward_onnx (line 103) | def forward_onnx(self, clip_input: torch.Tensor, images: torch.Tensor): FILE: src/diffusers/pipelines/stable_diffusion/safety_checker_flax.py function jax_cosine_distance (line 25) | def jax_cosine_distance(emb_1, emb_2, eps=1e-12): class FlaxStableDiffusionSafetyCheckerModule (line 31) | class FlaxStableDiffusionSafetyCheckerModule(nn.Module): method setup (line 35) | def setup(self): method __call__ (line 47) | def __call__(self, clip_input): class FlaxStableDiffusionSafetyChecker (line 71) | class FlaxStableDiffusionSafetyChecker(FlaxPreTrainedModel): method __init__ (line 76) | def __init__( method init_weights (line 90) | def init_weights(self, rng: jax.Array, input_shape: Tuple, params: Fro... method __call__ (line 101) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion/stable_unclip_image_normalizer.py class StableUnCLIPImageNormalizer (line 24) | class StableUnCLIPImageNormalizer(ModelMixin, ConfigMixin): method __init__ (line 33) | def __init__( method to (line 42) | def to( method scale (line 51) | def scale(self, embeds): method unscale (line 55) | def unscale(self, embeds): FILE: src/diffusers/pipelines/stable_diffusion_3/pipeline_output.py class StableDiffusion3PipelineOutput (line 11) | class StableDiffusion3PipelineOutput(BaseOutput): FILE: src/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3.py function retrieve_timesteps (line 72) | def retrieve_timesteps( class StableDiffusion3Pipeline (line 131) | class StableDiffusion3Pipeline(DiffusionPipeline, SD3LoraLoaderMixin, Fr... method __init__ (line 169) | def __init__( method _get_t5_prompt_embeds (line 207) | def _get_t5_prompt_embeds( method _get_clip_prompt_embeds (line 263) | def _get_clip_prompt_embeds( method encode_prompt (line 318) | def encode_prompt( method check_inputs (line 511) | def check_inputs( method prepare_latents (line 601) | def prepare_latents( method guidance_scale (line 633) | def guidance_scale(self): method clip_skip (line 637) | def clip_skip(self): method do_classifier_free_guidance (line 644) | def do_classifier_free_guidance(self): method joint_attention_kwargs (line 648) | def joint_attention_kwargs(self): method num_timesteps (line 652) | def num_timesteps(self): method interrupt (line 656) | def interrupt(self): method __call__ (line 661) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3_img2img.py function retrieve_latents (line 79) | def retrieve_latents( function retrieve_timesteps (line 93) | def retrieve_timesteps( class StableDiffusion3Img2ImgPipeline (line 152) | class StableDiffusion3Img2ImgPipeline(DiffusionPipeline): method __init__ (line 190) | def __init__( method _get_t5_prompt_embeds (line 223) | def _get_t5_prompt_embeds( method _get_clip_prompt_embeds (line 280) | def _get_clip_prompt_embeds( method encode_prompt (line 336) | def encode_prompt( method check_inputs (line 529) | def check_inputs( method get_timesteps (line 618) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 629) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method guidance_scale (line 680) | def guidance_scale(self): method clip_skip (line 684) | def clip_skip(self): method do_classifier_free_guidance (line 691) | def do_classifier_free_guidance(self): method num_timesteps (line 695) | def num_timesteps(self): method interrupt (line 699) | def interrupt(self): method __call__ (line 704) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3_inpaint.py function retrieve_latents (line 78) | def retrieve_latents( function retrieve_timesteps (line 92) | def retrieve_timesteps( class StableDiffusion3InpaintPipeline (line 151) | class StableDiffusion3InpaintPipeline(DiffusionPipeline): method __init__ (line 189) | def __init__( method _get_t5_prompt_embeds (line 229) | def _get_t5_prompt_embeds( method _get_clip_prompt_embeds (line 286) | def _get_clip_prompt_embeds( method encode_prompt (line 342) | def encode_prompt( method check_inputs (line 536) | def check_inputs( method get_timesteps (line 626) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 637) | def prepare_latents( method _encode_vae_image (line 698) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method prepare_mask_latents (line 712) | def prepare_mask_latents( method guidance_scale (line 772) | def guidance_scale(self): method clip_skip (line 776) | def clip_skip(self): method do_classifier_free_guidance (line 783) | def do_classifier_free_guidance(self): method num_timesteps (line 787) | def num_timesteps(self): method interrupt (line 791) | def interrupt(self): method __call__ (line 796) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_attend_and_excite/pipeline_stable_diffusion_attend_and_excite.py class AttentionStore (line 82) | class AttentionStore: method get_empty_store (line 84) | def get_empty_store(): method __call__ (line 87) | def __call__(self, attn, is_cross: bool, place_in_unet: str): method between_steps (line 97) | def between_steps(self): method get_average_attention (line 101) | def get_average_attention(self): method aggregate_attention (line 105) | def aggregate_attention(self, from_where: List[str]) -> torch.Tensor: method reset (line 117) | def reset(self): method __init__ (line 122) | def __init__(self, attn_res): class AttendExciteAttnProcessor (line 135) | class AttendExciteAttnProcessor: method __init__ (line 136) | def __init__(self, attnstore, place_in_unet): method __call__ (line 141) | def __call__(self, attn: Attention, hidden_states, encoder_hidden_stat... class StableDiffusionAttendAndExcitePipeline (line 173) | class StableDiffusionAttendAndExcitePipeline(DiffusionPipeline, StableDi... method __init__ (line 207) | def __init__( method _encode_prompt (line 250) | def _encode_prompt( method encode_prompt (line 283) | def encode_prompt( method run_safety_checker (line 466) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 481) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 493) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 510) | def check_inputs( method prepare_latents (line 584) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _compute_max_attention_per_index (line 607) | def _compute_max_attention_per_index( method _aggregate_and_get_max_attention_per_token (line 629) | def _aggregate_and_get_max_attention_per_token( method _compute_loss (line 644) | def _compute_loss(max_attention_per_index: List[torch.Tensor]) -> torc... method _update_latent (line 651) | def _update_latent(latents: torch.Tensor, loss: torch.Tensor, step_siz... method _perform_iterative_refinement_step (line 657) | def _perform_iterative_refinement_step( method register_attention_control (line 711) | def register_attention_control(self): method get_indices (line 730) | def get_indices(self, prompt: str) -> Dict[str, int]: method __call__ (line 738) | def __call__( class GaussianSmoothing (line 1035) | class GaussianSmoothing(torch.nn.Module): method __init__ (line 1048) | def __init__( method forward (line 1089) | def forward(self, input): FILE: src/diffusers/pipelines/stable_diffusion_diffedit/pipeline_stable_diffusion_diffedit.py class DiffEditInversionPipelineOutput (line 51) | class DiffEditInversionPipelineOutput(BaseOutput): function auto_corr_loss (line 139) | def auto_corr_loss(hidden_states, generator=None): function kl_divergence (line 155) | def kl_divergence(hidden_states): function preprocess (line 160) | def preprocess(image): function preprocess_mask (line 183) | def preprocess_mask(mask, batch_size: int = 1): class StableDiffusionDiffEditPipeline (line 236) | class StableDiffusionDiffEditPipeline( method __init__ (line 281) | def __init__( method _encode_prompt (line 375) | def _encode_prompt( method encode_prompt (line 408) | def encode_prompt( method run_safety_checker (line 591) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 606) | def prepare_extra_step_kwargs(self, generator, eta): method decode_latents (line 624) | def decode_latents(self, latents): method check_inputs (line 635) | def check_inputs( method check_source_inputs (line 683) | def check_source_inputs( method get_timesteps (line 718) | def get_timesteps(self, num_inference_steps, strength, device): method get_inverse_timesteps (line 727) | def get_inverse_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 741) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method prepare_image_latents (line 763) | def prepare_image_latents(self, image, batch_size, dtype, device, gene... method get_epsilon (line 812) | def get_epsilon(self, model_output: torch.Tensor, sample: torch.Tensor... method generate_mask (line 831) | def generate_mask( method invert (line 1050) | def invert( method __call__ (line 1288) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_gligen/pipeline_stable_diffusion_gligen.py class StableDiffusionGLIGENPipeline (line 102) | class StableDiffusionGLIGENPipeline(DiffusionPipeline, StableDiffusionMi... method __init__ (line 133) | def __init__( method _encode_prompt (line 176) | def _encode_prompt( method encode_prompt (line 209) | def encode_prompt( method run_safety_checker (line 392) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 406) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 423) | def check_inputs( method prepare_latents (line 479) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method enable_fuser (line 501) | def enable_fuser(self, enabled=True): method draw_inpaint_mask_from_boxes (line 506) | def draw_inpaint_mask_from_boxes(self, boxes, size): method crop (line 514) | def crop(self, im, new_width, new_height): method target_size_center_crop (line 522) | def target_size_center_crop(self, im, new_hw): method __call__ (line 530) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_gligen/pipeline_stable_diffusion_gligen_text_image.py class StableDiffusionGLIGENTextImagePipeline (line 148) | class StableDiffusionGLIGENTextImagePipeline(DiffusionPipeline, StableDi... method __init__ (line 185) | def __init__( method encode_prompt (line 234) | def encode_prompt( method run_safety_checker (line 417) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 432) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 450) | def check_inputs( method prepare_latents (line 503) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method enable_fuser (line 525) | def enable_fuser(self, enabled=True): method draw_inpaint_mask_from_boxes (line 530) | def draw_inpaint_mask_from_boxes(self, boxes, size): method crop (line 542) | def crop(self, im, new_width, new_height): method target_size_center_crop (line 553) | def target_size_center_crop(self, im, new_hw): method complete_mask (line 562) | def complete_mask(self, has_mask, max_objs, device): method get_clip_feature (line 578) | def get_clip_feature(self, input, normalize_constant, device, is_image... method get_cross_attention_kwargs_with_grounded (line 602) | def get_cross_attention_kwargs_with_grounded( method get_cross_attention_kwargs_without_grounded (line 666) | def get_cross_attention_kwargs_without_grounded(self, hidden_size, rep... method __call__ (line 691) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_k_diffusion/pipeline_stable_diffusion_k_diffusion.py class ModelWrapper (line 36) | class ModelWrapper: method __init__ (line 37) | def __init__(self, model, alphas_cumprod): method apply_model (line 41) | def apply_model(self, *args, **kwargs): class StableDiffusionKDiffusionPipeline (line 50) | class StableDiffusionKDiffusionPipeline( method __init__ (line 95) | def __init__( method set_scheduler (line 136) | def set_scheduler(self, scheduler_type: str): method _encode_prompt (line 150) | def _encode_prompt( method encode_prompt (line 183) | def encode_prompt( method run_safety_checker (line 366) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 381) | def decode_latents(self, latents): method check_inputs (line 392) | def check_inputs( method prepare_latents (line 444) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 462) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_k_diffusion/pipeline_stable_diffusion_xl_k_diffusion.py class ModelWrapper (line 76) | class ModelWrapper: method __init__ (line 77) | def __init__(self, model, alphas_cumprod): method apply_model (line 81) | def apply_model(self, *args, **kwargs): class StableDiffusionXLKDiffusionPipeline (line 90) | class StableDiffusionXLKDiffusionPipeline( method __init__ (line 148) | def __init__( method set_scheduler (line 185) | def set_scheduler(self, scheduler_type: str): method encode_prompt (line 199) | def encode_prompt( method check_inputs (line 433) | def check_inputs( method prepare_latents (line 497) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 517) | def _get_add_time_ids( method upcast_vae (line 536) | def upcast_vae(self): method guidance_scale (line 555) | def guidance_scale(self): method clip_skip (line 559) | def clip_skip(self): method do_classifier_free_guidance (line 566) | def do_classifier_free_guidance(self): method __call__ (line 571) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_ldm3d/pipeline_stable_diffusion_ldm3d.py function rescale_noise_cfg (line 63) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 78) | def retrieve_timesteps( class LDM3DPipelineOutput (line 138) | class LDM3DPipelineOutput(BaseOutput): class StableDiffusionLDM3DPipeline (line 159) | class StableDiffusionLDM3DPipeline( method __init__ (line 205) | def __init__( method _encode_prompt (line 250) | def _encode_prompt( method encode_prompt (line 283) | def encode_prompt( method encode_image (line 466) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 491) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 536) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 552) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 570) | def check_inputs( method prepare_latents (line 639) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 662) | def get_guidance_scale_embedding( method guidance_scale (line 693) | def guidance_scale(self): method guidance_rescale (line 697) | def guidance_rescale(self): method clip_skip (line 701) | def clip_skip(self): method do_classifier_free_guidance (line 708) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 712) | def cross_attention_kwargs(self): method num_timesteps (line 716) | def num_timesteps(self): method interrupt (line 720) | def interrupt(self): method __call__ (line 725) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_panorama/pipeline_stable_diffusion_panorama.py function rescale_noise_cfg (line 63) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 78) | def retrieve_timesteps( class StableDiffusionPanoramaPipeline (line 137) | class StableDiffusionPanoramaPipeline( method __init__ (line 181) | def __init__( method _encode_prompt (line 226) | def _encode_prompt( method encode_prompt (line 259) | def encode_prompt( method encode_image (line 442) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 467) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 513) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 528) | def decode_latents(self, latents): method decode_latents_with_padding (line 539) | def decode_latents_with_padding(self, latents: torch.Tensor, padding: ... method prepare_extra_step_kwargs (line 566) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 584) | def check_inputs( method prepare_latents (line 654) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 677) | def get_guidance_scale_embedding( method get_views (line 707) | def get_views( method guidance_scale (line 750) | def guidance_scale(self): method guidance_rescale (line 754) | def guidance_rescale(self): method cross_attention_kwargs (line 758) | def cross_attention_kwargs(self): method clip_skip (line 762) | def clip_skip(self): method do_classifier_free_guidance (line 766) | def do_classifier_free_guidance(self): method num_timesteps (line 770) | def num_timesteps(self): method interrupt (line 774) | def interrupt(self): method __call__ (line 779) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_safe/__init__.py class SafetyConfig (line 21) | class SafetyConfig(object): FILE: src/diffusers/pipelines/stable_diffusion_safe/pipeline_output.py class StableDiffusionSafePipelineOutput (line 13) | class StableDiffusionSafePipelineOutput(BaseOutput): FILE: src/diffusers/pipelines/stable_diffusion_safe/pipeline_stable_diffusion_safe.py class StableDiffusionPipelineSafe (line 25) | class StableDiffusionPipelineSafe(DiffusionPipeline, StableDiffusionMixi... method __init__ (line 58) | def __init__( method safety_concept (line 156) | def safety_concept(self): method safety_concept (line 166) | def safety_concept(self, concept): method _encode_prompt (line 176) | def _encode_prompt( method run_safety_checker (line 311) | def run_safety_checker(self, image, device, dtype, enable_safety_guida... method decode_latents (line 335) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 347) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 365) | def check_inputs( method prepare_latents (line 418) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method perform_safety_guidance (line 440) | def perform_safety_guidance( method encode_image (line 483) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method __call__ (line 508) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_safe/safety_checker.py function cosine_distance (line 25) | def cosine_distance(image_embeds, text_embeds): class SafeStableDiffusionSafetyChecker (line 31) | class SafeStableDiffusionSafetyChecker(PreTrainedModel): method __init__ (line 36) | def __init__(self, config: CLIPConfig): method forward (line 49) | def forward(self, clip_input, images): method forward_onnx (line 88) | def forward_onnx(self, clip_input: torch.Tensor, images: torch.Tensor): FILE: src/diffusers/pipelines/stable_diffusion_sag/pipeline_stable_diffusion_sag.py class CrossAttnStoreProcessor (line 61) | class CrossAttnStoreProcessor: method __init__ (line 62) | def __init__(self): method __call__ (line 65) | def __call__( class StableDiffusionSAGPipeline (line 101) | class StableDiffusionSAGPipeline(DiffusionPipeline, StableDiffusionMixin... method __init__ (line 136) | def __init__( method _encode_prompt (line 165) | def _encode_prompt( method encode_prompt (line 198) | def encode_prompt( method encode_image (line 381) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 405) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 440) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 455) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 467) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 485) | def check_inputs( method prepare_latents (line 538) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 562) | def __call__( method sag_masking (line 866) | def sag_masking(self, original_latents, attn_map, map_size, t, eps): method pred_x0 (line 896) | def pred_x0(self, sample, model_output, timestep): method pred_epsilon (line 916) | def pred_epsilon(self, sample, model_output, timestep): function gaussian_blur_2d (line 936) | def gaussian_blur_2d(img, kernel_size, sigma): FILE: src/diffusers/pipelines/stable_diffusion_xl/pipeline_flax_stable_diffusion_xl.py class FlaxStableDiffusionXLPipeline (line 42) | class FlaxStableDiffusionXLPipeline(FlaxDiffusionPipeline): method __init__ (line 43) | def __init__( method prepare_inputs (line 70) | def prepare_inputs(self, prompt: Union[str, List[str]]): method __call__ (line 88) | def __call__( method get_embeddings (line 147) | def get_embeddings(self, prompt_ids: jnp.array, params): method _get_add_time_ids (line 164) | def _get_add_time_ids(self, original_size, crops_coords_top_left, targ... method _generate (line 169) | def _generate( function _p_generate (line 284) | def _p_generate( FILE: src/diffusers/pipelines/stable_diffusion_xl/pipeline_output.py class StableDiffusionXLPipelineOutput (line 11) | class StableDiffusionXLPipelineOutput(BaseOutput): class FlaxStableDiffusionXLPipelineOutput (line 28) | class FlaxStableDiffusionXLPipelineOutput(BaseOutput): FILE: src/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl.py function rescale_noise_cfg (line 89) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 104) | def retrieve_timesteps( class StableDiffusionXLPipeline (line 163) | class StableDiffusionXLPipeline( method __init__ (line 235) | def __init__( method encode_prompt (line 275) | def encode_prompt( method encode_image (line 510) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 535) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 581) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 598) | def check_inputs( method prepare_latents (line 695) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 717) | def _get_add_time_ids( method upcast_vae (line 735) | def upcast_vae(self): method get_guidance_scale_embedding (line 754) | def get_guidance_scale_embedding( method guidance_scale (line 785) | def guidance_scale(self): method guidance_rescale (line 789) | def guidance_rescale(self): method clip_skip (line 793) | def clip_skip(self): method do_classifier_free_guidance (line 800) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 804) | def cross_attention_kwargs(self): method denoising_end (line 808) | def denoising_end(self): method num_timesteps (line 812) | def num_timesteps(self): method interrupt (line 816) | def interrupt(self): method __call__ (line 821) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl_img2img.py function rescale_noise_cfg (line 92) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_latents (line 107) | def retrieve_latents( function retrieve_timesteps (line 121) | def retrieve_timesteps( class StableDiffusionXLImg2ImgPipeline (line 180) | class StableDiffusionXLImg2ImgPipeline( method __init__ (line 255) | def __init__( method encode_prompt (line 296) | def encode_prompt( method prepare_extra_step_kwargs (line 531) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 548) | def check_inputs( method get_timesteps (line 638) | def get_timesteps(self, num_inference_steps, strength, device, denoisi... method prepare_latents (line 674) | def prepare_latents( method encode_image (line 761) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 786) | def prepare_ip_adapter_image_embeds( method _get_add_time_ids (line 831) | def _get_add_time_ids( method upcast_vae (line 883) | def upcast_vae(self): method get_guidance_scale_embedding (line 901) | def get_guidance_scale_embedding( method guidance_scale (line 932) | def guidance_scale(self): method guidance_rescale (line 936) | def guidance_rescale(self): method clip_skip (line 940) | def clip_skip(self): method do_classifier_free_guidance (line 947) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 951) | def cross_attention_kwargs(self): method denoising_end (line 955) | def denoising_end(self): method denoising_start (line 959) | def denoising_start(self): method num_timesteps (line 963) | def num_timesteps(self): method interrupt (line 967) | def interrupt(self): method __call__ (line 972) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl_inpaint.py function rescale_noise_cfg (line 103) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function mask_pil_to_torch (line 117) | def mask_pil_to_torch(mask, height, width): function retrieve_latents (line 134) | def retrieve_latents( function retrieve_timesteps (line 148) | def retrieve_timesteps( class StableDiffusionXLInpaintPipeline (line 207) | class StableDiffusionXLInpaintPipeline( method __init__ (line 285) | def __init__( method encode_image (line 329) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 354) | def prepare_ip_adapter_image_embeds( method encode_prompt (line 400) | def encode_prompt( method prepare_extra_step_kwargs (line 635) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 652) | def check_inputs( method prepare_latents (line 756) | def prepare_latents( method _encode_vae_image (line 822) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method prepare_mask_latents (line 845) | def prepare_mask_latents( method get_timesteps (line 899) | def get_timesteps(self, num_inference_steps, strength, device, denoisi... method _get_add_time_ids (line 936) | def _get_add_time_ids( method upcast_vae (line 988) | def upcast_vae(self): method get_guidance_scale_embedding (line 1006) | def get_guidance_scale_embedding( method guidance_scale (line 1037) | def guidance_scale(self): method guidance_rescale (line 1041) | def guidance_rescale(self): method clip_skip (line 1045) | def clip_skip(self): method do_classifier_free_guidance (line 1052) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 1056) | def cross_attention_kwargs(self): method denoising_end (line 1060) | def denoising_end(self): method denoising_start (line 1064) | def denoising_start(self): method num_timesteps (line 1068) | def num_timesteps(self): method interrupt (line 1072) | def interrupt(self): method __call__ (line 1077) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl_instruct_pix2pix.py function retrieve_latents (line 91) | def retrieve_latents( function rescale_noise_cfg (line 104) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class StableDiffusionXLInstructPix2PixPipeline (line 118) | class StableDiffusionXLInstructPix2PixPipeline( method __init__ (line 177) | def __init__( method encode_prompt (line 214) | def encode_prompt( method prepare_extra_step_kwargs (line 420) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 437) | def check_inputs( method prepare_latents (line 486) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method prepare_image_latents (line 508) | def prepare_image_latents( method _get_add_time_ids (line 566) | def _get_add_time_ids( method upcast_vae (line 585) | def upcast_vae(self): method __call__ (line 605) | def __call__( FILE: src/diffusers/pipelines/stable_diffusion_xl/watermark.py class StableDiffusionXLWatermarker (line 17) | class StableDiffusionXLWatermarker: method __init__ (line 18) | def __init__(self): method apply_watermark (line 24) | def apply_watermark(self, images: torch.Tensor): FILE: src/diffusers/pipelines/stable_video_diffusion/pipeline_stable_video_diffusion.py function _append_dims (line 57) | def _append_dims(x, target_dims): function retrieve_timesteps (line 66) | def retrieve_timesteps( class StableVideoDiffusionPipelineOutput (line 126) | class StableVideoDiffusionPipelineOutput(BaseOutput): class StableVideoDiffusionPipeline (line 139) | class StableVideoDiffusionPipeline(DiffusionPipeline): method __init__ (line 163) | def __init__( method _encode_image (line 183) | def _encode_image( method _encode_vae_image (line 231) | def _encode_vae_image( method _get_add_time_ids (line 254) | def _get_add_time_ids( method decode_latents (line 282) | def decode_latents(self, latents: torch.Tensor, num_frames: int, decod... method check_inputs (line 311) | def check_inputs(self, image, height, width): method prepare_latents (line 325) | def prepare_latents( method guidance_scale (line 360) | def guidance_scale(self): method do_classifier_free_guidance (line 367) | def do_classifier_free_guidance(self): method num_timesteps (line 373) | def num_timesteps(self): method __call__ (line 378) | def __call__( function _resize_with_antialiasing (line 622) | def _resize_with_antialiasing(input, size, interpolation="bicubic", alig... function _compute_padding (line 651) | def _compute_padding(kernel_size): function _filter2d (line 674) | def _filter2d(input, kernel): function _gaussian (line 697) | def _gaussian(window_size: int, sigma): function _gaussian_blur2d (line 713) | def _gaussian_blur2d(input, kernel_size, sigma): FILE: src/diffusers/pipelines/t2i_adapter/pipeline_stable_diffusion_adapter.py class StableDiffusionAdapterPipelineOutput (line 45) | class StableDiffusionAdapterPipelineOutput(BaseOutput): function _preprocess_adapter_image (line 94) | def _preprocess_adapter_image(image, height, width): function retrieve_timesteps (line 122) | def retrieve_timesteps( class StableDiffusionAdapterPipeline (line 181) | class StableDiffusionAdapterPipeline(DiffusionPipeline, StableDiffusionM... method __init__ (line 219) | def __init__( method _encode_prompt (line 267) | def _encode_prompt( method encode_prompt (line 300) | def encode_prompt( method run_safety_checker (line 483) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 498) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 510) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 527) | def check_inputs( method prepare_latents (line 587) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _default_height_width (line 609) | def _default_height_width(self, height, width, image): method get_guidance_scale_embedding (line 637) | def get_guidance_scale_embedding( method guidance_scale (line 668) | def guidance_scale(self): method do_classifier_free_guidance (line 675) | def do_classifier_free_guidance(self): method __call__ (line 680) | def __call__( FILE: src/diffusers/pipelines/t2i_adapter/pipeline_stable_diffusion_xl_adapter.py function _preprocess_adapter_image (line 93) | def _preprocess_adapter_image(image, height, width): function rescale_noise_cfg (line 121) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 136) | def retrieve_timesteps( class StableDiffusionXLAdapterPipeline (line 195) | class StableDiffusionXLAdapterPipeline( method __init__ (line 254) | def __init__( method encode_prompt (line 288) | def encode_prompt( method encode_image (line 523) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 548) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 594) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 612) | def check_inputs( method prepare_latents (line 709) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 732) | def _get_add_time_ids( method upcast_vae (line 751) | def upcast_vae(self): method _default_height_width (line 769) | def _default_height_width(self, height, width, image): method get_guidance_scale_embedding (line 797) | def get_guidance_scale_embedding( method guidance_scale (line 828) | def guidance_scale(self): method do_classifier_free_guidance (line 835) | def do_classifier_free_guidance(self): method __call__ (line 840) | def __call__( FILE: src/diffusers/pipelines/text_to_video_synthesis/pipeline_output.py class TextToVideoSDPipelineOutput (line 14) | class TextToVideoSDPipelineOutput(BaseOutput): FILE: src/diffusers/pipelines/text_to_video_synthesis/pipeline_text_to_video_synth.py class TextToVideoSDPipeline (line 61) | class TextToVideoSDPipeline( method __init__ (line 91) | def __init__( method _encode_prompt (line 112) | def _encode_prompt( method encode_prompt (line 145) | def encode_prompt( method decode_latents (line 327) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 340) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 358) | def check_inputs( method prepare_latents (line 410) | def prepare_latents( method __call__ (line 437) | def __call__( FILE: src/diffusers/pipelines/text_to_video_synthesis/pipeline_text_to_video_synth_img2img.py function retrieve_latents (line 83) | def retrieve_latents( class VideoToVideoSDPipeline (line 96) | class VideoToVideoSDPipeline( method __init__ (line 126) | def __init__( method _encode_prompt (line 147) | def _encode_prompt( method encode_prompt (line 180) | def encode_prompt( method decode_latents (line 363) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 376) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 393) | def check_inputs( method get_timesteps (line 445) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 456) | def prepare_latents(self, video, timestep, batch_size, dtype, device, ... method __call__ (line 502) | def __call__( FILE: src/diffusers/pipelines/text_to_video_synthesis/pipeline_text_to_video_zero.py function rearrange_0 (line 27) | def rearrange_0(tensor, f): function rearrange_1 (line 33) | def rearrange_1(tensor): function rearrange_3 (line 38) | def rearrange_3(tensor, f): function rearrange_4 (line 43) | def rearrange_4(tensor): class CrossFrameAttnProcessor (line 48) | class CrossFrameAttnProcessor: method __init__ (line 58) | def __init__(self, batch_size=2): method __call__ (line 61) | def __call__(self, attn, hidden_states, encoder_hidden_states=None, at... class CrossFrameAttnProcessor2_0 (line 107) | class CrossFrameAttnProcessor2_0: method __init__ (line 117) | def __init__(self, batch_size=2): method __call__ (line 122) | def __call__(self, attn, hidden_states, encoder_hidden_states=None, at... class TextToVideoPipelineOutput (line 183) | class TextToVideoPipelineOutput(BaseOutput): function coords_grid (line 200) | def coords_grid(batch, ht, wd, device): function warp_single_latent (line 207) | def warp_single_latent(latent, reference_flow): function create_motion_field (line 234) | def create_motion_field(motion_field_strength_x, motion_field_strength_y... function create_motion_field_and_warp_latents (line 257) | def create_motion_field_and_warp_latents(motion_field_strength_x, motion... class TextToVideoZeroPipeline (line 284) | class TextToVideoZeroPipeline( method __init__ (line 313) | def __init__( method forward_loop (line 346) | def forward_loop(self, x_t0, t0, t1, generator): method backward_loop (line 370) | def backward_loop( method check_inputs (line 446) | def check_inputs( method prepare_latents (line 499) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 522) | def __call__( method run_safety_checker (line 760) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 775) | def prepare_extra_step_kwargs(self, generator, eta): method encode_prompt (line 793) | def encode_prompt( method decode_latents (line 975) | def decode_latents(self, latents): FILE: src/diffusers/pipelines/text_to_video_synthesis/pipeline_text_to_video_zero_sdxl.py function rearrange_0 (line 49) | def rearrange_0(tensor, f): function rearrange_1 (line 56) | def rearrange_1(tensor): function rearrange_3 (line 62) | def rearrange_3(tensor, f): function rearrange_4 (line 68) | def rearrange_4(tensor): class CrossFrameAttnProcessor (line 74) | class CrossFrameAttnProcessor: method __init__ (line 84) | def __init__(self, batch_size=2): method __call__ (line 87) | def __call__(self, attn, hidden_states, encoder_hidden_states=None, at... class CrossFrameAttnProcessor2_0 (line 134) | class CrossFrameAttnProcessor2_0: method __init__ (line 144) | def __init__(self, batch_size=2): method __call__ (line 149) | def __call__(self, attn, hidden_states, encoder_hidden_states=None, at... class TextToVideoSDXLPipelineOutput (line 210) | class TextToVideoSDXLPipelineOutput(BaseOutput): function coords_grid (line 224) | def coords_grid(batch, ht, wd, device): function warp_single_latent (line 232) | def warp_single_latent(latent, reference_flow): function create_motion_field (line 260) | def create_motion_field(motion_field_strength_x, motion_field_strength_y... function create_motion_field_and_warp_latents (line 284) | def create_motion_field_and_warp_latents(motion_field_strength_x, motion... function rescale_noise_cfg (line 312) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class TextToVideoZeroSDXLPipeline (line 326) | class TextToVideoZeroSDXLPipeline( method __init__ (line 373) | def __init__( method prepare_extra_step_kwargs (line 413) | def prepare_extra_step_kwargs(self, generator, eta): method upcast_vae (line 431) | def upcast_vae(self): method _get_add_time_ids (line 450) | def _get_add_time_ids( method prepare_latents (line 469) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method check_inputs (line 491) | def check_inputs( method encode_prompt (line 571) | def encode_prompt( method forward_loop (line 806) | def forward_loop(self, x_t0, t0, t1, generator): method backward_loop (line 830) | def backward_loop( method __call__ (line 916) | def __call__( FILE: src/diffusers/pipelines/unclip/pipeline_unclip.py class UnCLIPPipeline (line 34) | class UnCLIPPipeline(DiffusionPipeline): method __init__ (line 81) | def __init__( method prepare_latents (line 109) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method _encode_prompt (line 120) | def _encode_prompt( method __call__ (line 212) | def __call__( FILE: src/diffusers/pipelines/unclip/pipeline_unclip_image_variation.py class UnCLIPImageVariationPipeline (line 39) | class UnCLIPImageVariationPipeline(DiffusionPipeline): method __init__ (line 82) | def __init__( method prepare_latents (line 111) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method _encode_prompt (line 122) | def _encode_prompt(self, prompt, device, num_images_per_prompt, do_cla... method _encode_image (line 185) | def _encode_image(self, image, device, num_images_per_prompt, image_em... method __call__ (line 200) | def __call__( FILE: src/diffusers/pipelines/unclip/text_proj.py class UnCLIPTextProjModel (line 22) | class UnCLIPTextProjModel(ModelMixin, ConfigMixin): method __init__ (line 31) | def __init__( method forward (line 55) | def forward(self, *, image_embeddings, prompt_embeds, text_encoder_hid... FILE: src/diffusers/pipelines/unidiffuser/modeling_text_decoder.py class UniDiffuserTextDecoder (line 14) | class UniDiffuserTextDecoder(ModelMixin, ConfigMixin, ModuleUtilsMixin): method __init__ (line 67) | def __init__( method forward (line 131) | def forward( method get_dummy_token (line 163) | def get_dummy_token(self, batch_size: int, device: torch.device) -> to... method encode (line 166) | def encode(self, prefix): method generate_captions (line 170) | def generate_captions(self, features, eos_token_id, device): method generate_beam (line 202) | def generate_beam( FILE: src/diffusers/pipelines/unidiffuser/modeling_uvit.py function _no_grad_trunc_normal_ (line 20) | def _no_grad_trunc_normal_(tensor, mean, std, a, b): function trunc_normal_ (line 57) | def trunc_normal_(tensor, mean=0.0, std=1.0, a=-2.0, b=2.0): class PatchEmbed (line 76) | class PatchEmbed(nn.Module): method __init__ (line 79) | def __init__( method forward (line 110) | def forward(self, latent): class SkipBlock (line 122) | class SkipBlock(nn.Module): method __init__ (line 123) | def __init__(self, dim: int): method forward (line 131) | def forward(self, x, skip): class UTransformerBlock (line 141) | class UTransformerBlock(nn.Module): method __init__ (line 175) | def __init__( method forward (line 252) | def forward( class UniDiffuserBlock (line 334) | class UniDiffuserBlock(nn.Module): method __init__ (line 370) | def __init__( method forward (line 447) | def forward( class UTransformer2DModel (line 531) | class UTransformer2DModel(ModelMixin, ConfigMixin): method __init__ (line 590) | def __init__( method forward (line 728) | def forward( class UniDiffuserModel (line 833) | class UniDiffuserModel(ModelMixin, ConfigMixin): method __init__ (line 898) | def __init__( method no_weight_decay (line 1035) | def no_weight_decay(self): method forward (line 1038) | def forward( FILE: src/diffusers/pipelines/unidiffuser/pipeline_unidiffuser.py class ImageTextPipelineOutput (line 34) | class ImageTextPipelineOutput(BaseOutput): class UniDiffuserPipeline (line 51) | class UniDiffuserPipeline(DiffusionPipeline): method __init__ (line 88) | def __init__( method prepare_extra_step_kwargs (line 139) | def prepare_extra_step_kwargs(self, generator, eta): method _infer_mode (line 156) | def _infer_mode(self, prompt, prompt_embeds, image, latents, prompt_la... method enable_vae_slicing (line 215) | def enable_vae_slicing(self): method disable_vae_slicing (line 223) | def disable_vae_slicing(self): method enable_vae_tiling (line 231) | def enable_vae_tiling(self): method disable_vae_tiling (line 240) | def disable_vae_tiling(self): method set_text_mode (line 248) | def set_text_mode(self): method set_image_mode (line 252) | def set_image_mode(self): method set_text_to_image_mode (line 256) | def set_text_to_image_mode(self): method set_image_to_text_mode (line 260) | def set_image_to_text_mode(self): method set_joint_mode (line 264) | def set_joint_mode(self): method reset_mode (line 268) | def reset_mode(self): method _infer_batch_size (line 272) | def _infer_batch_size( method _encode_prompt (line 349) | def _encode_prompt( method encode_prompt (line 382) | def encode_prompt( method encode_image_vae_latents (line 566) | def encode_image_vae_latents( method encode_image_clip_latents (line 626) | def encode_image_clip_latents( method prepare_text_latents (line 682) | def prepare_text_latents( method prepare_image_vae_latents (line 706) | def prepare_image_vae_latents( method prepare_image_clip_latents (line 741) | def prepare_image_clip_latents( method decode_text_latents (line 763) | def decode_text_latents(self, text_latents, device): method _split (line 774) | def _split(self, x, height, width): method _combine (line 790) | def _combine(self, img_vae, img_clip): method _split_joint (line 799) | def _split_joint(self, x, height, width): method _combine_joint (line 818) | def _combine_joint(self, img_vae, img_clip, text): method _get_noise_pred (line 829) | def _get_noise_pred( method check_latents_shape (line 951) | def check_latents_shape(self, latents_name, latents, expected_shape): method check_inputs (line 967) | def check_inputs( method __call__ (line 1081) | def __call__( FILE: src/diffusers/pipelines/wuerstchen/modeling_paella_vq_model.py class MixingResidualBlock (line 28) | class MixingResidualBlock(nn.Module): method __init__ (line 33) | def __init__(self, inp_channels, embed_dim): method forward (line 49) | def forward(self, x): class PaellaVQModel (line 58) | class PaellaVQModel(ModelMixin, ConfigMixin): method __init__ (line 77) | def __init__( method encode (line 133) | def encode(self, x: torch.Tensor, return_dict: bool = True) -> VQEncod... method decode (line 143) | def decode( method forward (line 158) | def forward(self, sample: torch.Tensor, return_dict: bool = True) -> U... FILE: src/diffusers/pipelines/wuerstchen/modeling_wuerstchen_common.py class WuerstchenLayerNorm (line 7) | class WuerstchenLayerNorm(nn.LayerNorm): method __init__ (line 8) | def __init__(self, *args, **kwargs): method forward (line 11) | def forward(self, x): class TimestepBlock (line 17) | class TimestepBlock(nn.Module): method __init__ (line 18) | def __init__(self, c, c_timestep): method forward (line 23) | def forward(self, x, t): class ResBlock (line 28) | class ResBlock(nn.Module): method __init__ (line 29) | def __init__(self, c, c_skip=0, kernel_size=3, dropout=0.0): method forward (line 38) | def forward(self, x, x_skip=None): class GlobalResponseNorm (line 48) | class GlobalResponseNorm(nn.Module): method __init__ (line 49) | def __init__(self, dim): method forward (line 54) | def forward(self, x): class AttnBlock (line 60) | class AttnBlock(nn.Module): method __init__ (line 61) | def __init__(self, c, c_cond, nhead, self_attn=True, dropout=0.0): method forward (line 69) | def forward(self, x, kv): FILE: src/diffusers/pipelines/wuerstchen/modeling_wuerstchen_diffnext.py class WuerstchenDiffNeXt (line 27) | class WuerstchenDiffNeXt(ModelMixin, ConfigMixin): method __init__ (line 29) | def __init__( method _init_weights (line 124) | def _init_weights(self, m): method gen_r_embedding (line 146) | def gen_r_embedding(self, r, max_positions=10000): method gen_c_embeddings (line 157) | def gen_c_embeddings(self, clip): method _down_encode (line 162) | def _down_encode(self, x, r_embed, effnet, clip=None): method _up_decode (line 186) | def _up_decode(self, level_outputs, r_embed, effnet, clip=None): method forward (line 214) | def forward(self, x, r, effnet, clip=None, x_cat=None, eps=1e-3, retur... class ResBlockStageB (line 235) | class ResBlockStageB(nn.Module): method __init__ (line 236) | def __init__(self, c, c_skip=0, kernel_size=3, dropout=0.0): method forward (line 248) | def forward(self, x, x_skip=None): FILE: src/diffusers/pipelines/wuerstchen/modeling_wuerstchen_prior.py class WuerstchenPrior (line 36) | class WuerstchenPrior(ModelMixin, ConfigMixin, UNet2DConditionLoadersMix... method __init__ (line 41) | def __init__(self, c_in=16, c=1280, c_cond=1024, c_r=64, depth=16, nhe... method attn_processors (line 67) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 91) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 126) | def set_default_attn_processor(self): method _set_gradient_checkpointing (line 141) | def _set_gradient_checkpointing(self, module, value=False): method gen_r_embedding (line 144) | def gen_r_embedding(self, r, max_positions=10000): method forward (line 155) | def forward(self, x, r, c): FILE: src/diffusers/pipelines/wuerstchen/pipeline_wuerstchen.py class WuerstchenDecoderPipeline (line 51) | class WuerstchenDecoderPipeline(DiffusionPipeline): method __init__ (line 83) | def __init__( method prepare_latents (line 103) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method encode_prompt (line 114) | def encode_prompt( method guidance_scale (line 197) | def guidance_scale(self): method do_classifier_free_guidance (line 201) | def do_classifier_free_guidance(self): method num_timesteps (line 205) | def num_timesteps(self): method __call__ (line 210) | def __call__( FILE: src/diffusers/pipelines/wuerstchen/pipeline_wuerstchen_combined.py class WuerstchenCombinedPipeline (line 43) | class WuerstchenCombinedPipeline(DiffusionPipeline): method __init__ (line 73) | def __init__( method enable_xformers_memory_efficient_attention (line 112) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_model_cpu_offload (line 115) | def enable_model_cpu_offload(self, gpu_id: Optional[int] = None, devic... method enable_sequential_cpu_offload (line 125) | def enable_sequential_cpu_offload(self, gpu_id: Optional[int] = None, ... method progress_bar (line 135) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 139) | def set_progress_bar_config(self, **kwargs): method __call__ (line 145) | def __call__( FILE: src/diffusers/pipelines/wuerstchen/pipeline_wuerstchen_prior.py class WuerstchenPriorPipelineOutput (line 52) | class WuerstchenPriorPipelineOutput(BaseOutput): class WuerstchenPriorPipeline (line 65) | class WuerstchenPriorPipeline(DiffusionPipeline, StableDiffusionLoraLoad... method __init__ (line 100) | def __init__( method prepare_latents (line 122) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method encode_prompt (line 133) | def encode_prompt( method check_inputs (line 228) | def check_inputs( method guidance_scale (line 270) | def guidance_scale(self): method do_classifier_free_guidance (line 274) | def do_classifier_free_guidance(self): method num_timesteps (line 278) | def num_timesteps(self): method __call__ (line 283) | def __call__( FILE: src/diffusers/schedulers/deprecated/scheduling_karras_ve.py class KarrasVeOutput (line 29) | class KarrasVeOutput(BaseOutput): class KarrasVeScheduler (line 49) | class KarrasVeScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 82) | def __init__( method scale_model_input (line 99) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 116) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method add_noise_to_input (line 138) | def add_noise_to_input( method step (line 164) | def step( method step_correct (line 203) | def step_correct( method add_noise (line 242) | def add_noise(self, original_samples, noise, timesteps): FILE: src/diffusers/schedulers/deprecated/scheduling_sde_vp.py class ScoreSdeVpScheduler (line 27) | class ScoreSdeVpScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 46) | def __init__(self, num_train_timesteps=2000, beta_min=0.1, beta_max=20... method set_timesteps (line 51) | def set_timesteps(self, num_inference_steps, device: Union[str, torch.... method step_pred (line 63) | def step_pred(self, score, x, t, generator=None): method __len__ (line 108) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_amused.py function gumbel_noise (line 12) | def gumbel_noise(t, generator=None): function mask_by_random_topk (line 18) | def mask_by_random_topk(mask_len, probs, temperature=1.0, generator=None): class AmusedSchedulerOutput (line 27) | class AmusedSchedulerOutput(BaseOutput): class AmusedScheduler (line 44) | class AmusedScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 50) | def __init__( method set_timesteps (line 58) | def set_timesteps( method step (line 71) | def step( method add_noise (line 140) | def add_noise(self, sample, timesteps, generator=None): FILE: src/diffusers/schedulers/scheduling_consistency_decoder.py function betas_for_alpha_bar (line 14) | def betas_for_alpha_bar( class ConsistencyDecoderSchedulerOutput (line 59) | class ConsistencyDecoderSchedulerOutput(BaseOutput): class ConsistencyDecoderScheduler (line 72) | class ConsistencyDecoderScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 76) | def __init__( method set_timesteps (line 97) | def set_timesteps( method init_noise_sigma (line 113) | def init_noise_sigma(self): method scale_model_input (line 116) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method step (line 133) | def step( FILE: src/diffusers/schedulers/scheduling_consistency_models.py class CMStochasticIterativeSchedulerOutput (line 31) | class CMStochasticIterativeSchedulerOutput(BaseOutput): class CMStochasticIterativeScheduler (line 44) | class CMStochasticIterativeScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 77) | def __init__( method step_index (line 105) | def step_index(self): method begin_index (line 112) | def begin_index(self): method set_begin_index (line 119) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 129) | def scale_model_input(self, sample: torch.Tensor, timestep: Union[floa... method sigma_to_t (line 154) | def sigma_to_t(self, sigmas: Union[float, np.ndarray]): method set_timesteps (line 173) | def set_timesteps( method _convert_to_karras (line 248) | def _convert_to_karras(self, ramp): method get_scalings (line 260) | def get_scalings(self, sigma): method get_scalings_for_boundary_condition (line 267) | def get_scalings_for_boundary_condition(self, sigma): method index_for_timestep (line 295) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 310) | def _init_step_index(self, timestep): method step (line 318) | def step( method add_noise (line 412) | def add_noise( method __len__ (line 445) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_cosine_dpmsolver_multistep.py class CosineDPMSolverMultistepScheduler (line 28) | class CosineDPMSolverMultistepScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 76) | def __init__( method init_noise_sigma (line 116) | def init_noise_sigma(self): method step_index (line 121) | def step_index(self): method begin_index (line 128) | def begin_index(self): method set_begin_index (line 135) | def set_begin_index(self, begin_index: int = 0): method precondition_inputs (line 146) | def precondition_inputs(self, sample, sigma): method precondition_noise (line 151) | def precondition_noise(self, sigma): method precondition_outputs (line 158) | def precondition_outputs(self, sample, model_output, sigma): method scale_model_input (line 174) | def scale_model_input(self, sample: torch.Tensor, timestep: Union[floa... method set_timesteps (line 198) | def set_timesteps(self, num_inference_steps: int = None, device: Union... method _compute_karras_sigmas (line 245) | def _compute_karras_sigmas(self, ramp, sigma_min=None, sigma_max=None)... method _compute_exponential_sigmas (line 257) | def _compute_exponential_sigmas(self, ramp, sigma_min=None, sigma_max=... method _sigma_to_t (line 268) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 291) | def _sigma_to_alpha_sigma_t(self, sigma): method convert_model_output (line 297) | def convert_model_output( method dpm_solver_first_order_update (line 329) | def dpm_solver_first_order_update( method multistep_dpm_solver_second_order_update (line 364) | def multistep_dpm_solver_second_order_update( method index_for_timestep (line 423) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 443) | def _init_step_index(self, timestep): method step (line 455) | def step( method add_noise (line 538) | def add_noise( method __len__ (line 571) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_ddim.py class DDIMSchedulerOutput (line 33) | class DDIMSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 51) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 95) | def rescale_zero_terminal_snr(betas): class DDIMScheduler (line 131) | class DDIMScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 185) | def __init__( method scale_model_input (line 236) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method _get_variance (line 253) | def _get_variance(self, timestep, prev_timestep): method _threshold_sample (line 264) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method set_timesteps (line 297) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method step (line 342) | def step( method add_noise (line 471) | def add_noise( method get_velocity (line 498) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 517) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_ddim_cogvideox.py class DDIMSchedulerOutput (line 33) | class DDIMSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 51) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 95) | def rescale_zero_terminal_snr(alphas_cumprod): class CogVideoXDDIMScheduler (line 126) | class CogVideoXDDIMScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 180) | def __init__( method _get_variance (line 233) | def _get_variance(self, timestep, prev_timestep): method scale_model_input (line 243) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 260) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method step (line 305) | def step( method add_noise (line 402) | def add_noise( method get_velocity (line 429) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 448) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_ddim_flax.py class DDIMSchedulerState (line 36) | class DDIMSchedulerState: method create (line 46) | def create( class FlaxDDIMSchedulerOutput (line 62) | class FlaxDDIMSchedulerOutput(FlaxSchedulerOutput): class FlaxDDIMScheduler (line 66) | class FlaxDDIMScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 110) | def has_state(self): method __init__ (line 114) | def __init__( method create_state (line 130) | def create_state(self, common: Optional[CommonSchedulerState] = None) ... method scale_model_input (line 154) | def scale_model_input( method set_timesteps (line 168) | def set_timesteps( method _get_variance (line 190) | def _get_variance(self, state: DDIMSchedulerState, timestep, prev_time... method step (line 202) | def step( method add_noise (line 295) | def add_noise( method get_velocity (line 304) | def get_velocity( method __len__ (line 313) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_ddim_inverse.py class DDIMSchedulerOutput (line 31) | class DDIMSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 49) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 94) | def rescale_zero_terminal_snr(betas): class DDIMInverseScheduler (line 130) | class DDIMInverseScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 177) | def __init__( method scale_model_input (line 234) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 251) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method step (line 289) | def step( method __len__ (line 373) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_ddim_parallel.py class DDIMParallelSchedulerOutput (line 33) | class DDIMParallelSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 51) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 96) | def rescale_zero_terminal_snr(betas): class DDIMParallelScheduler (line 132) | class DDIMParallelScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 192) | def __init__( method scale_model_input (line 244) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method _get_variance (line 261) | def _get_variance(self, timestep, prev_timestep=None): method _batch_get_variance (line 274) | def _batch_get_variance(self, t, prev_t): method _threshold_sample (line 286) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method set_timesteps (line 320) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method step (line 365) | def step( method batch_step_no_noise (line 487) | def batch_step_no_noise( method add_noise (line 596) | def add_noise( method get_velocity (line 623) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 642) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_ddpm.py class DDPMSchedulerOutput (line 31) | class DDPMSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 48) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 93) | def rescale_zero_terminal_snr(betas): class DDPMScheduler (line 129) | class DDPMScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 181) | def __init__( method scale_model_input (line 234) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 251) | def set_timesteps( method _get_variance (line 326) | def _get_variance(self, t, predicted_variance=None, variance_type=None): method _threshold_sample (line 366) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method step (line 399) | def step( method add_noise (line 499) | def add_noise( method get_velocity (line 525) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 544) | def __len__(self): method previous_timestep (line 547) | def previous_timestep(self, timestep): FILE: src/diffusers/schedulers/scheduling_ddpm_flax.py class DDPMSchedulerState (line 36) | class DDPMSchedulerState: method create (line 45) | def create(cls, common: CommonSchedulerState, init_noise_sigma: jnp.nd... class FlaxDDPMSchedulerOutput (line 50) | class FlaxDDPMSchedulerOutput(FlaxSchedulerOutput): class FlaxDDPMScheduler (line 54) | class FlaxDDPMScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 92) | def has_state(self): method __init__ (line 96) | def __init__( method create_state (line 110) | def create_state(self, common: Optional[CommonSchedulerState] = None) ... method scale_model_input (line 125) | def scale_model_input( method set_timesteps (line 139) | def set_timesteps( method _get_variance (line 162) | def _get_variance(self, state: DDPMSchedulerState, t, predicted_varian... method step (line 195) | def step( method add_noise (line 284) | def add_noise( method get_velocity (line 293) | def get_velocity( method __len__ (line 302) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_ddpm_parallel.py class DDPMParallelSchedulerOutput (line 32) | class DDPMParallelSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 50) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 95) | def rescale_zero_terminal_snr(betas): class DDPMParallelScheduler (line 131) | class DDPMParallelScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 189) | def __init__( method scale_model_input (line 243) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 261) | def set_timesteps( method _get_variance (line 337) | def _get_variance(self, t, predicted_variance=None, variance_type=None): method _threshold_sample (line 378) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method step (line 411) | def step( method batch_step_no_noise (line 507) | def batch_step_no_noise( method add_noise (line 588) | def add_noise( method get_velocity (line 615) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 634) | def __len__(self): method previous_timestep (line 638) | def previous_timestep(self, timestep): FILE: src/diffusers/schedulers/scheduling_ddpm_wuerstchen.py class DDPMWuerstchenSchedulerOutput (line 31) | class DDPMWuerstchenSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 44) | def betas_for_alpha_bar( class DDPMWuerstchenScheduler (line 88) | class DDPMWuerstchenScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 106) | def __init__( method _alpha_cumprod (line 118) | def _alpha_cumprod(self, t, device): method scale_model_input (line 128) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 142) | def set_timesteps( method step (line 164) | def step( method add_noise (line 210) | def add_noise( method __len__ (line 224) | def __len__(self): method previous_timestep (line 227) | def previous_timestep(self, timestep): FILE: src/diffusers/schedulers/scheduling_deis_multistep.py function betas_for_alpha_bar (line 30) | def betas_for_alpha_bar( class DEISMultistepScheduler (line 74) | class DEISMultistepScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 125) | def __init__( method step_index (line 192) | def step_index(self): method begin_index (line 199) | def begin_index(self): method set_begin_index (line 206) | def set_begin_index(self, begin_index: int = 0): method set_timesteps (line 216) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method _threshold_sample (line 279) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method _sigma_to_t (line 313) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 337) | def _sigma_to_alpha_sigma_t(self, sigma): method _convert_to_karras (line 344) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method convert_model_output (line 369) | def convert_model_output( method deis_first_order_update (line 426) | def deis_first_order_update( method multistep_deis_second_order_update (line 484) | def multistep_deis_second_order_update( method multistep_deis_third_order_update (line 553) | def multistep_deis_third_order_update( method index_for_timestep (line 642) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 662) | def _init_step_index(self, timestep): method step (line 674) | def step( method scale_model_input (line 739) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method add_noise (line 755) | def add_noise( method __len__ (line 789) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_dpm_cogvideox.py class DDIMSchedulerOutput (line 34) | class DDIMSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 52) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 96) | def rescale_zero_terminal_snr(alphas_cumprod): class CogVideoXDPMScheduler (line 127) | class CogVideoXDPMScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 181) | def __init__( method _get_variance (line 234) | def _get_variance(self, timestep, prev_timestep): method scale_model_input (line 244) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 261) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method get_variables (line 306) | def get_variables(self, alpha_prod_t, alpha_prod_t_prev, alpha_prod_t_... method get_mult (line 319) | def get_mult(self, h, r, alpha_prod_t, alpha_prod_t_prev, alpha_prod_t... method step (line 330) | def step( method add_noise (line 442) | def add_noise( method get_velocity (line 469) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 488) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_dpmsolver_multistep.py function betas_for_alpha_bar (line 30) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 75) | def rescale_zero_terminal_snr(betas): class DPMSolverMultistepScheduler (line 111) | class DPMSolverMultistepScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 192) | def __init__( method step_index (line 283) | def step_index(self): method begin_index (line 290) | def begin_index(self): method set_begin_index (line 296) | def set_begin_index(self, begin_index: int = 0): method set_timesteps (line 306) | def set_timesteps( method _threshold_sample (line 411) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method _sigma_to_t (line 445) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 468) | def _sigma_to_alpha_sigma_t(self, sigma): method _convert_to_karras (line 475) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method _convert_to_lu (line 500) | def _convert_to_lu(self, in_lambdas: torch.Tensor, num_inference_steps... method convert_model_output (line 513) | def convert_model_output( method dpm_solver_first_order_update (line 612) | def dpm_solver_first_order_update( method multistep_dpm_solver_second_order_update (line 681) | def multistep_dpm_solver_second_order_update( method multistep_dpm_solver_third_order_update (line 804) | def multistep_dpm_solver_third_order_update( method index_for_timestep (line 889) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 908) | def _init_step_index(self, timestep): method step (line 920) | def step( method scale_model_input (line 1009) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method add_noise (line 1024) | def add_noise( method __len__ (line 1058) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_dpmsolver_multistep_flax.py class DPMSolverMultistepSchedulerState (line 35) | class DPMSolverMultistepSchedulerState: method create (line 53) | def create( class FlaxDPMSolverMultistepSchedulerOutput (line 73) | class FlaxDPMSolverMultistepSchedulerOutput(FlaxSchedulerOutput): class FlaxDPMSolverMultistepScheduler (line 77) | class FlaxDPMSolverMultistepScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 150) | def has_state(self): method __init__ (line 154) | def __init__( method create_state (line 174) | def create_state(self, common: Optional[CommonSchedulerState] = None) ... method set_timesteps (line 203) | def set_timesteps( method convert_model_output (line 257) | def convert_model_output( method dpm_solver_first_order_update (line 327) | def dpm_solver_first_order_update( method multistep_dpm_solver_second_order_update (line 362) | def multistep_dpm_solver_second_order_update( method multistep_dpm_solver_third_order_update (line 422) | def multistep_dpm_solver_third_order_update( method step (line 478) | def step( method scale_model_input (line 615) | def scale_model_input( method add_noise (line 633) | def add_noise( method __len__ (line 642) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_dpmsolver_multistep_inverse.py function betas_for_alpha_bar (line 30) | def betas_for_alpha_bar( class DPMSolverMultistepInverseScheduler (line 74) | class DPMSolverMultistepInverseScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 144) | def __init__( method step_index (line 218) | def step_index(self): method set_timesteps (line 224) | def set_timesteps(self, num_inference_steps: int = None, device: Union... method _threshold_sample (line 298) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method _sigma_to_t (line 332) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 356) | def _sigma_to_alpha_sigma_t(self, sigma): method _convert_to_karras (line 363) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method convert_model_output (line 389) | def convert_model_output( method dpm_solver_first_order_update (line 489) | def dpm_solver_first_order_update( method multistep_dpm_solver_second_order_update (line 559) | def multistep_dpm_solver_second_order_update( method multistep_dpm_solver_third_order_update (line 683) | def multistep_dpm_solver_third_order_update( method _init_step_index (line 768) | def _init_step_index(self, timestep): method step (line 787) | def step( method scale_model_input (line 870) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method add_noise (line 885) | def add_noise( method __len__ (line 920) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_dpmsolver_sde.py class BatchedBrownianTree (line 26) | class BatchedBrownianTree: method __init__ (line 29) | def __init__(self, x, t0, t1, seed=None, **kwargs): method sort (line 44) | def sort(a, b): method __call__ (line 47) | def __call__(self, t0, t1): class BrownianTreeNoiseSampler (line 53) | class BrownianTreeNoiseSampler: method __init__ (line 68) | def __init__(self, x, sigma_min, sigma_max, seed=None, transform=lambd... method __call__ (line 73) | def __call__(self, sigma, sigma_next): function betas_for_alpha_bar (line 79) | def betas_for_alpha_bar( class DPMSolverSDEScheduler (line 123) | class DPMSolverSDEScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 163) | def __init__( method index_for_timestep (line 202) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 217) | def _init_step_index(self, timestep): method init_noise_sigma (line 226) | def init_noise_sigma(self): method step_index (line 234) | def step_index(self): method begin_index (line 241) | def begin_index(self): method set_begin_index (line 248) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 258) | def scale_model_input( method set_timesteps (line 285) | def set_timesteps( method _second_order_timesteps (line 358) | def _second_order_timesteps(self, sigmas, log_sigmas): method _sigma_to_t (line 374) | def _sigma_to_t(self, sigma, log_sigmas): method _convert_to_karras (line 398) | def _convert_to_karras(self, in_sigmas: torch.Tensor) -> torch.Tensor: method state_in_first_order (line 412) | def state_in_first_order(self): method step (line 415) | def step( method add_noise (line 527) | def add_noise( method __len__ (line 560) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_dpmsolver_singlestep.py function betas_for_alpha_bar (line 33) | def betas_for_alpha_bar( class DPMSolverSinglestepScheduler (line 77) | class DPMSolverSinglestepScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 141) | def __init__( method get_order_list (line 217) | def get_order_list(self, num_inference_steps: int) -> List[int]: method step_index (line 254) | def step_index(self): method begin_index (line 261) | def begin_index(self): method set_begin_index (line 268) | def set_begin_index(self, begin_index: int = 0): method set_timesteps (line 278) | def set_timesteps( method _threshold_sample (line 365) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method _sigma_to_t (line 399) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 423) | def _sigma_to_alpha_sigma_t(self, sigma): method _convert_to_karras (line 430) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method convert_model_output (line 455) | def convert_model_output( method dpm_solver_first_order_update (line 552) | def dpm_solver_first_order_update( method singlestep_dpm_solver_second_order_update (line 616) | def singlestep_dpm_solver_second_order_update( method singlestep_dpm_solver_third_order_update (line 727) | def singlestep_dpm_solver_third_order_update( method singlestep_dpm_solver_update (line 831) | def singlestep_dpm_solver_update( method index_for_timestep (line 895) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 915) | def _init_step_index(self, timestep): method step (line 927) | def step( method scale_model_input (line 998) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method add_noise (line 1014) | def add_noise( method __len__ (line 1048) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_edm_dpmsolver_multistep.py class EDMDPMSolverMultistepScheduler (line 28) | class EDMDPMSolverMultistepScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 92) | def __init__( method init_noise_sigma (line 148) | def init_noise_sigma(self): method step_index (line 153) | def step_index(self): method begin_index (line 160) | def begin_index(self): method set_begin_index (line 167) | def set_begin_index(self, begin_index: int = 0): method precondition_inputs (line 178) | def precondition_inputs(self, sample, sigma): method precondition_noise (line 184) | def precondition_noise(self, sigma): method precondition_outputs (line 193) | def precondition_outputs(self, sample, model_output, sigma): method scale_model_input (line 209) | def scale_model_input(self, sample: torch.Tensor, timestep: Union[floa... method set_timesteps (line 233) | def set_timesteps(self, num_inference_steps: int = None, device: Union... method _compute_karras_sigmas (line 277) | def _compute_karras_sigmas(self, ramp, sigma_min=None, sigma_max=None)... method _compute_exponential_sigmas (line 289) | def _compute_exponential_sigmas(self, ramp, sigma_min=None, sigma_max=... method _threshold_sample (line 300) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method _sigma_to_t (line 334) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 357) | def _sigma_to_alpha_sigma_t(self, sigma): method convert_model_output (line 363) | def convert_model_output( method dpm_solver_first_order_update (line 398) | def dpm_solver_first_order_update( method multistep_dpm_solver_second_order_update (line 436) | def multistep_dpm_solver_second_order_update( method multistep_dpm_solver_third_order_update (line 507) | def multistep_dpm_solver_third_order_update( method index_for_timestep (line 562) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 582) | def _init_step_index(self, timestep): method step (line 594) | def step( method add_noise (line 673) | def add_noise( method __len__ (line 706) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_edm_euler.py class EDMEulerSchedulerOutput (line 32) | class EDMEulerSchedulerOutput(BaseOutput): class EDMEulerScheduler (line 49) | class EDMEulerScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 86) | def __init__( method init_noise_sigma (line 119) | def init_noise_sigma(self): method step_index (line 124) | def step_index(self): method begin_index (line 131) | def begin_index(self): method set_begin_index (line 138) | def set_begin_index(self, begin_index: int = 0): method precondition_inputs (line 148) | def precondition_inputs(self, sample, sigma): method precondition_noise (line 153) | def precondition_noise(self, sigma): method precondition_outputs (line 161) | def precondition_outputs(self, sample, model_output, sigma): method scale_model_input (line 176) | def scale_model_input(self, sample: torch.Tensor, timestep: Union[floa... method set_timesteps (line 200) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method _compute_karras_sigmas (line 227) | def _compute_karras_sigmas(self, ramp, sigma_min=None, sigma_max=None)... method _compute_exponential_sigmas (line 238) | def _compute_exponential_sigmas(self, ramp, sigma_min=None, sigma_max=... method index_for_timestep (line 249) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 264) | def _init_step_index(self, timestep): method step (line 272) | def step( method add_noise (line 368) | def add_noise( method __len__ (line 401) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_euler_ancestral_discrete.py class EulerAncestralDiscreteSchedulerOutput (line 33) | class EulerAncestralDiscreteSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 51) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 96) | def rescale_zero_terminal_snr(betas): class EulerAncestralDiscreteScheduler (line 132) | class EulerAncestralDiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 170) | def __init__( method init_noise_sigma (line 221) | def init_noise_sigma(self): method step_index (line 229) | def step_index(self): method begin_index (line 236) | def begin_index(self): method set_begin_index (line 243) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 253) | def scale_model_input(self, sample: torch.Tensor, timestep: Union[floa... method set_timesteps (line 277) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method index_for_timestep (line 322) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 337) | def _init_step_index(self, timestep): method step (line 345) | def step( method add_noise (line 445) | def add_noise( method __len__ (line 478) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_euler_discrete.py class EulerDiscreteSchedulerOutput (line 33) | class EulerDiscreteSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 51) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 96) | def rescale_zero_terminal_snr(betas): class EulerDiscreteScheduler (line 132) | class EulerDiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 179) | def __init__( method init_noise_sigma (line 244) | def init_noise_sigma(self): method step_index (line 253) | def step_index(self): method begin_index (line 260) | def begin_index(self): method set_begin_index (line 267) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 277) | def scale_model_input(self, sample: torch.Tensor, timestep: Union[floa... method set_timesteps (line 301) | def set_timesteps( method _sigma_to_t (line 422) | def _sigma_to_t(self, sigma, log_sigmas): method _convert_to_karras (line 446) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method index_for_timestep (line 471) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 485) | def _init_step_index(self, timestep): method step (line 493) | def step( method add_noise (line 601) | def add_noise( method get_velocity (line 634) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 671) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_euler_discrete_flax.py class EulerDiscreteSchedulerState (line 32) | class EulerDiscreteSchedulerState: method create (line 42) | def create( class FlaxEulerDiscreteSchedulerOutput (line 49) | class FlaxEulerDiscreteSchedulerOutput(FlaxSchedulerOutput): class FlaxEulerDiscreteScheduler (line 53) | class FlaxEulerDiscreteScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 87) | def has_state(self): method __init__ (line 91) | def __init__( method create_state (line 104) | def create_state(self, common: Optional[CommonSchedulerState] = None) ... method scale_model_input (line 126) | def scale_model_input(self, state: EulerDiscreteSchedulerState, sample... method set_timesteps (line 148) | def set_timesteps( method step (line 189) | def step( method add_noise (line 250) | def add_noise( method __len__ (line 264) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_flow_match_euler_discrete.py class FlowMatchEulerDiscreteSchedulerOutput (line 31) | class FlowMatchEulerDiscreteSchedulerOutput(BaseOutput): class FlowMatchEulerDiscreteScheduler (line 44) | class FlowMatchEulerDiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 65) | def __init__( method step_index (line 93) | def step_index(self): method begin_index (line 100) | def begin_index(self): method set_begin_index (line 107) | def set_begin_index(self, begin_index: int = 0): method scale_noise (line 117) | def scale_noise( method _sigma_to_t (line 165) | def _sigma_to_t(self, sigma): method time_shift (line 168) | def time_shift(self, mu: float, sigma: float, t: torch.Tensor): method set_timesteps (line 171) | def set_timesteps( method index_for_timestep (line 213) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 227) | def _init_step_index(self, timestep): method step (line 235) | def step( method step_forward (line 310) | def step_forward( method __len__ (line 359) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_flow_match_heun_discrete.py class FlowMatchHeunDiscreteSchedulerOutput (line 31) | class FlowMatchHeunDiscreteSchedulerOutput(BaseOutput): class FlowMatchHeunDiscreteScheduler (line 44) | class FlowMatchHeunDiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 65) | def __init__( method step_index (line 86) | def step_index(self): method begin_index (line 93) | def begin_index(self): method set_begin_index (line 100) | def set_begin_index(self, begin_index: int = 0): method scale_noise (line 110) | def scale_noise( method _sigma_to_t (line 137) | def _sigma_to_t(self, sigma): method set_timesteps (line 140) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method index_for_timestep (line 174) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 188) | def _init_step_index(self, timestep): method state_in_first_order (line 197) | def state_in_first_order(self): method step (line 200) | def step( method __len__ (line 320) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_heun_discrete.py function betas_for_alpha_bar (line 26) | def betas_for_alpha_bar( class HeunDiscreteScheduler (line 70) | class HeunDiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 111) | def __init__( method index_for_timestep (line 152) | def index_for_timestep(self, timestep, schedule_timesteps=None): method init_noise_sigma (line 167) | def init_noise_sigma(self): method step_index (line 175) | def step_index(self): method begin_index (line 182) | def begin_index(self): method set_begin_index (line 189) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 199) | def scale_model_input( method set_timesteps (line 225) | def set_timesteps( method _sigma_to_t (line 308) | def _sigma_to_t(self, sigma, log_sigmas): method _convert_to_karras (line 332) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method state_in_first_order (line 358) | def state_in_first_order(self): method _init_step_index (line 362) | def _init_step_index(self, timestep): method step (line 370) | def step( method add_noise (line 470) | def add_noise( method __len__ (line 503) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_ipndm.py class IPNDMScheduler (line 25) | class IPNDMScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 42) | def __init__( method step_index (line 62) | def step_index(self): method begin_index (line 69) | def begin_index(self): method set_begin_index (line 76) | def set_begin_index(self, begin_index: int = 0): method set_timesteps (line 86) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method index_for_timestep (line 115) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 130) | def _init_step_index(self, timestep): method step (line 138) | def step( method scale_model_input (line 196) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method _get_prev_sample (line 211) | def _get_prev_sample(self, sample, timestep_index, prev_timestep_index... method __len__ (line 223) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_k_dpm_2_ancestral_discrete.py function betas_for_alpha_bar (line 27) | def betas_for_alpha_bar( class KDPM2AncestralDiscreteScheduler (line 71) | class KDPM2AncestralDiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 109) | def __init__( method init_noise_sigma (line 144) | def init_noise_sigma(self): method step_index (line 152) | def step_index(self): method begin_index (line 159) | def begin_index(self): method set_begin_index (line 166) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 176) | def scale_model_input( method set_timesteps (line 206) | def set_timesteps( method _sigma_to_t (line 300) | def _sigma_to_t(self, sigma, log_sigmas): method _convert_to_karras (line 324) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method state_in_first_order (line 350) | def state_in_first_order(self): method index_for_timestep (line 354) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 369) | def _init_step_index(self, timestep): method step (line 377) | def step( method add_noise (line 478) | def add_noise( method __len__ (line 511) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_k_dpm_2_discrete.py function betas_for_alpha_bar (line 26) | def betas_for_alpha_bar( class KDPM2DiscreteScheduler (line 70) | class KDPM2DiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 108) | def __init__( method init_noise_sigma (line 144) | def init_noise_sigma(self): method step_index (line 152) | def step_index(self): method begin_index (line 159) | def begin_index(self): method set_begin_index (line 166) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 176) | def scale_model_input( method set_timesteps (line 206) | def set_timesteps( method state_in_first_order (line 285) | def state_in_first_order(self): method index_for_timestep (line 289) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 304) | def _init_step_index(self, timestep): method _sigma_to_t (line 313) | def _sigma_to_t(self, sigma, log_sigmas): method _convert_to_karras (line 337) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method step (line 362) | def step( method add_noise (line 453) | def add_noise( method __len__ (line 486) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_karras_ve_flax.py class KarrasVeSchedulerState (line 30) | class KarrasVeSchedulerState: method create (line 37) | def create(cls): class FlaxKarrasVeOutput (line 42) | class FlaxKarrasVeOutput(BaseOutput): class FlaxKarrasVeScheduler (line 60) | class FlaxKarrasVeScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 92) | def has_state(self): method __init__ (line 96) | def __init__( method create_state (line 107) | def create_state(self): method set_timesteps (line 110) | def set_timesteps( method add_noise_to_input (line 138) | def add_noise_to_input( method step (line 164) | def step( method step_correct (line 200) | def step_correct( method add_noise (line 237) | def add_noise(self, state: KarrasVeSchedulerState, original_samples, n... FILE: src/diffusers/schedulers/scheduling_lcm.py class LCMSchedulerOutput (line 35) | class LCMSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 53) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 98) | def rescale_zero_terminal_snr(betas: torch.Tensor) -> torch.Tensor: class LCMScheduler (line 134) | class LCMScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 196) | def __init__( method index_for_timestep (line 254) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 269) | def _init_step_index(self, timestep): method step_index (line 278) | def step_index(self): method begin_index (line 282) | def begin_index(self): method set_begin_index (line 289) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 299) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method _threshold_sample (line 316) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method set_timesteps (line 349) | def set_timesteps( method get_scalings_for_boundary_condition_discrete (line 490) | def get_scalings_for_boundary_condition_discrete(self, timestep): method step (line 498) | def step( method add_noise (line 595) | def add_noise( method get_velocity (line 622) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 641) | def __len__(self): method previous_timestep (line 645) | def previous_timestep(self, timestep): FILE: src/diffusers/schedulers/scheduling_lms_discrete.py class LMSDiscreteSchedulerOutput (line 30) | class LMSDiscreteSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 48) | def betas_for_alpha_bar( class LMSDiscreteScheduler (line 92) | class LMSDiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 129) | def __init__( method init_noise_sigma (line 173) | def init_noise_sigma(self): method step_index (line 181) | def step_index(self): method begin_index (line 188) | def begin_index(self): method set_begin_index (line 195) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 205) | def scale_model_input(self, sample: torch.Tensor, timestep: Union[floa... method get_lms_coefficient (line 229) | def get_lms_coefficient(self, order, t, current_order): method set_timesteps (line 251) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method index_for_timestep (line 304) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 319) | def _init_step_index(self, timestep): method _sigma_to_t (line 328) | def _sigma_to_t(self, sigma, log_sigmas): method _convert_to_karras (line 352) | def _convert_to_karras(self, in_sigmas: torch.Tensor) -> torch.Tensor: method step (line 365) | def step( method add_noise (line 443) | def add_noise( method __len__ (line 476) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_lms_discrete_flax.py class LMSDiscreteSchedulerState (line 33) | class LMSDiscreteSchedulerState: method create (line 46) | def create( class FlaxLMSSchedulerOutput (line 53) | class FlaxLMSSchedulerOutput(FlaxSchedulerOutput): class FlaxLMSDiscreteScheduler (line 57) | class FlaxLMSDiscreteScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 90) | def has_state(self): method __init__ (line 94) | def __init__( method create_state (line 106) | def create_state(self, common: Optional[CommonSchedulerState] = None) ... method scale_model_input (line 123) | def scale_model_input(self, state: LMSDiscreteSchedulerState, sample: ... method get_lms_coefficient (line 145) | def get_lms_coefficient(self, state: LMSDiscreteSchedulerState, order,... method set_timesteps (line 167) | def set_timesteps( method step (line 203) | def step( method add_noise (line 268) | def add_noise( method __len__ (line 282) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_pndm.py function betas_for_alpha_bar (line 28) | def betas_for_alpha_bar( class PNDMScheduler (line 72) | class PNDMScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 114) | def __init__( method set_timesteps (line 166) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method step (line 226) | def step( method step_prk (line 259) | def step_prk( method step_plms (line 319) | def step_plms( method scale_model_input (line 390) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method _get_prev_sample (line 405) | def _get_prev_sample(self, sample, timestep, prev_timestep, model_outp... method add_noise (line 449) | def add_noise( method __len__ (line 475) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_pndm_flax.py class PNDMSchedulerState (line 35) | class PNDMSchedulerState: method create (line 53) | def create( class FlaxPNDMSchedulerOutput (line 69) | class FlaxPNDMSchedulerOutput(FlaxSchedulerOutput): class FlaxPNDMScheduler (line 73) | class FlaxPNDMScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 117) | def has_state(self): method __init__ (line 121) | def __init__( method create_state (line 141) | def create_state(self, common: Optional[CommonSchedulerState] = None) ... method set_timesteps (line 165) | def set_timesteps(self, state: PNDMSchedulerState, num_inference_steps... method scale_model_input (line 220) | def scale_model_input( method step (line 237) | def step( method step_prk (line 292) | def step_prk( method step_plms (line 360) | def step_plms( method _get_prev_sample (line 454) | def _get_prev_sample(self, state: PNDMSchedulerState, sample, timestep... method add_noise (line 499) | def add_noise( method __len__ (line 508) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_repaint.py class RePaintSchedulerOutput (line 29) | class RePaintSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 47) | def betas_for_alpha_bar( class RePaintScheduler (line 91) | class RePaintScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 121) | def __init__( method scale_model_input (line 163) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 180) | def set_timesteps( method _get_variance (line 227) | def _get_variance(self, t): method step (line 246) | def step( method undo_step (line 335) | def undo_step(self, sample, timestep, generator=None): method add_noise (line 352) | def add_noise( method __len__ (line 360) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_sasolver.py function betas_for_alpha_bar (line 31) | def betas_for_alpha_bar( class SASolverScheduler (line 75) | class SASolverScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 142) | def __init__( method step_index (line 218) | def step_index(self): method begin_index (line 225) | def begin_index(self): method set_begin_index (line 232) | def set_begin_index(self, begin_index: int = 0): method set_timesteps (line 242) | def set_timesteps(self, num_inference_steps: int = None, device: Union... method _threshold_sample (line 308) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method _sigma_to_t (line 342) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 366) | def _sigma_to_alpha_sigma_t(self, sigma): method _convert_to_karras (line 373) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method convert_model_output (line 398) | def convert_model_output( method get_coefficients_exponential_negative (line 490) | def get_coefficients_exponential_negative(self, order, interval_start,... method get_coefficients_exponential_positive (line 514) | def get_coefficients_exponential_positive(self, order, interval_start,... method lagrange_polynomial_coefficient (line 558) | def lagrange_polynomial_coefficient(self, order, lambda_list): method get_coefficients_fn (line 667) | def get_coefficients_fn(self, order, interval_start, interval_end, lam... method stochastic_adams_bashforth_update (line 687) | def stochastic_adams_bashforth_update( method stochastic_adams_moulton_update (line 814) | def stochastic_adams_moulton_update( method index_for_timestep (line 948) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 968) | def _init_step_index(self, timestep): method step (line 980) | def step( method scale_model_input (line 1082) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method add_noise (line 1098) | def add_noise( method __len__ (line 1124) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_sde_ve.py class SdeVeOutput (line 30) | class SdeVeOutput(BaseOutput): class ScoreSdeVeScheduler (line 46) | class ScoreSdeVeScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 72) | def __init__( method scale_model_input (line 89) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 106) | def set_timesteps( method set_sigmas (line 125) | def set_sigmas( method get_adjacent_sigma (line 153) | def get_adjacent_sigma(self, timesteps, t): method step_pred (line 160) | def step_pred( method step_correct (line 228) | def step_correct( method add_noise (line 283) | def add_noise( method __len__ (line 300) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_sde_ve_flax.py class ScoreSdeVeSchedulerState (line 30) | class ScoreSdeVeSchedulerState: method create (line 37) | def create(cls): class FlaxSdeVeOutput (line 42) | class FlaxSdeVeOutput(FlaxSchedulerOutput): class FlaxScoreSdeVeScheduler (line 60) | class FlaxScoreSdeVeScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 85) | def has_state(self): method __init__ (line 89) | def __init__( method create_state (line 100) | def create_state(self): method set_timesteps (line 110) | def set_timesteps( method set_sigmas (line 129) | def set_sigmas( method get_adjacent_sigma (line 164) | def get_adjacent_sigma(self, state, timesteps, t): method step_pred (line 167) | def step_pred( method step_correct (line 227) | def step_correct( method __len__ (line 279) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_tcd.py class TCDSchedulerOutput (line 35) | class TCDSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 52) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 97) | def rescale_zero_terminal_snr(betas: torch.Tensor) -> torch.Tensor: class TCDScheduler (line 133) | class TCDScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 197) | def __init__( method index_for_timestep (line 255) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 270) | def _init_step_index(self, timestep): method step_index (line 279) | def step_index(self): method begin_index (line 283) | def begin_index(self): method set_begin_index (line 290) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 300) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method _get_variance (line 318) | def _get_variance(self, timestep, prev_timestep): method _threshold_sample (line 329) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method set_timesteps (line 362) | def set_timesteps( method step (line 525) | def step( method add_noise (line 632) | def add_noise( method get_velocity (line 659) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 678) | def __len__(self): method previous_timestep (line 682) | def previous_timestep(self, timestep): FILE: src/diffusers/schedulers/scheduling_unclip.py class UnCLIPSchedulerOutput (line 30) | class UnCLIPSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 48) | def betas_for_alpha_bar( class UnCLIPScheduler (line 92) | class UnCLIPScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 122) | def __init__( method scale_model_input (line 149) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 163) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method _get_variance (line 180) | def _get_variance(self, t, prev_timestep=None, predicted_variance=None... method step (line 216) | def step( method add_noise (line 328) | def add_noise( FILE: src/diffusers/schedulers/scheduling_unipc_multistep.py function betas_for_alpha_bar (line 30) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 75) | def rescale_zero_terminal_snr(betas): class UniPCMultistepScheduler (line 111) | class UniPCMultistepScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 180) | def __init__( method step_index (line 258) | def step_index(self): method begin_index (line 265) | def begin_index(self): method set_begin_index (line 272) | def set_begin_index(self, begin_index: int = 0): method set_timesteps (line 282) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method _threshold_sample (line 363) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method _sigma_to_t (line 397) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 421) | def _sigma_to_alpha_sigma_t(self, sigma): method _convert_to_karras (line 428) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method convert_model_output (line 453) | def convert_model_output( method multistep_uni_p_bh_update (line 523) | def multistep_uni_p_bh_update( method multistep_uni_c_bh_update (line 652) | def multistep_uni_c_bh_update( method index_for_timestep (line 790) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 810) | def _init_step_index(self, timestep): method step (line 822) | def step( method scale_model_input (line 903) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method add_noise (line 919) | def add_noise( method __len__ (line 953) | def __len__(self): FILE: src/diffusers/schedulers/scheduling_utils.py class KarrasDiffusionSchedulers (line 33) | class KarrasDiffusionSchedulers(Enum): class SchedulerOutput (line 61) | class SchedulerOutput(BaseOutput): class SchedulerMixin (line 74) | class SchedulerMixin(PushToHubMixin): method from_pretrained (line 96) | def from_pretrained( method save_pretrained (line 159) | def save_pretrained(self, save_directory: Union[str, os.PathLike], pus... method compatibles (line 177) | def compatibles(self): method _get_compatibles (line 187) | def _get_compatibles(cls): FILE: src/diffusers/schedulers/scheduling_utils_flax.py class FlaxKarrasDiffusionSchedulers (line 35) | class FlaxKarrasDiffusionSchedulers(Enum): class FlaxSchedulerOutput (line 45) | class FlaxSchedulerOutput(BaseOutput): class FlaxSchedulerMixin (line 58) | class FlaxSchedulerMixin(PushToHubMixin): method from_pretrained (line 75) | def from_pretrained( method save_pretrained (line 152) | def save_pretrained(self, save_directory: Union[str, os.PathLike], pus... method compatibles (line 170) | def compatibles(self): method _get_compatibles (line 180) | def _get_compatibles(cls): function broadcast_to_shape_from_left (line 189) | def broadcast_to_shape_from_left(x: jnp.ndarray, shape: Tuple[int]) -> j... function betas_for_alpha_bar (line 194) | def betas_for_alpha_bar(num_diffusion_timesteps: int, max_beta=0.999, dt... class CommonSchedulerState (line 224) | class CommonSchedulerState: method create (line 230) | def create(cls, scheduler): function get_sqrt_alpha_prod (line 264) | def get_sqrt_alpha_prod( function add_noise_common (line 280) | def add_noise_common( function get_velocity_common (line 288) | def get_velocity_common(state: CommonSchedulerState, sample: jnp.ndarray... FILE: src/diffusers/schedulers/scheduling_vq_diffusion.py class VQDiffusionSchedulerOutput (line 28) | class VQDiffusionSchedulerOutput(BaseOutput): function index_to_log_onehot (line 41) | def index_to_log_onehot(x: torch.LongTensor, num_classes: int) -> torch.... function gumbel_noised (line 62) | def gumbel_noised(logits: torch.Tensor, generator: Optional[torch.Genera... function alpha_schedules (line 72) | def alpha_schedules(num_diffusion_timesteps: int, alpha_cum_start=0.9999... function gamma_schedules (line 88) | def gamma_schedules(num_diffusion_timesteps: int, gamma_cum_start=0.0000... class VQDiffusionScheduler (line 106) | class VQDiffusionScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 132) | def __init__( method set_timesteps (line 178) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method step (line 200) | def step( method q_posterior (line 245) | def q_posterior(self, log_p_x_0, x_t, t): method log_Q_t_transitioning_to_known_class (line 356) | def log_Q_t_transitioning_to_known_class( method apply_cumulative_transitions (line 455) | def apply_cumulative_transitions(self, q, t): FILE: src/diffusers/training_utils.py function set_seed (line 36) | def set_seed(seed: int): function compute_snr (line 52) | def compute_snr(noise_scheduler, timesteps): function resolve_interpolation_mode (line 78) | def resolve_interpolation_mode(interpolation_type: str): function compute_dream_and_update_latents (line 122) | def compute_dream_and_update_latents( function unet_lora_state_dict (line 176) | def unet_lora_state_dict(unet: UNet2DConditionModel) -> Dict[str, torch.... function cast_training_params (line 195) | def cast_training_params(model: Union[torch.nn.Module, List[torch.nn.Mod... function _set_state_dict_into_text_encoder (line 205) | def _set_state_dict_into_text_encoder( function compute_density_for_timestep_sampling (line 224) | def compute_density_for_timestep_sampling( function compute_loss_weighting_for_sd3 (line 245) | def compute_loss_weighting_for_sd3(weighting_scheme: str, sigmas=None): class EMAModel (line 263) | class EMAModel: method __init__ (line 268) | def __init__( method from_pretrained (line 353) | def from_pretrained(cls, path, model_cls, foreach=False) -> "EMAModel": method save_pretrained (line 362) | def save_pretrained(self, path): method get_decay (line 377) | def get_decay(self, optimization_step: int) -> float: method step (line 397) | def step(self, parameters: Iterable[torch.nn.Parameter]): method copy_to (line 456) | def copy_to(self, parameters: Iterable[torch.nn.Parameter]) -> None: method pin_memory (line 475) | def pin_memory(self) -> None: method to (line 483) | def to(self, device=None, dtype=None, non_blocking=False) -> None: method state_dict (line 497) | def state_dict(self) -> dict: method store (line 516) | def store(self, parameters: Iterable[torch.nn.Parameter]) -> None: method restore (line 525) | def restore(self, parameters: Iterable[torch.nn.Parameter]) -> None: method load_state_dict (line 548) | def load_state_dict(self, state_dict: dict) -> None: FILE: src/diffusers/utils/__init__.py function check_min_version (line 124) | def check_min_version(min_version): FILE: src/diffusers/utils/accelerate_utils.py function apply_forward_hook (line 27) | def apply_forward_hook(method): FILE: src/diffusers/utils/deprecation_utils.py function deprecate (line 8) | def deprecate(*args, take_from: Optional[Union[Dict, Any]] = None, stand... FILE: src/diffusers/utils/doc_utils.py function replace_example_docstring (line 21) | def replace_example_docstring(example_docstring): FILE: src/diffusers/utils/dummy_flax_and_transformers_objects.py class FlaxStableDiffusionControlNetPipeline (line 5) | class FlaxStableDiffusionControlNetPipeline(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class FlaxStableDiffusionImg2ImgPipeline (line 20) | class FlaxStableDiffusionImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): class FlaxStableDiffusionInpaintPipeline (line 35) | class FlaxStableDiffusionInpaintPipeline(metaclass=DummyObject): method __init__ (line 38) | def __init__(self, *args, **kwargs): method from_config (line 42) | def from_config(cls, *args, **kwargs): method from_pretrained (line 46) | def from_pretrained(cls, *args, **kwargs): class FlaxStableDiffusionPipeline (line 50) | class FlaxStableDiffusionPipeline(metaclass=DummyObject): method __init__ (line 53) | def __init__(self, *args, **kwargs): method from_config (line 57) | def from_config(cls, *args, **kwargs): method from_pretrained (line 61) | def from_pretrained(cls, *args, **kwargs): class FlaxStableDiffusionXLPipeline (line 65) | class FlaxStableDiffusionXLPipeline(metaclass=DummyObject): method __init__ (line 68) | def __init__(self, *args, **kwargs): method from_config (line 72) | def from_config(cls, *args, **kwargs): method from_pretrained (line 76) | def from_pretrained(cls, *args, **kwargs): FILE: src/diffusers/utils/dummy_flax_objects.py class FlaxControlNetModel (line 5) | class FlaxControlNetModel(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class FlaxModelMixin (line 20) | class FlaxModelMixin(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): class FlaxUNet2DConditionModel (line 35) | class FlaxUNet2DConditionModel(metaclass=DummyObject): method __init__ (line 38) | def __init__(self, *args, **kwargs): method from_config (line 42) | def from_config(cls, *args, **kwargs): method from_pretrained (line 46) | def from_pretrained(cls, *args, **kwargs): class FlaxAutoencoderKL (line 50) | class FlaxAutoencoderKL(metaclass=DummyObject): method __init__ (line 53) | def __init__(self, *args, **kwargs): method from_config (line 57) | def from_config(cls, *args, **kwargs): method from_pretrained (line 61) | def from_pretrained(cls, *args, **kwargs): class FlaxDiffusionPipeline (line 65) | class FlaxDiffusionPipeline(metaclass=DummyObject): method __init__ (line 68) | def __init__(self, *args, **kwargs): method from_config (line 72) | def from_config(cls, *args, **kwargs): method from_pretrained (line 76) | def from_pretrained(cls, *args, **kwargs): class FlaxDDIMScheduler (line 80) | class FlaxDDIMScheduler(metaclass=DummyObject): method __init__ (line 83) | def __init__(self, *args, **kwargs): method from_config (line 87) | def from_config(cls, *args, **kwargs): method from_pretrained (line 91) | def from_pretrained(cls, *args, **kwargs): class FlaxDDPMScheduler (line 95) | class FlaxDDPMScheduler(metaclass=DummyObject): method __init__ (line 98) | def __init__(self, *args, **kwargs): method from_config (line 102) | def from_config(cls, *args, **kwargs): method from_pretrained (line 106) | def from_pretrained(cls, *args, **kwargs): class FlaxDPMSolverMultistepScheduler (line 110) | class FlaxDPMSolverMultistepScheduler(metaclass=DummyObject): method __init__ (line 113) | def __init__(self, *args, **kwargs): method from_config (line 117) | def from_config(cls, *args, **kwargs): method from_pretrained (line 121) | def from_pretrained(cls, *args, **kwargs): class FlaxEulerDiscreteScheduler (line 125) | class FlaxEulerDiscreteScheduler(metaclass=DummyObject): method __init__ (line 128) | def __init__(self, *args, **kwargs): method from_config (line 132) | def from_config(cls, *args, **kwargs): method from_pretrained (line 136) | def from_pretrained(cls, *args, **kwargs): class FlaxKarrasVeScheduler (line 140) | class FlaxKarrasVeScheduler(metaclass=DummyObject): method __init__ (line 143) | def __init__(self, *args, **kwargs): method from_config (line 147) | def from_config(cls, *args, **kwargs): method from_pretrained (line 151) | def from_pretrained(cls, *args, **kwargs): class FlaxLMSDiscreteScheduler (line 155) | class FlaxLMSDiscreteScheduler(metaclass=DummyObject): method __init__ (line 158) | def __init__(self, *args, **kwargs): method from_config (line 162) | def from_config(cls, *args, **kwargs): method from_pretrained (line 166) | def from_pretrained(cls, *args, **kwargs): class FlaxPNDMScheduler (line 170) | class FlaxPNDMScheduler(metaclass=DummyObject): method __init__ (line 173) | def __init__(self, *args, **kwargs): method from_config (line 177) | def from_config(cls, *args, **kwargs): method from_pretrained (line 181) | def from_pretrained(cls, *args, **kwargs): class FlaxSchedulerMixin (line 185) | class FlaxSchedulerMixin(metaclass=DummyObject): method __init__ (line 188) | def __init__(self, *args, **kwargs): method from_config (line 192) | def from_config(cls, *args, **kwargs): method from_pretrained (line 196) | def from_pretrained(cls, *args, **kwargs): class FlaxScoreSdeVeScheduler (line 200) | class FlaxScoreSdeVeScheduler(metaclass=DummyObject): method __init__ (line 203) | def __init__(self, *args, **kwargs): method from_config (line 207) | def from_config(cls, *args, **kwargs): method from_pretrained (line 211) | def from_pretrained(cls, *args, **kwargs): FILE: src/diffusers/utils/dummy_note_seq_objects.py class MidiProcessor (line 5) | class MidiProcessor(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): FILE: src/diffusers/utils/dummy_onnx_objects.py class OnnxRuntimeModel (line 5) | class OnnxRuntimeModel(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): FILE: src/diffusers/utils/dummy_pt_objects.py class AsymmetricAutoencoderKL (line 5) | class AsymmetricAutoencoderKL(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class AuraFlowTransformer2DModel (line 20) | class AuraFlowTransformer2DModel(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): class AutoencoderKL (line 35) | class AutoencoderKL(metaclass=DummyObject): method __init__ (line 38) | def __init__(self, *args, **kwargs): method from_config (line 42) | def from_config(cls, *args, **kwargs): method from_pretrained (line 46) | def from_pretrained(cls, *args, **kwargs): class AutoencoderKLCogVideoX (line 50) | class AutoencoderKLCogVideoX(metaclass=DummyObject): method __init__ (line 53) | def __init__(self, *args, **kwargs): method from_config (line 57) | def from_config(cls, *args, **kwargs): method from_pretrained (line 61) | def from_pretrained(cls, *args, **kwargs): class AutoencoderKLTemporalDecoder (line 65) | class AutoencoderKLTemporalDecoder(metaclass=DummyObject): method __init__ (line 68) | def __init__(self, *args, **kwargs): method from_config (line 72) | def from_config(cls, *args, **kwargs): method from_pretrained (line 76) | def from_pretrained(cls, *args, **kwargs): class AutoencoderOobleck (line 80) | class AutoencoderOobleck(metaclass=DummyObject): method __init__ (line 83) | def __init__(self, *args, **kwargs): method from_config (line 87) | def from_config(cls, *args, **kwargs): method from_pretrained (line 91) | def from_pretrained(cls, *args, **kwargs): class AutoencoderTiny (line 95) | class AutoencoderTiny(metaclass=DummyObject): method __init__ (line 98) | def __init__(self, *args, **kwargs): method from_config (line 102) | def from_config(cls, *args, **kwargs): method from_pretrained (line 106) | def from_pretrained(cls, *args, **kwargs): class CogVideoXTransformer3DModel (line 110) | class CogVideoXTransformer3DModel(metaclass=DummyObject): method __init__ (line 113) | def __init__(self, *args, **kwargs): method from_config (line 117) | def from_config(cls, *args, **kwargs): method from_pretrained (line 121) | def from_pretrained(cls, *args, **kwargs): class ConsistencyDecoderVAE (line 125) | class ConsistencyDecoderVAE(metaclass=DummyObject): method __init__ (line 128) | def __init__(self, *args, **kwargs): method from_config (line 132) | def from_config(cls, *args, **kwargs): method from_pretrained (line 136) | def from_pretrained(cls, *args, **kwargs): class ControlNetModel (line 140) | class ControlNetModel(metaclass=DummyObject): method __init__ (line 143) | def __init__(self, *args, **kwargs): method from_config (line 147) | def from_config(cls, *args, **kwargs): method from_pretrained (line 151) | def from_pretrained(cls, *args, **kwargs): class ControlNetXSAdapter (line 155) | class ControlNetXSAdapter(metaclass=DummyObject): method __init__ (line 158) | def __init__(self, *args, **kwargs): method from_config (line 162) | def from_config(cls, *args, **kwargs): method from_pretrained (line 166) | def from_pretrained(cls, *args, **kwargs): class DiTTransformer2DModel (line 170) | class DiTTransformer2DModel(metaclass=DummyObject): method __init__ (line 173) | def __init__(self, *args, **kwargs): method from_config (line 177) | def from_config(cls, *args, **kwargs): method from_pretrained (line 181) | def from_pretrained(cls, *args, **kwargs): class FluxTransformer2DModel (line 185) | class FluxTransformer2DModel(metaclass=DummyObject): method __init__ (line 188) | def __init__(self, *args, **kwargs): method from_config (line 192) | def from_config(cls, *args, **kwargs): method from_pretrained (line 196) | def from_pretrained(cls, *args, **kwargs): class HunyuanDiT2DControlNetModel (line 200) | class HunyuanDiT2DControlNetModel(metaclass=DummyObject): method __init__ (line 203) | def __init__(self, *args, **kwargs): method from_config (line 207) | def from_config(cls, *args, **kwargs): method from_pretrained (line 211) | def from_pretrained(cls, *args, **kwargs): class HunyuanDiT2DModel (line 215) | class HunyuanDiT2DModel(metaclass=DummyObject): method __init__ (line 218) | def __init__(self, *args, **kwargs): method from_config (line 222) | def from_config(cls, *args, **kwargs): method from_pretrained (line 226) | def from_pretrained(cls, *args, **kwargs): class HunyuanDiT2DMultiControlNetModel (line 230) | class HunyuanDiT2DMultiControlNetModel(metaclass=DummyObject): method __init__ (line 233) | def __init__(self, *args, **kwargs): method from_config (line 237) | def from_config(cls, *args, **kwargs): method from_pretrained (line 241) | def from_pretrained(cls, *args, **kwargs): class I2VGenXLUNet (line 245) | class I2VGenXLUNet(metaclass=DummyObject): method __init__ (line 248) | def __init__(self, *args, **kwargs): method from_config (line 252) | def from_config(cls, *args, **kwargs): method from_pretrained (line 256) | def from_pretrained(cls, *args, **kwargs): class Kandinsky3UNet (line 260) | class Kandinsky3UNet(metaclass=DummyObject): method __init__ (line 263) | def __init__(self, *args, **kwargs): method from_config (line 267) | def from_config(cls, *args, **kwargs): method from_pretrained (line 271) | def from_pretrained(cls, *args, **kwargs): class LatteTransformer3DModel (line 275) | class LatteTransformer3DModel(metaclass=DummyObject): method __init__ (line 278) | def __init__(self, *args, **kwargs): method from_config (line 282) | def from_config(cls, *args, **kwargs): method from_pretrained (line 286) | def from_pretrained(cls, *args, **kwargs): class LuminaNextDiT2DModel (line 290) | class LuminaNextDiT2DModel(metaclass=DummyObject): method __init__ (line 293) | def __init__(self, *args, **kwargs): method from_config (line 297) | def from_config(cls, *args, **kwargs): method from_pretrained (line 301) | def from_pretrained(cls, *args, **kwargs): class ModelMixin (line 305) | class ModelMixin(metaclass=DummyObject): method __init__ (line 308) | def __init__(self, *args, **kwargs): method from_config (line 312) | def from_config(cls, *args, **kwargs): method from_pretrained (line 316) | def from_pretrained(cls, *args, **kwargs): class MotionAdapter (line 320) | class MotionAdapter(metaclass=DummyObject): method __init__ (line 323) | def __init__(self, *args, **kwargs): method from_config (line 327) | def from_config(cls, *args, **kwargs): method from_pretrained (line 331) | def from_pretrained(cls, *args, **kwargs): class MultiAdapter (line 335) | class MultiAdapter(metaclass=DummyObject): method __init__ (line 338) | def __init__(self, *args, **kwargs): method from_config (line 342) | def from_config(cls, *args, **kwargs): method from_pretrained (line 346) | def from_pretrained(cls, *args, **kwargs): class PixArtTransformer2DModel (line 350) | class PixArtTransformer2DModel(metaclass=DummyObject): method __init__ (line 353) | def __init__(self, *args, **kwargs): method from_config (line 357) | def from_config(cls, *args, **kwargs): method from_pretrained (line 361) | def from_pretrained(cls, *args, **kwargs): class PriorTransformer (line 365) | class PriorTransformer(metaclass=DummyObject): method __init__ (line 368) | def __init__(self, *args, **kwargs): method from_config (line 372) | def from_config(cls, *args, **kwargs): method from_pretrained (line 376) | def from_pretrained(cls, *args, **kwargs): class SD3ControlNetModel (line 380) | class SD3ControlNetModel(metaclass=DummyObject): method __init__ (line 383) | def __init__(self, *args, **kwargs): method from_config (line 387) | def from_config(cls, *args, **kwargs): method from_pretrained (line 391) | def from_pretrained(cls, *args, **kwargs): class SD3MultiControlNetModel (line 395) | class SD3MultiControlNetModel(metaclass=DummyObject): method __init__ (line 398) | def __init__(self, *args, **kwargs): method from_config (line 402) | def from_config(cls, *args, **kwargs): method from_pretrained (line 406) | def from_pretrained(cls, *args, **kwargs): class SD3Transformer2DModel (line 410) | class SD3Transformer2DModel(metaclass=DummyObject): method __init__ (line 413) | def __init__(self, *args, **kwargs): method from_config (line 417) | def from_config(cls, *args, **kwargs): method from_pretrained (line 421) | def from_pretrained(cls, *args, **kwargs): class SparseControlNetModel (line 425) | class SparseControlNetModel(metaclass=DummyObject): method __init__ (line 428) | def __init__(self, *args, **kwargs): method from_config (line 432) | def from_config(cls, *args, **kwargs): method from_pretrained (line 436) | def from_pretrained(cls, *args, **kwargs): class StableAudioDiTModel (line 440) | class StableAudioDiTModel(metaclass=DummyObject): method __init__ (line 443) | def __init__(self, *args, **kwargs): method from_config (line 447) | def from_config(cls, *args, **kwargs): method from_pretrained (line 451) | def from_pretrained(cls, *args, **kwargs): class T2IAdapter (line 455) | class T2IAdapter(metaclass=DummyObject): method __init__ (line 458) | def __init__(self, *args, **kwargs): method from_config (line 462) | def from_config(cls, *args, **kwargs): method from_pretrained (line 466) | def from_pretrained(cls, *args, **kwargs): class T5FilmDecoder (line 470) | class T5FilmDecoder(metaclass=DummyObject): method __init__ (line 473) | def __init__(self, *args, **kwargs): method from_config (line 477) | def from_config(cls, *args, **kwargs): method from_pretrained (line 481) | def from_pretrained(cls, *args, **kwargs): class Transformer2DModel (line 485) | class Transformer2DModel(metaclass=DummyObject): method __init__ (line 488) | def __init__(self, *args, **kwargs): method from_config (line 492) | def from_config(cls, *args, **kwargs): method from_pretrained (line 496) | def from_pretrained(cls, *args, **kwargs): class UNet1DModel (line 500) | class UNet1DModel(metaclass=DummyObject): method __init__ (line 503) | def __init__(self, *args, **kwargs): method from_config (line 507) | def from_config(cls, *args, **kwargs): method from_pretrained (line 511) | def from_pretrained(cls, *args, **kwargs): class UNet2DConditionModel (line 515) | class UNet2DConditionModel(metaclass=DummyObject): method __init__ (line 518) | def __init__(self, *args, **kwargs): method from_config (line 522) | def from_config(cls, *args, **kwargs): method from_pretrained (line 526) | def from_pretrained(cls, *args, **kwargs): class UNet2DModel (line 530) | class UNet2DModel(metaclass=DummyObject): method __init__ (line 533) | def __init__(self, *args, **kwargs): method from_config (line 537) | def from_config(cls, *args, **kwargs): method from_pretrained (line 541) | def from_pretrained(cls, *args, **kwargs): class UNet3DConditionModel (line 545) | class UNet3DConditionModel(metaclass=DummyObject): method __init__ (line 548) | def __init__(self, *args, **kwargs): method from_config (line 552) | def from_config(cls, *args, **kwargs): method from_pretrained (line 556) | def from_pretrained(cls, *args, **kwargs): class UNetControlNetXSModel (line 560) | class UNetControlNetXSModel(metaclass=DummyObject): method __init__ (line 563) | def __init__(self, *args, **kwargs): method from_config (line 567) | def from_config(cls, *args, **kwargs): method from_pretrained (line 571) | def from_pretrained(cls, *args, **kwargs): class UNetMotionModel (line 575) | class UNetMotionModel(metaclass=DummyObject): method __init__ (line 578) | def __init__(self, *args, **kwargs): method from_config (line 582) | def from_config(cls, *args, **kwargs): method from_pretrained (line 586) | def from_pretrained(cls, *args, **kwargs): class UNetSpatioTemporalConditionModel (line 590) | class UNetSpatioTemporalConditionModel(metaclass=DummyObject): method __init__ (line 593) | def __init__(self, *args, **kwargs): method from_config (line 597) | def from_config(cls, *args, **kwargs): method from_pretrained (line 601) | def from_pretrained(cls, *args, **kwargs): class UVit2DModel (line 605) | class UVit2DModel(metaclass=DummyObject): method __init__ (line 608) | def __init__(self, *args, **kwargs): method from_config (line 612) | def from_config(cls, *args, **kwargs): method from_pretrained (line 616) | def from_pretrained(cls, *args, **kwargs): class VQModel (line 620) | class VQModel(metaclass=DummyObject): method __init__ (line 623) | def __init__(self, *args, **kwargs): method from_config (line 627) | def from_config(cls, *args, **kwargs): method from_pretrained (line 631) | def from_pretrained(cls, *args, **kwargs): function get_constant_schedule (line 635) | def get_constant_schedule(*args, **kwargs): function get_constant_schedule_with_warmup (line 639) | def get_constant_schedule_with_warmup(*args, **kwargs): function get_cosine_schedule_with_warmup (line 643) | def get_cosine_schedule_with_warmup(*args, **kwargs): function get_cosine_with_hard_restarts_schedule_with_warmup (line 647) | def get_cosine_with_hard_restarts_schedule_with_warmup(*args, **kwargs): function get_linear_schedule_with_warmup (line 651) | def get_linear_schedule_with_warmup(*args, **kwargs): function get_polynomial_decay_schedule_with_warmup (line 655) | def get_polynomial_decay_schedule_with_warmup(*args, **kwargs): function get_scheduler (line 659) | def get_scheduler(*args, **kwargs): class AudioPipelineOutput (line 663) | class AudioPipelineOutput(metaclass=DummyObject): method __init__ (line 666) | def __init__(self, *args, **kwargs): method from_config (line 670) | def from_config(cls, *args, **kwargs): method from_pretrained (line 674) | def from_pretrained(cls, *args, **kwargs): class AutoPipelineForImage2Image (line 678) | class AutoPipelineForImage2Image(metaclass=DummyObject): method __init__ (line 681) | def __init__(self, *args, **kwargs): method from_config (line 685) | def from_config(cls, *args, **kwargs): method from_pretrained (line 689) | def from_pretrained(cls, *args, **kwargs): class AutoPipelineForInpainting (line 693) | class AutoPipelineForInpainting(metaclass=DummyObject): method __init__ (line 696) | def __init__(self, *args, **kwargs): method from_config (line 700) | def from_config(cls, *args, **kwargs): method from_pretrained (line 704) | def from_pretrained(cls, *args, **kwargs): class AutoPipelineForText2Image (line 708) | class AutoPipelineForText2Image(metaclass=DummyObject): method __init__ (line 711) | def __init__(self, *args, **kwargs): method from_config (line 715) | def from_config(cls, *args, **kwargs): method from_pretrained (line 719) | def from_pretrained(cls, *args, **kwargs): class BlipDiffusionControlNetPipeline (line 723) | class BlipDiffusionControlNetPipeline(metaclass=DummyObject): method __init__ (line 726) | def __init__(self, *args, **kwargs): method from_config (line 730) | def from_config(cls, *args, **kwargs): method from_pretrained (line 734) | def from_pretrained(cls, *args, **kwargs): class BlipDiffusionPipeline (line 738) | class BlipDiffusionPipeline(metaclass=DummyObject): method __init__ (line 741) | def __init__(self, *args, **kwargs): method from_config (line 745) | def from_config(cls, *args, **kwargs): method from_pretrained (line 749) | def from_pretrained(cls, *args, **kwargs): class CLIPImageProjection (line 753) | class CLIPImageProjection(metaclass=DummyObject): method __init__ (line 756) | def __init__(self, *args, **kwargs): method from_config (line 760) | def from_config(cls, *args, **kwargs): method from_pretrained (line 764) | def from_pretrained(cls, *args, **kwargs): class ConsistencyModelPipeline (line 768) | class ConsistencyModelPipeline(metaclass=DummyObject): method __init__ (line 771) | def __init__(self, *args, **kwargs): method from_config (line 775) | def from_config(cls, *args, **kwargs): method from_pretrained (line 779) | def from_pretrained(cls, *args, **kwargs): class DanceDiffusionPipeline (line 783) | class DanceDiffusionPipeline(metaclass=DummyObject): method __init__ (line 786) | def __init__(self, *args, **kwargs): method from_config (line 790) | def from_config(cls, *args, **kwargs): method from_pretrained (line 794) | def from_pretrained(cls, *args, **kwargs): class DDIMPipeline (line 798) | class DDIMPipeline(metaclass=DummyObject): method __init__ (line 801) | def __init__(self, *args, **kwargs): method from_config (line 805) | def from_config(cls, *args, **kwargs): method from_pretrained (line 809) | def from_pretrained(cls, *args, **kwargs): class DDPMPipeline (line 813) | class DDPMPipeline(metaclass=DummyObject): method __init__ (line 816) | def __init__(self, *args, **kwargs): method from_config (line 820) | def from_config(cls, *args, **kwargs): method from_pretrained (line 824) | def from_pretrained(cls, *args, **kwargs): class DiffusionPipeline (line 828) | class DiffusionPipeline(metaclass=DummyObject): method __init__ (line 831) | def __init__(self, *args, **kwargs): method from_config (line 835) | def from_config(cls, *args, **kwargs): method from_pretrained (line 839) | def from_pretrained(cls, *args, **kwargs): class DiTPipeline (line 843) | class DiTPipeline(metaclass=DummyObject): method __init__ (line 846) | def __init__(self, *args, **kwargs): method from_config (line 850) | def from_config(cls, *args, **kwargs): method from_pretrained (line 854) | def from_pretrained(cls, *args, **kwargs): class ImagePipelineOutput (line 858) | class ImagePipelineOutput(metaclass=DummyObject): method __init__ (line 861) | def __init__(self, *args, **kwargs): method from_config (line 865) | def from_config(cls, *args, **kwargs): method from_pretrained (line 869) | def from_pretrained(cls, *args, **kwargs): class KarrasVePipeline (line 873) | class KarrasVePipeline(metaclass=DummyObject): method __init__ (line 876) | def __init__(self, *args, **kwargs): method from_config (line 880) | def from_config(cls, *args, **kwargs): method from_pretrained (line 884) | def from_pretrained(cls, *args, **kwargs): class LDMPipeline (line 888) | class LDMPipeline(metaclass=DummyObject): method __init__ (line 891) | def __init__(self, *args, **kwargs): method from_config (line 895) | def from_config(cls, *args, **kwargs): method from_pretrained (line 899) | def from_pretrained(cls, *args, **kwargs): class LDMSuperResolutionPipeline (line 903) | class LDMSuperResolutionPipeline(metaclass=DummyObject): method __init__ (line 906) | def __init__(self, *args, **kwargs): method from_config (line 910) | def from_config(cls, *args, **kwargs): method from_pretrained (line 914) | def from_pretrained(cls, *args, **kwargs): class PNDMPipeline (line 918) | class PNDMPipeline(metaclass=DummyObject): method __init__ (line 921) | def __init__(self, *args, **kwargs): method from_config (line 925) | def from_config(cls, *args, **kwargs): method from_pretrained (line 929) | def from_pretrained(cls, *args, **kwargs): class RePaintPipeline (line 933) | class RePaintPipeline(metaclass=DummyObject): method __init__ (line 936) | def __init__(self, *args, **kwargs): method from_config (line 940) | def from_config(cls, *args, **kwargs): method from_pretrained (line 944) | def from_pretrained(cls, *args, **kwargs): class ScoreSdeVePipeline (line 948) | class ScoreSdeVePipeline(metaclass=DummyObject): method __init__ (line 951) | def __init__(self, *args, **kwargs): method from_config (line 955) | def from_config(cls, *args, **kwargs): method from_pretrained (line 959) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionMixin (line 963) | class StableDiffusionMixin(metaclass=DummyObject): method __init__ (line 966) | def __init__(self, *args, **kwargs): method from_config (line 970) | def from_config(cls, *args, **kwargs): method from_pretrained (line 974) | def from_pretrained(cls, *args, **kwargs): class AmusedScheduler (line 978) | class AmusedScheduler(metaclass=DummyObject): method __init__ (line 981) | def __init__(self, *args, **kwargs): method from_config (line 985) | def from_config(cls, *args, **kwargs): method from_pretrained (line 989) | def from_pretrained(cls, *args, **kwargs): class CMStochasticIterativeScheduler (line 993) | class CMStochasticIterativeScheduler(metaclass=DummyObject): method __init__ (line 996) | def __init__(self, *args, **kwargs): method from_config (line 1000) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1004) | def from_pretrained(cls, *args, **kwargs): class CogVideoXDDIMScheduler (line 1008) | class CogVideoXDDIMScheduler(metaclass=DummyObject): method __init__ (line 1011) | def __init__(self, *args, **kwargs): method from_config (line 1015) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1019) | def from_pretrained(cls, *args, **kwargs): class CogVideoXDPMScheduler (line 1023) | class CogVideoXDPMScheduler(metaclass=DummyObject): method __init__ (line 1026) | def __init__(self, *args, **kwargs): method from_config (line 1030) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1034) | def from_pretrained(cls, *args, **kwargs): class DDIMInverseScheduler (line 1038) | class DDIMInverseScheduler(metaclass=DummyObject): method __init__ (line 1041) | def __init__(self, *args, **kwargs): method from_config (line 1045) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1049) | def from_pretrained(cls, *args, **kwargs): class DDIMParallelScheduler (line 1053) | class DDIMParallelScheduler(metaclass=DummyObject): method __init__ (line 1056) | def __init__(self, *args, **kwargs): method from_config (line 1060) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1064) | def from_pretrained(cls, *args, **kwargs): class DDIMScheduler (line 1068) | class DDIMScheduler(metaclass=DummyObject): method __init__ (line 1071) | def __init__(self, *args, **kwargs): method from_config (line 1075) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1079) | def from_pretrained(cls, *args, **kwargs): class DDPMParallelScheduler (line 1083) | class DDPMParallelScheduler(metaclass=DummyObject): method __init__ (line 1086) | def __init__(self, *args, **kwargs): method from_config (line 1090) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1094) | def from_pretrained(cls, *args, **kwargs): class DDPMScheduler (line 1098) | class DDPMScheduler(metaclass=DummyObject): method __init__ (line 1101) | def __init__(self, *args, **kwargs): method from_config (line 1105) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1109) | def from_pretrained(cls, *args, **kwargs): class DDPMWuerstchenScheduler (line 1113) | class DDPMWuerstchenScheduler(metaclass=DummyObject): method __init__ (line 1116) | def __init__(self, *args, **kwargs): method from_config (line 1120) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1124) | def from_pretrained(cls, *args, **kwargs): class DEISMultistepScheduler (line 1128) | class DEISMultistepScheduler(metaclass=DummyObject): method __init__ (line 1131) | def __init__(self, *args, **kwargs): method from_config (line 1135) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1139) | def from_pretrained(cls, *args, **kwargs): class DPMSolverMultistepInverseScheduler (line 1143) | class DPMSolverMultistepInverseScheduler(metaclass=DummyObject): method __init__ (line 1146) | def __init__(self, *args, **kwargs): method from_config (line 1150) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1154) | def from_pretrained(cls, *args, **kwargs): class DPMSolverMultistepScheduler (line 1158) | class DPMSolverMultistepScheduler(metaclass=DummyObject): method __init__ (line 1161) | def __init__(self, *args, **kwargs): method from_config (line 1165) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1169) | def from_pretrained(cls, *args, **kwargs): class DPMSolverSinglestepScheduler (line 1173) | class DPMSolverSinglestepScheduler(metaclass=DummyObject): method __init__ (line 1176) | def __init__(self, *args, **kwargs): method from_config (line 1180) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1184) | def from_pretrained(cls, *args, **kwargs): class EDMDPMSolverMultistepScheduler (line 1188) | class EDMDPMSolverMultistepScheduler(metaclass=DummyObject): method __init__ (line 1191) | def __init__(self, *args, **kwargs): method from_config (line 1195) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1199) | def from_pretrained(cls, *args, **kwargs): class EDMEulerScheduler (line 1203) | class EDMEulerScheduler(metaclass=DummyObject): method __init__ (line 1206) | def __init__(self, *args, **kwargs): method from_config (line 1210) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1214) | def from_pretrained(cls, *args, **kwargs): class EulerAncestralDiscreteScheduler (line 1218) | class EulerAncestralDiscreteScheduler(metaclass=DummyObject): method __init__ (line 1221) | def __init__(self, *args, **kwargs): method from_config (line 1225) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1229) | def from_pretrained(cls, *args, **kwargs): class EulerDiscreteScheduler (line 1233) | class EulerDiscreteScheduler(metaclass=DummyObject): method __init__ (line 1236) | def __init__(self, *args, **kwargs): method from_config (line 1240) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1244) | def from_pretrained(cls, *args, **kwargs): class FlowMatchEulerDiscreteScheduler (line 1248) | class FlowMatchEulerDiscreteScheduler(metaclass=DummyObject): method __init__ (line 1251) | def __init__(self, *args, **kwargs): method from_config (line 1255) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1259) | def from_pretrained(cls, *args, **kwargs): class FlowMatchHeunDiscreteScheduler (line 1263) | class FlowMatchHeunDiscreteScheduler(metaclass=DummyObject): method __init__ (line 1266) | def __init__(self, *args, **kwargs): method from_config (line 1270) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1274) | def from_pretrained(cls, *args, **kwargs): class HeunDiscreteScheduler (line 1278) | class HeunDiscreteScheduler(metaclass=DummyObject): method __init__ (line 1281) | def __init__(self, *args, **kwargs): method from_config (line 1285) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1289) | def from_pretrained(cls, *args, **kwargs): class IPNDMScheduler (line 1293) | class IPNDMScheduler(metaclass=DummyObject): method __init__ (line 1296) | def __init__(self, *args, **kwargs): method from_config (line 1300) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1304) | def from_pretrained(cls, *args, **kwargs): class KarrasVeScheduler (line 1308) | class KarrasVeScheduler(metaclass=DummyObject): method __init__ (line 1311) | def __init__(self, *args, **kwargs): method from_config (line 1315) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1319) | def from_pretrained(cls, *args, **kwargs): class KDPM2AncestralDiscreteScheduler (line 1323) | class KDPM2AncestralDiscreteScheduler(metaclass=DummyObject): method __init__ (line 1326) | def __init__(self, *args, **kwargs): method from_config (line 1330) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1334) | def from_pretrained(cls, *args, **kwargs): class KDPM2DiscreteScheduler (line 1338) | class KDPM2DiscreteScheduler(metaclass=DummyObject): method __init__ (line 1341) | def __init__(self, *args, **kwargs): method from_config (line 1345) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1349) | def from_pretrained(cls, *args, **kwargs): class LCMScheduler (line 1353) | class LCMScheduler(metaclass=DummyObject): method __init__ (line 1356) | def __init__(self, *args, **kwargs): method from_config (line 1360) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1364) | def from_pretrained(cls, *args, **kwargs): class PNDMScheduler (line 1368) | class PNDMScheduler(metaclass=DummyObject): method __init__ (line 1371) | def __init__(self, *args, **kwargs): method from_config (line 1375) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1379) | def from_pretrained(cls, *args, **kwargs): class RePaintScheduler (line 1383) | class RePaintScheduler(metaclass=DummyObject): method __init__ (line 1386) | def __init__(self, *args, **kwargs): method from_config (line 1390) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1394) | def from_pretrained(cls, *args, **kwargs): class SASolverScheduler (line 1398) | class SASolverScheduler(metaclass=DummyObject): method __init__ (line 1401) | def __init__(self, *args, **kwargs): method from_config (line 1405) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1409) | def from_pretrained(cls, *args, **kwargs): class SchedulerMixin (line 1413) | class SchedulerMixin(metaclass=DummyObject): method __init__ (line 1416) | def __init__(self, *args, **kwargs): method from_config (line 1420) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1424) | def from_pretrained(cls, *args, **kwargs): class ScoreSdeVeScheduler (line 1428) | class ScoreSdeVeScheduler(metaclass=DummyObject): method __init__ (line 1431) | def __init__(self, *args, **kwargs): method from_config (line 1435) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1439) | def from_pretrained(cls, *args, **kwargs): class TCDScheduler (line 1443) | class TCDScheduler(metaclass=DummyObject): method __init__ (line 1446) | def __init__(self, *args, **kwargs): method from_config (line 1450) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1454) | def from_pretrained(cls, *args, **kwargs): class UnCLIPScheduler (line 1458) | class UnCLIPScheduler(metaclass=DummyObject): method __init__ (line 1461) | def __init__(self, *args, **kwargs): method from_config (line 1465) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1469) | def from_pretrained(cls, *args, **kwargs): class UniPCMultistepScheduler (line 1473) | class UniPCMultistepScheduler(metaclass=DummyObject): method __init__ (line 1476) | def __init__(self, *args, **kwargs): method from_config (line 1480) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1484) | def from_pretrained(cls, *args, **kwargs): class VQDiffusionScheduler (line 1488) | class VQDiffusionScheduler(metaclass=DummyObject): method __init__ (line 1491) | def __init__(self, *args, **kwargs): method from_config (line 1495) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1499) | def from_pretrained(cls, *args, **kwargs): class EMAModel (line 1503) | class EMAModel(metaclass=DummyObject): method __init__ (line 1506) | def __init__(self, *args, **kwargs): method from_config (line 1510) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1514) | def from_pretrained(cls, *args, **kwargs): FILE: src/diffusers/utils/dummy_torch_and_librosa_objects.py class AudioDiffusionPipeline (line 5) | class AudioDiffusionPipeline(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class Mel (line 20) | class Mel(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): FILE: src/diffusers/utils/dummy_torch_and_scipy_objects.py class LMSDiscreteScheduler (line 5) | class LMSDiscreteScheduler(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): FILE: src/diffusers/utils/dummy_torch_and_torchsde_objects.py class CosineDPMSolverMultistepScheduler (line 5) | class CosineDPMSolverMultistepScheduler(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class DPMSolverSDEScheduler (line 20) | class DPMSolverSDEScheduler(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): FILE: src/diffusers/utils/dummy_torch_and_transformers_and_k_diffusion_objects.py class StableDiffusionKDiffusionPipeline (line 5) | class StableDiffusionKDiffusionPipeline(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLKDiffusionPipeline (line 20) | class StableDiffusionXLKDiffusionPipeline(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): FILE: src/diffusers/utils/dummy_torch_and_transformers_and_onnx_objects.py class OnnxStableDiffusionImg2ImgPipeline (line 5) | class OnnxStableDiffusionImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class OnnxStableDiffusionInpaintPipeline (line 20) | class OnnxStableDiffusionInpaintPipeline(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): class OnnxStableDiffusionInpaintPipelineLegacy (line 35) | class OnnxStableDiffusionInpaintPipelineLegacy(metaclass=DummyObject): method __init__ (line 38) | def __init__(self, *args, **kwargs): method from_config (line 42) | def from_config(cls, *args, **kwargs): method from_pretrained (line 46) | def from_pretrained(cls, *args, **kwargs): class OnnxStableDiffusionPipeline (line 50) | class OnnxStableDiffusionPipeline(metaclass=DummyObject): method __init__ (line 53) | def __init__(self, *args, **kwargs): method from_config (line 57) | def from_config(cls, *args, **kwargs): method from_pretrained (line 61) | def from_pretrained(cls, *args, **kwargs): class OnnxStableDiffusionUpscalePipeline (line 65) | class OnnxStableDiffusionUpscalePipeline(metaclass=DummyObject): method __init__ (line 68) | def __init__(self, *args, **kwargs): method from_config (line 72) | def from_config(cls, *args, **kwargs): method from_pretrained (line 76) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionOnnxPipeline (line 80) | class StableDiffusionOnnxPipeline(metaclass=DummyObject): method __init__ (line 83) | def __init__(self, *args, **kwargs): method from_config (line 87) | def from_config(cls, *args, **kwargs): method from_pretrained (line 91) | def from_pretrained(cls, *args, **kwargs): FILE: src/diffusers/utils/dummy_torch_and_transformers_and_sentencepiece_objects.py class KolorsImg2ImgPipeline (line 5) | class KolorsImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class KolorsPAGPipeline (line 20) | class KolorsPAGPipeline(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): class KolorsPipeline (line 35) | class KolorsPipeline(metaclass=DummyObject): method __init__ (line 38) | def __init__(self, *args, **kwargs): method from_config (line 42) | def from_config(cls, *args, **kwargs): method from_pretrained (line 46) | def from_pretrained(cls, *args, **kwargs): FILE: src/diffusers/utils/dummy_torch_and_transformers_objects.py class AltDiffusionImg2ImgPipeline (line 5) | class AltDiffusionImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class AltDiffusionPipeline (line 20) | class AltDiffusionPipeline(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): class AmusedImg2ImgPipeline (line 35) | class AmusedImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 38) | def __init__(self, *args, **kwargs): method from_config (line 42) | def from_config(cls, *args, **kwargs): method from_pretrained (line 46) | def from_pretrained(cls, *args, **kwargs): class AmusedInpaintPipeline (line 50) | class AmusedInpaintPipeline(metaclass=DummyObject): method __init__ (line 53) | def __init__(self, *args, **kwargs): method from_config (line 57) | def from_config(cls, *args, **kwargs): method from_pretrained (line 61) | def from_pretrained(cls, *args, **kwargs): class AmusedPipeline (line 65) | class AmusedPipeline(metaclass=DummyObject): method __init__ (line 68) | def __init__(self, *args, **kwargs): method from_config (line 72) | def from_config(cls, *args, **kwargs): method from_pretrained (line 76) | def from_pretrained(cls, *args, **kwargs): class AnimateDiffControlNetPipeline (line 80) | class AnimateDiffControlNetPipeline(metaclass=DummyObject): method __init__ (line 83) | def __init__(self, *args, **kwargs): method from_config (line 87) | def from_config(cls, *args, **kwargs): method from_pretrained (line 91) | def from_pretrained(cls, *args, **kwargs): class AnimateDiffPAGPipeline (line 95) | class AnimateDiffPAGPipeline(metaclass=DummyObject): method __init__ (line 98) | def __init__(self, *args, **kwargs): method from_config (line 102) | def from_config(cls, *args, **kwargs): method from_pretrained (line 106) | def from_pretrained(cls, *args, **kwargs): class AnimateDiffPipeline (line 110) | class AnimateDiffPipeline(metaclass=DummyObject): method __init__ (line 113) | def __init__(self, *args, **kwargs): method from_config (line 117) | def from_config(cls, *args, **kwargs): method from_pretrained (line 121) | def from_pretrained(cls, *args, **kwargs): class AnimateDiffSDXLPipeline (line 125) | class AnimateDiffSDXLPipeline(metaclass=DummyObject): method __init__ (line 128) | def __init__(self, *args, **kwargs): method from_config (line 132) | def from_config(cls, *args, **kwargs): method from_pretrained (line 136) | def from_pretrained(cls, *args, **kwargs): class AnimateDiffSparseControlNetPipeline (line 140) | class AnimateDiffSparseControlNetPipeline(metaclass=DummyObject): method __init__ (line 143) | def __init__(self, *args, **kwargs): method from_config (line 147) | def from_config(cls, *args, **kwargs): method from_pretrained (line 151) | def from_pretrained(cls, *args, **kwargs): class AnimateDiffVideoToVideoPipeline (line 155) | class AnimateDiffVideoToVideoPipeline(metaclass=DummyObject): method __init__ (line 158) | def __init__(self, *args, **kwargs): method from_config (line 162) | def from_config(cls, *args, **kwargs): method from_pretrained (line 166) | def from_pretrained(cls, *args, **kwargs): class AudioLDM2Pipeline (line 170) | class AudioLDM2Pipeline(metaclass=DummyObject): method __init__ (line 173) | def __init__(self, *args, **kwargs): method from_config (line 177) | def from_config(cls, *args, **kwargs): method from_pretrained (line 181) | def from_pretrained(cls, *args, **kwargs): class AudioLDM2ProjectionModel (line 185) | class AudioLDM2ProjectionModel(metaclass=DummyObject): method __init__ (line 188) | def __init__(self, *args, **kwargs): method from_config (line 192) | def from_config(cls, *args, **kwargs): method from_pretrained (line 196) | def from_pretrained(cls, *args, **kwargs): class AudioLDM2UNet2DConditionModel (line 200) | class AudioLDM2UNet2DConditionModel(metaclass=DummyObject): method __init__ (line 203) | def __init__(self, *args, **kwargs): method from_config (line 207) | def from_config(cls, *args, **kwargs): method from_pretrained (line 211) | def from_pretrained(cls, *args, **kwargs): class AudioLDMPipeline (line 215) | class AudioLDMPipeline(metaclass=DummyObject): method __init__ (line 218) | def __init__(self, *args, **kwargs): method from_config (line 222) | def from_config(cls, *args, **kwargs): method from_pretrained (line 226) | def from_pretrained(cls, *args, **kwargs): class AuraFlowPipeline (line 230) | class AuraFlowPipeline(metaclass=DummyObject): method __init__ (line 233) | def __init__(self, *args, **kwargs): method from_config (line 237) | def from_config(cls, *args, **kwargs): method from_pretrained (line 241) | def from_pretrained(cls, *args, **kwargs): class CLIPImageProjection (line 245) | class CLIPImageProjection(metaclass=DummyObject): method __init__ (line 248) | def __init__(self, *args, **kwargs): method from_config (line 252) | def from_config(cls, *args, **kwargs): method from_pretrained (line 256) | def from_pretrained(cls, *args, **kwargs): class CogVideoXPipeline (line 260) | class CogVideoXPipeline(metaclass=DummyObject): method __init__ (line 263) | def __init__(self, *args, **kwargs): method from_config (line 267) | def from_config(cls, *args, **kwargs): method from_pretrained (line 271) | def from_pretrained(cls, *args, **kwargs): class CycleDiffusionPipeline (line 275) | class CycleDiffusionPipeline(metaclass=DummyObject): method __init__ (line 278) | def __init__(self, *args, **kwargs): method from_config (line 282) | def from_config(cls, *args, **kwargs): method from_pretrained (line 286) | def from_pretrained(cls, *args, **kwargs): class FluxPipeline (line 290) | class FluxPipeline(metaclass=DummyObject): method __init__ (line 293) | def __init__(self, *args, **kwargs): method from_config (line 297) | def from_config(cls, *args, **kwargs): method from_pretrained (line 301) | def from_pretrained(cls, *args, **kwargs): class HunyuanDiTControlNetPipeline (line 305) | class HunyuanDiTControlNetPipeline(metaclass=DummyObject): method __init__ (line 308) | def __init__(self, *args, **kwargs): method from_config (line 312) | def from_config(cls, *args, **kwargs): method from_pretrained (line 316) | def from_pretrained(cls, *args, **kwargs): class HunyuanDiTPAGPipeline (line 320) | class HunyuanDiTPAGPipeline(metaclass=DummyObject): method __init__ (line 323) | def __init__(self, *args, **kwargs): method from_config (line 327) | def from_config(cls, *args, **kwargs): method from_pretrained (line 331) | def from_pretrained(cls, *args, **kwargs): class HunyuanDiTPipeline (line 335) | class HunyuanDiTPipeline(metaclass=DummyObject): method __init__ (line 338) | def __init__(self, *args, **kwargs): method from_config (line 342) | def from_config(cls, *args, **kwargs): method from_pretrained (line 346) | def from_pretrained(cls, *args, **kwargs): class I2VGenXLPipeline (line 350) | class I2VGenXLPipeline(metaclass=DummyObject): method __init__ (line 353) | def __init__(self, *args, **kwargs): method from_config (line 357) | def from_config(cls, *args, **kwargs): method from_pretrained (line 361) | def from_pretrained(cls, *args, **kwargs): class IFImg2ImgPipeline (line 365) | class IFImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 368) | def __init__(self, *args, **kwargs): method from_config (line 372) | def from_config(cls, *args, **kwargs): method from_pretrained (line 376) | def from_pretrained(cls, *args, **kwargs): class IFImg2ImgSuperResolutionPipeline (line 380) | class IFImg2ImgSuperResolutionPipeline(metaclass=DummyObject): method __init__ (line 383) | def __init__(self, *args, **kwargs): method from_config (line 387) | def from_config(cls, *args, **kwargs): method from_pretrained (line 391) | def from_pretrained(cls, *args, **kwargs): class IFInpaintingPipeline (line 395) | class IFInpaintingPipeline(metaclass=DummyObject): method __init__ (line 398) | def __init__(self, *args, **kwargs): method from_config (line 402) | def from_config(cls, *args, **kwargs): method from_pretrained (line 406) | def from_pretrained(cls, *args, **kwargs): class IFInpaintingSuperResolutionPipeline (line 410) | class IFInpaintingSuperResolutionPipeline(metaclass=DummyObject): method __init__ (line 413) | def __init__(self, *args, **kwargs): method from_config (line 417) | def from_config(cls, *args, **kwargs): method from_pretrained (line 421) | def from_pretrained(cls, *args, **kwargs): class IFPipeline (line 425) | class IFPipeline(metaclass=DummyObject): method __init__ (line 428) | def __init__(self, *args, **kwargs): method from_config (line 432) | def from_config(cls, *args, **kwargs): method from_pretrained (line 436) | def from_pretrained(cls, *args, **kwargs): class IFSuperResolutionPipeline (line 440) | class IFSuperResolutionPipeline(metaclass=DummyObject): method __init__ (line 443) | def __init__(self, *args, **kwargs): method from_config (line 447) | def from_config(cls, *args, **kwargs): method from_pretrained (line 451) | def from_pretrained(cls, *args, **kwargs): class ImageTextPipelineOutput (line 455) | class ImageTextPipelineOutput(metaclass=DummyObject): method __init__ (line 458) | def __init__(self, *args, **kwargs): method from_config (line 462) | def from_config(cls, *args, **kwargs): method from_pretrained (line 466) | def from_pretrained(cls, *args, **kwargs): class Kandinsky3Img2ImgPipeline (line 470) | class Kandinsky3Img2ImgPipeline(metaclass=DummyObject): method __init__ (line 473) | def __init__(self, *args, **kwargs): method from_config (line 477) | def from_config(cls, *args, **kwargs): method from_pretrained (line 481) | def from_pretrained(cls, *args, **kwargs): class Kandinsky3Pipeline (line 485) | class Kandinsky3Pipeline(metaclass=DummyObject): method __init__ (line 488) | def __init__(self, *args, **kwargs): method from_config (line 492) | def from_config(cls, *args, **kwargs): method from_pretrained (line 496) | def from_pretrained(cls, *args, **kwargs): class KandinskyCombinedPipeline (line 500) | class KandinskyCombinedPipeline(metaclass=DummyObject): method __init__ (line 503) | def __init__(self, *args, **kwargs): method from_config (line 507) | def from_config(cls, *args, **kwargs): method from_pretrained (line 511) | def from_pretrained(cls, *args, **kwargs): class KandinskyImg2ImgCombinedPipeline (line 515) | class KandinskyImg2ImgCombinedPipeline(metaclass=DummyObject): method __init__ (line 518) | def __init__(self, *args, **kwargs): method from_config (line 522) | def from_config(cls, *args, **kwargs): method from_pretrained (line 526) | def from_pretrained(cls, *args, **kwargs): class KandinskyImg2ImgPipeline (line 530) | class KandinskyImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 533) | def __init__(self, *args, **kwargs): method from_config (line 537) | def from_config(cls, *args, **kwargs): method from_pretrained (line 541) | def from_pretrained(cls, *args, **kwargs): class KandinskyInpaintCombinedPipeline (line 545) | class KandinskyInpaintCombinedPipeline(metaclass=DummyObject): method __init__ (line 548) | def __init__(self, *args, **kwargs): method from_config (line 552) | def from_config(cls, *args, **kwargs): method from_pretrained (line 556) | def from_pretrained(cls, *args, **kwargs): class KandinskyInpaintPipeline (line 560) | class KandinskyInpaintPipeline(metaclass=DummyObject): method __init__ (line 563) | def __init__(self, *args, **kwargs): method from_config (line 567) | def from_config(cls, *args, **kwargs): method from_pretrained (line 571) | def from_pretrained(cls, *args, **kwargs): class KandinskyPipeline (line 575) | class KandinskyPipeline(metaclass=DummyObject): method __init__ (line 578) | def __init__(self, *args, **kwargs): method from_config (line 582) | def from_config(cls, *args, **kwargs): method from_pretrained (line 586) | def from_pretrained(cls, *args, **kwargs): class KandinskyPriorPipeline (line 590) | class KandinskyPriorPipeline(metaclass=DummyObject): method __init__ (line 593) | def __init__(self, *args, **kwargs): method from_config (line 597) | def from_config(cls, *args, **kwargs): method from_pretrained (line 601) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22CombinedPipeline (line 605) | class KandinskyV22CombinedPipeline(metaclass=DummyObject): method __init__ (line 608) | def __init__(self, *args, **kwargs): method from_config (line 612) | def from_config(cls, *args, **kwargs): method from_pretrained (line 616) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22ControlnetImg2ImgPipeline (line 620) | class KandinskyV22ControlnetImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 623) | def __init__(self, *args, **kwargs): method from_config (line 627) | def from_config(cls, *args, **kwargs): method from_pretrained (line 631) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22ControlnetPipeline (line 635) | class KandinskyV22ControlnetPipeline(metaclass=DummyObject): method __init__ (line 638) | def __init__(self, *args, **kwargs): method from_config (line 642) | def from_config(cls, *args, **kwargs): method from_pretrained (line 646) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22Img2ImgCombinedPipeline (line 650) | class KandinskyV22Img2ImgCombinedPipeline(metaclass=DummyObject): method __init__ (line 653) | def __init__(self, *args, **kwargs): method from_config (line 657) | def from_config(cls, *args, **kwargs): method from_pretrained (line 661) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22Img2ImgPipeline (line 665) | class KandinskyV22Img2ImgPipeline(metaclass=DummyObject): method __init__ (line 668) | def __init__(self, *args, **kwargs): method from_config (line 672) | def from_config(cls, *args, **kwargs): method from_pretrained (line 676) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22InpaintCombinedPipeline (line 680) | class KandinskyV22InpaintCombinedPipeline(metaclass=DummyObject): method __init__ (line 683) | def __init__(self, *args, **kwargs): method from_config (line 687) | def from_config(cls, *args, **kwargs): method from_pretrained (line 691) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22InpaintPipeline (line 695) | class KandinskyV22InpaintPipeline(metaclass=DummyObject): method __init__ (line 698) | def __init__(self, *args, **kwargs): method from_config (line 702) | def from_config(cls, *args, **kwargs): method from_pretrained (line 706) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22Pipeline (line 710) | class KandinskyV22Pipeline(metaclass=DummyObject): method __init__ (line 713) | def __init__(self, *args, **kwargs): method from_config (line 717) | def from_config(cls, *args, **kwargs): method from_pretrained (line 721) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22PriorEmb2EmbPipeline (line 725) | class KandinskyV22PriorEmb2EmbPipeline(metaclass=DummyObject): method __init__ (line 728) | def __init__(self, *args, **kwargs): method from_config (line 732) | def from_config(cls, *args, **kwargs): method from_pretrained (line 736) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22PriorPipeline (line 740) | class KandinskyV22PriorPipeline(metaclass=DummyObject): method __init__ (line 743) | def __init__(self, *args, **kwargs): method from_config (line 747) | def from_config(cls, *args, **kwargs): method from_pretrained (line 751) | def from_pretrained(cls, *args, **kwargs): class LatentConsistencyModelImg2ImgPipeline (line 755) | class LatentConsistencyModelImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 758) | def __init__(self, *args, **kwargs): method from_config (line 762) | def from_config(cls, *args, **kwargs): method from_pretrained (line 766) | def from_pretrained(cls, *args, **kwargs): class LatentConsistencyModelPipeline (line 770) | class LatentConsistencyModelPipeline(metaclass=DummyObject): method __init__ (line 773) | def __init__(self, *args, **kwargs): method from_config (line 777) | def from_config(cls, *args, **kwargs): method from_pretrained (line 781) | def from_pretrained(cls, *args, **kwargs): class LattePipeline (line 785) | class LattePipeline(metaclass=DummyObject): method __init__ (line 788) | def __init__(self, *args, **kwargs): method from_config (line 792) | def from_config(cls, *args, **kwargs): method from_pretrained (line 796) | def from_pretrained(cls, *args, **kwargs): class LDMTextToImagePipeline (line 800) | class LDMTextToImagePipeline(metaclass=DummyObject): method __init__ (line 803) | def __init__(self, *args, **kwargs): method from_config (line 807) | def from_config(cls, *args, **kwargs): method from_pretrained (line 811) | def from_pretrained(cls, *args, **kwargs): class LEditsPPPipelineStableDiffusion (line 815) | class LEditsPPPipelineStableDiffusion(metaclass=DummyObject): method __init__ (line 818) | def __init__(self, *args, **kwargs): method from_config (line 822) | def from_config(cls, *args, **kwargs): method from_pretrained (line 826) | def from_pretrained(cls, *args, **kwargs): class LEditsPPPipelineStableDiffusionXL (line 830) | class LEditsPPPipelineStableDiffusionXL(metaclass=DummyObject): method __init__ (line 833) | def __init__(self, *args, **kwargs): method from_config (line 837) | def from_config(cls, *args, **kwargs): method from_pretrained (line 841) | def from_pretrained(cls, *args, **kwargs): class LuminaText2ImgPipeline (line 845) | class LuminaText2ImgPipeline(metaclass=DummyObject): method __init__ (line 848) | def __init__(self, *args, **kwargs): method from_config (line 852) | def from_config(cls, *args, **kwargs): method from_pretrained (line 856) | def from_pretrained(cls, *args, **kwargs): class MarigoldDepthPipeline (line 860) | class MarigoldDepthPipeline(metaclass=DummyObject): method __init__ (line 863) | def __init__(self, *args, **kwargs): method from_config (line 867) | def from_config(cls, *args, **kwargs): method from_pretrained (line 871) | def from_pretrained(cls, *args, **kwargs): class MarigoldNormalsPipeline (line 875) | class MarigoldNormalsPipeline(metaclass=DummyObject): method __init__ (line 878) | def __init__(self, *args, **kwargs): method from_config (line 882) | def from_config(cls, *args, **kwargs): method from_pretrained (line 886) | def from_pretrained(cls, *args, **kwargs): class MusicLDMPipeline (line 890) | class MusicLDMPipeline(metaclass=DummyObject): method __init__ (line 893) | def __init__(self, *args, **kwargs): method from_config (line 897) | def from_config(cls, *args, **kwargs): method from_pretrained (line 901) | def from_pretrained(cls, *args, **kwargs): class PaintByExamplePipeline (line 905) | class PaintByExamplePipeline(metaclass=DummyObject): method __init__ (line 908) | def __init__(self, *args, **kwargs): method from_config (line 912) | def from_config(cls, *args, **kwargs): method from_pretrained (line 916) | def from_pretrained(cls, *args, **kwargs): class PIAPipeline (line 920) | class PIAPipeline(metaclass=DummyObject): method __init__ (line 923) | def __init__(self, *args, **kwargs): method from_config (line 927) | def from_config(cls, *args, **kwargs): method from_pretrained (line 931) | def from_pretrained(cls, *args, **kwargs): class PixArtAlphaPipeline (line 935) | class PixArtAlphaPipeline(metaclass=DummyObject): method __init__ (line 938) | def __init__(self, *args, **kwargs): method from_config (line 942) | def from_config(cls, *args, **kwargs): method from_pretrained (line 946) | def from_pretrained(cls, *args, **kwargs): class PixArtSigmaPAGPipeline (line 950) | class PixArtSigmaPAGPipeline(metaclass=DummyObject): method __init__ (line 953) | def __init__(self, *args, **kwargs): method from_config (line 957) | def from_config(cls, *args, **kwargs): method from_pretrained (line 961) | def from_pretrained(cls, *args, **kwargs): class PixArtSigmaPipeline (line 965) | class PixArtSigmaPipeline(metaclass=DummyObject): method __init__ (line 968) | def __init__(self, *args, **kwargs): method from_config (line 972) | def from_config(cls, *args, **kwargs): method from_pretrained (line 976) | def from_pretrained(cls, *args, **kwargs): class SemanticStableDiffusionPipeline (line 980) | class SemanticStableDiffusionPipeline(metaclass=DummyObject): method __init__ (line 983) | def __init__(self, *args, **kwargs): method from_config (line 987) | def from_config(cls, *args, **kwargs): method from_pretrained (line 991) | def from_pretrained(cls, *args, **kwargs): class ShapEImg2ImgPipeline (line 995) | class ShapEImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 998) | def __init__(self, *args, **kwargs): method from_config (line 1002) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1006) | def from_pretrained(cls, *args, **kwargs): class ShapEPipeline (line 1010) | class ShapEPipeline(metaclass=DummyObject): method __init__ (line 1013) | def __init__(self, *args, **kwargs): method from_config (line 1017) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1021) | def from_pretrained(cls, *args, **kwargs): class StableAudioPipeline (line 1025) | class StableAudioPipeline(metaclass=DummyObject): method __init__ (line 1028) | def __init__(self, *args, **kwargs): method from_config (line 1032) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1036) | def from_pretrained(cls, *args, **kwargs): class StableAudioProjectionModel (line 1040) | class StableAudioProjectionModel(metaclass=DummyObject): method __init__ (line 1043) | def __init__(self, *args, **kwargs): method from_config (line 1047) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1051) | def from_pretrained(cls, *args, **kwargs): class StableCascadeCombinedPipeline (line 1055) | class StableCascadeCombinedPipeline(metaclass=DummyObject): method __init__ (line 1058) | def __init__(self, *args, **kwargs): method from_config (line 1062) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1066) | def from_pretrained(cls, *args, **kwargs): class StableCascadeDecoderPipeline (line 1070) | class StableCascadeDecoderPipeline(metaclass=DummyObject): method __init__ (line 1073) | def __init__(self, *args, **kwargs): method from_config (line 1077) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1081) | def from_pretrained(cls, *args, **kwargs): class StableCascadePriorPipeline (line 1085) | class StableCascadePriorPipeline(metaclass=DummyObject): method __init__ (line 1088) | def __init__(self, *args, **kwargs): method from_config (line 1092) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1096) | def from_pretrained(cls, *args, **kwargs): class StableDiffusion3ControlNetPipeline (line 1100) | class StableDiffusion3ControlNetPipeline(metaclass=DummyObject): method __init__ (line 1103) | def __init__(self, *args, **kwargs): method from_config (line 1107) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1111) | def from_pretrained(cls, *args, **kwargs): class StableDiffusion3Img2ImgPipeline (line 1115) | class StableDiffusion3Img2ImgPipeline(metaclass=DummyObject): method __init__ (line 1118) | def __init__(self, *args, **kwargs): method from_config (line 1122) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1126) | def from_pretrained(cls, *args, **kwargs): class StableDiffusion3InpaintPipeline (line 1130) | class StableDiffusion3InpaintPipeline(metaclass=DummyObject): method __init__ (line 1133) | def __init__(self, *args, **kwargs): method from_config (line 1137) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1141) | def from_pretrained(cls, *args, **kwargs): class StableDiffusion3PAGPipeline (line 1145) | class StableDiffusion3PAGPipeline(metaclass=DummyObject): method __init__ (line 1148) | def __init__(self, *args, **kwargs): method from_config (line 1152) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1156) | def from_pretrained(cls, *args, **kwargs): class StableDiffusion3Pipeline (line 1160) | class StableDiffusion3Pipeline(metaclass=DummyObject): method __init__ (line 1163) | def __init__(self, *args, **kwargs): method from_config (line 1167) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1171) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionAdapterPipeline (line 1175) | class StableDiffusionAdapterPipeline(metaclass=DummyObject): method __init__ (line 1178) | def __init__(self, *args, **kwargs): method from_config (line 1182) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1186) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionAttendAndExcitePipeline (line 1190) | class StableDiffusionAttendAndExcitePipeline(metaclass=DummyObject): method __init__ (line 1193) | def __init__(self, *args, **kwargs): method from_config (line 1197) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1201) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionControlNetImg2ImgPipeline (line 1205) | class StableDiffusionControlNetImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 1208) | def __init__(self, *args, **kwargs): method from_config (line 1212) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1216) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionControlNetInpaintPipeline (line 1220) | class StableDiffusionControlNetInpaintPipeline(metaclass=DummyObject): method __init__ (line 1223) | def __init__(self, *args, **kwargs): method from_config (line 1227) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1231) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionControlNetPAGPipeline (line 1235) | class StableDiffusionControlNetPAGPipeline(metaclass=DummyObject): method __init__ (line 1238) | def __init__(self, *args, **kwargs): method from_config (line 1242) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1246) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionControlNetPipeline (line 1250) | class StableDiffusionControlNetPipeline(metaclass=DummyObject): method __init__ (line 1253) | def __init__(self, *args, **kwargs): method from_config (line 1257) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1261) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionControlNetXSPipeline (line 1265) | class StableDiffusionControlNetXSPipeline(metaclass=DummyObject): method __init__ (line 1268) | def __init__(self, *args, **kwargs): method from_config (line 1272) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1276) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionDepth2ImgPipeline (line 1280) | class StableDiffusionDepth2ImgPipeline(metaclass=DummyObject): method __init__ (line 1283) | def __init__(self, *args, **kwargs): method from_config (line 1287) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1291) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionDiffEditPipeline (line 1295) | class StableDiffusionDiffEditPipeline(metaclass=DummyObject): method __init__ (line 1298) | def __init__(self, *args, **kwargs): method from_config (line 1302) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1306) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionGLIGENPipeline (line 1310) | class StableDiffusionGLIGENPipeline(metaclass=DummyObject): method __init__ (line 1313) | def __init__(self, *args, **kwargs): method from_config (line 1317) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1321) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionGLIGENTextImagePipeline (line 1325) | class StableDiffusionGLIGENTextImagePipeline(metaclass=DummyObject): method __init__ (line 1328) | def __init__(self, *args, **kwargs): method from_config (line 1332) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1336) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionImageVariationPipeline (line 1340) | class StableDiffusionImageVariationPipeline(metaclass=DummyObject): method __init__ (line 1343) | def __init__(self, *args, **kwargs): method from_config (line 1347) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1351) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionImg2ImgPipeline (line 1355) | class StableDiffusionImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 1358) | def __init__(self, *args, **kwargs): method from_config (line 1362) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1366) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionInpaintPipeline (line 1370) | class StableDiffusionInpaintPipeline(metaclass=DummyObject): method __init__ (line 1373) | def __init__(self, *args, **kwargs): method from_config (line 1377) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1381) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionInpaintPipelineLegacy (line 1385) | class StableDiffusionInpaintPipelineLegacy(metaclass=DummyObject): method __init__ (line 1388) | def __init__(self, *args, **kwargs): method from_config (line 1392) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1396) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionInstructPix2PixPipeline (line 1400) | class StableDiffusionInstructPix2PixPipeline(metaclass=DummyObject): method __init__ (line 1403) | def __init__(self, *args, **kwargs): method from_config (line 1407) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1411) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionLatentUpscalePipeline (line 1415) | class StableDiffusionLatentUpscalePipeline(metaclass=DummyObject): method __init__ (line 1418) | def __init__(self, *args, **kwargs): method from_config (line 1422) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1426) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionLDM3DPipeline (line 1430) | class StableDiffusionLDM3DPipeline(metaclass=DummyObject): method __init__ (line 1433) | def __init__(self, *args, **kwargs): method from_config (line 1437) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1441) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionModelEditingPipeline (line 1445) | class StableDiffusionModelEditingPipeline(metaclass=DummyObject): method __init__ (line 1448) | def __init__(self, *args, **kwargs): method from_config (line 1452) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1456) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionPAGPipeline (line 1460) | class StableDiffusionPAGPipeline(metaclass=DummyObject): method __init__ (line 1463) | def __init__(self, *args, **kwargs): method from_config (line 1467) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1471) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionPanoramaPipeline (line 1475) | class StableDiffusionPanoramaPipeline(metaclass=DummyObject): method __init__ (line 1478) | def __init__(self, *args, **kwargs): method from_config (line 1482) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1486) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionParadigmsPipeline (line 1490) | class StableDiffusionParadigmsPipeline(metaclass=DummyObject): method __init__ (line 1493) | def __init__(self, *args, **kwargs): method from_config (line 1497) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1501) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionPipeline (line 1505) | class StableDiffusionPipeline(metaclass=DummyObject): method __init__ (line 1508) | def __init__(self, *args, **kwargs): method from_config (line 1512) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1516) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionPipelineSafe (line 1520) | class StableDiffusionPipelineSafe(metaclass=DummyObject): method __init__ (line 1523) | def __init__(self, *args, **kwargs): method from_config (line 1527) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1531) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionPix2PixZeroPipeline (line 1535) | class StableDiffusionPix2PixZeroPipeline(metaclass=DummyObject): method __init__ (line 1538) | def __init__(self, *args, **kwargs): method from_config (line 1542) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1546) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionSAGPipeline (line 1550) | class StableDiffusionSAGPipeline(metaclass=DummyObject): method __init__ (line 1553) | def __init__(self, *args, **kwargs): method from_config (line 1557) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1561) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionUpscalePipeline (line 1565) | class StableDiffusionUpscalePipeline(metaclass=DummyObject): method __init__ (line 1568) | def __init__(self, *args, **kwargs): method from_config (line 1572) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1576) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLAdapterPipeline (line 1580) | class StableDiffusionXLAdapterPipeline(metaclass=DummyObject): method __init__ (line 1583) | def __init__(self, *args, **kwargs): method from_config (line 1587) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1591) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLControlNetImg2ImgPipeline (line 1595) | class StableDiffusionXLControlNetImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 1598) | def __init__(self, *args, **kwargs): method from_config (line 1602) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1606) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLControlNetInpaintPipeline (line 1610) | class StableDiffusionXLControlNetInpaintPipeline(metaclass=DummyObject): method __init__ (line 1613) | def __init__(self, *args, **kwargs): method from_config (line 1617) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1621) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLControlNetPAGPipeline (line 1625) | class StableDiffusionXLControlNetPAGPipeline(metaclass=DummyObject): method __init__ (line 1628) | def __init__(self, *args, **kwargs): method from_config (line 1632) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1636) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLControlNetPipeline (line 1640) | class StableDiffusionXLControlNetPipeline(metaclass=DummyObject): method __init__ (line 1643) | def __init__(self, *args, **kwargs): method from_config (line 1647) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1651) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLControlNetXSPipeline (line 1655) | class StableDiffusionXLControlNetXSPipeline(metaclass=DummyObject): method __init__ (line 1658) | def __init__(self, *args, **kwargs): method from_config (line 1662) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1666) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLImg2ImgPipeline (line 1670) | class StableDiffusionXLImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 1673) | def __init__(self, *args, **kwargs): method from_config (line 1677) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1681) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLInpaintPipeline (line 1685) | class StableDiffusionXLInpaintPipeline(metaclass=DummyObject): method __init__ (line 1688) | def __init__(self, *args, **kwargs): method from_config (line 1692) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1696) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLInstructPix2PixPipeline (line 1700) | class StableDiffusionXLInstructPix2PixPipeline(metaclass=DummyObject): method __init__ (line 1703) | def __init__(self, *args, **kwargs): method from_config (line 1707) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1711) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLPAGImg2ImgPipeline (line 1715) | class StableDiffusionXLPAGImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 1718) | def __init__(self, *args, **kwargs): method from_config (line 1722) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1726) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLPAGInpaintPipeline (line 1730) | class StableDiffusionXLPAGInpaintPipeline(metaclass=DummyObject): method __init__ (line 1733) | def __init__(self, *args, **kwargs): method from_config (line 1737) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1741) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLPAGPipeline (line 1745) | class StableDiffusionXLPAGPipeline(metaclass=DummyObject): method __init__ (line 1748) | def __init__(self, *args, **kwargs): method from_config (line 1752) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1756) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLPipeline (line 1760) | class StableDiffusionXLPipeline(metaclass=DummyObject): method __init__ (line 1763) | def __init__(self, *args, **kwargs): method from_config (line 1767) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1771) | def from_pretrained(cls, *args, **kwargs): class StableUnCLIPImg2ImgPipeline (line 1775) | class StableUnCLIPImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 1778) | def __init__(self, *args, **kwargs): method from_config (line 1782) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1786) | def from_pretrained(cls, *args, **kwargs): class StableUnCLIPPipeline (line 1790) | class StableUnCLIPPipeline(metaclass=DummyObject): method __init__ (line 1793) | def __init__(self, *args, **kwargs): method from_config (line 1797) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1801) | def from_pretrained(cls, *args, **kwargs): class StableVideoDiffusionPipeline (line 1805) | class StableVideoDiffusionPipeline(metaclass=DummyObject): method __init__ (line 1808) | def __init__(self, *args, **kwargs): method from_config (line 1812) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1816) | def from_pretrained(cls, *args, **kwargs): class TextToVideoSDPipeline (line 1820) | class TextToVideoSDPipeline(metaclass=DummyObject): method __init__ (line 1823) | def __init__(self, *args, **kwargs): method from_config (line 1827) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1831) | def from_pretrained(cls, *args, **kwargs): class TextToVideoZeroPipeline (line 1835) | class TextToVideoZeroPipeline(metaclass=DummyObject): method __init__ (line 1838) | def __init__(self, *args, **kwargs): method from_config (line 1842) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1846) | def from_pretrained(cls, *args, **kwargs): class TextToVideoZeroSDXLPipeline (line 1850) | class TextToVideoZeroSDXLPipeline(metaclass=DummyObject): method __init__ (line 1853) | def __init__(self, *args, **kwargs): method from_config (line 1857) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1861) | def from_pretrained(cls, *args, **kwargs): class UnCLIPImageVariationPipeline (line 1865) | class UnCLIPImageVariationPipeline(metaclass=DummyObject): method __init__ (line 1868) | def __init__(self, *args, **kwargs): method from_config (line 1872) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1876) | def from_pretrained(cls, *args, **kwargs): class UnCLIPPipeline (line 1880) | class UnCLIPPipeline(metaclass=DummyObject): method __init__ (line 1883) | def __init__(self, *args, **kwargs): method from_config (line 1887) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1891) | def from_pretrained(cls, *args, **kwargs): class UniDiffuserModel (line 1895) | class UniDiffuserModel(metaclass=DummyObject): method __init__ (line 1898) | def __init__(self, *args, **kwargs): method from_config (line 1902) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1906) | def from_pretrained(cls, *args, **kwargs): class UniDiffuserPipeline (line 1910) | class UniDiffuserPipeline(metaclass=DummyObject): method __init__ (line 1913) | def __init__(self, *args, **kwargs): method from_config (line 1917) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1921) | def from_pretrained(cls, *args, **kwargs): class UniDiffuserTextDecoder (line 1925) | class UniDiffuserTextDecoder(metaclass=DummyObject): method __init__ (line 1928) | def __init__(self, *args, **kwargs): method from_config (line 1932) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1936) | def from_pretrained(cls, *args, **kwargs): class VersatileDiffusionDualGuidedPipeline (line 1940) | class VersatileDiffusionDualGuidedPipeline(metaclass=DummyObject): method __init__ (line 1943) | def __init__(self, *args, **kwargs): method from_config (line 1947) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1951) | def from_pretrained(cls, *args, **kwargs): class VersatileDiffusionImageVariationPipeline (line 1955) | class VersatileDiffusionImageVariationPipeline(metaclass=DummyObject): method __init__ (line 1958) | def __init__(self, *args, **kwargs): method from_config (line 1962) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1966) | def from_pretrained(cls, *args, **kwargs): class VersatileDiffusionPipeline (line 1970) | class VersatileDiffusionPipeline(metaclass=DummyObject): method __init__ (line 1973) | def __init__(self, *args, **kwargs): method from_config (line 1977) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1981) | def from_pretrained(cls, *args, **kwargs): class VersatileDiffusionTextToImagePipeline (line 1985) | class VersatileDiffusionTextToImagePipeline(metaclass=DummyObject): method __init__ (line 1988) | def __init__(self, *args, **kwargs): method from_config (line 1992) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1996) | def from_pretrained(cls, *args, **kwargs): class VideoToVideoSDPipeline (line 2000) | class VideoToVideoSDPipeline(metaclass=DummyObject): method __init__ (line 2003) | def __init__(self, *args, **kwargs): method from_config (line 2007) | def from_config(cls, *args, **kwargs): method from_pretrained (line 2011) | def from_pretrained(cls, *args, **kwargs): class VQDiffusionPipeline (line 2015) | class VQDiffusionPipeline(metaclass=DummyObject): method __init__ (line 2018) | def __init__(self, *args, **kwargs): method from_config (line 2022) | def from_config(cls, *args, **kwargs): method from_pretrained (line 2026) | def from_pretrained(cls, *args, **kwargs): class WuerstchenCombinedPipeline (line 2030) | class WuerstchenCombinedPipeline(metaclass=DummyObject): method __init__ (line 2033) | def __init__(self, *args, **kwargs): method from_config (line 2037) | def from_config(cls, *args, **kwargs): method from_pretrained (line 2041) | def from_pretrained(cls, *args, **kwargs): class WuerstchenDecoderPipeline (line 2045) | class WuerstchenDecoderPipeline(metaclass=DummyObject): method __init__ (line 2048) | def __init__(self, *args, **kwargs): method from_config (line 2052) | def from_config(cls, *args, **kwargs): method from_pretrained (line 2056) | def from_pretrained(cls, *args, **kwargs): class WuerstchenPriorPipeline (line 2060) | class WuerstchenPriorPipeline(metaclass=DummyObject): method __init__ (line 2063) | def __init__(self, *args, **kwargs): method from_config (line 2067) | def from_config(cls, *args, **kwargs): method from_pretrained (line 2071) | def from_pretrained(cls, *args, **kwargs): FILE: src/diffusers/utils/dummy_transformers_and_torch_and_note_seq_objects.py class SpectrogramDiffusionPipeline (line 5) | class SpectrogramDiffusionPipeline(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): FILE: src/diffusers/utils/dynamic_modules_utils.py function get_diffusers_versions (line 42) | def get_diffusers_versions(): function init_hf_modules (line 48) | def init_hf_modules(): function create_dynamic_module (line 63) | def create_dynamic_module(name: Union[str, os.PathLike]): function get_relative_imports (line 78) | def get_relative_imports(module_file): function get_relative_import_files (line 96) | def get_relative_import_files(module_file): function check_imports (line 125) | def check_imports(filename): function get_class_in_module (line 157) | def get_class_in_module(class_name, module_path): function find_pipeline_class (line 169) | def find_pipeline_class(loaded_module): function get_cached_module_file (line 197) | def get_cached_module_file( function get_class_from_dynamic_module (line 374) | def get_class_from_dynamic_module( FILE: src/diffusers/utils/export_utils.py function buffered_writer (line 22) | def buffered_writer(raw_f): function export_to_gif (line 28) | def export_to_gif(image: List[PIL.Image.Image], output_gif_path: str = N... function export_to_ply (line 43) | def export_to_ply(mesh, output_ply_path: str = None): function export_to_obj (line 95) | def export_to_obj(mesh, output_obj_path: str = None): function export_to_video (line 115) | def export_to_video( FILE: src/diffusers/utils/hub_utils.py function http_user_agent (line 75) | def http_user_agent(user_agent: Union[Dict, str, None] = None) -> str: function load_or_create_model_card (line 99) | def load_or_create_model_card( function populate_model_card (line 169) | def populate_model_card(model_card: ModelCard, tags: Union[str, List[str... function extract_commit_hash (line 185) | def extract_commit_hash(resolved_file: Optional[str], commit_hash: Optio... function move_cache (line 209) | def move_cache(old_cache_dir: Optional[str] = None, new_cache_dir: Optio... function _add_variant (line 271) | def _add_variant(weights_name: str, variant: Optional[str] = None) -> str: function _get_model_file (line 282) | def _get_model_file( function _check_if_shards_exist_locally (line 402) | def _check_if_shards_exist_locally(local_dir, subfolder, original_shard_... function _get_checkpoint_shard_files (line 413) | def _get_checkpoint_shard_files( class PushToHubMixin (line 505) | class PushToHubMixin: method _upload_folder (line 510) | def _upload_folder( method push_to_hub (line 534) | def push_to_hub( FILE: src/diffusers/utils/import_utils.py function is_timm_available (line 320) | def is_timm_available(): function is_torch_available (line 334) | def is_torch_available(): function is_torch_xla_available (line 338) | def is_torch_xla_available(): function is_torch_npu_available (line 342) | def is_torch_npu_available(): function is_flax_available (line 346) | def is_flax_available(): function is_transformers_available (line 350) | def is_transformers_available(): function is_inflect_available (line 354) | def is_inflect_available(): function is_unidecode_available (line 358) | def is_unidecode_available(): function is_onnx_available (line 362) | def is_onnx_available(): function is_opencv_available (line 366) | def is_opencv_available(): function is_scipy_available (line 370) | def is_scipy_available(): function is_librosa_available (line 374) | def is_librosa_available(): function is_xformers_available (line 378) | def is_xformers_available(): function is_accelerate_available (line 382) | def is_accelerate_available(): function is_k_diffusion_available (line 386) | def is_k_diffusion_available(): function is_note_seq_available (line 390) | def is_note_seq_available(): function is_wandb_available (line 394) | def is_wandb_available(): function is_tensorboard_available (line 398) | def is_tensorboard_available(): function is_compel_available (line 402) | def is_compel_available(): function is_ftfy_available (line 406) | def is_ftfy_available(): function is_bs4_available (line 410) | def is_bs4_available(): function is_torchsde_available (line 414) | def is_torchsde_available(): function is_invisible_watermark_available (line 418) | def is_invisible_watermark_available(): function is_peft_available (line 422) | def is_peft_available(): function is_torchvision_available (line 426) | def is_torchvision_available(): function is_matplotlib_available (line 430) | def is_matplotlib_available(): function is_safetensors_available (line 434) | def is_safetensors_available(): function is_bitsandbytes_available (line 438) | def is_bitsandbytes_available(): function is_google_colab (line 442) | def is_google_colab(): function is_sentencepiece_available (line 446) | def is_sentencepiece_available(): function requires_backends (line 606) | def requires_backends(obj, backends): class DummyObject (line 637) | class DummyObject(type): method __getattr__ (line 643) | def __getattr__(cls, key): function compare_versions (line 650) | def compare_versions(library_or_version: Union[str, Version], operation:... function is_torch_version (line 670) | def is_torch_version(operation: str, version: str): function is_transformers_version (line 682) | def is_transformers_version(operation: str, version: str): function is_accelerate_version (line 696) | def is_accelerate_version(operation: str, version: str): function is_peft_version (line 710) | def is_peft_version(operation: str, version: str): function is_k_diffusion_version (line 724) | def is_k_diffusion_version(operation: str, version: str): function get_objects_from_module (line 738) | def get_objects_from_module(module): class OptionalDependencyNotAvailable (line 758) | class OptionalDependencyNotAvailable(BaseException): class _LazyModule (line 762) | class _LazyModule(ModuleType): method __init__ (line 769) | def __init__(self, name, module_file, import_structure, module_spec=No... method __dir__ (line 786) | def __dir__(self): method __getattr__ (line 795) | def __getattr__(self, name: str) -> Any: method _get_module (line 809) | def _get_module(self, module_name: str): method __reduce__ (line 818) | def __reduce__(self): FILE: src/diffusers/utils/loading_utils.py function load_image (line 12) | def load_image( function load_video (line 55) | def load_video( FILE: src/diffusers/utils/logging.py function _get_default_logging_level (line 52) | def _get_default_logging_level() -> int: function _get_library_name (line 69) | def _get_library_name() -> str: function _get_library_root_logger (line 73) | def _get_library_root_logger() -> logging.Logger: function _configure_library_root_logger (line 77) | def _configure_library_root_logger() -> None: function _reset_library_root_logger (line 96) | def _reset_library_root_logger() -> None: function get_log_levels_dict (line 109) | def get_log_levels_dict() -> Dict[str, int]: function get_logger (line 113) | def get_logger(name: Optional[str] = None) -> logging.Logger: function get_verbosity (line 127) | def get_verbosity() -> int: function set_verbosity (line 147) | def set_verbosity(verbosity: int) -> None: function set_verbosity_info (line 166) | def set_verbosity_info() -> None: function set_verbosity_warning (line 171) | def set_verbosity_warning() -> None: function set_verbosity_debug (line 176) | def set_verbosity_debug() -> None: function set_verbosity_error (line 181) | def set_verbosity_error() -> None: function disable_default_handler (line 186) | def disable_default_handler() -> None: function enable_default_handler (line 195) | def enable_default_handler() -> None: function add_handler (line 204) | def add_handler(handler: logging.Handler) -> None: function remove_handler (line 213) | def remove_handler(handler: logging.Handler) -> None: function disable_propagation (line 222) | def disable_propagation() -> None: function enable_propagation (line 231) | def enable_propagation() -> None: function enable_explicit_format (line 241) | def enable_explicit_format() -> None: function reset_format (line 256) | def reset_format() -> None: function warning_advice (line 268) | def warning_advice(self, *args, **kwargs) -> None: class EmptyTqdm (line 282) | class EmptyTqdm: method __init__ (line 285) | def __init__(self, *args, **kwargs): # pylint: disable=unused-argument method __iter__ (line 288) | def __iter__(self): method __getattr__ (line 291) | def __getattr__(self, _): method __enter__ (line 299) | def __enter__(self): method __exit__ (line 302) | def __exit__(self, type_, value, traceback): class _tqdm_cls (line 306) | class _tqdm_cls: method __call__ (line 307) | def __call__(self, *args, **kwargs): method set_lock (line 313) | def set_lock(self, *args, **kwargs): method get_lock (line 318) | def get_lock(self): function is_progress_bar_enabled (line 326) | def is_progress_bar_enabled() -> bool: function enable_progress_bar (line 332) | def enable_progress_bar() -> None: function disable_progress_bar (line 338) | def disable_progress_bar() -> None: FILE: src/diffusers/utils/outputs.py function is_tensor (line 27) | def is_tensor(x) -> bool: class BaseOutput (line 40) | class BaseOutput(OrderedDict): method __init_subclass__ (line 54) | def __init_subclass__(cls) -> None: method __post_init__ (line 76) | def __post_init__(self) -> None: method __delitem__ (line 95) | def __delitem__(self, *args, **kwargs): method setdefault (line 98) | def setdefault(self, *args, **kwargs): method pop (line 101) | def pop(self, *args, **kwargs): method update (line 104) | def update(self, *args, **kwargs): method __getitem__ (line 107) | def __getitem__(self, k: Any) -> Any: method __setattr__ (line 114) | def __setattr__(self, name: Any, value: Any) -> None: method __setitem__ (line 120) | def __setitem__(self, key, value): method __reduce__ (line 126) | def __reduce__(self): method to_tuple (line 133) | def to_tuple(self) -> Tuple[Any, ...]: FILE: src/diffusers/utils/peft_utils.py function recurse_remove_peft_layers (line 31) | def recurse_remove_peft_layers(model): function scale_lora_layers (line 103) | def scale_lora_layers(model, weight): function unscale_lora_layers (line 123) | def unscale_lora_layers(model, weight: Optional[float] = None): function get_peft_kwargs (line 150) | def get_peft_kwargs(rank_dict, network_alpha_dict, peft_state_dict, is_u... function get_adapter_name (line 195) | def get_adapter_name(model): function set_adapter_layers (line 204) | def set_adapter_layers(model, enabled=True): function delete_adapter_layers (line 216) | def delete_adapter_layers(model, adapter_name): function set_weights_and_activate_adapters (line 238) | def set_weights_and_activate_adapters(model, adapter_names, weights): function check_peft_version (line 278) | def check_peft_version(min_version: str) -> None: FILE: src/diffusers/utils/pil_utils.py function pt_to_pil (line 27) | def pt_to_pil(images): function numpy_to_pil (line 37) | def numpy_to_pil(images): function make_image_grid (line 53) | def make_image_grid(images: List[PIL.Image.Image], rows: int, cols: int,... FILE: src/diffusers/utils/state_dict_utils.py class StateDictType (line 26) | class StateDictType(enum.Enum): function convert_state_dict (line 141) | def convert_state_dict(state_dict, mapping): function convert_state_dict_to_peft (line 174) | def convert_state_dict_to_peft(state_dict, original_type=None, **kwargs): function convert_state_dict_to_diffusers (line 201) | def convert_state_dict_to_diffusers(state_dict, original_type=None, **kw... function convert_unet_state_dict_to_peft (line 248) | def convert_unet_state_dict_to_peft(state_dict): function convert_all_state_dict_to_peft (line 256) | def convert_all_state_dict_to_peft(state_dict): function convert_state_dict_to_kohya (line 275) | def convert_state_dict_to_kohya(state_dict, original_type=None, **kwargs): FILE: src/diffusers/utils/testing_utils.py function torch_all_close (line 94) | def torch_all_close(a, b, *args, **kwargs): function numpy_cosine_similarity_distance (line 102) | def numpy_cosine_similarity_distance(a, b): function print_tensor_test (line 109) | def print_tensor_test( function get_tests_dir (line 135) | def get_tests_dir(append_path=None): function str_to_bool (line 158) | def str_to_bool(value) -> int: function parse_flag_from_env (line 172) | def parse_flag_from_env(key, default=False): function floats_tensor (line 193) | def floats_tensor(shape, scale=1.0, rng=None, name=None): function slow (line 209) | def slow(test_case): function nightly (line 219) | def nightly(test_case): function is_torch_compile (line 229) | def is_torch_compile(test_case): function require_torch (line 239) | def require_torch(test_case): function require_torch_2 (line 246) | def require_torch_2(test_case): function require_torch_gpu (line 255) | def require_torch_gpu(test_case): function require_torch_accelerator (line 263) | def require_torch_accelerator(test_case): function require_torch_multi_gpu (line 270) | def require_torch_multi_gpu(test_case): function require_torch_accelerator_with_fp16 (line 284) | def require_torch_accelerator_with_fp16(test_case): function require_torch_accelerator_with_fp64 (line 291) | def require_torch_accelerator_with_fp64(test_case): function require_torch_accelerator_with_training (line 298) | def require_torch_accelerator_with_training(test_case): function skip_mps (line 306) | def skip_mps(test_case): function require_flax (line 311) | def require_flax(test_case): function require_compel (line 318) | def require_compel(test_case): function require_onnxruntime (line 326) | def require_onnxruntime(test_case): function require_note_seq (line 333) | def require_note_seq(test_case): function require_torchsde (line 340) | def require_torchsde(test_case): function require_peft_backend (line 347) | def require_peft_backend(test_case): function require_timm (line 355) | def require_timm(test_case): function require_peft_version_greater (line 362) | def require_peft_version_greater(peft_version): function require_accelerate_version_greater (line 379) | def require_accelerate_version_greater(accelerate_version): function deprecate_after_peft_backend (line 391) | def deprecate_after_peft_backend(test_case): function get_python_version (line 398) | def get_python_version(): function load_numpy (line 404) | def load_numpy(arry: Union[str, np.ndarray], local_path: Optional[str] =... function load_pt (line 430) | def load_pt(url: str): function load_image (line 437) | def load_image(image: Union[str, PIL.Image.Image]) -> PIL.Image.Image: function preprocess_image (line 468) | def preprocess_image(image: PIL.Image, batch_size: int): function export_to_gif (line 478) | def export_to_gif(image: List[PIL.Image.Image], output_gif_path: str = N... function buffered_writer (line 494) | def buffered_writer(raw_f): function export_to_ply (line 500) | def export_to_ply(mesh, output_ply_path: str = None): function export_to_obj (line 552) | def export_to_obj(mesh, output_obj_path: str = None): function export_to_video (line 572) | def export_to_video(video_frames: List[np.ndarray], output_video_path: s... function load_hf_numpy (line 589) | def load_hf_numpy(path) -> np.ndarray: function pytest_addoption_shared (line 604) | def pytest_addoption_shared(parser): function pytest_terminal_summary_main (line 623) | def pytest_terminal_summary_main(tr, id): function is_flaky (line 753) | def is_flaky(max_attempts: int = 5, wait_before_retry: Optional[float] =... function run_test_in_subprocess (line 790) | def run_test_in_subprocess(test_case, target_func, inputs=None, timeout=... class CaptureLogger (line 833) | class CaptureLogger: method __init__ (line 854) | def __init__(self, logger): method __enter__ (line 860) | def __enter__(self): method __exit__ (line 864) | def __exit__(self, *exc): method __repr__ (line 868) | def __repr__(self): function enable_full_determinism (line 872) | def enable_full_determinism(): function disable_full_determinism (line 890) | def disable_full_determinism(): function _is_torch_fp16_available (line 897) | def _is_torch_fp16_available(device): function _is_torch_fp64_available (line 919) | def _is_torch_fp64_available(device): function _device_agnostic_dispatch (line 953) | def _device_agnostic_dispatch(device: str, dispatch_table: Dict[str, Cal... function backend_manual_seed (line 968) | def backend_manual_seed(device: str, seed: int): function backend_empty_cache (line 972) | def backend_empty_cache(device: str): function backend_device_count (line 976) | def backend_device_count(device: str): function backend_supports_training (line 982) | def backend_supports_training(device: str): function update_mapping_from_spec (line 995) | def update_mapping_from_spec(device_fn_dict: Dict[str, Callable], attrib... FILE: src/diffusers/utils/torch_utils.py function maybe_allow_in_graph (line 34) | def maybe_allow_in_graph(cls): function randn_tensor (line 38) | def randn_tensor( function is_compiled_module (line 86) | def is_compiled_module(module) -> bool: function fourier_filter (line 93) | def fourier_filter(x_in: "torch.Tensor", threshold: int, scale: int) -> ... function apply_freeu (line 124) | def apply_freeu( FILE: src/diffusers/utils/versions.py function _compare_versions (line 37) | def _compare_versions(op, got_ver, want_ver, requirement, pkg, hint): function require_version (line 49) | def require_version(requirement: str, hint: Optional[str] = None) -> None: function require_version_core (line 114) | def require_version_core(requirement): FILE: src/diffusers/video_processor.py class VideoProcessor (line 25) | class VideoProcessor(VaeImageProcessor): method preprocess_video (line 28) | def preprocess_video(self, video, height: Optional[int] = None, width:... method postprocess_video (line 89) | def postprocess_video(