SYMBOL INDEX (14100 symbols across 926 files) FILE: app.py function generate (line 16) | def generate(image_mask, prompt, num_inference_steps=50, image_guidance_... FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/EdgeSAM/common.py class MLPBlock (line 14) | class MLPBlock(nn.Module): method __init__ (line 15) | def __init__( method forward (line 26) | def forward(self, x: torch.Tensor) -> torch.Tensor: class LayerNorm2d (line 32) | class LayerNorm2d(nn.Module): method __init__ (line 33) | def __init__(self, num_channels: int, eps: float = 1e-6) -> None: method forward (line 39) | def forward(self, x: torch.Tensor) -> torch.Tensor: function val2list (line 47) | def val2list(x: list or tuple or any, repeat_time=1) -> list: function val2tuple (line 53) | def val2tuple(x: list or tuple or any, min_len: int = 1, idx_repeat: int... function list_sum (line 63) | def list_sum(x: list) -> any: function resize (line 67) | def resize( class UpSampleLayer (line 88) | class UpSampleLayer(nn.Module): method __init__ (line 89) | def __init__( method forward (line 102) | def forward(self, x: torch.Tensor) -> torch.Tensor: class OpSequential (line 106) | class OpSequential(nn.Module): method __init__ (line 107) | def __init__(self, op_list): method forward (line 115) | def forward(self, x: torch.Tensor) -> torch.Tensor: FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/EdgeSAM/rep_vit.py function _make_divisible (line 103) | def _make_divisible(v, divisor, min_value=None): class Conv2d_BN (line 128) | class Conv2d_BN(torch.nn.Sequential): method __init__ (line 129) | def __init__(self, a, b, ks=1, stride=1, pad=0, dilation=1, method fuse (line 139) | def fuse(self): class Residual (line 154) | class Residual(torch.nn.Module): method __init__ (line 155) | def __init__(self, m, drop=0.): method forward (line 160) | def forward(self, x): method fuse (line 168) | def fuse(self): class RepVGGDW (line 187) | class RepVGGDW(torch.nn.Module): method __init__ (line 188) | def __init__(self, ed) -> None: method forward (line 194) | def forward(self, x): method fuse (line 198) | def fuse(self): class RepViTBlock (line 220) | class RepViTBlock(nn.Module): method __init__ (line 221) | def __init__(self, inp, hidden_dim, oup, kernel_size, stride, use_se, ... method forward (line 257) | def forward(self, x): class BN_Linear (line 264) | class BN_Linear(torch.nn.Sequential): method __init__ (line 265) | def __init__(self, a, b, bias=True, std=0.02): method fuse (line 274) | def fuse(self): class RepViT (line 290) | class RepViT(nn.Module): method __init__ (line 297) | def __init__(self, arch, img_size=1024, upsample_mode='bicubic'): method forward (line 339) | def forward(self, x): function rep_vit_m1 (line 361) | def rep_vit_m1(img_size=1024, **kwargs): function rep_vit_m2 (line 365) | def rep_vit_m2(img_size=1024, **kwargs): function rep_vit_m3 (line 369) | def rep_vit_m3(img_size=1024, **kwargs): FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/EdgeSAM/setup_edge_sam.py function build_edge_sam (line 21) | def build_edge_sam(checkpoint=None, upsample_mode="bicubic"): function _build_sam_encoder (line 35) | def _build_sam_encoder( function _build_sam (line 58) | def _build_sam( FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/FastSAM/tools.py function convert_box_xywh_to_xyxy (line 10) | def convert_box_xywh_to_xyxy(box): function segment_image (line 18) | def segment_image(image, bbox): function format_results (line 35) | def format_results(result, filter=0): function filter_masks (line 53) | def filter_masks(annotations): # filte the overlap mask function get_bbox_from_mask (line 71) | def get_bbox_from_mask(mask): function fast_process (line 91) | def fast_process( function fast_show_mask (line 180) | def fast_show_mask( function fast_show_mask_gpu (line 246) | def fast_show_mask_gpu( function retriev (line 313) | def retriev( function crop_image (line 327) | def crop_image(annotations, image_path): function box_prompt (line 351) | def box_prompt(masks, bbox, target_height, target_width): function point_prompt (line 379) | def point_prompt(masks, points, pointlabel, target_height, target_width)... function text_prompt (line 402) | def text_prompt(annotations, args): FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/LightHQSAM/setup_light_hqsam.py function setup_model (line 4) | def setup_model(): FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/LightHQSAM/tiny_vit_sam.py class Conv2d_BN (line 21) | class Conv2d_BN(torch.nn.Sequential): method __init__ (line 22) | def __init__(self, a, b, ks=1, stride=1, pad=0, dilation=1, method fuse (line 33) | def fuse(self): class DropPath (line 46) | class DropPath(TimmDropPath): method __init__ (line 47) | def __init__(self, drop_prob=None): method __repr__ (line 51) | def __repr__(self): class PatchEmbed (line 57) | class PatchEmbed(nn.Module): method __init__ (line 58) | def __init__(self, in_chans, embed_dim, resolution, activation): method forward (line 73) | def forward(self, x): class MBConv (line 77) | class MBConv(nn.Module): method __init__ (line 78) | def __init__(self, in_chans, out_chans, expand_ratio, method forward (line 99) | def forward(self, x): class PatchMerging (line 118) | class PatchMerging(nn.Module): method __init__ (line 119) | def __init__(self, input_resolution, dim, out_dim, activation): method forward (line 133) | def forward(self, x): class ConvLayer (line 150) | class ConvLayer(nn.Module): method __init__ (line 151) | def __init__(self, dim, input_resolution, depth, method forward (line 178) | def forward(self, x): class Mlp (line 189) | class Mlp(nn.Module): method __init__ (line 190) | def __init__(self, in_features, hidden_features=None, method forward (line 201) | def forward(self, x): class Attention (line 212) | class Attention(torch.nn.Module): method __init__ (line 213) | def __init__(self, dim, key_dim, num_heads=8, method train (line 251) | def train(self, mode=True): method forward (line 260) | def forward(self, x): # x (B,N,C) class TinyViTBlock (line 287) | class TinyViTBlock(nn.Module): method __init__ (line 303) | def __init__(self, dim, input_resolution, num_heads, window_size=7, method forward (line 335) | def forward(self, x): method extra_repr (line 378) | def extra_repr(self) -> str: class BasicLayer (line 383) | class BasicLayer(nn.Module): method __init__ (line 402) | def __init__(self, dim, input_resolution, depth, num_heads, window_size, method forward (line 436) | def forward(self, x): method extra_repr (line 446) | def extra_repr(self) -> str: class LayerNorm2d (line 449) | class LayerNorm2d(nn.Module): method __init__ (line 450) | def __init__(self, num_channels: int, eps: float = 1e-6) -> None: method forward (line 456) | def forward(self, x: torch.Tensor) -> torch.Tensor: class TinyViT (line 462) | class TinyViT(nn.Module): method __init__ (line 463) | def __init__(self, img_size=224, in_chans=3, num_classes=1000, method set_layer_lr_decay (line 553) | def set_layer_lr_decay(self, layer_lr_decay): method _init_weights (line 587) | def _init_weights(self, m): method no_weight_decay_keywords (line 597) | def no_weight_decay_keywords(self): method forward_features (line 600) | def forward_features(self, x): method forward (line 621) | def forward(self, x): function register_tiny_vit_model (line 640) | def register_tiny_vit_model(fn): function tiny_vit_5m_224 (line 666) | def tiny_vit_5m_224(pretrained=False, num_classes=1000, drop_path_rate=0... function tiny_vit_11m_224 (line 678) | def tiny_vit_11m_224(pretrained=False, num_classes=1000, drop_path_rate=... function tiny_vit_21m_224 (line 690) | def tiny_vit_21m_224(pretrained=False, num_classes=1000, drop_path_rate=... function tiny_vit_21m_384 (line 702) | def tiny_vit_21m_384(pretrained=False, num_classes=1000, drop_path_rate=... function tiny_vit_21m_512 (line 715) | def tiny_vit_21m_512(pretrained=False, num_classes=1000, drop_path_rate=... FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/MobileSAM/setup_mobile_sam.py function setup_model (line 4) | def setup_model(): FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/MobileSAM/tiny_vit_sam.py class Conv2d_BN (line 21) | class Conv2d_BN(torch.nn.Sequential): method __init__ (line 22) | def __init__(self, a, b, ks=1, stride=1, pad=0, dilation=1, method fuse (line 33) | def fuse(self): class DropPath (line 46) | class DropPath(TimmDropPath): method __init__ (line 47) | def __init__(self, drop_prob=None): method __repr__ (line 51) | def __repr__(self): class PatchEmbed (line 57) | class PatchEmbed(nn.Module): method __init__ (line 58) | def __init__(self, in_chans, embed_dim, resolution, activation): method forward (line 73) | def forward(self, x): class MBConv (line 77) | class MBConv(nn.Module): method __init__ (line 78) | def __init__(self, in_chans, out_chans, expand_ratio, method forward (line 99) | def forward(self, x): class PatchMerging (line 118) | class PatchMerging(nn.Module): method __init__ (line 119) | def __init__(self, input_resolution, dim, out_dim, activation): method forward (line 133) | def forward(self, x): class ConvLayer (line 150) | class ConvLayer(nn.Module): method __init__ (line 151) | def __init__(self, dim, input_resolution, depth, method forward (line 178) | def forward(self, x): class Mlp (line 189) | class Mlp(nn.Module): method __init__ (line 190) | def __init__(self, in_features, hidden_features=None, method forward (line 201) | def forward(self, x): class Attention (line 212) | class Attention(torch.nn.Module): method __init__ (line 213) | def __init__(self, dim, key_dim, num_heads=8, method train (line 251) | def train(self, mode=True): method forward (line 258) | def forward(self, x): # x (B,N,C) class TinyViTBlock (line 285) | class TinyViTBlock(nn.Module): method __init__ (line 301) | def __init__(self, dim, input_resolution, num_heads, window_size=7, method forward (line 333) | def forward(self, x): method extra_repr (line 376) | def extra_repr(self) -> str: class BasicLayer (line 381) | class BasicLayer(nn.Module): method __init__ (line 400) | def __init__(self, dim, input_resolution, depth, num_heads, window_size, method forward (line 434) | def forward(self, x): method extra_repr (line 444) | def extra_repr(self) -> str: class LayerNorm2d (line 447) | class LayerNorm2d(nn.Module): method __init__ (line 448) | def __init__(self, num_channels: int, eps: float = 1e-6) -> None: method forward (line 454) | def forward(self, x: torch.Tensor) -> torch.Tensor: class TinyViT (line 460) | class TinyViT(nn.Module): method __init__ (line 461) | def __init__(self, img_size=224, in_chans=3, num_classes=1000, method set_layer_lr_decay (line 551) | def set_layer_lr_decay(self, layer_lr_decay): method _init_weights (line 585) | def _init_weights(self, m): method no_weight_decay_keywords (line 595) | def no_weight_decay_keywords(self): method forward_features (line 598) | def forward_features(self, x): method forward (line 614) | def forward(self, x): function register_tiny_vit_model (line 632) | def register_tiny_vit_model(fn): function tiny_vit_5m_224 (line 658) | def tiny_vit_5m_224(pretrained=False, num_classes=1000, drop_path_rate=0... function tiny_vit_11m_224 (line 670) | def tiny_vit_11m_224(pretrained=False, num_classes=1000, drop_path_rate=... function tiny_vit_21m_224 (line 682) | def tiny_vit_21m_224(pretrained=False, num_classes=1000, drop_path_rate=... function tiny_vit_21m_384 (line 694) | def tiny_vit_21m_384(pretrained=False, num_classes=1000, drop_path_rate=... function tiny_vit_21m_512 (line 707) | def tiny_vit_21m_512(pretrained=False, num_classes=1000, drop_path_rate=... FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/RepViTSAM/repvit.py function _make_divisible (line 7) | def _make_divisible(v, divisor, min_value=None): class LayerNorm2d (line 32) | class LayerNorm2d(nn.Module): method __init__ (line 33) | def __init__(self, num_channels: int, eps: float = 1e-6) -> None: method forward (line 39) | def forward(self, x: torch.Tensor) -> torch.Tensor: class Conv2d_BN (line 46) | class Conv2d_BN(torch.nn.Sequential): method __init__ (line 47) | def __init__(self, a, b, ks=1, stride=1, pad=0, dilation=1, method fuse (line 57) | def fuse(self): class Residual (line 70) | class Residual(torch.nn.Module): method __init__ (line 71) | def __init__(self, m, drop=0.): method forward (line 76) | def forward(self, x): method fuse (line 84) | def fuse(self): class RepVGGDW (line 103) | class RepVGGDW(torch.nn.Module): method __init__ (line 104) | def __init__(self, ed) -> None: method forward (line 111) | def forward(self, x): method fuse (line 115) | def fuse(self): class RepViTBlock (line 144) | class RepViTBlock(nn.Module): method __init__ (line 145) | def __init__(self, inp, hidden_dim, oup, kernel_size, stride, use_se, ... method forward (line 188) | def forward(self, x): class BN_Linear (line 192) | class BN_Linear(torch.nn.Sequential): method __init__ (line 193) | def __init__(self, a, b, bias=True, std=0.02): method fuse (line 202) | def fuse(self): class Classfier (line 217) | class Classfier(nn.Module): method __init__ (line 218) | def __init__(self, dim, num_classes, distillation=True): method forward (line 225) | def forward(self, x): method fuse (line 235) | def fuse(self): class RepViT (line 247) | class RepViT(nn.Module): method __init__ (line 248) | def __init__(self, cfgs, num_classes=1000, distillation=False, img_siz... method forward (line 288) | def forward(self, x): function repvit (line 299) | def repvit(pretrained=False, num_classes = 1000, distillation=False, **k... FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/RepViTSAM/setup_repvit_sam.py function build_sam_repvit (line 13) | def build_sam_repvit(checkpoint=None): FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/grounded_edge_sam.py function segment (line 75) | def segment(sam_predictor: SamPredictor, image: np.ndarray, xyxy: np.nda... FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/grounded_efficient_sam.py function efficient_sam_box_prompt_segment (line 71) | def efficient_sam_box_prompt_segment(image, pts_sampled, model): FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/grounded_fast_sam.py function parse_args (line 9) | def parse_args(): function main (line 68) | def main(args): FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/grounded_light_hqsam.py function segment (line 77) | def segment(sam_predictor: SamPredictor, image: np.ndarray, xyxy: np.nda... FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/grounded_mobile_sam.py function parse_args (line 12) | def parse_args(): function main (line 42) | def main(args): FILE: data_generation/Grounded-Segment-Anything/EfficientSAM/grounded_repvit_sam.py function segment (line 75) | def segment(sam_predictor: SamPredictor, image: np.ndarray, xyxy: np.nda... FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/demo/gradio_app.py function load_model_hf (line 42) | def load_model_hf(model_config_path, repo_id, filename, device='cpu'): function image_transform_grounding (line 54) | def image_transform_grounding(init_image): function image_transform_grounding_for_vis (line 63) | def image_transform_grounding_for_vis(init_image): function run_grounding (line 72) | def run_grounding(input_image, grounding_caption, box_threshold, text_th... FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/demo/inference_on_a_image.py function plot_boxes_to_image (line 16) | def plot_boxes_to_image(image_pil, tgt): function load_image (line 57) | def load_image(image_path): function load_model (line 72) | def load_model(model_config_path, model_checkpoint_path, cpu_only=False): function get_grounding_output (line 83) | def get_grounding_output(model, image, caption, box_threshold, text_thre... FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/datasets/transforms.py function crop (line 17) | def crop(image, target, region): function hflip (line 68) | def hflip(image, target): function resize (line 87) | def resize(image, target, size, max_size=None): function pad (line 149) | def pad(image, target, padding): class ResizeDebug (line 162) | class ResizeDebug(object): method __init__ (line 163) | def __init__(self, size): method __call__ (line 166) | def __call__(self, img, target): class RandomCrop (line 170) | class RandomCrop(object): method __init__ (line 171) | def __init__(self, size): method __call__ (line 174) | def __call__(self, img, target): class RandomSizeCrop (line 179) | class RandomSizeCrop(object): method __init__ (line 180) | def __init__(self, min_size: int, max_size: int, respect_boxes: bool =... method __call__ (line 187) | def __call__(self, img: PIL.Image.Image, target: dict): class CenterCrop (line 204) | class CenterCrop(object): method __init__ (line 205) | def __init__(self, size): method __call__ (line 208) | def __call__(self, img, target): class RandomHorizontalFlip (line 216) | class RandomHorizontalFlip(object): method __init__ (line 217) | def __init__(self, p=0.5): method __call__ (line 220) | def __call__(self, img, target): class RandomResize (line 226) | class RandomResize(object): method __init__ (line 227) | def __init__(self, sizes, max_size=None): method __call__ (line 232) | def __call__(self, img, target=None): class RandomPad (line 237) | class RandomPad(object): method __init__ (line 238) | def __init__(self, max_pad): method __call__ (line 241) | def __call__(self, img, target): class RandomSelect (line 247) | class RandomSelect(object): method __init__ (line 253) | def __init__(self, transforms1, transforms2, p=0.5): method __call__ (line 258) | def __call__(self, img, target): class ToTensor (line 264) | class ToTensor(object): method __call__ (line 265) | def __call__(self, img, target): class RandomErasing (line 269) | class RandomErasing(object): method __init__ (line 270) | def __init__(self, *args, **kwargs): method __call__ (line 273) | def __call__(self, img, target): class Normalize (line 277) | class Normalize(object): method __init__ (line 278) | def __init__(self, mean, std): method __call__ (line 282) | def __call__(self, image, target=None): class Compose (line 296) | class Compose(object): method __init__ (line 297) | def __init__(self, transforms): method __call__ (line 300) | def __call__(self, image, target): method __repr__ (line 305) | def __repr__(self): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/backbone/backbone.py class FrozenBatchNorm2d (line 33) | class FrozenBatchNorm2d(torch.nn.Module): method __init__ (line 42) | def __init__(self, n): method _load_from_state_dict (line 49) | def _load_from_state_dict( method forward (line 60) | def forward(self, x): class BackboneBase (line 73) | class BackboneBase(nn.Module): method __init__ (line 74) | def __init__( method forward (line 107) | def forward(self, tensor_list: NestedTensor): class Backbone (line 119) | class Backbone(BackboneBase): method __init__ (line 122) | def __init__( class Joiner (line 146) | class Joiner(nn.Sequential): method __init__ (line 147) | def __init__(self, backbone, position_embedding): method forward (line 150) | def forward(self, tensor_list: NestedTensor): function build_backbone (line 162) | def build_backbone(args): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/backbone/position_encoding.py class PositionEmbeddingSine (line 30) | class PositionEmbeddingSine(nn.Module): method __init__ (line 36) | def __init__(self, num_pos_feats=64, temperature=10000, normalize=Fals... method forward (line 47) | def forward(self, tensor_list: NestedTensor): class PositionEmbeddingSineHW (line 78) | class PositionEmbeddingSineHW(nn.Module): method __init__ (line 84) | def __init__( method forward (line 98) | def forward(self, tensor_list: NestedTensor): class PositionEmbeddingLearned (line 134) | class PositionEmbeddingLearned(nn.Module): method __init__ (line 139) | def __init__(self, num_pos_feats=256): method reset_parameters (line 145) | def reset_parameters(self): method forward (line 149) | def forward(self, tensor_list: NestedTensor): function build_position_encoding (line 171) | def build_position_encoding(args): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/backbone/swin_transformer.py class Mlp (line 24) | class Mlp(nn.Module): method __init__ (line 27) | def __init__( method forward (line 38) | def forward(self, x): function window_partition (line 47) | def window_partition(x, window_size): function window_reverse (line 61) | def window_reverse(windows, window_size, H, W): class WindowAttention (line 77) | class WindowAttention(nn.Module): method __init__ (line 90) | def __init__( method forward (line 134) | def forward(self, x, mask=None): class SwinTransformerBlock (line 177) | class SwinTransformerBlock(nn.Module): method __init__ (line 194) | def __init__( method forward (line 238) | def forward(self, x, mask_matrix): class PatchMerging (line 301) | class PatchMerging(nn.Module): method __init__ (line 308) | def __init__(self, dim, norm_layer=nn.LayerNorm): method forward (line 314) | def forward(self, x, H, W): class BasicLayer (line 343) | class BasicLayer(nn.Module): method __init__ (line 361) | def __init__( method forward (line 409) | def forward(self, x, H, W): class PatchEmbed (line 459) | class PatchEmbed(nn.Module): method __init__ (line 468) | def __init__(self, patch_size=4, in_chans=3, embed_dim=96, norm_layer=... method forward (line 482) | def forward(self, x): class SwinTransformer (line 501) | class SwinTransformer(nn.Module): method __init__ (line 530) | def __init__( method _freeze_stages (line 636) | def _freeze_stages(self): method forward_raw (line 678) | def forward_raw(self, x): method forward (line 712) | def forward(self, tensor_list: NestedTensor): method train (line 756) | def train(self, mode=True): function build_swin_transformer (line 762) | def build_swin_transformer(modelname, pretrain_img_size, **kw): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/bertwarper.py class BertModelWarper (line 17) | class BertModelWarper(nn.Module): method __init__ (line 18) | def __init__(self, bert_model): method forward (line 31) | def forward( class TextEncoderShell (line 169) | class TextEncoderShell(nn.Module): method __init__ (line 170) | def __init__(self, text_encoder): method forward (line 175) | def forward(self, **kw): function generate_masks_with_special_tokens (line 180) | def generate_masks_with_special_tokens(tokenized, special_tokens_list, t... function generate_masks_with_special_tokens_and_transfer_map (line 224) | def generate_masks_with_special_tokens_and_transfer_map(tokenized, speci... FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/csrc/MsDeformAttn/ms_deform_attn.h function namespace (line 19) | namespace groundingdino { FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/csrc/MsDeformAttn/ms_deform_attn_cpu.cpp type groundingdino (line 16) | namespace groundingdino { function ms_deform_attn_cpu_forward (line 18) | at::Tensor function ms_deform_attn_cpu_backward (line 30) | std::vector FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/csrc/MsDeformAttn/ms_deform_attn_cpu.h function namespace (line 14) | namespace groundingdino { FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/csrc/MsDeformAttn/ms_deform_attn_cuda.h function namespace (line 14) | namespace groundingdino { FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/csrc/vision.cpp type groundingdino (line 5) | namespace groundingdino { function get_cuda_version (line 11) | std::string get_cuda_version() { function get_compiler_version (line 32) | std::string get_compiler_version() { function PYBIND11_MODULE (line 53) | PYBIND11_MODULE(TORCH_EXTENSION_NAME, m) { FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/fuse_modules.py class FeatureResizer (line 14) | class FeatureResizer(nn.Module): method __init__ (line 20) | def __init__(self, input_feat_size, output_feat_size, dropout, do_ln=T... method forward (line 28) | def forward(self, encoder_features): function l1norm (line 36) | def l1norm(X, dim, eps=1e-8): function l2norm (line 43) | def l2norm(X, dim, eps=1e-8): function func_attention (line 50) | def func_attention(query, context, smooth=1, raw_feature_norm="softmax",... class BiMultiHeadAttention (line 99) | class BiMultiHeadAttention(nn.Module): method __init__ (line 100) | def __init__(self, v_dim, l_dim, embed_dim, num_heads, dropout=0.1, cf... method _shape (line 129) | def _shape(self, tensor: torch.Tensor, seq_len: int, bsz: int): method _reset_parameters (line 132) | def _reset_parameters(self): method forward (line 146) | def forward(self, v, l, attention_mask_v=None, attention_mask_l=None): class BiAttentionBlock (line 252) | class BiAttentionBlock(nn.Module): method __init__ (line 253) | def __init__( method forward (line 286) | def forward(self, v, l, attention_mask_v=None, attention_mask_l=None): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/groundingdino.py class GroundingDINO (line 51) | class GroundingDINO(nn.Module): method __init__ (line 54) | def __init__( method _reset_parameters (line 203) | def _reset_parameters(self): method init_ref_points (line 209) | def init_ref_points(self, use_num_queries): method forward (line 212) | def forward(self, samples: NestedTensor, targets: List = None, **kw): method _set_aux_loss (line 352) | def _set_aux_loss(self, outputs_class, outputs_coord): function build_groundingdino (line 363) | def build_groundingdino(args): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/ms_deform_attn.py function _is_power_of_2 (line 35) | def _is_power_of_2(n): class MultiScaleDeformableAttnFunction (line 41) | class MultiScaleDeformableAttnFunction(Function): method forward (line 43) | def forward( method backward (line 72) | def backward(ctx, grad_output): function multi_scale_deformable_attn_pytorch (line 93) | def multi_scale_deformable_attn_pytorch( class MultiScaleDeformableAttention (line 136) | class MultiScaleDeformableAttention(nn.Module): method __init__ (line 154) | def __init__( method _reset_parameters (line 194) | def _reset_parameters(self): method init_weights (line 197) | def init_weights(self): method freeze_sampling_offsets (line 222) | def freeze_sampling_offsets(self): method freeze_attention_weights (line 227) | def freeze_attention_weights(self): method forward (line 232) | def forward( function create_dummy_class (line 363) | def create_dummy_class(klass, dependency, message=""): function create_dummy_func (line 392) | def create_dummy_func(func, dependency, message=""): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/transformer.py class Transformer (line 40) | class Transformer(nn.Module): method __init__ (line 41) | def __init__( method _reset_parameters (line 189) | def _reset_parameters(self): method get_valid_ratio (line 199) | def get_valid_ratio(self, mask): method init_ref_points (line 208) | def init_ref_points(self, use_num_queries): method forward (line 211) | def forward(self, srcs, masks, refpoint_embed, pos_embeds, tgt, attn_m... class TransformerEncoder (line 406) | class TransformerEncoder(nn.Module): method __init__ (line 407) | def __init__( method get_reference_points (line 466) | def get_reference_points(spatial_shapes, valid_ratios, device): method forward (line 482) | def forward( class TransformerDecoder (line 599) | class TransformerDecoder(nn.Module): method __init__ (line 600) | def __init__( method forward (line 634) | def forward( class DeformableTransformerEncoderLayer (line 739) | class DeformableTransformerEncoderLayer(nn.Module): method __init__ (line 740) | def __init__( method with_pos_embed (line 772) | def with_pos_embed(tensor, pos): method forward_ffn (line 775) | def forward_ffn(self, src): method forward (line 781) | def forward( class DeformableTransformerDecoderLayer (line 803) | class DeformableTransformerDecoderLayer(nn.Module): method __init__ (line 804) | def __init__( method rm_self_attn_modules (line 853) | def rm_self_attn_modules(self): method with_pos_embed (line 859) | def with_pos_embed(tensor, pos): method forward_ffn (line 862) | def forward_ffn(self, tgt): method forward (line 869) | def forward( function build_transformer (line 931) | def build_transformer(args): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/transformer_vanilla.py class TextTransformer (line 33) | class TextTransformer(nn.Module): method __init__ (line 34) | def __init__(self, num_layers, d_model=256, nheads=8, dim_feedforward=... method forward (line 47) | def forward(self, memory_text: torch.Tensor, text_attention_mask: torc... class TransformerEncoderLayer (line 72) | class TransformerEncoderLayer(nn.Module): method __init__ (line 73) | def __init__( method with_pos_embed (line 98) | def with_pos_embed(self, tensor, pos: Optional[Tensor]): method forward (line 101) | def forward( FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/GroundingDINO/utils.py function _get_clones (line 16) | def _get_clones(module, N, layer_share=False): function get_sine_pos_embed (line 24) | def get_sine_pos_embed( function gen_encoder_output_proposals (line 56) | def gen_encoder_output_proposals( class RandomBoxPerturber (line 120) | class RandomBoxPerturber: method __init__ (line 121) | def __init__( method __call__ (line 128) | def __call__(self, refanchors: Tensor) -> Tensor: function sigmoid_focal_loss (line 139) | def sigmoid_focal_loss( class MLP (line 172) | class MLP(nn.Module): method __init__ (line 175) | def __init__(self, input_dim, hidden_dim, output_dim, num_layers): method forward (line 183) | def forward(self, x): function _get_activation_fn (line 189) | def _get_activation_fn(activation, d_model=256, batch_dim=0): function gen_sineembed_for_position (line 205) | def gen_sineembed_for_position(pos_tensor): class ContrastiveEmbed (line 235) | class ContrastiveEmbed(nn.Module): method __init__ (line 236) | def __init__(self, max_text_len=256): method forward (line 244) | def forward(self, x, text_dict): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/__init__.py function build_model (line 11) | def build_model(args): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/models/registry.py class Registry (line 18) | class Registry(object): method __init__ (line 19) | def __init__(self, name): method __repr__ (line 23) | def __repr__(self): method __len__ (line 29) | def __len__(self): method name (line 33) | def name(self): method module_dict (line 37) | def module_dict(self): method get (line 40) | def get(self, key): method registe_with_name (line 43) | def registe_with_name(self, module_name=None, force=False): method register (line 46) | def register(self, module_build_function, module_name=None, force=False): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/util/box_ops.py function box_cxcywh_to_xyxy (line 9) | def box_cxcywh_to_xyxy(x): function box_xyxy_to_cxcywh (line 15) | def box_xyxy_to_cxcywh(x): function box_iou (line 22) | def box_iou(boxes1, boxes2): function generalized_box_iou (line 39) | def generalized_box_iou(boxes1, boxes2): function box_iou_pairwise (line 66) | def box_iou_pairwise(boxes1, boxes2): function generalized_box_iou_pairwise (line 82) | def generalized_box_iou_pairwise(boxes1, boxes2): function masks_to_boxes (line 107) | def masks_to_boxes(masks): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/util/get_tokenlizer.py function get_tokenlizer (line 4) | def get_tokenlizer(text_encoder_type): function get_pretrained_language_model (line 21) | def get_pretrained_language_model(text_encoder_type): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/util/inference.py function preprocess_caption (line 22) | def preprocess_caption(caption: str) -> str: function load_model (line 29) | def load_model(model_config_path: str, model_checkpoint_path: str, devic... function load_image (line 39) | def load_image(image_path: str) -> Tuple[np.array, torch.Tensor]: function predict (line 53) | def predict( function annotate (line 88) | def annotate(image_source: np.ndarray, boxes: torch.Tensor, logits: torc... class Model (line 111) | class Model: method __init__ (line 113) | def __init__( method predict_with_caption (line 126) | def predict_with_caption( method predict_with_classes (line 167) | def predict_with_classes( method preprocess_image (line 213) | def preprocess_image(image_bgr: np.ndarray) -> torch.Tensor: method post_process_result (line 226) | def post_process_result( method phrases2classes (line 238) | def phrases2classes(phrases: List[str], classes: List[str]) -> np.ndar... method find_index (line 249) | def find_index(string, lst): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/util/logger.py class _ColorfulFormatter (line 10) | class _ColorfulFormatter(logging.Formatter): method __init__ (line 11) | def __init__(self, *args, **kwargs): method formatMessage (line 18) | def formatMessage(self, record): function setup_logger (line 32) | def setup_logger(output=None, distributed_rank=0, *, color=True, name="i... function _cached_log_stream (line 92) | def _cached_log_stream(filename): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/util/misc.py class SmoothedValue (line 33) | class SmoothedValue(object): method __init__ (line 38) | def __init__(self, window_size=20, fmt=None): method update (line 46) | def update(self, value, n=1): method synchronize_between_processes (line 51) | def synchronize_between_processes(self): method median (line 65) | def median(self): method avg (line 72) | def avg(self): method global_avg (line 77) | def global_avg(self): method max (line 85) | def max(self): method value (line 89) | def value(self): method __str__ (line 92) | def __str__(self): function _get_global_gloo_group (line 103) | def _get_global_gloo_group(): function all_gather_cpu (line 115) | def all_gather_cpu(data): function all_gather (line 173) | def all_gather(data): function reduce_dict (line 220) | def reduce_dict(input_dict, average=True): class MetricLogger (line 247) | class MetricLogger(object): method __init__ (line 248) | def __init__(self, delimiter="\t"): method update (line 252) | def update(self, **kwargs): method __getattr__ (line 259) | def __getattr__(self, attr): method __str__ (line 266) | def __str__(self): method synchronize_between_processes (line 275) | def synchronize_between_processes(self): method add_meter (line 279) | def add_meter(self, name, meter): method log_every (line 282) | def log_every(self, iterable, print_freq, header=None, logger=None): function get_sha (line 362) | def get_sha(): function collate_fn (line 383) | def collate_fn(batch): function _max_by_axis (line 390) | def _max_by_axis(the_list): class NestedTensor (line 399) | class NestedTensor(object): method __init__ (line 400) | def __init__(self, tensors, mask: Optional[Tensor]): method imgsize (line 416) | def imgsize(self): method to (line 425) | def to(self, device): method to_img_list_single (line 436) | def to_img_list_single(self, tensor, mask): method to_img_list (line 443) | def to_img_list(self): method device (line 460) | def device(self): method decompose (line 463) | def decompose(self): method __repr__ (line 466) | def __repr__(self): method shape (line 470) | def shape(self): function nested_tensor_from_tensor_list (line 474) | def nested_tensor_from_tensor_list(tensor_list: List[Tensor]): function _onnx_nested_tensor_from_tensor_list (line 502) | def _onnx_nested_tensor_from_tensor_list(tensor_list: List[Tensor]) -> N... function setup_for_distributed (line 532) | def setup_for_distributed(is_master): function is_dist_avail_and_initialized (line 548) | def is_dist_avail_and_initialized(): function get_world_size (line 556) | def get_world_size(): function get_rank (line 562) | def get_rank(): function is_main_process (line 568) | def is_main_process(): function save_on_master (line 572) | def save_on_master(*args, **kwargs): function init_distributed_mode (line 577) | def init_distributed_mode(args): function accuracy (line 638) | def accuracy(output, target, topk=(1,)): function accuracy_onehot (line 657) | def accuracy_onehot(pred, gt): function interpolate (line 669) | def interpolate(input, size=None, scale_factor=None, mode="nearest", ali... class color_sys (line 687) | class color_sys: method __init__ (line 688) | def __init__(self, num_colors) -> None: method __call__ (line 700) | def __call__(self, idx): function inverse_sigmoid (line 704) | def inverse_sigmoid(x, eps=1e-3): function clean_state_dict (line 711) | def clean_state_dict(state_dict): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/util/slconfig.py function check_file_exist (line 21) | def check_file_exist(filename, msg_tmpl='file "{}" does not exist'): class ConfigDict (line 26) | class ConfigDict(Dict): method __missing__ (line 27) | def __missing__(self, name): method __getattr__ (line 30) | def __getattr__(self, name): class SLConfig (line 42) | class SLConfig(object): method _validate_py_syntax (line 68) | def _validate_py_syntax(filename): method _file2dict (line 77) | def _file2dict(filename): method _merge_a_into_b (line 140) | def _merge_a_into_b(a, b): method fromfile (line 184) | def fromfile(filename): method __init__ (line 188) | def __init__(self, cfg_dict=None, cfg_text=None, filename=None): method filename (line 209) | def filename(self): method text (line 213) | def text(self): method pretty_text (line 217) | def pretty_text(self): method __repr__ (line 310) | def __repr__(self): method __len__ (line 313) | def __len__(self): method __getattr__ (line 316) | def __getattr__(self, name): method __getitem__ (line 329) | def __getitem__(self, name): method __setattr__ (line 332) | def __setattr__(self, name, value): method __setitem__ (line 337) | def __setitem__(self, name, value): method __iter__ (line 342) | def __iter__(self): method dump (line 345) | def dump(self, file=None): method merge_from_dict (line 353) | def merge_from_dict(self, options): method __setstate__ (line 386) | def __setstate__(self, state): method copy (line 389) | def copy(self): method deepcopy (line 392) | def deepcopy(self): class DictAction (line 396) | class DictAction(Action): method _parse_int_float_bool (line 404) | def _parse_int_float_bool(val): method __call__ (line 419) | def __call__(self, parser, namespace, values, option_string=None): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/util/slio.py class BaseFileHandler (line 23) | class BaseFileHandler(metaclass=ABCMeta): method load_from_fileobj (line 25) | def load_from_fileobj(self, file, **kwargs): method dump_to_fileobj (line 29) | def dump_to_fileobj(self, obj, file, **kwargs): method dump_to_str (line 33) | def dump_to_str(self, obj, **kwargs): method load_from_path (line 36) | def load_from_path(self, filepath, mode="r", **kwargs): method dump_to_path (line 40) | def dump_to_path(self, obj, filepath, mode="w", **kwargs): class JsonHandler (line 45) | class JsonHandler(BaseFileHandler): method load_from_fileobj (line 46) | def load_from_fileobj(self, file): method dump_to_fileobj (line 49) | def dump_to_fileobj(self, obj, file, **kwargs): method dump_to_str (line 52) | def dump_to_str(self, obj, **kwargs): class PickleHandler (line 56) | class PickleHandler(BaseFileHandler): method load_from_fileobj (line 57) | def load_from_fileobj(self, file, **kwargs): method load_from_path (line 60) | def load_from_path(self, filepath, **kwargs): method dump_to_str (line 63) | def dump_to_str(self, obj, **kwargs): method dump_to_fileobj (line 67) | def dump_to_fileobj(self, obj, file, **kwargs): method dump_to_path (line 71) | def dump_to_path(self, obj, filepath, **kwargs): class YamlHandler (line 75) | class YamlHandler(BaseFileHandler): method load_from_fileobj (line 76) | def load_from_fileobj(self, file, **kwargs): method dump_to_fileobj (line 80) | def dump_to_fileobj(self, obj, file, **kwargs): method dump_to_str (line 84) | def dump_to_str(self, obj, **kwargs): function is_str (line 102) | def is_str(x): function slload (line 110) | def slload(file, file_format=None, **kwargs): function sldump (line 143) | def sldump(obj, file=None, file_format=None, **kwargs): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/util/time_counter.py class TimeCounter (line 5) | class TimeCounter: method __init__ (line 6) | def __init__(self) -> None: method clear (line 9) | def clear(self): method timeit (line 13) | def timeit(self, name): class TimeHolder (line 19) | class TimeHolder: method __init__ (line 20) | def __init__(self) -> None: method update (line 23) | def update(self, _timedict: dict): method final_res (line 29) | def final_res(self): method __str__ (line 32) | def __str__(self): class AverageMeter (line 36) | class AverageMeter(object): method __init__ (line 39) | def __init__(self, name, fmt=":f", val_only=False): method reset (line 45) | def reset(self): method update (line 51) | def update(self, val, n=1): method __str__ (line 57) | def __str__(self): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/util/utils.py function slprint (line 15) | def slprint(x, name="x"): function clean_state_dict (line 29) | def clean_state_dict(state_dict): function renorm (line 38) | def renorm( class CocoClassMapper (line 66) | class CocoClassMapper: method __init__ (line 67) | def __init__(self) -> None: method origin2compact (line 153) | def origin2compact(self, idx): method compact2origin (line 156) | def compact2origin(self, idx): function to_device (line 160) | def to_device(item, device): function get_gaussian_mean (line 174) | def get_gaussian_mean(x, axis, other_axis, softmax=True): function get_expected_points_from_map (line 200) | def get_expected_points_from_map(hm, softmax=True): class Embedder (line 222) | class Embedder: method __init__ (line 223) | def __init__(self, **kwargs): method create_embedding_fn (line 227) | def create_embedding_fn(self): method embed (line 251) | def embed(self, inputs): function get_embedder (line 255) | def get_embedder(multires, i=0): class APOPMeter (line 275) | class APOPMeter: method __init__ (line 276) | def __init__(self) -> None: method update (line 282) | def update(self, pred, gt): method update_cm (line 293) | def update_cm(self, tp, fp, tn, fn): function inverse_sigmoid (line 300) | def inverse_sigmoid(x, eps=1e-5): function get_raw_dict (line 307) | def get_raw_dict(args): function stat_tensors (line 325) | def stat_tensors(tensor): class NiceRepr (line 340) | class NiceRepr: method __nice__ (line 374) | def __nice__(self): method __repr__ (line 384) | def __repr__(self): method __str__ (line 394) | def __str__(self): function ensure_rng (line 405) | def ensure_rng(rng=None): function random_boxes (line 436) | def random_boxes(num=1, scale=1, rng=None): class ModelEma (line 473) | class ModelEma(torch.nn.Module): method __init__ (line 474) | def __init__(self, model, decay=0.9997, device=None): method _update (line 487) | def _update(self, model, update_fn): method update (line 496) | def update(self, model): method set (line 499) | def set(self, model): class BestMetricSingle (line 503) | class BestMetricSingle: method __init__ (line 504) | def __init__(self, init_res=0.0, better="large") -> None: method isbetter (line 512) | def isbetter(self, new_res, old_res): method update (line 518) | def update(self, new_res, ep): method __str__ (line 525) | def __str__(self) -> str: method __repr__ (line 528) | def __repr__(self) -> str: method summary (line 531) | def summary(self) -> dict: class BestMetricHolder (line 538) | class BestMetricHolder: method __init__ (line 539) | def __init__(self, init_res=0.0, better="large", use_ema=False) -> None: method update (line 546) | def update(self, new_res, epoch, is_ema=False): method summary (line 560) | def summary(self): method __repr__ (line 570) | def __repr__(self) -> str: method __str__ (line 573) | def __str__(self) -> str: function targets_to (line 577) | def targets_to(targets: List[Dict[str, Any]], device): function get_phrases_from_posmap (line 599) | def get_phrases_from_posmap( FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/util/visualizer.py function renorm (line 22) | def renorm( class ColorMap (line 50) | class ColorMap: method __init__ (line 51) | def __init__(self, basergb=[255, 255, 0]): method __call__ (line 54) | def __call__(self, attnmap): function rainbow_text (line 66) | def rainbow_text(x, y, ls, lc, **kw): class COCOVisualizer (line 95) | class COCOVisualizer: method __init__ (line 96) | def __init__(self, coco=None, tokenlizer=None) -> None: method visualize (line 99) | def visualize(self, img, tgt, caption=None, dpi=180, savedir="vis"): method addtgt (line 135) | def addtgt(self, tgt): method showAnns (line 225) | def showAnns(self, anns, draw_bbox=False): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/groundingdino/util/vl_utils.py function create_positive_map_from_span (line 8) | def create_positive_map_from_span(tokenized, token_span, max_text_len=256): function build_captions_and_token_span (line 49) | def build_captions_and_token_span(cat_list, force_lowercase): function build_id2posspan_and_caption (line 90) | def build_id2posspan_and_caption(category_dict: dict): FILE: data_generation/Grounded-Segment-Anything/GroundingDINO/setup.py function write_version_file (line 44) | def write_version_file(): function get_extensions (line 56) | def get_extensions(): function parse_requirements (line 114) | def parse_requirements(fname="requirements.txt", with_version=True): FILE: data_generation/Grounded-Segment-Anything/automatic_label_demo.py function load_image (line 33) | def load_image(image_path): function generate_caption (line 48) | def generate_caption(raw_image, device): function generate_tags (line 59) | def generate_tags(caption, split=',', max_tokens=100, model="gpt-3.5-tur... function check_caption (line 82) | def check_caption(caption, pred_phrases, max_tokens=100, model="gpt-3.5-... function load_model (line 107) | def load_model(model_config_path, model_checkpoint_path, device): function get_grounding_output (line 118) | def get_grounding_output(model, image, caption, box_threshold, text_thre... function show_mask (line 153) | def show_mask(mask, ax, random_color=False): function show_box (line 163) | def show_box(box, ax, label): function save_mask_data (line 170) | def save_mask_data(output_dir, caption, mask_list, box_list, label_list): FILE: data_generation/Grounded-Segment-Anything/automatic_label_ram_demo.py function load_image (line 36) | def load_image(image_path): function check_tags_chinese (line 51) | def check_tags_chinese(tags_chinese, pred_phrases, max_tokens=100, model... function load_model (line 76) | def load_model(model_config_path, model_checkpoint_path, device): function get_grounding_output (line 87) | def get_grounding_output(model, image, caption, box_threshold, text_thre... function show_mask (line 122) | def show_mask(mask, ax, random_color=False): function show_box (line 132) | def show_box(box, ax, label): function save_mask_data (line 139) | def save_mask_data(output_dir, tags_chinese, mask_list, box_list, label_... FILE: data_generation/Grounded-Segment-Anything/automatic_label_simple_demo.py function segment (line 135) | def segment(sam_predictor: SamPredictor, image: np.ndarray, xyxy: np.nda... FILE: data_generation/Grounded-Segment-Anything/automatic_label_tag2text_demo.py function load_image (line 34) | def load_image(image_path): function generate_caption (line 49) | def generate_caption(raw_image, device): function generate_tags (line 60) | def generate_tags(caption, split=',', max_tokens=100, model="gpt-3.5-tur... function check_caption (line 83) | def check_caption(caption, pred_phrases, max_tokens=100, model="gpt-3.5-... function load_model (line 108) | def load_model(model_config_path, model_checkpoint_path, device): function get_grounding_output (line 119) | def get_grounding_output(model, image, caption, box_threshold, text_thre... function show_mask (line 154) | def show_mask(mask, ax, random_color=False): function show_box (line 164) | def show_box(box, ax, label): function save_mask_data (line 171) | def save_mask_data(output_dir, caption, mask_list, box_list, label_list): FILE: data_generation/Grounded-Segment-Anything/chatbot.py function seed_everything (line 131) | def seed_everything(seed): function prompts (line 139) | def prompts(name, description): function blend_gt2pt (line 148) | def blend_gt2pt(old_image, new_image, sigma=0.15, steps=100): function cut_dialogue_history (line 205) | def cut_dialogue_history(history_memory, keep_last_n_words=500): function get_new_image_name (line 221) | def get_new_image_name(org_img_name, func_name="update"): class MaskFormer (line 238) | class MaskFormer: method __init__ (line 239) | def __init__(self, device): method inference (line 245) | def inference(self, image_path, text): class ImageEditing (line 268) | class ImageEditing: method __init__ (line 269) | def __init__(self, device): method inference_replace (line 283) | def inference_replace(self, inputs): class InstructPix2Pix (line 299) | class InstructPix2Pix: method __init__ (line 300) | def __init__(self, device): method inference (line 314) | def inference(self, inputs): class Text2Image (line 327) | class Text2Image: method __init__ (line 328) | def __init__(self, device): method inference (line 343) | def inference(self, text): class ImageCaptioning (line 353) | class ImageCaptioning: method __init__ (line 354) | def __init__(self, device): method inference (line 365) | def inference(self, image_path): class Image2Canny (line 373) | class Image2Canny: method __init__ (line 374) | def __init__(self, device): method inference (line 384) | def inference(self, inputs): class CannyText2Image (line 397) | class CannyText2Image: method __init__ (line 398) | def __init__(self, device): method inference (line 419) | def inference(self, inputs): class Image2Line (line 434) | class Image2Line: method __init__ (line 435) | def __init__(self, device): method inference (line 444) | def inference(self, inputs): class LineText2Image (line 453) | class LineText2Image: method __init__ (line 454) | def __init__(self, device): method inference (line 477) | def inference(self, inputs): class Image2Hed (line 492) | class Image2Hed: method __init__ (line 493) | def __init__(self, device): method inference (line 502) | def inference(self, inputs): class HedText2Image (line 511) | class HedText2Image: method __init__ (line 512) | def __init__(self, device): method inference (line 535) | def inference(self, inputs): class Image2Scribble (line 550) | class Image2Scribble: method __init__ (line 551) | def __init__(self, device): method inference (line 560) | def inference(self, inputs): class ScribbleText2Image (line 569) | class ScribbleText2Image: method __init__ (line 570) | def __init__(self, device): method inference (line 591) | def inference(self, inputs): class Image2Pose (line 606) | class Image2Pose: method __init__ (line 607) | def __init__(self, device): method inference (line 615) | def inference(self, inputs): class PoseText2Image (line 624) | class PoseText2Image: method __init__ (line 625) | def __init__(self, device): method inference (line 649) | def inference(self, inputs): class Image2Seg (line 664) | class Image2Seg: method __init__ (line 665) | def __init__(self, device): method inference (line 713) | def inference(self, inputs): class SegText2Image (line 731) | class SegText2Image: method __init__ (line 732) | def __init__(self, device): method inference (line 753) | def inference(self, inputs): class Image2Depth (line 768) | class Image2Depth: method __init__ (line 769) | def __init__(self, device): method inference (line 777) | def inference(self, inputs): class DepthText2Image (line 790) | class DepthText2Image: method __init__ (line 791) | def __init__(self, device): method inference (line 812) | def inference(self, inputs): class Image2Normal (line 827) | class Image2Normal: method __init__ (line 828) | def __init__(self, device): method inference (line 837) | def inference(self, inputs): class NormalText2Image (line 861) | class NormalText2Image: method __init__ (line 862) | def __init__(self, device): method inference (line 883) | def inference(self, inputs): class VisualQuestionAnswering (line 898) | class VisualQuestionAnswering: method __init__ (line 899) | def __init__(self, device): method inference (line 911) | def inference(self, inputs): class InfinityOutPainting (line 922) | class InfinityOutPainting: method __init__ (line 924) | def __init__(self, ImageCaptioning, ImageEditing, VisualQuestionAnswer... method get_BLIP_vqa (line 933) | def get_BLIP_vqa(self, image, question): method get_BLIP_caption (line 941) | def get_BLIP_caption(self, image): method check_prompt (line 948) | def check_prompt(self, prompt): method get_imagine_caption (line 954) | def get_imagine_caption(self, image, imagine): method resize_image (line 971) | def resize_image(self, image, max_size=1000000, multiple=8): method dowhile (line 978) | def dowhile(self, original_img, tosize, expand_ratio, imagine, usr_pro... method inference (line 1006) | def inference(self, inputs): class Grounded_dino_sam_inpainting (line 1020) | class Grounded_dino_sam_inpainting: method __init__ (line 1021) | def __init__(self, device): method inference_caption (line 1045) | def inference_caption(self, image_path): method _detect_object (line 1052) | def _detect_object(self, image_path, text_prompt, func_name): method inference_detect_one_object (line 1079) | def inference_detect_one_object(self, inputs): method inference_detect_multi_object (line 1093) | def inference_detect_multi_object(self, inputs): method inference_segment_anything (line 1106) | def inference_segment_anything(self, image_path): method _segment_object (line 1129) | def _segment_object(self, image_path, text_prompt, func_name): method inference_segment_one_object (line 1187) | def inference_segment_one_object(self, inputs): method inference_segment_multi_object (line 1201) | def inference_segment_multi_object(self, inputs): method inference_auto_segment_object (line 1213) | def inference_auto_segment_object(self, image_path): method _inpainting (line 1227) | def _inpainting(self, image_path, to_be_replaced_txt, replace_with_txt... method inference_replace (line 1267) | def inference_replace(self, inputs): class ConversationBot (line 1277) | class ConversationBot: method __init__ (line 1278) | def __init__(self, load_dict): method run_text (line 1306) | def run_text(self, text, state): method run_image (line 1316) | def run_image(self, image, state, txt, lang): method init_agent (line 1350) | def init_agent(self, openai_api_key, lang): function speech_recognition (line 1373) | def speech_recognition(speech_file): FILE: data_generation/Grounded-Segment-Anything/gradio_app.py function show_anns (line 34) | def show_anns(anns): function generate_caption (line 59) | def generate_caption(processor, blip_model, raw_image): function generate_tags (line 66) | def generate_tags(caption, split=',', max_tokens=100, model="gpt-3.5-tur... function transform_image (line 83) | def transform_image(image_pil): function load_model (line 96) | def load_model(model_config_path, model_checkpoint_path, device): function get_grounding_output (line 107) | def get_grounding_output(model, image, caption, box_threshold, text_thre... function draw_mask (line 143) | def draw_mask(mask, draw, random_color=False): function draw_box (line 154) | def draw_box(box, draw, label): function run_grounded_sam (line 189) | def run_grounded_sam(input_image, text_prompt, task_type, inpaint_prompt... FILE: data_generation/Grounded-Segment-Anything/grounded_sam_demo.py function load_image (line 32) | def load_image(image_path): function load_model (line 47) | def load_model(model_config_path, model_checkpoint_path, device): function get_grounding_output (line 58) | def get_grounding_output(model, image, caption, box_threshold, text_thre... function show_mask (line 93) | def show_mask(mask, ax, random_color=False): function show_box (line 103) | def show_box(box, ax, label): function save_mask_data (line 110) | def save_mask_data(output_dir, mask_list, box_list, label_list): FILE: data_generation/Grounded-Segment-Anything/grounded_sam_inpainting_demo.py function load_image (line 31) | def load_image(image_path): function load_model (line 46) | def load_model(model_config_path, model_checkpoint_path, device): function get_grounding_output (line 57) | def get_grounding_output(model, image, caption, box_threshold, text_thre... function show_mask (line 92) | def show_mask(mask, ax, random_color=False): function show_box (line 102) | def show_box(box, ax, label): FILE: data_generation/Grounded-Segment-Anything/grounded_sam_osx_demo.py function load_image (line 35) | def load_image(image_path): function load_model (line 50) | def load_model(model_config_path, model_checkpoint_path, device): function get_grounding_output (line 61) | def get_grounding_output(model, image, caption, box_threshold, text_thre... function show_mask (line 97) | def show_mask(mask, ax, random_color=False): function show_box (line 106) | def show_box(box, ax, label): function save_mask_data (line 116) | def save_mask_data(output_dir, mask_list, box_list, label_list): function bbox_resize (line 144) | def bbox_resize(bbox, scale=1.0): function mesh_recovery (line 150) | def mesh_recovery(original_img, bboxes): FILE: data_generation/Grounded-Segment-Anything/grounded_sam_simple_demo.py function segment (line 76) | def segment(sam_predictor: SamPredictor, image: np.ndarray, xyxy: np.nda... FILE: data_generation/Grounded-Segment-Anything/grounded_sam_visam.py class Colors (line 28) | class Colors: method __init__ (line 30) | def __init__(self): method __call__ (line 37) | def __call__(self, i, bgr=False): method hex2rgb (line 42) | def hex2rgb(h): # rgb order (PIL) class ListImgDataset (line 49) | class ListImgDataset(Dataset): method __init__ (line 50) | def __init__(self, mot_path, img_list, det_db) -> None: method load_img_from_file (line 64) | def load_img_from_file(self, f_path): method init_img (line 79) | def init_img(self, img, proposals): method __len__ (line 92) | def __len__(self): method __getitem__ (line 95) | def __getitem__(self, index): class Detector (line 100) | class Detector(object): method __init__ (line 101) | def __init__(self, args, model, vid, sam_predictor=None): method filter_dt_by_score (line 123) | def filter_dt_by_score(dt_instances: Instances, prob_threshold: float)... method filter_dt_by_area (line 129) | def filter_dt_by_area(dt_instances: Instances, area_threshold: float) ... method detect (line 135) | def detect(self, prob_threshold=0.6, area_threshold=100, vis=False): class RuntimeTrackerBase (line 205) | class RuntimeTrackerBase(object): method __init__ (line 206) | def __init__(self, score_thresh=0.6, filter_score_thresh=0.5, miss_tol... method clear (line 212) | def clear(self): method update (line 215) | def update(self, track_instances: Instances): FILE: data_generation/Grounded-Segment-Anything/grounded_sam_whisper_demo.py function load_image (line 28) | def load_image(image_path): function load_model (line 43) | def load_model(model_config_path, model_checkpoint_path, device): function get_grounding_output (line 54) | def get_grounding_output(model, image, caption, box_threshold, text_thre... function show_mask (line 88) | def show_mask(mask, ax, random_color=False): function show_box (line 98) | def show_box(box, ax, label): function save_mask_data (line 105) | def save_mask_data(output_dir, mask_list, box_list, label_list): function speech_recognition (line 134) | def speech_recognition(speech_file, model): FILE: data_generation/Grounded-Segment-Anything/grounded_sam_whisper_inpainting_demo.py function load_image (line 38) | def load_image(image_path): function load_model (line 53) | def load_model(model_config_path, model_checkpoint_path, device): function get_grounding_output (line 64) | def get_grounding_output(model, image, caption, box_threshold, text_thre... function show_mask (line 99) | def show_mask(mask, ax, random_color=False): function show_box (line 109) | def show_box(box, ax, label): function speech_recognition (line 116) | def speech_recognition(speech_file, model): function filter_prompts_with_chatgpt (line 138) | def filter_prompts_with_chatgpt(caption, max_tokens=100, model="gpt-3.5-... FILE: data_generation/Grounded-Segment-Anything/playground/DeepFloyd/inpaint.py function download_image (line 10) | def download_image(url): FILE: data_generation/Grounded-Segment-Anything/playground/ImageBind_SAM/audio_referring_seg_demo.py function retriev_vision_and_audio (line 57) | def retriev_vision_and_audio(elements, audio_list): FILE: data_generation/Grounded-Segment-Anything/playground/ImageBind_SAM/data.py function waveform2melspec (line 29) | def waveform2melspec(waveform, sample_rate, num_mel_bins, target_length): function get_clip_timepoints (line 68) | def get_clip_timepoints(clip_sampler, duration): function load_and_transform_vision_data_from_pil_image (line 78) | def load_and_transform_vision_data_from_pil_image(img_list, device): function load_and_transform_vision_data (line 102) | def load_and_transform_vision_data(image_paths, device): function load_and_transform_text (line 129) | def load_and_transform_text(text, device): function load_and_transform_audio_data (line 138) | def load_and_transform_audio_data( function get_clip_timepoints (line 188) | def get_clip_timepoints(clip_sampler, duration): function crop_boxes (line 199) | def crop_boxes(boxes, x_offset, y_offset): function uniform_crop (line 218) | def uniform_crop(images, size, spatial_idx, boxes=None, scale_size=None): class SpatialCrop (line 277) | class SpatialCrop(nn.Module): method __init__ (line 286) | def __init__(self, crop_size: int = 224, num_crops: int = 3): method forward (line 298) | def forward(self, videos): function load_and_transform_video_data (line 320) | def load_and_transform_video_data( FILE: data_generation/Grounded-Segment-Anything/playground/ImageBind_SAM/demo.py function retriev_vision_and_text (line 58) | def retriev_vision_and_text(elements, text_list): function retriev_vision_and_audio (line 68) | def retriev_vision_and_audio(elements, audio_list): FILE: data_generation/Grounded-Segment-Anything/playground/ImageBind_SAM/image_referring_seg_demo.py function retriev_vision_and_vision (line 65) | def retriev_vision_and_vision(elements): FILE: data_generation/Grounded-Segment-Anything/playground/ImageBind_SAM/models/helpers.py class Normalize (line 17) | class Normalize(nn.Module): method __init__ (line 18) | def __init__(self, dim: int) -> None: method forward (line 22) | def forward(self, x): class LearnableLogitScaling (line 26) | class LearnableLogitScaling(nn.Module): method __init__ (line 27) | def __init__( method forward (line 43) | def forward(self, x): method extra_repr (line 46) | def extra_repr(self): class EinOpsRearrange (line 51) | class EinOpsRearrange(nn.Module): method __init__ (line 52) | def __init__(self, rearrange_expr: str, **kwargs) -> None: method forward (line 57) | def forward(self, x): class VerboseNNModule (line 62) | class VerboseNNModule(nn.Module): method get_readable_tensor_repr (line 68) | def get_readable_tensor_repr(name: str, tensor: torch.Tensor) -> str: method extra_repr (line 81) | def extra_repr(self) -> str: function cast_if_src_dtype (line 100) | def cast_if_src_dtype( class QuickGELU (line 110) | class QuickGELU(nn.Module): method forward (line 112) | def forward(self, x: torch.Tensor): class SelectElement (line 116) | class SelectElement(nn.Module): method __init__ (line 117) | def __init__(self, index) -> None: method forward (line 121) | def forward(self, x): class SelectEOSAndProject (line 126) | class SelectEOSAndProject(nn.Module): method __init__ (line 131) | def __init__(self, proj: nn.Module) -> None: method forward (line 135) | def forward(self, x, seq_len): FILE: data_generation/Grounded-Segment-Anything/playground/ImageBind_SAM/models/imagebind_model.py class ImageBindModel (line 48) | class ImageBindModel(nn.Module): method __init__ (line 49) | def __init__( method _create_modality_preprocessors (line 142) | def _create_modality_preprocessors( method _create_modality_trunks (line 276) | def _create_modality_trunks( method _create_modality_heads (line 377) | def _create_modality_heads( method _create_modality_postprocessors (line 429) | def _create_modality_postprocessors(self, out_embed_dim): method forward (line 455) | def forward(self, inputs): function imagebind_huge (line 490) | def imagebind_huge(pretrained=False): FILE: data_generation/Grounded-Segment-Anything/playground/ImageBind_SAM/models/multimodal_preprocessors.py function get_sinusoid_encoding_table (line 27) | def get_sinusoid_encoding_table(n_position, d_hid): function interpolate_pos_encoding_2d (line 46) | def interpolate_pos_encoding_2d(target_spatial_size, pos_embed): function interpolate_pos_encoding (line 66) | def interpolate_pos_encoding( function _get_pos_embedding (line 105) | def _get_pos_embedding( class PatchEmbedGeneric (line 122) | class PatchEmbedGeneric(nn.Module): method __init__ (line 127) | def __init__(self, proj_stem, norm_layer: Optional[nn.Module] = None): method get_patch_layout (line 138) | def get_patch_layout(self, img_size): method forward (line 152) | def forward(self, x): class SpatioTemporalPosEmbeddingHelper (line 161) | class SpatioTemporalPosEmbeddingHelper(VerboseNNModule): method __init__ (line 162) | def __init__( method get_pos_embedding (line 184) | def get_pos_embedding(self, vision_input, all_vision_tokens): class RGBDTPreprocessor (line 196) | class RGBDTPreprocessor(VerboseNNModule): method __init__ (line 197) | def __init__( method init_parameters (line 237) | def init_parameters(self, init_param_style): method tokenize_input_and_cls_pos (line 256) | def tokenize_input_and_cls_pos(self, input, stem, mask): method forward (line 274) | def forward(self, vision=None, depth=None, patch_mask=None): class AudioPreprocessor (line 302) | class AudioPreprocessor(RGBDTPreprocessor): method __init__ (line 303) | def __init__(self, audio_stem: PatchEmbedGeneric, **kwargs) -> None: method forward (line 306) | def forward(self, audio=None): class ThermalPreprocessor (line 310) | class ThermalPreprocessor(RGBDTPreprocessor): method __init__ (line 311) | def __init__(self, thermal_stem: PatchEmbedGeneric, **kwargs) -> None: method forward (line 314) | def forward(self, thermal=None): function build_causal_attention_mask (line 318) | def build_causal_attention_mask(context_length): class TextPreprocessor (line 327) | class TextPreprocessor(VerboseNNModule): method __init__ (line 328) | def __init__( method init_parameters (line 363) | def init_parameters(self, init_param_style="openclip"): method forward (line 379) | def forward(self, text): class Im2Video (line 407) | class Im2Video(nn.Module): method __init__ (line 410) | def __init__(self, time_dim=2): method forward (line 414) | def forward(self, x): class PadIm2Video (line 424) | class PadIm2Video(Im2Video): method __init__ (line 425) | def __init__(self, ntimes, pad_type, time_dim=2): method forward (line 432) | def forward(self, x): function bytes_to_unicode (line 448) | def bytes_to_unicode(): function get_pairs (line 474) | def get_pairs(word): function basic_clean (line 486) | def basic_clean(text): function whitespace_clean (line 492) | def whitespace_clean(text): class SimpleTokenizer (line 498) | class SimpleTokenizer(object): method __init__ (line 499) | def __init__(self, bpe_path: str, context_length=77): method bpe (line 526) | def bpe(self, token): method encode (line 567) | def encode(self, text): method decode (line 577) | def decode(self, tokens): method __call__ (line 586) | def __call__(self, texts, context_length=None): class IMUPreprocessor (line 607) | class IMUPreprocessor(VerboseNNModule): method __init__ (line 608) | def __init__( method init_parameters (line 637) | def init_parameters(self, init_param_style): method tokenize_input_and_cls_pos (line 652) | def tokenize_input_and_cls_pos(self, input, stem): method forward (line 667) | def forward(self, imu): FILE: data_generation/Grounded-Segment-Anything/playground/ImageBind_SAM/models/transformer.py class Attention (line 27) | class Attention(nn.Module): method __init__ (line 28) | def __init__( method forward (line 49) | def forward(self, x): class Mlp (line 72) | class Mlp(nn.Module): method __init__ (line 73) | def __init__( method forward (line 89) | def forward(self, x): class MultiheadAttention (line 98) | class MultiheadAttention(nn.MultiheadAttention): method forward (line 99) | def forward(self, x: torch.Tensor, attn_mask: torch.Tensor): class ViTAttention (line 103) | class ViTAttention(Attention): method forward (line 104) | def forward(self, x: torch.Tensor, attn_mask: torch.Tensor): class BlockWithMasking (line 109) | class BlockWithMasking(nn.Module): method __init__ (line 110) | def __init__( method forward (line 163) | def forward(self, x: torch.Tensor, attn_mask: torch.Tensor): class SimpleTransformer (line 180) | class SimpleTransformer(nn.Module): method __init__ (line 181) | def __init__( method _init_weights (line 234) | def _init_weights(self, m): method forward (line 249) | def forward( FILE: data_generation/Grounded-Segment-Anything/playground/ImageBind_SAM/text_referring_seg_demo.py function retriev_vision_and_text (line 57) | def retriev_vision_and_text(elements, text_list): FILE: data_generation/Grounded-Segment-Anything/playground/ImageBind_SAM/utils.py function segment_image (line 4) | def segment_image(image, segmentation_mask): function convert_box_xywh_to_xyxy (line 17) | def convert_box_xywh_to_xyxy(box): function get_indices_of_values_above_threshold (line 25) | def get_indices_of_values_above_threshold(values, threshold): FILE: data_generation/Grounded-Segment-Anything/playground/LaMa/lama_inpaint_demo.py function download_image (line 9) | def download_image(url): FILE: data_generation/Grounded-Segment-Anything/playground/LaMa/sam_lama.py function download_image (line 18) | def download_image(url): function dilate_mask (line 73) | def dilate_mask(mask, dilate_factor=15): function save_array_to_img (line 82) | def save_array_to_img(img_arr, img_p): FILE: data_generation/Grounded-Segment-Anything/playground/PaintByExample/paint_by_example.py function download_image (line 13) | def download_image(url): FILE: data_generation/Grounded-Segment-Anything/playground/PaintByExample/sam_paint_by_example.py function download_image (line 16) | def download_image(url): FILE: data_generation/Grounded-Segment-Anything/playground/RePaint/repaint.py function download_image (line 10) | def download_image(url): FILE: data_generation/Grounded-Segment-Anything/predict.py class ModelOutput (line 42) | class ModelOutput(BaseModel): class Predictor (line 49) | class Predictor(BasePredictor): method setup (line 50) | def setup(self): method predict (line 87) | def predict( function get_grounding_output (line 205) | def get_grounding_output( function load_image (line 244) | def load_image(image_path): function load_model (line 259) | def load_model(model_config_path, model_checkpoint_path, device): function show_mask (line 272) | def show_mask(mask, ax, random_color=False): function show_box (line 282) | def show_box(box, ax, label): FILE: data_generation/Grounded-Segment-Anything/segment_anything/scripts/amg.py function write_masks_to_folder (line 152) | def write_masks_to_folder(masks: List[Dict[str, Any]], path: str) -> None: function get_amg_kwargs (line 177) | def get_amg_kwargs(args): function main (line 195) | def main(args: argparse.Namespace) -> None: FILE: data_generation/Grounded-Segment-Anything/segment_anything/scripts/export_onnx_model.py function run_export (line 97) | def run_export( function to_numpy (line 173) | def to_numpy(tensor): FILE: data_generation/Grounded-Segment-Anything/segment_anything/segment_anything/automatic_mask_generator.py class SamAutomaticMaskGenerator (line 35) | class SamAutomaticMaskGenerator: method __init__ (line 36) | def __init__( method generate (line 137) | def generate(self, image: np.ndarray) -> List[Dict[str, Any]]: method _generate_masks (line 197) | def _generate_masks(self, image: np.ndarray) -> MaskData: method _process_crop (line 225) | def _process_crop( method _process_batch (line 266) | def _process_batch( method postprocess_small_regions (line 324) | def postprocess_small_regions( FILE: data_generation/Grounded-Segment-Anything/segment_anything/segment_anything/build_sam.py function build_sam_vit_h (line 14) | def build_sam_vit_h(checkpoint=None): function build_sam_vit_l (line 27) | def build_sam_vit_l(checkpoint=None): function build_sam_vit_b (line 37) | def build_sam_vit_b(checkpoint=None): function _build_sam (line 55) | def _build_sam( FILE: data_generation/Grounded-Segment-Anything/segment_anything/segment_anything/build_sam_hq.py function build_sam_hq_vit_h (line 14) | def build_sam_hq_vit_h(checkpoint=None): function build_sam_hq_vit_l (line 27) | def build_sam_hq_vit_l(checkpoint=None): function build_sam_hq_vit_b (line 37) | def build_sam_hq_vit_b(checkpoint=None): function _build_sam (line 55) | def _build_sam( FILE: data_generation/Grounded-Segment-Anything/segment_anything/segment_anything/modeling/common.py class MLPBlock (line 13) | class MLPBlock(nn.Module): method __init__ (line 14) | def __init__( method forward (line 25) | def forward(self, x: torch.Tensor) -> torch.Tensor: class LayerNorm2d (line 31) | class LayerNorm2d(nn.Module): method __init__ (line 32) | def __init__(self, num_channels: int, eps: float = 1e-6) -> None: method forward (line 38) | def forward(self, x: torch.Tensor) -> torch.Tensor: FILE: data_generation/Grounded-Segment-Anything/segment_anything/segment_anything/modeling/image_encoder.py class ImageEncoderViT (line 17) | class ImageEncoderViT(nn.Module): method __init__ (line 18) | def __init__( method forward (line 106) | def forward(self, x: torch.Tensor) -> torch.Tensor: class Block (line 122) | class Block(nn.Module): method __init__ (line 125) | def __init__( method forward (line 169) | def forward(self, x: torch.Tensor) -> torch.Tensor: class Attention (line 188) | class Attention(nn.Module): method __init__ (line 191) | def __init__( method forward (line 227) | def forward(self, x: torch.Tensor) -> torch.Tensor: function window_partition (line 246) | def window_partition(x: torch.Tensor, window_size: int) -> Tuple[torch.T... function window_unpartition (line 270) | def window_unpartition( function get_rel_pos (line 295) | def get_rel_pos(q_size: int, k_size: int, rel_pos: torch.Tensor) -> torc... function add_decomposed_rel_pos (line 328) | def add_decomposed_rel_pos( class PatchEmbed (line 367) | class PatchEmbed(nn.Module): method __init__ (line 372) | def __init__( method forward (line 394) | def forward(self, x: torch.Tensor) -> torch.Tensor: FILE: data_generation/Grounded-Segment-Anything/segment_anything/segment_anything/modeling/mask_decoder.py class MaskDecoder (line 16) | class MaskDecoder(nn.Module): method __init__ (line 17) | def __init__( method forward (line 71) | def forward( method predict_masks (line 114) | def predict_masks( class MLP (line 156) | class MLP(nn.Module): method __init__ (line 157) | def __init__( method forward (line 173) | def forward(self, x): FILE: data_generation/Grounded-Segment-Anything/segment_anything/segment_anything/modeling/mask_decoder_hq.py class MaskDecoderHQ (line 17) | class MaskDecoderHQ(nn.Module): method __init__ (line 18) | def __init__( method forward (line 99) | def forward( method predict_masks (line 158) | def predict_masks( class MLP (line 210) | class MLP(nn.Module): method __init__ (line 211) | def __init__( method forward (line 227) | def forward(self, x): FILE: data_generation/Grounded-Segment-Anything/segment_anything/segment_anything/modeling/prompt_encoder.py class PromptEncoder (line 16) | class PromptEncoder(nn.Module): method __init__ (line 17) | def __init__( method get_dense_pe (line 62) | def get_dense_pe(self) -> torch.Tensor: method _embed_points (line 73) | def _embed_points( method _embed_boxes (line 93) | def _embed_boxes(self, boxes: torch.Tensor) -> torch.Tensor: method _embed_masks (line 102) | def _embed_masks(self, masks: torch.Tensor) -> torch.Tensor: method _get_batch_size (line 107) | def _get_batch_size( method _get_device (line 125) | def _get_device(self) -> torch.device: method forward (line 128) | def forward( class PositionEmbeddingRandom (line 171) | class PositionEmbeddingRandom(nn.Module): method __init__ (line 176) | def __init__(self, num_pos_feats: int = 64, scale: Optional[float] = N... method _pe_encoding (line 185) | def _pe_encoding(self, coords: torch.Tensor) -> torch.Tensor: method forward (line 194) | def forward(self, size: Tuple[int, int]) -> torch.Tensor: method forward_with_coords (line 207) | def forward_with_coords( FILE: data_generation/Grounded-Segment-Anything/segment_anything/segment_anything/modeling/sam.py class Sam (line 18) | class Sam(nn.Module): method __init__ (line 22) | def __init__( method device (line 50) | def device(self) -> Any: method forward (line 54) | def forward( method postprocess_masks (line 133) | def postprocess_masks( method preprocess (line 164) | def preprocess(self, x: torch.Tensor) -> torch.Tensor: FILE: data_generation/Grounded-Segment-Anything/segment_anything/segment_anything/modeling/transformer.py class TwoWayTransformer (line 16) | class TwoWayTransformer(nn.Module): method __init__ (line 17) | def __init__( method forward (line 62) | def forward( class TwoWayAttentionBlock (line 109) | class TwoWayAttentionBlock(nn.Module): method __init__ (line 110) | def __init__( method forward (line 151) | def forward( class Attention (line 185) | class Attention(nn.Module): method __init__ (line 191) | def __init__( method _separate_heads (line 208) | def _separate_heads(self, x: Tensor, num_heads: int) -> Tensor: method _recombine_heads (line 213) | def _recombine_heads(self, x: Tensor) -> Tensor: method forward (line 218) | def forward(self, q: Tensor, k: Tensor, v: Tensor) -> Tensor: FILE: data_generation/Grounded-Segment-Anything/segment_anything/segment_anything/predictor.py class SamPredictor (line 17) | class SamPredictor: method __init__ (line 18) | def __init__( method set_image (line 34) | def set_image( method set_torch_image (line 65) | def set_torch_image( method predict (line 94) | def predict( method predict_torch (line 173) | def predict_torch( method get_image_embedding (line 252) | def get_image_embedding(self) -> torch.Tensor: method device (line 266) | def device(self) -> torch.device: method reset_image (line 269) | def reset_image(self) -> None: FILE: data_generation/Grounded-Segment-Anything/segment_anything/segment_anything/utils/amg.py class MaskData (line 16) | class MaskData: method __init__ (line 22) | def __init__(self, **kwargs) -> None: method __setitem__ (line 29) | def __setitem__(self, key: str, item: Any) -> None: method __delitem__ (line 35) | def __delitem__(self, key: str) -> None: method __getitem__ (line 38) | def __getitem__(self, key: str) -> Any: method items (line 41) | def items(self) -> ItemsView[str, Any]: method filter (line 44) | def filter(self, keep: torch.Tensor) -> None: method cat (line 59) | def cat(self, new_stats: "MaskData") -> None: method to_numpy (line 72) | def to_numpy(self) -> None: function is_box_near_crop_edge (line 78) | def is_box_near_crop_edge( function box_xyxy_to_xywh (line 91) | def box_xyxy_to_xywh(box_xyxy: torch.Tensor) -> torch.Tensor: function batch_iterator (line 98) | def batch_iterator(batch_size: int, *args) -> Generator[List[Any], None,... function mask_to_rle_pytorch (line 107) | def mask_to_rle_pytorch(tensor: torch.Tensor) -> List[Dict[str, Any]]: function rle_to_mask (line 138) | def rle_to_mask(rle: Dict[str, Any]) -> np.ndarray: function area_from_rle (line 152) | def area_from_rle(rle: Dict[str, Any]) -> int: function calculate_stability_score (line 156) | def calculate_stability_score( function build_point_grid (line 179) | def build_point_grid(n_per_side: int) -> np.ndarray: function build_all_layer_point_grids (line 189) | def build_all_layer_point_grids( function generate_crop_boxes (line 200) | def generate_crop_boxes( function uncrop_boxes_xyxy (line 237) | def uncrop_boxes_xyxy(boxes: torch.Tensor, crop_box: List[int]) -> torch... function uncrop_points (line 246) | def uncrop_points(points: torch.Tensor, crop_box: List[int]) -> torch.Te... function uncrop_masks (line 255) | def uncrop_masks( function remove_small_regions (line 267) | def remove_small_regions( function coco_encode_rle (line 294) | def coco_encode_rle(uncompressed_rle: Dict[str, Any]) -> Dict[str, Any]: function batched_mask_to_box (line 303) | def batched_mask_to_box(masks: torch.Tensor) -> torch.Tensor: FILE: data_generation/Grounded-Segment-Anything/segment_anything/segment_anything/utils/onnx.py class SamOnnxModel (line 17) | class SamOnnxModel(nn.Module): method __init__ (line 25) | def __init__( method resize_longest_image_size (line 42) | def resize_longest_image_size( method _embed_points (line 51) | def _embed_points(self, point_coords: torch.Tensor, point_labels: torc... method _embed_masks (line 69) | def _embed_masks(self, input_mask: torch.Tensor, has_mask_input: torch... method mask_postprocessing (line 76) | def mask_postprocessing(self, masks: torch.Tensor, orig_im_size: torch... method select_masks (line 92) | def select_masks( method forward (line 108) | def forward( FILE: data_generation/Grounded-Segment-Anything/segment_anything/segment_anything/utils/transforms.py class ResizeLongestSide (line 16) | class ResizeLongestSide: method __init__ (line 23) | def __init__(self, target_length: int) -> None: method apply_image (line 26) | def apply_image(self, image: np.ndarray) -> np.ndarray: method apply_coords (line 33) | def apply_coords(self, coords: np.ndarray, original_size: Tuple[int, .... method apply_boxes (line 47) | def apply_boxes(self, boxes: np.ndarray, original_size: Tuple[int, ...... method apply_image_torch (line 55) | def apply_image_torch(self, image: torch.Tensor) -> torch.Tensor: method apply_coords_torch (line 67) | def apply_coords_torch( method apply_boxes_torch (line 83) | def apply_boxes_torch( method get_preprocess_shape (line 94) | def get_preprocess_shape(oldh: int, oldw: int, long_side_length: int) ... FILE: data_generation/Grounded-Segment-Anything/voxelnext_3d_box/model.py class VoxelNeXt (line 12) | class VoxelNeXt(nn.Module): method __init__ (line 13) | def __init__(self, model_cfg): class Model (line 39) | class Model(nn.Module): method __init__ (line 40) | def __init__(self, model_cfg, device="cuda"): method image_embedding (line 56) | def image_embedding(self, image): method point_embedding (line 59) | def point_embedding(self, data_dict, image_id): method generate_3D_box (line 85) | def generate_3D_box(self, lidar2img_rt, mask, voxel_coords, pred_dicts... method forward (line 118) | def forward(self, image, point_dict, prompt_point, lidar2img_rt, image... FILE: data_generation/Grounded-Segment-Anything/voxelnext_3d_box/models/data_processor.py function mask_points_by_range (line 12) | def mask_points_by_range(points, limit_range): class VoxelGeneratorWrapper (line 18) | class VoxelGeneratorWrapper(): method __init__ (line 19) | def __init__(self, vsize_xyz, coors_range_xyz, num_point_features, max... method generate (line 47) | def generate(self, points): class DataProcessor (line 66) | class DataProcessor(object): method __init__ (line 67) | def __init__(self, processor_configs, point_cloud_range, training, num... method mask_points_and_boxes_outside_range (line 81) | def mask_points_and_boxes_outside_range(self, data_dict=None, config=N... method shuffle_points (line 91) | def shuffle_points(self, data_dict=None, config=None): method transform_points_to_voxels_placeholder (line 103) | def transform_points_to_voxels_placeholder(self, data_dict=None, confi... method double_flip (line 113) | def double_flip(self, points): method transform_points_to_voxels (line 129) | def transform_points_to_voxels(self, data_dict=None, config=None): method sample_points (line 156) | def sample_points(self, data_dict=None, config=None): method calculate_grid_size (line 188) | def calculate_grid_size(self, data_dict=None, config=None): method forward (line 196) | def forward(self, data_dict): FILE: data_generation/Grounded-Segment-Anything/voxelnext_3d_box/models/mean_vfe.py class MeanVFE (line 4) | class MeanVFE(nn.Module): method __init__ (line 5) | def __init__(self): method forward (line 8) | def forward(self, batch_dict, **kwargs): FILE: data_generation/Grounded-Segment-Anything/voxelnext_3d_box/models/spconv_backbone_voxelnext.py function replace_feature (line 9) | def replace_feature(out, new_features): function post_act_block (line 13) | def post_act_block(in_channels, out_channels, kernel_size, indice_key=No... class SparseBasicBlock (line 35) | class SparseBasicBlock(spconv.SparseModule): method __init__ (line 38) | def __init__(self, inplanes, planes, stride=1, norm_fn=None, downsampl... method forward (line 55) | def forward(self, x): class VoxelResBackBone8xVoxelNeXt (line 74) | class VoxelResBackBone8xVoxelNeXt(nn.Module): method __init__ (line 75) | def __init__(self, input_channels, grid_size, **kwargs): method bev_out (line 153) | def bev_out(self, x_conv, index): method track_voxels_2d (line 174) | def track_voxels_2d(self, x, x_downsample, index, kernel_size=3): method index_from_sparse (line 214) | def index_from_sparse(self, feature, indices, x_target, _2d=False): method forward (line 229) | def forward(self, batch_dict): FILE: data_generation/Grounded-Segment-Anything/voxelnext_3d_box/models/voxelnext_head.py class SeparateHead (line 10) | class SeparateHead(nn.Module): method __init__ (line 11) | def __init__(self, input_channels, sep_head_dict, kernel_size, use_bia... method forward (line 30) | def forward(self, x): class VoxelNeXtHead (line 38) | class VoxelNeXtHead(nn.Module): method __init__ (line 39) | def __init__(self, class_names, point_cloud_range, voxel_size, kernel_... method generate_predicted_boxes (line 76) | def generate_predicted_boxes(self, batch_size, pred_dicts, voxel_indic... method _get_voxel_infos (line 137) | def _get_voxel_infos(self, x): method forward (line 152) | def forward(self, data_dict): FILE: data_generation/Grounded-Segment-Anything/voxelnext_3d_box/utils/centernet_utils.py function _topk_1d (line 6) | def _topk_1d(scores, batch_size, batch_idx, obj, K=40, nuscenes=False): function gather_feat_idx (line 40) | def gather_feat_idx(feats, inds, batch_size, batch_idx): function decode_bbox_from_voxels_nuscenes (line 53) | def decode_bbox_from_voxels_nuscenes(batch_size, indices, obj, rot_cos, ... FILE: data_generation/Grounded-Segment-Anything/voxelnext_3d_box/utils/config.py function merge_new_config (line 4) | def merge_new_config(config, new_config): function cfg_from_yaml_file (line 23) | def cfg_from_yaml_file(cfg_file, config): FILE: data_generation/Grounded-Segment-Anything/voxelnext_3d_box/utils/image_projection.py function get_data_info (line 5) | def get_data_info(info, cam_type): function _proj_voxel_image (line 23) | def _proj_voxel_image(voxel_coords, lidar2img_rt, voxel_size, point_clou... function _draw_image (line 41) | def _draw_image(points_image, image_path, depth): function _draw_mask (line 51) | def _draw_mask(image_path, mask, color=None): function _draw_3dbox (line 62) | def _draw_3dbox(box, lidar2img_rt, image, mask=None, color=None, output_... FILE: data_generation/data_generation.py function generate_Contours_mask (line 59) | def generate_Contours_mask(pil_img): function find_contours_number (line 92) | def find_contours_number(mask_image: np.array): function check_mask_size (line 96) | def check_mask_size(mask, min_size=0.01, max_size=0.9): function create_mask_image (line 105) | def create_mask_image(image_source, detected_boxes): function to_pil (line 133) | def to_pil(image: torch.Tensor) -> Image.Image: function gd_load_image (line 138) | def gd_load_image(image_path) -> Tuple[np.array, torch.Tensor]: function load_image (line 157) | def load_image(image_path, target_size=512): function compare_prompts (line 188) | def compare_prompts(prompt1, prompt2): function generate_images (line 194) | def generate_images(args,pipe, prompts, init_image, cross_attention_kwar... function generate_contour_mask (line 260) | def generate_contour_mask(mask): function detect (line 277) | def detect(image, text_prompt, model, box_threshold = 0.3, text_threshol... function segment (line 287) | def segment(image, sam_model, boxes,device): function save_tsv (line 302) | def save_tsv(args, shard_id, shard, device, global_data): class NoneImageDataset (line 789) | class NoneImageDataset(Dataset): method __init__ (line 790) | def __init__(self, url_data): method __len__ (line 793) | def __len__(self): method __getitem__ (line 796) | def __getitem__(self, idx): class OpenImageDataset (line 803) | class OpenImageDataset(Dataset): method __init__ (line 804) | def __init__(self, url_data): method __len__ (line 807) | def __len__(self): method __getitem__ (line 810) | def __getitem__(self, idx): class RegionDataset (line 834) | class RegionDataset(Dataset): method __init__ (line 835) | def __init__(self, url_data): method __len__ (line 838) | def __len__(self): method __getitem__ (line 841) | def __getitem__(self, idx): function collate_fn (line 885) | def collate_fn(batch): function load_jsonl (line 930) | def load_jsonl(file_path): function load_json (line 936) | def load_json(file_path): function load_groundingdino_model (line 939) | def load_groundingdino_model(model_config_path, model_checkpoint_path, d... function main (line 949) | def main(): FILE: data_generation/ldm/data/base.py class Txt2ImgIterableBaseDataset (line 5) | class Txt2ImgIterableBaseDataset(IterableDataset): method __init__ (line 9) | def __init__(self, num_records=0, valid_ids=None, size=256): method __len__ (line 18) | def __len__(self): method __iter__ (line 22) | def __iter__(self): FILE: data_generation/ldm/data/imagenet.py function synset2idx (line 20) | def synset2idx(path_to_yaml="data/index_synset.yaml"): class ImageNetBase (line 26) | class ImageNetBase(Dataset): method __init__ (line 27) | def __init__(self, config=None): method __len__ (line 39) | def __len__(self): method __getitem__ (line 42) | def __getitem__(self, i): method _prepare (line 45) | def _prepare(self): method _filter_relpaths (line 48) | def _filter_relpaths(self, relpaths): method _prepare_synset_to_human (line 66) | def _prepare_synset_to_human(self): method _prepare_idx_to_synset (line 74) | def _prepare_idx_to_synset(self): method _prepare_human_to_integer_label (line 80) | def _prepare_human_to_integer_label(self): method _load (line 93) | def _load(self): class ImageNetTrain (line 134) | class ImageNetTrain(ImageNetBase): method __init__ (line 145) | def __init__(self, process_images=True, data_root=None, **kwargs): method _prepare (line 150) | def _prepare(self): class ImageNetValidation (line 197) | class ImageNetValidation(ImageNetBase): method __init__ (line 211) | def __init__(self, process_images=True, data_root=None, **kwargs): method _prepare (line 216) | def _prepare(self): class ImageNetSR (line 272) | class ImageNetSR(Dataset): method __init__ (line 273) | def __init__(self, size=None, method __len__ (line 336) | def __len__(self): method __getitem__ (line 339) | def __getitem__(self, i): class ImageNetSRTrain (line 375) | class ImageNetSRTrain(ImageNetSR): method __init__ (line 376) | def __init__(self, **kwargs): method get_base (line 379) | def get_base(self): class ImageNetSRValidation (line 386) | class ImageNetSRValidation(ImageNetSR): method __init__ (line 387) | def __init__(self, **kwargs): method get_base (line 390) | def get_base(self): FILE: data_generation/ldm/data/lsun.py class LSUNBase (line 9) | class LSUNBase(Dataset): method __init__ (line 10) | def __init__(self, method __len__ (line 36) | def __len__(self): method __getitem__ (line 39) | def __getitem__(self, i): class LSUNChurchesTrain (line 62) | class LSUNChurchesTrain(LSUNBase): method __init__ (line 63) | def __init__(self, **kwargs): class LSUNChurchesValidation (line 67) | class LSUNChurchesValidation(LSUNBase): method __init__ (line 68) | def __init__(self, flip_p=0., **kwargs): class LSUNBedroomsTrain (line 73) | class LSUNBedroomsTrain(LSUNBase): method __init__ (line 74) | def __init__(self, **kwargs): class LSUNBedroomsValidation (line 78) | class LSUNBedroomsValidation(LSUNBase): method __init__ (line 79) | def __init__(self, flip_p=0.0, **kwargs): class LSUNCatsTrain (line 84) | class LSUNCatsTrain(LSUNBase): method __init__ (line 85) | def __init__(self, **kwargs): class LSUNCatsValidation (line 89) | class LSUNCatsValidation(LSUNBase): method __init__ (line 90) | def __init__(self, flip_p=0., **kwargs): FILE: data_generation/ldm/lr_scheduler.py class LambdaWarmUpCosineScheduler (line 4) | class LambdaWarmUpCosineScheduler: method __init__ (line 8) | def __init__(self, warm_up_steps, lr_min, lr_max, lr_start, max_decay_... method schedule (line 17) | def schedule(self, n, **kwargs): method __call__ (line 32) | def __call__(self, n, **kwargs): class LambdaWarmUpCosineScheduler2 (line 36) | class LambdaWarmUpCosineScheduler2: method __init__ (line 41) | def __init__(self, warm_up_steps, f_min, f_max, f_start, cycle_lengths... method find_in_interval (line 52) | def find_in_interval(self, n): method schedule (line 59) | def schedule(self, n, **kwargs): method __call__ (line 77) | def __call__(self, n, **kwargs): class LambdaLinearScheduler (line 81) | class LambdaLinearScheduler(LambdaWarmUpCosineScheduler2): method schedule (line 83) | def schedule(self, n, **kwargs): FILE: data_generation/ldm/models/autoencoder.py class VQModel (line 14) | class VQModel(pl.LightningModule): method __init__ (line 15) | def __init__(self, method ema_scope (line 64) | def ema_scope(self, context=None): method init_from_ckpt (line 78) | def init_from_ckpt(self, path, ignore_keys=list()): method on_train_batch_end (line 92) | def on_train_batch_end(self, *args, **kwargs): method encode (line 96) | def encode(self, x): method encode_to_prequant (line 102) | def encode_to_prequant(self, x): method decode (line 107) | def decode(self, quant): method decode_code (line 112) | def decode_code(self, code_b): method forward (line 117) | def forward(self, input, return_pred_indices=False): method get_input (line 124) | def get_input(self, batch, k): method training_step (line 142) | def training_step(self, batch, batch_idx, optimizer_idx): method validation_step (line 164) | def validation_step(self, batch, batch_idx): method _validation_step (line 170) | def _validation_step(self, batch, batch_idx, suffix=""): method configure_optimizers (line 197) | def configure_optimizers(self): method get_last_layer (line 230) | def get_last_layer(self): method log_images (line 233) | def log_images(self, batch, only_inputs=False, plot_ema=False, **kwargs): method to_rgb (line 255) | def to_rgb(self, x): class VQModelInterface (line 264) | class VQModelInterface(VQModel): method __init__ (line 265) | def __init__(self, embed_dim, *args, **kwargs): method encode (line 269) | def encode(self, x): method decode (line 274) | def decode(self, h, force_not_quantize=False): class AutoencoderKL (line 285) | class AutoencoderKL(pl.LightningModule): method __init__ (line 286) | def __init__(self, method init_from_ckpt (line 313) | def init_from_ckpt(self, path, ignore_keys=list()): method encode (line 324) | def encode(self, x): method decode (line 330) | def decode(self, z): method forward (line 335) | def forward(self, input, sample_posterior=True): method get_input (line 344) | def get_input(self, batch, k): method training_step (line 351) | def training_step(self, batch, batch_idx, optimizer_idx): method validation_step (line 372) | def validation_step(self, batch, batch_idx): method configure_optimizers (line 386) | def configure_optimizers(self): method get_last_layer (line 397) | def get_last_layer(self): method log_images (line 401) | def log_images(self, batch, only_inputs=False, **kwargs): method to_rgb (line 417) | def to_rgb(self, x): class IdentityFirstStage (line 426) | class IdentityFirstStage(torch.nn.Module): method __init__ (line 427) | def __init__(self, *args, vq_interface=False, **kwargs): method encode (line 431) | def encode(self, x, *args, **kwargs): method decode (line 434) | def decode(self, x, *args, **kwargs): method quantize (line 437) | def quantize(self, x, *args, **kwargs): method forward (line 442) | def forward(self, x, *args, **kwargs): FILE: data_generation/ldm/models/diffusion/classifier.py function disabled_train (line 22) | def disabled_train(self, mode=True): class NoisyLatentImageClassifier (line 28) | class NoisyLatentImageClassifier(pl.LightningModule): method __init__ (line 30) | def __init__(self, method init_from_ckpt (line 70) | def init_from_ckpt(self, path, ignore_keys=list(), only_model=False): method load_diffusion (line 88) | def load_diffusion(self): method load_classifier (line 95) | def load_classifier(self, ckpt_path, pool): method get_x_noisy (line 110) | def get_x_noisy(self, x, t, noise=None): method forward (line 120) | def forward(self, x_noisy, t, *args, **kwargs): method get_input (line 124) | def get_input(self, batch, k): method get_conditioning (line 133) | def get_conditioning(self, batch, k=None): method compute_top_k (line 150) | def compute_top_k(self, logits, labels, k, reduction="mean"): method on_train_epoch_start (line 157) | def on_train_epoch_start(self): method write_logs (line 162) | def write_logs(self, loss, logits, targets): method shared_step (line 179) | def shared_step(self, batch, t=None): method training_step (line 198) | def training_step(self, batch, batch_idx): method reset_noise_accs (line 202) | def reset_noise_accs(self): method on_validation_start (line 206) | def on_validation_start(self): method validation_step (line 210) | def validation_step(self, batch, batch_idx): method configure_optimizers (line 220) | def configure_optimizers(self): method log_images (line 238) | def log_images(self, batch, N=8, *args, **kwargs): FILE: data_generation/ldm/models/diffusion/ddim.py class DDIMSampler (line 12) | class DDIMSampler(object): method __init__ (line 13) | def __init__(self, model, schedule="linear", **kwargs): method register_buffer (line 19) | def register_buffer(self, name, attr): method make_schedule (line 25) | def make_schedule(self, ddim_num_steps, ddim_discretize="uniform", ddi... method sample (line 57) | def sample(self, method ddim_sampling (line 114) | def ddim_sampling(self, cond, shape, method p_sample_ddim (line 166) | def p_sample_ddim(self, x, c, t, index, repeat_noise=False, use_origin... method stochastic_encode (line 207) | def stochastic_encode(self, x0, t, use_original_steps=False, noise=None): method decode (line 223) | def decode(self, x_latent, cond, t_start, unconditional_guidance_scale... FILE: data_generation/ldm/models/diffusion/ddpm.py function disabled_train (line 34) | def disabled_train(self, mode=True): function uniform_on_device (line 40) | def uniform_on_device(r1, r2, shape, device): class DDPM (line 44) | class DDPM(pl.LightningModule): method __init__ (line 46) | def __init__(self, method register_schedule (line 117) | def register_schedule(self, given_betas=None, beta_schedule="linear", ... method ema_scope (line 172) | def ema_scope(self, context=None): method init_from_ckpt (line 186) | def init_from_ckpt(self, path, ignore_keys=list(), only_model=False): method q_mean_variance (line 204) | def q_mean_variance(self, x_start, t): method predict_start_from_noise (line 216) | def predict_start_from_noise(self, x_t, t, noise): method q_posterior (line 222) | def q_posterior(self, x_start, x_t, t): method p_mean_variance (line 231) | def p_mean_variance(self, x, t, clip_denoised: bool): method p_sample (line 244) | def p_sample(self, x, t, clip_denoised=True, repeat_noise=False): method p_sample_loop (line 253) | def p_sample_loop(self, shape, return_intermediates=False): method sample (line 268) | def sample(self, batch_size=16, return_intermediates=False): method q_sample (line 274) | def q_sample(self, x_start, t, noise=None): method get_loss (line 279) | def get_loss(self, pred, target, mean=True): method p_losses (line 294) | def p_losses(self, x_start, t, noise=None): method forward (line 323) | def forward(self, x, *args, **kwargs): method get_input (line 329) | def get_input(self, batch, k): method shared_step (line 337) | def shared_step(self, batch): method training_step (line 342) | def training_step(self, batch, batch_idx): method validation_step (line 358) | def validation_step(self, batch, batch_idx): method on_train_batch_end (line 366) | def on_train_batch_end(self, *args, **kwargs): method _get_rows_from_list (line 370) | def _get_rows_from_list(self, samples): method log_images (line 378) | def log_images(self, batch, N=8, n_row=2, sample=True, return_keys=Non... method configure_optimizers (line 415) | def configure_optimizers(self): class LatentDiffusion (line 424) | class LatentDiffusion(DDPM): method __init__ (line 426) | def __init__(self, method make_cond_schedule (line 471) | def make_cond_schedule(self, ): method on_train_batch_start (line 478) | def on_train_batch_start(self, batch, batch_idx, dataloader_idx): method register_schedule (line 493) | def register_schedule(self, method instantiate_first_stage (line 502) | def instantiate_first_stage(self, config): method instantiate_cond_stage (line 509) | def instantiate_cond_stage(self, config): method _get_denoise_row_from_list (line 530) | def _get_denoise_row_from_list(self, samples, desc='', force_no_decode... method get_first_stage_encoding (line 542) | def get_first_stage_encoding(self, encoder_posterior): method get_learned_conditioning (line 551) | def get_learned_conditioning(self, c): method meshgrid (line 564) | def meshgrid(self, h, w): method delta_border (line 571) | def delta_border(self, h, w): method get_weighting (line 585) | def get_weighting(self, h, w, Ly, Lx, device): method get_fold_unfold (line 601) | def get_fold_unfold(self, x, kernel_size, stride, uf=1, df=1): # todo... method get_input (line 654) | def get_input(self, batch, k, return_first_stage_outputs=False, force_... method decode_first_stage (line 706) | def decode_first_stage(self, z, predict_cids=False, force_not_quantize... method differentiable_decode_first_stage (line 766) | def differentiable_decode_first_stage(self, z, predict_cids=False, for... method encode_first_stage (line 826) | def encode_first_stage(self, x): method shared_step (line 865) | def shared_step(self, batch, **kwargs): method forward (line 870) | def forward(self, x, c, *args, **kwargs): method _rescale_annotations (line 881) | def _rescale_annotations(self, bboxes, crop_coordinates): # TODO: mov... method apply_model (line 891) | def apply_model(self, x_noisy, t, cond, return_ids=False): method _predict_eps_from_xstart (line 994) | def _predict_eps_from_xstart(self, x_t, t, pred_xstart): method _prior_bpd (line 998) | def _prior_bpd(self, x_start): method p_losses (line 1012) | def p_losses(self, x_start, cond, t, noise=None): method p_mean_variance (line 1047) | def p_mean_variance(self, x, c, t, clip_denoised: bool, return_codeboo... method p_sample (line 1079) | def p_sample(self, x, c, t, clip_denoised=False, repeat_noise=False, method progressive_denoising (line 1110) | def progressive_denoising(self, cond, shape, verbose=True, callback=No... method p_sample_loop (line 1166) | def p_sample_loop(self, cond, shape, return_intermediates=False, method sample (line 1217) | def sample(self, cond, batch_size=16, return_intermediates=False, x_T=... method sample_log (line 1235) | def sample_log(self,cond,batch_size,ddim, ddim_steps,**kwargs): method log_images (line 1251) | def log_images(self, batch, N=8, n_row=4, sample=True, ddim_steps=200,... method configure_optimizers (line 1361) | def configure_optimizers(self): method to_rgb (line 1386) | def to_rgb(self, x): class DiffusionWrapper (line 1395) | class DiffusionWrapper(pl.LightningModule): method __init__ (line 1396) | def __init__(self, diff_model_config, conditioning_key): method forward (line 1402) | def forward(self, x, t, c_concat: list = None, c_crossattn: list = None): class Layout2ImgDiffusion (line 1424) | class Layout2ImgDiffusion(LatentDiffusion): method __init__ (line 1426) | def __init__(self, cond_stage_key, *args, **kwargs): method log_images (line 1430) | def log_images(self, batch, N=8, *args, **kwargs): FILE: data_generation/ldm/models/diffusion/dpm_solver/dpm_solver.py class NoiseScheduleVP (line 6) | class NoiseScheduleVP: method __init__ (line 7) | def __init__( method marginal_log_mean_coeff (line 125) | def marginal_log_mean_coeff(self, t): method marginal_alpha (line 138) | def marginal_alpha(self, t): method marginal_std (line 144) | def marginal_std(self, t): method marginal_lambda (line 150) | def marginal_lambda(self, t): method inverse_lambda (line 158) | def inverse_lambda(self, lamb): function model_wrapper (line 177) | def model_wrapper( class DPM_Solver (line 351) | class DPM_Solver: method __init__ (line 352) | def __init__(self, model_fn, noise_schedule, predict_x0=False, thresho... method noise_prediction_fn (line 380) | def noise_prediction_fn(self, x, t): method data_prediction_fn (line 386) | def data_prediction_fn(self, x, t): method model_fn (line 401) | def model_fn(self, x, t): method get_time_steps (line 410) | def get_time_steps(self, skip_type, t_T, t_0, N, device): method get_orders_and_timesteps_for_singlestep_solver (line 439) | def get_orders_and_timesteps_for_singlestep_solver(self, steps, order,... method denoise_to_zero_fn (line 498) | def denoise_to_zero_fn(self, x, s): method dpm_solver_first_update (line 504) | def dpm_solver_first_update(self, x, s, t, model_s=None, return_interm... method singlestep_dpm_solver_second_update (line 551) | def singlestep_dpm_solver_second_update(self, x, s, t, r1=0.5, model_s... method singlestep_dpm_solver_third_update (line 633) | def singlestep_dpm_solver_third_update(self, x, s, t, r1=1./3., r2=2./... method multistep_dpm_solver_second_update (line 755) | def multistep_dpm_solver_second_update(self, x, model_prev_list, t_pre... method multistep_dpm_solver_third_update (line 812) | def multistep_dpm_solver_third_update(self, x, model_prev_list, t_prev... method singlestep_dpm_solver_update (line 859) | def singlestep_dpm_solver_update(self, x, s, t, order, return_intermed... method multistep_dpm_solver_update (line 885) | def multistep_dpm_solver_update(self, x, model_prev_list, t_prev_list,... method dpm_solver_adaptive (line 909) | def dpm_solver_adaptive(self, x, order, t_T, t_0, h_init=0.05, atol=0.... method sample (line 965) | def sample(self, x, steps=20, t_start=None, t_end=None, order=3, skip_... function interpolate_fn (line 1132) | def interpolate_fn(x, xp, yp): function expand_dims (line 1174) | def expand_dims(v, dims): FILE: data_generation/ldm/models/diffusion/dpm_solver/sampler.py class DPMSolverSampler (line 8) | class DPMSolverSampler(object): method __init__ (line 9) | def __init__(self, model, **kwargs): method register_buffer (line 15) | def register_buffer(self, name, attr): method sample (line 22) | def sample(self, FILE: data_generation/ldm/models/diffusion/plms.py class PLMSSampler (line 11) | class PLMSSampler(object): method __init__ (line 12) | def __init__(self, model, schedule="linear", **kwargs): method register_buffer (line 18) | def register_buffer(self, name, attr): method make_schedule (line 24) | def make_schedule(self, ddim_num_steps, ddim_discretize="uniform", ddi... method sample (line 58) | def sample(self, method plms_sampling (line 115) | def plms_sampling(self, cond, shape, method p_sample_plms (line 173) | def p_sample_plms(self, x, c, t, index, repeat_noise=False, use_origin... FILE: data_generation/ldm/modules/attention.py function exists (line 11) | def exists(val): function uniq (line 15) | def uniq(arr): function default (line 19) | def default(val, d): function max_neg_value (line 25) | def max_neg_value(t): function init_ (line 29) | def init_(tensor): class GEGLU (line 37) | class GEGLU(nn.Module): method __init__ (line 38) | def __init__(self, dim_in, dim_out): method forward (line 42) | def forward(self, x): class FeedForward (line 47) | class FeedForward(nn.Module): method __init__ (line 48) | def __init__(self, dim, dim_out=None, mult=4, glu=False, dropout=0.): method forward (line 63) | def forward(self, x): function zero_module (line 67) | def zero_module(module): function Normalize (line 76) | def Normalize(in_channels): class LinearAttention (line 80) | class LinearAttention(nn.Module): method __init__ (line 81) | def __init__(self, dim, heads=4, dim_head=32): method forward (line 88) | def forward(self, x): class SpatialSelfAttention (line 99) | class SpatialSelfAttention(nn.Module): method __init__ (line 100) | def __init__(self, in_channels): method forward (line 126) | def forward(self, x): class CrossAttention (line 152) | class CrossAttention(nn.Module): method __init__ (line 153) | def __init__(self, query_dim, context_dim=None, heads=8, dim_head=64, ... method forward (line 170) | def forward(self, x, context=None, mask=None): class BasicTransformerBlock (line 196) | class BasicTransformerBlock(nn.Module): method __init__ (line 197) | def __init__(self, dim, n_heads, d_head, dropout=0., context_dim=None,... method forward (line 208) | def forward(self, x, context=None): method _forward (line 211) | def _forward(self, x, context=None): class SpatialTransformer (line 218) | class SpatialTransformer(nn.Module): method __init__ (line 226) | def __init__(self, in_channels, n_heads, d_head, method forward (line 250) | def forward(self, x, context=None): FILE: data_generation/ldm/modules/diffusionmodules/model.py function get_timestep_embedding (line 12) | def get_timestep_embedding(timesteps, embedding_dim): function nonlinearity (line 33) | def nonlinearity(x): function Normalize (line 38) | def Normalize(in_channels, num_groups=32): class Upsample (line 42) | class Upsample(nn.Module): method __init__ (line 43) | def __init__(self, in_channels, with_conv): method forward (line 53) | def forward(self, x): class Downsample (line 60) | class Downsample(nn.Module): method __init__ (line 61) | def __init__(self, in_channels, with_conv): method forward (line 72) | def forward(self, x): class ResnetBlock (line 82) | class ResnetBlock(nn.Module): method __init__ (line 83) | def __init__(self, *, in_channels, out_channels=None, conv_shortcut=Fa... method forward (line 121) | def forward(self, x, temb): class LinAttnBlock (line 144) | class LinAttnBlock(LinearAttention): method __init__ (line 146) | def __init__(self, in_channels): class AttnBlock (line 150) | class AttnBlock(nn.Module): method __init__ (line 151) | def __init__(self, in_channels): method forward (line 178) | def forward(self, x): function make_attn (line 205) | def make_attn(in_channels, attn_type="vanilla"): class Model (line 216) | class Model(nn.Module): method __init__ (line 217) | def __init__(self, *, ch, out_ch, ch_mult=(1,2,4,8), num_res_blocks, method forward (line 316) | def forward(self, x, t=None, context=None): method get_last_layer (line 364) | def get_last_layer(self): class Encoder (line 368) | class Encoder(nn.Module): method __init__ (line 369) | def __init__(self, *, ch, out_ch, ch_mult=(1,2,4,8), num_res_blocks, method forward (line 434) | def forward(self, x): class Decoder (line 462) | class Decoder(nn.Module): method __init__ (line 463) | def __init__(self, *, ch, out_ch, ch_mult=(1,2,4,8), num_res_blocks, method forward (line 535) | def forward(self, z): class SimpleDecoder (line 571) | class SimpleDecoder(nn.Module): method __init__ (line 572) | def __init__(self, in_channels, out_channels, *args, **kwargs): method forward (line 594) | def forward(self, x): class UpsampleDecoder (line 607) | class UpsampleDecoder(nn.Module): method __init__ (line 608) | def __init__(self, in_channels, out_channels, ch, num_res_blocks, reso... method forward (line 641) | def forward(self, x): class LatentRescaler (line 655) | class LatentRescaler(nn.Module): method __init__ (line 656) | def __init__(self, factor, in_channels, mid_channels, out_channels, de... method forward (line 680) | def forward(self, x): class MergedRescaleEncoder (line 692) | class MergedRescaleEncoder(nn.Module): method __init__ (line 693) | def __init__(self, in_channels, ch, resolution, out_ch, num_res_blocks, method forward (line 705) | def forward(self, x): class MergedRescaleDecoder (line 711) | class MergedRescaleDecoder(nn.Module): method __init__ (line 712) | def __init__(self, z_channels, out_ch, resolution, num_res_blocks, att... method forward (line 722) | def forward(self, x): class Upsampler (line 728) | class Upsampler(nn.Module): method __init__ (line 729) | def __init__(self, in_size, out_size, in_channels, out_channels, ch_mu... method forward (line 741) | def forward(self, x): class Resize (line 747) | class Resize(nn.Module): method __init__ (line 748) | def __init__(self, in_channels=None, learned=False, mode="bilinear"): method forward (line 763) | def forward(self, x, scale_factor=1.0): class FirstStagePostProcessor (line 770) | class FirstStagePostProcessor(nn.Module): method __init__ (line 772) | def __init__(self, ch_mult:list, in_channels, method instantiate_pretrained (line 807) | def instantiate_pretrained(self, config): method encode_with_pretrained (line 816) | def encode_with_pretrained(self,x): method forward (line 822) | def forward(self,x): FILE: data_generation/ldm/modules/diffusionmodules/openaimodel.py function convert_module_to_f16 (line 24) | def convert_module_to_f16(x): function convert_module_to_f32 (line 27) | def convert_module_to_f32(x): class AttentionPool2d (line 32) | class AttentionPool2d(nn.Module): method __init__ (line 37) | def __init__( method forward (line 51) | def forward(self, x): class TimestepBlock (line 62) | class TimestepBlock(nn.Module): method forward (line 68) | def forward(self, x, emb): class TimestepEmbedSequential (line 74) | class TimestepEmbedSequential(nn.Sequential, TimestepBlock): method forward (line 80) | def forward(self, x, emb, context=None): class Upsample (line 91) | class Upsample(nn.Module): method __init__ (line 100) | def __init__(self, channels, use_conv, dims=2, out_channels=None, padd... method forward (line 109) | def forward(self, x): class TransposedUpsample (line 121) | class TransposedUpsample(nn.Module): method __init__ (line 123) | def __init__(self, channels, out_channels=None, ks=5): method forward (line 130) | def forward(self,x): class Downsample (line 134) | class Downsample(nn.Module): method __init__ (line 143) | def __init__(self, channels, use_conv, dims=2, out_channels=None,paddi... method forward (line 158) | def forward(self, x): class ResBlock (line 163) | class ResBlock(TimestepBlock): method __init__ (line 179) | def __init__( method forward (line 243) | def forward(self, x, emb): method _forward (line 255) | def _forward(self, x, emb): class AttentionBlock (line 278) | class AttentionBlock(nn.Module): method __init__ (line 285) | def __init__( method forward (line 314) | def forward(self, x): method _forward (line 318) | def _forward(self, x): function count_flops_attn (line 327) | def count_flops_attn(model, _x, y): class QKVAttentionLegacy (line 347) | class QKVAttentionLegacy(nn.Module): method __init__ (line 352) | def __init__(self, n_heads): method forward (line 356) | def forward(self, qkv): method count_flops (line 375) | def count_flops(model, _x, y): class QKVAttention (line 379) | class QKVAttention(nn.Module): method __init__ (line 384) | def __init__(self, n_heads): method forward (line 388) | def forward(self, qkv): method count_flops (line 409) | def count_flops(model, _x, y): class UNetModel (line 413) | class UNetModel(nn.Module): method __init__ (line 443) | def __init__( method convert_to_fp16 (line 694) | def convert_to_fp16(self): method convert_to_fp32 (line 702) | def convert_to_fp32(self): method forward (line 710) | def forward(self, x, timesteps=None, context=None, y=None,**kwargs): class EncoderUNetModel (line 745) | class EncoderUNetModel(nn.Module): method __init__ (line 751) | def __init__( method convert_to_fp16 (line 924) | def convert_to_fp16(self): method convert_to_fp32 (line 931) | def convert_to_fp32(self): method forward (line 938) | def forward(self, x, timesteps): FILE: data_generation/ldm/modules/diffusionmodules/util.py function make_beta_schedule (line 21) | def make_beta_schedule(schedule, n_timestep, linear_start=1e-4, linear_e... function make_ddim_timesteps (line 46) | def make_ddim_timesteps(ddim_discr_method, num_ddim_timesteps, num_ddpm_... function make_ddim_sampling_parameters (line 63) | def make_ddim_sampling_parameters(alphacums, ddim_timesteps, eta, verbos... function betas_for_alpha_bar (line 77) | def betas_for_alpha_bar(num_diffusion_timesteps, alpha_bar, max_beta=0.9... function extract_into_tensor (line 96) | def extract_into_tensor(a, t, x_shape): function checkpoint (line 102) | def checkpoint(func, inputs, params, flag): class CheckpointFunction (line 119) | class CheckpointFunction(torch.autograd.Function): method forward (line 121) | def forward(ctx, run_function, length, *args): method backward (line 131) | def backward(ctx, *output_grads): function timestep_embedding (line 151) | def timestep_embedding(timesteps, dim, max_period=10000, repeat_only=Fal... function zero_module (line 174) | def zero_module(module): function scale_module (line 183) | def scale_module(module, scale): function mean_flat (line 192) | def mean_flat(tensor): function normalization (line 199) | def normalization(channels): class SiLU (line 209) | class SiLU(nn.Module): method forward (line 210) | def forward(self, x): class GroupNorm32 (line 214) | class GroupNorm32(nn.GroupNorm): method forward (line 215) | def forward(self, x): function conv_nd (line 218) | def conv_nd(dims, *args, **kwargs): function linear (line 231) | def linear(*args, **kwargs): function avg_pool_nd (line 238) | def avg_pool_nd(dims, *args, **kwargs): class HybridConditioner (line 251) | class HybridConditioner(nn.Module): method __init__ (line 253) | def __init__(self, c_concat_config, c_crossattn_config): method forward (line 258) | def forward(self, c_concat, c_crossattn): function noise_like (line 264) | def noise_like(shape, device, repeat=False): FILE: data_generation/ldm/modules/distributions/distributions.py class AbstractDistribution (line 5) | class AbstractDistribution: method sample (line 6) | def sample(self): method mode (line 9) | def mode(self): class DiracDistribution (line 13) | class DiracDistribution(AbstractDistribution): method __init__ (line 14) | def __init__(self, value): method sample (line 17) | def sample(self): method mode (line 20) | def mode(self): class DiagonalGaussianDistribution (line 24) | class DiagonalGaussianDistribution(object): method __init__ (line 25) | def __init__(self, parameters, deterministic=False): method sample (line 35) | def sample(self): method kl (line 39) | def kl(self, other=None): method nll (line 53) | def nll(self, sample, dims=[1,2,3]): method mode (line 61) | def mode(self): function normal_kl (line 65) | def normal_kl(mean1, logvar1, mean2, logvar2): FILE: data_generation/ldm/modules/ema.py class LitEma (line 5) | class LitEma(nn.Module): method __init__ (line 6) | def __init__(self, model, decay=0.9999, use_num_upates=True): method forward (line 25) | def forward(self,model): method copy_to (line 46) | def copy_to(self, model): method store (line 55) | def store(self, parameters): method restore (line 64) | def restore(self, parameters): FILE: data_generation/ldm/modules/encoders/modules.py class AbstractEncoder (line 12) | class AbstractEncoder(nn.Module): method __init__ (line 13) | def __init__(self): method encode (line 16) | def encode(self, *args, **kwargs): class ClassEmbedder (line 21) | class ClassEmbedder(nn.Module): method __init__ (line 22) | def __init__(self, embed_dim, n_classes=1000, key='class'): method forward (line 27) | def forward(self, batch, key=None): class TransformerEmbedder (line 36) | class TransformerEmbedder(AbstractEncoder): method __init__ (line 38) | def __init__(self, n_embed, n_layer, vocab_size, max_seq_len=77, devic... method forward (line 44) | def forward(self, tokens): method encode (line 49) | def encode(self, x): class BERTTokenizer (line 53) | class BERTTokenizer(AbstractEncoder): method __init__ (line 55) | def __init__(self, device="cuda", vq_interface=True, max_length=77): method forward (line 63) | def forward(self, text): method encode (line 70) | def encode(self, text): method decode (line 76) | def decode(self, text): class BERTEmbedder (line 80) | class BERTEmbedder(AbstractEncoder): method __init__ (line 82) | def __init__(self, n_embed, n_layer, vocab_size=30522, max_seq_len=77, method forward (line 93) | def forward(self, text): method encode (line 101) | def encode(self, text): class SpatialRescaler (line 106) | class SpatialRescaler(nn.Module): method __init__ (line 107) | def __init__(self, method forward (line 125) | def forward(self,x): method encode (line 134) | def encode(self, x): class FrozenCLIPEmbedder (line 137) | class FrozenCLIPEmbedder(AbstractEncoder): method __init__ (line 139) | def __init__(self, version="/home/zhaohaozhe/model2/clip-vit-large-pat... method freeze (line 147) | def freeze(self): method forward (line 152) | def forward(self, text): method encode (line 161) | def encode(self, text): class FrozenCLIPTextEmbedder (line 165) | class FrozenCLIPTextEmbedder(nn.Module): method __init__ (line 169) | def __init__(self, version='ViT-L/14', device="cuda", max_length=77, n... method freeze (line 177) | def freeze(self): method forward (line 182) | def forward(self, text): method encode (line 189) | def encode(self, text): class FrozenClipImageEmbedder (line 197) | class FrozenClipImageEmbedder(nn.Module): method __init__ (line 201) | def __init__( method preprocess (line 216) | def preprocess(self, x): method forward (line 226) | def forward(self, x): FILE: data_generation/ldm/modules/image_degradation/bsrgan.py function modcrop_np (line 29) | def modcrop_np(img, sf): function analytic_kernel (line 49) | def analytic_kernel(k): function anisotropic_Gaussian (line 65) | def anisotropic_Gaussian(ksize=15, theta=np.pi, l1=6, l2=6): function gm_blur_kernel (line 86) | def gm_blur_kernel(mean, cov, size=15): function shift_pixel (line 99) | def shift_pixel(x, sf, upper_left=True): function blur (line 128) | def blur(x, k): function gen_kernel (line 145) | def gen_kernel(k_size=np.array([15, 15]), scale_factor=np.array([4, 4]),... function fspecial_gaussian (line 187) | def fspecial_gaussian(hsize, sigma): function fspecial_laplacian (line 201) | def fspecial_laplacian(alpha): function fspecial (line 210) | def fspecial(filter_type, *args, **kwargs): function bicubic_degradation (line 228) | def bicubic_degradation(x, sf=3): function srmd_degradation (line 240) | def srmd_degradation(x, k, sf=3): function dpsr_degradation (line 262) | def dpsr_degradation(x, k, sf=3): function classical_degradation (line 284) | def classical_degradation(x, k, sf=3): function add_sharpening (line 299) | def add_sharpening(img, weight=0.5, radius=50, threshold=10): function add_blur (line 325) | def add_blur(img, sf=4): function add_resize (line 339) | def add_resize(img, sf=4): function add_Gaussian_noise (line 369) | def add_Gaussian_noise(img, noise_level1=2, noise_level2=25): function add_speckle_noise (line 386) | def add_speckle_noise(img, noise_level1=2, noise_level2=25): function add_Poisson_noise (line 404) | def add_Poisson_noise(img): function add_JPEG_noise (line 418) | def add_JPEG_noise(img): function random_crop (line 427) | def random_crop(lq, hq, sf=4, lq_patchsize=64): function degradation_bsrgan (line 438) | def degradation_bsrgan(img, sf=4, lq_patchsize=72, isp_model=None): function degradation_bsrgan_variant (line 530) | def degradation_bsrgan_variant(image, sf=4, isp_model=None): function degradation_bsrgan_plus (line 617) | def degradation_bsrgan_plus(img, sf=4, shuffle_prob=0.5, use_sharp=True,... FILE: data_generation/ldm/modules/image_degradation/bsrgan_light.py function modcrop_np (line 29) | def modcrop_np(img, sf): function analytic_kernel (line 49) | def analytic_kernel(k): function anisotropic_Gaussian (line 65) | def anisotropic_Gaussian(ksize=15, theta=np.pi, l1=6, l2=6): function gm_blur_kernel (line 86) | def gm_blur_kernel(mean, cov, size=15): function shift_pixel (line 99) | def shift_pixel(x, sf, upper_left=True): function blur (line 128) | def blur(x, k): function gen_kernel (line 145) | def gen_kernel(k_size=np.array([15, 15]), scale_factor=np.array([4, 4]),... function fspecial_gaussian (line 187) | def fspecial_gaussian(hsize, sigma): function fspecial_laplacian (line 201) | def fspecial_laplacian(alpha): function fspecial (line 210) | def fspecial(filter_type, *args, **kwargs): function bicubic_degradation (line 228) | def bicubic_degradation(x, sf=3): function srmd_degradation (line 240) | def srmd_degradation(x, k, sf=3): function dpsr_degradation (line 262) | def dpsr_degradation(x, k, sf=3): function classical_degradation (line 284) | def classical_degradation(x, k, sf=3): function add_sharpening (line 299) | def add_sharpening(img, weight=0.5, radius=50, threshold=10): function add_blur (line 325) | def add_blur(img, sf=4): function add_resize (line 343) | def add_resize(img, sf=4): function add_Gaussian_noise (line 373) | def add_Gaussian_noise(img, noise_level1=2, noise_level2=25): function add_speckle_noise (line 390) | def add_speckle_noise(img, noise_level1=2, noise_level2=25): function add_Poisson_noise (line 408) | def add_Poisson_noise(img): function add_JPEG_noise (line 422) | def add_JPEG_noise(img): function random_crop (line 431) | def random_crop(lq, hq, sf=4, lq_patchsize=64): function degradation_bsrgan (line 442) | def degradation_bsrgan(img, sf=4, lq_patchsize=72, isp_model=None): function degradation_bsrgan_variant (line 534) | def degradation_bsrgan_variant(image, sf=4, isp_model=None): FILE: data_generation/ldm/modules/image_degradation/utils_image.py function is_image_file (line 29) | def is_image_file(filename): function get_timestamp (line 33) | def get_timestamp(): function imshow (line 37) | def imshow(x, title=None, cbar=False, figsize=None): function surf (line 47) | def surf(Z, cmap='rainbow', figsize=None): function get_image_paths (line 67) | def get_image_paths(dataroot): function _get_paths_from_images (line 74) | def _get_paths_from_images(path): function patches_from_image (line 93) | def patches_from_image(img, p_size=512, p_overlap=64, p_max=800): function imssave (line 112) | def imssave(imgs, img_path): function split_imageset (line 125) | def split_imageset(original_dataroot, taget_dataroot, n_channels=3, p_si... function mkdir (line 153) | def mkdir(path): function mkdirs (line 158) | def mkdirs(paths): function mkdir_and_rename (line 166) | def mkdir_and_rename(path): function imread_uint (line 185) | def imread_uint(path, n_channels=3): function imsave (line 203) | def imsave(img, img_path): function imwrite (line 209) | def imwrite(img, img_path): function read_img (line 220) | def read_img(path): function uint2single (line 249) | def uint2single(img): function single2uint (line 254) | def single2uint(img): function uint162single (line 259) | def uint162single(img): function single2uint16 (line 264) | def single2uint16(img): function uint2tensor4 (line 275) | def uint2tensor4(img): function uint2tensor3 (line 282) | def uint2tensor3(img): function tensor2uint (line 289) | def tensor2uint(img): function single2tensor3 (line 302) | def single2tensor3(img): function single2tensor4 (line 307) | def single2tensor4(img): function tensor2single (line 312) | def tensor2single(img): function tensor2single3 (line 320) | def tensor2single3(img): function single2tensor5 (line 329) | def single2tensor5(img): function single32tensor5 (line 333) | def single32tensor5(img): function single42tensor4 (line 337) | def single42tensor4(img): function tensor2img (line 342) | def tensor2img(tensor, out_type=np.uint8, min_max=(0, 1)): function augment_img (line 380) | def augment_img(img, mode=0): function augment_img_tensor4 (line 401) | def augment_img_tensor4(img, mode=0): function augment_img_tensor (line 422) | def augment_img_tensor(img, mode=0): function augment_img_np3 (line 441) | def augment_img_np3(img, mode=0): function augment_imgs (line 469) | def augment_imgs(img_list, hflip=True, rot=True): function modcrop (line 494) | def modcrop(img_in, scale): function shave (line 510) | def shave(img_in, border=0): function rgb2ycbcr (line 529) | def rgb2ycbcr(img, only_y=True): function ycbcr2rgb (line 553) | def ycbcr2rgb(img): function bgr2ycbcr (line 573) | def bgr2ycbcr(img, only_y=True): function channel_convert (line 597) | def channel_convert(in_c, tar_type, img_list): function calculate_psnr (line 621) | def calculate_psnr(img1, img2, border=0): function calculate_ssim (line 642) | def calculate_ssim(img1, img2, border=0): function ssim (line 669) | def ssim(img1, img2): function cubic (line 700) | def cubic(x): function calculate_weights_indices (line 708) | def calculate_weights_indices(in_length, out_length, scale, kernel, kern... function imresize (line 766) | def imresize(img, scale, antialiasing=True): function imresize_np (line 839) | def imresize_np(img, scale, antialiasing=True): FILE: data_generation/ldm/modules/losses/contperceptual.py class LPIPSWithDiscriminator (line 7) | class LPIPSWithDiscriminator(nn.Module): method __init__ (line 8) | def __init__(self, disc_start, logvar_init=0.0, kl_weight=1.0, pixello... method calculate_adaptive_weight (line 32) | def calculate_adaptive_weight(self, nll_loss, g_loss, last_layer=None): method forward (line 45) | def forward(self, inputs, reconstructions, posteriors, optimizer_idx, FILE: data_generation/ldm/modules/losses/vqperceptual.py function hinge_d_loss_with_exemplar_weights (line 11) | def hinge_d_loss_with_exemplar_weights(logits_real, logits_fake, weights): function adopt_weight (line 20) | def adopt_weight(weight, global_step, threshold=0, value=0.): function measure_perplexity (line 26) | def measure_perplexity(predicted_indices, n_embed): function l1 (line 35) | def l1(x, y): function l2 (line 39) | def l2(x, y): class VQLPIPSWithDiscriminator (line 43) | class VQLPIPSWithDiscriminator(nn.Module): method __init__ (line 44) | def __init__(self, disc_start, codebook_weight=1.0, pixelloss_weight=1.0, method calculate_adaptive_weight (line 85) | def calculate_adaptive_weight(self, nll_loss, g_loss, last_layer=None): method forward (line 98) | def forward(self, codebook_loss, inputs, reconstructions, optimizer_idx, FILE: data_generation/ldm/modules/x_transformer.py class AbsolutePositionalEmbedding (line 25) | class AbsolutePositionalEmbedding(nn.Module): method __init__ (line 26) | def __init__(self, dim, max_seq_len): method init_ (line 31) | def init_(self): method forward (line 34) | def forward(self, x): class FixedPositionalEmbedding (line 39) | class FixedPositionalEmbedding(nn.Module): method __init__ (line 40) | def __init__(self, dim): method forward (line 45) | def forward(self, x, seq_dim=1, offset=0): function exists (line 54) | def exists(val): function default (line 58) | def default(val, d): function always (line 64) | def always(val): function not_equals (line 70) | def not_equals(val): function equals (line 76) | def equals(val): function max_neg_value (line 82) | def max_neg_value(tensor): function pick_and_pop (line 88) | def pick_and_pop(keys, d): function group_dict_by_key (line 93) | def group_dict_by_key(cond, d): function string_begins_with (line 102) | def string_begins_with(prefix, str): function group_by_key_prefix (line 106) | def group_by_key_prefix(prefix, d): function groupby_prefix_and_trim (line 110) | def groupby_prefix_and_trim(prefix, d): class Scale (line 117) | class Scale(nn.Module): method __init__ (line 118) | def __init__(self, value, fn): method forward (line 123) | def forward(self, x, **kwargs): class Rezero (line 128) | class Rezero(nn.Module): method __init__ (line 129) | def __init__(self, fn): method forward (line 134) | def forward(self, x, **kwargs): class ScaleNorm (line 139) | class ScaleNorm(nn.Module): method __init__ (line 140) | def __init__(self, dim, eps=1e-5): method forward (line 146) | def forward(self, x): class RMSNorm (line 151) | class RMSNorm(nn.Module): method __init__ (line 152) | def __init__(self, dim, eps=1e-8): method forward (line 158) | def forward(self, x): class Residual (line 163) | class Residual(nn.Module): method forward (line 164) | def forward(self, x, residual): class GRUGating (line 168) | class GRUGating(nn.Module): method __init__ (line 169) | def __init__(self, dim): method forward (line 173) | def forward(self, x, residual): class GEGLU (line 184) | class GEGLU(nn.Module): method __init__ (line 185) | def __init__(self, dim_in, dim_out): method forward (line 189) | def forward(self, x): class FeedForward (line 194) | class FeedForward(nn.Module): method __init__ (line 195) | def __init__(self, dim, dim_out=None, mult=4, glu=False, dropout=0.): method forward (line 210) | def forward(self, x): class Attention (line 215) | class Attention(nn.Module): method __init__ (line 216) | def __init__( method forward (line 268) | def forward( class AttentionLayers (line 370) | class AttentionLayers(nn.Module): method __init__ (line 371) | def __init__( method forward (line 481) | def forward( class Encoder (line 541) | class Encoder(AttentionLayers): method __init__ (line 542) | def __init__(self, **kwargs): class TransformerWrapper (line 548) | class TransformerWrapper(nn.Module): method __init__ (line 549) | def __init__( method init_ (line 595) | def init_(self): method forward (line 598) | def forward( FILE: data_generation/ldm/util.py function log_txt_as_img (line 17) | def log_txt_as_img(wh, xc, size=10): function ismap (line 41) | def ismap(x): function isimage (line 47) | def isimage(x): function exists (line 53) | def exists(x): function default (line 57) | def default(val, d): function mean_flat (line 63) | def mean_flat(tensor): function count_params (line 71) | def count_params(model, verbose=False): function instantiate_from_config (line 78) | def instantiate_from_config(config): function get_obj_from_str (line 88) | def get_obj_from_str(string, reload=False): function _do_parallel_data_prefetch (line 96) | def _do_parallel_data_prefetch(func, Q, data, idx, idx_to_fn=False): function parallel_data_prefetch (line 108) | def parallel_data_prefetch( FILE: data_generation/metrics/clip_similarity.py class ClipSimilarity (line 14) | class ClipSimilarity(nn.Module): method __init__ (line 15) | def __init__(self, name: str = "ViT-L/14"): method encode_text (line 28) | def encode_text(self, text: list[str]) -> torch.Tensor: method encode_image (line 34) | def encode_image(self, image: torch.Tensor) -> torch.Tensor: # Input ... method get_dinov2_embedding (line 41) | def get_dinov2_embedding(self, image: torch.Tensor) -> torch.Tensor: method get_ssim_sim (line 47) | def get_ssim_sim(self, image1,image2): method forward (line 62) | def forward( FILE: data_generation/metrics/compute_metrics.py class CFGDenoiser (line 34) | class CFGDenoiser(nn.Module): method __init__ (line 35) | def __init__(self, model): method forward (line 39) | def forward(self, z, sigma, cond, uncond, text_cfg_scale, image_cfg_sc... function load_model_from_config (line 50) | def load_model_from_config(config, ckpt, vae_ckpt=None, verbose=False): class ImageEditor (line 73) | class ImageEditor(nn.Module): method __init__ (line 74) | def __init__(self, config, ckpt, vae_ckpt=None): method forward (line 84) | def forward( function compute_metrics (line 117) | def compute_metrics(config, function plot_metrics (line 186) | def plot_metrics(metrics_file, output_path): function main (line 205) | def main(): FILE: data_generation/processors.py class P2PCrossAttnProcessor (line 11) | class P2PCrossAttnProcessor: method __init__ (line 12) | def __init__(self, controller, place_in_unet): method __call__ (line 17) | def __call__(self, attn: Attention, hidden_states, encoder_hidden_stat... function create_controller (line 48) | def create_controller( class AttentionControl (line 129) | class AttentionControl(abc.ABC): method step_callback (line 130) | def step_callback(self, x_t): method between_steps (line 133) | def between_steps(self): method num_uncond_att_layers (line 137) | def num_uncond_att_layers(self): method forward (line 141) | def forward(self, attn, is_cross: bool, place_in_unet: str): method __call__ (line 144) | def __call__(self, attn, is_cross: bool, place_in_unet: str): method reset (line 155) | def reset(self): method __init__ (line 159) | def __init__(self, attn_res=None): class EmptyControl (line 166) | class EmptyControl(AttentionControl): method forward (line 167) | def forward(self, attn, is_cross: bool, place_in_unet: str): class AttentionStore (line 171) | class AttentionStore(AttentionControl): method get_empty_store (line 173) | def get_empty_store(): method forward (line 176) | def forward(self, attn, is_cross: bool, place_in_unet: str): method between_steps (line 182) | def between_steps(self): method get_average_attention (line 191) | def get_average_attention(self): method reset (line 197) | def reset(self): method __init__ (line 202) | def __init__(self, attn_res=None): class LocalBlend (line 209) | class LocalBlend: method __call__ (line 210) | def __call__(self, x_t, attention_store): method __init__ (line 229) | def __init__( class AttentionControlEdit (line 246) | class AttentionControlEdit(AttentionStore, abc.ABC): method step_callback (line 247) | def step_callback(self, x_t): method replace_self_attention (line 252) | def replace_self_attention(self, attn_base, att_replace): method replace_cross_attention (line 259) | def replace_cross_attention(self, attn_base, att_replace): method forward (line 262) | def forward(self, attn, is_cross: bool, place_in_unet: str): method __init__ (line 280) | def __init__( class AttentionReplace (line 307) | class AttentionReplace(AttentionControlEdit): method replace_cross_attention (line 308) | def replace_cross_attention(self, attn_base, att_replace): method __init__ (line 311) | def __init__( class AttentionRefine (line 328) | class AttentionRefine(AttentionControlEdit): method replace_cross_attention (line 329) | def replace_cross_attention(self, attn_base, att_replace): method __init__ (line 334) | def __init__( class AttentionReweight (line 353) | class AttentionReweight(AttentionControlEdit): method replace_cross_attention (line 354) | def replace_cross_attention(self, attn_base, att_replace): method __init__ (line 360) | def __init__( function update_alpha_time_word (line 381) | def update_alpha_time_word( function get_time_words_attention_alpha (line 395) | def get_time_words_attention_alpha( function get_word_inds (line 416) | def get_word_inds(text: str, word_place: int, tokenizer): function get_replacement_mapper_ (line 438) | def get_replacement_mapper_(x: str, y: str, tokenizer, max_len=77): function get_replacement_mapper (line 477) | def get_replacement_mapper(prompts, tokenizer, max_len=77): function get_equalizer (line 487) | def get_equalizer( class ScoreParams (line 501) | class ScoreParams: method __init__ (line 502) | def __init__(self, gap, match, mismatch): method mis_match_char (line 507) | def mis_match_char(self, x, y): function get_matrix (line 514) | def get_matrix(size_x, size_y, gap): function get_traceback_matrix (line 521) | def get_traceback_matrix(size_x, size_y): function global_align (line 529) | def global_align(x, y, score): function get_aligned_sequences (line 547) | def get_aligned_sequences(x, y, trace_back): function get_mapper (line 575) | def get_mapper(x: str, y: str, tokenizer, max_len=77): function get_refinement_mapper (line 589) | def get_refinement_mapper(prompts, tokenizer, max_len=77): FILE: data_generation/prompt_to_prompt_pipeline.py function rescale_noise_cfg (line 7) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class Prompt2PromptPipeline (line 21) | class Prompt2PromptPipeline(StableDiffusionXLPipeline): method check_inputs (line 49) | def check_inputs( method _aggregate_and_get_attention_maps_per_token (line 122) | def _aggregate_and_get_attention_maps_per_token(self, with_softmax): method _get_attention_maps_list (line 134) | def _get_attention_maps_list( method __call__ (line 148) | def __call__( method register_attention_control (line 452) | def register_attention_control(self, controller): FILE: data_generation/sdxl_p2p_pipeline.py function rescale_noise_cfg (line 10) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class Prompt2PromptPipeline (line 24) | class Prompt2PromptPipeline(StableDiffusionXLPipeline): method check_inputs (line 52) | def check_inputs( method _aggregate_and_get_attention_maps_per_token (line 125) | def _aggregate_and_get_attention_maps_per_token(self, with_softmax): method _get_attention_maps_list (line 137) | def _get_attention_maps_list( method __call__ (line 151) | def __call__( method register_attention_control (line 455) | def register_attention_control(self, controller): function retrieve_timesteps (line 486) | def retrieve_timesteps( class Prompt2PromptImg2ImgPipeline (line 529) | class Prompt2PromptImg2ImgPipeline(StableDiffusionXLImg2ImgPipeline): method check_inputs (line 557) | def check_inputs( method _aggregate_and_get_attention_maps_per_token (line 624) | def _aggregate_and_get_attention_maps_per_token(self, with_softmax): method _get_attention_maps_list (line 636) | def _get_attention_maps_list( method __call__ (line 650) | def __call__( method register_attention_control (line 1040) | def register_attention_control(self, controller): class Prompt2PromptInpaintPipeline (line 1068) | class Prompt2PromptInpaintPipeline(StableDiffusionXLInpaintPipeline): method check_inputs (line 1096) | def check_inputs( method _aggregate_and_get_attention_maps_per_token (line 1175) | def _aggregate_and_get_attention_maps_per_token(self, with_softmax): method _get_attention_maps_list (line 1187) | def _get_attention_maps_list( method __call__ (line 1201) | def __call__( method register_attention_control (line 1776) | def register_attention_control(self, controller): FILE: data_generation/util.py class P2PCrossAttnProcessor (line 11) | class P2PCrossAttnProcessor: method __init__ (line 12) | def __init__(self, controller, place_in_unet): method __call__ (line 17) | def __call__(self, attn: Attention, hidden_states, encoder_hidden_stat... function create_controller (line 48) | def create_controller( class AttentionControl (line 129) | class AttentionControl(abc.ABC): method step_callback (line 130) | def step_callback(self, x_t): method between_steps (line 133) | def between_steps(self): method num_uncond_att_layers (line 137) | def num_uncond_att_layers(self): method forward (line 141) | def forward(self, attn, is_cross: bool, place_in_unet: str): method __call__ (line 144) | def __call__(self, attn, is_cross: bool, place_in_unet: str): method reset (line 155) | def reset(self): method __init__ (line 159) | def __init__(self, attn_res=None): class EmptyControl (line 166) | class EmptyControl(AttentionControl): method forward (line 167) | def forward(self, attn, is_cross: bool, place_in_unet: str): class AttentionStore (line 171) | class AttentionStore(AttentionControl): method get_empty_store (line 173) | def get_empty_store(): method forward (line 176) | def forward(self, attn, is_cross: bool, place_in_unet: str): method between_steps (line 182) | def between_steps(self): method get_average_attention (line 191) | def get_average_attention(self): method reset (line 197) | def reset(self): method __init__ (line 202) | def __init__(self, attn_res=None): class LocalBlend (line 209) | class LocalBlend: method __call__ (line 210) | def __call__(self, x_t, attention_store): method __init__ (line 229) | def __init__( class AttentionControlEdit (line 246) | class AttentionControlEdit(AttentionStore, abc.ABC): method step_callback (line 247) | def step_callback(self, x_t): method replace_self_attention (line 252) | def replace_self_attention(self, attn_base, att_replace): method replace_cross_attention (line 259) | def replace_cross_attention(self, attn_base, att_replace): method forward (line 262) | def forward(self, attn, is_cross: bool, place_in_unet: str): method __init__ (line 280) | def __init__( class AttentionReplace (line 307) | class AttentionReplace(AttentionControlEdit): method replace_cross_attention (line 308) | def replace_cross_attention(self, attn_base, att_replace): method __init__ (line 314) | def __init__( class AttentionRefine (line 332) | class AttentionRefine(AttentionControlEdit): method replace_cross_attention (line 333) | def replace_cross_attention(self, attn_base, att_replace): method __init__ (line 338) | def __init__( class AttentionReweight (line 357) | class AttentionReweight(AttentionControlEdit): method replace_cross_attention (line 358) | def replace_cross_attention(self, attn_base, att_replace): method __init__ (line 364) | def __init__( function update_alpha_time_word (line 385) | def update_alpha_time_word( function get_time_words_attention_alpha (line 399) | def get_time_words_attention_alpha( function get_word_inds (line 420) | def get_word_inds(text: str, word_place: int, tokenizer): function get_replacement_mapper_ (line 442) | def get_replacement_mapper_(x: str, y: str, tokenizer, max_len=77): function get_replacement_mapper (line 482) | def get_replacement_mapper(prompts, tokenizer, max_len=77,torch_dtype=to... function get_equalizer (line 500) | def get_equalizer( class ScoreParams (line 514) | class ScoreParams: method __init__ (line 515) | def __init__(self, gap, match, mismatch): method mis_match_char (line 520) | def mis_match_char(self, x, y): function get_matrix (line 527) | def get_matrix(size_x, size_y, gap): function get_traceback_matrix (line 534) | def get_traceback_matrix(size_x, size_y): function global_align (line 542) | def global_align(x, y, score): function get_aligned_sequences (line 560) | def get_aligned_sequences(x, y, trace_back): function get_mapper (line 588) | def get_mapper(x: str, y: str, tokenizer, max_len=77): function get_refinement_mapper (line 602) | def get_refinement_mapper(prompts, tokenizer, max_len=77): FILE: diffusers/benchmarks/base_classes.py class BaseBenchmak (line 49) | class BaseBenchmak: method __init__ (line 52) | def __init__(self, args): method run_inference (line 55) | def run_inference(self, args): method benchmark (line 58) | def benchmark(self, args): method get_result_filepath (line 61) | def get_result_filepath(self, args): class TextToImageBenchmark (line 73) | class TextToImageBenchmark(BaseBenchmak): method __init__ (line 76) | def __init__(self, args): method run_inference (line 97) | def run_inference(self, pipe, args): method benchmark (line 104) | def benchmark(self, args): class TurboTextToImageBenchmark (line 124) | class TurboTextToImageBenchmark(TextToImageBenchmark): method __init__ (line 125) | def __init__(self, args): method run_inference (line 128) | def run_inference(self, pipe, args): class LCMLoRATextToImageBenchmark (line 137) | class LCMLoRATextToImageBenchmark(TextToImageBenchmark): method __init__ (line 140) | def __init__(self, args): method get_result_filepath (line 147) | def get_result_filepath(self, args): method run_inference (line 158) | def run_inference(self, pipe, args): method benchmark (line 166) | def benchmark(self, args): class ImageToImageBenchmark (line 186) | class ImageToImageBenchmark(TextToImageBenchmark): method __init__ (line 191) | def __init__(self, args): method run_inference (line 195) | def run_inference(self, pipe, args): class TurboImageToImageBenchmark (line 204) | class TurboImageToImageBenchmark(ImageToImageBenchmark): method __init__ (line 205) | def __init__(self, args): method run_inference (line 208) | def run_inference(self, pipe, args): class InpaintingBenchmark (line 219) | class InpaintingBenchmark(ImageToImageBenchmark): method __init__ (line 224) | def __init__(self, args): method run_inference (line 229) | def run_inference(self, pipe, args): class IPAdapterTextToImageBenchmark (line 239) | class IPAdapterTextToImageBenchmark(TextToImageBenchmark): method __init__ (line 243) | def __init__(self, args): method run_inference (line 259) | def run_inference(self, pipe, args): class ControlNetBenchmark (line 268) | class ControlNetBenchmark(TextToImageBenchmark): method __init__ (line 276) | def __init__(self, args): method run_inference (line 294) | def run_inference(self, pipe, args): class ControlNetSDXLBenchmark (line 303) | class ControlNetSDXLBenchmark(ControlNetBenchmark): method __init__ (line 307) | def __init__(self, args): class T2IAdapterBenchmark (line 311) | class T2IAdapterBenchmark(ControlNetBenchmark): method __init__ (line 319) | def __init__(self, args): class T2IAdapterSDXLBenchmark (line 338) | class T2IAdapterSDXLBenchmark(T2IAdapterBenchmark): method __init__ (line 345) | def __init__(self, args): FILE: diffusers/benchmarks/push_results.py function has_previous_benchmark (line 13) | def has_previous_benchmark() -> str: function filter_float (line 22) | def filter_float(value): function push_to_hf_dataset (line 28) | def push_to_hf_dataset(): FILE: diffusers/benchmarks/run_all.py class SubprocessCallException (line 14) | class SubprocessCallException(Exception): function run_command (line 19) | def run_command(command: List[str], return_stdout=False): function main (line 36) | def main(): FILE: diffusers/benchmarks/utils.py class BenchmarkInfo (line 35) | class BenchmarkInfo: function flush (line 40) | def flush(): function bytes_to_giga_bytes (line 48) | def bytes_to_giga_bytes(bytes): function benchmark_fn (line 52) | def benchmark_fn(f, *args, **kwargs): function generate_csv_dict (line 61) | def generate_csv_dict( function write_to_csv (line 80) | def write_to_csv(file_name: str, data_dict: Dict[str, Union[str, bool, f... function collate_csv (line 88) | def collate_csv(input_files: List[str], output_file: str): FILE: diffusers/examples/advanced_diffusion_training/train_dreambooth_lora_sd15_advanced.py function save_model_card (line 79) | def save_model_card( function import_model_class_from_model_name_or_path (line 209) | def import_model_class_from_model_name_or_path( function parse_args (line 229) | def parse_args(input_args=None): class TokenEmbeddingsHandler (line 712) | class TokenEmbeddingsHandler: method __init__ (line 713) | def __init__(self, text_encoders, tokenizers): method initialize_new_tokens (line 721) | def initialize_new_tokens(self, inserting_toks: List[str]): method save_embeddings (line 762) | def save_embeddings(self, file_path: str): method dtype (line 782) | def dtype(self): method device (line 786) | def device(self): method retract_embeddings (line 790) | def retract_embeddings(self): class DreamBoothDataset (line 811) | class DreamBoothDataset(Dataset): method __init__ (line 817) | def __init__( method __len__ (line 928) | def __len__(self): method __getitem__ (line 931) | def __getitem__(self, index): function collate_fn (line 966) | def collate_fn(examples, with_prior_preservation=False): class PromptDataset (line 983) | class PromptDataset(Dataset): method __init__ (line 986) | def __init__(self, prompt, num_samples): method __len__ (line 990) | def __len__(self): method __getitem__ (line 993) | def __getitem__(self, index): function tokenize_prompt (line 1000) | def tokenize_prompt(tokenizer, prompt, add_special_tokens=False): function encode_prompt (line 1014) | def encode_prompt(text_encoders, tokenizers, prompt, text_input_ids_list... function main (line 1031) | def main(args): FILE: diffusers/examples/advanced_diffusion_training/train_dreambooth_lora_sdxl_advanced.py function determine_scheduler_type (line 86) | def determine_scheduler_type(pretrained_model_name_or_path, revision): function save_model_card (line 100) | def save_model_card( function import_model_class_from_model_name_or_path (line 231) | def import_model_class_from_model_name_or_path( function parse_args (line 251) | def parse_args(input_args=None): function is_belong_to_blocks (line 766) | def is_belong_to_blocks(key, blocks): function get_unet_lora_target_modules (line 776) | def get_unet_lora_target_modules(unet, use_blora, target_blocks=None): class TokenEmbeddingsHandler (line 800) | class TokenEmbeddingsHandler: method __init__ (line 801) | def __init__(self, text_encoders, tokenizers): method initialize_new_tokens (line 809) | def initialize_new_tokens(self, inserting_toks: List[str]): method save_embeddings (line 849) | def save_embeddings(self, file_path: str): method dtype (line 869) | def dtype(self): method device (line 873) | def device(self): method retract_embeddings (line 877) | def retract_embeddings(self): class DreamBoothDataset (line 898) | class DreamBoothDataset(Dataset): method __init__ (line 904) | def __init__( method __len__ (line 1080) | def __len__(self): method __getitem__ (line 1083) | def __getitem__(self, index): function collate_fn (line 1112) | def collate_fn(examples, with_prior_preservation=False): class PromptDataset (line 1138) | class PromptDataset(Dataset): method __init__ (line 1141) | def __init__(self, prompt, num_samples): method __len__ (line 1145) | def __len__(self): method __getitem__ (line 1148) | def __getitem__(self, index): function tokenize_prompt (line 1155) | def tokenize_prompt(tokenizer, prompt, add_special_tokens=False): function encode_prompt (line 1169) | def encode_prompt(text_encoders, tokenizers, prompt, text_input_ids_list... function main (line 1197) | def main(args): FILE: diffusers/examples/amused/train_amused.py function parse_args (line 54) | def parse_args(): class InstanceDataRootDataset (line 301) | class InstanceDataRootDataset(Dataset): method __init__ (line 302) | def __init__( method __len__ (line 312) | def __len__(self): method __getitem__ (line 315) | def __getitem__(self, index): class InstanceDataImageDataset (line 325) | class InstanceDataImageDataset(Dataset): method __init__ (line 326) | def __init__( method __len__ (line 335) | def __len__(self): method __getitem__ (line 340) | def __getitem__(self, index): class HuggingFaceDataset (line 344) | class HuggingFaceDataset(Dataset): method __init__ (line 345) | def __init__( method __len__ (line 361) | def __len__(self): method __getitem__ (line 364) | def __getitem__(self, index): function process_image (line 379) | def process_image(image, size): function tokenize_prompt (line 402) | def tokenize_prompt(tokenizer, prompt): function encode_prompt (line 412) | def encode_prompt(text_encoder, input_ids): function main (line 419) | def main(args): function save_checkpoint (line 946) | def save_checkpoint(args, accelerator, global_step): FILE: diffusers/examples/community/bit_diffusion.py function decimal_to_bits (line 15) | def decimal_to_bits(x, bits=BITS): function bits_to_decimal (line 31) | def bits_to_decimal(x, bits=BITS): function ddim_bit_scheduler_step (line 45) | def ddim_bit_scheduler_step( function ddpm_bit_scheduler_step (line 135) | def ddpm_bit_scheduler_step( class BitDiffusion (line 213) | class BitDiffusion(DiffusionPipeline): method __init__ (line 214) | def __init__( method __call__ (line 229) | def __call__( FILE: diffusers/examples/community/checkpoint_merger.py class CheckpointMergerPipeline (line 15) | class CheckpointMergerPipeline(DiffusionPipeline): method __init__ (line 36) | def __init__(self): method _compare_model_configs (line 40) | def _compare_model_configs(self, dict0, dict1): method _remove_meta_keys (line 51) | def _remove_meta_keys(self, config_dict: Dict): method merge (line 62) | def merge(self, pretrained_model_name_or_path_list: List[Union[str, os... method weighted_sum (line 268) | def weighted_sum(theta0, theta1, theta2, alpha): method sigmoid (line 273) | def sigmoid(theta0, theta1, theta2, alpha): method inv_sigmoid (line 279) | def inv_sigmoid(theta0, theta1, theta2, alpha): method add_difference (line 286) | def add_difference(theta0, theta1, theta2, alpha): FILE: diffusers/examples/community/clip_guided_images_mixing_stable_diffusion.py function preprocess (line 26) | def preprocess(image, w, h): function slerp (line 44) | def slerp(t, v0, v1, DOT_THRESHOLD=0.9995): function spherical_dist_loss (line 69) | def spherical_dist_loss(x, y): function set_requires_grad (line 75) | def set_requires_grad(model, value): class CLIPGuidedImagesMixingStableDiffusion (line 80) | class CLIPGuidedImagesMixingStableDiffusion(DiffusionPipeline, StableDif... method __init__ (line 81) | def __init__( method freeze_vae (line 116) | def freeze_vae(self): method unfreeze_vae (line 119) | def unfreeze_vae(self): method freeze_unet (line 122) | def freeze_unet(self): method unfreeze_unet (line 125) | def unfreeze_unet(self): method get_timesteps (line 128) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 137) | def prepare_latents(self, image, timestep, batch_size, dtype, device, ... method get_image_description (line 163) | def get_image_description(self, image): method get_clip_image_embeddings (line 170) | def get_clip_image_embeddings(self, image, batch_size): method cond_fn (line 179) | def cond_fn( method __call__ (line 234) | def __call__( FILE: diffusers/examples/community/clip_guided_stable_diffusion.py class MakeCutouts (line 22) | class MakeCutouts(nn.Module): method __init__ (line 23) | def __init__(self, cut_size, cut_power=1.0): method forward (line 29) | def forward(self, pixel_values, num_cutouts): function spherical_dist_loss (line 43) | def spherical_dist_loss(x, y): function set_requires_grad (line 49) | def set_requires_grad(model, value): class CLIPGuidedStableDiffusion (line 54) | class CLIPGuidedStableDiffusion(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 60) | def __init__( method freeze_vae (line 92) | def freeze_vae(self): method unfreeze_vae (line 95) | def unfreeze_vae(self): method freeze_unet (line 98) | def freeze_unet(self): method unfreeze_unet (line 101) | def unfreeze_unet(self): method cond_fn (line 105) | def cond_fn( method __call__ (line 169) | def __call__( FILE: diffusers/examples/community/clip_guided_stable_diffusion_img2img.py function preprocess (line 78) | def preprocess(image, w, h): class MakeCutouts (line 96) | class MakeCutouts(nn.Module): method __init__ (line 97) | def __init__(self, cut_size, cut_power=1.0): method forward (line 103) | def forward(self, pixel_values, num_cutouts): function spherical_dist_loss (line 117) | def spherical_dist_loss(x, y): function set_requires_grad (line 123) | def set_requires_grad(model, value): class CLIPGuidedStableDiffusion (line 128) | class CLIPGuidedStableDiffusion(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 134) | def __init__( method freeze_vae (line 166) | def freeze_vae(self): method unfreeze_vae (line 169) | def unfreeze_vae(self): method freeze_unet (line 172) | def freeze_unet(self): method unfreeze_unet (line 175) | def unfreeze_unet(self): method get_timesteps (line 178) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 187) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method cond_fn (line 240) | def cond_fn( method __call__ (line 304) | def __call__( FILE: diffusers/examples/community/composable_stable_diffusion.py class ComposableStableDiffusionPipeline (line 42) | class ComposableStableDiffusionPipeline(DiffusionPipeline, StableDiffusi... method __init__ (line 72) | def __init__( method _encode_prompt (line 168) | def _encode_prompt(self, prompt, device, num_images_per_prompt, do_cla... method run_safety_checker (line 273) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 283) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 291) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 308) | def check_inputs(self, prompt, height, width, callback_steps): method prepare_latents (line 323) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 346) | def __call__( FILE: diffusers/examples/community/ddim_noise_comparative_analysis.py function preprocess (line 35) | def preprocess(image): class DDIMNoiseComparativeAnalysisPipeline (line 46) | class DDIMNoiseComparativeAnalysisPipeline(DiffusionPipeline): method __init__ (line 58) | def __init__(self, unet, scheduler): method check_inputs (line 66) | def check_inputs(self, strength): method get_timesteps (line 70) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 79) | def prepare_latents(self, image, timestep, batch_size, dtype, device, ... method __call__ (line 104) | def __call__( FILE: diffusers/examples/community/dps_pipeline.py class DPSPipeline (line 27) | class DPSPipeline(DiffusionPipeline): method __init__ (line 44) | def __init__(self, unet, scheduler): method __call__ (line 49) | def __call__( class SuperResolutionOperator (line 167) | class SuperResolutionOperator(nn.Module): method __init__ (line 168) | def __init__(self, in_shape, scale_factor): method forward (line 371) | def forward(self, data, **kwargs): class GaussialBlurOperator (line 375) | class GaussialBlurOperator(nn.Module): method __init__ (line 376) | def __init__(self, kernel_size, intensity): method forward (line 421) | def forward(self, data, **kwargs): method transpose (line 424) | def transpose(self, data, **kwargs): method get_kernel (line 427) | def get_kernel(self): function RMSELoss (line 431) | def RMSELoss(yhat, y): FILE: diffusers/examples/community/edict_pipeline.py class EDICTPipeline (line 15) | class EDICTPipeline(DiffusionPipeline): method __init__ (line 16) | def __init__( method _encode_prompt (line 41) | def _encode_prompt( method denoise_mixing_layer (line 73) | def denoise_mixing_layer(self, x: torch.Tensor, y: torch.Tensor): method noise_mixing_layer (line 79) | def noise_mixing_layer(self, x: torch.Tensor, y: torch.Tensor): method _get_alpha_and_beta (line 85) | def _get_alpha_and_beta(self, t: torch.Tensor): method noise_step (line 93) | def noise_step( method denoise_step (line 112) | def denoise_step( method decode_latents (line 131) | def decode_latents(self, latents: torch.Tensor): method prepare_latents (line 138) | def prepare_latents( method __call__ (line 189) | def __call__( FILE: diffusers/examples/community/fresco_v2v.py function clear_cache (line 52) | def clear_cache(): function coords_grid (line 57) | def coords_grid(b, h, w, homogeneous=False, device=None): function bilinear_sample (line 76) | def bilinear_sample(img, sample_coords, mode="bilinear", padding_mode="z... class Dilate (line 100) | class Dilate: method __init__ (line 101) | def __init__(self, kernel_size=7, channels=1, device="cpu"): method __call__ (line 110) | def __call__(self, x): function flow_warp (line 115) | def flow_warp(feature, flow, mask=False, mode="bilinear", padding_mode="... function forward_backward_consistency_check (line 124) | def forward_backward_consistency_check(fwd_flow, bwd_flow, alpha=0.01, b... function numpy2tensor (line 146) | def numpy2tensor(img): function calc_mean_std (line 153) | def calc_mean_std(feat, eps=1e-5, chunk=1): function adaptive_instance_normalization (line 165) | def adaptive_instance_normalization(content_feat, style_feat, chunk=1): function optimize_feature (line 175) | def optimize_feature( function warp_tensor (line 263) | def warp_tensor(sample, flows, occs, saliency, unet_chunk_size): function my_forward (line 300) | def my_forward( function get_single_mapping_ind (line 641) | def get_single_mapping_ind(bwd_flow, bwd_occ, imgs, scale=1.0): function get_mapping_ind (line 697) | def get_mapping_ind(bwd_flows, bwd_occs, imgs, scale=1.0): function apply_FRESCO_opt (line 736) | def apply_FRESCO_opt( function get_intraframe_paras (line 757) | def get_intraframe_paras(pipe, imgs, frescoProc, prompt_embeds, do_class... function get_flow_and_interframe_paras (line 806) | def get_flow_and_interframe_paras(flow_model, imgs): class AttentionControl (line 863) | class AttentionControl: method __init__ (line 872) | def __init__(self): method get_empty_store (line 886) | def get_empty_store(): method clear_store (line 891) | def clear_store(self): method enable_store (line 899) | def enable_store(self): method disable_store (line 902) | def disable_store(self): method enable_intraattn (line 906) | def enable_intraattn(self): method disable_intraattn (line 913) | def disable_intraattn(self): method disable_cfattn (line 918) | def disable_cfattn(self): method enable_cfattn (line 922) | def enable_cfattn(self, attn_mask=None): method disable_interattn (line 936) | def disable_interattn(self): method enable_interattn (line 940) | def enable_interattn(self, interattn_paras=None): method disable_controller (line 954) | def disable_controller(self): method enable_controller (line 959) | def enable_controller(self, interattn_paras=None, attn_mask=None): method forward (line 964) | def forward(self, context): method __call__ (line 976) | def __call__(self, context): class FRESCOAttnProcessor2_0 (line 981) | class FRESCOAttnProcessor2_0: method __init__ (line 1002) | def __init__(self, unet_chunk_size=2, controller=None): method __call__ (line 1008) | def __call__( function apply_FRESCO_attn (line 1193) | def apply_FRESCO_attn(pipe): function retrieve_latents (line 1209) | def retrieve_latents( function prepare_image (line 1222) | def prepare_image(image): class FrescoV2VPipeline (line 1246) | class FrescoV2VPipeline(StableDiffusionControlNetImg2ImgPipeline): method __init__ (line 1289) | def __init__( method _encode_prompt (line 1383) | def _encode_prompt( method encode_prompt (line 1416) | def encode_prompt( method encode_image (line 1598) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 1623) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 1675) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 1690) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 1702) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 1719) | def check_inputs( method check_image (line 1877) | def check_image(self, image, prompt, prompt_embeds): method prepare_control_image (line 1915) | def prepare_control_image( method get_timesteps (line 1946) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 1958) | def prepare_latents( method guidance_scale (line 2024) | def guidance_scale(self): method clip_skip (line 2028) | def clip_skip(self): method do_classifier_free_guidance (line 2035) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 2039) | def cross_attention_kwargs(self): method num_timesteps (line 2043) | def num_timesteps(self): method __call__ (line 2047) | def __call__( FILE: diffusers/examples/community/gluegen.py class TranslatorBase (line 29) | class TranslatorBase(nn.Module): method __init__ (line 30) | def __init__(self, num_tok, dim, dim_out, mult=2): method forward (line 58) | def forward(self, x): class TranslatorBaseNoLN (line 78) | class TranslatorBaseNoLN(nn.Module): method __init__ (line 79) | def __init__(self, num_tok, dim, dim_out, mult=2): method forward (line 101) | def forward(self, x): class TranslatorNoLN (line 121) | class TranslatorNoLN(nn.Module): method __init__ (line 122) | def __init__(self, num_tok, dim, dim_out, mult=2, depth=5): method forward (line 130) | def forward(self, x): function rescale_noise_cfg (line 139) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 153) | def retrieve_timesteps( class GlueGenStableDiffusionPipeline (line 197) | class GlueGenStableDiffusionPipeline(DiffusionPipeline, StableDiffusionM... method __init__ (line 198) | def __init__( method load_language_adapter (line 228) | def load_language_adapter( method _adapt_language (line 245) | def _adapt_language(self, prompt_embeds: torch.Tensor): method encode_prompt (line 250) | def encode_prompt( method run_safety_checker (line 433) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 447) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 464) | def check_inputs( method prepare_latents (line 502) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 525) | def get_guidance_scale_embedding(self, w, embedding_dim=512, dtype=tor... method guidance_scale (line 554) | def guidance_scale(self): method guidance_rescale (line 558) | def guidance_rescale(self): method clip_skip (line 562) | def clip_skip(self): method do_classifier_free_guidance (line 569) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 573) | def cross_attention_kwargs(self): method num_timesteps (line 577) | def num_timesteps(self): method interrupt (line 581) | def interrupt(self): method __call__ (line 585) | def __call__( FILE: diffusers/examples/community/hd_painter.py class RASGAttnProcessor (line 20) | class RASGAttnProcessor: method __init__ (line 21) | def __init__(self, mask, token_idx, scale_factor): method __call__ (line 28) | def __call__( class PAIntAAttnProcessor (line 100) | class PAIntAAttnProcessor: method __init__ (line 101) | def __init__(self, transformer_block, mask, token_idx, do_classifier_f... method __call__ (line 111) | def __call__( class StableDiffusionHDPainterPipeline (line 401) | class StableDiffusionHDPainterPipeline(StableDiffusionInpaintPipeline): method get_tokenized_prompt (line 402) | def get_tokenized_prompt(self, prompt): method init_attn_processors (line 406) | def init_attn_processors( method __call__ (line 452) | def __call__( class GaussianSmoothing (line 896) | class GaussianSmoothing(nn.Module): method __init__ (line 910) | def __init__(self, channels, kernel_size, sigma, dim=2): method forward (line 944) | def forward(self, input): function get_attention_scores (line 955) | def get_attention_scores( FILE: diffusers/examples/community/iadb.py class IADBScheduler (line 11) | class IADBScheduler(SchedulerMixin, ConfigMixin): method step (line 18) | def step( method set_timesteps (line 51) | def set_timesteps(self, num_inference_steps: int): method add_noise (line 54) | def add_noise( method __len__ (line 62) | def __len__(self): class IADBPipeline (line 66) | class IADBPipeline(DiffusionPipeline): method __init__ (line 78) | def __init__(self, unet, scheduler): method __call__ (line 84) | def __call__( FILE: diffusers/examples/community/imagic_stable_diffusion.py function preprocess (line 51) | def preprocess(image): class ImagicStableDiffusionPipeline (line 61) | class ImagicStableDiffusionPipeline(DiffusionPipeline, StableDiffusionMi... method __init__ (line 89) | def __init__( method train (line 110) | def train( method __call__ (line 311) | def __call__( FILE: diffusers/examples/community/img2img_inpainting.py function prepare_mask_and_masked_image (line 21) | def prepare_mask_and_masked_image(image, mask): function check_size (line 38) | def check_size(image, height, width): function overlay_inner_image (line 48) | def overlay_inner_image(image, inner_image, paste_offset: Tuple[int] = (... class ImageToImageInpaintingPipeline (line 58) | class ImageToImageInpaintingPipeline(DiffusionPipeline): method __init__ (line 86) | def __init__( method __call__ (line 133) | def __call__( FILE: diffusers/examples/community/instaflow_one_step.py function rescale_noise_cfg (line 41) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class InstaFlowPipeline (line 55) | class InstaFlowPipeline( method __init__ (line 95) | def __init__( method _encode_prompt (line 185) | def _encode_prompt( method encode_prompt (line 215) | def encode_prompt( method run_safety_checker (line 374) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 388) | def decode_latents(self, latents): method merge_dW_to_unet (line 399) | def merge_dW_to_unet(pipe, dW_dict, alpha=1.0): method prepare_extra_step_kwargs (line 406) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 423) | def check_inputs( method prepare_latents (line 470) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 493) | def __call__( FILE: diffusers/examples/community/interpolate_stable_diffusion.py function slerp (line 22) | def slerp(t, v0, v1, DOT_THRESHOLD=0.9995): class StableDiffusionWalkPipeline (line 49) | class StableDiffusionWalkPipeline(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 77) | def __init__( method __call__ (line 124) | def __call__( method embed_text (line 377) | def embed_text(self, text): method get_noise (line 390) | def get_noise(self, seed, dtype=torch.float32, height=512, width=512): method walk (line 399) | def walk( FILE: diffusers/examples/community/ip_adapter_face_id.py class IPAdapterFullImageProjection (line 55) | class IPAdapterFullImageProjection(nn.Module): method __init__ (line 56) | def __init__(self, image_embed_dim=1024, cross_attention_dim=1024, mul... method forward (line 65) | def forward(self, image_embeds: torch.Tensor): function rescale_noise_cfg (line 71) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 85) | def retrieve_timesteps( class IPAdapterFaceIDStableDiffusionPipeline (line 129) | class IPAdapterFaceIDStableDiffusionPipeline( method __init__ (line 175) | def __init__( method load_ip_adapter_face_id (line 267) | def load_ip_adapter_face_id(self, pretrained_model_name_or_path_or_dic... method convert_ip_adapter_image_proj_to_diffusers (line 306) | def convert_ip_adapter_image_proj_to_diffusers(self, state_dict): method _load_ip_adapter_weights (line 330) | def _load_ip_adapter_weights(self, state_dict): method set_ip_adapter_scale (line 442) | def set_ip_adapter_scale(self, scale): method _encode_prompt (line 448) | def _encode_prompt( method encode_prompt (line 480) | def encode_prompt( method run_safety_checker (line 661) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 675) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 686) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 703) | def check_inputs( method prepare_latents (line 755) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 778) | def get_guidance_scale_embedding(self, w, embedding_dim=512, dtype=tor... method guidance_scale (line 807) | def guidance_scale(self): method guidance_rescale (line 811) | def guidance_rescale(self): method clip_skip (line 815) | def clip_skip(self): method do_classifier_free_guidance (line 822) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 826) | def cross_attention_kwargs(self): method num_timesteps (line 830) | def num_timesteps(self): method interrupt (line 834) | def interrupt(self): method __call__ (line 838) | def __call__( FILE: diffusers/examples/community/kohya_hires_fix.py class UNet2DConditionModelHighResFix (line 35) | class UNet2DConditionModelHighResFix(UNet2DConditionModel): method __init__ (line 49) | def __init__(self, high_res_fix: List[Dict] = [{"timestep": 600, "scal... method _resize (line 55) | def _resize(cls, sample, target=None, scale_factor=1, mode="bicubic"): method forward (line 68) | def forward( method from_unet (line 361) | def from_unet(cls, unet: UNet2DConditionModel, high_res_fix: list): class StableDiffusionHighResFixPipeline (line 390) | class StableDiffusionHighResFixPipeline(StableDiffusionPipeline): method __init__ (line 430) | def __init__( FILE: diffusers/examples/community/latent_consistency_img2img.py class LatentConsistencyModelImg2ImgPipeline (line 39) | class LatentConsistencyModelImg2ImgPipeline(DiffusionPipeline): method __init__ (line 42) | def __init__( method _encode_prompt (line 75) | def _encode_prompt( method run_safety_checker (line 153) | def run_safety_checker(self, image, device, dtype): method prepare_latents (line 167) | def prepare_latents( method get_w_embedding (line 243) | def get_w_embedding(self, w, embedding_dim=512, dtype=torch.float32): method get_timesteps (line 266) | def get_timesteps(self, num_inference_steps, strength, device): method __call__ (line 276) | def __call__( class LCMSchedulerOutput (line 394) | class LCMSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 411) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 451) | def rescale_zero_terminal_snr(betas): class LCMSchedulerWithTimestamp (line 484) | class LCMSchedulerWithTimestamp(SchedulerMixin, ConfigMixin): method __init__ (line 539) | def __init__( method scale_model_input (line 590) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method _get_variance (line 605) | def _get_variance(self, timestep, prev_timestep): method _threshold_sample (line 616) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method set_timesteps (line 649) | def set_timesteps( method get_scalings_for_boundary_condition_discrete (line 678) | def get_scalings_for_boundary_condition_discrete(self, t): method step (line 686) | def step( method add_noise (line 778) | def add_noise( method get_velocity (line 802) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 820) | def __len__(self): FILE: diffusers/examples/community/latent_consistency_interpolate.py function lerp (line 76) | def lerp( function slerp (line 126) | def slerp( class LatentConsistencyModelWalkPipeline (line 192) | class LatentConsistencyModelWalkPipeline( method __init__ (line 234) | def __init__( method encode_prompt (line 277) | def encode_prompt( method run_safety_checker (line 459) | def run_safety_checker(self, image, device, dtype): method prepare_latents (line 474) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 496) | def get_guidance_scale_embedding(self, w, embedding_dim=512, dtype=tor... method prepare_extra_step_kwargs (line 525) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 543) | def check_inputs( method interpolate_embedding (line 581) | def interpolate_embedding( method interpolate_latent (line 612) | def interpolate_latent( method guidance_scale (line 638) | def guidance_scale(self): method cross_attention_kwargs (line 642) | def cross_attention_kwargs(self): method clip_skip (line 646) | def clip_skip(self): method num_timesteps (line 650) | def num_timesteps(self): method __call__ (line 655) | def __call__( FILE: diffusers/examples/community/latent_consistency_txt2img.py class LatentConsistencyModelPipeline (line 37) | class LatentConsistencyModelPipeline(DiffusionPipeline): method __init__ (line 40) | def __init__( method _encode_prompt (line 73) | def _encode_prompt( method run_safety_checker (line 151) | def run_safety_checker(self, image, device, dtype): method prepare_latents (line 165) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_w_embedding (line 180) | def get_w_embedding(self, w, embedding_dim=512, dtype=torch.float32): method __call__ (line 204) | def __call__( class LCMSchedulerOutput (line 309) | class LCMSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 326) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 366) | def rescale_zero_terminal_snr(betas): class LCMScheduler (line 399) | class LCMScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 451) | def __init__( method scale_model_input (line 502) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method _get_variance (line 517) | def _get_variance(self, timestep, prev_timestep): method _threshold_sample (line 528) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method set_timesteps (line 561) | def set_timesteps(self, num_inference_steps: int, lcm_origin_steps: in... method get_scalings_for_boundary_condition_discrete (line 586) | def get_scalings_for_boundary_condition_discrete(self, t): method step (line 594) | def step( method add_noise (line 686) | def add_noise( method get_velocity (line 710) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 728) | def __len__(self): FILE: diffusers/examples/community/llm_grounded_diffusion.py function convert_attn_keys (line 123) | def convert_attn_keys(key): function scale_proportion (line 136) | def scale_proportion(obj_box, H, W): class AttnProcessorWithHook (line 149) | class AttnProcessorWithHook(AttnProcessor2_0): method __init__ (line 150) | def __init__( method __call__ (line 167) | def __call__( class LLMGroundedDiffusionPipeline (line 270) | class LLMGroundedDiffusionPipeline( method __init__ (line 319) | def __init__( method attn_hook (line 416) | def attn_hook(self, name, query, key, value, attention_probs): method convert_box (line 421) | def convert_box(cls, box, height, width): method _parse_response_with_negative (line 431) | def _parse_response_with_negative(cls, text): method parse_llm_response (line 469) | def parse_llm_response(cls, response, canvas_height=512, canvas_width=... method check_inputs (line 480) | def check_inputs( method register_attn_hooks (line 538) | def register_attn_hooks(self, unet): method enable_fuser (line 573) | def enable_fuser(self, enabled=True): method enable_attn_hook (line 578) | def enable_attn_hook(self, enabled=True): method get_token_map (line 583) | def get_token_map(self, prompt, padding="do_not_pad", verbose=False): method get_phrase_indices (line 599) | def get_phrase_indices( method add_ca_loss_per_attn_map_to_loss (line 648) | def add_ca_loss_per_attn_map_to_loss( method compute_ca_loss (line 701) | def compute_ca_loss( method __call__ (line 745) | def __call__( method latent_lmd_guidance (line 1082) | def latent_lmd_guidance( method _encode_prompt (line 1190) | def _encode_prompt( method encode_prompt (line 1223) | def encode_prompt( method encode_image (line 1407) | def encode_image(self, image, device, num_images_per_prompt): method run_safety_checker (line 1421) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 1436) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 1448) | def prepare_extra_step_kwargs(self, generator, eta): method prepare_latents (line 1466) | def prepare_latents( method get_guidance_scale_embedding (line 1499) | def get_guidance_scale_embedding(self, w, embedding_dim=512, dtype=tor... method guidance_scale (line 1529) | def guidance_scale(self): method guidance_rescale (line 1534) | def guidance_rescale(self): method clip_skip (line 1539) | def clip_skip(self): method do_classifier_free_guidance (line 1547) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 1552) | def cross_attention_kwargs(self): method num_timesteps (line 1557) | def num_timesteps(self): FILE: diffusers/examples/community/lpw_stable_diffusion.py function parse_prompt_attention (line 51) | def parse_prompt_attention(text): function get_prompts_with_weights (line 137) | def get_prompts_with_weights(pipe: DiffusionPipeline, prompt: List[str],... function pad_tokens_and_weights (line 172) | def pad_tokens_and_weights(tokens, weights, max_length, bos, eos, pad, n... function get_unweighted_text_embeddings (line 197) | def get_unweighted_text_embeddings( function get_weighted_text_embeddings (line 237) | def get_weighted_text_embeddings( function preprocess_image (line 370) | def preprocess_image(image, batch_size): function preprocess_mask (line 380) | def preprocess_mask(mask, batch_size, scale_factor=8): class StableDiffusionLongPromptWeightingPipeline (line 411) | class StableDiffusionLongPromptWeightingPipeline( method __init__ (line 446) | def __init__( method _encode_prompt (line 538) | def _encode_prompt( method check_inputs (line 615) | def check_inputs( method get_timesteps (line 666) | def get_timesteps(self, num_inference_steps, strength, device, is_text... method run_safety_checker (line 678) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 688) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 696) | def prepare_extra_step_kwargs(self, generator, eta): method prepare_latents (line 713) | def prepare_latents( method __call__ (line 766) | def __call__( method text2img (line 1024) | def text2img( method img2img (line 1138) | def img2img( method inpaint (line 1250) | def inpaint( FILE: diffusers/examples/community/lpw_stable_diffusion_onnx.py function parse_prompt_attention (line 78) | def parse_prompt_attention(text): function get_prompts_with_weights (line 164) | def get_prompts_with_weights(pipe, prompt: List[str], max_length: int): function pad_tokens_and_weights (line 199) | def pad_tokens_and_weights(tokens, weights, max_length, bos, eos, pad, n... function get_unweighted_text_embeddings (line 224) | def get_unweighted_text_embeddings( function get_weighted_text_embeddings (line 265) | def get_weighted_text_embeddings( function preprocess_image (line 407) | def preprocess_image(image): function preprocess_mask (line 416) | def preprocess_mask(mask, scale_factor=8): class OnnxStableDiffusionLongPromptWeightingPipeline (line 428) | class OnnxStableDiffusionLongPromptWeightingPipeline(OnnxStableDiffusion... method __init__ (line 439) | def __init__( method __init__ (line 466) | def __init__( method __init__additional__ (line 489) | def __init__additional__(self): method _encode_prompt (line 493) | def _encode_prompt( method check_inputs (line 544) | def check_inputs(self, prompt, height, width, strength, callback_steps): method get_timesteps (line 562) | def get_timesteps(self, num_inference_steps, strength, is_text2img): method run_safety_checker (line 575) | def run_safety_checker(self, image): method decode_latents (line 593) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 604) | def prepare_extra_step_kwargs(self, generator, eta): method prepare_latents (line 621) | def prepare_latents(self, image, timestep, batch_size, height, width, ... method __call__ (line 654) | def __call__( method text2img (line 870) | def text2img( method img2img (line 962) | def img2img( method inpaint (line 1053) | def inpaint( FILE: diffusers/examples/community/lpw_stable_diffusion_xl.py function parse_prompt_attention (line 51) | def parse_prompt_attention(text): function get_prompts_tokens_with_weights (line 153) | def get_prompts_tokens_with_weights(clip_tokenizer: CLIPTokenizer, promp... function group_tokens_and_weights (line 203) | def group_tokens_and_weights(token_ids: list, weights: list, pad_last_bl... function get_weighted_text_embeddings_sdxl (line 255) | def get_weighted_text_embeddings_sdxl( function rescale_noise_cfg (line 474) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_latents (line 489) | def retrieve_latents( function retrieve_timesteps (line 503) | def retrieve_timesteps( class SDXLLongPromptWeightingPipeline (line 547) | class SDXLLongPromptWeightingPipeline( method __init__ (line 615) | def __init__( method enable_model_cpu_offload (line 657) | def enable_model_cpu_offload(self, gpu_id=0): method encode_prompt (line 688) | def encode_prompt( method encode_image (line 880) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_extra_step_kwargs (line 905) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 922) | def check_inputs( method get_timesteps (line 1005) | def get_timesteps(self, num_inference_steps, strength, device, denoisi... method prepare_latents (line 1041) | def prepare_latents( method _encode_vae_image (line 1197) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method prepare_mask_latents (line 1220) | def prepare_mask_latents( method _get_add_time_ids (line 1273) | def _get_add_time_ids(self, original_size, crops_coords_top_left, targ... method upcast_vae (line 1290) | def upcast_vae(self): method get_guidance_scale_embedding (line 1310) | def get_guidance_scale_embedding(self, w, embedding_dim=512, dtype=tor... method guidance_scale (line 1339) | def guidance_scale(self): method guidance_rescale (line 1343) | def guidance_rescale(self): method clip_skip (line 1347) | def clip_skip(self): method do_classifier_free_guidance (line 1354) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 1358) | def cross_attention_kwargs(self): method denoising_end (line 1362) | def denoising_end(self): method denoising_start (line 1366) | def denoising_start(self): method num_timesteps (line 1370) | def num_timesteps(self): method __call__ (line 1375) | def __call__( method text2img (line 1913) | def text2img( method img2img (line 1986) | def img2img( method inpaint (line 2063) | def inpaint( method load_lora_weights (line 2145) | def load_lora_weights(self, pretrained_model_name_or_path_or_dict: Uni... method save_lora_weights (line 2177) | def save_lora_weights( method _remove_text_encoder_monkey_patch (line 2210) | def _remove_text_encoder_monkey_patch(self): FILE: diffusers/examples/community/magic_mix.py class MagicMixPipeline (line 19) | class MagicMixPipeline(DiffusionPipeline): method __init__ (line 20) | def __init__( method encode (line 33) | def encode(self, img): method decode (line 40) | def decode(self, latent): method prep_text (line 50) | def prep_text(self, prompt): method __call__ (line 73) | def __call__( FILE: diffusers/examples/community/marigold_depth_estimation.py class MarigoldDepthOutput (line 49) | class MarigoldDepthOutput(BaseOutput): function get_pil_resample_method (line 67) | def get_pil_resample_method(method_str: str) -> Resampling: class MarigoldPipeline (line 80) | class MarigoldPipeline(DiffusionPipeline): method __init__ (line 104) | def __init__( method __call__ (line 125) | def __call__( method _check_inference_step (line 283) | def _check_inference_step(self, n_step: int): method _encode_empty_text (line 302) | def _encode_empty_text(self): method single_infer (line 318) | def single_infer( method encode_rgb (line 394) | def encode_rgb(self, rgb_in: torch.Tensor) -> torch.Tensor: method decode_depth (line 413) | def decode_depth(self, depth_latent: torch.Tensor) -> torch.Tensor: method resize_max_res (line 434) | def resize_max_res(img: Image.Image, max_edge_resolution: int, resampl... method colorize_depth_maps (line 459) | def colorize_depth_maps(depth_map, min_depth, max_depth, cmap="Spectra... method chw2hwc (line 498) | def chw2hwc(chw): method _find_batch_size (line 507) | def _find_batch_size(ensemble_size: int, input_res: int, dtype: torch.... method ensemble_depths (line 565) | def ensemble_depths( FILE: diffusers/examples/community/masked_stable_diffusion_img2img.py class MaskedStableDiffusionImg2ImgPipeline (line 11) | class MaskedStableDiffusionImg2ImgPipeline(StableDiffusionImg2ImgPipeline): method __call__ (line 15) | def __call__( method _make_latent_mask (line 239) | def _make_latent_mask(self, latents, mask): FILE: diffusers/examples/community/mixture_canvas.py function preprocess_image (line 20) | def preprocess_image(image): class CanvasRegion (line 38) | class CanvasRegion: method __post_init__ (line 48) | def __post_init__(self): method width (line 74) | def width(self): method height (line 78) | def height(self): method get_region_generator (line 81) | def get_region_generator(self, device="cpu"): method __dict__ (line 87) | def __dict__(self): class MaskModes (line 91) | class MaskModes(Enum): class DiffusionRegion (line 100) | class DiffusionRegion(CanvasRegion): class Text2ImageRegion (line 107) | class Text2ImageRegion(DiffusionRegion): method __post_init__ (line 117) | def __post_init__(self): method tokenize_prompt (line 135) | def tokenize_prompt(self, tokenizer): method encode_prompt (line 145) | def encode_prompt(self, text_encoder, device): class Image2ImageRegion (line 154) | class Image2ImageRegion(DiffusionRegion): method __post_init__ (line 160) | def __post_init__(self): method encode_reference_image (line 169) | def encode_reference_image(self, encoder, device, generator, cpu_vae=F... method __dict__ (line 182) | def __dict__(self): class RerollModes (line 191) | class RerollModes(Enum): class RerollRegion (line 199) | class RerollRegion(CanvasRegion): class MaskWeightsBuilder (line 206) | class MaskWeightsBuilder: method compute_mask_weights (line 212) | def compute_mask_weights(self, region: DiffusionRegion) -> torch.tensor: method _constant_weights (line 221) | def _constant_weights(self, region: DiffusionRegion) -> torch.tensor: method _gaussian_weights (line 227) | def _gaussian_weights(self, region: DiffusionRegion) -> torch.tensor: method _quartic_weights (line 247) | def _quartic_weights(self, region: DiffusionRegion) -> torch.tensor: class StableDiffusionCanvasPipeline (line 267) | class StableDiffusionCanvasPipeline(DiffusionPipeline, StableDiffusionMi... method __init__ (line 270) | def __init__( method decode_latents (line 291) | def decode_latents(self, latents, cpu_vae=False): method get_latest_timestep_img2img (line 309) | def get_latest_timestep_img2img(self, num_inference_steps, strength): method __call__ (line 322) | def __call__( FILE: diffusers/examples/community/mixture_tiling.py function _tile2pixel_indices (line 51) | def _tile2pixel_indices(tile_row, tile_col, tile_width, tile_height, til... function _pixel2latent_indices (line 67) | def _pixel2latent_indices(px_row_init, px_row_end, px_col_init, px_col_e... function _tile2latent_indices (line 72) | def _tile2latent_indices(tile_row, tile_col, tile_width, tile_height, ti... function _tile2latent_exclusive_indices (line 87) | def _tile2latent_exclusive_indices( class StableDiffusionExtrasMixin (line 116) | class StableDiffusionExtrasMixin: method decode_latents (line 119) | def decode_latents(self, latents, cpu_vae=False): class StableDiffusionTilingPipeline (line 138) | class StableDiffusionTilingPipeline(DiffusionPipeline, StableDiffusionEx... method __init__ (line 139) | def __init__( class SeedTilesMode (line 160) | class SeedTilesMode(Enum): method __call__ (line 167) | def __call__( method _gaussian_weights (line 384) | def _gaussian_weights(self, tile_width, tile_height, nbatches): FILE: diffusers/examples/community/multilingual_stable_diffusion.py function detect_language (line 26) | def detect_language(pipe, prompt, batch_size): function translate_prompt (line 41) | def translate_prompt(prompt, translation_tokenizer, translation_model, d... class MultilingualStableDiffusion (line 51) | class MultilingualStableDiffusion(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 86) | def __init__( method __call__ (line 139) | def __call__( FILE: diffusers/examples/community/one_step_unet.py class UnetSchedulerOneForwardPipeline (line 7) | class UnetSchedulerOneForwardPipeline(DiffusionPipeline): method __init__ (line 8) | def __init__(self, unet, scheduler): method __call__ (line 13) | def __call__(self): FILE: diffusers/examples/community/pipeline_animatediff_controlnet.py function tensor2vid (line 95) | def tensor2vid(video: torch.Tensor, processor, output_type="np"): class AnimateDiffControlNetPipeline (line 116) | class AnimateDiffControlNetPipeline( method __init__ (line 151) | def __init__( method encode_prompt (line 194) | def encode_prompt( method encode_image (line 376) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 401) | def prepare_ip_adapter_image_embeds( method decode_latents (line 436) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 460) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 478) | def check_inputs( method check_image (line 625) | def check_image(self, image, prompt, prompt_embeds): method prepare_latents (line 663) | def prepare_latents( method prepare_image (line 689) | def prepare_image( method guidance_scale (line 720) | def guidance_scale(self): method clip_skip (line 724) | def clip_skip(self): method do_classifier_free_guidance (line 731) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 735) | def cross_attention_kwargs(self): method num_timesteps (line 739) | def num_timesteps(self): method __call__ (line 743) | def __call__( FILE: diffusers/examples/community/pipeline_animatediff_img2video.py function lerp (line 70) | def lerp( function slerp (line 107) | def slerp( function tensor2vid (line 162) | def tensor2vid(video: torch.Tensor, processor, output_type="np"): function retrieve_latents (line 184) | def retrieve_latents( function retrieve_timesteps (line 198) | def retrieve_timesteps( class AnimateDiffImgToVideoPipeline (line 242) | class AnimateDiffImgToVideoPipeline( method __init__ (line 276) | def __init__( method encode_prompt (line 311) | def encode_prompt( method encode_image (line 493) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 518) | def prepare_ip_adapter_image_embeds( method decode_latents (line 553) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 577) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 594) | def check_inputs( method prepare_latents (line 655) | def prepare_latents( method __call__ (line 734) | def __call__( FILE: diffusers/examples/community/pipeline_demofusion_sdxl.py function gaussian_kernel (line 63) | def gaussian_kernel(kernel_size=3, sigma=1.0, channels=3): function gaussian_filter (line 73) | def gaussian_filter(latents, kernel_size=3, sigma=1.0): function rescale_noise_cfg (line 82) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class DemoFusionSDXLPipeline (line 96) | class DemoFusionSDXLPipeline( method __init__ (line 146) | def __init__( method encode_prompt (line 181) | def encode_prompt( method prepare_extra_step_kwargs (line 377) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 394) | def check_inputs( method prepare_latents (line 479) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 501) | def _get_add_time_ids(self, original_size, crops_coords_top_left, targ... method get_views (line 517) | def get_views(self, height, width, window_size=128, stride=64, random_... method tiled_decode (line 567) | def tiled_decode(self, latents, current_height, current_width): method upcast_vae (line 610) | def upcast_vae(self): method __call__ (line 631) | def __call__( method load_lora_weights (line 1289) | def load_lora_weights(self, pretrained_model_name_or_path_or_dict: Uni... method save_lora_weights (line 1351) | def save_lora_weights( method _remove_text_encoder_monkey_patch (line 1390) | def _remove_text_encoder_monkey_patch(self): FILE: diffusers/examples/community/pipeline_fabric.py class FabricCrossAttnProcessor (line 57) | class FabricCrossAttnProcessor: method __init__ (line 58) | def __init__(self): method __call__ (line 61) | def __call__( class FabricPipeline (line 118) | class FabricPipeline(DiffusionPipeline): method __init__ (line 142) | def __init__( method _encode_prompt (line 186) | def _encode_prompt( method get_unet_hidden_states (line 347) | def get_unet_hidden_states(self, z_all, t, prompt_embd): method unet_forward_with_cached_hidden_states (line 370) | def unet_forward_with_cached_hidden_states( method preprocess_feedback_images (line 454) | def preprocess_feedback_images(self, images, vae, dim, device, dtype, ... method check_inputs (line 461) | def check_inputs( method __call__ (line 494) | def __call__( method image_to_tensor (line 742) | def image_to_tensor(self, image: Union[str, Image.Image], dim: tuple, ... FILE: diffusers/examples/community/pipeline_null_text_inversion.py function retrieve_timesteps (line 15) | def retrieve_timesteps( class NullTextPipeline (line 58) | class NullTextPipeline(StableDiffusionPipeline): method get_noise_pred (line 59) | def get_noise_pred(self, latents, t, context): method get_noise_pred_single (line 68) | def get_noise_pred_single(self, latents, t, context): method image2latent (line 73) | def image2latent(self, image_path): method latent2image (line 83) | def latent2image(self, latents): method prev_step (line 89) | def prev_step(self, model_output, timestep, sample): method next_step (line 101) | def next_step(self, model_output, timestep, sample): method null_optimization (line 114) | def null_optimization(self, latents, context, num_inner_steps, epsilon): method ddim_inversion_loop (line 149) | def ddim_inversion_loop(self, latent, context): method get_context (line 162) | def get_context(self, prompt): method invert (line 178) | def invert( method __call__ (line 194) | def __call__( FILE: diffusers/examples/community/pipeline_prompt2prompt.py function rescale_noise_cfg (line 61) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class Prompt2PromptPipeline (line 75) | class Prompt2PromptPipeline( method __init__ (line 120) | def __init__( method _encode_prompt (line 213) | def _encode_prompt( method encode_prompt (line 246) | def encode_prompt( method run_safety_checker (line 430) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 445) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 463) | def check_inputs( method prepare_latents (line 523) | def prepare_latents( method __call__ (line 556) | def __call__( method register_attention_control (line 774) | def register_attention_control(self, controller): class P2PCrossAttnProcessor (line 799) | class P2PCrossAttnProcessor: method __init__ (line 800) | def __init__(self, controller, place_in_unet): method __call__ (line 805) | def __call__( function create_controller (line 842) | def create_controller( class AttentionControl (line 926) | class AttentionControl(abc.ABC): method step_callback (line 927) | def step_callback(self, x_t): method between_steps (line 930) | def between_steps(self): method num_uncond_att_layers (line 934) | def num_uncond_att_layers(self): method forward (line 938) | def forward(self, attn, is_cross: bool, place_in_unet: str): method __call__ (line 941) | def __call__(self, attn, is_cross: bool, place_in_unet: str): method reset (line 952) | def reset(self): method __init__ (line 956) | def __init__(self): class EmptyControl (line 962) | class EmptyControl(AttentionControl): method forward (line 963) | def forward(self, attn, is_cross: bool, place_in_unet: str): class AttentionStore (line 967) | class AttentionStore(AttentionControl): method get_empty_store (line 969) | def get_empty_store(): method forward (line 979) | def forward(self, attn, is_cross: bool, place_in_unet: str): method between_steps (line 985) | def between_steps(self): method get_average_attention (line 994) | def get_average_attention(self): method reset (line 1000) | def reset(self): method __init__ (line 1005) | def __init__(self): class LocalBlend (line 1011) | class LocalBlend: method __call__ (line 1012) | def __call__(self, x_t, attention_store): method __init__ (line 1026) | def __init__( class AttentionControlEdit (line 1048) | class AttentionControlEdit(AttentionStore, abc.ABC): method step_callback (line 1049) | def step_callback(self, x_t): method replace_self_attention (line 1054) | def replace_self_attention(self, attn_base, att_replace): method replace_cross_attention (line 1061) | def replace_cross_attention(self, attn_base, att_replace): method forward (line 1064) | def forward(self, attn, is_cross: bool, place_in_unet: str): method __init__ (line 1083) | def __init__( class AttentionReplace (line 1109) | class AttentionReplace(AttentionControlEdit): method replace_cross_attention (line 1110) | def replace_cross_attention(self, attn_base, att_replace): method __init__ (line 1113) | def __init__( class AttentionRefine (line 1135) | class AttentionRefine(AttentionControlEdit): method replace_cross_attention (line 1136) | def replace_cross_attention(self, attn_base, att_replace): method __init__ (line 1141) | def __init__( class AttentionReweight (line 1165) | class AttentionReweight(AttentionControlEdit): method replace_cross_attention (line 1166) | def replace_cross_attention(self, attn_base, att_replace): method __init__ (line 1172) | def __init__( function update_alpha_time_word (line 1198) | def update_alpha_time_word( function get_time_words_attention_alpha (line 1215) | def get_time_words_attention_alpha( function get_word_inds (line 1240) | def get_word_inds(text: str, word_place: int, tokenizer): function get_replacement_mapper_ (line 1262) | def get_replacement_mapper_(x: str, y: str, tokenizer, max_len=77): function get_replacement_mapper (line 1300) | def get_replacement_mapper(prompts, tokenizer, max_len=77): function get_equalizer (line 1310) | def get_equalizer( class ScoreParams (line 1327) | class ScoreParams: method __init__ (line 1328) | def __init__(self, gap, match, mismatch): method mis_match_char (line 1333) | def mis_match_char(self, x, y): function get_matrix (line 1340) | def get_matrix(size_x, size_y, gap): function get_traceback_matrix (line 1347) | def get_traceback_matrix(size_x, size_y): function global_align (line 1355) | def global_align(x, y, score): function get_aligned_sequences (line 1373) | def get_aligned_sequences(x, y, trace_back): function get_mapper (line 1401) | def get_mapper(x: str, y: str, tokenizer, max_len=77): function get_refinement_mapper (line 1415) | def get_refinement_mapper(prompts, tokenizer, max_len=77): FILE: diffusers/examples/community/pipeline_sdxl_style_aligned.py function expand_first (line 138) | def expand_first(feat: torch.Tensor, scale: float = 1.0) -> torch.Tensor: function concat_first (line 149) | def concat_first(feat: torch.Tensor, dim: int = 2, scale: float = 1.0) -... function calc_mean_std (line 154) | def calc_mean_std(feat: torch.Tensor, eps: float = 1e-5) -> Tuple[torch.... function adain (line 160) | def adain(feat: torch.Tensor) -> torch.Tensor: function get_switch_vec (line 169) | def get_switch_vec(total_num_layers, level): class SharedAttentionProcessor (line 186) | class SharedAttentionProcessor(AttnProcessor2_0): method __init__ (line 187) | def __init__( method shifted_scaled_dot_product_attention (line 207) | def shifted_scaled_dot_product_attention( method shared_call (line 215) | def shared_call( method __call__ (line 288) | def __call__( function rescale_noise_cfg (line 317) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 332) | def retrieve_timesteps( function retrieve_latents (line 377) | def retrieve_latents( class StyleAlignedSDXLPipeline (line 390) | class StyleAlignedSDXLPipeline( method __init__ (line 465) | def __init__( method encode_prompt (line 508) | def encode_prompt( method encode_image (line 743) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_extra_step_kwargs (line 768) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 785) | def check_inputs( method get_timesteps (line 864) | def get_timesteps(self, num_inference_steps, strength, device, denoisi... method prepare_latents (line 900) | def prepare_latents( method prepare_mask_latents (line 1056) | def prepare_mask_latents( method _encode_vae_image (line 1109) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method _get_add_time_ids (line 1132) | def _get_add_time_ids(self, original_size, crops_coords_top_left, targ... method upcast_vae (line 1148) | def upcast_vae(self): method _enable_shared_attention_processors (line 1168) | def _enable_shared_attention_processors( method _disable_shared_attention_processors (line 1205) | def _disable_shared_attention_processors(self): method _register_shared_norm (line 1217) | def _register_shared_norm(self, share_group_norm: bool = True, share_l... method style_aligned_enabled (line 1254) | def style_aligned_enabled(self): method enable_style_aligned (line 1258) | def enable_style_aligned( method disable_style_aligned (line 1305) | def disable_style_aligned(self): method get_guidance_scale_embedding (line 1315) | def get_guidance_scale_embedding(self, w, embedding_dim=512, dtype=tor... method guidance_scale (line 1344) | def guidance_scale(self): method guidance_rescale (line 1348) | def guidance_rescale(self): method clip_skip (line 1352) | def clip_skip(self): method do_classifier_free_guidance (line 1359) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 1363) | def cross_attention_kwargs(self): method denoising_end (line 1367) | def denoising_end(self): method denoising_start (line 1371) | def denoising_start(self): method num_timesteps (line 1375) | def num_timesteps(self): method interrupt (line 1379) | def interrupt(self): method __call__ (line 1384) | def __call__( FILE: diffusers/examples/community/pipeline_stable_diffusion_boxdiff.py class GaussianSmoothing (line 66) | class GaussianSmoothing(nn.Module): method __init__ (line 81) | def __init__(self, channels, kernel_size, sigma, dim=2): method forward (line 115) | def forward(self, input): class AttendExciteCrossAttnProcessor (line 126) | class AttendExciteCrossAttnProcessor: method __init__ (line 127) | def __init__(self, attnstore, place_in_unet): method __call__ (line 132) | def __call__( class AttentionControl (line 166) | class AttentionControl(abc.ABC): method step_callback (line 167) | def step_callback(self, x_t): method between_steps (line 170) | def between_steps(self): method forward (line 178) | def forward(self, attn, is_cross: bool, place_in_unet: str): method __call__ (line 181) | def __call__(self, attn, is_cross: bool, place_in_unet: str): method reset (line 190) | def reset(self): method __init__ (line 194) | def __init__(self): class AttentionStore (line 200) | class AttentionStore(AttentionControl): method get_empty_store (line 202) | def get_empty_store(): method forward (line 205) | def forward(self, attn, is_cross: bool, place_in_unet: str): method between_steps (line 211) | def between_steps(self): method get_average_attention (line 224) | def get_average_attention(self): method get_average_global_attention (line 228) | def get_average_global_attention(self): method reset (line 234) | def reset(self): method __init__ (line 240) | def __init__(self, save_global_store=False): function aggregate_attention (line 254) | def aggregate_attention( function register_attention_control (line 277) | def register_attention_control(model, controller): function rescale_noise_cfg (line 302) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 316) | def retrieve_timesteps( class StableDiffusionBoxDiffPipeline (line 360) | class StableDiffusionBoxDiffPipeline( method __init__ (line 401) | def __init__( method enable_vae_slicing (line 493) | def enable_vae_slicing(self): method disable_vae_slicing (line 500) | def disable_vae_slicing(self): method enable_vae_tiling (line 507) | def enable_vae_tiling(self): method disable_vae_tiling (line 515) | def disable_vae_tiling(self): method _encode_prompt (line 522) | def _encode_prompt( method encode_prompt (line 554) | def encode_prompt( method encode_image (line 735) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method run_safety_checker (line 759) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 773) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 784) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 801) | def check_inputs( method prepare_latents (line 869) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method enable_freeu (line 886) | def enable_freeu(self, s1: float, s2: float, b1: float, b2: float): method disable_freeu (line 908) | def disable_freeu(self): method fuse_qkv_projections (line 913) | def fuse_qkv_projections(self, unet: bool = True, vae: bool = True): method unfuse_qkv_projections (line 945) | def unfuse_qkv_projections(self, unet: bool = True, vae: bool = True): method get_guidance_scale_embedding (line 974) | def get_guidance_scale_embedding(self, w, embedding_dim=512, dtype=tor... method guidance_scale (line 1003) | def guidance_scale(self): method guidance_rescale (line 1007) | def guidance_rescale(self): method clip_skip (line 1011) | def clip_skip(self): method do_classifier_free_guidance (line 1018) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 1022) | def cross_attention_kwargs(self): method num_timesteps (line 1026) | def num_timesteps(self): method interrupt (line 1030) | def interrupt(self): method _compute_max_attention_per_index (line 1033) | def _compute_max_attention_per_index( method _aggregate_and_get_max_attention_per_token (line 1116) | def _aggregate_and_get_max_attention_per_token( method _compute_loss (line 1151) | def _compute_loss( method _update_latent (line 1168) | def _update_latent(latents: torch.Tensor, loss: torch.Tensor, step_siz... method _perform_iterative_refinement_step (line 1174) | def _perform_iterative_refinement_step( method __call__ (line 1285) | def __call__( FILE: diffusers/examples/community/pipeline_stable_diffusion_pag.py class PAGIdentitySelfAttnProcessor (line 48) | class PAGIdentitySelfAttnProcessor: method __init__ (line 53) | def __init__(self): method __call__ (line 57) | def __call__( class PAGCFGIdentitySelfAttnProcessor (line 162) | class PAGCFGIdentitySelfAttnProcessor: method __init__ (line 167) | def __init__(self): method __call__ (line 171) | def __call__( function rescale_noise_cfg (line 274) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 288) | def retrieve_timesteps( class StableDiffusionPAGPipeline (line 330) | class StableDiffusionPAGPipeline( method __init__ (line 368) | def __init__( method enable_vae_slicing (line 460) | def enable_vae_slicing(self): method disable_vae_slicing (line 467) | def disable_vae_slicing(self): method enable_vae_tiling (line 474) | def enable_vae_tiling(self): method disable_vae_tiling (line 482) | def disable_vae_tiling(self): method _encode_prompt (line 489) | def _encode_prompt( method encode_prompt (line 521) | def encode_prompt( method encode_image (line 701) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 725) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 759) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 773) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 784) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 801) | def check_inputs( method prepare_latents (line 860) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method enable_freeu (line 882) | def enable_freeu(self, s1: float, s2: float, b1: float, b2: float): method disable_freeu (line 901) | def disable_freeu(self): method fuse_qkv_projections (line 906) | def fuse_qkv_projections(self, unet: bool = True, vae: bool = True): method unfuse_qkv_projections (line 934) | def unfuse_qkv_projections(self, unet: bool = True, vae: bool = True): method get_guidance_scale_embedding (line 958) | def get_guidance_scale_embedding(self, w, embedding_dim=512, dtype=tor... method pred_z0 (line 984) | def pred_z0(self, sample, model_output, timestep): method pred_x0 (line 1004) | def pred_x0(self, latents, noise_pred, t, generator, device, prompt_em... method guidance_scale (line 1014) | def guidance_scale(self): method guidance_rescale (line 1018) | def guidance_rescale(self): method clip_skip (line 1022) | def clip_skip(self): method do_classifier_free_guidance (line 1029) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 1033) | def cross_attention_kwargs(self): method num_timesteps (line 1037) | def num_timesteps(self): method interrupt (line 1041) | def interrupt(self): method pag_scale (line 1045) | def pag_scale(self): method do_perturbed_attention_guidance (line 1049) | def do_perturbed_attention_guidance(self): method pag_adaptive_scaling (line 1053) | def pag_adaptive_scaling(self): method do_pag_adaptive_scaling (line 1057) | def do_pag_adaptive_scaling(self): method pag_applied_layers_index (line 1061) | def pag_applied_layers_index(self): method __call__ (line 1066) | def __call__( FILE: diffusers/examples/community/pipeline_stable_diffusion_upscale_ldm3d.py class StableDiffusionUpscaleLDM3DPipeline (line 71) | class StableDiffusionUpscaleLDM3DPipeline( method __init__ (line 111) | def __init__( method _encode_prompt (line 160) | def _encode_prompt( method encode_prompt (line 193) | def encode_prompt( method run_safety_checker (line 374) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 390) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 407) | def check_inputs( method prepare_latents (line 493) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 526) | def __call__( FILE: diffusers/examples/community/pipeline_stable_diffusion_xl_controlnet_adapter.py function _preprocess_adapter_image (line 118) | def _preprocess_adapter_image(image, height, width): function rescale_noise_cfg (line 146) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class StableDiffusionXLControlNetAdapterPipeline (line 160) | class StableDiffusionXLControlNetAdapterPipeline( method __init__ (line 204) | def __init__( method encode_prompt (line 242) | def encode_prompt( method prepare_extra_step_kwargs (line 477) | def prepare_extra_step_kwargs(self, generator, eta): method check_image (line 495) | def check_image(self, image, prompt, prompt_embeds): method check_inputs (line 533) | def check_inputs( method check_conditions (line 612) | def check_conditions( method prepare_latents (line 753) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 776) | def _get_add_time_ids( method upcast_vae (line 795) | def upcast_vae(self): method _default_height_width (line 815) | def _default_height_width(self, height, width, image): method prepare_control_image (line 842) | def prepare_control_image( method __call__ (line 874) | def __call__( FILE: diffusers/examples/community/pipeline_stable_diffusion_xl_controlnet_adapter_inpaint.py function _preprocess_adapter_image (line 137) | def _preprocess_adapter_image(image, height, width): function mask_pil_to_torch (line 164) | def mask_pil_to_torch(mask, height, width): function prepare_mask_and_masked_image (line 180) | def prepare_mask_and_masked_image(image, mask, height, width, return_ima... function rescale_noise_cfg (line 293) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class StableDiffusionXLControlNetAdapterInpaintPipeline (line 307) | class StableDiffusionXLControlNetAdapterInpaintPipeline( method __init__ (line 350) | def __init__( method encode_prompt (line 390) | def encode_prompt( method prepare_extra_step_kwargs (line 625) | def prepare_extra_step_kwargs(self, generator, eta): method check_image (line 643) | def check_image(self, image, prompt, prompt_embeds): method check_inputs (line 681) | def check_inputs( method check_conditions (line 760) | def check_conditions( method prepare_latents (line 900) | def prepare_latents( method _encode_vae_image (line 966) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method prepare_mask_latents (line 989) | def prepare_mask_latents( method get_timesteps (line 1050) | def get_timesteps(self, num_inference_steps, strength, device, denoisi... method _get_add_time_ids (line 1086) | def _get_add_time_ids( method upcast_vae (line 1133) | def upcast_vae(self): method _default_height_width (line 1153) | def _default_height_width(self, height, width, image): method prepare_control_image (line 1180) | def prepare_control_image( method __call__ (line 1212) | def __call__( FILE: diffusers/examples/community/pipeline_stable_diffusion_xl_differential_img2img.py function rescale_noise_cfg (line 94) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_latents (line 109) | def retrieve_latents( function retrieve_timesteps (line 123) | def retrieve_timesteps( class StableDiffusionXLDifferentialImg2ImgPipeline (line 167) | class StableDiffusionXLDifferentialImg2ImgPipeline( method __init__ (line 233) | def __init__( method encode_prompt (line 274) | def encode_prompt( method prepare_extra_step_kwargs (line 509) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 526) | def check_inputs( method get_timesteps (line 616) | def get_timesteps(self, num_inference_steps, strength, device, denoisi... method prepare_latents (line 652) | def prepare_latents( method encode_image (line 721) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 746) | def prepare_ip_adapter_image_embeds( method _get_add_time_ids (line 797) | def _get_add_time_ids( method upcast_vae (line 849) | def upcast_vae(self): method get_guidance_scale_embedding (line 869) | def get_guidance_scale_embedding( method guidance_scale (line 900) | def guidance_scale(self): method guidance_rescale (line 904) | def guidance_rescale(self): method clip_skip (line 908) | def clip_skip(self): method do_classifier_free_guidance (line 915) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 919) | def cross_attention_kwargs(self): method denoising_end (line 923) | def denoising_end(self): method denoising_start (line 927) | def denoising_start(self): method num_timesteps (line 931) | def num_timesteps(self): method interrupt (line 935) | def interrupt(self): method __call__ (line 940) | def __call__( FILE: diffusers/examples/community/pipeline_stable_diffusion_xl_instandid_img2img.py function FeedForward (line 55) | def FeedForward(dim, mult=4): function reshape_tensor (line 65) | def reshape_tensor(x, heads): class PerceiverAttention (line 76) | class PerceiverAttention(nn.Module): method __init__ (line 77) | def __init__(self, *, dim, dim_head=64, heads=8): method forward (line 91) | def forward(self, x, latents): class Resampler (line 123) | class Resampler(nn.Module): method __init__ (line 124) | def __init__( method forward (line 155) | def forward(self, x): class AttnProcessor (line 167) | class AttnProcessor(nn.Module): method __init__ (line 172) | def __init__( method __call__ (line 179) | def __call__( class IPAttnProcessor (line 240) | class IPAttnProcessor(nn.Module): method __init__ (line 254) | def __init__(self, hidden_size, cross_attention_dim=None, scale=1.0, n... method __call__ (line 265) | def __call__( method _memory_efficient_attention_xformers (line 351) | def _memory_efficient_attention_xformers(self, query, key, value, atte... function draw_kps (line 415) | def draw_kps(image_pil, kps, color_list=[(255, 0, 0), (0, 255, 0), (0, 0... class StableDiffusionXLInstantIDImg2ImgPipeline (line 446) | class StableDiffusionXLInstantIDImg2ImgPipeline(StableDiffusionXLControl... method cuda (line 447) | def cuda(self, dtype=torch.float16, use_xformers=False): method load_ip_adapter_instantid (line 467) | def load_ip_adapter_instantid(self, model_ckpt, image_emb_dim=512, num... method set_image_proj_model (line 471) | def set_image_proj_model(self, model_ckpt, image_emb_dim=512, num_toke... method set_ip_adapter (line 493) | def set_ip_adapter(self, model_ckpt, num_tokens, scale): method set_ip_adapter_scale (line 523) | def set_ip_adapter_scale(self, scale): method _encode_prompt_image_emb (line 529) | def _encode_prompt_image_emb(self, prompt_image_emb, device, dtype, do... method __call__ (line 548) | def __call__( FILE: diffusers/examples/community/pipeline_stable_diffusion_xl_instantid.py function FeedForward (line 55) | def FeedForward(dim, mult=4): function reshape_tensor (line 65) | def reshape_tensor(x, heads): class PerceiverAttention (line 76) | class PerceiverAttention(nn.Module): method __init__ (line 77) | def __init__(self, *, dim, dim_head=64, heads=8): method forward (line 91) | def forward(self, x, latents): class Resampler (line 123) | class Resampler(nn.Module): method __init__ (line 124) | def __init__( method forward (line 155) | def forward(self, x): class AttnProcessor (line 167) | class AttnProcessor(nn.Module): method __init__ (line 172) | def __init__( method __call__ (line 179) | def __call__( class IPAttnProcessor (line 240) | class IPAttnProcessor(nn.Module): method __init__ (line 254) | def __init__(self, hidden_size, cross_attention_dim=None, scale=1.0, n... method __call__ (line 265) | def __call__( method _memory_efficient_attention_xformers (line 351) | def _memory_efficient_attention_xformers(self, query, key, value, atte... function draw_kps (line 415) | def draw_kps(image_pil, kps, color_list=[(255, 0, 0), (0, 255, 0), (0, 0... class StableDiffusionXLInstantIDPipeline (line 446) | class StableDiffusionXLInstantIDPipeline(StableDiffusionXLControlNetPipe... method cuda (line 447) | def cuda(self, dtype=torch.float16, use_xformers=False): method load_ip_adapter_instantid (line 467) | def load_ip_adapter_instantid(self, model_ckpt, image_emb_dim=512, num... method set_image_proj_model (line 471) | def set_image_proj_model(self, model_ckpt, image_emb_dim=512, num_toke... method set_ip_adapter (line 493) | def set_ip_adapter(self, model_ckpt, num_tokens, scale): method set_ip_adapter_scale (line 523) | def set_ip_adapter_scale(self, scale): method _encode_prompt_image_emb (line 529) | def _encode_prompt_image_emb(self, prompt_image_emb, device, dtype, do... method __call__ (line 548) | def __call__( FILE: diffusers/examples/community/pipeline_stable_diffusion_xl_ipex.py function rescale_noise_cfg (line 104) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 119) | def retrieve_timesteps( class StableDiffusionXLPipelineIpex (line 163) | class StableDiffusionXLPipelineIpex( method __init__ (line 230) | def __init__( method encode_prompt (line 270) | def encode_prompt( method encode_image (line 505) | def encode_image(self, image, device, num_images_per_prompt): method prepare_extra_step_kwargs (line 519) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 536) | def check_inputs( method prepare_latents (line 616) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 638) | def _get_add_time_ids( method upcast_vae (line 657) | def upcast_vae(self): method get_guidance_scale_embedding (line 677) | def get_guidance_scale_embedding(self, w, embedding_dim=512, dtype=tor... method guidance_scale (line 706) | def guidance_scale(self): method guidance_rescale (line 710) | def guidance_rescale(self): method clip_skip (line 714) | def clip_skip(self): method do_classifier_free_guidance (line 721) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 725) | def cross_attention_kwargs(self): method denoising_end (line 729) | def denoising_end(self): method num_timesteps (line 733) | def num_timesteps(self): method __call__ (line 738) | def __call__( method prepare_for_ipex (line 1177) | def prepare_for_ipex( FILE: diffusers/examples/community/pipeline_zero1to3.py class CCProjection (line 58) | class CCProjection(ModelMixin, ConfigMixin): method __init__ (line 59) | def __init__(self, in_channel=772, out_channel=768): method forward (line 65) | def forward(self, x): class Zero1to3StableDiffusionPipeline (line 69) | class Zero1to3StableDiffusionPipeline(DiffusionPipeline, StableDiffusion... method __init__ (line 98) | def __init__( method _encode_prompt (line 188) | def _encode_prompt( method CLIP_preprocess (line 326) | def CLIP_preprocess(self, x): method _encode_image (line 345) | def _encode_image(self, image, device, num_images_per_prompt, do_class... method _encode_pose (line 402) | def _encode_pose(self, pose, device, num_images_per_prompt, do_classif... method _encode_image_with_pose (line 430) | def _encode_image_with_pose(self, image, pose, device, num_images_per_... method run_safety_checker (line 442) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 452) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 460) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 477) | def check_inputs(self, image, height, width, callback_steps): method prepare_latents (line 499) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method prepare_img_latents (line 521) | def prepare_img_latents(self, image, batch_size, dtype, device, genera... method __call__ (line 596) | def __call__( FILE: diffusers/examples/community/regional_prompting_stable_diffusion.py class RegionalPromptingStableDiffusionPipeline (line 24) | class RegionalPromptingStableDiffusionPipeline(StableDiffusionPipeline): method __init__ (line 64) | def __init__( method __call__ (line 96) | def __call__( function promptsmaker (line 406) | def promptsmaker(prompts, batch): function make_cells (line 426) | def make_cells(ratios): function make_emblist (line 456) | def make_emblist(self, prompts): function split_dims (line 469) | def split_dims(xs, height, width): function get_attn_maps (line 485) | def get_attn_maps(self, attn): function reset_attnmaps (line 508) | def reset_attnmaps(self): # init parameters in every batch function saveattnmaps (line 517) | def saveattnmaps(self, output, h, w, th, step, regions): function makepmask (line 531) | def makepmask( function tokendealer (line 551) | def tokendealer(self, all_prompts): function scaled_dot_product_attention (line 588) | def scaled_dot_product_attention( FILE: diffusers/examples/community/rerender_a_video.py function coords_grid (line 40) | def coords_grid(b, h, w, homogeneous=False, device=None): function bilinear_sample (line 59) | def bilinear_sample(img, sample_coords, mode="bilinear", padding_mode="z... function flow_warp (line 83) | def flow_warp(feature, flow, mask=False, mode="bilinear", padding_mode="... function forward_backward_consistency_check (line 92) | def forward_backward_consistency_check(fwd_flow, bwd_flow, alpha=0.01, b... function get_warped_and_mask (line 115) | def get_warped_and_mask(flow_model, image1, image2, image3=None, pixel_c... class TextToVideoSDPipelineOutput (line 140) | class TextToVideoSDPipelineOutput(BaseOutput): function find_flat_region (line 154) | def find_flat_region(mask): class AttnState (line 165) | class AttnState: method __init__ (line 170) | def __init__(self): method state (line 174) | def state(self): method timestep (line 178) | def timestep(self): method set_timestep (line 181) | def set_timestep(self, t): method reset (line 184) | def reset(self): method to_load (line 188) | def to_load(self): method to_load_and_store_prev (line 191) | def to_load_and_store_prev(self): class CrossFrameAttnProcessor (line 195) | class CrossFrameAttnProcessor(AttnProcessor): method __init__ (line 203) | def __init__(self, attn_state: AttnState): method __call__ (line 209) | def __call__(self, attn: Attention, hidden_states, encoder_hidden_stat... function prepare_image (line 230) | def prepare_image(image): class RerenderAVideoPipeline (line 254) | class RerenderAVideoPipeline(StableDiffusionControlNetImg2ImgPipeline): method __init__ (line 291) | def __init__( method check_inputs (line 384) | def check_inputs( method prepare_control_image (line 490) | def prepare_control_image( method get_timesteps (line 521) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 531) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method __call__ (line 589) | def __call__( class InputPadder (line 1176) | class InputPadder: method __init__ (line 1179) | def __init__(self, dims, mode="sintel", padding_factor=8): method pad (line 1188) | def pad(self, *inputs): method unpad (line 1191) | def unpad(self, x): FILE: diffusers/examples/community/run_onnx_controlnet.py function prepare_image (line 77) | def prepare_image(image): class OnnxStableDiffusionControlNetImg2ImgPipeline (line 101) | class OnnxStableDiffusionControlNetImg2ImgPipeline(DiffusionPipeline): method __init__ (line 109) | def __init__( method _encode_prompt (line 134) | def _encode_prompt( method decode_latents (line 238) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 252) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 269) | def check_inputs( method check_image (line 378) | def check_image(self, image, prompt, prompt_embeds): method prepare_control_image (line 416) | def prepare_control_image( method get_timesteps (line 447) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 456) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method __call__ (line 504) | def __call__( FILE: diffusers/examples/community/run_tensorrt_controlnet.py function load_engine (line 40) | def load_engine(trt_runtime, engine_path): class TensorRTModel (line 47) | class TensorRTModel: method __init__ (line 48) | def __init__( method __call__ (line 105) | def __call__(self, **kwargs): function prepare_image (line 181) | def prepare_image(image): class TensorRTStableDiffusionControlNetImg2ImgPipeline (line 205) | class TensorRTStableDiffusionControlNetImg2ImgPipeline(DiffusionPipeline): method __init__ (line 213) | def __init__( method _encode_prompt (line 238) | def _encode_prompt( method decode_latents (line 342) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 356) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 373) | def check_inputs( method check_image (line 482) | def check_image(self, image, prompt, prompt_embeds): method prepare_control_image (line 520) | def prepare_control_image( method get_timesteps (line 551) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 560) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method __call__ (line 608) | def __call__( FILE: diffusers/examples/community/scheduling_ufogen.py class UFOGenSchedulerOutput (line 32) | class UFOGenSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 50) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 95) | def rescale_zero_terminal_snr(betas): class UFOGenScheduler (line 131) | class UFOGenScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 187) | def __init__( method scale_model_input (line 243) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 260) | def set_timesteps( method _threshold_sample (line 343) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method step (line 376) | def step( method add_noise (line 462) | def add_noise( method get_velocity (line 486) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 504) | def __len__(self): method previous_timestep (line 508) | def previous_timestep(self, timestep): FILE: diffusers/examples/community/sd_text2img_k_diffusion.py class ModelWrapper (line 30) | class ModelWrapper: method __init__ (line 31) | def __init__(self, model, alphas_cumprod): method apply_model (line 35) | def apply_model(self, *args, **kwargs): class StableDiffusionPipeline (line 44) | class StableDiffusionPipeline(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 74) | def __init__( method set_sampler (line 114) | def set_sampler(self, scheduler_type: str): method set_scheduler (line 118) | def set_scheduler(self, scheduler_type: str): method _encode_prompt (line 123) | def _encode_prompt(self, prompt, device, num_images_per_prompt, do_cla... method run_safety_checker (line 228) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 238) | def decode_latents(self, latents): method check_inputs (line 246) | def check_inputs(self, prompt, height, width, callback_steps): method prepare_latents (line 261) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 278) | def __call__( FILE: diffusers/examples/community/sde_drag.py class SdeDragPipeline (line 26) | class SdeDragPipeline(DiffusionPipeline): method __init__ (line 50) | def __init__( method __call__ (line 63) | def __call__( method train_lora (line 207) | def train_lora(self, prompt, image, lora_step=100, lora_rank=16, gener... method _tokenize_prompt (line 332) | def _tokenize_prompt(self, prompt, tokenizer_max_length=None): method _encode_prompt (line 348) | def _encode_prompt(self, input_ids, attention_mask, text_encoder_use_a... method _get_text_embed (line 365) | def _get_text_embed(self, prompt): method _copy_and_paste (line 376) | def _copy_and_paste( method _get_img_latent (line 412) | def _get_img_latent(self, image, height=None, weight=None): method _get_eps (line 425) | def _get_eps(self, latent, timestep, guidance_scale, text_embeddings, ... method _forward_sde (line 450) | def _forward_sde( method _sample (line 486) | def _sample( method _forward (line 536) | def _forward(self, latent, steps, t0, lora_scale_min, text_embeddings,... method _backward (line 572) | def _backward( FILE: diffusers/examples/community/seed_resize_stable_diffusion.py class SeedResizeStableDiffusionPipeline (line 23) | class SeedResizeStableDiffusionPipeline(DiffusionPipeline, StableDiffusi... method __init__ (line 51) | def __init__( method __call__ (line 73) | def __call__( FILE: diffusers/examples/community/speech_to_image_diffusion.py class SpeechToImagePipeline (line 30) | class SpeechToImagePipeline(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 31) | def __init__( method __call__ (line 67) | def __call__( FILE: diffusers/examples/community/stable_diffusion_comparison.py class StableDiffusionComparisonPipeline (line 26) | class StableDiffusionComparisonPipeline(DiffusionPipeline, StableDiffusi... method __init__ (line 54) | def __init__( method layers (line 84) | def layers(self) -> Dict[str, Any]: method text2img_sd1_1 (line 88) | def text2img_sd1_1( method text2img_sd1_2 (line 125) | def text2img_sd1_2( method text2img_sd1_3 (line 162) | def text2img_sd1_3( method text2img_sd1_4 (line 199) | def text2img_sd1_4( method _call_ (line 236) | def _call_( FILE: diffusers/examples/community/stable_diffusion_controlnet_img2img.py function prepare_image (line 64) | def prepare_image(image): function prepare_controlnet_conditioning_image (line 88) | def prepare_controlnet_conditioning_image( class StableDiffusionControlNetImg2ImgPipeline (line 132) | class StableDiffusionControlNetImg2ImgPipeline(DiffusionPipeline, Stable... method __init__ (line 139) | def __init__( method _encode_prompt (line 185) | def _encode_prompt( method run_safety_checker (line 322) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 332) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 340) | def prepare_extra_step_kwargs(self, generator, eta): method check_controlnet_conditioning_image (line 357) | def check_controlnet_conditioning_image(self, image, prompt, prompt_em... method check_inputs (line 393) | def check_inputs( method get_timesteps (line 524) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 533) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method _default_height_width (line 574) | def _default_height_width(self, height, width, image): method __call__ (line 598) | def __call__( FILE: diffusers/examples/community/stable_diffusion_controlnet_inpaint.py function prepare_image (line 127) | def prepare_image(image): function prepare_mask_image (line 151) | def prepare_mask_image(mask_image): function prepare_controlnet_conditioning_image (line 186) | def prepare_controlnet_conditioning_image( class StableDiffusionControlNetInpaintPipeline (line 230) | class StableDiffusionControlNetInpaintPipeline(DiffusionPipeline, Stable... method __init__ (line 237) | def __init__( method _encode_prompt (line 284) | def _encode_prompt( method run_safety_checker (line 421) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 431) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 439) | def prepare_extra_step_kwargs(self, generator, eta): method check_controlnet_conditioning_image (line 456) | def check_controlnet_conditioning_image(self, image, prompt, prompt_em... method check_inputs (line 492) | def check_inputs( method prepare_latents (line 637) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method prepare_mask_latents (line 660) | def prepare_mask_latents(self, mask_image, batch_size, height, width, ... method prepare_masked_image_latents (line 683) | def prepare_masked_image_latents( method _default_height_width (line 717) | def _default_height_width(self, height, width, image): method __call__ (line 741) | def __call__( FILE: diffusers/examples/community/stable_diffusion_controlnet_inpaint_img2img.py function prepare_image (line 126) | def prepare_image(image): function prepare_mask_image (line 150) | def prepare_mask_image(mask_image): function prepare_controlnet_conditioning_image (line 185) | def prepare_controlnet_conditioning_image( class StableDiffusionControlNetInpaintImg2ImgPipeline (line 219) | class StableDiffusionControlNetInpaintImg2ImgPipeline(DiffusionPipeline,... method __init__ (line 226) | def __init__( method _encode_prompt (line 269) | def _encode_prompt( method run_safety_checker (line 406) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 416) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 424) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 441) | def check_inputs( method get_timesteps (line 597) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 606) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method prepare_mask_latents (line 647) | def prepare_mask_latents(self, mask_image, batch_size, height, width, ... method prepare_masked_image_latents (line 670) | def prepare_masked_image_latents( method _default_height_width (line 704) | def _default_height_width(self, height, width, image): method __call__ (line 728) | def __call__( FILE: diffusers/examples/community/stable_diffusion_controlnet_reference.py function torch_dfs (line 60) | def torch_dfs(model: torch.nn.Module): class StableDiffusionControlNetReferencePipeline (line 67) | class StableDiffusionControlNetReferencePipeline(StableDiffusionControlN... method prepare_ref_latents (line 68) | def prepare_ref_latents(self, refimage, batch_size, dtype, device, gen... method __call__ (line 99) | def __call__( FILE: diffusers/examples/community/stable_diffusion_ipex.py class StableDiffusionIPEXPipeline (line 63) | class StableDiffusionIPEXPipeline( method __init__ (line 95) | def __init__( method get_input_example (line 184) | def get_input_example(self, prompt, height=None, width=None, guidance_... method prepare_for_ipex (line 244) | def prepare_for_ipex(self, promt, dtype=torch.float32, height=None, wi... method _encode_prompt (line 307) | def _encode_prompt( method run_safety_checker (line 453) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 463) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 471) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 488) | def check_inputs( method prepare_latents (line 535) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 559) | def __call__( FILE: diffusers/examples/community/stable_diffusion_mega.py class StableDiffusionMegaPipeline (line 27) | class StableDiffusionMegaPipeline(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 57) | def __init__( method components (line 95) | def components(self) -> Dict[str, Any]: method inpaint (line 99) | def inpaint( method img2img (line 134) | def img2img( method text2img (line 169) | def text2img( FILE: diffusers/examples/community/stable_diffusion_reference.py function torch_dfs (line 62) | def torch_dfs(model: torch.nn.Module): class StableDiffusionReferencePipeline (line 78) | class StableDiffusionReferencePipeline( method __init__ (line 117) | def __init__( method _default_height_width (line 221) | def _default_height_width( method check_inputs (line 264) | def check_inputs( method _encode_prompt (line 353) | def _encode_prompt( method encode_prompt (line 403) | def encode_prompt( method prepare_latents (line 585) | def prepare_latents( method prepare_extra_step_kwargs (line 634) | def prepare_extra_step_kwargs( method prepare_image (line 663) | def prepare_image( method prepare_ref_latents (line 733) | def prepare_ref_latents( method run_safety_checker (line 784) | def run_safety_checker( method __call__ (line 813) | def __call__( FILE: diffusers/examples/community/stable_diffusion_repaint.py function prepare_mask_and_masked_image (line 42) | def prepare_mask_and_masked_image(image, mask): class StableDiffusionRepaintPipeline (line 142) | class StableDiffusionRepaintPipeline( method __init__ (line 177) | def __init__( method _encode_prompt (line 281) | def _encode_prompt( method run_safety_checker (line 427) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 438) | def prepare_extra_step_kwargs(self, generator, eta): method decode_latents (line 456) | def decode_latents(self, latents): method check_inputs (line 465) | def check_inputs( method prepare_latents (line 513) | def prepare_latents( method prepare_mask_latents (line 545) | def prepare_mask_latents( method __call__ (line 606) | def __call__( FILE: diffusers/examples/community/stable_diffusion_tensorrt_img2img.py function device_view (line 91) | def device_view(t): function preprocess_image (line 95) | def preprocess_image(image): class Engine (line 108) | class Engine: method __init__ (line 109) | def __init__(self, engine_path): method __del__ (line 116) | def __del__(self): method build (line 123) | def build( method load (line 158) | def load(self): method activate (line 162) | def activate(self): method allocate_buffers (line 165) | def allocate_buffers(self, shape_dict=None, device="cuda"): method infer (line 179) | def infer(self, feed_dict, stream): class Optimizer (line 194) | class Optimizer: method __init__ (line 195) | def __init__(self, onnx_graph): method cleanup (line 198) | def cleanup(self, return_onnx=False): method select_outputs (line 203) | def select_outputs(self, keep, names=None): method fold_constants (line 209) | def fold_constants(self, return_onnx=False): method infer_shapes (line 215) | def infer_shapes(self, return_onnx=False): class BaseModel (line 227) | class BaseModel: method __init__ (line 228) | def __init__(self, model, fp16=False, device="cuda", max_batch_size=16... method get_model (line 244) | def get_model(self): method get_input_names (line 247) | def get_input_names(self): method get_output_names (line 250) | def get_output_names(self): method get_dynamic_axes (line 253) | def get_dynamic_axes(self): method get_sample_input (line 256) | def get_sample_input(self, batch_size, image_height, image_width): method get_input_profile (line 259) | def get_input_profile(self, batch_size, image_height, image_width, sta... method get_shape_dict (line 262) | def get_shape_dict(self, batch_size, image_height, image_width): method optimize (line 265) | def optimize(self, onnx_graph): method check_dims (line 273) | def check_dims(self, batch_size, image_height, image_width): method get_minmax_dims (line 282) | def get_minmax_dims(self, batch_size, image_height, image_width, stati... function getOnnxPath (line 309) | def getOnnxPath(model_name, onnx_dir, opt=True): function getEnginePath (line 313) | def getEnginePath(model_name, engine_dir): function build_engines (line 317) | def build_engines( function runEngine (line 411) | def runEngine(engine, feed_dict, stream): class CLIP (line 415) | class CLIP(BaseModel): method __init__ (line 416) | def __init__(self, model, device, max_batch_size, embedding_dim): method get_input_names (line 422) | def get_input_names(self): method get_output_names (line 425) | def get_output_names(self): method get_dynamic_axes (line 428) | def get_dynamic_axes(self): method get_input_profile (line 431) | def get_input_profile(self, batch_size, image_height, image_width, sta... method get_shape_dict (line 440) | def get_shape_dict(self, batch_size, image_height, image_width): method get_sample_input (line 447) | def get_sample_input(self, batch_size, image_height, image_width): method optimize (line 451) | def optimize(self, onnx_graph): function make_CLIP (line 462) | def make_CLIP(model, device, max_batch_size, embedding_dim, inpaint=False): class UNet (line 466) | class UNet(BaseModel): method __init__ (line 467) | def __init__( method get_input_names (line 481) | def get_input_names(self): method get_output_names (line 484) | def get_output_names(self): method get_dynamic_axes (line 487) | def get_dynamic_axes(self): method get_input_profile (line 494) | def get_input_profile(self, batch_size, image_height, image_width, sta... method get_shape_dict (line 521) | def get_shape_dict(self, batch_size, image_height, image_width): method get_sample_input (line 529) | def get_sample_input(self, batch_size, image_height, image_width): function make_UNet (line 541) | def make_UNet(model, device, max_batch_size, embedding_dim, inpaint=False): class VAE (line 552) | class VAE(BaseModel): method __init__ (line 553) | def __init__(self, model, device, max_batch_size, embedding_dim): method get_input_names (line 559) | def get_input_names(self): method get_output_names (line 562) | def get_output_names(self): method get_dynamic_axes (line 565) | def get_dynamic_axes(self): method get_input_profile (line 568) | def get_input_profile(self, batch_size, image_height, image_width, sta... method get_shape_dict (line 590) | def get_shape_dict(self, batch_size, image_height, image_width): method get_sample_input (line 597) | def get_sample_input(self, batch_size, image_height, image_width): function make_VAE (line 602) | def make_VAE(model, device, max_batch_size, embedding_dim, inpaint=False): class TorchVAEEncoder (line 606) | class TorchVAEEncoder(torch.nn.Module): method __init__ (line 607) | def __init__(self, model): method forward (line 611) | def forward(self, x): class VAEEncoder (line 615) | class VAEEncoder(BaseModel): method __init__ (line 616) | def __init__(self, model, device, max_batch_size, embedding_dim): method get_model (line 622) | def get_model(self): method get_input_names (line 626) | def get_input_names(self): method get_output_names (line 629) | def get_output_names(self): method get_dynamic_axes (line 632) | def get_dynamic_axes(self): method get_input_profile (line 635) | def get_input_profile(self, batch_size, image_height, image_width, sta... method get_shape_dict (line 661) | def get_shape_dict(self, batch_size, image_height, image_width): method get_sample_input (line 668) | def get_sample_input(self, batch_size, image_height, image_width): function make_VAEEncoder (line 673) | def make_VAEEncoder(model, device, max_batch_size, embedding_dim, inpain... class TensorRTStableDiffusionImg2ImgPipeline (line 677) | class TensorRTStableDiffusionImg2ImgPipeline(StableDiffusionImg2ImgPipel... method __init__ (line 705) | def __init__( method __loadModels (line 764) | def __loadModels(self): method set_cached_folder (line 784) | def set_cached_folder(cls, pretrained_model_name_or_path: Optional[Uni... method to (line 806) | def to(self, torch_device: Optional[Union[str, torch.device]] = None, ... method __initialize_timesteps (line 837) | def __initialize_timesteps(self, timesteps, strength): method __preprocess_images (line 846) | def __preprocess_images(self, batch_size, images=()): method __encode_image (line 854) | def __encode_image(self, init_image): method __encode_prompt (line 861) | def __encode_prompt(self, prompt, negative_prompt): method __denoise_latent (line 914) | def __denoise_latent( method __decode_latent (line 947) | def __decode_latent(self, latents): method __loadResources (line 952) | def __loadResources(self, image_height, image_width, batch_size): method __call__ (line 962) | def __call__( FILE: diffusers/examples/community/stable_diffusion_tensorrt_inpaint.py function device_view (line 91) | def device_view(t): function preprocess_image (line 95) | def preprocess_image(image): class Engine (line 108) | class Engine: method __init__ (line 109) | def __init__(self, engine_path): method __del__ (line 116) | def __del__(self): method build (line 123) | def build( method load (line 158) | def load(self): method activate (line 162) | def activate(self): method allocate_buffers (line 165) | def allocate_buffers(self, shape_dict=None, device="cuda"): method infer (line 179) | def infer(self, feed_dict, stream): class Optimizer (line 194) | class Optimizer: method __init__ (line 195) | def __init__(self, onnx_graph): method cleanup (line 198) | def cleanup(self, return_onnx=False): method select_outputs (line 203) | def select_outputs(self, keep, names=None): method fold_constants (line 209) | def fold_constants(self, return_onnx=False): method infer_shapes (line 215) | def infer_shapes(self, return_onnx=False): class BaseModel (line 227) | class BaseModel: method __init__ (line 228) | def __init__(self, model, fp16=False, device="cuda", max_batch_size=16... method get_model (line 244) | def get_model(self): method get_input_names (line 247) | def get_input_names(self): method get_output_names (line 250) | def get_output_names(self): method get_dynamic_axes (line 253) | def get_dynamic_axes(self): method get_sample_input (line 256) | def get_sample_input(self, batch_size, image_height, image_width): method get_input_profile (line 259) | def get_input_profile(self, batch_size, image_height, image_width, sta... method get_shape_dict (line 262) | def get_shape_dict(self, batch_size, image_height, image_width): method optimize (line 265) | def optimize(self, onnx_graph): method check_dims (line 273) | def check_dims(self, batch_size, image_height, image_width): method get_minmax_dims (line 282) | def get_minmax_dims(self, batch_size, image_height, image_width, stati... function getOnnxPath (line 309) | def getOnnxPath(model_name, onnx_dir, opt=True): function getEnginePath (line 313) | def getEnginePath(model_name, engine_dir): function build_engines (line 317) | def build_engines( function runEngine (line 411) | def runEngine(engine, feed_dict, stream): class CLIP (line 415) | class CLIP(BaseModel): method __init__ (line 416) | def __init__(self, model, device, max_batch_size, embedding_dim): method get_input_names (line 422) | def get_input_names(self): method get_output_names (line 425) | def get_output_names(self): method get_dynamic_axes (line 428) | def get_dynamic_axes(self): method get_input_profile (line 431) | def get_input_profile(self, batch_size, image_height, image_width, sta... method get_shape_dict (line 440) | def get_shape_dict(self, batch_size, image_height, image_width): method get_sample_input (line 447) | def get_sample_input(self, batch_size, image_height, image_width): method optimize (line 451) | def optimize(self, onnx_graph): function make_CLIP (line 462) | def make_CLIP(model, device, max_batch_size, embedding_dim, inpaint=False): class UNet (line 466) | class UNet(BaseModel): method __init__ (line 467) | def __init__( method get_input_names (line 481) | def get_input_names(self): method get_output_names (line 484) | def get_output_names(self): method get_dynamic_axes (line 487) | def get_dynamic_axes(self): method get_input_profile (line 494) | def get_input_profile(self, batch_size, image_height, image_width, sta... method get_shape_dict (line 521) | def get_shape_dict(self, batch_size, image_height, image_width): method get_sample_input (line 529) | def get_sample_input(self, batch_size, image_height, image_width): function make_UNet (line 541) | def make_UNet(model, device, max_batch_size, embedding_dim, inpaint=Fals... class VAE (line 552) | class VAE(BaseModel): method __init__ (line 553) | def __init__(self, model, device, max_batch_size, embedding_dim): method get_input_names (line 559) | def get_input_names(self): method get_output_names (line 562) | def get_output_names(self): method get_dynamic_axes (line 565) | def get_dynamic_axes(self): method get_input_profile (line 568) | def get_input_profile(self, batch_size, image_height, image_width, sta... method get_shape_dict (line 590) | def get_shape_dict(self, batch_size, image_height, image_width): method get_sample_input (line 597) | def get_sample_input(self, batch_size, image_height, image_width): function make_VAE (line 602) | def make_VAE(model, device, max_batch_size, embedding_dim, inpaint=False): class TorchVAEEncoder (line 606) | class TorchVAEEncoder(torch.nn.Module): method __init__ (line 607) | def __init__(self, model): method forward (line 611) | def forward(self, x): class VAEEncoder (line 615) | class VAEEncoder(BaseModel): method __init__ (line 616) | def __init__(self, model, device, max_batch_size, embedding_dim): method get_model (line 622) | def get_model(self): method get_input_names (line 626) | def get_input_names(self): method get_output_names (line 629) | def get_output_names(self): method get_dynamic_axes (line 632) | def get_dynamic_axes(self): method get_input_profile (line 635) | def get_input_profile(self, batch_size, image_height, image_width, sta... method get_shape_dict (line 661) | def get_shape_dict(self, batch_size, image_height, image_width): method get_sample_input (line 668) | def get_sample_input(self, batch_size, image_height, image_width): function make_VAEEncoder (line 673) | def make_VAEEncoder(model, device, max_batch_size, embedding_dim, inpain... class TensorRTStableDiffusionInpaintPipeline (line 677) | class TensorRTStableDiffusionInpaintPipeline(StableDiffusionInpaintPipel... method __init__ (line 705) | def __init__( method __loadModels (line 764) | def __loadModels(self): method set_cached_folder (line 784) | def set_cached_folder(cls, pretrained_model_name_or_path: Optional[Uni... method to (line 806) | def to(self, torch_device: Optional[Union[str, torch.device]] = None, ... method __initialize_timesteps (line 837) | def __initialize_timesteps(self, num_inference_steps, strength): method __preprocess_images (line 846) | def __preprocess_images(self, batch_size, images=()): method __encode_image (line 854) | def __encode_image(self, init_image): method __encode_prompt (line 861) | def __encode_prompt(self, prompt, negative_prompt): method __denoise_latent (line 914) | def __denoise_latent( method __decode_latent (line 947) | def __decode_latent(self, latents): method __loadResources (line 952) | def __loadResources(self, image_height, image_width, batch_size): method __call__ (line 962) | def __call__( FILE: diffusers/examples/community/stable_diffusion_tensorrt_txt2img.py function device_view (line 89) | def device_view(t): class Engine (line 93) | class Engine: method __init__ (line 94) | def __init__(self, engine_path): method __del__ (line 101) | def __del__(self): method build (line 108) | def build( method load (line 143) | def load(self): method activate (line 147) | def activate(self): method allocate_buffers (line 150) | def allocate_buffers(self, shape_dict=None, device="cuda"): method infer (line 164) | def infer(self, feed_dict, stream): class Optimizer (line 179) | class Optimizer: method __init__ (line 180) | def __init__(self, onnx_graph): method cleanup (line 183) | def cleanup(self, return_onnx=False): method select_outputs (line 188) | def select_outputs(self, keep, names=None): method fold_constants (line 194) | def fold_constants(self, return_onnx=False): method infer_shapes (line 200) | def infer_shapes(self, return_onnx=False): class BaseModel (line 212) | class BaseModel: method __init__ (line 213) | def __init__(self, model, fp16=False, device="cuda", max_batch_size=16... method get_model (line 229) | def get_model(self): method get_input_names (line 232) | def get_input_names(self): method get_output_names (line 235) | def get_output_names(self): method get_dynamic_axes (line 238) | def get_dynamic_axes(self): method get_sample_input (line 241) | def get_sample_input(self, batch_size, image_height, image_width): method get_input_profile (line 244) | def get_input_profile(self, batch_size, image_height, image_width, sta... method get_shape_dict (line 247) | def get_shape_dict(self, batch_size, image_height, image_width): method optimize (line 250) | def optimize(self, onnx_graph): method check_dims (line 258) | def check_dims(self, batch_size, image_height, image_width): method get_minmax_dims (line 267) | def get_minmax_dims(self, batch_size, image_height, image_width, stati... function getOnnxPath (line 294) | def getOnnxPath(model_name, onnx_dir, opt=True): function getEnginePath (line 298) | def getEnginePath(model_name, engine_dir): function build_engines (line 302) | def build_engines( function runEngine (line 396) | def runEngine(engine, feed_dict, stream): class CLIP (line 400) | class CLIP(BaseModel): method __init__ (line 401) | def __init__(self, model, device, max_batch_size, embedding_dim): method get_input_names (line 407) | def get_input_names(self): method get_output_names (line 410) | def get_output_names(self): method get_dynamic_axes (line 413) | def get_dynamic_axes(self): method get_input_profile (line 416) | def get_input_profile(self, batch_size, image_height, image_width, sta... method get_shape_dict (line 425) | def get_shape_dict(self, batch_size, image_height, image_width): method get_sample_input (line 432) | def get_sample_input(self, batch_size, image_height, image_width): method optimize (line 436) | def optimize(self, onnx_graph): function make_CLIP (line 447) | def make_CLIP(model, device, max_batch_size, embedding_dim, inpaint=False): class UNet (line 451) | class UNet(BaseModel): method __init__ (line 452) | def __init__( method get_input_names (line 466) | def get_input_names(self): method get_output_names (line 469) | def get_output_names(self): method get_dynamic_axes (line 472) | def get_dynamic_axes(self): method get_input_profile (line 479) | def get_input_profile(self, batch_size, image_height, image_width, sta... method get_shape_dict (line 506) | def get_shape_dict(self, batch_size, image_height, image_width): method get_sample_input (line 514) | def get_sample_input(self, batch_size, image_height, image_width): function make_UNet (line 526) | def make_UNet(model, device, max_batch_size, embedding_dim, inpaint=False): class VAE (line 537) | class VAE(BaseModel): method __init__ (line 538) | def __init__(self, model, device, max_batch_size, embedding_dim): method get_input_names (line 544) | def get_input_names(self): method get_output_names (line 547) | def get_output_names(self): method get_dynamic_axes (line 550) | def get_dynamic_axes(self): method get_input_profile (line 553) | def get_input_profile(self, batch_size, image_height, image_width, sta... method get_shape_dict (line 575) | def get_shape_dict(self, batch_size, image_height, image_width): method get_sample_input (line 582) | def get_sample_input(self, batch_size, image_height, image_width): function make_VAE (line 587) | def make_VAE(model, device, max_batch_size, embedding_dim, inpaint=False): class TensorRTStableDiffusionPipeline (line 591) | class TensorRTStableDiffusionPipeline(StableDiffusionPipeline): method __init__ (line 619) | def __init__( method __loadModels (line 678) | def __loadModels(self): method set_cached_folder (line 696) | def set_cached_folder(cls, pretrained_model_name_or_path: Optional[Uni... method to (line 718) | def to(self, torch_device: Optional[Union[str, torch.device]] = None, ... method __encode_prompt (line 749) | def __encode_prompt(self, prompt, negative_prompt): method __denoise_latent (line 802) | def __denoise_latent( method __decode_latent (line 835) | def __decode_latent(self, latents): method __loadResources (line 840) | def __loadResources(self, image_height, image_width, batch_size): method __call__ (line 850) | def __call__( FILE: diffusers/examples/community/stable_diffusion_xl_reference.py function torch_dfs (line 51) | def torch_dfs(model: torch.nn.Module): function rescale_noise_cfg (line 61) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class StableDiffusionXLReferencePipeline (line 75) | class StableDiffusionXLReferencePipeline(StableDiffusionXLPipeline): method _default_height_width (line 76) | def _default_height_width(self, height, width, image): method prepare_image (line 101) | def prepare_image( method prepare_ref_latents (line 154) | def prepare_ref_latents(self, refimage, batch_size, dtype, device, gen... method __call__ (line 189) | def __call__( FILE: diffusers/examples/community/stable_unclip.py function _encode_image (line 18) | def _encode_image(self, image, device, num_images_per_prompt, do_classif... class StableUnCLIPPipeline (line 39) | class StableUnCLIPPipeline(DiffusionPipeline): method __init__ (line 40) | def __init__( method _encode_prompt (line 68) | def _encode_prompt( method _execution_device (line 153) | def _execution_device(self): method prepare_latents (line 170) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method to (line 181) | def to(self, torch_device: Optional[Union[str, torch.device]] = None): method __call__ (line 186) | def __call__( FILE: diffusers/examples/community/text_inpainting.py class TextInpainting (line 26) | class TextInpainting(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 60) | def __init__( method __call__ (line 125) | def __call__( FILE: diffusers/examples/community/tiled_upscaling.py function make_transparency_mask (line 29) | def make_transparency_mask(size, overlap_pixels, remove_borders=[]): function clamp (line 52) | def clamp(n, smallest, largest): function clamp_rect (line 56) | def clamp_rect(rect: [int], min: [int], max: [int]): function add_overlap_rect (line 65) | def add_overlap_rect(rect: [int], overlap: int, image_size: [int]): function squeeze_tile (line 75) | def squeeze_tile(tile, original_image, original_slice, slice_x): function unsqueeze_tile (line 87) | def unsqueeze_tile(tile, original_image_slice): function next_divisible (line 93) | def next_divisible(n, d): class StableDiffusionTiledUpscalePipeline (line 98) | class StableDiffusionTiledUpscalePipeline(StableDiffusionUpscalePipeline): method __init__ (line 125) | def __init__( method _process_tile (line 145) | def _process_tile(self, original_image_slice, x, y, tile_size, tile_bo... method __call__ (line 185) | def __call__( function main (line 282) | def main(): FILE: diffusers/examples/community/unclip_image_interpolation.py function slerp (line 29) | def slerp(val, low, high): class UnCLIPImageInterpolationPipeline (line 41) | class UnCLIPImageInterpolationPipeline(DiffusionPipeline): method __init__ (line 88) | def __init__( method prepare_latents (line 117) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method _encode_prompt (line 129) | def _encode_prompt(self, prompt, device, num_images_per_prompt, do_cla... method _encode_image (line 193) | def _encode_image(self, image, device, num_images_per_prompt, image_em... method __call__ (line 208) | def __call__( FILE: diffusers/examples/community/unclip_text_interpolation.py function slerp (line 25) | def slerp(val, low, high): class UnCLIPTextInterpolationPipeline (line 37) | class UnCLIPTextInterpolationPipeline(DiffusionPipeline): method __init__ (line 82) | def __init__( method prepare_latents (line 111) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method _encode_prompt (line 123) | def _encode_prompt( method __call__ (line 215) | def __call__( FILE: diffusers/examples/community/wildcard_stable_diffusion.py function get_filename (line 25) | def get_filename(path: str): function read_wildcard_values (line 30) | def read_wildcard_values(path: str): function grab_wildcard_values (line 35) | def grab_wildcard_values(wildcard_option_dict: Dict[str, List[str]] = {}... function replace_prompt_with_wildcards (line 45) | def replace_prompt_with_wildcards( class WildcardStableDiffusionOutput (line 62) | class WildcardStableDiffusionOutput(StableDiffusionPipelineOutput): class WildcardStableDiffusionPipeline (line 66) | class WildcardStableDiffusionPipeline(DiffusionPipeline, StableDiffusion... method __init__ (line 111) | def __init__( method __call__ (line 158) | def __call__( FILE: diffusers/examples/conftest.py function pytest_addoption (line 34) | def pytest_addoption(parser): function pytest_terminal_summary (line 40) | def pytest_terminal_summary(terminalreporter): FILE: diffusers/examples/consistency_distillation/test_lcm_lora.py class TextToImageLCM (line 35) | class TextToImageLCM(ExamplesTestsAccelerate): method test_text_to_image_lcm_lora_sdxl (line 36) | def test_text_to_image_lcm_lora_sdxl(self): method test_text_to_image_lcm_lora_sdxl_checkpointing (line 63) | def test_text_to_image_lcm_lora_sdxl_checkpointing(self): FILE: diffusers/examples/consistency_distillation/train_lcm_distill_lora_sd_wds.py function get_module_kohya_state_dict (line 81) | def get_module_kohya_state_dict(module, prefix: str, dtype: torch.dtype,... function filter_keys (line 98) | def filter_keys(key_set): function group_by_keys_nothrow (line 105) | def group_by_keys_nothrow(data, keys=base_plus_ext, lcase=True, suffixes... function tarfile_to_samples_nothrow (line 133) | def tarfile_to_samples_nothrow(src, handler=wds.warn_and_continue): class WebdatasetFilter (line 141) | class WebdatasetFilter: method __init__ (line 142) | def __init__(self, min_size=1024, max_pwatermark=0.5): method __call__ (line 146) | def __call__(self, x): class SDText2ImageDataset (line 161) | class SDText2ImageDataset: method __init__ (line 162) | def __init__( method train_dataset (line 232) | def train_dataset(self): method train_dataloader (line 236) | def train_dataloader(self): function log_validation (line 240) | def log_validation(vae, unet, args, accelerator, weight_dtype, step): function guidance_scale_embedding (line 326) | def guidance_scale_embedding(w, embedding_dim=512, dtype=torch.float32): function append_dims (line 355) | def append_dims(x, target_dims): function scalings_for_boundary_conditions (line 364) | def scalings_for_boundary_conditions(timestep, sigma_data=0.5, timestep_... function get_predicted_original_sample (line 372) | def get_predicted_original_sample(model_output, timesteps, sample, predi... function get_predicted_noise (line 391) | def get_predicted_noise(model_output, timesteps, sample, prediction_type... function extract_into_tensor (line 409) | def extract_into_tensor(a, t, x_shape): class DDIMSolver (line 415) | class DDIMSolver: method __init__ (line 416) | def __init__(self, alpha_cumprods, timesteps=1000, ddim_timesteps=50): method to (line 429) | def to(self, device): method ddim_step (line 435) | def ddim_step(self, pred_x0, pred_noise, timestep_index): function update_ema (line 443) | def update_ema(target_params, source_params, rate=0.99): function import_model_class_from_model_name_or_path (line 456) | def import_model_class_from_model_name_or_path( function parse_args (line 476) | def parse_args(): function encode_prompt (line 851) | def encode_prompt(prompt_batch, text_encoder, tokenizer, proportion_empt... function main (line 876) | def main(args): FILE: diffusers/examples/consistency_distillation/train_lcm_distill_lora_sdxl.py class DDIMSolver (line 78) | class DDIMSolver: method __init__ (line 79) | def __init__(self, alpha_cumprods, timesteps=1000, ddim_timesteps=50): method to (line 93) | def to(self, device): method ddim_step (line 99) | def ddim_step(self, pred_x0, pred_noise, timestep_index): function log_validation (line 106) | def log_validation(vae, args, accelerator, weight_dtype, step, unet=None... function append_dims (line 198) | def append_dims(x, target_dims): function scalings_for_boundary_conditions (line 207) | def scalings_for_boundary_conditions(timestep, sigma_data=0.5, timestep_... function get_predicted_original_sample (line 215) | def get_predicted_original_sample(model_output, timesteps, sample, predi... function get_predicted_noise (line 234) | def get_predicted_noise(model_output, timesteps, sample, prediction_type... function extract_into_tensor (line 252) | def extract_into_tensor(a, t, x_shape): function import_model_class_from_model_name_or_path (line 258) | def import_model_class_from_model_name_or_path( function parse_args (line 278) | def parse_args(): function encode_prompt (line 670) | def encode_prompt(prompt_batch, text_encoders, tokenizers, is_train=True): function main (line 708) | def main(args): FILE: diffusers/examples/consistency_distillation/train_lcm_distill_lora_sdxl_wds.py function get_module_kohya_state_dict (line 87) | def get_module_kohya_state_dict(module, prefix: str, dtype: torch.dtype,... function filter_keys (line 104) | def filter_keys(key_set): function group_by_keys_nothrow (line 111) | def group_by_keys_nothrow(data, keys=base_plus_ext, lcase=True, suffixes... function tarfile_to_samples_nothrow (line 139) | def tarfile_to_samples_nothrow(src, handler=wds.warn_and_continue): class WebdatasetFilter (line 147) | class WebdatasetFilter: method __init__ (line 148) | def __init__(self, min_size=1024, max_pwatermark=0.5): method __call__ (line 152) | def __call__(self, x): class SDXLText2ImageDataset (line 167) | class SDXLText2ImageDataset: method __init__ (line 168) | def __init__( method train_dataset (line 250) | def train_dataset(self): method train_dataloader (line 254) | def train_dataloader(self): function log_validation (line 258) | def log_validation(vae, unet, args, accelerator, weight_dtype, step): function append_dims (line 342) | def append_dims(x, target_dims): function scalings_for_boundary_conditions (line 351) | def scalings_for_boundary_conditions(timestep, sigma_data=0.5, timestep_... function get_predicted_original_sample (line 359) | def get_predicted_original_sample(model_output, timesteps, sample, predi... function get_predicted_noise (line 378) | def get_predicted_noise(model_output, timesteps, sample, prediction_type... function extract_into_tensor (line 396) | def extract_into_tensor(a, t, x_shape): class DDIMSolver (line 402) | class DDIMSolver: method __init__ (line 403) | def __init__(self, alpha_cumprods, timesteps=1000, ddim_timesteps=50): method to (line 417) | def to(self, device): method ddim_step (line 423) | def ddim_step(self, pred_x0, pred_noise, timestep_index): function import_model_class_from_model_name_or_path (line 430) | def import_model_class_from_model_name_or_path( function parse_args (line 450) | def parse_args(): function encode_prompt (line 831) | def encode_prompt(prompt_batch, text_encoders, tokenizers, proportion_em... function main (line 871) | def main(args): FILE: diffusers/examples/consistency_distillation/train_lcm_distill_sd_wds.py function filter_keys (line 80) | def filter_keys(key_set): function group_by_keys_nothrow (line 87) | def group_by_keys_nothrow(data, keys=base_plus_ext, lcase=True, suffixes... function tarfile_to_samples_nothrow (line 115) | def tarfile_to_samples_nothrow(src, handler=wds.warn_and_continue): class WebdatasetFilter (line 123) | class WebdatasetFilter: method __init__ (line 124) | def __init__(self, min_size=1024, max_pwatermark=0.5): method __call__ (line 128) | def __call__(self, x): class SDText2ImageDataset (line 143) | class SDText2ImageDataset: method __init__ (line 144) | def __init__( method train_dataset (line 214) | def train_dataset(self): method train_dataloader (line 218) | def train_dataloader(self): function log_validation (line 222) | def log_validation(vae, unet, args, accelerator, weight_dtype, step, nam... function guidance_scale_embedding (line 304) | def guidance_scale_embedding(w, embedding_dim=512, dtype=torch.float32): function append_dims (line 333) | def append_dims(x, target_dims): function scalings_for_boundary_conditions (line 342) | def scalings_for_boundary_conditions(timestep, sigma_data=0.5, timestep_... function get_predicted_original_sample (line 350) | def get_predicted_original_sample(model_output, timesteps, sample, predi... function get_predicted_noise (line 369) | def get_predicted_noise(model_output, timesteps, sample, prediction_type... function extract_into_tensor (line 387) | def extract_into_tensor(a, t, x_shape): class DDIMSolver (line 393) | class DDIMSolver: method __init__ (line 394) | def __init__(self, alpha_cumprods, timesteps=1000, ddim_timesteps=50): method to (line 407) | def to(self, device): method ddim_step (line 413) | def ddim_step(self, pred_x0, pred_noise, timestep_index): function update_ema (line 421) | def update_ema(target_params, source_params, rate=0.99): function import_model_class_from_model_name_or_path (line 434) | def import_model_class_from_model_name_or_path( function parse_args (line 454) | def parse_args(): function encode_prompt (line 816) | def encode_prompt(prompt_batch, text_encoder, tokenizer, proportion_empt... function main (line 841) | def main(args): FILE: diffusers/examples/consistency_distillation/train_lcm_distill_sdxl_wds.py function filter_keys (line 86) | def filter_keys(key_set): function group_by_keys_nothrow (line 93) | def group_by_keys_nothrow(data, keys=base_plus_ext, lcase=True, suffixes... function tarfile_to_samples_nothrow (line 121) | def tarfile_to_samples_nothrow(src, handler=wds.warn_and_continue): class WebdatasetFilter (line 129) | class WebdatasetFilter: method __init__ (line 130) | def __init__(self, min_size=1024, max_pwatermark=0.5): method __call__ (line 134) | def __call__(self, x): class SDXLText2ImageDataset (line 149) | class SDXLText2ImageDataset: method __init__ (line 150) | def __init__( method train_dataset (line 232) | def train_dataset(self): method train_dataloader (line 236) | def train_dataloader(self): function log_validation (line 240) | def log_validation(vae, unet, args, accelerator, weight_dtype, step, nam... function append_dims (line 321) | def append_dims(x, target_dims): function scalings_for_boundary_conditions (line 330) | def scalings_for_boundary_conditions(timestep, sigma_data=0.5, timestep_... function get_predicted_original_sample (line 338) | def get_predicted_original_sample(model_output, timesteps, sample, predi... function get_predicted_noise (line 357) | def get_predicted_noise(model_output, timesteps, sample, prediction_type... function extract_into_tensor (line 375) | def extract_into_tensor(a, t, x_shape): function update_ema (line 382) | def update_ema(target_params, source_params, rate=0.99): function guidance_scale_embedding (line 396) | def guidance_scale_embedding(w, embedding_dim=512, dtype=torch.float32): class DDIMSolver (line 425) | class DDIMSolver: method __init__ (line 426) | def __init__(self, alpha_cumprods, timesteps=1000, ddim_timesteps=50): method to (line 440) | def to(self, device): method ddim_step (line 446) | def ddim_step(self, pred_x0, pred_noise, timestep_index): function import_model_class_from_model_name_or_path (line 453) | def import_model_class_from_model_name_or_path( function parse_args (line 473) | def parse_args(): function encode_prompt (line 841) | def encode_prompt(prompt_batch, text_encoders, tokenizers, proportion_em... function main (line 881) | def main(args): FILE: diffusers/examples/controlnet/test_controlnet.py class ControlNet (line 33) | class ControlNet(ExamplesTestsAccelerate): method test_controlnet_checkpointing_checkpoints_total_limit (line 34) | def test_controlnet_checkpointing_checkpoints_total_limit(self): method test_controlnet_checkpointing_checkpoints_total_limit_removes_multiple_checkpoints (line 57) | def test_controlnet_checkpointing_checkpoints_total_limit_removes_mult... class ControlNetSDXL (line 99) | class ControlNetSDXL(ExamplesTestsAccelerate): method test_controlnet_sdxl (line 100) | def test_controlnet_sdxl(self): FILE: diffusers/examples/controlnet/train_controlnet.py function image_grid (line 68) | def image_grid(imgs, rows, cols): function log_validation (line 79) | def log_validation( function import_model_class_from_model_name_or_path (line 190) | def import_model_class_from_model_name_or_path(pretrained_model_name_or_... function save_model_card (line 210) | def save_model_card(repo_id: str, image_logs=None, base_model=str, repo_... function parse_args (line 252) | def parse_args(input_args=None): function make_train_dataset (line 606) | def make_train_dataset(args, tokenizer, accelerator): function collate_fn (line 721) | def collate_fn(examples): function main (line 737) | def main(args): FILE: diffusers/examples/controlnet/train_controlnet_flax.py function log_validation (line 68) | def log_validation(pipeline, pipeline_params, controlnet_params, tokeniz... function save_model_card (line 136) | def save_model_card(repo_id: str, image_logs=None, base_model=str, repo_... function parse_args (line 179) | def parse_args(): function make_train_dataset (line 505) | def make_train_dataset(args, tokenizer, batch_size=None): function collate_fn (line 640) | def collate_fn(examples): function get_params_to_save (line 658) | def get_params_to_save(params): function main (line 662) | def main(): FILE: diffusers/examples/controlnet/train_controlnet_sdxl.py function log_validation (line 71) | def log_validation(vae, unet, controlnet, args, accelerator, weight_dtyp... function import_model_class_from_model_name_or_path (line 195) | def import_model_class_from_model_name_or_path( function save_model_card (line 215) | def save_model_card(repo_id: str, image_logs=None, base_model=str, repo_... function parse_args (line 258) | def parse_args(input_args=None): function get_train_dataset (line 633) | def get_train_dataset(args, accelerator): function encode_prompt (line 698) | def encode_prompt(prompt_batch, text_encoders, tokenizers, proportion_em... function prepare_train_dataset (line 738) | def prepare_train_dataset(dataset, accelerator): function collate_fn (line 774) | def collate_fn(examples): function main (line 794) | def main(args): FILE: diffusers/examples/custom_diffusion/retrieve.py function retrieve (line 25) | def retrieve(class_prompt, class_data_dir, num_class_images): function parse_args (line 77) | def parse_args(): FILE: diffusers/examples/custom_diffusion/test_custom_diffusion.py class CustomDiffusion (line 33) | class CustomDiffusion(ExamplesTestsAccelerate): method test_custom_diffusion (line 34) | def test_custom_diffusion(self): method test_custom_diffusion_checkpointing_checkpoints_total_limit (line 59) | def test_custom_diffusion_checkpointing_checkpoints_total_limit(self): method test_custom_diffusion_checkpointing_checkpoints_total_limit_removes_multiple_checkpoints (line 81) | def test_custom_diffusion_checkpointing_checkpoints_total_limit_remove... FILE: diffusers/examples/custom_diffusion/train_custom_diffusion.py function freeze_params (line 71) | def freeze_params(params): function save_model_card (line 76) | def save_model_card(repo_id: str, images=None, base_model=str, prompt=st... function import_model_class_from_model_name_or_path (line 113) | def import_model_class_from_model_name_or_path(pretrained_model_name_or_... function collate_fn (line 133) | def collate_fn(examples, with_prior_preservation): class PromptDataset (line 154) | class PromptDataset(Dataset): method __init__ (line 157) | def __init__(self, prompt, num_samples): method __len__ (line 161) | def __len__(self): method __getitem__ (line 164) | def __getitem__(self, index): class CustomDiffusionDataset (line 171) | class CustomDiffusionDataset(Dataset): method __init__ (line 177) | def __init__( method __len__ (line 235) | def __len__(self): method preprocess (line 238) | def preprocess(self, image, scale, resample): method __getitem__ (line 259) | def __getitem__(self, index): function save_new_embed (line 310) | def save_new_embed(text_encoder, modifier_token_id, accelerator, args, o... function parse_args (line 325) | def parse_args(input_args=None): function main (line 661) | def main(args): FILE: diffusers/examples/dreambooth/test_dreambooth.py class DreamBooth (line 36) | class DreamBooth(ExamplesTestsAccelerate): method test_dreambooth (line 37) | def test_dreambooth(self): method test_dreambooth_if (line 60) | def test_dreambooth_if(self): method test_dreambooth_checkpointing (line 86) | def test_dreambooth_checkpointing(self): method test_dreambooth_checkpointing_checkpoints_total_limit (line 165) | def test_dreambooth_checkpointing_checkpoints_total_limit(self): method test_dreambooth_checkpointing_checkpoints_total_limit_removes_multiple_checkpoints (line 188) | def test_dreambooth_checkpointing_checkpoints_total_limit_removes_mult... FILE: diffusers/examples/dreambooth/test_dreambooth_lora.py class DreamBoothLoRA (line 37) | class DreamBoothLoRA(ExamplesTestsAccelerate): method test_dreambooth_lora (line 38) | def test_dreambooth_lora(self): method test_dreambooth_lora_with_text_encoder (line 70) | def test_dreambooth_lora_with_text_encoder(self): method test_dreambooth_lora_checkpointing_checkpoints_total_limit (line 104) | def test_dreambooth_lora_checkpointing_checkpoints_total_limit(self): method test_dreambooth_lora_checkpointing_checkpoints_total_limit_removes_multiple_checkpoints (line 127) | def test_dreambooth_lora_checkpointing_checkpoints_total_limit_removes... method test_dreambooth_lora_if_model (line 165) | def test_dreambooth_lora_if_model(self): class DreamBoothLoRASDXL (line 201) | class DreamBoothLoRASDXL(ExamplesTestsAccelerate): method test_dreambooth_lora_sdxl (line 202) | def test_dreambooth_lora_sdxl(self): method test_dreambooth_lora_sdxl_with_text_encoder (line 234) | def test_dreambooth_lora_sdxl_with_text_encoder(self): method test_dreambooth_lora_sdxl_custom_captions (line 270) | def test_dreambooth_lora_sdxl_custom_captions(self): method test_dreambooth_lora_sdxl_text_encoder_custom_captions (line 291) | def test_dreambooth_lora_sdxl_text_encoder_custom_captions(self): method test_dreambooth_lora_sdxl_checkpointing_checkpoints_total_limit (line 313) | def test_dreambooth_lora_sdxl_checkpointing_checkpoints_total_limit(se... method test_dreambooth_lora_sdxl_text_encoder_checkpointing_checkpoints_total_limit (line 345) | def test_dreambooth_lora_sdxl_text_encoder_checkpointing_checkpoints_t... FILE: diffusers/examples/dreambooth/test_dreambooth_lora_edm.py class DreamBoothLoRASDXLWithEDM (line 35) | class DreamBoothLoRASDXLWithEDM(ExamplesTestsAccelerate): method test_dreambooth_lora_sdxl_with_edm (line 36) | def test_dreambooth_lora_sdxl_with_edm(self): method test_dreambooth_lora_playground (line 69) | def test_dreambooth_lora_playground(self): FILE: diffusers/examples/dreambooth/train_dreambooth.py function save_model_card (line 71) | def save_model_card( function log_validation (line 115) | def log_validation( function import_model_class_from_model_name_or_path (line 207) | def import_model_class_from_model_name_or_path(pretrained_model_name_or_... function parse_args (line 231) | def parse_args(input_args=None): class DreamBoothDataset (line 610) | class DreamBoothDataset(Dataset): method __init__ (line 616) | def __init__( method __len__ (line 668) | def __len__(self): method __getitem__ (line 671) | def __getitem__(self, index): function collate_fn (line 709) | def collate_fn(examples, with_prior_preservation=False): class PromptDataset (line 744) | class PromptDataset(Dataset): method __init__ (line 747) | def __init__(self, prompt, num_samples): method __len__ (line 751) | def __len__(self): method __getitem__ (line 754) | def __getitem__(self, index): function model_has_vae (line 761) | def model_has_vae(args): function tokenize_prompt (line 771) | def tokenize_prompt(tokenizer, prompt, tokenizer_max_length=None): function encode_prompt (line 788) | def encode_prompt(text_encoder, input_ids, attention_mask, text_encoder_... function main (line 806) | def main(args): FILE: diffusers/examples/dreambooth/train_dreambooth_flax.py function parse_args (line 46) | def parse_args(): class DreamBoothDataset (line 220) | class DreamBoothDataset(Dataset): method __init__ (line 226) | def __init__( method __len__ (line 272) | def __len__(self): method __getitem__ (line 275) | def __getitem__(self, index): class PromptDataset (line 303) | class PromptDataset(Dataset): method __init__ (line 306) | def __init__(self, prompt, num_samples): method __len__ (line 310) | def __len__(self): method __getitem__ (line 313) | def __getitem__(self, index): function get_params_to_save (line 320) | def get_params_to_save(params): function main (line 324) | def main(): FILE: diffusers/examples/dreambooth/train_dreambooth_lora.py function save_model_card (line 78) | def save_model_card( function log_validation (line 119) | def log_validation( function import_model_class_from_model_name_or_path (line 184) | def import_model_class_from_model_name_or_path(pretrained_model_name_or_... function parse_args (line 208) | def parse_args(input_args=None): class DreamBoothDataset (line 550) | class DreamBoothDataset(Dataset): method __init__ (line 556) | def __init__( method __len__ (line 608) | def __len__(self): method __getitem__ (line 611) | def __getitem__(self, index): function collate_fn (line 649) | def collate_fn(examples, with_prior_preservation=False): class PromptDataset (line 682) | class PromptDataset(Dataset): method __init__ (line 685) | def __init__(self, prompt, num_samples): method __len__ (line 689) | def __len__(self): method __getitem__ (line 692) | def __getitem__(self, index): function tokenize_prompt (line 699) | def tokenize_prompt(tokenizer, prompt, tokenizer_max_length=None): function encode_prompt (line 716) | def encode_prompt(text_encoder, input_ids, attention_mask, text_encoder_... function main (line 734) | def main(args): FILE: diffusers/examples/dreambooth/train_dreambooth_lora_sd3.py function save_model_card (line 75) | def save_model_card( function load_text_encoders (line 136) | def load_text_encoders(class_one, class_two, class_three): function log_validation (line 149) | def log_validation( function import_model_class_from_model_name_or_path (line 193) | def import_model_class_from_model_name_or_path( function parse_args (line 212) | def parse_args(input_args=None): class DreamBoothDataset (line 621) | class DreamBoothDataset(Dataset): method __init__ (line 627) | def __init__( method __len__ (line 760) | def __len__(self): method __getitem__ (line 763) | def __getitem__(self, index): function collate_fn (line 790) | def collate_fn(examples, with_prior_preservation=False): class PromptDataset (line 807) | class PromptDataset(Dataset): method __init__ (line 810) | def __init__(self, prompt, num_samples): method __len__ (line 814) | def __len__(self): method __getitem__ (line 817) | def __getitem__(self, index): function tokenize_prompt (line 824) | def tokenize_prompt(tokenizer, prompt): function _encode_prompt_with_t5 (line 836) | def _encode_prompt_with_t5( function _encode_prompt_with_clip (line 870) | def _encode_prompt_with_clip( function encode_prompt (line 903) | def encode_prompt( function main (line 949) | def main(args): FILE: diffusers/examples/dreambooth/train_dreambooth_lora_sdxl.py function determine_scheduler_type (line 86) | def determine_scheduler_type(pretrained_model_name_or_path, revision): function save_model_card (line 100) | def save_model_card( function log_validation (line 177) | def log_validation( function import_model_class_from_model_name_or_path (line 240) | def import_model_class_from_model_name_or_path( function parse_args (line 260) | def parse_args(input_args=None): class DreamBoothDataset (line 700) | class DreamBoothDataset(Dataset): method __init__ (line 706) | def __init__( method __len__ (line 845) | def __len__(self): method __getitem__ (line 848) | def __getitem__(self, index): function collate_fn (line 879) | def collate_fn(examples, with_prior_preservation=False): class PromptDataset (line 905) | class PromptDataset(Dataset): method __init__ (line 908) | def __init__(self, prompt, num_samples): method __len__ (line 912) | def __len__(self): method __getitem__ (line 915) | def __getitem__(self, index): function tokenize_prompt (line 922) | def tokenize_prompt(tokenizer, prompt): function encode_prompt (line 935) | def encode_prompt(text_encoders, tokenizers, prompt, text_input_ids_list... function main (line 962) | def main(args): FILE: diffusers/examples/dreambooth/train_dreambooth_sd3.py function save_model_card (line 71) | def save_model_card( function load_text_encoders (line 135) | def load_text_encoders(class_one, class_two, class_three): function log_validation (line 148) | def log_validation( function import_model_class_from_model_name_or_path (line 192) | def import_model_class_from_model_name_or_path( function parse_args (line 211) | def parse_args(input_args=None): class DreamBoothDataset (line 619) | class DreamBoothDataset(Dataset): method __init__ (line 625) | def __init__( method __len__ (line 758) | def __len__(self): method __getitem__ (line 761) | def __getitem__(self, index): function collate_fn (line 788) | def collate_fn(examples, with_prior_preservation=False): class PromptDataset (line 805) | class PromptDataset(Dataset): method __init__ (line 808) | def __init__(self, prompt, num_samples): method __len__ (line 812) | def __len__(self): method __getitem__ (line 815) | def __getitem__(self, index): function tokenize_prompt (line 822) | def tokenize_prompt(tokenizer, prompt): function _encode_prompt_with_t5 (line 834) | def _encode_prompt_with_t5( function _encode_prompt_with_clip (line 868) | def _encode_prompt_with_clip( function encode_prompt (line 901) | def encode_prompt( function main (line 947) | def main(args): FILE: diffusers/examples/instruct_pix2pix/test_instruct_pix2pix.py class InstructPix2Pix (line 33) | class InstructPix2Pix(ExamplesTestsAccelerate): method test_instruct_pix2pix_checkpointing_checkpoints_total_limit (line 34) | def test_instruct_pix2pix_checkpointing_checkpoints_total_limit(self): method test_instruct_pix2pix_checkpointing_checkpoints_total_limit_removes_multiple_checkpoints (line 57) | def test_instruct_pix2pix_checkpointing_checkpoints_total_limit_remove... FILE: diffusers/examples/instruct_pix2pix/train_instruct_pix2pix.py function log_validation (line 70) | def log_validation( function parse_args (line 112) | def parse_args(): function convert_to_np (line 415) | def convert_to_np(image, resolution): function download_image (line 420) | def download_image(url): function main (line 427) | def main(): FILE: diffusers/examples/instruct_pix2pix/train_instruct_pix2pix_sdxl.py function log_validation (line 74) | def log_validation(pipeline, args, accelerator, generator, global_step, ... function import_model_class_from_model_name_or_path (line 123) | def import_model_class_from_model_name_or_path( function parse_args (line 143) | def parse_args(): function convert_to_np (line 473) | def convert_to_np(image, resolution): function main (line 480) | def main(): FILE: diffusers/examples/kandinsky2_2/text_to_image/train_text_to_image_decoder.py function save_model_card (line 60) | def save_model_card( function log_validation (line 137) | def log_validation(vae, image_encoder, image_processor, unet, args, acce... function parse_args (line 188) | def parse_args(): function main (line 441) | def main(): FILE: diffusers/examples/kandinsky2_2/text_to_image/train_text_to_image_lora_decoder.py function save_model_card (line 54) | def save_model_card(repo_id: str, images=None, base_model=str, dataset_n... function parse_args (line 82) | def parse_args(): function main (line 327) | def main(): FILE: diffusers/examples/kandinsky2_2/text_to_image/train_text_to_image_lora_prior.py function save_model_card (line 54) | def save_model_card(repo_id: str, images=None, base_model=str, dataset_n... function parse_args (line 82) | def parse_args(): function main (line 339) | def main(): FILE: diffusers/examples/kandinsky2_2/text_to_image/train_text_to_image_prior.py function save_model_card (line 63) | def save_model_card( function log_validation (line 140) | def log_validation( function parse_args (line 191) | def parse_args(): function main (line 442) | def main(): FILE: diffusers/examples/research_projects/colossalai/train_dreambooth_colossalai.py function import_model_class_from_model_name_or_path (line 33) | def import_model_class_from_model_name_or_path(pretrained_model_name_or_... function parse_args (line 53) | def parse_args(input_args=None): class DreamBoothDataset (line 250) | class DreamBoothDataset(Dataset): method __init__ (line 256) | def __init__( method __len__ (line 298) | def __len__(self): method __getitem__ (line 301) | def __getitem__(self, index): class PromptDataset (line 329) | class PromptDataset(Dataset): method __init__ (line 332) | def __init__(self, prompt, num_samples): method __len__ (line 336) | def __len__(self): method __getitem__ (line 339) | def __getitem__(self, index): function gemini_zero_dpp (line 347) | def gemini_zero_dpp(model: torch.nn.Module, placememt_policy: str = "aut... function main (line 356) | def main(args): FILE: diffusers/examples/research_projects/consistency_training/train_cm_ct_unconditional.py function _extract_into_tensor (line 59) | def _extract_into_tensor(arr, timesteps, broadcast_shape): function append_dims (line 77) | def append_dims(x, target_dims): function extract_into_tensor (line 85) | def extract_into_tensor(a, t, x_shape): function get_discretization_steps (line 91) | def get_discretization_steps(global_step: int, max_train_steps: int, s_0... function get_skip_steps (line 104) | def get_skip_steps(global_step, initial_skip: int = 1): function get_karras_sigmas (line 109) | def get_karras_sigmas( function get_discretized_lognormal_weights (line 129) | def get_discretized_lognormal_weights(noise_levels: torch.Tensor, p_mean... function get_loss_weighting_schedule (line 140) | def get_loss_weighting_schedule(noise_levels: torch.Tensor): function add_noise (line 147) | def add_noise(original_samples: torch.Tensor, noise: torch.Tensor, times... function get_noise_preconditioning (line 158) | def get_noise_preconditioning(sigmas, noise_precond_type: str = "cm"): function get_input_preconditioning (line 176) | def get_input_preconditioning(sigmas, sigma_data=0.5, input_precond_type... function scalings_for_boundary_conditions (line 191) | def scalings_for_boundary_conditions(timestep, sigma_data=0.5, timestep_... function log_validation (line 198) | def log_validation(unet, scheduler, args, accelerator, weight_dtype, ste... function parse_args (line 278) | def parse_args(): function main (line 786) | def main(args): FILE: diffusers/examples/research_projects/controlnet/train_controlnet_webdataset.py function filter_keys (line 78) | def filter_keys(key_set): function group_by_keys_nothrow (line 85) | def group_by_keys_nothrow(data, keys=base_plus_ext, lcase=True, suffixes... function tarfile_to_samples_nothrow (line 113) | def tarfile_to_samples_nothrow(src, handler=wds.warn_and_continue): function control_transform (line 121) | def control_transform(image): function canny_image_transform (line 134) | def canny_image_transform(example, resolution=1024): function depth_image_transform (line 153) | def depth_image_transform(example, feature_extractor, resolution=1024): class WebdatasetFilter (line 172) | class WebdatasetFilter: method __init__ (line 173) | def __init__(self, min_size=1024, max_pwatermark=0.5): method __call__ (line 177) | def __call__(self, x): class Text2ImageDataset (line 192) | class Text2ImageDataset: method __init__ (line 193) | def __init__( method train_dataset (line 292) | def train_dataset(self): method train_dataloader (line 296) | def train_dataloader(self): method eval_dataset (line 300) | def eval_dataset(self): method eval_dataloader (line 304) | def eval_dataloader(self): function image_grid (line 308) | def image_grid(imgs, rows, cols): function log_validation (line 319) | def log_validation(vae, unet, controlnet, args, accelerator, weight_dtyp... function import_model_class_from_model_name_or_path (line 419) | def import_model_class_from_model_name_or_path( function save_model_card (line 439) | def save_model_card(repo_id: str, image_logs=None, base_model=str, repo_... function parse_args (line 478) | def parse_args(input_args=None): function encode_prompt (line 861) | def encode_prompt(prompt_batch, text_encoders, tokenizers, proportion_em... function main (line 901) | def main(args): FILE: diffusers/examples/research_projects/diffusion_dpo/train_diffusion_dpo.py function import_model_class_from_model_name_or_path (line 72) | def import_model_class_from_model_name_or_path(pretrained_model_name_or_... function log_validation (line 88) | def log_validation(args, unet, accelerator, weight_dtype, epoch, is_fina... function parse_args (line 153) | def parse_args(input_args=None): function tokenize_captions (line 445) | def tokenize_captions(tokenizer, examples): function encode_prompt (line 459) | def encode_prompt(text_encoder, input_ids): function main (line 469) | def main(args): FILE: diffusers/examples/research_projects/diffusion_dpo/train_diffusion_dpo_sdxl.py function import_model_class_from_model_name_or_path (line 73) | def import_model_class_from_model_name_or_path( function log_validation (line 93) | def log_validation(args, unet, vae, accelerator, weight_dtype, epoch, is... function parse_args (line 172) | def parse_args(input_args=None): function tokenize_captions (line 472) | def tokenize_captions(tokenizers, examples): function encode_prompt (line 488) | def encode_prompt(text_encoders, text_input_ids_list): function main (line 511) | def main(args): FILE: diffusers/examples/research_projects/diffusion_orpo/train_diffusion_orpo_sdxl_lora.py function import_model_class_from_model_name_or_path (line 73) | def import_model_class_from_model_name_or_path( function log_validation (line 93) | def log_validation(args, unet, vae, accelerator, weight_dtype, epoch, is... function parse_args (line 170) | def parse_args(input_args=None): function tokenize_captions (line 462) | def tokenize_captions(tokenizers, examples): function encode_prompt (line 478) | def encode_prompt(text_encoders, text_input_ids_list): function main (line 501) | def main(args): FILE: diffusers/examples/research_projects/diffusion_orpo/train_diffusion_orpo_sdxl_lora_wds.py function import_model_class_from_model_name_or_path (line 71) | def import_model_class_from_model_name_or_path( function log_validation (line 91) | def log_validation(args, unet, vae, accelerator, weight_dtype, epoch, is... function parse_args (line 168) | def parse_args(input_args=None): function tokenize_captions (line 443) | def tokenize_captions(tokenizers, sample): function encode_prompt (line 463) | def encode_prompt(text_encoders, text_input_ids_list): function get_dataset (line 486) | def get_dataset(args): function get_loader (line 503) | def get_loader(args, tokenizer_one, tokenizer_two): function main (line 608) | def main(args): FILE: diffusers/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint.py function prepare_mask_and_masked_image (line 40) | def prepare_mask_and_masked_image(image, mask): function random_mask (line 58) | def random_mask(im_shape, ratio=1, mask_full_image=False): function parse_args (line 82) | def parse_args(): class DreamBoothDataset (line 297) | class DreamBoothDataset(Dataset): method __init__ (line 303) | def __init__( method __len__ (line 350) | def __len__(self): method __getitem__ (line 353) | def __getitem__(self, index): class PromptDataset (line 387) | class PromptDataset(Dataset): method __init__ (line 390) | def __init__(self, prompt, num_samples): method __len__ (line 394) | def __len__(self): method __getitem__ (line 397) | def __getitem__(self, index): function main (line 404) | def main(): FILE: diffusers/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint_lora.py function prepare_mask_and_masked_image (line 36) | def prepare_mask_and_masked_image(image, mask): function random_mask (line 54) | def random_mask(im_shape, ratio=1, mask_full_image=False): function parse_args (line 78) | def parse_args(): class DreamBoothDataset (line 296) | class DreamBoothDataset(Dataset): method __init__ (line 302) | def __init__( method __len__ (line 349) | def __len__(self): method __getitem__ (line 352) | def __getitem__(self, index): class PromptDataset (line 386) | class PromptDataset(Dataset): method __init__ (line 389) | def __init__(self, prompt, num_samples): method __len__ (line 393) | def __len__(self): method __getitem__ (line 396) | def __getitem__(self, index): function main (line 403) | def main(): FILE: diffusers/examples/research_projects/gligen/dataset.py function recalculate_box_and_verify_if_valid (line 9) | def recalculate_box_and_verify_if_valid(x, y, w, h, image_size, original... class COCODataset (line 22) | class COCODataset(torch.utils.data.Dataset): method __init__ (line 23) | def __init__( method __getitem__ (line 49) | def __getitem__(self, index): method __len__ (line 109) | def __len__(self): FILE: diffusers/examples/research_projects/gligen/train_gligen_text.py function log_validation (line 46) | def log_validation(vae, text_encoder, tokenizer, unet, noise_scheduler, ... function parse_args (line 98) | def parse_args(input_args=None): function main (line 282) | def main(args): FILE: diffusers/examples/research_projects/instructpix2pix_lora/train_instruct_pix2pix_lora.py function parse_args (line 66) | def parse_args(): function convert_to_np (line 375) | def convert_to_np(image, resolution): function download_image (line 380) | def download_image(url): function main (line 387) | def main(): FILE: diffusers/examples/research_projects/intel_opts/textual_inversion/textual_inversion_bf16.py function save_progress (line 59) | def save_progress(text_encoder, placeholder_token_id, accelerator, args,... function parse_args (line 66) | def parse_args(): class TextualInversionDataset (line 272) | class TextualInversionDataset(Dataset): method __init__ (line 273) | def __init__( method __len__ (line 312) | def __len__(self): method __getitem__ (line 315) | def __getitem__(self, i): function freeze_params (line 358) | def freeze_params(params): function main (line 363) | def main(): FILE: diffusers/examples/research_projects/intel_opts/textual_inversion_dfq/text2images.py function parse_args (line 13) | def parse_args(): function image_grid (line 55) | def image_grid(imgs, rows, cols): function generate_images (line 68) | def generate_images( FILE: diffusers/examples/research_projects/intel_opts/textual_inversion_dfq/textual_inversion.py function save_progress (line 49) | def save_progress(text_encoder, placeholder_token_id, accelerator, args,... function parse_args (line 56) | def parse_args(): class EMAModel (line 270) | class EMAModel: method __init__ (line 275) | def __init__(self, parameters: Iterable[torch.nn.Parameter], decay=0.9... method get_decay (line 282) | def get_decay(self, optimization_step): method step (line 290) | def step(self, parameters): method copy_to (line 305) | def copy_to(self, parameters: Iterable[torch.nn.Parameter]) -> None: method to (line 318) | def to(self, device=None, dtype=None) -> None: class TextualInversionDataset (line 330) | class TextualInversionDataset(Dataset): method __init__ (line 331) | def __init__( method __len__ (line 370) | def __len__(self): method __getitem__ (line 373) | def __getitem__(self, i): function freeze_params (line 416) | def freeze_params(params): function generate_images (line 421) | def generate_images(pipeline, prompt="", guidance_scale=7.5, num_inferen... function main (line 435) | def main(): FILE: diffusers/examples/research_projects/lora/train_text_to_image_lora.py function save_model_card (line 57) | def save_model_card(repo_id: str, images=None, base_model=str, dataset_n... function parse_args (line 86) | def parse_args(): function main (line 394) | def main(): FILE: diffusers/examples/research_projects/multi_subject_dreambooth/train_multi_subject_dreambooth.py function log_validation_images_to_tracker (line 54) | def log_validation_images_to_tracker( function generate_validation_images (line 76) | def generate_validation_images( function import_model_class_from_model_name_or_path (line 159) | def import_model_class_from_model_name_or_path(pretrained_model_name_or_... function parse_args (line 179) | def parse_args(input_args=None): class DreamBoothDataset (line 545) | class DreamBoothDataset(Dataset): method __init__ (line 551) | def __init__( method __len__ (line 604) | def __len__(self): method __getitem__ (line 607) | def __getitem__(self, index): function collate_fn (line 639) | def collate_fn(num_instances, examples, with_prior_preservation=False): class PromptDataset (line 666) | class PromptDataset(Dataset): method __init__ (line 669) | def __init__(self, prompt, num_samples): method __len__ (line 673) | def __len__(self): method __getitem__ (line 676) | def __getitem__(self, index): function main (line 683) | def main(args): FILE: diffusers/examples/research_projects/multi_subject_dreambooth_inpainting/train_multi_subject_dreambooth_inpainting.py function parse_args (line 43) | def parse_args(): function prepare_mask_and_masked_image (line 205) | def prepare_mask_and_masked_image(image, mask): class DreamBoothDataset (line 222) | class DreamBoothDataset(Dataset): method __init__ (line 223) | def __init__( method set_image (line 241) | def set_image(self, img, switch): method __len__ (line 252) | def __len__(self): method __getitem__ (line 255) | def __getitem__(self, index): function weighted_mask (line 289) | def weighted_mask(masks): function collate_fn (line 309) | def collate_fn(examples, tokenizer): function log_validation (line 335) | def log_validation(pipeline, text_encoder, unet, val_pairs, accelerator): function checkpoint (line 348) | def checkpoint(args, global_step, accelerator): function main (line 354) | def main(): FILE: diffusers/examples/research_projects/multi_token_textual_inversion/multi_token_clip.py class MultiTokenCLIPTokenizer (line 33) | class MultiTokenCLIPTokenizer(CLIPTokenizer): method __init__ (line 34) | def __init__(self, *args, **kwargs): method try_adding_tokens (line 38) | def try_adding_tokens(self, placeholder_token, *args, **kwargs): method add_placeholder_tokens (line 46) | def add_placeholder_tokens(self, placeholder_token, *args, num_vec_per... method replace_placeholder_tokens_in_text (line 66) | def replace_placeholder_tokens_in_text(self, text, vector_shuffle=Fals... method __call__ (line 88) | def __call__(self, text, *args, vector_shuffle=False, prop_tokens_to_l... method encode (line 97) | def encode(self, text, *args, vector_shuffle=False, prop_tokens_to_loa... FILE: diffusers/examples/research_projects/multi_token_textual_inversion/textual_inversion.py function add_tokens (line 82) | def add_tokens(tokenizer, text_encoder, placeholder_token, num_vec_per_t... function save_progress (line 100) | def save_progress(tokenizer, text_encoder, accelerator, save_path): function load_multitoken_tokenizer (line 110) | def load_multitoken_tokenizer(tokenizer, text_encoder, learned_embeds_di... function load_multitoken_tokenizer_from_automatic (line 122) | def load_multitoken_tokenizer_from_automatic(tokenizer, text_encoder, au... function get_mask (line 136) | def get_mask(tokenizer, accelerator): function parse_args (line 146) | def parse_args(): class TextualInversionDataset (line 456) | class TextualInversionDataset(Dataset): method __init__ (line 457) | def __init__( method __len__ (line 501) | def __len__(self): method __getitem__ (line 504) | def __getitem__(self, i): function main (line 549) | def main(): FILE: diffusers/examples/research_projects/multi_token_textual_inversion/textual_inversion_flax.py function parse_args (line 64) | def parse_args(): class TextualInversionDataset (line 242) | class TextualInversionDataset(Dataset): method __init__ (line 243) | def __init__( method __len__ (line 282) | def __len__(self): method __getitem__ (line 285) | def __getitem__(self, i): function resize_token_embeddings (line 328) | def resize_token_embeddings(model, new_num_tokens, initializer_token_id,... function get_params_to_save (line 348) | def get_params_to_save(params): function main (line 352) | def main(): FILE: diffusers/examples/research_projects/onnxruntime/text_to_image/train_text_to_image.py function log_validation (line 66) | def log_validation(vae, text_encoder, tokenizer, unet, args, accelerator... function parse_args (line 117) | def parse_args(): function main (line 412) | def main(): FILE: diffusers/examples/research_projects/onnxruntime/textual_inversion/textual_inversion.py function save_model_card (line 87) | def save_model_card(repo_id: str, images=None, base_model=str, repo_fold... function log_validation (line 117) | def log_validation(text_encoder, tokenizer, unet, vae, args, accelerator... function save_progress (line 163) | def save_progress(text_encoder, placeholder_token_ids, accelerator, args... function parse_args (line 174) | def parse_args(): class TextualInversionDataset (line 478) | class TextualInversionDataset(Dataset): method __init__ (line 479) | def __init__( method __len__ (line 518) | def __len__(self): method __getitem__ (line 521) | def __getitem__(self, i): function main (line 564) | def main(): FILE: diffusers/examples/research_projects/onnxruntime/unconditional_image_generation/train_unconditional.py function _extract_into_tensor (line 37) | def _extract_into_tensor(arr, timesteps, broadcast_shape): function parse_args (line 55) | def parse_args(): function main (line 279) | def main(args): FILE: diffusers/examples/research_projects/promptdiffusion/convert_original_promptdiffusion_to_diffusers.py function shave_segments (line 71) | def shave_segments(path, n_shave_prefix_segments=1): function renew_resnet_paths (line 81) | def renew_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_vae_resnet_paths (line 103) | def renew_vae_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_attention_paths (line 119) | def renew_attention_paths(old_list, n_shave_prefix_segments=0): function renew_vae_attention_paths (line 140) | def renew_vae_attention_paths(old_list, n_shave_prefix_segments=0): function assign_to_checkpoint (line 170) | def assign_to_checkpoint( function conv_attn_to_linear (line 225) | def conv_attn_to_linear(checkpoint): function create_unet_diffusers_config (line 237) | def create_unet_diffusers_config(original_config, image_size: int, contr... function create_vae_diffusers_config (line 346) | def create_vae_diffusers_config(original_config, image_size: int): function create_diffusers_schedular (line 370) | def create_diffusers_schedular(original_config): function create_ldm_bert_config (line 380) | def create_ldm_bert_config(original_config): function convert_ldm_unet_checkpoint (line 390) | def convert_ldm_unet_checkpoint( function convert_ldm_vae_checkpoint (line 704) | def convert_ldm_vae_checkpoint(checkpoint, config): function convert_ldm_bert_checkpoint (line 811) | def convert_ldm_bert_checkpoint(checkpoint, config): function convert_ldm_clip_checkpoint (line 861) | def convert_ldm_clip_checkpoint(checkpoint, local_files_only=False, text... function convert_paint_by_example_checkpoint (line 925) | def convert_paint_by_example_checkpoint(checkpoint, local_files_only=Fal... function convert_open_clip_checkpoint (line 992) | def convert_open_clip_checkpoint( function stable_unclip_image_encoder (line 1074) | def stable_unclip_image_encoder(original_config, local_files_only=False): function stable_unclip_image_noising_components (line 1111) | def stable_unclip_image_noising_components( function convert_controlnet_checkpoint (line 1156) | def convert_controlnet_checkpoint( function convert_promptdiffusion_checkpoint (line 1206) | def convert_promptdiffusion_checkpoint( function download_from_original_stable_diffusion_ckpt (line 1257) | def download_from_original_stable_diffusion_ckpt( function download_controlnet_from_original_ckpt (line 1926) | def download_controlnet_from_original_ckpt( function download_promptdiffusion_from_original_ckpt (line 1979) | def download_promptdiffusion_from_original_ckpt( function parse_bool (line 2089) | def parse_bool(string): FILE: diffusers/examples/research_projects/promptdiffusion/pipeline_prompt_diffusion.py function retrieve_timesteps (line 101) | def retrieve_timesteps( class PromptDiffusionPipeline (line 145) | class PromptDiffusionPipeline(DiffusionPipeline, TextualInversionLoaderM... method __init__ (line 189) | def __init__( method enable_vae_slicing (line 242) | def enable_vae_slicing(self): method disable_vae_slicing (line 250) | def disable_vae_slicing(self): method enable_vae_tiling (line 258) | def enable_vae_tiling(self): method disable_vae_tiling (line 267) | def disable_vae_tiling(self): method _encode_prompt (line 275) | def _encode_prompt( method encode_prompt (line 308) | def encode_prompt( method encode_image (line 490) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method run_safety_checker (line 515) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 530) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 542) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 559) | def check_inputs( method check_image (line 722) | def check_image(self, image, prompt, prompt_embeds): method prepare_image (line 760) | def prepare_image( method prepare_latents (line 791) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method enable_freeu (line 814) | def enable_freeu(self, s1: float, s2: float, b1: float, b2: float): method disable_freeu (line 837) | def disable_freeu(self): method get_guidance_scale_embedding (line 842) | def get_guidance_scale_embedding(self, w, embedding_dim=512, dtype=tor... method guidance_scale (line 871) | def guidance_scale(self): method clip_skip (line 875) | def clip_skip(self): method do_classifier_free_guidance (line 882) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 886) | def cross_attention_kwargs(self): method num_timesteps (line 890) | def num_timesteps(self): method __call__ (line 895) | def __call__( FILE: diffusers/examples/research_projects/promptdiffusion/promptdiffusioncontrolnet.py class PromptDiffusionControlNetModel (line 30) | class PromptDiffusionControlNetModel(ControlNetModel): method __init__ (line 102) | def __init__( method forward (line 182) | def forward( FILE: diffusers/examples/research_projects/rdm/pipeline_rdm.py class RDMPipeline (line 30) | class RDMPipeline(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 54) | def __init__( method _encode_prompt (line 85) | def _encode_prompt(self, prompt): method _encode_image (line 108) | def _encode_image(self, retrieved_images, batch_size): method prepare_latents (line 125) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method retrieve_images (line 147) | def retrieve_images(self, retrieved_images, prompt_embeds, knn=10): method __call__ (line 155) | def __call__( FILE: diffusers/examples/research_projects/rdm/retriever.py function normalize_images (line 17) | def normalize_images(images: List[Image.Image]): function preprocess_images (line 23) | def preprocess_images(images: List[np.array], feature_extractor: CLIPFea... class IndexConfig (line 40) | class IndexConfig(PretrainedConfig): method __init__ (line 41) | def __init__( class Index (line 64) | class Index: method __init__ (line 69) | def __init__(self, config: IndexConfig, dataset: Dataset): method set_index_name (line 77) | def set_index_name(self, index_name: str): method init_index (line 80) | def init_index(self): method build_index (line 95) | def build_index( method retrieve_imgs (line 115) | def retrieve_imgs(self, vec, k: int = 20): method retrieve_imgs_batch (line 119) | def retrieve_imgs_batch(self, vec, k: int = 20): method retrieve_indices (line 123) | def retrieve_indices(self, vec, k: int = 20): method retrieve_indices_batch (line 127) | def retrieve_indices_batch(self, vec, k: int = 20): class Retriever (line 132) | class Retriever: method __init__ (line 133) | def __init__( method from_pretrained (line 145) | def from_pretrained( method _build_index (line 158) | def _build_index( method save_pretrained (line 167) | def save_pretrained(self, save_directory): method init_retrieval (line 175) | def init_retrieval(self): method retrieve_imgs (line 179) | def retrieve_imgs(self, embeddings: np.ndarray, k: int): method retrieve_imgs_batch (line 182) | def retrieve_imgs_batch(self, embeddings: np.ndarray, k: int): method retrieve_indices (line 185) | def retrieve_indices(self, embeddings: np.ndarray, k: int): method retrieve_indices_batch (line 188) | def retrieve_indices_batch(self, embeddings: np.ndarray, k: int): method __call__ (line 191) | def __call__( function map_txt_to_clip_feature (line 199) | def map_txt_to_clip_feature(clip_model, tokenizer, prompt): function map_img_to_model_feature (line 221) | def map_img_to_model_feature(model, feature_extractor, imgs, device): function get_dataset_with_emb_from_model (line 233) | def get_dataset_with_emb_from_model(dataset, model, feature_extractor, i... function get_dataset_with_emb_from_clip_model (line 241) | def get_dataset_with_emb_from_clip_model( FILE: diffusers/examples/research_projects/realfill/train_realfill.py function make_mask (line 51) | def make_mask(images, resolution, times=30): function save_model_card (line 68) | def save_model_card( function log_validation (line 105) | def log_validation( function parse_args (line 168) | def parse_args(input_args=None): class RealFillDataset (line 426) | class RealFillDataset(Dataset): method __init__ (line 433) | def __init__( method __len__ (line 462) | def __len__(self): method __getitem__ (line 465) | def __getitem__(self, index): function collate_fn (line 502) | def collate_fn(examples): function main (line 534) | def main(args): FILE: diffusers/examples/research_projects/scheduled_huber_loss_training/dreambooth/train_dreambooth.py function save_model_card (line 71) | def save_model_card( function log_validation (line 115) | def log_validation( function import_model_class_from_model_name_or_path (line 207) | def import_model_class_from_model_name_or_path(pretrained_model_name_or_... function parse_args (line 231) | def parse_args(input_args=None): class DreamBoothDataset (line 630) | class DreamBoothDataset(Dataset): method __init__ (line 636) | def __init__( method __len__ (line 688) | def __len__(self): method __getitem__ (line 691) | def __getitem__(self, index): function collate_fn (line 729) | def collate_fn(examples, with_prior_preservation=False): class PromptDataset (line 764) | class PromptDataset(Dataset): method __init__ (line 767) | def __init__(self, prompt, num_samples): method __len__ (line 771) | def __len__(self): method __getitem__ (line 774) | def __getitem__(self, index): function model_has_vae (line 781) | def model_has_vae(args): function tokenize_prompt (line 791) | def tokenize_prompt(tokenizer, prompt, tokenizer_max_length=None): function encode_prompt (line 808) | def encode_prompt(text_encoder, input_ids, attention_mask, text_encoder_... function conditional_loss (line 827) | def conditional_loss( function main (line 853) | def main(args): FILE: diffusers/examples/research_projects/scheduled_huber_loss_training/dreambooth/train_dreambooth_lora.py function save_model_card (line 78) | def save_model_card( function log_validation (line 119) | def log_validation( function import_model_class_from_model_name_or_path (line 184) | def import_model_class_from_model_name_or_path(pretrained_model_name_or_... function parse_args (line 208) | def parse_args(input_args=None): class DreamBoothDataset (line 570) | class DreamBoothDataset(Dataset): method __init__ (line 576) | def __init__( method __len__ (line 628) | def __len__(self): method __getitem__ (line 631) | def __getitem__(self, index): function collate_fn (line 669) | def collate_fn(examples, with_prior_preservation=False): class PromptDataset (line 702) | class PromptDataset(Dataset): method __init__ (line 705) | def __init__(self, prompt, num_samples): method __len__ (line 709) | def __len__(self): method __getitem__ (line 712) | def __getitem__(self, index): function tokenize_prompt (line 719) | def tokenize_prompt(tokenizer, prompt, tokenizer_max_length=None): function encode_prompt (line 736) | def encode_prompt(text_encoder, input_ids, attention_mask, text_encoder_... function conditional_loss (line 755) | def conditional_loss( function main (line 781) | def main(args): FILE: diffusers/examples/research_projects/scheduled_huber_loss_training/dreambooth/train_dreambooth_lora_sdxl.py function determine_scheduler_type (line 87) | def determine_scheduler_type(pretrained_model_name_or_path, revision): function save_model_card (line 101) | def save_model_card( function log_validation (line 178) | def log_validation( function import_model_class_from_model_name_or_path (line 239) | def import_model_class_from_model_name_or_path( function parse_args (line 259) | def parse_args(input_args=None): class DreamBoothDataset (line 719) | class DreamBoothDataset(Dataset): method __init__ (line 725) | def __init__( method __len__ (line 864) | def __len__(self): method __getitem__ (line 867) | def __getitem__(self, index): function collate_fn (line 898) | def collate_fn(examples, with_prior_preservation=False): class PromptDataset (line 924) | class PromptDataset(Dataset): method __init__ (line 927) | def __init__(self, prompt, num_samples): method __len__ (line 931) | def __len__(self): method __getitem__ (line 934) | def __getitem__(self, index): function tokenize_prompt (line 941) | def tokenize_prompt(tokenizer, prompt): function encode_prompt (line 954) | def encode_prompt(text_encoders, tokenizers, prompt, text_input_ids_list... function conditional_loss (line 982) | def conditional_loss( function main (line 1053) | def main(args): FILE: diffusers/examples/research_projects/scheduled_huber_loss_training/text_to_image/train_text_to_image.py function save_model_card (line 68) | def save_model_card( function log_validation (line 140) | def log_validation(vae, text_encoder, tokenizer, unet, args, accelerator... function parse_args (line 194) | def parse_args(): function conditional_loss (line 518) | def conditional_loss( function main (line 544) | def main(): FILE: diffusers/examples/research_projects/scheduled_huber_loss_training/text_to_image/train_text_to_image_lora.py function save_model_card (line 60) | def save_model_card( function parse_args (line 101) | def parse_args(): function conditional_loss (line 414) | def conditional_loss( function main (line 440) | def main(): FILE: diffusers/examples/research_projects/scheduled_huber_loss_training/text_to_image/train_text_to_image_lora_sdxl.py function save_model_card (line 72) | def save_model_card( function import_model_class_from_model_name_or_path (line 119) | def import_model_class_from_model_name_or_path( function parse_args (line 139) | def parse_args(input_args=None): function tokenize_prompt (line 475) | def tokenize_prompt(tokenizer, prompt): function encode_prompt (line 488) | def encode_prompt(text_encoders, tokenizers, prompt, text_input_ids_list... function conditional_loss (line 516) | def conditional_loss( function main (line 536) | def main(args): FILE: diffusers/examples/research_projects/scheduled_huber_loss_training/text_to_image/train_text_to_image_sdxl.py function save_model_card (line 67) | def save_model_card( function import_model_class_from_model_name_or_path (line 112) | def import_model_class_from_model_name_or_path( function parse_args (line 132) | def parse_args(input_args=None): function encode_prompt (line 507) | def encode_prompt(batch, text_encoders, tokenizers, proportion_empty_pro... function compute_vae_encodings (line 549) | def compute_vae_encodings(batch, vae): function generate_timestep_weights (line 561) | def generate_timestep_weights(args, num_timesteps): function conditional_loss (line 603) | def conditional_loss( function main (line 629) | def main(args): FILE: diffusers/examples/research_projects/sdxl_flax/sdxl_single.py function tokenize_prompt (line 44) | def tokenize_prompt(prompt, neg_prompt): function replicate_all (line 58) | def replicate_all(prompt_ids, neg_prompt_ids, seed): function generate (line 67) | def generate( FILE: diffusers/examples/research_projects/sdxl_flax/sdxl_single_aot.py function tokenize_prompt (line 46) | def tokenize_prompt(prompt, neg_prompt): function replicate_all (line 60) | def replicate_all(prompt_ids, neg_prompt_ids, seed): function aot_compile (line 74) | def aot_compile( function generate (line 111) | def generate(prompt, negative_prompt, seed=default_seed, guidance_scale=... FILE: diffusers/examples/t2i_adapter/test_t2i_adapter.py class T2IAdapter (line 33) | class T2IAdapter(ExamplesTestsAccelerate): method test_t2i_adapter_sdxl (line 34) | def test_t2i_adapter_sdxl(self): FILE: diffusers/examples/t2i_adapter/train_t2i_adapter_sdxl.py function image_grid (line 68) | def image_grid(imgs, rows, cols): function log_validation (line 79) | def log_validation(vae, unet, adapter, args, accelerator, weight_dtype, ... function import_model_class_from_model_name_or_path (line 179) | def import_model_class_from_model_name_or_path( function save_model_card (line 199) | def save_model_card(repo_id: str, image_logs: dict = None, base_model: s... function parse_args (line 241) | def parse_args(input_args=None): function get_train_dataset (line 623) | def get_train_dataset(args, accelerator): function encode_prompt (line 688) | def encode_prompt(prompt_batch, text_encoders, tokenizers, proportion_em... function prepare_train_dataset (line 728) | def prepare_train_dataset(dataset, accelerator): function collate_fn (line 764) | def collate_fn(examples): function main (line 784) | def main(args): FILE: diffusers/examples/test_examples_utils.py class SubprocessCallException (line 27) | class SubprocessCallException(Exception): function run_command (line 31) | def run_command(command: List[str], return_stdout=False): class ExamplesTestsAccelerate (line 48) | class ExamplesTestsAccelerate(unittest.TestCase): method setUpClass (line 50) | def setUpClass(cls): method tearDownClass (line 59) | def tearDownClass(cls): FILE: diffusers/examples/text_to_image/test_text_to_image.py class TextToImage (line 37) | class TextToImage(ExamplesTestsAccelerate): method test_text_to_image (line 38) | def test_text_to_image(self): method test_text_to_image_checkpointing (line 62) | def test_text_to_image_checkpointing(self): method test_text_to_image_checkpointing_use_ema (line 144) | def test_text_to_image_checkpointing_use_ema(self): method test_text_to_image_checkpointing_checkpoints_total_limit (line 228) | def test_text_to_image_checkpointing_checkpoints_total_limit(self): method test_text_to_image_checkpointing_checkpoints_total_limit_removes_multiple_checkpoints (line 267) | def test_text_to_image_checkpointing_checkpoints_total_limit_removes_m... class TextToImageSDXL (line 342) | class TextToImageSDXL(ExamplesTestsAccelerate): method test_text_to_image_sdxl (line 343) | def test_text_to_image_sdxl(self): FILE: diffusers/examples/text_to_image/test_text_to_image_lora.py class TextToImageLoRA (line 38) | class TextToImageLoRA(ExamplesTestsAccelerate): method test_text_to_image_lora_sdxl_checkpointing_checkpoints_total_limit (line 39) | def test_text_to_image_lora_sdxl_checkpointing_checkpoints_total_limit... method test_text_to_image_lora_checkpointing_checkpoints_total_limit (line 76) | def test_text_to_image_lora_checkpointing_checkpoints_total_limit(self): method test_text_to_image_lora_checkpointing_checkpoints_total_limit_removes_multiple_checkpoints (line 119) | def test_text_to_image_lora_checkpointing_checkpoints_total_limit_remo... class TextToImageLoRASDXL (line 202) | class TextToImageLoRASDXL(ExamplesTestsAccelerate): method test_text_to_image_lora_sdxl (line 203) | def test_text_to_image_lora_sdxl(self): method test_text_to_image_lora_sdxl_with_text_encoder (line 229) | def test_text_to_image_lora_sdxl_with_text_encoder(self): method test_text_to_image_lora_sdxl_text_encoder_checkpointing_checkpoints_total_limit (line 264) | def test_text_to_image_lora_sdxl_text_encoder_checkpointing_checkpoint... FILE: diffusers/examples/text_to_image/train_text_to_image.py function save_model_card (line 69) | def save_model_card( function log_validation (line 141) | def log_validation(vae, text_encoder, tokenizer, unet, args, accelerator... function parse_args (line 200) | def parse_args(): function main (line 517) | def main(): FILE: diffusers/examples/text_to_image/train_text_to_image_flax.py function parse_args (line 57) | def parse_args(): function get_params_to_save (line 257) | def get_params_to_save(params): function main (line 261) | def main(): FILE: diffusers/examples/text_to_image/train_text_to_image_lora.py function save_model_card (line 64) | def save_model_card( function log_validation (line 105) | def log_validation( function parse_args (line 147) | def parse_args(): function main (line 439) | def main(): FILE: diffusers/examples/text_to_image/train_text_to_image_lora_sdxl.py function save_model_card (line 78) | def save_model_card( function log_validation (line 125) | def log_validation( function import_model_class_from_model_name_or_path (line 166) | def import_model_class_from_model_name_or_path( function parse_args (line 186) | def parse_args(input_args=None): function tokenize_prompt (line 505) | def tokenize_prompt(tokenizer, prompt): function encode_prompt (line 518) | def encode_prompt(text_encoders, tokenizers, prompt, text_input_ids_list... function main (line 545) | def main(args): FILE: diffusers/examples/text_to_image/train_text_to_image_sdxl.py function save_model_card (line 69) | def save_model_card( function import_model_class_from_model_name_or_path (line 114) | def import_model_class_from_model_name_or_path( function parse_args (line 134) | def parse_args(input_args=None): function encode_prompt (line 492) | def encode_prompt(batch, text_encoders, tokenizers, proportion_empty_pro... function compute_vae_encodings (line 534) | def compute_vae_encodings(batch, vae): function generate_timestep_weights (line 546) | def generate_timestep_weights(args, num_timesteps): function main (line 587) | def main(args): FILE: diffusers/examples/textual_inversion/test_textual_inversion.py class TextualInversion (line 33) | class TextualInversion(ExamplesTestsAccelerate): method test_textual_inversion (line 34) | def test_textual_inversion(self): method test_textual_inversion_checkpointing (line 60) | def test_textual_inversion_checkpointing(self): method test_textual_inversion_checkpointing_checkpoints_total_limit_removes_multiple_checkpoints (line 92) | def test_textual_inversion_checkpointing_checkpoints_total_limit_remov... FILE: diffusers/examples/textual_inversion/test_textual_inversion_sdxl.py class TextualInversionSdxl (line 33) | class TextualInversionSdxl(ExamplesTestsAccelerate): method test_textual_inversion_sdxl (line 34) | def test_textual_inversion_sdxl(self): method test_textual_inversion_sdxl_checkpointing (line 60) | def test_textual_inversion_sdxl_checkpointing(self): method test_textual_inversion_sdxl_checkpointing_checkpoints_total_limit_removes_multiple_checkpoints (line 92) | def test_textual_inversion_sdxl_checkpointing_checkpoints_total_limit_... FILE: diffusers/examples/textual_inversion/textual_inversion.py function save_model_card (line 89) | def save_model_card(repo_id: str, images: list = None, base_model: str =... function log_validation (line 122) | def log_validation(text_encoder, tokenizer, unet, vae, args, accelerator... function save_progress (line 174) | def save_progress(text_encoder, placeholder_token_ids, accelerator, args... function parse_args (line 189) | def parse_args(): class TextualInversionDataset (line 506) | class TextualInversionDataset(Dataset): method __init__ (line 507) | def __init__( method __len__ (line 546) | def __len__(self): method __getitem__ (line 549) | def __getitem__(self, i): function main (line 592) | def main(): FILE: diffusers/examples/textual_inversion/textual_inversion_flax.py function parse_args (line 64) | def parse_args(): class TextualInversionDataset (line 255) | class TextualInversionDataset(Dataset): method __init__ (line 256) | def __init__( method __len__ (line 295) | def __len__(self): method __getitem__ (line 298) | def __getitem__(self, i): function resize_token_embeddings (line 341) | def resize_token_embeddings(model, new_num_tokens, initializer_token_id,... function get_params_to_save (line 361) | def get_params_to_save(params): function main (line 365) | def main(): FILE: diffusers/examples/textual_inversion/textual_inversion_sdxl.py function save_model_card (line 84) | def save_model_card(repo_id: str, images=None, base_model=str, repo_fold... function log_validation (line 118) | def log_validation( function save_progress (line 178) | def save_progress(text_encoder, placeholder_token_ids, accelerator, args... function parse_args (line 193) | def parse_args(): class TextualInversionDataset (line 492) | class TextualInversionDataset(Dataset): method __init__ (line 493) | def __init__( method __len__ (line 535) | def __len__(self): method __getitem__ (line 538) | def __getitem__(self, i): function main (line 591) | def main(): FILE: diffusers/examples/unconditional_image_generation/test_unconditional.py class Unconditional (line 33) | class Unconditional(ExamplesTestsAccelerate): method test_train_unconditional (line 34) | def test_train_unconditional(self): method test_unconditional_checkpointing_checkpoints_total_limit (line 55) | def test_unconditional_checkpointing_checkpoints_total_limit(self): method test_unconditional_checkpointing_checkpoints_total_limit_removes_multiple_checkpoints (line 82) | def test_unconditional_checkpointing_checkpoints_total_limit_removes_m... FILE: diffusers/examples/unconditional_image_generation/train_unconditional.py function _extract_into_tensor (line 37) | def _extract_into_tensor(arr, timesteps, broadcast_shape): function parse_args (line 55) | def parse_args(): function main (line 275) | def main(args): FILE: diffusers/examples/vqgan/discriminator.py class Discriminator (line 13) | class Discriminator(ModelMixin, ConfigMixin): method __init__ (line 15) | def __init__(self, in_channels=3, cond_channels=0, hidden_channels=512... method forward (line 36) | def forward(self, x, cond=None): FILE: diffusers/examples/vqgan/test_vqgan.py class TextToImage (line 42) | class TextToImage(ExamplesTestsAccelerate): method test_vqmodel_config (line 44) | def test_vqmodel_config(self): method test_discriminator_config (line 71) | def test_discriminator_config(self): method get_vq_and_discriminator_configs (line 81) | def get_vq_and_discriminator_configs(self, tmpdir): method test_vqmodel (line 90) | def test_vqmodel(self): method test_vqmodel_checkpointing (line 117) | def test_vqmodel_checkpointing(self): method test_vqmodel_checkpointing_use_ema (line 201) | def test_vqmodel_checkpointing_use_ema(self): method test_vqmodel_checkpointing_checkpoints_total_limit (line 285) | def test_vqmodel_checkpointing_checkpoints_total_limit(self): method test_vqmodel_checkpointing_checkpoints_total_limit_removes_multiple_checkpoints (line 323) | def test_vqmodel_checkpointing_checkpoints_total_limit_removes_multipl... FILE: diffusers/examples/vqgan/train_vqgan.py class AverageMeter (line 58) | class AverageMeter(object): method __init__ (line 61) | def __init__(self): method reset (line 64) | def reset(self): method update (line 70) | def update(self, val, n=1): function _map_layer_to_idx (line 77) | def _map_layer_to_idx(backbone, layers, offset=0): function get_perceptual_loss (line 100) | def get_perceptual_loss(pixel_values, fmap, timm_model, timm_model_resol... function grad_layer_wrt_loss (line 119) | def grad_layer_wrt_loss(loss, layer): function gradient_penalty (line 128) | def gradient_penalty(images, output, weight=10): function log_validation (line 143) | def log_validation(model, args, validation_transform, accelerator, globa... function log_grad_norm (line 195) | def log_grad_norm(model, accelerator, global_step): function parse_args (line 203) | def parse_args(): function main (line 548) | def main(): FILE: diffusers/examples/wuerstchen/text_to_image/modeling_efficient_net_encoder.py class EfficientNetEncoder (line 8) | class EfficientNetEncoder(ModelMixin, ConfigMixin): method __init__ (line 10) | def __init__(self, c_latent=16, c_cond=1280, effnet="efficientnet_v2_s"): method forward (line 22) | def forward(self, x): FILE: diffusers/examples/wuerstchen/text_to_image/train_text_to_image_lora_prior.py function save_model_card (line 62) | def save_model_card( function log_validation (line 143) | def log_validation(text_encoder, tokenizer, prior, args, accelerator, we... function parse_args (line 195) | def parse_args(): function main (line 444) | def main(): FILE: diffusers/examples/wuerstchen/text_to_image/train_text_to_image_prior.py function save_model_card (line 63) | def save_model_card( function log_validation (line 140) | def log_validation(text_encoder, tokenizer, prior, args, accelerator, we... function parse_args (line 193) | def parse_args(): function main (line 442) | def main(): FILE: diffusers/scripts/conversion_ldm_uncond.py function convert_ldm_original (line 9) | def convert_ldm_original(checkpoint_path, config_path, output_path): FILE: diffusers/scripts/convert_amused.py function main (line 34) | def main(): function make_transformer (line 118) | def make_transformer(old_transformer, model_256): function make_vqvae (line 346) | def make_vqvae(old_vae): function convert_vae_block_state_dict (line 476) | def convert_vae_block_state_dict(old_state_dict, prefix_from, new_state_... FILE: diffusers/scripts/convert_animatediff_motion_lora_to_diffusers.py function convert_motion_module (line 7) | def convert_motion_module(original_state_dict): function get_args (line 26) | def get_args(): FILE: diffusers/scripts/convert_animatediff_motion_module_to_diffusers.py function convert_motion_module (line 9) | def convert_motion_module(original_state_dict): function get_args (line 28) | def get_args(): FILE: diffusers/scripts/convert_asymmetric_vqgan_to_diffusers.py function convert_asymmetric_autoencoder_kl_state_dict (line 64) | def convert_asymmetric_autoencoder_kl_state_dict(original_state_dict: Di... function get_asymmetric_autoencoder_kl_from_original_checkpoint (line 123) | def get_asymmetric_autoencoder_kl_from_original_checkpoint( FILE: diffusers/scripts/convert_blipdiffusion_to_diffusers.py function qformer_model_from_original_config (line 45) | def qformer_model_from_original_config(): function embeddings_from_original_checkpoint (line 50) | def embeddings_from_original_checkpoint(model, diffuser_embeddings_prefi... function proj_layer_from_original_checkpoint (line 75) | def proj_layer_from_original_checkpoint(model, diffuser_proj_prefix, ori... function attention_from_original_checkpoint (line 86) | def attention_from_original_checkpoint(model, diffuser_attention_prefix,... function output_layers_from_original_checkpoint (line 137) | def output_layers_from_original_checkpoint(model, diffuser_output_prefix... function encoder_from_original_checkpoint (line 150) | def encoder_from_original_checkpoint(model, diffuser_encoder_prefix, ori... function visual_encoder_layer_from_original_checkpoint (line 206) | def visual_encoder_layer_from_original_checkpoint(model, diffuser_prefix... function visual_encoder_from_original_checkpoint (line 233) | def visual_encoder_from_original_checkpoint(model, diffuser_prefix, orig... function qformer_original_checkpoint_to_diffusers_checkpoint (line 269) | def qformer_original_checkpoint_to_diffusers_checkpoint(model): function get_qformer (line 281) | def get_qformer(model): function load_checkpoint_to_model (line 293) | def load_checkpoint_to_model(checkpoint, model): function save_blip_diffusion_model (line 302) | def save_blip_diffusion_model(model, args): function main (line 333) | def main(args): FILE: diffusers/scripts/convert_consistency_decoder.py function _extract_into_tensor (line 26) | def _extract_into_tensor(arr, timesteps, broadcast_shape): function betas_for_alpha_bar (line 33) | def betas_for_alpha_bar(num_diffusion_timesteps, alpha_bar, max_beta=0.9... function _download (line 43) | def _download(url: str, root: str): class ConsistencyDecoder (line 81) | class ConsistencyDecoder: method __init__ (line 82) | def __init__(self, device="cuda:0", download_root=os.path.expanduser("... method round_timesteps (line 103) | def round_timesteps(timesteps, total_timesteps, n_distilled_steps, tru... method ldm_transform_latent (line 115) | def ldm_transform_latent(z, extra_scale_factor=1): method __call__ (line 129) | def __call__( function save_image (line 181) | def save_image(image, name): function load_image (line 192) | def load_image(uri, size=None, center_crop=False): class TimestepEmbedding_ (line 213) | class TimestepEmbedding_(nn.Module): method __init__ (line 214) | def __init__(self, n_time=1024, n_emb=320, n_out=1280) -> None: method forward (line 220) | def forward(self, x) -> torch.Tensor: class ImageEmbedding (line 227) | class ImageEmbedding(nn.Module): method __init__ (line 228) | def __init__(self, in_channels=7, out_channels=320) -> None: method forward (line 232) | def forward(self, x) -> torch.Tensor: class ImageUnembedding (line 236) | class ImageUnembedding(nn.Module): method __init__ (line 237) | def __init__(self, in_channels=320, out_channels=6) -> None: method forward (line 242) | def forward(self, x) -> torch.Tensor: class ConvResblock (line 246) | class ConvResblock(nn.Module): method __init__ (line 247) | def __init__(self, in_features=320, out_features=320) -> None: method forward (line 260) | def forward(self, x, t): class Downsample (line 276) | class Downsample(nn.Module): method __init__ (line 277) | def __init__(self, in_channels=320) -> None: method forward (line 287) | def forward(self, x, t) -> torch.Tensor: class Upsample (line 307) | class Upsample(nn.Module): method __init__ (line 308) | def __init__(self, in_channels=1024) -> None: method forward (line 318) | def forward(self, x, t) -> torch.Tensor: class ConvUNetVAE (line 336) | class ConvUNetVAE(nn.Module): method __init__ (line 337) | def __init__(self) -> None: method forward (line 435) | def forward(self, x, t, features) -> torch.Tensor: function rename_state_dict_key (line 483) | def rename_state_dict_key(k): function rename_state_dict (line 503) | def rename_state_dict(sd, embedding): function add_state_dict (line 1047) | def add_state_dict(prefix, mod): function new_constructor (line 1085) | def new_constructor(self, **kwargs): FILE: diffusers/scripts/convert_consistency_to_diffusers.py function str2bool (line 109) | def str2bool(v): function convert_resnet (line 123) | def convert_resnet(checkpoint, new_checkpoint, old_prefix, new_prefix, h... function convert_attention (line 142) | def convert_attention(checkpoint, new_checkpoint, old_prefix, new_prefix... function con_pt_to_diffuser (line 164) | def con_pt_to_diffuser(checkpoint_path: str, unet_config): FILE: diffusers/scripts/convert_dance_diffusion_to_diffusers.py function alpha_sigma_to_t (line 50) | def alpha_sigma_to_t(alpha, sigma): function get_crash_schedule (line 56) | def get_crash_schedule(t): class Object (line 62) | class Object(object): class DiffusionUncond (line 66) | class DiffusionUncond(nn.Module): method __init__ (line 67) | def __init__(self, global_args): function download (line 75) | def download(model_name): function convert_resconv_naming (line 141) | def convert_resconv_naming(name): function convert_attn_naming (line 152) | def convert_attn_naming(name): function rename (line 161) | def rename(input_string, max_depth=13): function rename_orig_weights (line 220) | def rename_orig_weights(state_dict): function transform_conv_attns (line 238) | def transform_conv_attns(new_state_dict, new_k, v): function main (line 258) | def main(args): FILE: diffusers/scripts/convert_ddpm_original_checkpoint_to_diffusers.py function shave_segments (line 9) | def shave_segments(path, n_shave_prefix_segments=1): function renew_resnet_paths (line 19) | def renew_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_attention_paths (line 35) | def renew_attention_paths(old_list, n_shave_prefix_segments=0, in_mid=Fa... function assign_to_checkpoint (line 56) | def assign_to_checkpoint( function convert_ddpm_checkpoint (line 99) | def convert_ddpm_checkpoint(checkpoint, config): function convert_vq_autoenc_checkpoint (line 235) | def convert_vq_autoenc_checkpoint(checkpoint, config): FILE: diffusers/scripts/convert_diffusers_sdxl_lora_to_webui.py function convert_and_save (line 28) | def convert_and_save(input_lora, output_lora=None): FILE: diffusers/scripts/convert_diffusers_to_original_sdxl.py function convert_unet_state_dict (line 96) | def convert_unet_state_dict(unet_state_dict): function reshape_weight_for_sd (line 168) | def reshape_weight_for_sd(w): function convert_vae_state_dict (line 176) | def convert_vae_state_dict(vae_state_dict): function convert_openclip_text_enc_state_dict (line 221) | def convert_openclip_text_enc_state_dict(text_enc_dict): function convert_openai_text_enc_state_dict (line 268) | def convert_openai_text_enc_state_dict(text_enc_dict): FILE: diffusers/scripts/convert_diffusers_to_original_stable_diffusion.py function convert_unet_state_dict (line 92) | def convert_unet_state_dict(unet_state_dict): function reshape_weight_for_sd (line 171) | def reshape_weight_for_sd(w): function convert_vae_state_dict (line 179) | def convert_vae_state_dict(vae_state_dict): function convert_text_enc_state_dict_v20 (line 233) | def convert_text_enc_state_dict_v20(text_enc_dict): function convert_text_enc_state_dict (line 280) | def convert_text_enc_state_dict(text_enc_dict): FILE: diffusers/scripts/convert_dit_to_diffusers.py function download_model (line 13) | def download_model(model_name): function main (line 26) | def main(args): FILE: diffusers/scripts/convert_gligen_to_diffusers.py function convert_open_clip_checkpoint (line 34) | def convert_open_clip_checkpoint(checkpoint): function convert_gligen_vae_checkpoint (line 82) | def convert_gligen_vae_checkpoint(checkpoint, config): function convert_gligen_unet_checkpoint (line 200) | def convert_gligen_unet_checkpoint(checkpoint, config, path=None, extrac... function create_vae_config (line 371) | def create_vae_config(original_config, image_size: int): function create_unet_config (line 393) | def create_unet_config(original_config, image_size: int, attention_type): function convert_gligen_to_diffusers (line 438) | def convert_gligen_to_diffusers( FILE: diffusers/scripts/convert_i2vgen_to_diffusers.py function assign_to_checkpoint (line 28) | def assign_to_checkpoint( function renew_attention_paths (line 79) | def renew_attention_paths(old_list, n_shave_prefix_segments=0): function shave_segments (line 91) | def shave_segments(path, n_shave_prefix_segments=1): function renew_temp_conv_paths (line 101) | def renew_temp_conv_paths(old_list, n_shave_prefix_segments=0): function renew_resnet_paths (line 112) | def renew_resnet_paths(old_list, n_shave_prefix_segments=0): function convert_ldm_unet_checkpoint (line 135) | def convert_ldm_unet_checkpoint(checkpoint, config, path=None, extract_e... FILE: diffusers/scripts/convert_if.py function parse_args (line 15) | def parse_args(): function main (line 55) | def main(args): function convert_stage_1_pipeline (line 72) | def convert_stage_1_pipeline(tokenizer, text_encoder, feature_extractor,... function convert_super_res_pipeline (line 97) | def convert_super_res_pipeline(tokenizer, text_encoder, feature_extracto... function get_stage_1_unet (line 138) | def get_stage_1_unet(unet_config, unet_checkpoint_path): function convert_safety_checker (line 158) | def convert_safety_checker(p_head_path, w_head_path): function create_unet_diffusers_config (line 210) | def create_unet_diffusers_config(original_unet_config, class_embed_type=... function convert_ldm_unet_checkpoint (line 298) | def convert_ldm_unet_checkpoint(unet_state_dict, config, path=None): function shave_segments (line 539) | def shave_segments(path, n_shave_prefix_segments=1): function renew_resnet_paths (line 549) | def renew_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_attention_paths (line 571) | def renew_attention_paths(old_list, n_shave_prefix_segments=0): function assign_attention_to_checkpoint (line 601) | def assign_attention_to_checkpoint(new_checkpoint, unet_state_dict, old_... function assign_to_checkpoint (line 649) | def assign_to_checkpoint(paths, checkpoint, old_checkpoint, additional_r... function split_attentions (line 678) | def split_attentions(*, weight, bias, split, chunk_size): function parse_list (line 703) | def parse_list(value): function get_super_res_unet (line 718) | def get_super_res_unet(unet_checkpoint_path, verify_param_count=True, sa... function superres_create_unet_diffusers_config (line 778) | def superres_create_unet_diffusers_config(original_unet_config): function superres_convert_ldm_unet_checkpoint (line 860) | def superres_convert_ldm_unet_checkpoint(unet_state_dict, config, path=N... function verify_param_count (line 1119) | def verify_param_count(orig_path, unet_diffusers_config): function assert_param_count (line 1176) | def assert_param_count(model_1, model_2): function superres_check_against_original (line 1182) | def superres_check_against_original(dump_path, unet_checkpoint_path): FILE: diffusers/scripts/convert_k_upscaler_to_diffusers.py function resnet_to_diffusers_checkpoint (line 13) | def resnet_to_diffusers_checkpoint(resnet, checkpoint, *, diffusers_resn... function self_attn_to_diffusers_checkpoint (line 39) | def self_attn_to_diffusers_checkpoint(checkpoint, *, diffusers_attention... function cross_attn_to_diffusers_checkpoint (line 65) | def cross_attn_to_diffusers_checkpoint( function block_to_diffusers_checkpoint (line 115) | def block_to_diffusers_checkpoint(block, checkpoint, block_idx, block_ty... function unet_to_diffusers_checkpoint (line 175) | def unet_to_diffusers_checkpoint(model, checkpoint): function unet_model_from_original_config (line 217) | def unet_model_from_original_config(original_config): function main (line 269) | def main(args): FILE: diffusers/scripts/convert_kakao_brain_unclip_to_diffusers.py function prior_model_from_original_config (line 45) | def prior_model_from_original_config(): function prior_original_checkpoint_to_diffusers_checkpoint (line 51) | def prior_original_checkpoint_to_diffusers_checkpoint(model, checkpoint,... function prior_attention_to_diffusers (line 172) | def prior_attention_to_diffusers( function prior_ff_to_diffusers (line 207) | def prior_ff_to_diffusers(checkpoint, *, diffusers_ff_prefix, original_f... function decoder_model_from_original_config (line 255) | def decoder_model_from_original_config(): function decoder_original_checkpoint_to_diffusers_checkpoint (line 261) | def decoder_original_checkpoint_to_diffusers_checkpoint(model, checkpoint): function text_proj_from_original_config (line 330) | def text_proj_from_original_config(): function text_proj_original_checkpoint_to_diffusers_checkpoint (line 343) | def text_proj_original_checkpoint_to_diffusers_checkpoint(checkpoint): function super_res_unet_first_steps_model_from_original_config (line 399) | def super_res_unet_first_steps_model_from_original_config(): function super_res_unet_first_steps_original_checkpoint_to_diffusers_checkpoint (line 405) | def super_res_unet_first_steps_original_checkpoint_to_diffusers_checkpoi... function super_res_unet_last_step_model_from_original_config (line 494) | def super_res_unet_last_step_model_from_original_config(): function super_res_unet_last_step_original_checkpoint_to_diffusers_checkpoint (line 500) | def super_res_unet_last_step_original_checkpoint_to_diffusers_checkpoint... function unet_time_embeddings (line 568) | def unet_time_embeddings(checkpoint, original_unet_prefix): function unet_conv_in (line 584) | def unet_conv_in(checkpoint, original_unet_prefix): function unet_conv_norm_out (line 598) | def unet_conv_norm_out(checkpoint, original_unet_prefix): function unet_conv_out (line 612) | def unet_conv_out(checkpoint, original_unet_prefix): function unet_downblock_to_diffusers_checkpoint (line 626) | def unet_downblock_to_diffusers_checkpoint( function unet_midblock_to_diffusers_checkpoint (line 685) | def unet_midblock_to_diffusers_checkpoint(model, checkpoint, *, original... function unet_upblock_to_diffusers_checkpoint (line 729) | def unet_upblock_to_diffusers_checkpoint( function resnet_to_diffusers_checkpoint (line 799) | def resnet_to_diffusers_checkpoint(checkpoint, *, diffusers_resnet_prefi... function attention_to_diffusers_checkpoint (line 826) | def attention_to_diffusers_checkpoint(checkpoint, *, diffusers_attention... function split_attentions (line 887) | def split_attentions(*, weight, bias, split, chunk_size): function text_encoder (line 919) | def text_encoder(): function prior (line 942) | def prior(*, args, checkpoint_map_location): function decoder (line 966) | def decoder(*, args, checkpoint_map_location): function super_res_unet (line 1001) | def super_res_unet(*, args, checkpoint_map_location): function load_checkpoint_to_model (line 1033) | def load_checkpoint_to_model(checkpoint, model, strict=False): FILE: diffusers/scripts/convert_kandinsky3_unet.py function convert_state_dict (line 36) | def convert_state_dict(unet_state_dict): function main (line 75) | def main(model_path, output_path): FILE: diffusers/scripts/convert_kandinsky_to_diffusers.py function prior_model_from_original_config (line 43) | def prior_model_from_original_config(): function prior_original_checkpoint_to_diffusers_checkpoint (line 49) | def prior_original_checkpoint_to_diffusers_checkpoint(model, checkpoint,... function prior_attention_to_diffusers (line 170) | def prior_attention_to_diffusers( function prior_ff_to_diffusers (line 205) | def prior_ff_to_diffusers(checkpoint, *, diffusers_ff_prefix, original_f... function unet_model_from_original_config (line 281) | def unet_model_from_original_config(): function unet_original_checkpoint_to_diffusers_checkpoint (line 287) | def unet_original_checkpoint_to_diffusers_checkpoint(model, checkpoint): function inpaint_unet_model_from_original_config (line 412) | def inpaint_unet_model_from_original_config(): function inpaint_unet_original_checkpoint_to_diffusers_checkpoint (line 418) | def inpaint_unet_original_checkpoint_to_diffusers_checkpoint(model, chec... function unet_time_embeddings (line 487) | def unet_time_embeddings(checkpoint): function unet_conv_in (line 503) | def unet_conv_in(checkpoint): function unet_add_embedding (line 516) | def unet_add_embedding(checkpoint): function unet_encoder_hid_proj (line 533) | def unet_encoder_hid_proj(checkpoint): function unet_conv_norm_out (line 549) | def unet_conv_norm_out(checkpoint): function unet_conv_out (line 563) | def unet_conv_out(checkpoint): function unet_downblock_to_diffusers_checkpoint (line 577) | def unet_downblock_to_diffusers_checkpoint( function unet_midblock_to_diffusers_checkpoint (line 636) | def unet_midblock_to_diffusers_checkpoint(model, checkpoint, *, num_head... function unet_upblock_to_diffusers_checkpoint (line 680) | def unet_upblock_to_diffusers_checkpoint( function resnet_to_diffusers_checkpoint (line 750) | def resnet_to_diffusers_checkpoint(checkpoint, *, diffusers_resnet_prefi... function attention_to_diffusers_checkpoint (line 777) | def attention_to_diffusers_checkpoint(checkpoint, *, diffusers_attention... function split_attentions (line 838) | def split_attentions(*, weight, bias, split, chunk_size): function prior (line 867) | def prior(*, args, checkpoint_map_location): function text2img (line 890) | def text2img(*, args, checkpoint_map_location): function inpaint_text2img (line 908) | def inpaint_text2img(*, args, checkpoint_map_location): function movq_model_from_original_config (line 946) | def movq_model_from_original_config(): function movq_encoder_to_diffusers_checkpoint (line 951) | def movq_encoder_to_diffusers_checkpoint(model, checkpoint): function movq_decoder_to_diffusers_checkpoint (line 1051) | def movq_decoder_to_diffusers_checkpoint(model, checkpoint): function movq_resnet_to_diffusers_checkpoint (line 1158) | def movq_resnet_to_diffusers_checkpoint(resnet, checkpoint, *, diffusers... function movq_resnet_to_diffusers_checkpoint_spatial_norm (line 1185) | def movq_resnet_to_diffusers_checkpoint_spatial_norm(resnet, checkpoint,... function movq_attention_to_diffusers_checkpoint (line 1220) | def movq_attention_to_diffusers_checkpoint(checkpoint, *, diffusers_atte... function movq_attention_to_diffusers_checkpoint_spatial_norm (line 1240) | def movq_attention_to_diffusers_checkpoint_spatial_norm(checkpoint, *, d... function movq_original_checkpoint_to_diffusers_checkpoint (line 1272) | def movq_original_checkpoint_to_diffusers_checkpoint(model, checkpoint): function movq (line 1302) | def movq(*, args, checkpoint_map_location): function load_checkpoint_to_model (line 1320) | def load_checkpoint_to_model(checkpoint, model, strict=False): FILE: diffusers/scripts/convert_ldm_original_checkpoint_to_diffusers.py function shave_segments (line 25) | def shave_segments(path, n_shave_prefix_segments=1): function renew_resnet_paths (line 35) | def renew_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_attention_paths (line 57) | def renew_attention_paths(old_list, n_shave_prefix_segments=0): function assign_to_checkpoint (line 78) | def assign_to_checkpoint( function convert_ldm_checkpoint (line 130) | def convert_ldm_checkpoint(checkpoint, config): FILE: diffusers/scripts/convert_lora_safetensor_to_diffusers.py function convert (line 26) | def convert(base_model_path, checkpoint_path, LORA_PREFIX_UNET, LORA_PRE... FILE: diffusers/scripts/convert_models_diffuser_to_diffusers.py function unet (line 15) | def unet(hor): function value_function (line 59) | def value_function(): FILE: diffusers/scripts/convert_ms_text_to_video_to_diffusers.py function assign_to_checkpoint (line 24) | def assign_to_checkpoint( function renew_attention_paths (line 75) | def renew_attention_paths(old_list, n_shave_prefix_segments=0): function shave_segments (line 96) | def shave_segments(path, n_shave_prefix_segments=1): function renew_temp_conv_paths (line 106) | def renew_temp_conv_paths(old_list, n_shave_prefix_segments=0): function renew_resnet_paths (line 117) | def renew_resnet_paths(old_list, n_shave_prefix_segments=0): function convert_ldm_unet_checkpoint (line 140) | def convert_ldm_unet_checkpoint(checkpoint, config, path=None, extract_e... FILE: diffusers/scripts/convert_music_spectrogram_to_diffusers.py function load_notes_encoder (line 19) | def load_notes_encoder(weights, model): function load_continuous_encoder (line 42) | def load_continuous_encoder(weights, model): function load_decoder (line 67) | def load_decoder(weights, model): function main (line 117) | def main(args): FILE: diffusers/scripts/convert_ncsnpp_original_checkpoint_to_diffusers.py function convert_ncsnpp_checkpoint (line 25) | def convert_ncsnpp_checkpoint(checkpoint, config): FILE: diffusers/scripts/convert_original_audioldm2_to_diffusers.py function shave_segments (line 54) | def shave_segments(path, n_shave_prefix_segments=1): function renew_resnet_paths (line 65) | def renew_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_vae_resnet_paths (line 88) | def renew_vae_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_attention_paths (line 105) | def renew_attention_paths(old_list): function renew_vae_attention_paths (line 126) | def renew_vae_attention_paths(old_list, n_shave_prefix_segments=0): function assign_to_checkpoint (line 156) | def assign_to_checkpoint( function conv_attn_to_linear (line 202) | def conv_attn_to_linear(checkpoint): function create_unet_diffusers_config (line 212) | def create_unet_diffusers_config(original_config, image_size: int): function create_vae_diffusers_config (line 258) | def create_vae_diffusers_config(original_config, checkpoint, image_size:... function create_diffusers_schedular (line 287) | def create_diffusers_schedular(original_config): function convert_ldm_unet_checkpoint (line 297) | def convert_ldm_unet_checkpoint(checkpoint, config, path=None, extract_e... function convert_ldm_vae_checkpoint (line 500) | def convert_ldm_vae_checkpoint(checkpoint, config): function convert_open_clap_checkpoint (line 633) | def convert_open_clap_checkpoint(checkpoint): function create_transformers_vocoder_config (line 692) | def create_transformers_vocoder_config(original_config): function extract_sub_model (line 714) | def extract_sub_model(checkpoint, key_prefix): function convert_hifigan_checkpoint (line 728) | def convert_hifigan_checkpoint(checkpoint, config): function convert_projection_checkpoint (line 748) | def convert_projection_checkpoint(checkpoint): function load_pipeline_from_original_AudioLDM2_ckpt (line 831) | def load_pipeline_from_original_AudioLDM2_ckpt( FILE: diffusers/scripts/convert_original_audioldm_to_diffusers.py function shave_segments (line 45) | def shave_segments(path, n_shave_prefix_segments=1): function renew_resnet_paths (line 56) | def renew_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_vae_resnet_paths (line 79) | def renew_vae_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_attention_paths (line 96) | def renew_attention_paths(old_list): function renew_vae_attention_paths (line 118) | def renew_vae_attention_paths(old_list, n_shave_prefix_segments=0): function assign_to_checkpoint (line 149) | def assign_to_checkpoint( function conv_attn_to_linear (line 201) | def conv_attn_to_linear(checkpoint): function create_unet_diffusers_config (line 213) | def create_unet_diffusers_config(original_config, image_size: int): function create_vae_diffusers_config (line 266) | def create_vae_diffusers_config(original_config, checkpoint, image_size:... function create_diffusers_schedular (line 295) | def create_diffusers_schedular(original_config): function convert_ldm_unet_checkpoint (line 306) | def convert_ldm_unet_checkpoint(checkpoint, config, path=None, extract_e... function convert_ldm_vae_checkpoint (line 484) | def convert_ldm_vae_checkpoint(checkpoint, config): function convert_open_clap_checkpoint (line 608) | def convert_open_clap_checkpoint(checkpoint): function create_transformers_vocoder_config (line 666) | def create_transformers_vocoder_config(original_config): function convert_hifigan_checkpoint (line 688) | def convert_hifigan_checkpoint(checkpoint, config): function load_pipeline_from_original_audioldm_ckpt (line 768) | def load_pipeline_from_original_audioldm_ckpt( FILE: diffusers/scripts/convert_original_controlnet_to_diffusers.py function parse_bool (line 80) | def parse_bool(string): FILE: diffusers/scripts/convert_original_musicldm_to_diffusers.py function shave_segments (line 46) | def shave_segments(path, n_shave_prefix_segments=1): function renew_resnet_paths (line 57) | def renew_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_vae_resnet_paths (line 80) | def renew_vae_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_attention_paths (line 97) | def renew_attention_paths(old_list): function renew_vae_attention_paths (line 118) | def renew_vae_attention_paths(old_list, n_shave_prefix_segments=0): function assign_to_checkpoint (line 149) | def assign_to_checkpoint( function conv_attn_to_linear (line 200) | def conv_attn_to_linear(checkpoint): function create_unet_diffusers_config (line 210) | def create_unet_diffusers_config(original_config, image_size: int): function create_vae_diffusers_config (line 263) | def create_vae_diffusers_config(original_config, checkpoint, image_size:... function create_diffusers_schedular (line 292) | def create_diffusers_schedular(original_config): function convert_ldm_unet_checkpoint (line 302) | def convert_ldm_unet_checkpoint(checkpoint, config, path=None, extract_e... function convert_ldm_vae_checkpoint (line 480) | def convert_ldm_vae_checkpoint(checkpoint, config): function convert_open_clap_checkpoint (line 613) | def convert_open_clap_checkpoint(checkpoint): function create_transformers_vocoder_config (line 672) | def create_transformers_vocoder_config(original_config): function convert_hifigan_checkpoint (line 694) | def convert_hifigan_checkpoint(checkpoint, config): function load_pipeline_from_original_MusicLDM_ckpt (line 774) | def load_pipeline_from_original_MusicLDM_ckpt( FILE: diffusers/scripts/convert_original_t2i_adapter.py function convert_adapter (line 26) | def convert_adapter(src_state, in_channels): function convert_light_adapter (line 103) | def convert_light_adapter(src_state): FILE: diffusers/scripts/convert_pixart_alpha_to_diffusers.py function main (line 15) | def main(args): FILE: diffusers/scripts/convert_pixart_sigma_to_diffusers.py function main (line 15) | def main(args): FILE: diffusers/scripts/convert_shap_e_to_diffusers.py function prior_model_from_original_config (line 52) | def prior_model_from_original_config(): function prior_original_checkpoint_to_diffusers_checkpoint (line 58) | def prior_original_checkpoint_to_diffusers_checkpoint(model, checkpoint): function prior_attention_to_diffusers (line 169) | def prior_attention_to_diffusers( function prior_ff_to_diffusers (line 204) | def prior_ff_to_diffusers(checkpoint, *, diffusers_ff_prefix, original_f... function prior_image_model_from_original_config (line 244) | def prior_image_model_from_original_config(): function prior_image_original_checkpoint_to_diffusers_checkpoint (line 250) | def prior_image_original_checkpoint_to_diffusers_checkpoint(model, check... function create_mc_lookup_table (line 830) | def create_mc_lookup_table(): function renderer_model_from_original_config (line 860) | def renderer_model_from_original_config(): function renderer_model_original_checkpoint_to_diffusers_checkpoint (line 871) | def renderer_model_original_checkpoint_to_diffusers_checkpoint(model, ch... function split_attentions (line 898) | def split_attentions(*, weight, bias, split, chunk_size): function prior (line 930) | def prior(*, args, checkpoint_map_location): function prior_image (line 948) | def prior_image(*, args, checkpoint_map_location): function renderer (line 967) | def renderer(*, args, checkpoint_map_location): function load_prior_checkpoint_to_model (line 991) | def load_prior_checkpoint_to_model(checkpoint, model): function load_checkpoint_to_model (line 1004) | def load_checkpoint_to_model(checkpoint, model, strict=False): FILE: diffusers/scripts/convert_stable_diffusion_checkpoint_to_onnx.py function onnx_export (line 31) | def onnx_export( function convert_models (line 71) | def convert_models(model_path: str, output_path: str, opset: int, fp16: ... FILE: diffusers/scripts/convert_stable_diffusion_controlnet_to_onnx.py class Optimizer (line 26) | class Optimizer: method __init__ (line 27) | def __init__(self, onnx_graph, verbose=False): method info (line 31) | def info(self, prefix): method cleanup (line 37) | def cleanup(self, return_onnx=False): method select_outputs (line 42) | def select_outputs(self, keep, names=None): method fold_constants (line 48) | def fold_constants(self, return_onnx=False): method infer_shapes (line 54) | def infer_shapes(self, return_onnx=False): function optimize (line 66) | def optimize(onnx_graph, name, verbose): class UNet2DConditionControlNetModel (line 80) | class UNet2DConditionControlNetModel(torch.nn.Module): method __init__ (line 81) | def __init__( method forward (line 90) | def forward( class UNet2DConditionXLControlNetModel (line 131) | class UNet2DConditionXLControlNetModel(torch.nn.Module): method __init__ (line 132) | def __init__( method forward (line 141) | def forward( function onnx_export (line 187) | def onnx_export( function convert_models (line 228) | def convert_models( FILE: diffusers/scripts/convert_stable_diffusion_controlnet_to_tensorrt.py function convert_models (line 7) | def convert_models(onnx_path: str, num_controlnet: int, output_path: str... FILE: diffusers/scripts/convert_svd_to_diffusers.py function create_unet_diffusers_config (line 10) | def create_unet_diffusers_config(original_config, image_size: int, contr... function assign_to_checkpoint (line 117) | def assign_to_checkpoint( function renew_attention_paths (line 186) | def renew_attention_paths(old_list, n_shave_prefix_segments=0): function shave_segments (line 213) | def shave_segments(path, n_shave_prefix_segments=1): function renew_resnet_paths (line 223) | def renew_resnet_paths(old_list, n_shave_prefix_segments=0): function convert_ldm_unet_checkpoint (line 247) | def convert_ldm_unet_checkpoint( function conv_attn_to_linear (line 544) | def conv_attn_to_linear(checkpoint): function renew_vae_resnet_paths (line 556) | def renew_vae_resnet_paths(old_list, n_shave_prefix_segments=0, is_tempo... function renew_vae_attention_paths (line 593) | def renew_vae_attention_paths(old_list, n_shave_prefix_segments=0): function convert_ldm_vae_checkpoint (line 623) | def convert_ldm_vae_checkpoint(checkpoint, config): FILE: diffusers/scripts/convert_unidiffuser_to_diffusers.py function shave_segments (line 37) | def shave_segments(path, n_shave_prefix_segments=1): function renew_vae_resnet_paths (line 48) | def renew_vae_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_vae_attention_paths (line 65) | def renew_vae_attention_paths(old_list, n_shave_prefix_segments=0): function conv_attn_to_linear (line 96) | def conv_attn_to_linear(checkpoint): function assign_to_checkpoint (line 110) | def assign_to_checkpoint( function create_vae_diffusers_config (line 170) | def create_vae_diffusers_config(config_type): function create_unidiffuser_unet_config (line 184) | def create_unidiffuser_unet_config(config_type, version): function create_text_decoder_config (line 201) | def create_text_decoder_config(config_type): function create_vae_diffusers_config_test (line 216) | def create_vae_diffusers_config_test(): function create_unidiffuser_unet_config_test (line 230) | def create_unidiffuser_unet_config_test(): function create_text_decoder_config_test (line 259) | def create_text_decoder_config_test(): function create_vae_diffusers_config_big (line 282) | def create_vae_diffusers_config_big(): function create_unidiffuser_unet_config_big (line 296) | def create_unidiffuser_unet_config_big(): function create_text_decoder_config_big (line 326) | def create_text_decoder_config_big(): function convert_vae_to_diffusers (line 348) | def convert_vae_to_diffusers(ckpt, diffusers_model, num_head_channels=1): function convert_uvit_block_to_diffusers_block (line 496) | def convert_uvit_block_to_diffusers_block( function convert_uvit_to_diffusers (line 539) | def convert_uvit_to_diffusers(ckpt, diffusers_model): function convert_caption_decoder_to_diffusers (line 602) | def convert_caption_decoder_to_diffusers(ckpt, diffusers_model): FILE: diffusers/scripts/convert_vae_diff_to_onnx.py function onnx_export (line 28) | def onnx_export( function convert_models (line 68) | def convert_models(model_path: str, output_path: str, opset: int, fp16: ... FILE: diffusers/scripts/convert_vae_pt_to_diffusers.py function custom_convert_ldm_vae_checkpoint (line 18) | def custom_convert_ldm_vae_checkpoint(checkpoint, config): function vae_pt_to_vae_diffuser (line 119) | def vae_pt_to_vae_diffuser( FILE: diffusers/scripts/convert_versatile_diffusion_to_diffusers.py function shave_segments (line 96) | def shave_segments(path, n_shave_prefix_segments=1): function renew_resnet_paths (line 106) | def renew_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_vae_resnet_paths (line 128) | def renew_vae_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_attention_paths (line 144) | def renew_attention_paths(old_list, n_shave_prefix_segments=0): function renew_vae_attention_paths (line 165) | def renew_vae_attention_paths(old_list, n_shave_prefix_segments=0): function assign_to_checkpoint (line 195) | def assign_to_checkpoint( function conv_attn_to_linear (line 247) | def conv_attn_to_linear(checkpoint): function create_image_unet_diffusers_config (line 259) | def create_image_unet_diffusers_config(unet_params): function create_text_unet_diffusers_config (line 298) | def create_text_unet_diffusers_config(unet_params): function create_vae_diffusers_config (line 337) | def create_vae_diffusers_config(vae_params): function create_diffusers_scheduler (line 359) | def create_diffusers_scheduler(original_config): function convert_vd_unet_checkpoint (line 369) | def convert_vd_unet_checkpoint(checkpoint, config, unet_key, extract_ema... function convert_vd_vae_checkpoint (line 574) | def convert_vd_vae_checkpoint(checkpoint, config): FILE: diffusers/scripts/convert_vq_diffusion_to_diffusers.py function vqvae_model_from_original_config (line 53) | def vqvae_model_from_original_config(original_config): function get_down_block_types (line 104) | def get_down_block_types(original_encoder_config): function get_up_block_types (line 125) | def get_up_block_types(original_decoder_config): function coerce_attn_resolutions (line 146) | def coerce_attn_resolutions(attn_resolutions): function coerce_resolution (line 157) | def coerce_resolution(resolution): function vqvae_original_checkpoint_to_diffusers_checkpoint (line 172) | def vqvae_original_checkpoint_to_diffusers_checkpoint(model, checkpoint): function vqvae_encoder_to_diffusers_checkpoint (line 203) | def vqvae_encoder_to_diffusers_checkpoint(model, checkpoint): function vqvae_decoder_to_diffusers_checkpoint (line 303) | def vqvae_decoder_to_diffusers_checkpoint(model, checkpoint): function vqvae_resnet_to_diffusers_checkpoint (line 406) | def vqvae_resnet_to_diffusers_checkpoint(resnet, checkpoint, *, diffuser... function vqvae_attention_to_diffusers_checkpoint (line 433) | def vqvae_attention_to_diffusers_checkpoint(checkpoint, *, diffusers_att... function transformer_model_from_original_config (line 464) | def transformer_model_from_original_config( function transformer_original_checkpoint_to_diffusers_checkpoint (line 530) | def transformer_original_checkpoint_to_diffusers_checkpoint(model, check... function transformer_ada_norm_to_diffusers_checkpoint (line 646) | def transformer_ada_norm_to_diffusers_checkpoint(checkpoint, *, diffuser... function transformer_attention_to_diffusers_checkpoint (line 654) | def transformer_attention_to_diffusers_checkpoint(checkpoint, *, diffuse... function transformer_feedforward_to_diffusers_checkpoint (line 671) | def transformer_feedforward_to_diffusers_checkpoint(checkpoint, *, diffu... function read_config_file (line 683) | def read_config_file(filename): FILE: diffusers/scripts/convert_zero123_to_diffusers.py function create_unet_diffusers_config (line 38) | def create_unet_diffusers_config(original_config, image_size: int, contr... function assign_to_checkpoint (line 143) | def assign_to_checkpoint( function shave_segments (line 198) | def shave_segments(path, n_shave_prefix_segments=1): function renew_resnet_paths (line 208) | def renew_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_attention_paths (line 230) | def renew_attention_paths(old_list, n_shave_prefix_segments=0): function convert_ldm_unet_checkpoint (line 251) | def convert_ldm_unet_checkpoint( function create_vae_diffusers_config (line 493) | def create_vae_diffusers_config(original_config, image_size: int): function convert_ldm_vae_checkpoint (line 517) | def convert_ldm_vae_checkpoint(checkpoint, config): function renew_vae_resnet_paths (line 624) | def renew_vae_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_vae_attention_paths (line 640) | def renew_vae_attention_paths(old_list, n_shave_prefix_segments=0): function conv_attn_to_linear (line 670) | def conv_attn_to_linear(checkpoint): function convert_from_original_zero123_ckpt (line 682) | def convert_from_original_zero123_ckpt(checkpoint_path, original_config_... FILE: diffusers/scripts/log_reports.py function main (line 17) | def main(slack_channel_name=None): FILE: diffusers/setup.py function deps_list (line 165) | def deps_list(*pkgs): class DepsTableUpdateCommand (line 169) | class DepsTableUpdateCommand(Command): method initialize_options (line 185) | def initialize_options(self): method finalize_options (line 188) | def finalize_options(self): method run (line 191) | def run(self): FILE: diffusers/src/diffusers/callbacks.py class PipelineCallback (line 7) | class PipelineCallback(ConfigMixin): method __init__ (line 22) | def __init__(self, cutoff_step_ratio=1.0, cutoff_step_index=None): method tensor_inputs (line 36) | def tensor_inputs(self) -> List[str]: method callback_fn (line 39) | def callback_fn(self, pipeline, step_index, timesteps, callback_kwargs... method __call__ (line 42) | def __call__(self, pipeline, step_index, timestep, callback_kwargs) ->... class MultiPipelineCallbacks (line 46) | class MultiPipelineCallbacks: method __init__ (line 52) | def __init__(self, callbacks: List[PipelineCallback]): method tensor_inputs (line 56) | def tensor_inputs(self) -> List[str]: method __call__ (line 59) | def __call__(self, pipeline, step_index, timestep, callback_kwargs) ->... class SDCFGCutoffCallback (line 69) | class SDCFGCutoffCallback(PipelineCallback): method callback_fn (line 79) | def callback_fn(self, pipeline, step_index, timestep, callback_kwargs)... class SDXLCFGCutoffCallback (line 98) | class SDXLCFGCutoffCallback(PipelineCallback): method callback_fn (line 108) | def callback_fn(self, pipeline, step_index, timestep, callback_kwargs)... class IPAdapterScaleCutoffCallback (line 135) | class IPAdapterScaleCutoffCallback(PipelineCallback): method callback_fn (line 145) | def callback_fn(self, pipeline, step_index, timestep, callback_kwargs)... FILE: diffusers/src/diffusers/commands/__init__.py class BaseDiffusersCLICommand (line 19) | class BaseDiffusersCLICommand(ABC): method register_subcommand (line 22) | def register_subcommand(parser: ArgumentParser): method run (line 26) | def run(self): FILE: diffusers/src/diffusers/commands/diffusers_cli.py function main (line 22) | def main(): FILE: diffusers/src/diffusers/commands/env.py function info_command_factory (line 37) | def info_command_factory(_): class EnvironmentCommand (line 41) | class EnvironmentCommand(BaseDiffusersCLICommand): method register_subcommand (line 43) | def register_subcommand(parser: ArgumentParser) -> None: method run (line 47) | def run(self) -> dict: method format_dict (line 183) | def format_dict(d: dict) -> str: FILE: diffusers/src/diffusers/commands/fp16_safetensors.py function conversion_command_factory (line 35) | def conversion_command_factory(args: Namespace): class FP16SafetensorsCommand (line 44) | class FP16SafetensorsCommand(BaseDiffusersCLICommand): method register_subcommand (line 46) | def register_subcommand(parser: ArgumentParser): method __init__ (line 66) | def __init__(self, ckpt_id: str, fp16: bool, use_safetensors: bool): method run (line 79) | def run(self): FILE: diffusers/src/diffusers/configuration_utils.py class FrozenDict (line 55) | class FrozenDict(OrderedDict): method __init__ (line 56) | def __init__(self, *args, **kwargs): method __delitem__ (line 64) | def __delitem__(self, *args, **kwargs): method setdefault (line 67) | def setdefault(self, *args, **kwargs): method pop (line 70) | def pop(self, *args, **kwargs): method update (line 73) | def update(self, *args, **kwargs): method __setattr__ (line 76) | def __setattr__(self, name, value): method __setitem__ (line 81) | def __setitem__(self, name, value): class ConfigMixin (line 87) | class ConfigMixin: method register_to_config (line 110) | def register_to_config(self, **kwargs): method __getattr__ (line 127) | def __getattr__(self, name: str) -> Any: method save_config (line 145) | def save_config(self, save_directory: Union[str, os.PathLike], push_to... method from_config (line 188) | def from_config(cls, config: Union[FrozenDict, Dict[str, Any]] = None,... method get_config_dict (line 278) | def get_config_dict(cls, *args, **kwargs): method load_config (line 288) | def load_config( method _get_init_keys (line 457) | def _get_init_keys(input_class): method extract_init_dict (line 461) | def extract_init_dict(cls, config_dict, **kwargs): method _dict_from_json_file (line 557) | def _dict_from_json_file(cls, json_file: Union[str, os.PathLike]): method __repr__ (line 562) | def __repr__(self): method config (line 566) | def config(self) -> Dict[str, Any]: method to_json_string (line 575) | def to_json_string(self) -> str: method to_json_file (line 601) | def to_json_file(self, json_file_path: Union[str, os.PathLike]): function register_to_config (line 613) | def register_to_config(init): function flax_register_to_config (line 663) | def flax_register_to_config(cls): class LegacyConfigMixin (line 711) | class LegacyConfigMixin(ConfigMixin): method from_config (line 718) | def from_config(cls, config: Union[FrozenDict, Dict[str, Any]] = None,... FILE: diffusers/src/diffusers/dependency_versions_check.py function dep_version_check (line 33) | def dep_version_check(pkg, hint=None): FILE: diffusers/src/diffusers/experimental/rl/value_guided_sampling.py class ValueGuidedRLPipeline (line 25) | class ValueGuidedRLPipeline(DiffusionPipeline): method __init__ (line 44) | def __init__( method normalize (line 71) | def normalize(self, x_in, key): method de_normalize (line 74) | def de_normalize(self, x_in, key): method to_torch (line 77) | def to_torch(self, x_in): method reset_x0 (line 84) | def reset_x0(self, x_in, cond, act_dim): method run_diffusion (line 89) | def run_diffusion(self, x, conditions, n_guide_steps, scale): method __call__ (line 122) | def __call__(self, obs, batch_size=64, planning_horizon=32, n_guide_st... FILE: diffusers/src/diffusers/image_processor.py function is_valid_image (line 41) | def is_valid_image(image): function is_valid_image_imagelist (line 45) | def is_valid_image_imagelist(images): class VaeImageProcessor (line 60) | class VaeImageProcessor(ConfigMixin): method __init__ (line 85) | def __init__( method numpy_to_pil (line 105) | def numpy_to_pil(images: np.ndarray) -> List[PIL.Image.Image]: method pil_to_numpy (line 121) | def pil_to_numpy(images: Union[List[PIL.Image.Image], PIL.Image.Image]... method numpy_to_pt (line 133) | def numpy_to_pt(images: np.ndarray) -> torch.Tensor: method pt_to_numpy (line 144) | def pt_to_numpy(images: torch.Tensor) -> np.ndarray: method normalize (line 152) | def normalize(images: Union[np.ndarray, torch.Tensor]) -> Union[np.nda... method denormalize (line 159) | def denormalize(images: Union[np.ndarray, torch.Tensor]) -> Union[np.n... method convert_to_rgb (line 166) | def convert_to_rgb(image: PIL.Image.Image) -> PIL.Image.Image: method convert_to_grayscale (line 175) | def convert_to_grayscale(image: PIL.Image.Image) -> PIL.Image.Image: method blur (line 184) | def blur(image: PIL.Image.Image, blur_factor: int = 4) -> PIL.Image.Im... method get_crop_region (line 193) | def get_crop_region(mask_image: PIL.Image.Image, width: int, height: i... method _resize_and_fill (line 282) | def _resize_and_fill( method _resize_and_crop (line 327) | def _resize_and_crop( method resize (line 353) | def resize( method binarize (line 409) | def binarize(self, image: PIL.Image.Image) -> PIL.Image.Image: method get_default_height_width (line 426) | def get_default_height_width( method preprocess (line 469) | def preprocess( method postprocess (line 596) | def postprocess( method apply_overlay (line 651) | def apply_overlay( class VaeImageProcessorLDM3D (line 687) | class VaeImageProcessorLDM3D(VaeImageProcessor): method __init__ (line 705) | def __init__( method numpy_to_pil (line 715) | def numpy_to_pil(images: np.ndarray) -> List[PIL.Image.Image]: method depth_pil_to_numpy (line 731) | def depth_pil_to_numpy(images: Union[List[PIL.Image.Image], PIL.Image.... method rgblike_to_depthmap (line 743) | def rgblike_to_depthmap(image: Union[np.ndarray, torch.Tensor]) -> Uni... method numpy_to_depth (line 753) | def numpy_to_depth(self, images: np.ndarray) -> List[PIL.Image.Image]: method postprocess (line 773) | def postprocess( method preprocess (line 828) | def preprocess( class IPAdapterMaskProcessor (line 928) | class IPAdapterMaskProcessor(VaeImageProcessor): method __init__ (line 951) | def __init__( method downsample (line 970) | def downsample(mask: torch.Tensor, batch_size: int, num_queries: int, ... class PixArtImageProcessor (line 1032) | class PixArtImageProcessor(VaeImageProcessor): method __init__ (line 1055) | def __init__( method classify_height_width_bin (line 1074) | def classify_height_width_bin(height: int, width: int, ratios: dict) -... method resize_and_crop_tensor (line 1082) | def resize_and_crop_tensor(samples: torch.Tensor, new_width: int, new_... FILE: diffusers/src/diffusers/loaders/__init__.py function text_encoder_lora_state_dict (line 7) | def text_encoder_lora_state_dict(text_encoder): function text_encoder_attn_modules (line 33) | def text_encoder_attn_modules(text_encoder): FILE: diffusers/src/diffusers/loaders/autoencoder.py class FromOriginalVAEMixin (line 23) | class FromOriginalVAEMixin: method from_single_file (line 30) | def from_single_file(cls, pretrained_model_link_or_path, **kwargs): FILE: diffusers/src/diffusers/loaders/controlnet.py class FromOriginalControlNetMixin (line 23) | class FromOriginalControlNetMixin: method from_single_file (line 30) | def from_single_file(cls, pretrained_model_link_or_path, **kwargs): FILE: diffusers/src/diffusers/loaders/ip_adapter.py class IPAdapterMixin (line 51) | class IPAdapterMixin: method load_ip_adapter (line 55) | def load_ip_adapter( method set_ip_adapter_scale (line 248) | def set_ip_adapter_scale(self, scale): method unload_ip_adapter (line 300) | def unload_ip_adapter(self): FILE: diffusers/src/diffusers/loaders/lora.py class LoraLoaderMixin (line 68) | class LoraLoaderMixin: method load_lora_weights (line 79) | def load_lora_weights( method lora_state_dict (line 139) | def lora_state_dict( method _best_guess_weight_name (line 295) | def _best_guess_weight_name( method _optionally_disable_offloading (line 336) | def _optionally_disable_offloading(cls, _pipeline): method load_lora_into_unet (line 371) | def load_lora_into_unet(cls, state_dict, network_alphas, unet, adapter... method load_lora_into_text_encoder (line 407) | def load_lora_into_text_encoder( method load_lora_into_transformer (line 524) | def load_lora_into_transformer(cls, state_dict, network_alphas, transf... method lora_scale (line 605) | def lora_scale(self) -> float: method _remove_text_encoder_monkey_patch (line 610) | def _remove_text_encoder_monkey_patch(self): method save_lora_weights (line 626) | def save_lora_weights( method write_lora_layers (line 691) | def write_lora_layers( method unload_lora_weights (line 724) | def unload_lora_weights(self): method fuse_lora (line 745) | def fuse_lora( method unfuse_lora (line 827) | def unfuse_lora(self, unfuse_unet: bool = True, unfuse_text_encoder: b... method set_adapters_for_text_encoder (line 865) | def set_adapters_for_text_encoder( method disable_lora_for_text_encoder (line 912) | def disable_lora_for_text_encoder(self, text_encoder: Optional["PreTra... method enable_lora_for_text_encoder (line 929) | def enable_lora_for_text_encoder(self, text_encoder: Optional["PreTrai... method set_adapters (line 945) | def set_adapters( method disable_lora (line 1020) | def disable_lora(self): method enable_lora (line 1034) | def enable_lora(self): method delete_adapters (line 1048) | def delete_adapters(self, adapter_names: Union[List[str], str]): method get_active_adapters (line 1072) | def get_active_adapters(self) -> List[str]: method get_list_adapters (line 1104) | def get_list_adapters(self) -> Dict[str, List[str]]: method set_lora_device (line 1127) | def set_lora_device(self, adapter_names: List[str], device: Union[torc... class StableDiffusionXLLoraLoaderMixin (line 1181) | class StableDiffusionXLLoraLoaderMixin(LoraLoaderMixin): method load_lora_weights (line 1185) | def load_lora_weights( method save_lora_weights (line 1263) | def save_lora_weights( method _remove_text_encoder_monkey_patch (line 1329) | def _remove_text_encoder_monkey_patch(self): class SD3LoraLoaderMixin (line 1342) | class SD3LoraLoaderMixin: method load_lora_weights (line 1350) | def load_lora_weights( method lora_state_dict (line 1396) | def lora_state_dict( method load_lora_into_transformer (line 1520) | def load_lora_into_transformer(cls, state_dict, transformer, adapter_n... method save_lora_weights (line 1593) | def save_lora_weights( method write_lora_layers (line 1645) | def write_lora_layers( method unload_lora_weights (line 1678) | def unload_lora_weights(self): method _optionally_disable_offloading (line 1697) | def _optionally_disable_offloading(cls, _pipeline): FILE: diffusers/src/diffusers/loaders/lora_conversion_utils.py function _maybe_map_sgm_blocks_to_diffusers (line 23) | def _maybe_map_sgm_blocks_to_diffusers(state_dict, unet_config, delimite... function _convert_kohya_lora_to_diffusers (line 126) | def _convert_kohya_lora_to_diffusers(state_dict, unet_name="unet", text_... FILE: diffusers/src/diffusers/loaders/peft.py class PeftAdapterMixin (line 20) | class PeftAdapterMixin: method add_adapter (line 36) | def add_adapter(self, adapter_config, adapter_name: str = "default") -... method set_adapter (line 74) | def set_adapter(self, adapter_name: Union[str, List[str]]) -> None: method disable_adapters (line 123) | def disable_adapters(self) -> None: method enable_adapters (line 145) | def enable_adapters(self) -> None: method active_adapters (line 168) | def active_adapters(self) -> List[str]: FILE: diffusers/src/diffusers/loaders/single_file.py function load_single_file_sub_model (line 51) | def load_single_file_sub_model( function _map_component_types_to_config_dict (line 178) | def _map_component_types_to_config_dict(component_types): function _infer_pipeline_config_dict (line 227) | def _infer_pipeline_config_dict(pipeline_class): function _download_diffusers_model_config_from_hub (line 239) | def _download_diffusers_model_config_from_hub( class FromSingleFileMixin (line 265) | class FromSingleFileMixin: method from_single_file (line 272) | def from_single_file(cls, pretrained_model_link_or_path, **kwargs): FILE: diffusers/src/diffusers/loaders/single_file_model.py function _get_mapping_function_kwargs (line 75) | def _get_mapping_function_kwargs(mapping_fn, **kwargs): class FromOriginalModelMixin (line 86) | class FromOriginalModelMixin: method from_single_file (line 93) | def from_single_file(cls, pretrained_model_link_or_path_or_dict: Optio... FILE: diffusers/src/diffusers/loaders/single_file_utils.py class SingleFileComponentError (line 262) | class SingleFileComponentError(Exception): method __init__ (line 263) | def __init__(self, message=None): function is_valid_url (line 268) | def is_valid_url(url): function _extract_repo_id_and_weights_name (line 276) | def _extract_repo_id_and_weights_name(pretrained_model_name_or_path): function _is_model_weights_in_cached_folder (line 296) | def _is_model_weights_in_cached_folder(cached_folder, name): function load_single_file_checkpoint (line 307) | def load_single_file_checkpoint( function fetch_original_config (line 343) | def fetch_original_config(original_config_file, local_files_only=False): function is_clip_model (line 365) | def is_clip_model(checkpoint): function is_clip_sdxl_model (line 372) | def is_clip_sdxl_model(checkpoint): function is_clip_sd3_model (line 379) | def is_clip_sd3_model(checkpoint): function is_open_clip_model (line 386) | def is_open_clip_model(checkpoint): function is_open_clip_sdxl_model (line 393) | def is_open_clip_sdxl_model(checkpoint): function is_open_clip_sd3_model (line 400) | def is_open_clip_sd3_model(checkpoint): function is_open_clip_sdxl_refiner_model (line 407) | def is_open_clip_sdxl_refiner_model(checkpoint): function is_clip_model_in_single_file (line 414) | def is_clip_model_in_single_file(class_obj, checkpoint): function infer_diffusers_model_type (line 433) | def infer_diffusers_model_type(checkpoint): function fetch_diffusers_config (line 494) | def fetch_diffusers_config(checkpoint): function set_image_size (line 501) | def set_image_size(checkpoint, image_size=None): function conv_attn_to_linear (line 512) | def conv_attn_to_linear(checkpoint): function create_unet_diffusers_config_from_ldm (line 524) | def create_unet_diffusers_config_from_ldm( function create_controlnet_diffusers_config_from_ldm (line 655) | def create_controlnet_diffusers_config_from_ldm(original_config, checkpo... function create_vae_diffusers_config_from_ldm (line 687) | def create_vae_diffusers_config_from_ldm(original_config, checkpoint, im... function update_unet_resnet_ldm_to_diffusers (line 738) | def update_unet_resnet_ldm_to_diffusers(ldm_keys, new_checkpoint, checkp... function update_unet_attention_ldm_to_diffusers (line 753) | def update_unet_attention_ldm_to_diffusers(ldm_keys, new_checkpoint, che... function update_vae_resnet_ldm_to_diffusers (line 759) | def update_vae_resnet_ldm_to_diffusers(keys, new_checkpoint, checkpoint,... function update_vae_attentions_ldm_to_diffusers (line 765) | def update_vae_attentions_ldm_to_diffusers(keys, new_checkpoint, checkpo... function convert_stable_cascade_unet_single_file_to_diffusers (line 791) | def convert_stable_cascade_unet_single_file_to_diffusers(checkpoint, **k... function convert_ldm_unet_checkpoint (line 847) | def convert_ldm_unet_checkpoint(checkpoint, config, extract_ema=False, *... function convert_controlnet_checkpoint (line 1014) | def convert_controlnet_checkpoint( function convert_ldm_vae_checkpoint (line 1138) | def convert_ldm_vae_checkpoint(checkpoint, config): function convert_ldm_clip_checkpoint (line 1238) | def convert_ldm_clip_checkpoint(checkpoint, remove_prefix=None): function convert_open_clip_checkpoint (line 1256) | def convert_open_clip_checkpoint( function create_diffusers_clip_model_from_ldm (line 1327) | def create_diffusers_clip_model_from_ldm( function _legacy_load_scheduler (line 1437) | def _legacy_load_scheduler( function _legacy_load_clip_tokenizer (line 1567) | def _legacy_load_clip_tokenizer(cls, checkpoint, config=None, local_file... function _legacy_load_safety_checker (line 1593) | def _legacy_load_safety_checker(local_files_only, torch_dtype): function swap_scale_shift (line 1611) | def swap_scale_shift(weight, dim): function convert_sd3_transformer_checkpoint_to_diffusers (line 1617) | def convert_sd3_transformer_checkpoint_to_diffusers(checkpoint, **kwargs): function is_t5_in_single_file (line 1761) | def is_t5_in_single_file(checkpoint): function convert_sd3_t5_checkpoint_to_diffusers (line 1768) | def convert_sd3_t5_checkpoint_to_diffusers(checkpoint): function create_diffusers_t5_model_from_checkpoint (line 1783) | def create_diffusers_t5_model_from_checkpoint( FILE: diffusers/src/diffusers/loaders/textual_inversion.py function load_textual_inversion_state_dicts (line 38) | def load_textual_inversion_state_dicts(pretrained_model_name_or_paths, *... class TextualInversionLoaderMixin (line 113) | class TextualInversionLoaderMixin: method maybe_convert_prompt (line 118) | def maybe_convert_prompt(self, prompt: Union[str, List[str]], tokenize... method _maybe_convert_prompt (line 145) | def _maybe_convert_prompt(self, prompt: str, tokenizer: "PreTrainedTok... method _check_text_inv_inputs (line 175) | def _check_text_inv_inputs(self, tokenizer, text_encoder, pretrained_m... method _retrieve_tokens_and_embeddings (line 199) | def _retrieve_tokens_and_embeddings(tokens, state_dicts, tokenizer): method _extend_tokens_and_embeddings (line 240) | def _extend_tokens_and_embeddings(tokens, embeddings, tokenizer): method load_textual_inversion (line 267) | def load_textual_inversion( method unload_textual_inversion (line 463) | def unload_textual_inversion( FILE: diffusers/src/diffusers/loaders/unet.py class UNet2DConditionLoadersMixin (line 66) | class UNet2DConditionLoadersMixin: method load_attn_procs (line 75) | def load_attn_procs(self, pretrained_model_name_or_path_or_dict: Union... method _process_custom_diffusion (line 248) | def _process_custom_diffusion(self, state_dict): method _process_lora (line 282) | def _process_lora(self, state_dict, unet_identifier_key, network_alpha... method _optionally_disable_offloading (line 366) | def _optionally_disable_offloading(cls, _pipeline): method save_attn_procs (line 400) | def save_attn_procs( method _get_custom_diffusion_state_dict (line 490) | def _get_custom_diffusion_state_dict(self): method fuse_lora (line 518) | def fuse_lora(self, lora_scale=1.0, safe_fusing=False, adapter_names=N... method _fuse_lora_apply (line 526) | def _fuse_lora_apply(self, module, adapter_names=None): method unfuse_lora (line 548) | def unfuse_lora(self): method _unfuse_lora_apply (line 553) | def _unfuse_lora_apply(self, module): method unload_lora (line 559) | def unload_lora(self): method set_adapters (line 569) | def set_adapters( method disable_lora (line 624) | def disable_lora(self): method enable_lora (line 647) | def enable_lora(self): method delete_adapters (line 670) | def delete_adapters(self, adapter_names: Union[List[str], str]): method _convert_ip_adapter_image_proj_to_diffusers (line 706) | def _convert_ip_adapter_image_proj_to_diffusers(self, state_dict, low_... method _convert_ip_adapter_attn_to_diffusers (line 923) | def _convert_ip_adapter_attn_to_diffusers(self, state_dicts, low_cpu_m... method _load_ip_adapter_weights (line 1017) | def _load_ip_adapter_weights(self, state_dicts, low_cpu_mem_usage=False): method _load_ip_adapter_loras (line 1040) | def _load_ip_adapter_loras(self, state_dicts): FILE: diffusers/src/diffusers/loaders/unet_loader_utils.py function _translate_into_actual_layer_name (line 27) | def _translate_into_actual_layer_name(name): function _maybe_expand_lora_scales (line 41) | def _maybe_expand_lora_scales( function _maybe_expand_lora_scales_for_one_adapter (line 64) | def _maybe_expand_lora_scales_for_one_adapter( FILE: diffusers/src/diffusers/loaders/utils.py class AttnProcsLayers (line 20) | class AttnProcsLayers(torch.nn.Module): method __init__ (line 21) | def __init__(self, state_dict: Dict[str, torch.Tensor]): FILE: diffusers/src/diffusers/models/activations.py function get_activation (line 36) | def get_activation(act_fn: str) -> nn.Module: class FP32SiLU (line 53) | class FP32SiLU(nn.Module): method __init__ (line 58) | def __init__(self): method forward (line 61) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: class GELU (line 65) | class GELU(nn.Module): method __init__ (line 76) | def __init__(self, dim_in: int, dim_out: int, approximate: str = "none... method gelu (line 81) | def gelu(self, gate: torch.Tensor) -> torch.Tensor: method forward (line 87) | def forward(self, hidden_states): class GEGLU (line 93) | class GEGLU(nn.Module): method __init__ (line 103) | def __init__(self, dim_in: int, dim_out: int, bias: bool = True): method gelu (line 107) | def gelu(self, gate: torch.Tensor) -> torch.Tensor: method forward (line 113) | def forward(self, hidden_states, *args, **kwargs): class ApproximateGELU (line 126) | class ApproximateGELU(nn.Module): method __init__ (line 137) | def __init__(self, dim_in: int, dim_out: int, bias: bool = True): method forward (line 141) | def forward(self, x: torch.Tensor) -> torch.Tensor: FILE: diffusers/src/diffusers/models/adapter.py class MultiAdapter (line 28) | class MultiAdapter(ModelMixin): method __init__ (line 41) | def __init__(self, adapters: List["T2IAdapter"]): method forward (line 76) | def forward(self, xs: torch.Tensor, adapter_weights: Optional[List[flo... method save_pretrained (line 103) | def save_pretrained( method from_pretrained (line 146) | def from_pretrained(cls, pretrained_model_path: Optional[Union[str, os... class T2IAdapter (line 217) | class T2IAdapter(ModelMixin, ConfigMixin): method __init__ (line 245) | def __init__( method forward (line 267) | def forward(self, x: torch.Tensor) -> List[torch.Tensor]: method total_downscale_factor (line 277) | def total_downscale_factor(self): method downscale_factor (line 281) | def downscale_factor(self): class FullAdapter (line 291) | class FullAdapter(nn.Module): method __init__ (line 296) | def __init__( method forward (line 322) | def forward(self, x: torch.Tensor) -> List[torch.Tensor]: class FullAdapterXL (line 341) | class FullAdapterXL(nn.Module): method __init__ (line 346) | def __init__( method forward (line 374) | def forward(self, x: torch.Tensor) -> List[torch.Tensor]: class AdapterBlock (line 391) | class AdapterBlock(nn.Module): method __init__ (line 407) | def __init__(self, in_channels: int, out_channels: int, num_res_blocks... method forward (line 422) | def forward(self, x: torch.Tensor) -> torch.Tensor: class AdapterResnetBlock (line 439) | class AdapterResnetBlock(nn.Module): method __init__ (line 448) | def __init__(self, channels: int): method forward (line 454) | def forward(self, x: torch.Tensor) -> torch.Tensor: class LightAdapter (line 469) | class LightAdapter(nn.Module): method __init__ (line 474) | def __init__( method forward (line 500) | def forward(self, x: torch.Tensor) -> List[torch.Tensor]: class LightAdapterBlock (line 516) | class LightAdapterBlock(nn.Module): method __init__ (line 532) | def __init__(self, in_channels: int, out_channels: int, num_res_blocks... method forward (line 544) | def forward(self, x: torch.Tensor) -> torch.Tensor: class LightAdapterResnetBlock (line 559) | class LightAdapterResnetBlock(nn.Module): method __init__ (line 569) | def __init__(self, channels: int): method forward (line 575) | def forward(self, x: torch.Tensor) -> torch.Tensor: FILE: diffusers/src/diffusers/models/attention.py function _chunked_feed_forward (line 31) | def _chunked_feed_forward(ff: nn.Module, hidden_states: torch.Tensor, ch... class GatedSelfAttentionDense (line 47) | class GatedSelfAttentionDense(nn.Module): method __init__ (line 58) | def __init__(self, query_dim: int, context_dim: int, n_heads: int, d_h... method forward (line 75) | def forward(self, x: torch.Tensor, objs: torch.Tensor) -> torch.Tensor: class JointTransformerBlock (line 89) | class JointTransformerBlock(nn.Module): method __init__ (line 103) | def __init__(self, dim, num_attention_heads, attention_head_dim, conte... method set_chunk_feed_forward (line 154) | def set_chunk_feed_forward(self, chunk_size: Optional[int], dim: int =... method forward (line 159) | def forward( class BasicTransformerBlock (line 213) | class BasicTransformerBlock(nn.Module): method __init__ (line 248) | def __init__( method set_chunk_feed_forward (line 402) | def set_chunk_feed_forward(self, chunk_size: Optional[int], dim: int =... method forward (line 407) | def forward( class TemporalBasicTransformerBlock (line 531) | class TemporalBasicTransformerBlock(nn.Module): method __init__ (line 543) | def __init__( method set_chunk_feed_forward (line 596) | def set_chunk_feed_forward(self, chunk_size: Optional[int], **kwargs): method forward (line 602) | def forward( class SkipFFTransformerBlock (line 660) | class SkipFFTransformerBlock(nn.Module): method __init__ (line 661) | def __init__( method forward (line 703) | def forward(self, hidden_states, encoder_hidden_states, cross_attentio... class FeedForward (line 732) | class FeedForward(nn.Module): method __init__ (line 746) | def __init__( method forward (line 782) | def forward(self, hidden_states: torch.Tensor, *args, **kwargs) -> tor... FILE: diffusers/src/diffusers/models/attention_flax.py function _query_chunk_attention (line 23) | def _query_chunk_attention(query, key, value, precision, key_chunk_size:... function jax_memory_efficient_attention (line 74) | def jax_memory_efficient_attention( class FlaxAttention (line 122) | class FlaxAttention(nn.Module): method setup (line 153) | def setup(self): method reshape_heads_to_batch_dim (line 165) | def reshape_heads_to_batch_dim(self, tensor): method reshape_batch_dim_to_heads (line 173) | def reshape_batch_dim_to_heads(self, tensor): method __call__ (line 181) | def __call__(self, hidden_states, context=None, deterministic=True): class FlaxBasicTransformerBlock (line 244) | class FlaxBasicTransformerBlock(nn.Module): method setup (line 279) | def setup(self): method __call__ (line 306) | def __call__(self, hidden_states, context, deterministic=True): class FlaxTransformer2DModel (line 328) | class FlaxTransformer2DModel(nn.Module): method setup (line 367) | def setup(self): method __call__ (line 409) | def __call__(self, hidden_states, context, deterministic=True): class FlaxFeedForward (line 434) | class FlaxFeedForward(nn.Module): method setup (line 456) | def setup(self): method __call__ (line 462) | def __call__(self, hidden_states, deterministic=True): class FlaxGEGLU (line 468) | class FlaxGEGLU(nn.Module): method setup (line 486) | def setup(self): method __call__ (line 491) | def __call__(self, hidden_states, deterministic=True): FILE: diffusers/src/diffusers/models/attention_processor.py class Attention (line 43) | class Attention(nn.Module): method __init__ (line 94) | def __init__( method set_use_npu_flash_attention (line 232) | def set_use_npu_flash_attention(self, use_npu_flash_attention: bool) -... method set_use_memory_efficient_attention_xformers (line 249) | def set_use_memory_efficient_attention_xformers( method set_attention_slice (line 383) | def set_attention_slice(self, slice_size: int) -> None: method set_processor (line 411) | def set_processor(self, processor: "AttnProcessor") -> None: method get_processor (line 431) | def get_processor(self, return_deprecated_lora: bool = False) -> "Atte... method forward (line 521) | def forward( method batch_to_head_dim (line 567) | def batch_to_head_dim(self, tensor: torch.Tensor) -> torch.Tensor: method head_to_batch_dim (line 584) | def head_to_batch_dim(self, tensor: torch.Tensor, out_dim: int = 3) ->... method get_attention_scores (line 611) | def get_attention_scores( method prepare_attention_mask (line 658) | def prepare_attention_mask( method norm_encoder_hidden_states (line 705) | def norm_encoder_hidden_states(self, encoder_hidden_states: torch.Tens... method fuse_projections (line 735) | def fuse_projections(self, fuse=True): class AttnProcessor (line 766) | class AttnProcessor: method __call__ (line 771) | def __call__( class CustomDiffusionAttnProcessor (line 838) | class CustomDiffusionAttnProcessor(nn.Module): method __init__ (line 857) | def __init__( method __call__ (line 883) | def __call__( class AttnAddedKVProcessor (line 942) | class AttnAddedKVProcessor: method __call__ (line 948) | def __call__( class AttnAddedKVProcessor2_0 (line 1009) | class AttnAddedKVProcessor2_0: method __init__ (line 1015) | def __init__(self): method __call__ (line 1021) | def __call__( class JointAttnProcessor2_0 (line 1085) | class JointAttnProcessor2_0: method __init__ (line 1088) | def __init__(self): method __call__ (line 1092) | def __call__( class FusedJointAttnProcessor2_0 (line 1160) | class FusedJointAttnProcessor2_0: method __init__ (line 1163) | def __init__(self): method __call__ (line 1167) | def __call__( class XFormersAttnAddedKVProcessor (line 1241) | class XFormersAttnAddedKVProcessor: method __init__ (line 1253) | def __init__(self, attention_op: Optional[Callable] = None): method __call__ (line 1256) | def __call__( class XFormersAttnProcessor (line 1312) | class XFormersAttnProcessor: method __init__ (line 1324) | def __init__(self, attention_op: Optional[Callable] = None): method __call__ (line 1327) | def __call__( class AttnProcessorNPU (line 1406) | class AttnProcessorNPU: method __init__ (line 1414) | def __init__(self): method __call__ (line 1418) | def __call__( class AttnProcessor2_0 (line 1515) | class AttnProcessor2_0: method __init__ (line 1520) | def __init__(self): method __call__ (line 1524) | def __call__( class HunyuanAttnProcessor2_0 (line 1604) | class HunyuanAttnProcessor2_0: method __init__ (line 1610) | def __init__(self): method __call__ (line 1614) | def __call__( class FusedAttnProcessor2_0 (line 1702) | class FusedAttnProcessor2_0: method __init__ (line 1715) | def __init__(self): method __call__ (line 1721) | def __call__( class CustomDiffusionXFormersAttnProcessor (line 1803) | class CustomDiffusionXFormersAttnProcessor(nn.Module): method __init__ (line 1826) | def __init__( method __call__ (line 1854) | def __call__( class CustomDiffusionAttnProcessor2_0 (line 1919) | class CustomDiffusionAttnProcessor2_0(nn.Module): method __init__ (line 1939) | def __init__( method __call__ (line 1965) | def __call__( class SlicedAttnProcessor (line 2033) | class SlicedAttnProcessor: method __init__ (line 2043) | def __init__(self, slice_size: int): method __call__ (line 2046) | def __call__( class SlicedAttnAddedKVProcessor (line 2120) | class SlicedAttnAddedKVProcessor: method __init__ (line 2130) | def __init__(self, slice_size): method __call__ (line 2133) | def __call__( class SpatialNorm (line 2212) | class SpatialNorm(nn.Module): method __init__ (line 2223) | def __init__( method forward (line 2233) | def forward(self, f: torch.Tensor, zq: torch.Tensor) -> torch.Tensor: class LoRAAttnProcessor (line 2241) | class LoRAAttnProcessor(nn.Module): method __init__ (line 2242) | def __init__( method __call__ (line 2279) | def __call__(self, attn: Attention, hidden_states: torch.Tensor, **kwa... class LoRAAttnProcessor2_0 (line 2300) | class LoRAAttnProcessor2_0(nn.Module): method __init__ (line 2301) | def __init__( method __call__ (line 2340) | def __call__(self, attn: Attention, hidden_states: torch.Tensor, **kwa... class LoRAXFormersAttnProcessor (line 2361) | class LoRAXFormersAttnProcessor(nn.Module): method __init__ (line 2383) | def __init__( method __call__ (line 2419) | def __call__(self, attn: Attention, hidden_states: torch.Tensor, **kwa... class LoRAAttnAddedKVProcessor (line 2440) | class LoRAAttnAddedKVProcessor(nn.Module): method __init__ (line 2458) | def __init__( method __call__ (line 2478) | def __call__(self, attn: Attention, hidden_states: torch.Tensor, **kwa... class IPAdapterAttnProcessor (line 2499) | class IPAdapterAttnProcessor(nn.Module): method __init__ (line 2514) | def __init__(self, hidden_size, cross_attention_dim=None, num_tokens=(... method __call__ (line 2537) | def __call__( class IPAdapterAttnProcessor2_0 (line 2697) | class IPAdapterAttnProcessor2_0(torch.nn.Module): method __init__ (line 2712) | def __init__(self, hidden_size, cross_attention_dim=None, num_tokens=(... method __call__ (line 2740) | def __call__( FILE: diffusers/src/diffusers/models/autoencoders/autoencoder_asym_kl.py class AsymmetricAutoencoderKL (line 26) | class AsymmetricAutoencoderKL(ModelMixin, ConfigMixin): method __init__ (line 64) | def __init__( method encode (line 115) | def encode(self, x: torch.Tensor, return_dict: bool = True) -> Union[A... method _decode (line 125) | def _decode( method decode (line 141) | def decode( method forward (line 156) | def forward( FILE: diffusers/src/diffusers/models/autoencoders/autoencoder_kl.py class AutoencoderKL (line 35) | class AutoencoderKL(ModelMixin, ConfigMixin, FromOriginalModelMixin): method __init__ (line 71) | def __init__( method _set_gradient_checkpointing (line 132) | def _set_gradient_checkpointing(self, module, value=False): method enable_tiling (line 136) | def enable_tiling(self, use_tiling: bool = True): method disable_tiling (line 144) | def disable_tiling(self): method enable_slicing (line 151) | def enable_slicing(self): method disable_slicing (line 158) | def disable_slicing(self): method attn_processors (line 167) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 191) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 226) | def set_default_attn_processor(self): method encode (line 242) | def encode( method _decode (line 278) | def _decode(self, z: torch.Tensor, return_dict: bool = True) -> Union[... method decode (line 293) | def decode( method blend_v (line 321) | def blend_v(self, a: torch.Tensor, b: torch.Tensor, blend_extent: int)... method blend_h (line 327) | def blend_h(self, a: torch.Tensor, b: torch.Tensor, blend_extent: int)... method tiled_encode (line 333) | def tiled_encode(self, x: torch.Tensor, return_dict: bool = True) -> A... method tiled_decode (line 387) | def tiled_decode(self, z: torch.Tensor, return_dict: bool = True) -> U... method forward (line 435) | def forward( method fuse_qkv_projections (line 464) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 488) | def unfuse_qkv_projections(self): FILE: diffusers/src/diffusers/models/autoencoders/autoencoder_kl_temporal_decoder.py class TemporalDecoder (line 29) | class TemporalDecoder(nn.Module): method __init__ (line 30) | def __init__( method forward (line 87) | def forward( class AutoencoderKLTemporalDecoder (line 164) | class AutoencoderKLTemporalDecoder(ModelMixin, ConfigMixin): method __init__ (line 197) | def __init__( method _set_gradient_checkpointing (line 239) | def _set_gradient_checkpointing(self, module, value=False): method attn_processors (line 245) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 269) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 303) | def set_default_attn_processor(self): method encode (line 317) | def encode( method decode (line 344) | def decode( method forward (line 373) | def forward( FILE: diffusers/src/diffusers/models/autoencoders/autoencoder_tiny.py class AutoencoderTinyOutput (line 29) | class AutoencoderTinyOutput(BaseOutput): class AutoencoderTiny (line 41) | class AutoencoderTiny(ModelMixin, ConfigMixin): method __init__ (line 98) | def __init__( method _set_gradient_checkpointing (line 156) | def _set_gradient_checkpointing(self, module, value: bool = False) -> ... method scale_latents (line 160) | def scale_latents(self, x: torch.Tensor) -> torch.Tensor: method unscale_latents (line 164) | def unscale_latents(self, x: torch.Tensor) -> torch.Tensor: method enable_slicing (line 168) | def enable_slicing(self) -> None: method disable_slicing (line 175) | def disable_slicing(self) -> None: method enable_tiling (line 182) | def enable_tiling(self, use_tiling: bool = True) -> None: method disable_tiling (line 190) | def disable_tiling(self) -> None: method _tiled_encode (line 197) | def _tiled_encode(self, x: torch.Tensor) -> torch.Tensor: method _tiled_decode (line 245) | def _tiled_decode(self, x: torch.Tensor) -> torch.Tensor: method encode (line 293) | def encode(self, x: torch.Tensor, return_dict: bool = True) -> Union[A... method decode (line 308) | def decode( method forward (line 322) | def forward( FILE: diffusers/src/diffusers/models/autoencoders/consistency_decoder_vae.py class ConsistencyDecoderVAEOutput (line 39) | class ConsistencyDecoderVAEOutput(BaseOutput): class ConsistencyDecoderVAE (line 52) | class ConsistencyDecoderVAE(ModelMixin, ConfigMixin): method __init__ (line 72) | def __init__( method enable_tiling (line 169) | def enable_tiling(self, use_tiling: bool = True): method disable_tiling (line 178) | def disable_tiling(self): method enable_slicing (line 186) | def enable_slicing(self): method disable_slicing (line 194) | def disable_slicing(self): method attn_processors (line 203) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 227) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 262) | def set_default_attn_processor(self): method encode (line 278) | def encode( method decode (line 313) | def decode( method blend_v (line 360) | def blend_v(self, a: torch.Tensor, b: torch.Tensor, blend_extent: int)... method blend_h (line 367) | def blend_h(self, a: torch.Tensor, b: torch.Tensor, blend_extent: int)... method tiled_encode (line 373) | def tiled_encode(self, x: torch.Tensor, return_dict: bool = True) -> U... method forward (line 428) | def forward( FILE: diffusers/src/diffusers/models/autoencoders/vae.py class DecoderOutput (line 34) | class DecoderOutput(BaseOutput): class Encoder (line 47) | class Encoder(nn.Module): method __init__ (line 71) | def __init__( method forward (line 140) | def forward(self, sample: torch.Tensor) -> torch.Tensor: class Decoder (line 185) | class Decoder(nn.Module): method __init__ (line 208) | def __init__( method forward (line 284) | def forward( class UpSample (line 350) | class UpSample(nn.Module): method __init__ (line 361) | def __init__( method forward (line 371) | def forward(self, x: torch.Tensor) -> torch.Tensor: class MaskConditionEncoder (line 378) | class MaskConditionEncoder(nn.Module): method __init__ (line 383) | def __init__( method forward (line 420) | def forward(self, x: torch.Tensor, mask=None) -> torch.Tensor: class MaskConditionDecoder (line 431) | class MaskConditionDecoder(nn.Module): method __init__ (line 454) | def __init__( method forward (line 535) | def forward( class VectorQuantizer (line 645) | class VectorQuantizer(nn.Module): method __init__ (line 654) | def __init__( method remap_to_used (line 691) | def remap_to_used(self, inds: torch.LongTensor) -> torch.LongTensor: method unmap_to_all (line 705) | def unmap_to_all(self, inds: torch.LongTensor) -> torch.LongTensor: method forward (line 715) | def forward(self, z: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor... method get_codebook_entry (line 749) | def get_codebook_entry(self, indices: torch.LongTensor, shape: Tuple[i... class DiagonalGaussianDistribution (line 767) | class DiagonalGaussianDistribution(object): method __init__ (line 768) | def __init__(self, parameters: torch.Tensor, deterministic: bool = Fal... method sample (line 780) | def sample(self, generator: Optional[torch.Generator] = None) -> torch... method kl (line 791) | def kl(self, other: "DiagonalGaussianDistribution" = None) -> torch.Te... method nll (line 810) | def nll(self, sample: torch.Tensor, dims: Tuple[int, ...] = [1, 2, 3])... method mode (line 819) | def mode(self) -> torch.Tensor: class EncoderTiny (line 823) | class EncoderTiny(nn.Module): method __init__ (line 841) | def __init__( method forward (line 877) | def forward(self, x: torch.Tensor) -> torch.Tensor: class DecoderTiny (line 899) | class DecoderTiny(nn.Module): method __init__ (line 919) | def __init__( method forward (line 960) | def forward(self, x: torch.Tensor) -> torch.Tensor: FILE: diffusers/src/diffusers/models/autoencoders/vq_model.py class VQEncoderOutput (line 28) | class VQEncoderOutput(BaseOutput): class VQModel (line 40) | class VQModel(ModelMixin, ConfigMixin): method __init__ (line 75) | def __init__( method encode (line 130) | def encode(self, x: torch.Tensor, return_dict: bool = True) -> VQEncod... method decode (line 140) | def decode( method forward (line 160) | def forward( FILE: diffusers/src/diffusers/models/controlnet.py class ControlNetOutput (line 47) | class ControlNetOutput(BaseOutput): class ControlNetConditioningEmbedding (line 66) | class ControlNetConditioningEmbedding(nn.Module): method __init__ (line 76) | def __init__( method forward (line 98) | def forward(self, conditioning): class ControlNetModel (line 111) | class ControlNetModel(ModelMixin, ConfigMixin, FromOriginalModelMixin): method __init__ (line 183) | def __init__( method from_unet (line 445) | def from_unet( method attn_processors (line 522) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 546) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 581) | def set_default_attn_processor(self): method set_attention_slice (line 597) | def set_attention_slice(self, slice_size: Union[str, int, List[int]]) ... method _set_gradient_checkpointing (line 662) | def _set_gradient_checkpointing(self, module, value: bool = False) -> ... method forward (line 666) | def forward( function zero_module (line 868) | def zero_module(module): FILE: diffusers/src/diffusers/models/controlnet_flax.py class FlaxControlNetOutput (line 34) | class FlaxControlNetOutput(BaseOutput): class FlaxControlNetConditioningEmbedding (line 47) | class FlaxControlNetConditioningEmbedding(nn.Module): method setup (line 52) | def setup(self) -> None: method __call__ (line 90) | def __call__(self, conditioning: jnp.ndarray) -> jnp.ndarray: class FlaxControlNetModel (line 104) | class FlaxControlNetModel(nn.Module, FlaxModelMixin, ConfigMixin): method init_weights (line 172) | def init_weights(self, rng: jax.Array) -> FrozenDict: method setup (line 186) | def setup(self) -> None: method __call__ (line 314) | def __call__( FILE: diffusers/src/diffusers/models/controlnet_xs.py class ControlNetXSOutput (line 52) | class ControlNetXSOutput(BaseOutput): class DownBlockControlNetXSAdapter (line 65) | class DownBlockControlNetXSAdapter(nn.Module): method __init__ (line 69) | def __init__( class MidBlockControlNetXSAdapter (line 85) | class MidBlockControlNetXSAdapter(nn.Module): method __init__ (line 89) | def __init__(self, midblock: UNetMidBlock2DCrossAttn, base_to_ctrl: nn... class UpBlockControlNetXSAdapter (line 96) | class UpBlockControlNetXSAdapter(nn.Module): method __init__ (line 99) | def __init__(self, ctrl_to_base: nn.ModuleList): function get_down_block_adapter (line 104) | def get_down_block_adapter( function get_mid_block_adapter (line 194) | def get_mid_block_adapter( function get_up_block_adapter (line 228) | def get_up_block_adapter( class ControlNetXSAdapter (line 242) | class ControlNetXSAdapter(ModelMixin, ConfigMixin): method __init__ (line 290) | def __init__( method from_unet (line 424) | def from_unet( method forward (line 499) | def forward(self, *args, **kwargs): class UNetControlNetXSModel (line 505) | class UNetControlNetXSModel(ModelMixin, ConfigMixin): method __init__ (line 522) | def __init__( method from_unet (line 704) | def from_unet( method freeze_unet_params (line 821) | def freeze_unet_params(self) -> None: method _set_gradient_checkpointing (line 850) | def _set_gradient_checkpointing(self, module, value=False): method attn_processors (line 856) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 880) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 915) | def set_default_attn_processor(self): method enable_freeu (line 931) | def enable_freeu(self, s1: float, s2: float, b1: float, b2: float): method disable_freeu (line 956) | def disable_freeu(self): method fuse_qkv_projections (line 965) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 989) | def unfuse_qkv_projections(self): method forward (line 1002) | def forward( class ControlNetXSCrossAttnDownBlock2D (line 1205) | class ControlNetXSCrossAttnDownBlock2D(nn.Module): method __init__ (line 1206) | def __init__( method from_modules (line 1324) | def from_modules(cls, base_downblock: CrossAttnDownBlock2D, ctrl_downb... method freeze_base_params (line 1386) | def freeze_base_params(self) -> None: method forward (line 1403) | def forward( class ControlNetXSCrossAttnMidBlock2D (line 1517) | class ControlNetXSCrossAttnMidBlock2D(nn.Module): method __init__ (line 1518) | def __init__( method from_modules (line 1570) | def from_modules( method freeze_base_params (line 1616) | def freeze_base_params(self) -> None: method forward (line 1627) | def forward( class ControlNetXSCrossAttnUpBlock2D (line 1664) | class ControlNetXSCrossAttnUpBlock2D(nn.Module): method __init__ (line 1665) | def __init__( method from_modules (line 1736) | def from_modules(cls, base_upblock: CrossAttnUpBlock2D, ctrl_upblock: ... method freeze_base_params (line 1791) | def freeze_base_params(self) -> None: method forward (line 1808) | def forward( function make_zero_conv (line 1897) | def make_zero_conv(in_channels, out_channels=None): function zero_module (line 1901) | def zero_module(module): function find_largest_factor (line 1907) | def find_largest_factor(number, max_factor): FILE: diffusers/src/diffusers/models/downsampling.py class Downsample1D (line 26) | class Downsample1D(nn.Module): method __init__ (line 42) | def __init__( method forward (line 64) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: class Downsample2D (line 69) | class Downsample2D(nn.Module): method __init__ (line 85) | def __init__( method forward (line 132) | def forward(self, hidden_states: torch.Tensor, *args, **kwargs) -> tor... class FirDownsample2D (line 152) | class FirDownsample2D(nn.Module): method __init__ (line 166) | def __init__( method _downsample_2d (line 181) | def _downsample_2d( method forward (line 247) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class KDownsample2D (line 258) | class KDownsample2D(nn.Module): method __init__ (line 265) | def __init__(self, pad_mode: str = "reflect"): method forward (line 272) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: function downsample_2d (line 288) | def downsample_2d( FILE: diffusers/src/diffusers/models/embeddings.py function get_timestep_embedding (line 27) | def get_timestep_embedding( function get_2d_sincos_pos_embed (line 70) | def get_2d_sincos_pos_embed( function get_2d_sincos_pos_embed_from_grid (line 92) | def get_2d_sincos_pos_embed_from_grid(embed_dim, grid): function get_1d_sincos_pos_embed_from_grid (line 104) | def get_1d_sincos_pos_embed_from_grid(embed_dim, pos): class PatchEmbed (line 125) | class PatchEmbed(nn.Module): method __init__ (line 128) | def __init__( method cropped_pos_embed (line 179) | def cropped_pos_embed(self, height, width): method forward (line 202) | def forward(self, latent): function get_2d_rotary_pos_embed (line 233) | def get_2d_rotary_pos_embed(embed_dim, crops_coords, grid_size, use_real... function get_2d_rotary_pos_embed_from_grid (line 261) | def get_2d_rotary_pos_embed_from_grid(embed_dim, grid, use_real=False): function get_1d_rotary_pos_embed (line 277) | def get_1d_rotary_pos_embed(dim: int, pos: Union[np.ndarray, int], theta... function apply_rotary_emb (line 310) | def apply_rotary_emb( class TimestepEmbedding (line 340) | class TimestepEmbedding(nn.Module): method __init__ (line 341) | def __init__( method forward (line 373) | def forward(self, sample, condition=None): class Timesteps (line 388) | class Timesteps(nn.Module): method __init__ (line 389) | def __init__(self, num_channels: int, flip_sin_to_cos: bool, downscale... method forward (line 395) | def forward(self, timesteps): class GaussianFourierProjection (line 405) | class GaussianFourierProjection(nn.Module): method __init__ (line 408) | def __init__( method forward (line 422) | def forward(self, x): class SinusoidalPositionalEmbedding (line 435) | class SinusoidalPositionalEmbedding(nn.Module): method __init__ (line 447) | def __init__(self, embed_dim: int, max_seq_length: int = 32): method forward (line 456) | def forward(self, x): class ImagePositionalEmbeddings (line 462) | class ImagePositionalEmbeddings(nn.Module): method __init__ (line 486) | def __init__( method forward (line 504) | def forward(self, index): class LabelEmbedding (line 527) | class LabelEmbedding(nn.Module): method __init__ (line 537) | def __init__(self, num_classes, hidden_size, dropout_prob): method token_drop (line 544) | def token_drop(self, labels, force_drop_ids=None): method forward (line 555) | def forward(self, labels: torch.LongTensor, force_drop_ids=None): class TextImageProjection (line 563) | class TextImageProjection(nn.Module): method __init__ (line 564) | def __init__( method forward (line 577) | def forward(self, text_embeds: torch.Tensor, image_embeds: torch.Tensor): class ImageProjection (line 590) | class ImageProjection(nn.Module): method __init__ (line 591) | def __init__( method forward (line 603) | def forward(self, image_embeds: torch.Tensor): class IPAdapterFullImageProjection (line 613) | class IPAdapterFullImageProjection(nn.Module): method __init__ (line 614) | def __init__(self, image_embed_dim=1024, cross_attention_dim=1024): method forward (line 621) | def forward(self, image_embeds: torch.Tensor): class IPAdapterFaceIDImageProjection (line 625) | class IPAdapterFaceIDImageProjection(nn.Module): method __init__ (line 626) | def __init__(self, image_embed_dim=1024, cross_attention_dim=1024, mul... method forward (line 635) | def forward(self, image_embeds: torch.Tensor): class CombinedTimestepLabelEmbeddings (line 641) | class CombinedTimestepLabelEmbeddings(nn.Module): method __init__ (line 642) | def __init__(self, num_classes, embedding_dim, class_dropout_prob=0.1): method forward (line 649) | def forward(self, timestep, class_labels, hidden_dtype=None): class CombinedTimestepTextProjEmbeddings (line 660) | class CombinedTimestepTextProjEmbeddings(nn.Module): method __init__ (line 661) | def __init__(self, embedding_dim, pooled_projection_dim): method forward (line 668) | def forward(self, timestep, pooled_projection): class HunyuanDiTAttentionPool (line 679) | class HunyuanDiTAttentionPool(nn.Module): method __init__ (line 682) | def __init__(self, spacial_dim: int, embed_dim: int, num_heads: int, o... method forward (line 691) | def forward(self, x): class HunyuanCombinedTimestepTextSizeStyleEmbedding (line 719) | class HunyuanCombinedTimestepTextSizeStyleEmbedding(nn.Module): method __init__ (line 720) | def __init__(self, embedding_dim, pooled_projection_dim=1024, seq_len=... method forward (line 739) | def forward(self, timestep, encoder_hidden_states, image_meta_size, st... class TextTimeEmbedding (line 761) | class TextTimeEmbedding(nn.Module): method __init__ (line 762) | def __init__(self, encoder_dim: int, time_embed_dim: int, num_heads: i... method forward (line 769) | def forward(self, hidden_states): class TextImageTimeEmbedding (line 777) | class TextImageTimeEmbedding(nn.Module): method __init__ (line 778) | def __init__(self, text_embed_dim: int = 768, image_embed_dim: int = 7... method forward (line 784) | def forward(self, text_embeds: torch.Tensor, image_embeds: torch.Tensor): class ImageTimeEmbedding (line 795) | class ImageTimeEmbedding(nn.Module): method __init__ (line 796) | def __init__(self, image_embed_dim: int = 768, time_embed_dim: int = 1... method forward (line 801) | def forward(self, image_embeds: torch.Tensor): class ImageHintTimeEmbedding (line 808) | class ImageHintTimeEmbedding(nn.Module): method __init__ (line 809) | def __init__(self, image_embed_dim: int = 768, time_embed_dim: int = 1... method forward (line 831) | def forward(self, image_embeds: torch.Tensor, hint: torch.Tensor): class AttentionPooling (line 839) | class AttentionPooling(nn.Module): method __init__ (line 842) | def __init__(self, num_heads, embed_dim, dtype=None): method forward (line 852) | def forward(self, x): function get_fourier_embeds_from_boundingbox (line 889) | def get_fourier_embeds_from_boundingbox(embed_dim, box): class GLIGENTextBoundingboxProjection (line 910) | class GLIGENTextBoundingboxProjection(nn.Module): method __init__ (line 911) | def __init__(self, positive_len, out_dim, feature_type="text-only", fo... method forward (line 952) | def forward( class PixArtAlphaCombinedTimestepSizeEmbeddings (line 1003) | class PixArtAlphaCombinedTimestepSizeEmbeddings(nn.Module): method __init__ (line 1011) | def __init__(self, embedding_dim, size_emb_dim, use_additional_conditi... method forward (line 1024) | def forward(self, timestep, resolution, aspect_ratio, batch_size, hidd... class PixArtAlphaTextProjection (line 1040) | class PixArtAlphaTextProjection(nn.Module): method __init__ (line 1047) | def __init__(self, in_features, hidden_size, out_features=None, act_fn... method forward (line 1062) | def forward(self, caption): class IPAdapterPlusImageProjectionBlock (line 1069) | class IPAdapterPlusImageProjectionBlock(nn.Module): method __init__ (line 1070) | def __init__( method forward (line 1093) | def forward(self, x, latents, residual): class IPAdapterPlusImageProjection (line 1102) | class IPAdapterPlusImageProjection(nn.Module): method __init__ (line 1118) | def __init__( method forward (line 1141) | def forward(self, x: torch.Tensor) -> torch.Tensor: class IPAdapterFaceIDPlusImageProjection (line 1161) | class IPAdapterFaceIDPlusImageProjection(nn.Module): method __init__ (line 1178) | def __init__( method forward (line 1213) | def forward(self, id_embeds: torch.Tensor) -> torch.Tensor: class MultiIPAdapterImageProjection (line 1241) | class MultiIPAdapterImageProjection(nn.Module): method __init__ (line 1242) | def __init__(self, IPAdapterImageProjectionLayers: Union[List[nn.Modul... method forward (line 1246) | def forward(self, image_embeds: List[torch.Tensor]): FILE: diffusers/src/diffusers/models/embeddings_flax.py function get_sinusoidal_embeddings (line 20) | def get_sinusoidal_embeddings( class FlaxTimestepEmbedding (line 58) | class FlaxTimestepEmbedding(nn.Module): method __call__ (line 73) | def __call__(self, temb): class FlaxTimesteps (line 80) | class FlaxTimesteps(nn.Module): method __call__ (line 94) | def __call__(self, timesteps): FILE: diffusers/src/diffusers/models/lora.py function text_encoder_attn_modules (line 41) | def text_encoder_attn_modules(text_encoder): function text_encoder_mlp_modules (line 55) | def text_encoder_mlp_modules(text_encoder): function adjust_lora_scale_text_encoder (line 69) | def adjust_lora_scale_text_encoder(text_encoder, lora_scale: float = 1.0): class PatchedLoraProjection (line 83) | class PatchedLoraProjection(torch.nn.Module): method __init__ (line 84) | def __init__(self, regular_linear_layer, lora_scale=1, network_alpha=N... method state_dict (line 111) | def state_dict(self, *args, destination=None, prefix="", keep_vars=Fal... method _fuse_lora (line 119) | def _fuse_lora(self, lora_scale=1.0, safe_fusing=False): method _unfuse_lora (line 151) | def _unfuse_lora(self): method forward (line 167) | def forward(self, input): class LoRALinearLayer (line 175) | class LoRALinearLayer(nn.Module): method __init__ (line 196) | def __init__( method forward (line 222) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class LoRAConv2dLayer (line 235) | class LoRAConv2dLayer(nn.Module): method __init__ (line 258) | def __init__( method forward (line 286) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class LoRACompatibleConv (line 299) | class LoRACompatibleConv(nn.Conv2d): method __init__ (line 304) | def __init__(self, *args, lora_layer: Optional[LoRAConv2dLayer] = None... method set_lora_layer (line 311) | def set_lora_layer(self, lora_layer: Optional[LoRAConv2dLayer]): method _fuse_lora (line 317) | def _fuse_lora(self, lora_scale: float = 1.0, safe_fusing: bool = False): method _unfuse_lora (line 351) | def _unfuse_lora(self): method forward (line 369) | def forward(self, hidden_states: torch.Tensor, scale: float = 1.0) -> ... class LoRACompatibleLinear (line 386) | class LoRACompatibleLinear(nn.Linear): method __init__ (line 391) | def __init__(self, *args, lora_layer: Optional[LoRALinearLayer] = None... method set_lora_layer (line 398) | def set_lora_layer(self, lora_layer: Optional[LoRALinearLayer]): method _fuse_lora (line 403) | def _fuse_lora(self, lora_scale: float = 1.0, safe_fusing: bool = False): method _unfuse_lora (line 435) | def _unfuse_lora(self): method forward (line 451) | def forward(self, hidden_states: torch.Tensor, scale: float = 1.0) -> ... FILE: diffusers/src/diffusers/models/model_loading_utils.py function _determine_device_map (line 56) | def _determine_device_map(model: torch.nn.Module, device_map, max_memory... function _fetch_remapped_cls_from_config (line 78) | def _fetch_remapped_cls_from_config(config, old_class): function load_state_dict (line 98) | def load_state_dict(checkpoint_file: Union[str, os.PathLike], variant: O... function load_model_dict_into_meta (line 133) | def load_model_dict_into_meta( function _load_state_dict_into_model (line 165) | def _load_state_dict_into_model(model_to_load, state_dict: OrderedDict) ... function _fetch_index_file (line 186) | def _fetch_index_file( FILE: diffusers/src/diffusers/models/modeling_flax_pytorch_utils.py function rename_key (line 29) | def rename_key(key): function rename_key_and_reshape_tensor (line 44) | def rename_key_and_reshape_tensor(pt_tuple_key, pt_tensor, random_flax_s... function convert_pytorch_state_dict_to_flax (line 107) | def convert_pytorch_state_dict_to_flax(pt_state_dict, flax_model, init_k... FILE: diffusers/src/diffusers/models/modeling_flax_utils.py class FlaxModelMixin (line 50) | class FlaxModelMixin(PushToHubMixin): method _from_config (line 65) | def _from_config(cls, config, **kwargs): method _cast_floating_to (line 71) | def _cast_floating_to(self, params: Union[Dict, FrozenDict], dtype: jn... method to_bf16 (line 95) | def to_bf16(self, params: Union[Dict, FrozenDict], mask: Any = None): method to_fp32 (line 134) | def to_fp32(self, params: Union[Dict, FrozenDict], mask: Any = None): method to_fp16 (line 161) | def to_fp16(self, params: Union[Dict, FrozenDict], mask: Any = None): method init_weights (line 200) | def init_weights(self, rng: jax.Array) -> Dict: method from_pretrained (line 205) | def from_pretrained( method save_pretrained (line 502) | def save_pretrained( FILE: diffusers/src/diffusers/models/modeling_outputs.py class AutoencoderKLOutput (line 7) | class AutoencoderKLOutput(BaseOutput): class Transformer2DModelOutput (line 21) | class Transformer2DModelOutput(BaseOutput): FILE: diffusers/src/diffusers/models/modeling_pytorch_flax_utils.py function load_flax_checkpoint_in_pytorch_model (line 37) | def load_flax_checkpoint_in_pytorch_model(pt_model, model_file): function load_flax_weights_in_pytorch_model (line 58) | def load_flax_weights_in_pytorch_model(pt_model, flax_state): FILE: diffusers/src/diffusers/models/modeling_utils.py function get_parameter_device (line 78) | def get_parameter_device(parameter: torch.nn.Module) -> torch.device: function get_parameter_dtype (line 94) | def get_parameter_dtype(parameter: torch.nn.Module) -> torch.dtype: class ModelMixin (line 116) | class ModelMixin(torch.nn.Module, PushToHubMixin): method __init__ (line 132) | def __init__(self): method __getattr__ (line 135) | def __getattr__(self, name: str) -> Any: method is_gradient_checkpointing (line 154) | def is_gradient_checkpointing(self) -> bool: method enable_gradient_checkpointing (line 160) | def enable_gradient_checkpointing(self) -> None: method disable_gradient_checkpointing (line 169) | def disable_gradient_checkpointing(self) -> None: method set_use_npu_flash_attention (line 177) | def set_use_npu_flash_attention(self, valid: bool) -> None: method enable_npu_flash_attention (line 193) | def enable_npu_flash_attention(self) -> None: method disable_npu_flash_attention (line 200) | def disable_npu_flash_attention(self) -> None: method set_use_memory_efficient_attention_xformers (line 207) | def set_use_memory_efficient_attention_xformers( method enable_xformers_memory_efficient_attention (line 224) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method disable_xformers_memory_efficient_attention (line 260) | def disable_xformers_memory_efficient_attention(self) -> None: method save_pretrained (line 266) | def save_pretrained( method from_pretrained (line 412) | def from_pretrained(cls, pretrained_model_name_or_path: Optional[Union... method _load_pretrained_model (line 866) | def _load_pretrained_model( method _get_signature_keys (line 970) | def _get_signature_keys(cls, obj): method _get_no_split_modules (line 979) | def _get_no_split_modules(self, device_map: str): method device (line 1009) | def device(self) -> torch.device: method dtype (line 1017) | def dtype(self) -> torch.dtype: method num_parameters (line 1023) | def num_parameters(self, only_trainable: bool = False, exclude_embeddi... method _convert_deprecated_attention_blocks (line 1061) | def _convert_deprecated_attention_blocks(self, state_dict: OrderedDict... method _temp_convert_self_to_deprecated_attention_blocks (line 1105) | def _temp_convert_self_to_deprecated_attention_blocks(self) -> None: method _undo_temp_convert_self_to_deprecated_attention_blocks (line 1132) | def _undo_temp_convert_self_to_deprecated_attention_blocks(self) -> None: class LegacyModelMixin (line 1156) | class LegacyModelMixin(ModelMixin): method from_pretrained (line 1164) | def from_pretrained(cls, pretrained_model_name_or_path: Optional[Union... FILE: diffusers/src/diffusers/models/normalization.py class AdaLayerNorm (line 28) | class AdaLayerNorm(nn.Module): method __init__ (line 37) | def __init__(self, embedding_dim: int, num_embeddings: int): method forward (line 44) | def forward(self, x: torch.Tensor, timestep: torch.Tensor) -> torch.Te... class AdaLayerNormZero (line 51) | class AdaLayerNormZero(nn.Module): method __init__ (line 60) | def __init__(self, embedding_dim: int, num_embeddings: Optional[int] =... method forward (line 71) | def forward( class AdaLayerNormSingle (line 87) | class AdaLayerNormSingle(nn.Module): method __init__ (line 98) | def __init__(self, embedding_dim: int, use_additional_conditions: bool... method forward (line 108) | def forward( class AdaGroupNorm (line 120) | class AdaGroupNorm(nn.Module): method __init__ (line 132) | def __init__( method forward (line 146) | def forward(self, x: torch.Tensor, emb: torch.Tensor) -> torch.Tensor: class AdaLayerNormContinuous (line 158) | class AdaLayerNormContinuous(nn.Module): method __init__ (line 159) | def __init__( method forward (line 183) | def forward(self, x: torch.Tensor, conditioning_embedding: torch.Tenso... class LayerNorm (line 196) | class LayerNorm(nn.Module): method __init__ (line 197) | def __init__(self, dim, eps: float = 1e-5, elementwise_affine: bool = ... method forward (line 214) | def forward(self, input): class RMSNorm (line 218) | class RMSNorm(nn.Module): method __init__ (line 219) | def __init__(self, dim, eps: float, elementwise_affine: bool = True): method forward (line 234) | def forward(self, hidden_states): class GlobalResponseNorm (line 250) | class GlobalResponseNorm(nn.Module): method __init__ (line 252) | def __init__(self, dim): method forward (line 257) | def forward(self, x): FILE: diffusers/src/diffusers/models/resnet.py class ResnetBlockCondNorm2D (line 44) | class ResnetBlockCondNorm2D(nn.Module): method __init__ (line 74) | def __init__( method forward (line 149) | def forward(self, input_tensor: torch.Tensor, temb: torch.Tensor, *arg... class ResnetBlock2D (line 189) | class ResnetBlock2D(nn.Module): method __init__ (line 220) | def __init__( method forward (line 320) | def forward(self, input_tensor: torch.Tensor, temb: torch.Tensor, *arg... function rearrange_dims (line 377) | def rearrange_dims(tensor: torch.Tensor) -> torch.Tensor: class Conv1dBlock (line 388) | class Conv1dBlock(nn.Module): method __init__ (line 400) | def __init__( method forward (line 414) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: class ResidualTemporalBlock1D (line 424) | class ResidualTemporalBlock1D(nn.Module): method __init__ (line 436) | def __init__( method forward (line 455) | def forward(self, inputs: torch.Tensor, t: torch.Tensor) -> torch.Tensor: class TemporalConvLayer (line 471) | class TemporalConvLayer(nn.Module): method __init__ (line 482) | def __init__( method forward (line 523) | def forward(self, hidden_states: torch.Tensor, num_frames: int = 1) ->... class TemporalResnetBlock (line 542) | class TemporalResnetBlock(nn.Module): method __init__ (line 554) | def __init__( method forward (line 608) | def forward(self, input_tensor: torch.Tensor, temb: torch.Tensor) -> t... class SpatioTemporalResBlock (line 635) | class SpatioTemporalResBlock(nn.Module): method __init__ (line 653) | def __init__( method forward (line 686) | def forward( class AlphaBlender (line 719) | class AlphaBlender(nn.Module): method __init__ (line 733) | def __init__( method get_alpha (line 753) | def get_alpha(self, image_only_indicator: torch.Tensor, ndims: int) ->... method forward (line 784) | def forward( FILE: diffusers/src/diffusers/models/resnet_flax.py class FlaxUpsample2D (line 19) | class FlaxUpsample2D(nn.Module): method setup (line 23) | def setup(self): method __call__ (line 32) | def __call__(self, hidden_states): class FlaxDownsample2D (line 43) | class FlaxDownsample2D(nn.Module): method setup (line 47) | def setup(self): method __call__ (line 56) | def __call__(self, hidden_states): class FlaxResnetBlock2D (line 63) | class FlaxResnetBlock2D(nn.Module): method setup (line 70) | def setup(self): method __call__ (line 106) | def __call__(self, hidden_states, temb, deterministic=True): FILE: diffusers/src/diffusers/models/transformers/dit_transformer_2d.py class DiTTransformer2DModel (line 31) | class DiTTransformer2DModel(ModelMixin, ConfigMixin): method __init__ (line 70) | def __init__( method _set_gradient_checkpointing (line 146) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 150) | def forward( FILE: diffusers/src/diffusers/models/transformers/dual_transformer_2d.py class DualTransformer2DModel (line 22) | class DualTransformer2DModel(nn.Module): method __init__ (line 49) | def __init__( method forward (line 98) | def forward( FILE: diffusers/src/diffusers/models/transformers/hunyuan_transformer_2d.py class FP32LayerNorm (line 38) | class FP32LayerNorm(nn.LayerNorm): method forward (line 39) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: class AdaLayerNormShift (line 46) | class AdaLayerNormShift(nn.Module): method __init__ (line 55) | def __init__(self, embedding_dim: int, elementwise_affine=True, eps=1e... method forward (line 61) | def forward(self, x: torch.Tensor, emb: torch.Tensor) -> torch.Tensor: class HunyuanDiTBlock (line 68) | class HunyuanDiTBlock(nn.Module): method __init__ (line 100) | def __init__( method set_chunk_feed_forward (line 170) | def set_chunk_feed_forward(self, chunk_size: Optional[int], dim: int =... method forward (line 175) | def forward( class HunyuanDiT2DModel (line 212) | class HunyuanDiT2DModel(ModelMixin, ConfigMixin): method __init__ (line 255) | def __init__( method fuse_qkv_projections (line 326) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 350) | def unfuse_qkv_projections(self): method attn_processors (line 365) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 389) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 423) | def set_default_attn_processor(self): method forward (line 429) | def forward( method enable_forward_chunking (line 535) | def enable_forward_chunking(self, chunk_size: Optional[int] = None, di... method disable_forward_chunking (line 565) | def disable_forward_chunking(self): FILE: diffusers/src/diffusers/models/transformers/pixart_transformer_2d.py class PixArtTransformer2DModel (line 31) | class PixArtTransformer2DModel(ModelMixin, ConfigMixin): method __init__ (line 83) | def __init__( method _set_gradient_checkpointing (line 185) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 189) | def forward( FILE: diffusers/src/diffusers/models/transformers/prior_transformer.py class PriorTransformerOutput (line 24) | class PriorTransformerOutput(BaseOutput): class PriorTransformer (line 36) | class PriorTransformer(ModelMixin, ConfigMixin, UNet2DConditionLoadersMi... method __init__ (line 74) | def __init__( method attn_processors (line 171) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 195) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 230) | def set_default_attn_processor(self): method forward (line 245) | def forward( method post_process_latents (line 378) | def post_process_latents(self, prior_latents): FILE: diffusers/src/diffusers/models/transformers/t5_film_transformer.py class T5FilmDecoder (line 26) | class T5FilmDecoder(ModelMixin, ConfigMixin): method __init__ (line 50) | def __init__( method encoder_decoder_mask (line 89) | def encoder_decoder_mask(self, query_input: torch.Tensor, key_input: t... method forward (line 93) | def forward(self, encodings_and_masks, decoder_input_tokens, decoder_n... class DecoderLayer (line 150) | class DecoderLayer(nn.Module): method __init__ (line 169) | def __init__( method forward (line 196) | def forward( class T5LayerSelfAttentionCond (line 228) | class T5LayerSelfAttentionCond(nn.Module): method __init__ (line 243) | def __init__(self, d_model: int, d_kv: int, num_heads: int, dropout_ra... method forward (line 250) | def forward( class T5LayerCrossAttention (line 270) | class T5LayerCrossAttention(nn.Module): method __init__ (line 287) | def __init__(self, d_model: int, d_kv: int, num_heads: int, dropout_ra... method forward (line 293) | def forward( class T5LayerFFCond (line 309) | class T5LayerFFCond(nn.Module): method __init__ (line 324) | def __init__(self, d_model: int, d_ff: int, dropout_rate: float, layer... method forward (line 331) | def forward(self, hidden_states: torch.Tensor, conditioning_emb: Optio... class T5DenseGatedActDense (line 341) | class T5DenseGatedActDense(nn.Module): method __init__ (line 354) | def __init__(self, d_model: int, d_ff: int, dropout_rate: float): method forward (line 362) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class T5LayerNorm (line 372) | class T5LayerNorm(nn.Module): method __init__ (line 383) | def __init__(self, hidden_size: int, eps: float = 1e-6): method forward (line 391) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class NewGELUActivation (line 407) | class NewGELUActivation(nn.Module): method forward (line 413) | def forward(self, input: torch.Tensor) -> torch.Tensor: class T5FiLMLayer (line 417) | class T5FiLMLayer(nn.Module): method __init__ (line 428) | def __init__(self, in_features: int, out_features: int): method forward (line 432) | def forward(self, x: torch.Tensor, conditioning_emb: torch.Tensor) -> ... FILE: diffusers/src/diffusers/models/transformers/transformer_2d.py class Transformer2DModelOutput (line 32) | class Transformer2DModelOutput(Transformer2DModelOutput): class Transformer2DModel (line 37) | class Transformer2DModel(LegacyModelMixin, LegacyConfigMixin): method __init__ (line 69) | def __init__( method _init_continuous_input (line 172) | def _init_continuous_input(self, norm_type): method _init_vectorized_inputs (line 209) | def _init_vectorized_inputs(self, norm_type): method _init_patched_inputs (line 249) | def _init_patched_inputs(self, norm_type): method _set_gradient_checkpointing (line 321) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 325) | def forward( method _operate_on_continuous_inputs (line 477) | def _operate_on_continuous_inputs(self, hidden_states): method _operate_on_patched_inputs (line 492) | def _operate_on_patched_inputs(self, hidden_states, encoder_hidden_sta... method _get_output_for_continuous_inputs (line 512) | def _get_output_for_continuous_inputs(self, hidden_states, residual, b... method _get_output_for_vectorized_inputs (line 527) | def _get_output_for_vectorized_inputs(self, hidden_states): method _get_output_for_patched_inputs (line 536) | def _get_output_for_patched_inputs( FILE: diffusers/src/diffusers/models/transformers/transformer_sd3.py class SD3Transformer2DModel (line 35) | class SD3Transformer2DModel(ModelMixin, ConfigMixin, PeftAdapterMixin, F... method __init__ (line 59) | def __init__( method enable_forward_chunking (line 111) | def enable_forward_chunking(self, chunk_size: Optional[int] = None, di... method attn_processors (line 142) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 166) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method fuse_qkv_projections (line 201) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 225) | def unfuse_qkv_projections(self): method _set_gradient_checkpointing (line 238) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 242) | def forward( FILE: diffusers/src/diffusers/models/transformers/transformer_temporal.py class TransformerTemporalModelOutput (line 29) | class TransformerTemporalModelOutput(BaseOutput): class TransformerTemporalModel (line 41) | class TransformerTemporalModel(ModelMixin, ConfigMixin): method __init__ (line 71) | def __init__( method forward (line 121) | def forward( class TransformerSpatioTemporalModel (line 203) | class TransformerSpatioTemporalModel(nn.Module): method __init__ (line 218) | def __init__( method forward (line 278) | def forward( FILE: diffusers/src/diffusers/models/unets/unet_1d.py class UNet1DOutput (line 29) | class UNet1DOutput(BaseOutput): class UNet1DModel (line 41) | class UNet1DModel(ModelMixin, ConfigMixin): method __init__ (line 75) | def __init__( method forward (line 195) | def forward( FILE: diffusers/src/diffusers/models/unets/unet_1d_blocks.py class DownResnetBlock1D (line 25) | class DownResnetBlock1D(nn.Module): method __init__ (line 26) | def __init__( method forward (line 69) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class UpResnetBlock1D (line 87) | class UpResnetBlock1D(nn.Module): method __init__ (line 88) | def __init__( method forward (line 129) | def forward( class ValueFunctionMidBlock1D (line 152) | class ValueFunctionMidBlock1D(nn.Module): method __init__ (line 153) | def __init__(self, in_channels: int, out_channels: int, embed_dim: int): method forward (line 164) | def forward(self, x: torch.Tensor, temb: Optional[torch.Tensor] = None... class MidResTemporalBlock1D (line 172) | class MidResTemporalBlock1D(nn.Module): method __init__ (line 173) | def __init__( method forward (line 212) | def forward(self, hidden_states: torch.Tensor, temb: torch.Tensor) -> ... class OutConv1DBlock (line 225) | class OutConv1DBlock(nn.Module): method __init__ (line 226) | def __init__(self, num_groups_out: int, out_channels: int, embed_dim: ... method forward (line 233) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class OutValueFunctionBlock (line 243) | class OutValueFunctionBlock(nn.Module): method __init__ (line 244) | def __init__(self, fc_dim: int, embed_dim: int, act_fn: str = "mish"): method forward (line 254) | def forward(self, hidden_states: torch.Tensor, temb: torch.Tensor) -> ... class Downsample1d (line 283) | class Downsample1d(nn.Module): method __init__ (line 284) | def __init__(self, kernel: str = "linear", pad_mode: str = "reflect"): method forward (line 291) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class Upsample1d (line 300) | class Upsample1d(nn.Module): method __init__ (line 301) | def __init__(self, kernel: str = "linear", pad_mode: str = "reflect"): method forward (line 308) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class SelfAttention1d (line 317) | class SelfAttention1d(nn.Module): method __init__ (line 318) | def __init__(self, in_channels: int, n_head: int = 1, dropout_rate: fl... method transpose_for_scores (line 332) | def transpose_for_scores(self, projection: torch.Tensor) -> torch.Tensor: method forward (line 338) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class ResConvBlock (line 375) | class ResConvBlock(nn.Module): method __init__ (line 376) | def __init__(self, in_channels: int, mid_channels: int, out_channels: ... method forward (line 393) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class UNetMidBlock1D (line 409) | class UNetMidBlock1D(nn.Module): method __init__ (line 410) | def __init__(self, mid_channels: int, in_channels: int, out_channels: ... method forward (line 438) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class AttnDownBlock1D (line 449) | class AttnDownBlock1D(nn.Module): method __init__ (line 450) | def __init__(self, out_channels: int, in_channels: int, mid_channels: ... method forward (line 469) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class DownBlock1D (line 479) | class DownBlock1D(nn.Module): method __init__ (line 480) | def __init__(self, out_channels: int, in_channels: int, mid_channels: ... method forward (line 493) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class DownBlock1DNoSkip (line 502) | class DownBlock1DNoSkip(nn.Module): method __init__ (line 503) | def __init__(self, out_channels: int, in_channels: int, mid_channels: ... method forward (line 515) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class AttnUpBlock1D (line 523) | class AttnUpBlock1D(nn.Module): method __init__ (line 524) | def __init__(self, in_channels: int, out_channels: int, mid_channels: ... method forward (line 543) | def forward( class UpBlock1D (line 561) | class UpBlock1D(nn.Module): method __init__ (line 562) | def __init__(self, in_channels: int, out_channels: int, mid_channels: ... method forward (line 575) | def forward( class UpBlock1DNoSkip (line 592) | class UpBlock1DNoSkip(nn.Module): method __init__ (line 593) | def __init__(self, in_channels: int, out_channels: int, mid_channels: ... method forward (line 605) | def forward( function get_down_block (line 626) | def get_down_block( function get_up_block (line 651) | def get_up_block( function get_mid_block (line 671) | def get_mid_block( function get_out_block (line 695) | def get_out_block( FILE: diffusers/src/diffusers/models/unets/unet_2d.py class UNet2DOutput (line 28) | class UNet2DOutput(BaseOutput): class UNet2DModel (line 40) | class UNet2DModel(ModelMixin, ConfigMixin): method __init__ (line 93) | def __init__( method forward (line 243) | def forward( FILE: diffusers/src/diffusers/models/unets/unet_2d_blocks.py function get_down_block (line 43) | def get_down_block( function get_mid_block (line 252) | def get_mid_block( function get_up_block (line 327) | def get_up_block( class AutoencoderTinyBlock (line 552) | class AutoencoderTinyBlock(nn.Module): method __init__ (line 568) | def __init__(self, in_channels: int, out_channels: int, act_fn: str): method forward (line 585) | def forward(self, x: torch.Tensor) -> torch.Tensor: class UNetMidBlock2D (line 589) | class UNetMidBlock2D(nn.Module): method __init__ (line 620) | def __init__( method forward (line 734) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class UNetMidBlock2DCrossAttn (line 744) | class UNetMidBlock2DCrossAttn(nn.Module): method __init__ (line 745) | def __init__( method forward (line 847) | def forward( class UNetMidBlock2DSimpleCrossAttn (line 902) | class UNetMidBlock2DSimpleCrossAttn(nn.Module): method __init__ (line 903) | def __init__( method forward (line 987) | def forward( class AttnDownBlock2D (line 1027) | class AttnDownBlock2D(nn.Module): method __init__ (line 1028) | def __init__( method forward (line 1119) | def forward( class CrossAttnDownBlock2D (line 1149) | class CrossAttnDownBlock2D(nn.Module): method __init__ (line 1150) | def __init__( method forward (line 1241) | def forward( class DownBlock2D (line 1312) | class DownBlock2D(nn.Module): method __init__ (line 1313) | def __init__( method forward (line 1364) | def forward( class DownEncoderBlock2D (line 1404) | class DownEncoderBlock2D(nn.Module): method __init__ (line 1405) | def __init__( method forward (line 1468) | def forward(self, hidden_states: torch.Tensor, *args, **kwargs) -> tor... class AttnDownEncoderBlock2D (line 1483) | class AttnDownEncoderBlock2D(nn.Module): method __init__ (line 1484) | def __init__( method forward (line 1570) | def forward(self, hidden_states: torch.Tensor, *args, **kwargs) -> tor... class AttnSkipDownBlock2D (line 1586) | class AttnSkipDownBlock2D(nn.Module): method __init__ (line 1587) | def __init__( method forward (line 1667) | def forward( class SkipDownBlock2D (line 1698) | class SkipDownBlock2D(nn.Module): method __init__ (line 1699) | def __init__( method forward (line 1758) | def forward( class ResnetDownsampleBlock2D (line 1788) | class ResnetDownsampleBlock2D(nn.Module): method __init__ (line 1789) | def __init__( method forward (line 1852) | def forward( class SimpleCrossAttnDownBlock2D (line 1892) | class SimpleCrossAttnDownBlock2D(nn.Module): method __init__ (line 1893) | def __init__( method forward (line 1987) | def forward( class KDownBlock2D (line 2053) | class KDownBlock2D(nn.Module): method __init__ (line 2054) | def __init__( method forward (line 2099) | def forward( class KCrossAttnDownBlock2D (line 2137) | class KCrossAttnDownBlock2D(nn.Module): method __init__ (line 2138) | def __init__( method forward (line 2202) | def forward( class AttnUpBlock2D (line 2267) | class AttnUpBlock2D(nn.Module): method __init__ (line 2268) | def __init__( method forward (line 2359) | def forward( class CrossAttnUpBlock2D (line 2391) | class CrossAttnUpBlock2D(nn.Module): method __init__ (line 2392) | def __init__( method forward (line 2482) | def forward( class UpBlock2D (line 2567) | class UpBlock2D(nn.Module): method __init__ (line 2568) | def __init__( method forward (line 2617) | def forward( class UpDecoderBlock2D (line 2682) | class UpDecoderBlock2D(nn.Module): method __init__ (line 2683) | def __init__( method forward (line 2744) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class AttnUpDecoderBlock2D (line 2755) | class AttnUpDecoderBlock2D(nn.Module): method __init__ (line 2756) | def __init__( method forward (line 2842) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class AttnSkipUpBlock2D (line 2854) | class AttnSkipUpBlock2D(nn.Module): method __init__ (line 2855) | def __init__( method forward (line 2948) | def forward( class SkipUpBlock2D (line 2988) | class SkipUpBlock2D(nn.Module): method __init__ (line 2989) | def __init__( method forward (line 3060) | def forward( class ResnetUpsampleBlock2D (line 3098) | class ResnetUpsampleBlock2D(nn.Module): method __init__ (line 3099) | def __init__( method forward (line 3167) | def forward( class SimpleCrossAttnUpBlock2D (line 3212) | class SimpleCrossAttnUpBlock2D(nn.Module): method __init__ (line 3213) | def __init__( method forward (line 3311) | def forward( class KUpBlock2D (line 3379) | class KUpBlock2D(nn.Module): method __init__ (line 3380) | def __init__( method forward (line 3429) | def forward( class KCrossAttnUpBlock2D (line 3473) | class KCrossAttnUpBlock2D(nn.Module): method __init__ (line 3474) | def __init__( method forward (line 3559) | def forward( class KAttentionBlock (line 3620) | class KAttentionBlock(nn.Module): method __init__ (line 3644) | def __init__( method _to_3d (line 3687) | def _to_3d(self, hidden_states: torch.Tensor, height: int, weight: int... method _to_4d (line 3690) | def _to_4d(self, hidden_states: torch.Tensor, height: int, weight: int... method forward (line 3693) | def forward( FILE: diffusers/src/diffusers/models/unets/unet_2d_blocks_flax.py class FlaxCrossAttnDownBlock2D (line 22) | class FlaxCrossAttnDownBlock2D(nn.Module): method setup (line 62) | def setup(self): method __call__ (line 96) | def __call__(self, hidden_states, temb, encoder_hidden_states, determi... class FlaxDownBlock2D (line 111) | class FlaxDownBlock2D(nn.Module): method setup (line 137) | def setup(self): method __call__ (line 155) | def __call__(self, hidden_states, temb, deterministic=True): class FlaxCrossAttnUpBlock2D (line 169) | class FlaxCrossAttnUpBlock2D(nn.Module): method setup (line 210) | def setup(self): method __call__ (line 245) | def __call__(self, hidden_states, res_hidden_states_tuple, temb, encod... class FlaxUpBlock2D (line 261) | class FlaxUpBlock2D(nn.Module): method setup (line 290) | def setup(self): method __call__ (line 310) | def __call__(self, hidden_states, res_hidden_states_tuple, temb, deter... class FlaxUNetMidBlock2DCrossAttn (line 325) | class FlaxUNetMidBlock2DCrossAttn(nn.Module): method setup (line 357) | def setup(self): method __call__ (line 394) | def __call__(self, hidden_states, temb, encoder_hidden_states, determi... FILE: diffusers/src/diffusers/models/unets/unet_2d_condition.py class UNet2DConditionOutput (line 58) | class UNet2DConditionOutput(BaseOutput): class UNet2DConditionModel (line 70) | class UNet2DConditionModel( method __init__ (line 170) | def __init__( method _check_config (line 486) | def _check_config( method _set_time_proj (line 538) | def _set_time_proj( method _set_encoder_hid_proj (line 566) | def _set_encoder_hid_proj( method _set_class_embedding (line 606) | def _set_class_embedding( method _set_add_embedding (line 643) | def _set_add_embedding( method _set_pos_net_if_use_gligen (line 683) | def _set_pos_net_if_use_gligen(self, attention_type: str, cross_attent... method attn_processors (line 697) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 720) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 754) | def set_default_attn_processor(self): method set_attention_slice (line 769) | def set_attention_slice(self, slice_size: Union[str, int, List[int]] =... method _set_gradient_checkpointing (line 834) | def _set_gradient_checkpointing(self, module, value=False): method enable_freeu (line 838) | def enable_freeu(self, s1: float, s2: float, b1: float, b2: float): method disable_freeu (line 862) | def disable_freeu(self): method fuse_qkv_projections (line 870) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 893) | def unfuse_qkv_projections(self): method get_time_embed (line 906) | def get_time_embed( method get_class_embed (line 932) | def get_class_embed(self, sample: torch.Tensor, class_labels: Optional... method get_aug_embed (line 948) | def get_aug_embed( method process_encoder_hidden_states (line 1000) | def process_encoder_hidden_states( method forward (line 1032) | def forward( FILE: diffusers/src/diffusers/models/unets/unet_2d_condition_flax.py class FlaxUNet2DConditionOutput (line 36) | class FlaxUNet2DConditionOutput(BaseOutput): class FlaxUNet2DConditionModel (line 49) | class FlaxUNet2DConditionModel(nn.Module, FlaxModelMixin, ConfigMixin): method init_weights (line 133) | def init_weights(self, rng: jax.Array) -> FrozenDict: method setup (line 165) | def setup(self) -> None: method __call__ (line 330) | def __call__( FILE: diffusers/src/diffusers/models/unets/unet_3d_blocks.py function get_down_block (line 41) | def get_down_block( function get_up_block (line 170) | def get_up_block( class UNetMidBlock3DCrossAttn (line 310) | class UNetMidBlock3DCrossAttn(nn.Module): method __init__ (line 311) | def __init__( method forward (line 412) | def forward( class CrossAttnDownBlock3D (line 444) | class CrossAttnDownBlock3D(nn.Module): method __init__ (line 445) | def __init__( method forward (line 545) | def forward( class DownBlock3D (line 586) | class DownBlock3D(nn.Module): method __init__ (line 587) | def __init__( method forward (line 652) | def forward( class CrossAttnUpBlock3D (line 675) | class CrossAttnUpBlock3D(nn.Module): method __init__ (line 676) | def __init__( method forward (line 770) | def forward( class UpBlock3D (line 832) | class UpBlock3D(nn.Module): method __init__ (line 833) | def __init__( method forward (line 892) | def forward( class DownBlockMotion (line 935) | class DownBlockMotion(nn.Module): method __init__ (line 936) | def __init__( method forward (line 1009) | def forward( class CrossAttnDownBlockMotion (line 1060) | class CrossAttnDownBlockMotion(nn.Module): method __init__ (line 1061) | def __init__( method forward (line 1175) | def forward( class CrossAttnUpBlockMotion (line 1250) | class CrossAttnUpBlockMotion(nn.Module): method __init__ (line 1251) | def __init__( method forward (line 1358) | def forward( class UpBlockMotion (line 1449) | class UpBlockMotion(nn.Module): method __init__ (line 1450) | def __init__( method forward (line 1519) | def forward( class UNetMidBlockCrossAttnMotion (line 1592) | class UNetMidBlockCrossAttnMotion(nn.Module): method __init__ (line 1593) | def __init__( method forward (line 1700) | def forward( class MidBlockTemporalDecoder (line 1768) | class MidBlockTemporalDecoder(nn.Module): method __init__ (line 1769) | def __init__( method forward (line 1812) | def forward( class UpBlockTemporalDecoder (line 1831) | class UpBlockTemporalDecoder(nn.Module): method __init__ (line 1832) | def __init__( method forward (line 1863) | def forward( class UNetMidBlockSpatioTemporal (line 1881) | class UNetMidBlockSpatioTemporal(nn.Module): method __init__ (line 1882) | def __init__( method forward (line 1936) | def forward( class DownBlockSpatioTemporal (line 1991) | class DownBlockSpatioTemporal(nn.Module): method __init__ (line 1992) | def __init__( method forward (line 2032) | def forward( class CrossAttnDownBlockSpatioTemporal (line 2081) | class CrossAttnDownBlockSpatioTemporal(nn.Module): method __init__ (line 2082) | def __init__( method forward (line 2142) | def forward( class UpBlockSpatioTemporal (line 2203) | class UpBlockSpatioTemporal(nn.Module): method __init__ (line 2204) | def __init__( method forward (line 2241) | def forward( class CrossAttnUpBlockSpatioTemporal (line 2292) | class CrossAttnUpBlockSpatioTemporal(nn.Module): method __init__ (line 2293) | def __init__( method forward (line 2350) | def forward( FILE: diffusers/src/diffusers/models/unets/unet_3d_condition.py class UNet3DConditionOutput (line 53) | class UNet3DConditionOutput(BaseOutput): class UNet3DConditionModel (line 65) | class UNet3DConditionModel(ModelMixin, ConfigMixin, UNet2DConditionLoade... method __init__ (line 101) | def __init__( method attn_processors (line 293) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attention_slice (line 317) | def set_attention_slice(self, slice_size: Union[str, int, List[int]]) ... method set_attn_processor (line 383) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method enable_forward_chunking (line 417) | def enable_forward_chunking(self, chunk_size: Optional[int] = None, di... method disable_forward_chunking (line 446) | def disable_forward_chunking(self): method set_default_attn_processor (line 458) | def set_default_attn_processor(self): method _set_gradient_checkpointing (line 473) | def _set_gradient_checkpointing(self, module, value: bool = False) -> ... method enable_freeu (line 478) | def enable_freeu(self, s1, s2, b1, b2): method disable_freeu (line 503) | def disable_freeu(self): method fuse_qkv_projections (line 512) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 536) | def unfuse_qkv_projections(self): method forward (line 549) | def forward( FILE: diffusers/src/diffusers/models/unets/unet_i2vgen_xl.py class I2VGenXLTransformerTemporalEncoder (line 51) | class I2VGenXLTransformerTemporalEncoder(nn.Module): method __init__ (line 52) | def __init__( method forward (line 82) | def forward( class I2VGenXLUNet (line 100) | class I2VGenXLUNet(ModelMixin, ConfigMixin, UNet2DConditionLoadersMixin): method __init__ (line 130) | def __init__( method attn_processors (line 322) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 346) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method enable_forward_chunking (line 381) | def enable_forward_chunking(self, chunk_size: Optional[int] = None, di... method disable_forward_chunking (line 411) | def disable_forward_chunking(self): method set_default_attn_processor (line 423) | def set_default_attn_processor(self): method _set_gradient_checkpointing (line 439) | def _set_gradient_checkpointing(self, module, value: bool = False) -> ... method enable_freeu (line 444) | def enable_freeu(self, s1, s2, b1, b2): method disable_freeu (line 469) | def disable_freeu(self): method fuse_qkv_projections (line 478) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 502) | def unfuse_qkv_projections(self): method forward (line 515) | def forward( FILE: diffusers/src/diffusers/models/unets/unet_kandinsky3.py class Kandinsky3UNetOutput (line 33) | class Kandinsky3UNetOutput(BaseOutput): class Kandinsky3EncoderProj (line 37) | class Kandinsky3EncoderProj(nn.Module): method __init__ (line 38) | def __init__(self, encoder_hid_dim, cross_attention_dim): method forward (line 43) | def forward(self, x): class Kandinsky3UNet (line 49) | class Kandinsky3UNet(ModelMixin, ConfigMixin): method __init__ (line 51) | def __init__( method attn_processors (line 145) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 168) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 202) | def set_default_attn_processor(self): method _set_gradient_checkpointing (line 208) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 212) | def forward(self, sample, timestep, encoder_hidden_states=None, encode... class Kandinsky3UpSampleBlock (line 254) | class Kandinsky3UpSampleBlock(nn.Module): method __init__ (line 255) | def __init__( method forward (line 313) | def forward(self, x, time_embed, context=None, context_mask=None, imag... class Kandinsky3DownSampleBlock (line 325) | class Kandinsky3DownSampleBlock(nn.Module): method __init__ (line 326) | def __init__( method forward (line 381) | def forward(self, x, time_embed, context=None, context_mask=None, imag... class Kandinsky3ConditionalGroupNorm (line 393) | class Kandinsky3ConditionalGroupNorm(nn.Module): method __init__ (line 394) | def __init__(self, groups, normalized_shape, context_dim): method forward (line 401) | def forward(self, x, context): class Kandinsky3Block (line 412) | class Kandinsky3Block(nn.Module): method __init__ (line 413) | def __init__(self, in_channels, out_channels, time_embed_dim, kernel_s... method forward (line 430) | def forward(self, x, time_embed): class Kandinsky3ResNetBlock (line 439) | class Kandinsky3ResNetBlock(nn.Module): method __init__ (line 440) | def __init__( method forward (line 471) | def forward(self, x, time_embed): class Kandinsky3AttentionPooling (line 483) | class Kandinsky3AttentionPooling(nn.Module): method __init__ (line 484) | def __init__(self, num_channels, context_dim, head_dim=64): method forward (line 494) | def forward(self, x, context, context_mask=None): class Kandinsky3AttentionBlock (line 500) | class Kandinsky3AttentionBlock(nn.Module): method __init__ (line 501) | def __init__(self, num_channels, time_embed_dim, context_dim=None, nor... method forward (line 520) | def forward(self, x, time_embed, context=None, context_mask=None, imag... FILE: diffusers/src/diffusers/models/unets/unet_motion_model.py class MotionModules (line 55) | class MotionModules(nn.Module): method __init__ (line 56) | def __init__( class MotionAdapter (line 86) | class MotionAdapter(ModelMixin, ConfigMixin): method __init__ (line 88) | def __init__( method forward (line 177) | def forward(self, sample): class UNetMotionModel (line 181) | class UNetMotionModel(ModelMixin, ConfigMixin, UNet2DConditionLoadersMix... method __init__ (line 193) | def __init__( method from_unet2d (line 432) | def from_unet2d( method freeze_unet2d_params (line 558) | def freeze_unet2d_params(self) -> None: method load_motion_modules (line 582) | def load_motion_modules(self, motion_adapter: Optional[MotionAdapter])... method save_motion_modules (line 592) | def save_motion_modules( method attn_processors (line 629) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 653) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method enable_forward_chunking (line 688) | def enable_forward_chunking(self, chunk_size: Optional[int] = None, di... method disable_forward_chunking (line 718) | def disable_forward_chunking(self) -> None: method set_default_attn_processor (line 730) | def set_default_attn_processor(self) -> None: method _set_gradient_checkpointing (line 745) | def _set_gradient_checkpointing(self, module, value: bool = False) -> ... method enable_freeu (line 750) | def enable_freeu(self, s1: float, s2: float, b1: float, b2: float) -> ... method disable_freeu (line 775) | def disable_freeu(self) -> None: method fuse_qkv_projections (line 784) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 808) | def unfuse_qkv_projections(self): method forward (line 821) | def forward( FILE: diffusers/src/diffusers/models/unets/unet_spatio_temporal_condition.py class UNetSpatioTemporalConditionOutput (line 20) | class UNetSpatioTemporalConditionOutput(BaseOutput): class UNetSpatioTemporalConditionModel (line 32) | class UNetSpatioTemporalConditionModel(ModelMixin, ConfigMixin, UNet2DCo... method __init__ (line 71) | def __init__( method attn_processors (line 249) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 276) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 310) | def set_default_attn_processor(self): method _set_gradient_checkpointing (line 323) | def _set_gradient_checkpointing(self, module, value=False): method enable_forward_chunking (line 328) | def enable_forward_chunking(self, chunk_size: Optional[int] = None, di... method forward (line 357) | def forward( FILE: diffusers/src/diffusers/models/unets/unet_stable_cascade.py class SDCascadeLayerNorm (line 31) | class SDCascadeLayerNorm(nn.LayerNorm): method __init__ (line 32) | def __init__(self, *args, **kwargs): method forward (line 35) | def forward(self, x): class SDCascadeTimestepBlock (line 41) | class SDCascadeTimestepBlock(nn.Module): method __init__ (line 42) | def __init__(self, c, c_timestep, conds=[]): method forward (line 50) | def forward(self, x, t): class SDCascadeResBlock (line 59) | class SDCascadeResBlock(nn.Module): method __init__ (line 60) | def __init__(self, c, c_skip=0, kernel_size=3, dropout=0.0): method forward (line 72) | def forward(self, x, x_skip=None): class GlobalResponseNorm (line 82) | class GlobalResponseNorm(nn.Module): method __init__ (line 83) | def __init__(self, dim): method forward (line 88) | def forward(self, x): class SDCascadeAttnBlock (line 94) | class SDCascadeAttnBlock(nn.Module): method __init__ (line 95) | def __init__(self, c, c_cond, nhead, self_attn=True, dropout=0.0): method forward (line 103) | def forward(self, x, kv): class UpDownBlock2d (line 113) | class UpDownBlock2d(nn.Module): method __init__ (line 114) | def __init__(self, in_channels, out_channels, mode, enabled=True): method forward (line 126) | def forward(self, x): class StableCascadeUNetOutput (line 133) | class StableCascadeUNetOutput(BaseOutput): class StableCascadeUNet (line 137) | class StableCascadeUNet(ModelMixin, ConfigMixin, FromOriginalModelMixin): method __init__ (line 141) | def __init__( method _set_gradient_checkpointing (line 390) | def _set_gradient_checkpointing(self, value=False): method _init_weights (line 393) | def _init_weights(self, m): method get_timestep_ratio_embedding (line 422) | def get_timestep_ratio_embedding(self, timestep_ratio, max_positions=1... method get_clip_embeddings (line 436) | def get_clip_embeddings(self, clip_txt_pooled, clip_txt=None, clip_img... method _down_encode (line 454) | def _down_encode(self, x, r_embed, clip): method _up_decode (line 505) | def _up_decode(self, level_outputs, r_embed, clip): method forward (line 568) | def forward( FILE: diffusers/src/diffusers/models/unets/uvit_2d.py class UVit2DModel (line 39) | class UVit2DModel(ModelMixin, ConfigMixin, PeftAdapterMixin): method __init__ (line 43) | def __init__( method _set_gradient_checkpointing (line 151) | def _set_gradient_checkpointing(self, module, value: bool = False) -> ... method forward (line 154) | def forward(self, input_ids, encoder_hidden_states, pooled_text_emb, m... method attn_processors (line 217) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 241) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 276) | def set_default_attn_processor(self): class UVit2DConvEmbed (line 292) | class UVit2DConvEmbed(nn.Module): method __init__ (line 293) | def __init__(self, in_channels, block_out_channels, vocab_size, elemen... method forward (line 299) | def forward(self, input_ids): class UVitBlock (line 307) | class UVitBlock(nn.Module): method __init__ (line 308) | def __init__( method forward (line 386) | def forward(self, x, pooled_text_emb, encoder_hidden_states, cross_att... class ConvNextBlock (line 406) | class ConvNextBlock(nn.Module): method __init__ (line 407) | def __init__( method forward (line 427) | def forward(self, x, cond_embeds): class ConvMlmLayer (line 451) | class ConvMlmLayer(nn.Module): method __init__ (line 452) | def __init__( method forward (line 466) | def forward(self, hidden_states): FILE: diffusers/src/diffusers/models/upsampling.py class Upsample1D (line 25) | class Upsample1D(nn.Module): method __init__ (line 41) | def __init__( method forward (line 62) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: class Upsample2D (line 75) | class Upsample2D(nn.Module): method __init__ (line 91) | def __init__( method forward (line 141) | def forward(self, hidden_states: torch.Tensor, output_size: Optional[i... class FirUpsample2D (line 187) | class FirUpsample2D(nn.Module): method __init__ (line 201) | def __init__( method _upsample_2d (line 216) | def _upsample_2d( method forward (line 311) | def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: class KUpsample2D (line 321) | class KUpsample2D(nn.Module): method __init__ (line 328) | def __init__(self, pad_mode: str = "reflect"): method forward (line 335) | def forward(self, inputs: torch.Tensor) -> torch.Tensor: function upfirdn2d_native (line 351) | def upfirdn2d_native( function upsample_2d (line 401) | def upsample_2d( FILE: diffusers/src/diffusers/models/vae_flax.py class FlaxDecoderOutput (line 33) | class FlaxDecoderOutput(BaseOutput): class FlaxAutoencoderKLOutput (line 48) | class FlaxAutoencoderKLOutput(BaseOutput): class FlaxUpsample2D (line 61) | class FlaxUpsample2D(nn.Module): method setup (line 75) | def setup(self): method __call__ (line 84) | def __call__(self, hidden_states): class FlaxDownsample2D (line 95) | class FlaxDownsample2D(nn.Module): method setup (line 109) | def setup(self): method __call__ (line 118) | def __call__(self, hidden_states): class FlaxResnetBlock2D (line 125) | class FlaxResnetBlock2D(nn.Module): method setup (line 151) | def setup(self): method __call__ (line 185) | def __call__(self, hidden_states, deterministic=True): class FlaxAttentionBlock (line 202) | class FlaxAttentionBlock(nn.Module): method setup (line 223) | def setup(self): method transpose_for_scores (line 232) | def transpose_for_scores(self, projection): method __call__ (line 240) | def __call__(self, hidden_states): class FlaxDownEncoderBlock2D (line 275) | class FlaxDownEncoderBlock2D(nn.Module): method setup (line 304) | def setup(self): method __call__ (line 322) | def __call__(self, hidden_states, deterministic=True): class FlaxUpDecoderBlock2D (line 332) | class FlaxUpDecoderBlock2D(nn.Module): method setup (line 361) | def setup(self): method __call__ (line 379) | def __call__(self, hidden_states, deterministic=True): class FlaxUNetMidBlock2D (line 389) | class FlaxUNetMidBlock2D(nn.Module): method setup (line 415) | def setup(self): method __call__ (line 452) | def __call__(self, hidden_states, deterministic=True): class FlaxEncoder (line 461) | class FlaxEncoder(nn.Module): method setup (line 506) | def setup(self): method __call__ (line 555) | def __call__(self, sample, deterministic: bool = True): class FlaxDecoder (line 574) | class FlaxDecoder(nn.Module): method setup (line 618) | def setup(self): method __call__ (line 671) | def __call__(self, sample, deterministic: bool = True): class FlaxDiagonalGaussianDistribution (line 689) | class FlaxDiagonalGaussianDistribution(object): method __init__ (line 690) | def __init__(self, parameters, deterministic=False): method sample (line 700) | def sample(self, key): method kl (line 703) | def kl(self, other=None): method nll (line 715) | def nll(self, sample, axis=[1, 2, 3]): method mode (line 722) | def mode(self): class FlaxAutoencoderKL (line 727) | class FlaxAutoencoderKL(nn.Module, FlaxModelMixin, ConfigMixin): method setup (line 790) | def setup(self): method init_weights (line 827) | def init_weights(self, rng: jax.Array) -> FrozenDict: method encode (line 837) | def encode(self, sample, deterministic: bool = True, return_dict: bool... method decode (line 849) | def decode(self, latents, deterministic: bool = True, return_dict: boo... method __call__ (line 863) | def __call__(self, sample, sample_posterior=False, deterministic: bool... FILE: diffusers/src/diffusers/models/vq_model.py class VQEncoderOutput (line 18) | class VQEncoderOutput(VQEncoderOutput): class VQModel (line 23) | class VQModel(VQModel): FILE: diffusers/src/diffusers/optimization.py class SchedulerType (line 30) | class SchedulerType(Enum): function get_constant_schedule (line 40) | def get_constant_schedule(optimizer: Optimizer, last_epoch: int = -1) ->... function get_constant_schedule_with_warmup (line 56) | def get_constant_schedule_with_warmup(optimizer: Optimizer, num_warmup_s... function get_piecewise_constant_schedule (line 81) | def get_piecewise_constant_schedule(optimizer: Optimizer, step_rules: st... function get_linear_schedule_with_warmup (line 123) | def get_linear_schedule_with_warmup( function get_cosine_schedule_with_warmup (line 154) | def get_cosine_schedule_with_warmup( function get_cosine_with_hard_restarts_schedule_with_warmup (line 188) | def get_cosine_with_hard_restarts_schedule_with_warmup( function get_polynomial_decay_schedule_with_warmup (line 223) | def get_polynomial_decay_schedule_with_warmup( function get_scheduler (line 289) | def get_scheduler( FILE: diffusers/src/diffusers/pipelines/amused/pipeline_amused.py class AmusedPipeline (line 42) | class AmusedPipeline(DiffusionPipeline): method __init__ (line 52) | def __init__( method __call__ (line 74) | def __call__( FILE: diffusers/src/diffusers/pipelines/amused/pipeline_amused_img2img.py class AmusedImg2ImgPipeline (line 52) | class AmusedImg2ImgPipeline(DiffusionPipeline): method __init__ (line 67) | def __init__( method __call__ (line 89) | def __call__( FILE: diffusers/src/diffusers/pipelines/amused/pipeline_amused_inpaint.py class AmusedInpaintPipeline (line 60) | class AmusedInpaintPipeline(DiffusionPipeline): method __init__ (line 75) | def __init__( method __call__ (line 105) | def __call__( FILE: diffusers/src/diffusers/pipelines/animatediff/pipeline_animatediff.py class AnimateDiffPipeline (line 68) | class AnimateDiffPipeline( method __init__ (line 108) | def __init__( method encode_prompt (line 144) | def encode_prompt( method encode_image (line 327) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 352) | def prepare_ip_adapter_image_embeds( method decode_latents (line 404) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 417) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 435) | def check_inputs( method prepare_latents (line 505) | def prepare_latents( method guidance_scale (line 531) | def guidance_scale(self): method clip_skip (line 535) | def clip_skip(self): method do_classifier_free_guidance (line 542) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 546) | def cross_attention_kwargs(self): method num_timesteps (line 550) | def num_timesteps(self): method __call__ (line 555) | def __call__( FILE: diffusers/src/diffusers/pipelines/animatediff/pipeline_animatediff_sdxl.py function rescale_noise_cfg (line 117) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 132) | def retrieve_timesteps( class AnimateDiffSDXLPipeline (line 191) | class AnimateDiffSDXLPipeline( method __init__ (line 261) | def __init__( method encode_prompt (line 306) | def encode_prompt( method encode_image (line 541) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 566) | def prepare_ip_adapter_image_embeds( method decode_latents (line 618) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 631) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 648) | def check_inputs( method prepare_latents (line 721) | def prepare_latents( method _get_add_time_ids (line 746) | def _get_add_time_ids( method upcast_vae (line 764) | def upcast_vae(self): method get_guidance_scale_embedding (line 785) | def get_guidance_scale_embedding( method guidance_scale (line 816) | def guidance_scale(self): method guidance_rescale (line 820) | def guidance_rescale(self): method clip_skip (line 824) | def clip_skip(self): method do_classifier_free_guidance (line 831) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 835) | def cross_attention_kwargs(self): method denoising_end (line 839) | def denoising_end(self): method num_timesteps (line 843) | def num_timesteps(self): method interrupt (line 847) | def interrupt(self): method __call__ (line 852) | def __call__( FILE: diffusers/src/diffusers/pipelines/animatediff/pipeline_animatediff_video2video.py function retrieve_latents (line 99) | def retrieve_latents( function retrieve_timesteps (line 113) | def retrieve_timesteps( class AnimateDiffVideoToVideoPipeline (line 172) | class AnimateDiffVideoToVideoPipeline( method __init__ (line 212) | def __init__( method encode_prompt (line 248) | def encode_prompt( method encode_image (line 431) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 456) | def prepare_ip_adapter_image_embeds( method decode_latents (line 508) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 521) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 538) | def check_inputs( method get_timesteps (line 610) | def get_timesteps(self, num_inference_steps, timesteps, strength, devi... method prepare_latents (line 619) | def prepare_latents( method guidance_scale (line 707) | def guidance_scale(self): method clip_skip (line 711) | def clip_skip(self): method do_classifier_free_guidance (line 718) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 722) | def cross_attention_kwargs(self): method num_timesteps (line 726) | def num_timesteps(self): method __call__ (line 730) | def __call__( FILE: diffusers/src/diffusers/pipelines/animatediff/pipeline_output.py class AnimateDiffPipelineOutput (line 12) | class AnimateDiffPipelineOutput(BaseOutput): FILE: diffusers/src/diffusers/pipelines/audioldm/pipeline_audioldm.py class AudioLDMPipeline (line 52) | class AudioLDMPipeline(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 78) | def __init__( method _encode_prompt (line 99) | def _encode_prompt( method decode_latents (line 238) | def decode_latents(self, latents): method mel_spectrogram_to_waveform (line 243) | def mel_spectrogram_to_waveform(self, mel_spectrogram): method prepare_extra_step_kwargs (line 253) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 270) | def check_inputs( method prepare_latents (line 329) | def prepare_latents(self, batch_size, num_channels_latents, height, dt... method __call__ (line 353) | def __call__( FILE: diffusers/src/diffusers/pipelines/audioldm2/modeling_audioldm2.py function add_special_tokens (line 47) | def add_special_tokens(hidden_states, attention_mask, sos_token, eos_tok... class AudioLDM2ProjectionModelOutput (line 63) | class AudioLDM2ProjectionModelOutput(BaseOutput): class AudioLDM2ProjectionModel (line 82) | class AudioLDM2ProjectionModel(ModelMixin, ConfigMixin): method __init__ (line 98) | def __init__( method forward (line 126) | def forward( class AudioLDM2UNet2DConditionModel (line 167) | class AudioLDM2UNet2DConditionModel(ModelMixin, ConfigMixin, UNet2DCondi... method __init__ (line 242) | def __init__( method attn_processors (line 536) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 560) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 595) | def set_default_attn_processor(self): method set_attention_slice (line 611) | def set_attention_slice(self, slice_size): method _set_gradient_checkpointing (line 677) | def _set_gradient_checkpointing(self, module, value=False): method forward (line 681) | def forward( function get_down_block (line 892) | def get_down_block( function get_up_block (line 949) | def get_up_block( class CrossAttnDownBlock2D (line 1006) | class CrossAttnDownBlock2D(nn.Module): method __init__ (line 1007) | def __init__( method forward (line 1092) | def forward( class UNetMidBlock2DCrossAttn (line 1185) | class UNetMidBlock2DCrossAttn(nn.Module): method __init__ (line 1186) | def __init__( method forward (line 1271) | def forward( class CrossAttnUpBlock2D (line 1356) | class CrossAttnUpBlock2D(nn.Module): method __init__ (line 1357) | def __init__( method forward (line 1438) | def forward( FILE: diffusers/src/diffusers/pipelines/audioldm2/pipeline_audioldm2.py function prepare_inputs_for_generation (line 118) | def prepare_inputs_for_generation( class AudioLDM2Pipeline (line 136) | class AudioLDM2Pipeline(DiffusionPipeline): method __init__ (line 181) | def __init__( method enable_vae_slicing (line 213) | def enable_vae_slicing(self): method disable_vae_slicing (line 221) | def disable_vae_slicing(self): method enable_model_cpu_offload (line 228) | def enable_model_cpu_offload(self, gpu_id=0): method generate_language_model (line 265) | def generate_language_model( method encode_prompt (line 306) | def encode_prompt( method mel_spectrogram_to_waveform (line 639) | def mel_spectrogram_to_waveform(self, mel_spectrogram): method score_waveforms (line 648) | def score_waveforms(self, text, audio, num_waveforms_per_prompt, devic... method prepare_extra_step_kwargs (line 673) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 690) | def check_inputs( method prepare_latents (line 789) | def prepare_latents(self, batch_size, num_channels_latents, height, dt... method __call__ (line 813) | def __call__( FILE: diffusers/src/diffusers/pipelines/auto_pipeline.py function _get_connected_pipeline (line 146) | def _get_connected_pipeline(pipeline_cls): function _get_task_class (line 160) | def _get_task_class(mapping, pipeline_class_name, throw_error_if_not_exi... class AutoPipelineForText2Image (line 178) | class AutoPipelineForText2Image(ConfigMixin): method __init__ (line 196) | def __init__(self, *args, **kwargs): method from_pretrained (line 205) | def from_pretrained(cls, pretrained_model_or_path, **kwargs): method from_pipe (line 350) | def from_pipe(cls, pipeline, **kwargs): class AutoPipelineForImage2Image (line 450) | class AutoPipelineForImage2Image(ConfigMixin): method __init__ (line 468) | def __init__(self, *args, **kwargs): method from_pretrained (line 477) | def from_pretrained(cls, pretrained_model_or_path, **kwargs): method from_pipe (line 623) | def from_pipe(cls, pipeline, **kwargs): class AutoPipelineForInpainting (line 727) | class AutoPipelineForInpainting(ConfigMixin): method __init__ (line 745) | def __init__(self, *args, **kwargs): method from_pretrained (line 754) | def from_pretrained(cls, pretrained_model_or_path, **kwargs): method from_pipe (line 899) | def from_pipe(cls, pipeline, **kwargs): FILE: diffusers/src/diffusers/pipelines/blip_diffusion/blip_image_processing.py class BlipImageProcessor (line 49) | class BlipImageProcessor(BaseImageProcessor): method __init__ (line 86) | def __init__( method resize (line 116) | def resize( method preprocess (line 164) | def preprocess( method postprocess (line 301) | def postprocess(self, sample: torch.Tensor, output_type: str = "pil"): FILE: diffusers/src/diffusers/pipelines/blip_diffusion/modeling_blip2.py class Blip2TextEmbeddings (line 47) | class Blip2TextEmbeddings(nn.Module): method __init__ (line 50) | def __init__(self, config): method forward (line 66) | def forward( class Blip2VisionEmbeddings (line 101) | class Blip2VisionEmbeddings(nn.Module): method __init__ (line 102) | def __init__(self, config: Blip2VisionConfig): method forward (line 120) | def forward(self, pixel_values: torch.Tensor) -> torch.Tensor: class Blip2QFormerEncoder (line 133) | class Blip2QFormerEncoder(nn.Module): method __init__ (line 134) | def __init__(self, config): method forward (line 142) | def forward( class Blip2QFormerLayer (line 236) | class Blip2QFormerLayer(nn.Module): method __init__ (line 237) | def __init__(self, config, layer_idx): method forward (line 256) | def forward( method feed_forward_chunk (line 327) | def feed_forward_chunk(self, attention_output): method feed_forward_chunk_query (line 332) | def feed_forward_chunk_query(self, attention_output): class ProjLayer (line 339) | class ProjLayer(nn.Module): method __init__ (line 340) | def __init__(self, in_dim, out_dim, hidden_dim, drop_p=0.1, eps=1e-12): method forward (line 351) | def forward(self, x): class Blip2VisionModel (line 361) | class Blip2VisionModel(Blip2PreTrainedModel): method __init__ (line 365) | def __init__(self, config: Blip2VisionConfig): method forward (line 377) | def forward( method get_input_embeddings (line 421) | def get_input_embeddings(self): class Blip2QFormerModel (line 426) | class Blip2QFormerModel(Blip2PreTrainedModel): method __init__ (line 431) | def __init__(self, config: Blip2Config): method get_input_embeddings (line 454) | def get_input_embeddings(self): method set_input_embeddings (line 457) | def set_input_embeddings(self, value): method _prune_heads (line 460) | def _prune_heads(self, heads_to_prune): method get_extended_attention_mask (line 468) | def get_extended_attention_mask( method forward (line 513) | def forward( FILE: diffusers/src/diffusers/pipelines/blip_diffusion/modeling_ctx_clip.py function _expand_mask (line 25) | def _expand_mask(mask: torch.Tensor, dtype: torch.dtype, tgt_len: Option... class ContextCLIPTextModel (line 42) | class ContextCLIPTextModel(CLIPPreTrainedModel): method __init__ (line 47) | def __init__(self, config: CLIPTextConfig): method forward (line 53) | def forward( class ContextCLIPTextTransformer (line 76) | class ContextCLIPTextTransformer(nn.Module): method __init__ (line 77) | def __init__(self, config: CLIPTextConfig): method forward (line 85) | def forward( method _build_causal_attention_mask (line 162) | def _build_causal_attention_mask(self, bsz, seq_len, dtype): class ContextCLIPTextEmbeddings (line 172) | class ContextCLIPTextEmbeddings(nn.Module): method __init__ (line 173) | def __init__(self, config: CLIPTextConfig): method forward (line 183) | def forward( FILE: diffusers/src/diffusers/pipelines/blip_diffusion/pipeline_blip_diffusion.py class BlipDiffusionPipeline (line 75) | class BlipDiffusionPipeline(DiffusionPipeline): method __init__ (line 103) | def __init__( method get_query_embeddings (line 129) | def get_query_embeddings(self, input_image, src_subject): method _build_prompt (line 133) | def _build_prompt(self, prompts, tgt_subjects, prompt_strength=1.0, pr... method prepare_latents (line 143) | def prepare_latents(self, batch_size, num_channels, height, width, dty... method encode_prompt (line 160) | def encode_prompt(self, query_embeds, prompt, device=None): method __call__ (line 188) | def __call__( FILE: diffusers/src/diffusers/pipelines/consistency_models/pipeline_consistency_models.py class ConsistencyModelPipeline (line 63) | class ConsistencyModelPipeline(DiffusionPipeline): method __init__ (line 80) | def __init__(self, unet: UNet2DModel, scheduler: CMStochasticIterative... method prepare_latents (line 90) | def prepare_latents(self, batch_size, num_channels, height, width, dty... method postprocess_image (line 108) | def postprocess_image(self, sample: torch.Tensor, output_type: str = "... method prepare_class_labels (line 128) | def prepare_class_labels(self, batch_size, device, class_labels=None): method check_inputs (line 144) | def check_inputs(self, num_inference_steps, timesteps, latents, batch_... method __call__ (line 169) | def __call__( FILE: diffusers/src/diffusers/pipelines/controlnet/multicontrolnet.py class MultiControlNetModel (line 15) | class MultiControlNetModel(ModelMixin): method __init__ (line 28) | def __init__(self, controlnets: Union[List[ControlNetModel], Tuple[Con... method forward (line 32) | def forward( method save_pretrained (line 75) | def save_pretrained( method from_pretrained (line 114) | def from_pretrained(cls, pretrained_model_path: Optional[Union[str, os... FILE: diffusers/src/diffusers/pipelines/controlnet/pipeline_controlnet.py function retrieve_timesteps (line 96) | def retrieve_timesteps( class StableDiffusionControlNetPipeline (line 155) | class StableDiffusionControlNetPipeline( method __init__ (line 205) | def __init__( method _encode_prompt (line 258) | def _encode_prompt( method encode_prompt (line 291) | def encode_prompt( method encode_image (line 474) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 499) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 551) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 566) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 578) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 595) | def check_inputs( method check_image (line 758) | def check_image(self, image, prompt, prompt_embeds): method prepare_image (line 795) | def prepare_image( method prepare_latents (line 826) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 849) | def get_guidance_scale_embedding( method guidance_scale (line 880) | def guidance_scale(self): method clip_skip (line 884) | def clip_skip(self): method do_classifier_free_guidance (line 891) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 895) | def cross_attention_kwargs(self): method num_timesteps (line 899) | def num_timesteps(self): method __call__ (line 904) | def __call__( FILE: diffusers/src/diffusers/pipelines/controlnet/pipeline_controlnet_blip_diffusion.py class BlipDiffusionControlNetPipeline (line 82) | class BlipDiffusionControlNetPipeline(DiffusionPipeline): method __init__ (line 112) | def __init__( method get_query_embeddings (line 140) | def get_query_embeddings(self, input_image, src_subject): method _build_prompt (line 144) | def _build_prompt(self, prompts, tgt_subjects, prompt_strength=1.0, pr... method prepare_latents (line 154) | def prepare_latents(self, batch_size, num_channels, height, width, dty... method encode_prompt (line 171) | def encode_prompt(self, query_embeds, prompt, device=None): method prepare_control_image (line 198) | def prepare_control_image( method __call__ (line 236) | def __call__( FILE: diffusers/src/diffusers/pipelines/controlnet/pipeline_controlnet_img2img.py function retrieve_latents (line 96) | def retrieve_latents( function prepare_image (line 109) | def prepare_image(image): class StableDiffusionControlNetImg2ImgPipeline (line 133) | class StableDiffusionControlNetImg2ImgPipeline( method __init__ (line 183) | def __init__( method _encode_prompt (line 236) | def _encode_prompt( method encode_prompt (line 269) | def encode_prompt( method encode_image (line 452) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 477) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 529) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 544) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 556) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 573) | def check_inputs( method check_image (line 731) | def check_image(self, image, prompt, prompt_embeds): method prepare_control_image (line 769) | def prepare_control_image( method get_timesteps (line 800) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 812) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method guidance_scale (line 871) | def guidance_scale(self): method clip_skip (line 875) | def clip_skip(self): method do_classifier_free_guidance (line 882) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 886) | def cross_attention_kwargs(self): method num_timesteps (line 890) | def num_timesteps(self): method __call__ (line 895) | def __call__( FILE: diffusers/src/diffusers/pipelines/controlnet/pipeline_controlnet_inpaint.py function retrieve_latents (line 108) | def retrieve_latents( class StableDiffusionControlNetInpaintPipeline (line 121) | class StableDiffusionControlNetInpaintPipeline( method __init__ (line 182) | def __init__( method _encode_prompt (line 238) | def _encode_prompt( method encode_prompt (line 271) | def encode_prompt( method encode_image (line 454) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 479) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 531) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 546) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 558) | def prepare_extra_step_kwargs(self, generator, eta): method get_timesteps (line 576) | def get_timesteps(self, num_inference_steps, strength, device): method check_inputs (line 587) | def check_inputs( method check_image (line 766) | def check_image(self, image, prompt, prompt_embeds): method prepare_control_image (line 803) | def prepare_control_image( method prepare_latents (line 838) | def prepare_latents( method prepare_mask_latents (line 902) | def prepare_mask_latents( method _encode_vae_image (line 948) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method guidance_scale (line 963) | def guidance_scale(self): method clip_skip (line 967) | def clip_skip(self): method do_classifier_free_guidance (line 974) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 978) | def cross_attention_kwargs(self): method num_timesteps (line 982) | def num_timesteps(self): method __call__ (line 987) | def __call__( FILE: diffusers/src/diffusers/pipelines/controlnet/pipeline_controlnet_inpaint_sd_xl.py function retrieve_latents (line 70) | def retrieve_latents( function rescale_noise_cfg (line 140) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class StableDiffusionXLControlNetInpaintPipeline (line 154) | class StableDiffusionXLControlNetInpaintPipeline( method __init__ (line 222) | def __init__( method encode_prompt (line 274) | def encode_prompt( method encode_image (line 509) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 534) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 586) | def prepare_extra_step_kwargs(self, generator, eta): method check_image (line 603) | def check_image(self, image, prompt, prompt_embeds): method check_inputs (line 640) | def check_inputs( method prepare_control_image (line 857) | def prepare_control_image( method prepare_latents (line 891) | def prepare_latents( method _encode_vae_image (line 958) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method prepare_mask_latents (line 981) | def prepare_mask_latents( method get_timesteps (line 1029) | def get_timesteps(self, num_inference_steps, strength, device, denoisi... method _get_add_time_ids (line 1065) | def _get_add_time_ids( method upcast_vae (line 1112) | def upcast_vae(self): method guidance_scale (line 1132) | def guidance_scale(self): method clip_skip (line 1136) | def clip_skip(self): method do_classifier_free_guidance (line 1143) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 1147) | def cross_attention_kwargs(self): method num_timesteps (line 1151) | def num_timesteps(self): method __call__ (line 1156) | def __call__( FILE: diffusers/src/diffusers/pipelines/controlnet/pipeline_controlnet_sd_xl.py function retrieve_timesteps (line 119) | def retrieve_timesteps( class StableDiffusionXLControlNetPipeline (line 178) | class StableDiffusionXLControlNetPipeline( method __init__ (line 249) | def __init__( method encode_prompt (line 296) | def encode_prompt( method encode_image (line 531) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 556) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 608) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 625) | def check_inputs( method check_image (line 815) | def check_image(self, image, prompt, prompt_embeds): method prepare_image (line 853) | def prepare_image( method prepare_latents (line 884) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 907) | def _get_add_time_ids( method upcast_vae (line 926) | def upcast_vae(self): method get_guidance_scale_embedding (line 946) | def get_guidance_scale_embedding( method guidance_scale (line 977) | def guidance_scale(self): method clip_skip (line 981) | def clip_skip(self): method do_classifier_free_guidance (line 988) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 992) | def cross_attention_kwargs(self): method denoising_end (line 996) | def denoising_end(self): method num_timesteps (line 1000) | def num_timesteps(self): method __call__ (line 1005) | def __call__( FILE: diffusers/src/diffusers/pipelines/controlnet/pipeline_controlnet_sd_xl_img2img.py function retrieve_latents (line 148) | def retrieve_latents( class StableDiffusionXLControlNetImg2ImgPipeline (line 161) | class StableDiffusionXLControlNetImg2ImgPipeline( method __init__ (line 241) | def __init__( method encode_prompt (line 290) | def encode_prompt( method encode_image (line 525) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 550) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 602) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 619) | def check_inputs( method check_image (line 821) | def check_image(self, image, prompt, prompt_embeds): method prepare_control_image (line 859) | def prepare_control_image( method get_timesteps (line 890) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 902) | def prepare_latents( method _get_add_time_ids (line 982) | def _get_add_time_ids( method upcast_vae (line 1034) | def upcast_vae(self): method guidance_scale (line 1054) | def guidance_scale(self): method clip_skip (line 1058) | def clip_skip(self): method do_classifier_free_guidance (line 1065) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 1069) | def cross_attention_kwargs(self): method num_timesteps (line 1073) | def num_timesteps(self): method __call__ (line 1078) | def __call__( FILE: diffusers/src/diffusers/pipelines/controlnet/pipeline_flax_controlnet.py class FlaxStableDiffusionControlNetPipeline (line 111) | class FlaxStableDiffusionControlNetPipeline(FlaxDiffusionPipeline): method __init__ (line 141) | def __init__( method prepare_text_inputs (line 180) | def prepare_text_inputs(self, prompt: Union[str, List[str]]): method prepare_image_inputs (line 194) | def prepare_image_inputs(self, image: Union[Image.Image, List[Image.Im... method _get_has_nsfw_concepts (line 205) | def _get_has_nsfw_concepts(self, features, params): method _run_safety_checker (line 209) | def _run_safety_checker(self, images, safety_model_params, jit=False): method _generate (line 239) | def _generate( method __call__ (line 349) | def __call__( function _p_generate (line 488) | def _p_generate( function _p_get_has_nsfw_concepts (line 514) | def _p_get_has_nsfw_concepts(pipe, features, params): function unshard (line 518) | def unshard(x: jnp.ndarray): function preprocess (line 525) | def preprocess(image, dtype): FILE: diffusers/src/diffusers/pipelines/controlnet_xs/pipeline_controlnet_xs.py class StableDiffusionControlNetXSPipeline (line 92) | class StableDiffusionControlNetXSPipeline( method __init__ (line 134) | def __init__( method _encode_prompt (line 185) | def _encode_prompt( method encode_prompt (line 218) | def encode_prompt( method run_safety_checker (line 401) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 416) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 428) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 445) | def check_inputs( method check_image (line 515) | def check_image(self, image, prompt, prompt_embeds): method prepare_image (line 552) | def prepare_image( method prepare_latents (line 582) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method guidance_scale (line 606) | def guidance_scale(self): method clip_skip (line 611) | def clip_skip(self): method do_classifier_free_guidance (line 616) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 621) | def cross_attention_kwargs(self): method num_timesteps (line 626) | def num_timesteps(self): method __call__ (line 631) | def __call__( FILE: diffusers/src/diffusers/pipelines/controlnet_xs/pipeline_controlnet_xs_sd_xl.py class StableDiffusionXLControlNetXSPipeline (line 108) | class StableDiffusionXLControlNetXSPipeline( method __init__ (line 171) | def __init__( method encode_prompt (line 216) | def encode_prompt( method prepare_extra_step_kwargs (line 451) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 468) | def check_inputs( method check_image (line 565) | def check_image(self, image, prompt, prompt_embeds): method prepare_image (line 602) | def prepare_image( method prepare_latents (line 632) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 654) | def _get_add_time_ids( method upcast_vae (line 673) | def upcast_vae(self): method guidance_scale (line 694) | def guidance_scale(self): method clip_skip (line 699) | def clip_skip(self): method do_classifier_free_guidance (line 704) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 709) | def cross_attention_kwargs(self): method num_timesteps (line 714) | def num_timesteps(self): method __call__ (line 719) | def __call__( FILE: diffusers/src/diffusers/pipelines/dance_diffusion/pipeline_dance_diffusion.py class DanceDiffusionPipeline (line 28) | class DanceDiffusionPipeline(DiffusionPipeline): method __init__ (line 45) | def __init__(self, unet, scheduler): method __call__ (line 50) | def __call__( FILE: diffusers/src/diffusers/pipelines/ddim/pipeline_ddim.py class DDIMPipeline (line 24) | class DDIMPipeline(DiffusionPipeline): method __init__ (line 41) | def __init__(self, unet, scheduler): method __call__ (line 50) | def __call__( FILE: diffusers/src/diffusers/pipelines/ddpm/pipeline_ddpm.py class DDPMPipeline (line 24) | class DDPMPipeline(DiffusionPipeline): method __init__ (line 41) | def __init__(self, unet, scheduler): method __call__ (line 46) | def __call__( FILE: diffusers/src/diffusers/pipelines/deepfloyd_if/pipeline_if.py class IFPipeline (line 87) | class IFPipeline(DiffusionPipeline, LoraLoaderMixin): method __init__ (line 119) | def __init__( method encode_prompt (line 160) | def encode_prompt( method run_safety_checker (line 313) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 327) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 344) | def check_inputs( method prepare_intermediate_images (line 386) | def prepare_intermediate_images(self, batch_size, num_channels, height... method _text_preprocessing (line 400) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 424) | def _clean_caption(self, caption): method __call__ (line 540) | def __call__( FILE: diffusers/src/diffusers/pipelines/deepfloyd_if/pipeline_if_img2img.py function resize (line 39) | def resize(images: PIL.Image.Image, img_size: int) -> PIL.Image.Image: class IFImg2ImgPipeline (line 111) | class IFImg2ImgPipeline(DiffusionPipeline, LoraLoaderMixin): method __init__ (line 143) | def __init__( method encode_prompt (line 184) | def encode_prompt( method run_safety_checker (line 338) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 352) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 369) | def check_inputs( method _text_preprocessing (line 443) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 468) | def _clean_caption(self, caption): method preprocess_image (line 582) | def preprocess_image(self, image: PIL.Image.Image) -> torch.Tensor: method get_timesteps (line 619) | def get_timesteps(self, num_inference_steps, strength): method prepare_intermediate_images (line 630) | def prepare_intermediate_images( method __call__ (line 654) | def __call__( FILE: diffusers/src/diffusers/pipelines/deepfloyd_if/pipeline_if_img2img_superresolution.py function resize (line 42) | def resize(images: PIL.Image.Image, img_size: int) -> PIL.Image.Image: class IFImg2ImgSuperResolutionPipeline (line 114) | class IFImg2ImgSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin): method __init__ (line 147) | def __init__( method _text_preprocessing (line 195) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 220) | def _clean_caption(self, caption): method encode_prompt (line 336) | def encode_prompt( method run_safety_checker (line 490) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 504) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 521) | def check_inputs( method preprocess_original_image (line 631) | def preprocess_original_image(self, image: PIL.Image.Image) -> torch.T... method preprocess_image (line 668) | def preprocess_image(self, image: PIL.Image.Image, num_images_per_prom... method get_timesteps (line 700) | def get_timesteps(self, num_inference_steps, strength): method prepare_intermediate_images (line 712) | def prepare_intermediate_images( method __call__ (line 736) | def __call__( FILE: diffusers/src/diffusers/pipelines/deepfloyd_if/pipeline_if_inpainting.py function resize (line 40) | def resize(images: PIL.Image.Image, img_size: int) -> PIL.Image.Image: class IFInpaintingPipeline (line 114) | class IFInpaintingPipeline(DiffusionPipeline, LoraLoaderMixin): method __init__ (line 146) | def __init__( method encode_prompt (line 188) | def encode_prompt( method run_safety_checker (line 342) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 356) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 373) | def check_inputs( method _text_preprocessing (line 483) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 508) | def _clean_caption(self, caption): method preprocess_image (line 623) | def preprocess_image(self, image: PIL.Image.Image) -> torch.Tensor: method preprocess_mask_image (line 659) | def preprocess_mask_image(self, mask_image) -> torch.Tensor: method get_timesteps (line 709) | def get_timesteps(self, num_inference_steps, strength): method prepare_intermediate_images (line 720) | def prepare_intermediate_images( method __call__ (line 746) | def __call__( FILE: diffusers/src/diffusers/pipelines/deepfloyd_if/pipeline_if_inpainting_superresolution.py function resize (line 42) | def resize(images: PIL.Image.Image, img_size: int) -> PIL.Image.Image: class IFInpaintingSuperResolutionPipeline (line 116) | class IFInpaintingSuperResolutionPipeline(DiffusionPipeline, LoraLoaderM... method __init__ (line 149) | def __init__( method _text_preprocessing (line 197) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 222) | def _clean_caption(self, caption): method encode_prompt (line 338) | def encode_prompt( method run_safety_checker (line 492) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 506) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 523) | def check_inputs( method preprocess_original_image (line 667) | def preprocess_original_image(self, image: PIL.Image.Image) -> torch.T... method preprocess_image (line 704) | def preprocess_image(self, image: PIL.Image.Image, num_images_per_prom... method preprocess_mask_image (line 736) | def preprocess_mask_image(self, mask_image) -> torch.Tensor: method get_timesteps (line 786) | def get_timesteps(self, num_inference_steps, strength): method prepare_intermediate_images (line 798) | def prepare_intermediate_images( method __call__ (line 824) | def __call__( FILE: diffusers/src/diffusers/pipelines/deepfloyd_if/pipeline_if_superresolution.py class IFSuperResolutionPipeline (line 72) | class IFSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin): method __init__ (line 105) | def __init__( method _text_preprocessing (line 153) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 178) | def _clean_caption(self, caption): method encode_prompt (line 294) | def encode_prompt( method run_safety_checker (line 448) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 462) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 479) | def check_inputs( method prepare_intermediate_images (line 559) | def prepare_intermediate_images(self, batch_size, num_channels, height... method preprocess_image (line 573) | def preprocess_image(self, image, num_images_per_prompt, device): method __call__ (line 606) | def __call__( FILE: diffusers/src/diffusers/pipelines/deepfloyd_if/pipeline_output.py class IFPipelineOutput (line 11) | class IFPipelineOutput(BaseOutput): FILE: diffusers/src/diffusers/pipelines/deepfloyd_if/safety_checker.py class IFSafetyChecker (line 12) | class IFSafetyChecker(PreTrainedModel): method __init__ (line 17) | def __init__(self, config: CLIPConfig): method forward (line 26) | def forward(self, clip_input, images, p_threshold=0.5, w_threshold=0.5): FILE: diffusers/src/diffusers/pipelines/deepfloyd_if/watermark.py class IFWatermarker (line 12) | class IFWatermarker(ModelMixin, ConfigMixin): method __init__ (line 13) | def __init__(self): method apply_watermark (line 19) | def apply_watermark(self, images: List[PIL.Image.Image], sample_size=N... FILE: diffusers/src/diffusers/pipelines/deprecated/alt_diffusion/modeling_roberta_series.py class TransformationModelOutput (line 11) | class TransformationModelOutput(ModelOutput): class RobertaSeriesConfig (line 39) | class RobertaSeriesConfig(XLMRobertaConfig): method __init__ (line 40) | def __init__( class RobertaSeriesModelWithTransformation (line 58) | class RobertaSeriesModelWithTransformation(RobertaPreTrainedModel): method __init__ (line 64) | def __init__(self, config): method forward (line 74) | def forward( FILE: diffusers/src/diffusers/pipelines/deprecated/alt_diffusion/pipeline_alt_diffusion.py function rescale_noise_cfg (line 62) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 77) | def retrieve_timesteps( class AltDiffusionPipeline (line 136) | class AltDiffusionPipeline( method __init__ (line 182) | def __init__( method _encode_prompt (line 274) | def _encode_prompt( method encode_prompt (line 306) | def encode_prompt( method encode_image (line 487) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method run_safety_checker (line 511) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 525) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 536) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 553) | def check_inputs( method prepare_latents (line 605) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 627) | def get_guidance_scale_embedding(self, w, embedding_dim=512, dtype=tor... method guidance_scale (line 656) | def guidance_scale(self): method guidance_rescale (line 660) | def guidance_rescale(self): method clip_skip (line 664) | def clip_skip(self): method do_classifier_free_guidance (line 671) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 675) | def cross_attention_kwargs(self): method num_timesteps (line 679) | def num_timesteps(self): method __call__ (line 684) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/alt_diffusion/pipeline_alt_diffusion_img2img.py function retrieve_latents (line 79) | def retrieve_latents( function preprocess (line 93) | def preprocess(image): function retrieve_timesteps (line 117) | def retrieve_timesteps( class AltDiffusionImg2ImgPipeline (line 176) | class AltDiffusionImg2ImgPipeline( method __init__ (line 222) | def __init__( method _encode_prompt (line 314) | def _encode_prompt( method encode_prompt (line 346) | def encode_prompt( method encode_image (line 527) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method run_safety_checker (line 551) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 565) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 576) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 593) | def check_inputs( method get_timesteps (line 644) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 653) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method get_guidance_scale_embedding (line 711) | def get_guidance_scale_embedding(self, w, embedding_dim=512, dtype=tor... method guidance_scale (line 740) | def guidance_scale(self): method clip_skip (line 744) | def clip_skip(self): method do_classifier_free_guidance (line 751) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 755) | def cross_attention_kwargs(self): method num_timesteps (line 759) | def num_timesteps(self): method __call__ (line 764) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/alt_diffusion/pipeline_output.py class AltDiffusionPipelineOutput (line 14) | class AltDiffusionPipelineOutput(BaseOutput): FILE: diffusers/src/diffusers/pipelines/deprecated/audio_diffusion/mel.py class Mel (line 37) | class Mel(ConfigMixin, SchedulerMixin): method __init__ (line 59) | def __init__( method set_resolution (line 80) | def set_resolution(self, x_res: int, y_res: int): method load_audio (line 94) | def load_audio(self, audio_file: str = None, raw_audio: np.ndarray = N... method get_number_of_slices (line 112) | def get_number_of_slices(self) -> int: method get_audio_slice (line 121) | def get_audio_slice(self, slice: int = 0) -> np.ndarray: method get_sample_rate (line 134) | def get_sample_rate(self) -> int: method audio_slice_to_image (line 143) | def audio_slice_to_image(self, slice: int) -> Image.Image: method image_to_audio (line 162) | def image_to_audio(self, image: Image.Image) -> np.ndarray: FILE: diffusers/src/diffusers/pipelines/deprecated/audio_diffusion/pipeline_audio_diffusion.py class AudioDiffusionPipeline (line 30) | class AudioDiffusionPipeline(DiffusionPipeline): method __init__ (line 51) | def __init__( method get_default_steps (line 61) | def get_default_steps(self) -> int: method __call__ (line 71) | def __call__( method encode (line 271) | def encode(self, images: List[Image.Image], steps: int = 50) -> np.nda... method slerp (line 312) | def slerp(x0: torch.Tensor, x1: torch.Tensor, alpha: float) -> torch.T... FILE: diffusers/src/diffusers/pipelines/deprecated/latent_diffusion_uncond/pipeline_latent_diffusion_uncond.py class LDMPipeline (line 26) | class LDMPipeline(DiffusionPipeline): method __init__ (line 42) | def __init__(self, vqvae: VQModel, unet: UNet2DModel, scheduler: DDIMS... method __call__ (line 47) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/pndm/pipeline_pndm.py class PNDMPipeline (line 26) | class PNDMPipeline(DiffusionPipeline): method __init__ (line 43) | def __init__(self, unet: UNet2DModel, scheduler: PNDMScheduler): method __call__ (line 51) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/repaint/pipeline_repaint.py function _preprocess_image (line 33) | def _preprocess_image(image: Union[List, PIL.Image.Image, torch.Tensor]): function _preprocess_mask (line 56) | def _preprocess_mask(mask: Union[List, PIL.Image.Image, torch.Tensor]): class RePaintPipeline (line 76) | class RePaintPipeline(DiffusionPipeline): method __init__ (line 94) | def __init__(self, unet, scheduler): method __call__ (line 99) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/score_sde_ve/pipeline_score_sde_ve.py class ScoreSdeVePipeline (line 25) | class ScoreSdeVePipeline(DiffusionPipeline): method __init__ (line 42) | def __init__(self, unet: UNet2DModel, scheduler: ScoreSdeVeScheduler): method __call__ (line 47) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/spectrogram_diffusion/continuous_encoder.py class SpectrogramContEncoder (line 29) | class SpectrogramContEncoder(ModelMixin, ConfigMixin, ModuleUtilsMixin): method __init__ (line 31) | def __init__( method forward (line 71) | def forward(self, encoder_inputs, encoder_inputs_mask): FILE: diffusers/src/diffusers/pipelines/deprecated/spectrogram_diffusion/midi_utils.py class NoteRepresentationConfig (line 84) | class NoteRepresentationConfig: class NoteEventData (line 92) | class NoteEventData: class NoteEncodingState (line 101) | class NoteEncodingState: class EventRange (line 109) | class EventRange: class Event (line 116) | class Event: class Tokenizer (line 121) | class Tokenizer: method __init__ (line 122) | def __init__(self, regular_ids: int): method encode (line 127) | def encode(self, token_ids): class Codec (line 145) | class Codec: method __init__ (line 156) | def __init__(self, max_shift_steps: int, steps_per_second: float, even... method num_classes (line 172) | def num_classes(self) -> int: method is_shift_event_index (line 178) | def is_shift_event_index(self, index: int) -> bool: method max_shift_steps (line 182) | def max_shift_steps(self) -> int: method encode_event (line 185) | def encode_event(self, event: Event) -> int: method event_type_range (line 200) | def event_type_range(self, event_type: str) -> Tuple[int, int]: method decode_event_index (line 210) | def decode_event_index(self, index: int) -> Event: class ProgramGranularity (line 222) | class ProgramGranularity: function drop_programs (line 228) | def drop_programs(tokens, codec: Codec): function programs_to_midi_classes (line 234) | def programs_to_midi_classes(tokens, codec): function frame (line 254) | def frame(signal, frame_length, frame_step, pad_end=False, pad_value=0, ... function program_to_slakh_program (line 272) | def program_to_slakh_program(program): function audio_to_frames (line 279) | def audio_to_frames( function note_sequence_to_onsets_and_offsets_and_programs (line 302) | def note_sequence_to_onsets_and_offsets_and_programs( function num_velocity_bins_from_codec (line 332) | def num_velocity_bins_from_codec(codec: Codec): function segment (line 339) | def segment(a, n): function velocity_to_bin (line 343) | def velocity_to_bin(velocity, num_velocity_bins): function note_event_data_to_events (line 350) | def note_event_data_to_events( function note_encoding_state_to_events (line 382) | def note_encoding_state_to_events(state: NoteEncodingState) -> Sequence[... function encode_and_index_events (line 392) | def encode_and_index_events( function extract_sequence_with_indices (line 498) | def extract_sequence_with_indices(features, state_events_end_token=None,... function map_midi_programs (line 524) | def map_midi_programs( function run_length_encode_shifts_fn (line 534) | def run_length_encode_shifts_fn( function note_representation_processor_chain (line 604) | def note_representation_processor_chain(features, codec: Codec, note_rep... class MidiProcessor (line 619) | class MidiProcessor: method __init__ (line 620) | def __init__(self): method __call__ (line 635) | def __call__(self, midi: Union[bytes, os.PathLike, str]): FILE: diffusers/src/diffusers/pipelines/deprecated/spectrogram_diffusion/notes_encoder.py class SpectrogramNotesEncoder (line 25) | class SpectrogramNotesEncoder(ModelMixin, ConfigMixin, ModuleUtilsMixin): method __init__ (line 27) | def __init__( method forward (line 69) | def forward(self, encoder_input_tokens, encoder_inputs_mask): FILE: diffusers/src/diffusers/pipelines/deprecated/spectrogram_diffusion/pipeline_spectrogram_diffusion.py class SpectrogramDiffusionPipeline (line 41) | class SpectrogramDiffusionPipeline(DiffusionPipeline): method __init__ (line 60) | def __init__( method scale_features (line 83) | def scale_features(self, features, output_range=(-1.0, 1.0), clip=False): method scale_to_features (line 93) | def scale_to_features(self, outputs, input_range=(-1.0, 1.0), clip=Fal... method encode (line 102) | def encode(self, input_tokens, continuous_inputs, continuous_mask): method decode (line 114) | def decode(self, encodings_and_masks, input_tokens, noise_time): method __call__ (line 130) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/stable_diffusion_variants/pipeline_cycle_diffusion.py function preprocess (line 41) | def preprocess(image): function retrieve_latents (line 65) | def retrieve_latents( function posterior_sample (line 78) | def posterior_sample(scheduler, latents, timestep, clean_latents, genera... function compute_noise (line 105) | def compute_noise(scheduler, prev_latents, latents, timestep, noise_pred... class CycleDiffusionPipeline (line 139) | class CycleDiffusionPipeline(DiffusionPipeline, TextualInversionLoaderMi... method __init__ (line 174) | def __init__( method _encode_prompt (line 251) | def _encode_prompt( method encode_prompt (line 284) | def encode_prompt( method check_inputs (line 466) | def check_inputs( method prepare_extra_step_kwargs (line 518) | def prepare_extra_step_kwargs(self, generator, eta): method run_safety_checker (line 536) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 551) | def decode_latents(self, latents): method get_timesteps (line 563) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 574) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method __call__ (line 630) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/stable_diffusion_variants/pipeline_onnx_stable_diffusion_inpaint_legacy.py function preprocess (line 20) | def preprocess(image): function preprocess_mask (line 29) | def preprocess_mask(mask, scale_factor=8): class OnnxStableDiffusionInpaintPipelineLegacy (line 41) | class OnnxStableDiffusionInpaintPipelineLegacy(DiffusionPipeline): method __init__ (line 82) | def __init__( method _encode_prompt (line 152) | def _encode_prompt( method check_inputs (line 255) | def check_inputs( method __call__ (line 297) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/stable_diffusion_variants/pipeline_stable_diffusion_inpaint_legacy.py function preprocess_image (line 40) | def preprocess_image(image, batch_size): function preprocess_mask (line 50) | def preprocess_mask(mask, batch_size, scale_factor=8): class StableDiffusionInpaintPipelineLegacy (line 81) | class StableDiffusionInpaintPipelineLegacy( method __init__ (line 123) | def __init__( method _encode_prompt (line 221) | def _encode_prompt( method encode_prompt (line 254) | def encode_prompt( method run_safety_checker (line 437) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 452) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 464) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 481) | def check_inputs( method get_timesteps (line 532) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 541) | def prepare_latents(self, image, timestep, num_images_per_prompt, dtyp... method __call__ (line 558) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/stable_diffusion_variants/pipeline_stable_diffusion_model_editing.py class StableDiffusionModelEditingPipeline (line 39) | class StableDiffusionModelEditingPipeline( method __init__ (line 81) | def __init__( method _encode_prompt (line 159) | def _encode_prompt( method encode_prompt (line 192) | def encode_prompt( method run_safety_checker (line 375) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 390) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 402) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 419) | def check_inputs( method prepare_latents (line 472) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method edit_model (line 495) | def edit_model( method __call__ (line 613) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/stable_diffusion_variants/pipeline_stable_diffusion_paradigms.py class StableDiffusionParadigmsPipeline (line 65) | class StableDiffusionParadigmsPipeline( method __init__ (line 104) | def __init__( method _encode_prompt (line 150) | def _encode_prompt( method encode_prompt (line 183) | def encode_prompt( method run_safety_checker (line 366) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 381) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 398) | def check_inputs( method prepare_latents (line 451) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _cumsum (line 473) | def _cumsum(self, input, dim, debug=False): method __call__ (line 483) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/stable_diffusion_variants/pipeline_stable_diffusion_pix2pix_zero.py class Pix2PixInversionPipelineOutput (line 58) | class Pix2PixInversionPipelineOutput(BaseOutput, TextualInversionLoaderM... function preprocess (line 180) | def preprocess(image): function prepare_unet (line 203) | def prepare_unet(unet: UNet2DConditionModel): class Pix2PixZeroL2Loss (line 220) | class Pix2PixZeroL2Loss: method __init__ (line 221) | def __init__(self): method compute_loss (line 224) | def compute_loss(self, predictions, targets): class Pix2PixZeroAttnProcessor (line 228) | class Pix2PixZeroAttnProcessor: method __init__ (line 232) | def __init__(self, is_pix2pix_zero=False): method __call__ (line 237) | def __call__( class StableDiffusionPix2PixZeroPipeline (line 283) | class StableDiffusionPix2PixZeroPipeline(DiffusionPipeline, StableDiffus... method __init__ (line 324) | def __init__( method _encode_prompt (line 373) | def _encode_prompt( method encode_prompt (line 406) | def encode_prompt( method run_safety_checker (line 589) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 604) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 616) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 633) | def check_inputs( method prepare_latents (line 664) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method generate_caption (line 687) | def generate_caption(self, images): method construct_direction (line 706) | def construct_direction(self, embs_source: torch.Tensor, embs_target: ... method get_embeds (line 711) | def get_embeds(self, prompt: List[str], batch_size: int = 16) -> torch... method prepare_image_latents (line 730) | def prepare_image_latents(self, image, batch_size, dtype, device, gene... method get_epsilon (line 779) | def get_epsilon(self, model_output: torch.Tensor, sample: torch.Tensor... method auto_corr_loss (line 796) | def auto_corr_loss(self, hidden_states, generator=None): method kl_divergence (line 811) | def kl_divergence(self, hidden_states): method __call__ (line 818) | def __call__( method invert (line 1109) | def invert( FILE: diffusers/src/diffusers/pipelines/deprecated/stochastic_karras_ve/pipeline_stochastic_karras_ve.py class KarrasVePipeline (line 25) | class KarrasVePipeline(DiffusionPipeline): method __init__ (line 40) | def __init__(self, unet: UNet2DModel, scheduler: KarrasVeScheduler): method __call__ (line 45) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/versatile_diffusion/modeling_text_unet.py function get_down_block (line 44) | def get_down_block( function get_up_block (line 109) | def get_up_block( class FourierEmbedder (line 175) | class FourierEmbedder(nn.Module): method __init__ (line 176) | def __init__(self, num_freqs=64, temperature=100): method __call__ (line 186) | def __call__(self, x): class GLIGENTextBoundingboxProjection (line 191) | class GLIGENTextBoundingboxProjection(nn.Module): method __init__ (line 192) | def __init__(self, positive_len, out_dim, feature_type, fourier_freqs=8): method forward (line 233) | def forward( class UNetFlatConditionModel (line 271) | class UNetFlatConditionModel(ModelMixin, ConfigMixin): method __init__ (line 369) | def __init__( method attn_processors (line 829) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 852) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 886) | def set_default_attn_processor(self): method set_attention_slice (line 901) | def set_attention_slice(self, slice_size): method _set_gradient_checkpointing (line 966) | def _set_gradient_checkpointing(self, module, value=False): method enable_freeu (line 970) | def enable_freeu(self, s1, s2, b1, b2): method disable_freeu (line 994) | def disable_freeu(self): method fuse_qkv_projections (line 1002) | def fuse_qkv_projections(self): method unfuse_qkv_projections (line 1025) | def unfuse_qkv_projections(self): method unload_lora (line 1038) | def unload_lora(self): method forward (line 1049) | def forward( class LinearMultiDim (line 1427) | class LinearMultiDim(nn.Linear): method __init__ (line 1428) | def __init__(self, in_features, out_features=None, second_dim=4, *args... method forward (line 1437) | def forward(self, input_tensor, *args, **kwargs): class ResnetBlockFlat (line 1446) | class ResnetBlockFlat(nn.Module): method __init__ (line 1447) | def __init__( method forward (line 1507) | def forward(self, input_tensor, temb): class DownBlockFlat (line 1540) | class DownBlockFlat(nn.Module): method __init__ (line 1541) | def __init__( method forward (line 1592) | def forward( class CrossAttnDownBlockFlat (line 1628) | class CrossAttnDownBlockFlat(nn.Module): method __init__ (line 1629) | def __init__( method forward (line 1720) | def forward( class UpBlockFlat (line 1788) | class UpBlockFlat(nn.Module): method __init__ (line 1789) | def __init__( method forward (line 1838) | def forward( class CrossAttnUpBlockFlat (line 1904) | class CrossAttnUpBlockFlat(nn.Module): method __init__ (line 1905) | def __init__( method forward (line 1995) | def forward( class UNetMidBlockFlat (line 2081) | class UNetMidBlockFlat(nn.Module): method __init__ (line 2112) | def __init__( method forward (line 2226) | def forward(self, hidden_states: torch.Tensor, temb: Optional[torch.Te... class UNetMidBlockFlatCrossAttn (line 2237) | class UNetMidBlockFlatCrossAttn(nn.Module): method __init__ (line 2238) | def __init__( method forward (line 2340) | def forward( class UNetMidBlockFlatSimpleCrossAttn (line 2396) | class UNetMidBlockFlatSimpleCrossAttn(nn.Module): method __init__ (line 2397) | def __init__( method forward (line 2481) | def forward( FILE: diffusers/src/diffusers/pipelines/deprecated/versatile_diffusion/pipeline_versatile_diffusion.py class VersatileDiffusionPipeline (line 20) | class VersatileDiffusionPipeline(DiffusionPipeline): method __init__ (line 56) | def __init__( method image_variation (line 82) | def image_variation( method text_to_image (line 194) | def text_to_image( method dual_guided (line 302) | def dual_guided( FILE: diffusers/src/diffusers/pipelines/deprecated/versatile_diffusion/pipeline_versatile_diffusion_dual_guided.py class VersatileDiffusionDualGuidedPipeline (line 41) | class VersatileDiffusionDualGuidedPipeline(DiffusionPipeline): method __init__ (line 75) | def __init__( method remove_unused_weights (line 106) | def remove_unused_weights(self): method _convert_to_dual_attention (line 109) | def _convert_to_dual_attention(self): method _revert_dual_attention (line 143) | def _revert_dual_attention(self): method _encode_text_prompt (line 156) | def _encode_text_prompt(self, prompt, device, num_images_per_prompt, d... method _encode_image_prompt (line 247) | def _encode_image_prompt(self, prompt, device, num_images_per_prompt, ... method decode_latents (line 303) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 315) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 332) | def check_inputs(self, prompt, image, height, width, callback_steps): method prepare_latents (line 350) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method set_transformer_params (line 372) | def set_transformer_params(self, mix_ratio: float = 0.5, condition_typ... method __call__ (line 386) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/versatile_diffusion/pipeline_versatile_diffusion_image_variation.py class VersatileDiffusionImageVariationPipeline (line 35) | class VersatileDiffusionImageVariationPipeline(DiffusionPipeline): method __init__ (line 64) | def __init__( method _encode_prompt (line 83) | def _encode_prompt(self, prompt, device, num_images_per_prompt, do_cla... method decode_latents (line 163) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 175) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 193) | def check_inputs(self, image, height, width, callback_steps): method prepare_latents (line 216) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 239) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/versatile_diffusion/pipeline_versatile_diffusion_text_to_image.py class VersatileDiffusionTextToImagePipeline (line 34) | class VersatileDiffusionTextToImagePipeline(DiffusionPipeline): method __init__ (line 67) | def __init__( method _swap_unet_attention_blocks (line 91) | def _swap_unet_attention_blocks(self): method remove_unused_weights (line 104) | def remove_unused_weights(self): method _encode_prompt (line 107) | def _encode_prompt(self, prompt, device, num_images_per_prompt, do_cla... method decode_latents (line 220) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 232) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 249) | def check_inputs( method prepare_latents (line 302) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 325) | def __call__( FILE: diffusers/src/diffusers/pipelines/deprecated/vq_diffusion/pipeline_vq_diffusion.py class LearnedClassifierFreeSamplingEmbeddings (line 30) | class LearnedClassifierFreeSamplingEmbeddings(ModelMixin, ConfigMixin): method __init__ (line 36) | def __init__(self, learnable: bool, hidden_size: Optional[int] = None,... class VQDiffusionPipeline (line 52) | class VQDiffusionPipeline(DiffusionPipeline): method __init__ (line 80) | def __init__( method _encode_prompt (line 100) | def _encode_prompt(self, prompt, num_images_per_prompt, do_classifier_... method __call__ (line 164) | def __call__( method truncate (line 304) | def truncate(self, log_p_x_0: torch.Tensor, truncation_rate: float) ->... FILE: diffusers/src/diffusers/pipelines/dit/pipeline_dit.py class DiTPipeline (line 31) | class DiTPipeline(DiffusionPipeline): method __init__ (line 49) | def __init__( method get_label_ids (line 67) | def get_label_ids(self, label: Union[str, List[str]]) -> List[int]: method __call__ (line 93) | def __call__( FILE: diffusers/src/diffusers/pipelines/free_init_utils.py class FreeInitMixin (line 24) | class FreeInitMixin: method enable_free_init (line 27) | def enable_free_init( method disable_free_init (line 66) | def disable_free_init(self): method free_init_enabled (line 71) | def free_init_enabled(self): method _get_free_init_freq_filter (line 74) | def _get_free_init_freq_filter( method _apply_freq_filter (line 118) | def _apply_freq_filter(self, x: torch.Tensor, noise: torch.Tensor, low... method _apply_free_init (line 138) | def _apply_free_init( FILE: diffusers/src/diffusers/pipelines/hunyuandit/pipeline_hunyuandit.py function map_to_standard_shapes (line 98) | def map_to_standard_shapes(target_width, target_height): function get_resize_crop_region_for_grid (line 106) | def get_resize_crop_region_for_grid(src, tgt_size): function rescale_noise_cfg (line 127) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class HunyuanDiTPipeline (line 141) | class HunyuanDiTPipeline(DiffusionPipeline): method __init__ (line 188) | def __init__( method encode_prompt (line 242) | def encode_prompt( method run_safety_checker (line 408) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 423) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 440) | def check_inputs( method prepare_latents (line 517) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method guidance_scale (line 540) | def guidance_scale(self): method guidance_rescale (line 544) | def guidance_rescale(self): method do_classifier_free_guidance (line 551) | def do_classifier_free_guidance(self): method num_timesteps (line 555) | def num_timesteps(self): method interrupt (line 559) | def interrupt(self): method __call__ (line 564) | def __call__( FILE: diffusers/src/diffusers/pipelines/i2vgen_xl/pipeline_i2vgen_xl.py class I2VGenXLPipelineOutput (line 75) | class I2VGenXLPipelineOutput(BaseOutput): class I2VGenXLPipeline (line 90) | class I2VGenXLPipeline( method __init__ (line 115) | def __init__( method guidance_scale (line 141) | def guidance_scale(self): method do_classifier_free_guidance (line 148) | def do_classifier_free_guidance(self): method encode_prompt (line 151) | def encode_prompt( method _encode_image (line 320) | def _encode_image(self, image, device, num_videos_per_prompt): method decode_latents (line 352) | def decode_latents(self, latents, decode_chunk_size=None): method prepare_extra_step_kwargs (line 375) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 392) | def check_inputs( method prepare_image_latents (line 441) | def prepare_image_latents( method prepare_latents (line 474) | def prepare_latents( method __call__ (line 501) | def __call__( function _convert_pt_to_pil (line 734) | def _convert_pt_to_pil(image: Union[torch.Tensor, List[torch.Tensor]]): function _resize_bilinear (line 749) | def _resize_bilinear( function _center_crop_wide (line 762) | def _center_crop_wide( FILE: diffusers/src/diffusers/pipelines/kandinsky/pipeline_kandinsky.py function get_new_h_w (line 66) | def get_new_h_w(h, w, scale_factor=8): class KandinskyPipeline (line 76) | class KandinskyPipeline(DiffusionPipeline): method __init__ (line 98) | def __init__( method prepare_latents (line 118) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method _encode_prompt (line 129) | def _encode_prompt( method __call__ (line 233) | def __call__( FILE: diffusers/src/diffusers/pipelines/kandinsky/pipeline_kandinsky_combined.py class KandinskyCombinedPipeline (line 113) | class KandinskyCombinedPipeline(DiffusionPipeline): method __init__ (line 148) | def __init__( method enable_xformers_memory_efficient_attention (line 193) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_sequential_cpu_offload (line 196) | def enable_sequential_cpu_offload(self, gpu_id=0): method progress_bar (line 206) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 211) | def set_progress_bar_config(self, **kwargs): method __call__ (line 217) | def __call__( class KandinskyImg2ImgCombinedPipeline (line 331) | class KandinskyImg2ImgCombinedPipeline(DiffusionPipeline): method __init__ (line 366) | def __init__( method enable_xformers_memory_efficient_attention (line 411) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_sequential_cpu_offload (line 414) | def enable_sequential_cpu_offload(self, gpu_id=0): method progress_bar (line 425) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 430) | def set_progress_bar_config(self, **kwargs): method __call__ (line 436) | def __call__( class KandinskyInpaintCombinedPipeline (line 572) | class KandinskyInpaintCombinedPipeline(DiffusionPipeline): method __init__ (line 607) | def __init__( method enable_xformers_memory_efficient_attention (line 652) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_sequential_cpu_offload (line 655) | def enable_sequential_cpu_offload(self, gpu_id=0): method progress_bar (line 666) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 671) | def set_progress_bar_config(self, **kwargs): method __call__ (line 677) | def __call__( FILE: diffusers/src/diffusers/pipelines/kandinsky/pipeline_kandinsky_img2img.py function get_new_h_w (line 79) | def get_new_h_w(h, w, scale_factor=8): function prepare_image (line 89) | def prepare_image(pil_image, w=512, h=512): class KandinskyImg2ImgPipeline (line 98) | class KandinskyImg2ImgPipeline(DiffusionPipeline): method __init__ (line 120) | def __init__( method get_timesteps (line 139) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 148) | def prepare_latents(self, latents, latent_timestep, shape, dtype, devi... method _encode_prompt (line 164) | def _encode_prompt( method add_noise (line 267) | def add_noise( method __call__ (line 295) | def __call__( FILE: diffusers/src/diffusers/pipelines/kandinsky/pipeline_kandinsky_inpaint.py function get_new_h_w (line 88) | def get_new_h_w(h, w, scale_factor=8): function prepare_mask (line 98) | def prepare_mask(masks): function prepare_mask_and_masked_image (line 122) | def prepare_mask_and_masked_image(image, mask, height, width): class KandinskyInpaintPipeline (line 240) | class KandinskyInpaintPipeline(DiffusionPipeline): method __init__ (line 262) | def __init__( method prepare_latents (line 283) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method _encode_prompt (line 294) | def _encode_prompt( method __call__ (line 398) | def __call__( FILE: diffusers/src/diffusers/pipelines/kandinsky/pipeline_kandinsky_prior.py class KandinskyPriorPipelineOutput (line 113) | class KandinskyPriorPipelineOutput(BaseOutput): class KandinskyPriorPipeline (line 128) | class KandinskyPriorPipeline(DiffusionPipeline): method __init__ (line 152) | def __init__( method interpolate (line 174) | def interpolate( method prepare_latents (line 280) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method get_zero_embed (line 291) | def get_zero_embed(self, batch_size=1, device=None): method _encode_prompt (line 300) | def _encode_prompt( method __call__ (line 399) | def __call__( FILE: diffusers/src/diffusers/pipelines/kandinsky/text_encoder.py class MCLIPConfig (line 5) | class MCLIPConfig(XLMRobertaConfig): method __init__ (line 8) | def __init__(self, transformerDimSize=1024, imageDimSize=768, **kwargs): class MultilingualCLIP (line 14) | class MultilingualCLIP(PreTrainedModel): method __init__ (line 17) | def __init__(self, config, *args, **kwargs): method forward (line 24) | def forward(self, input_ids, attention_mask): FILE: diffusers/src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2.py function downscale_height_and_width (line 54) | def downscale_height_and_width(height, width, scale_factor=8): class KandinskyV22Pipeline (line 64) | class KandinskyV22Pipeline(DiffusionPipeline): method __init__ (line 83) | def __init__( method prepare_latents (line 99) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method guidance_scale (line 111) | def guidance_scale(self): method do_classifier_free_guidance (line 115) | def do_classifier_free_guidance(self): method num_timesteps (line 119) | def num_timesteps(self): method __call__ (line 124) | def __call__( FILE: diffusers/src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_combined.py class KandinskyV22CombinedPipeline (line 107) | class KandinskyV22CombinedPipeline(DiffusionPipeline): method __init__ (line 140) | def __init__( method enable_xformers_memory_efficient_attention (line 179) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_sequential_cpu_offload (line 182) | def enable_sequential_cpu_offload(self, gpu_id: Optional[int] = None, ... method progress_bar (line 193) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 198) | def set_progress_bar_config(self, **kwargs): method __call__ (line 204) | def __call__( class KandinskyV22Img2ImgCombinedPipeline (line 335) | class KandinskyV22Img2ImgCombinedPipeline(DiffusionPipeline): method __init__ (line 368) | def __init__( method enable_xformers_memory_efficient_attention (line 407) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_model_cpu_offload (line 410) | def enable_model_cpu_offload(self, gpu_id: Optional[int] = None, devic... method enable_sequential_cpu_offload (line 420) | def enable_sequential_cpu_offload(self, gpu_id: Optional[int] = None, ... method progress_bar (line 431) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 436) | def set_progress_bar_config(self, **kwargs): method __call__ (line 442) | def __call__( class KandinskyV22InpaintCombinedPipeline (line 584) | class KandinskyV22InpaintCombinedPipeline(DiffusionPipeline): method __init__ (line 617) | def __init__( method enable_xformers_memory_efficient_attention (line 656) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_sequential_cpu_offload (line 659) | def enable_sequential_cpu_offload(self, gpu_id: Optional[int] = None, ... method progress_bar (line 670) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 675) | def set_progress_bar_config(self, **kwargs): method __call__ (line 681) | def __call__( FILE: diffusers/src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_controlnet.py function downscale_height_and_width (line 96) | def downscale_height_and_width(height, width, scale_factor=8): class KandinskyV22ControlnetPipeline (line 106) | class KandinskyV22ControlnetPipeline(DiffusionPipeline): method __init__ (line 124) | def __init__( method prepare_latents (line 140) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method __call__ (line 152) | def __call__( FILE: diffusers/src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_controlnet_img2img.py function downscale_height_and_width (line 100) | def downscale_height_and_width(height, width, scale_factor=8): function prepare_image (line 111) | def prepare_image(pil_image, w=512, h=512): class KandinskyV22ControlnetImg2ImgPipeline (line 120) | class KandinskyV22ControlnetImg2ImgPipeline(DiffusionPipeline): method __init__ (line 138) | def __init__( method get_timesteps (line 154) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 164) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method __call__ (line 207) | def __call__( FILE: diffusers/src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_img2img.py function downscale_height_and_width (line 72) | def downscale_height_and_width(height, width, scale_factor=8): function prepare_image (line 83) | def prepare_image(pil_image, w=512, h=512): class KandinskyV22Img2ImgPipeline (line 92) | class KandinskyV22Img2ImgPipeline(DiffusionPipeline): method __init__ (line 111) | def __init__( method get_timesteps (line 127) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 136) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method guidance_scale (line 179) | def guidance_scale(self): method do_classifier_free_guidance (line 183) | def do_classifier_free_guidance(self): method num_timesteps (line 187) | def num_timesteps(self): method __call__ (line 191) | def __call__( FILE: diffusers/src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_inpainting.py function downscale_height_and_width (line 81) | def downscale_height_and_width(height, width, scale_factor=8): function prepare_mask (line 92) | def prepare_mask(masks): function prepare_mask_and_masked_image (line 117) | def prepare_mask_and_masked_image(image, mask, height, width): class KandinskyV22InpaintPipeline (line 235) | class KandinskyV22InpaintPipeline(DiffusionPipeline): method __init__ (line 254) | def __init__( method prepare_latents (line 271) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method guidance_scale (line 283) | def guidance_scale(self): method do_classifier_free_guidance (line 287) | def do_classifier_free_guidance(self): method num_timesteps (line 291) | def num_timesteps(self): method __call__ (line 295) | def __call__( FILE: diffusers/src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_prior.py class KandinskyV22PriorPipeline (line 84) | class KandinskyV22PriorPipeline(DiffusionPipeline): method __init__ (line 111) | def __init__( method interpolate (line 133) | def interpolate( method prepare_latents (line 239) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method get_zero_embed (line 251) | def get_zero_embed(self, batch_size=1, device=None): method _encode_prompt (line 261) | def _encode_prompt( method do_classifier_free_guidance (line 359) | def do_classifier_free_guidance(self): method guidance_scale (line 363) | def guidance_scale(self): method num_timesteps (line 367) | def num_timesteps(self): method __call__ (line 372) | def __call__( FILE: diffusers/src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_prior_emb2emb.py class KandinskyV22PriorEmb2EmbPipeline (line 102) | class KandinskyV22PriorEmb2EmbPipeline(DiffusionPipeline): method __init__ (line 126) | def __init__( method get_timesteps (line 146) | def get_timesteps(self, num_inference_steps, strength, device): method interpolate (line 157) | def interpolate( method _encode_image (line 245) | def _encode_image( method prepare_latents (line 262) | def prepare_latents(self, emb, timestep, batch_size, num_images_per_pr... method get_zero_embed (line 289) | def get_zero_embed(self, batch_size=1, device=None): method _encode_prompt (line 299) | def _encode_prompt( method __call__ (line 398) | def __call__( FILE: diffusers/src/diffusers/pipelines/kandinsky3/convert_kandinsky3_unet.py function convert_state_dict (line 36) | def convert_state_dict(unet_state_dict): function main (line 75) | def main(model_path, output_path): FILE: diffusers/src/diffusers/pipelines/kandinsky3/pipeline_kandinsky3.py function downscale_height_and_width (line 40) | def downscale_height_and_width(height, width, scale_factor=8): class Kandinsky3Pipeline (line 50) | class Kandinsky3Pipeline(DiffusionPipeline, LoraLoaderMixin): method __init__ (line 60) | def __init__( method process_embeds (line 74) | def process_embeds(self, embeddings, attention_mask, cut_context): method encode_prompt (line 83) | def encode_prompt( method prepare_latents (line 232) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method check_inputs (line 243) | def check_inputs( method guidance_scale (line 314) | def guidance_scale(self): method do_classifier_free_guidance (line 318) | def do_classifier_free_guidance(self): method num_timesteps (line 322) | def num_timesteps(self): method __call__ (line 327) | def __call__( FILE: diffusers/src/diffusers/pipelines/kandinsky3/pipeline_kandinsky3_img2img.py function downscale_height_and_width (line 47) | def downscale_height_and_width(height, width, scale_factor=8): function prepare_image (line 57) | def prepare_image(pil_image): class Kandinsky3Img2ImgPipeline (line 65) | class Kandinsky3Img2ImgPipeline(DiffusionPipeline, LoraLoaderMixin): method __init__ (line 75) | def __init__( method get_timesteps (line 89) | def get_timesteps(self, num_inference_steps, strength, device): method _process_embeds (line 98) | def _process_embeds(self, embeddings, attention_mask, cut_context): method encode_prompt (line 108) | def encode_prompt( method prepare_latents (line 257) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method prepare_extra_step_kwargs (line 300) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 317) | def check_inputs( method guidance_scale (line 390) | def guidance_scale(self): method do_classifier_free_guidance (line 394) | def do_classifier_free_guidance(self): method num_timesteps (line 398) | def num_timesteps(self): method __call__ (line 403) | def __call__( FILE: diffusers/src/diffusers/pipelines/latent_consistency_models/pipeline_latent_consistency_img2img.py function retrieve_latents (line 47) | def retrieve_latents( function retrieve_timesteps (line 61) | def retrieve_timesteps( class LatentConsistencyModelImg2ImgPipeline (line 146) | class LatentConsistencyModelImg2ImgPipeline( method __init__ (line 194) | def __init__( method encode_prompt (line 233) | def encode_prompt( method encode_image (line 416) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 441) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 493) | def run_safety_checker(self, image, device, dtype): method prepare_latents (line 508) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method get_guidance_scale_embedding (line 567) | def get_guidance_scale_embedding( method prepare_extra_step_kwargs (line 598) | def prepare_extra_step_kwargs(self, generator, eta): method get_timesteps (line 616) | def get_timesteps(self, num_inference_steps, strength, device): method check_inputs (line 627) | def check_inputs( method guidance_scale (line 681) | def guidance_scale(self): method cross_attention_kwargs (line 685) | def cross_attention_kwargs(self): method clip_skip (line 689) | def clip_skip(self): method do_classifier_free_guidance (line 693) | def do_classifier_free_guidance(self): method num_timesteps (line 697) | def num_timesteps(self): method __call__ (line 702) | def __call__( FILE: diffusers/src/diffusers/pipelines/latent_consistency_models/pipeline_latent_consistency_text2img.py function retrieve_timesteps (line 65) | def retrieve_timesteps( class LatentConsistencyModelPipeline (line 124) | class LatentConsistencyModelPipeline( method __init__ (line 172) | def __init__( method encode_prompt (line 217) | def encode_prompt( method encode_image (line 400) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 425) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 477) | def run_safety_checker(self, image, device, dtype): method prepare_latents (line 492) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 514) | def get_guidance_scale_embedding( method prepare_extra_step_kwargs (line 545) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 563) | def check_inputs( method guidance_scale (line 618) | def guidance_scale(self): method cross_attention_kwargs (line 622) | def cross_attention_kwargs(self): method clip_skip (line 626) | def clip_skip(self): method do_classifier_free_guidance (line 630) | def do_classifier_free_guidance(self): method num_timesteps (line 634) | def num_timesteps(self): method __call__ (line 639) | def __call__( FILE: diffusers/src/diffusers/pipelines/latent_diffusion/pipeline_latent_diffusion.py class LDMTextToImagePipeline (line 32) | class LDMTextToImagePipeline(DiffusionPipeline): method __init__ (line 55) | def __init__( method __call__ (line 68) | def __call__( class LDMBertConfig (line 242) | class LDMBertConfig(PretrainedConfig): method __init__ (line 247) | def __init__( function _expand_mask (line 289) | def _expand_mask(mask: torch.Tensor, dtype: torch.dtype, tgt_len: Option... class LDMBertAttention (line 304) | class LDMBertAttention(nn.Module): method __init__ (line 307) | def __init__( method _shape (line 331) | def _shape(self, tensor: torch.Tensor, seq_len: int, bsz: int): method forward (line 334) | def forward( class LDMBertEncoderLayer (line 448) | class LDMBertEncoderLayer(nn.Module): method __init__ (line 449) | def __init__(self, config: LDMBertConfig): method forward (line 466) | def forward( class LDMBertPreTrainedModel (line 518) | class LDMBertPreTrainedModel(PreTrainedModel): method _init_weights (line 524) | def _init_weights(self, module): method _set_gradient_checkpointing (line 535) | def _set_gradient_checkpointing(self, module, value=False): method dummy_inputs (line 540) | def dummy_inputs(self): class LDMBertEncoder (line 550) | class LDMBertEncoder(LDMBertPreTrainedModel): method __init__ (line 560) | def __init__(self, config: LDMBertConfig): method get_input_embeddings (line 578) | def get_input_embeddings(self): method set_input_embeddings (line 581) | def set_input_embeddings(self, value): method forward (line 584) | def forward( class LDMBertModel (line 717) | class LDMBertModel(LDMBertPreTrainedModel): method __init__ (line 720) | def __init__(self, config: LDMBertConfig): method forward (line 725) | def forward( FILE: diffusers/src/diffusers/pipelines/latent_diffusion/pipeline_latent_diffusion_superresolution.py function preprocess (line 23) | def preprocess(image): class LDMSuperResolutionPipeline (line 33) | class LDMSuperResolutionPipeline(DiffusionPipeline): method __init__ (line 51) | def __init__( method __call__ (line 68) | def __call__( FILE: diffusers/src/diffusers/pipelines/ledits_pp/pipeline_leditspp_stable_diffusion.py class LeditsAttentionStore (line 63) | class LeditsAttentionStore: method get_empty_store (line 65) | def get_empty_store(): method __call__ (line 68) | def __call__(self, attn, is_cross: bool, place_in_unet: str, editing_p... method forward (line 77) | def forward(self, attn, is_cross: bool, place_in_unet: str): method between_steps (line 82) | def between_steps(self, store_step=True): method get_attention (line 100) | def get_attention(self, step: int): method aggregate_attention (line 110) | def aggregate_attention( method __init__ (line 133) | def __init__(self, average: bool, batch_size=1, max_resolution=16, max... class LeditsGaussianSmoothing (line 148) | class LeditsGaussianSmoothing: method __init__ (line 149) | def __init__(self, device): method __call__ (line 169) | def __call__(self, input): class LEDITSCrossAttnProcessor (line 180) | class LEDITSCrossAttnProcessor: method __init__ (line 181) | def __init__(self, attention_store, place_in_unet, pnp, editing_prompts): method __call__ (line 187) | def __call__( function rescale_noise_cfg (line 236) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class LEditsPPPipelineStableDiffusion (line 250) | class LEditsPPPipelineStableDiffusion( method __init__ (line 287) | def __init__( method run_safety_checker (line 389) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 404) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 416) | def prepare_extra_step_kwargs(self, eta, generator=None): method check_inputs (line 434) | def check_inputs( method prepare_latents (line 462) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method prepare_unet (line 474) | def prepare_unet(self, attention_store, PnP: bool = False): method encode_prompt (line 498) | def encode_prompt( method guidance_rescale (line 686) | def guidance_rescale(self): method clip_skip (line 690) | def clip_skip(self): method cross_attention_kwargs (line 694) | def cross_attention_kwargs(self): method __call__ (line 699) | def __call__( method invert (line 1199) | def invert( method encode_image (line 1347) | def encode_image(self, image, dtype=None, height=None, width=None, res... function compute_noise_ddim (line 1367) | def compute_noise_ddim(scheduler, prev_latents, latents, timestep, noise... function compute_noise_sde_dpm_pp_2nd (line 1405) | def compute_noise_sde_dpm_pp_2nd(scheduler, prev_latents, latents, times... function compute_noise (line 1489) | def compute_noise(scheduler, *args): FILE: diffusers/src/diffusers/pipelines/ledits_pp/pipeline_leditspp_stable_diffusion_xl.py class LeditsAttentionStore (line 110) | class LeditsAttentionStore: method get_empty_store (line 112) | def get_empty_store(): method __call__ (line 115) | def __call__(self, attn, is_cross: bool, place_in_unet: str, editing_p... method forward (line 124) | def forward(self, attn, is_cross: bool, place_in_unet: str): method between_steps (line 129) | def between_steps(self, store_step=True): method get_attention (line 147) | def get_attention(self, step: int): method aggregate_attention (line 157) | def aggregate_attention( method __init__ (line 180) | def __init__(self, average: bool, batch_size=1, max_resolution=16, max... class LeditsGaussianSmoothing (line 195) | class LeditsGaussianSmoothing: method __init__ (line 196) | def __init__(self, device): method __call__ (line 216) | def __call__(self, input): class LEDITSCrossAttnProcessor (line 228) | class LEDITSCrossAttnProcessor: method __init__ (line 229) | def __init__(self, attention_store, place_in_unet, pnp, editing_prompts): method __call__ (line 235) | def __call__( class LEditsPPPipelineStableDiffusionXL (line 283) | class LEditsPPPipelineStableDiffusionXL( method __init__ (line 356) | def __init__( method encode_prompt (line 406) | def encode_prompt( method prepare_extra_step_kwargs (line 627) | def prepare_extra_step_kwargs(self, eta, generator=None): method check_inputs (line 644) | def check_inputs( method prepare_latents (line 668) | def prepare_latents(self, device, latents): method _get_add_time_ids (line 675) | def _get_add_time_ids( method upcast_vae (line 694) | def upcast_vae(self): method get_guidance_scale_embedding (line 714) | def get_guidance_scale_embedding( method guidance_scale (line 745) | def guidance_scale(self): method guidance_rescale (line 749) | def guidance_rescale(self): method clip_skip (line 753) | def clip_skip(self): method do_classifier_free_guidance (line 760) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 764) | def cross_attention_kwargs(self): method denoising_end (line 768) | def denoising_end(self): method num_timesteps (line 772) | def num_timesteps(self): method prepare_unet (line 776) | def prepare_unet(self, attention_store, PnP: bool = False): method __call__ (line 802) | def __call__( method encode_image (line 1404) | def encode_image(self, image, dtype=None, height=None, width=None, res... method invert (line 1433) | def invert( function rescale_noise_cfg (line 1649) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function compute_noise_ddim (line 1664) | def compute_noise_ddim(scheduler, prev_latents, latents, timestep, noise... function compute_noise_sde_dpm_pp_2nd (line 1703) | def compute_noise_sde_dpm_pp_2nd(scheduler, prev_latents, latents, times... function compute_noise (line 1788) | def compute_noise(scheduler, *args): FILE: diffusers/src/diffusers/pipelines/ledits_pp/pipeline_output.py class LEditsPPDiffusionPipelineOutput (line 11) | class LEditsPPDiffusionPipelineOutput(BaseOutput): class LEditsPPInversionPipelineOutput (line 29) | class LEditsPPInversionPipelineOutput(BaseOutput): FILE: diffusers/src/diffusers/pipelines/marigold/marigold_image_processing.py class MarigoldImageProcessor (line 19) | class MarigoldImageProcessor(ConfigMixin): method __init__ (line 23) | def __init__( method expand_tensor_or_array (line 32) | def expand_tensor_or_array(images: Union[torch.Tensor, np.ndarray]) ->... method pt_to_numpy (line 51) | def pt_to_numpy(images: torch.Tensor) -> np.ndarray: method numpy_to_pt (line 59) | def numpy_to_pt(images: np.ndarray) -> torch.Tensor: method resize_antialias (line 74) | def resize_antialias( method resize_to_max_edge (line 90) | def resize_to_max_edge(image: torch.Tensor, max_edge_sz: int, mode: st... method pad_image (line 111) | def pad_image(image: torch.Tensor, align: int) -> Tuple[torch.Tensor, ... method unpad_image (line 127) | def unpad_image(image: torch.Tensor, padding: Tuple[int, int]) -> torc... method load_image_canonical (line 144) | def load_image_canonical( method check_image_values_range (line 190) | def check_image_values_range(image: torch.Tensor) -> None: method preprocess (line 198) | def preprocess( method colormap (line 240) | def colormap( method visualize_depth (line 370) | def visualize_depth( method export_depth_to_16bit_png (line 435) | def export_depth_to_16bit_png( method visualize_normals (line 471) | def visualize_normals( method visualize_uncertainty (line 532) | def visualize_uncertainty( FILE: diffusers/src/diffusers/pipelines/marigold/pipeline_marigold_depth.py class MarigoldDepthOutput (line 75) | class MarigoldDepthOutput(BaseOutput): class MarigoldDepthPipeline (line 96) | class MarigoldDepthPipeline(DiffusionPipeline): method __init__ (line 141) | def __init__( method check_inputs (line 188) | def check_inputs( method progress_bar (line 315) | def progress_bar(self, iterable=None, total=None, desc=None, leave=True): method __call__ (line 335) | def __call__( method prepare_latents (line 608) | def prepare_latents( method decode_prediction (line 645) | def decode_prediction(self, pred_latent: torch.Tensor) -> torch.Tensor: method ensemble_depth (line 660) | def ensemble_depth( FILE: diffusers/src/diffusers/pipelines/marigold/pipeline_marigold_normals.py class MarigoldNormalsOutput (line 70) | class MarigoldNormalsOutput(BaseOutput): class MarigoldNormalsPipeline (line 91) | class MarigoldNormalsPipeline(DiffusionPipeline): method __init__ (line 131) | def __init__( method check_inputs (line 174) | def check_inputs( method progress_bar (line 299) | def progress_bar(self, iterable=None, total=None, desc=None, leave=True): method __call__ (line 319) | def __call__( method prepare_latents (line 581) | def prepare_latents( method decode_prediction (line 618) | def decode_prediction(self, pred_latent: torch.Tensor) -> torch.Tensor: method normalize_normals (line 637) | def normalize_normals(normals: torch.Tensor, eps: float = 1e-6) -> tor... method ensemble_normals (line 647) | def ensemble_normals( FILE: diffusers/src/diffusers/pipelines/musicldm/pipeline_musicldm.py class MusicLDMPipeline (line 67) | class MusicLDMPipeline(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 93) | def __init__( method _encode_prompt (line 116) | def _encode_prompt( method mel_spectrogram_to_waveform (line 250) | def mel_spectrogram_to_waveform(self, mel_spectrogram): method score_waveforms (line 260) | def score_waveforms(self, text, audio, num_waveforms_per_prompt, devic... method prepare_extra_step_kwargs (line 285) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 303) | def check_inputs( method prepare_latents (line 362) | def prepare_latents(self, batch_size, num_channels_latents, height, dt... method enable_model_cpu_offload (line 384) | def enable_model_cpu_offload(self, gpu_id=0): method __call__ (line 420) | def __call__( FILE: diffusers/src/diffusers/pipelines/onnx_utils.py class OnnxRuntimeModel (line 52) | class OnnxRuntimeModel: method __init__ (line 53) | def __init__(self, model=None, **kwargs): method __call__ (line 59) | def __call__(self, **kwargs): method load_model (line 64) | def load_model(path: Union[str, Path], provider=None, sess_options=None): method _save_pretrained (line 80) | def _save_pretrained(self, save_directory: Union[str, Path], file_name... method save_pretrained (line 111) | def save_pretrained( method _from_pretrained (line 135) | def _from_pretrained( method from_pretrained (line 196) | def from_pretrained( FILE: diffusers/src/diffusers/pipelines/paint_by_example/image_encoder.py class PaintByExampleImageEncoder (line 25) | class PaintByExampleImageEncoder(CLIPPreTrainedModel): method __init__ (line 26) | def __init__(self, config, proj_size=None): method forward (line 38) | def forward(self, pixel_values, return_uncond_vector=False): class PaintByExampleMapper (line 50) | class PaintByExampleMapper(nn.Module): method __init__ (line 51) | def __init__(self, config): method forward (line 63) | def forward(self, hidden_states): FILE: diffusers/src/diffusers/pipelines/paint_by_example/pipeline_paint_by_example.py function retrieve_latents (line 38) | def retrieve_latents( function prepare_mask_and_masked_image (line 51) | def prepare_mask_and_masked_image(image, mask): class PaintByExamplePipeline (line 151) | class PaintByExamplePipeline(DiffusionPipeline, StableDiffusionMixin): method __init__ (line 192) | def __init__( method run_safety_checker (line 217) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 232) | def prepare_extra_step_kwargs(self, generator, eta): method decode_latents (line 250) | def decode_latents(self, latents): method check_inputs (line 262) | def check_inputs(self, image, height, width, callback_steps): method prepare_latents (line 285) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method prepare_mask_latents (line 308) | def prepare_mask_latents( method _encode_vae_image (line 354) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method _encode_image (line 368) | def _encode_image(self, image, device, num_images_per_prompt, do_class... method __call__ (line 394) | def __call__( FILE: diffusers/src/diffusers/pipelines/pia/pipeline_pia.py function prepare_mask_coef_by_statistics (line 93) | def prepare_mask_coef_by_statistics(num_frames: int, cond_frame: int, mo... class PIAPipelineOutput (line 112) | class PIAPipelineOutput(BaseOutput): class PIAPipeline (line 126) | class PIAPipeline( method __init__ (line 167) | def __init__( method encode_prompt (line 203) | def encode_prompt( method encode_image (line 386) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method decode_latents (line 411) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 424) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 441) | def check_inputs( method prepare_ip_adapter_image_embeds (line 505) | def prepare_ip_adapter_image_embeds( method prepare_latents (line 557) | def prepare_latents( method prepare_masked_condition (line 582) | def prepare_masked_condition( method get_timesteps (line 634) | def get_timesteps(self, num_inference_steps, strength, device): method guidance_scale (line 646) | def guidance_scale(self): method clip_skip (line 650) | def clip_skip(self): method do_classifier_free_guidance (line 657) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 661) | def cross_attention_kwargs(self): method num_timesteps (line 665) | def num_timesteps(self): method __call__ (line 670) | def __call__( FILE: diffusers/src/diffusers/pipelines/pipeline_flax_utils.py function import_flax_or_no_model (line 74) | def import_flax_or_no_model(module, class_name): class FlaxImagePipelineOutput (line 88) | class FlaxImagePipelineOutput(BaseOutput): class FlaxDiffusionPipeline (line 101) | class FlaxDiffusionPipeline(ConfigMixin, PushToHubMixin): method register_modules (line 118) | def register_modules(self, **kwargs): method save_pretrained (line 151) | def save_pretrained( method from_pretrained (line 231) | def from_pretrained(cls, pretrained_model_name_or_path: Optional[Union... method _get_signature_keys (line 543) | def _get_signature_keys(cls, obj): method components (line 552) | def components(self) -> Dict[str, Any]: method numpy_to_pil (line 589) | def numpy_to_pil(images): method progress_bar (line 605) | def progress_bar(self, iterable): method set_progress_bar_config (line 615) | def set_progress_bar_config(self, **kwargs): FILE: diffusers/src/diffusers/pipelines/pipeline_loading_utils.py function is_safetensors_compatible (line 92) | def is_safetensors_compatible(filenames, variant=None, passed_components... function variant_compatible_siblings (line 140) | def variant_compatible_siblings(filenames, variant=None) -> Union[List[o... function warn_deprecated_model_variant (line 208) | def warn_deprecated_model_variant(pretrained_model_name_or_path, token, ... function _unwrap_model (line 230) | def _unwrap_model(model): function maybe_raise_or_warn (line 244) | def maybe_raise_or_warn( function get_class_obj_and_candidates (line 275) | def get_class_obj_and_candidates( function _get_custom_pipeline_class (line 302) | def _get_custom_pipeline_class( function _get_pipeline_class (line 335) | def _get_pipeline_class( function _load_empty_model (line 386) | def _load_empty_model( function _assign_components_to_devices (line 472) | def _assign_components_to_devices( function _get_final_device_map (line 502) | def _get_final_device_map(device_map, pipeline_class, passed_class_obj, ... function load_sub_model (line 589) | def load_sub_model( function _fetch_class_library_tuple (line 727) | def _fetch_class_library_tuple(module): FILE: diffusers/src/diffusers/pipelines/pipeline_utils.py class ImagePipelineOutput (line 101) | class ImagePipelineOutput(BaseOutput): class AudioPipelineOutput (line 115) | class AudioPipelineOutput(BaseOutput): class DiffusionPipeline (line 127) | class DiffusionPipeline(ConfigMixin, PushToHubMixin): method register_modules (line 153) | def register_modules(self, **kwargs): method __setattr__ (line 168) | def __setattr__(self, name: str, value: Any): method save_pretrained (line 183) | def save_pretrained( method to (line 303) | def to(self, *args, **kwargs): method device (line 449) | def device(self) -> torch.device: method dtype (line 464) | def dtype(self) -> torch.dtype: method from_pretrained (line 480) | def from_pretrained(cls, pretrained_model_name_or_path: Optional[Union... method name_or_path (line 981) | def name_or_path(self) -> str: method _execution_device (line 985) | def _execution_device(self): method remove_all_hooks (line 1006) | def remove_all_hooks(self): method enable_model_cpu_offload (line 1015) | def enable_model_cpu_offload(self, gpu_id: Optional[int] = None, devic... method maybe_free_model_hooks (line 1093) | def maybe_free_model_hooks(self): method enable_sequential_cpu_offload (line 1107) | def enable_sequential_cpu_offload(self, gpu_id: Optional[int] = None, ... method reset_device_map (line 1168) | def reset_device_map(self): method download (line 1183) | def download(cls, pretrained_model_name, **kwargs) -> Union[str, os.Pa... method _get_signature_keys (line 1555) | def _get_signature_keys(cls, obj): method _get_signature_types (line 1570) | def _get_signature_types(cls): method components (line 1582) | def components(self) -> Dict[str, Any]: method numpy_to_pil (line 1618) | def numpy_to_pil(images): method progress_bar (line 1624) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 1639) | def set_progress_bar_config(self, **kwargs): method enable_xformers_memory_efficient_attention (line 1642) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method disable_xformers_memory_efficient_attention (line 1677) | def disable_xformers_memory_efficient_attention(self): method set_use_memory_efficient_attention_xformers (line 1683) | def set_use_memory_efficient_attention_xformers( method enable_attention_slicing (line 1703) | def enable_attention_slicing(self, slice_size: Optional[Union[str, int... method disable_attention_slicing (line 1743) | def disable_attention_slicing(self): method set_attention_slice (line 1751) | def set_attention_slice(self, slice_size: Optional[int]): method from_pipe (line 1760) | def from_pipe(cls, pipeline, **kwargs): class StableDiffusionMixin (line 1883) | class StableDiffusionMixin: method enable_vae_slicing (line 1888) | def enable_vae_slicing(self): method disable_vae_slicing (line 1895) | def disable_vae_slicing(self): method enable_vae_tiling (line 1902) | def enable_vae_tiling(self): method disable_vae_tiling (line 1910) | def disable_vae_tiling(self): method enable_freeu (line 1917) | def enable_freeu(self, s1: float, s2: float, b1: float, b2: float): method disable_freeu (line 1939) | def disable_freeu(self): method fuse_qkv_projections (line 1943) | def fuse_qkv_projections(self, unet: bool = True, vae: bool = True): method unfuse_qkv_projections (line 1974) | def unfuse_qkv_projections(self, unet: bool = True, vae: bool = True): FILE: diffusers/src/diffusers/pipelines/pixart_alpha/pipeline_pixart_alpha.py function retrieve_timesteps (line 173) | def retrieve_timesteps( class PixArtAlphaPipeline (line 232) | class PixArtAlphaPipeline(DiffusionPipeline): method __init__ (line 274) | def __init__( method encode_prompt (line 292) | def encode_prompt( method prepare_extra_step_kwargs (line 435) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 452) | def check_inputs( method _text_preprocessing (line 520) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 545) | def _clean_caption(self, caption): method prepare_latents (line 660) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 684) | def __call__( FILE: diffusers/src/diffusers/pipelines/pixart_alpha/pipeline_pixart_sigma.py function retrieve_timesteps (line 117) | def retrieve_timesteps( class PixArtSigmaPipeline (line 176) | class PixArtSigmaPipeline(DiffusionPipeline): method __init__ (line 200) | def __init__( method encode_prompt (line 218) | def encode_prompt( method prepare_extra_step_kwargs (line 361) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 379) | def check_inputs( method _text_preprocessing (line 447) | def _text_preprocessing(self, text, clean_caption=False): method _clean_caption (line 472) | def _clean_caption(self, caption): method prepare_latents (line 587) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 611) | def __call__( FILE: diffusers/src/diffusers/pipelines/semantic_stable_diffusion/pipeline_output.py class SemanticStableDiffusionPipelineOutput (line 11) | class SemanticStableDiffusionPipelineOutput(BaseOutput): FILE: diffusers/src/diffusers/pipelines/semantic_stable_diffusion/pipeline_semantic_stable_diffusion.py class SemanticStableDiffusionPipeline (line 21) | class SemanticStableDiffusionPipeline(DiffusionPipeline, StableDiffusion... method __init__ (line 52) | def __init__( method run_safety_checker (line 95) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 110) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 122) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 140) | def check_inputs( method prepare_latents (line 193) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 216) | def __call__( FILE: diffusers/src/diffusers/pipelines/shap_e/camera.py class DifferentiableProjectiveCamera (line 23) | class DifferentiableProjectiveCamera: method __post_init__ (line 38) | def __post_init__(self): method resolution (line 43) | def resolution(self): method fov (line 46) | def fov(self): method get_image_coords (line 49) | def get_image_coords(self) -> torch.Tensor: method camera_rays (line 64) | def camera_rays(self): method get_camera_rays (line 76) | def get_camera_rays(self, coords: torch.Tensor) -> torch.Tensor: method resize_image (line 105) | def resize_image(self, width: int, height: int) -> "DifferentiableProj... function create_pan_cameras (line 122) | def create_pan_cameras(size: int) -> DifferentiableProjectiveCamera: FILE: diffusers/src/diffusers/pipelines/shap_e/pipeline_shap_e.py class ShapEPipelineOutput (line 67) | class ShapEPipelineOutput(BaseOutput): class ShapEPipeline (line 79) | class ShapEPipeline(DiffusionPipeline): method __init__ (line 103) | def __init__( method prepare_latents (line 122) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method _encode_prompt (line 133) | def _encode_prompt( method __call__ (line 184) | def __call__( FILE: diffusers/src/diffusers/pipelines/shap_e/pipeline_shap_e_img2img.py class ShapEPipelineOutput (line 68) | class ShapEPipelineOutput(BaseOutput): class ShapEImg2ImgPipeline (line 80) | class ShapEImg2ImgPipeline(DiffusionPipeline): method __init__ (line 104) | def __init__( method prepare_latents (line 123) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method _encode_image (line 134) | def _encode_image( method __call__ (line 166) | def __call__( FILE: diffusers/src/diffusers/pipelines/shap_e/renderer.py function sample_pmf (line 30) | def sample_pmf(pmf: torch.Tensor, n_samples: int) -> torch.Tensor: function posenc_nerf (line 53) | def posenc_nerf(x: torch.Tensor, min_deg: int = 0, max_deg: int = 15) ->... function encode_position (line 70) | def encode_position(position): function encode_direction (line 74) | def encode_direction(position, direction=None): function _sanitize_name (line 81) | def _sanitize_name(x: str) -> str: function integrate_samples (line 85) | def integrate_samples(volume_range, ts, density, channels): function volume_query_points (line 119) | def volume_query_points(volume, grid_size): function _convert_srgb_to_linear (line 128) | def _convert_srgb_to_linear(u: torch.Tensor): function _create_flat_edge_indices (line 132) | def _create_flat_edge_indices( class VoidNeRFModel (line 214) | class VoidNeRFModel(nn.Module): method __init__ (line 219) | def __init__(self, background, channel_scale=255.0): method forward (line 225) | def forward(self, position): class VolumeRange (line 237) | class VolumeRange: method __post_init__ (line 242) | def __post_init__(self): method partition (line 245) | def partition(self, ts): class BoundingBoxVolume (line 269) | class BoundingBoxVolume(nn.Module): method __init__ (line 274) | def __init__( method intersect (line 300) | def intersect( class StratifiedRaySampler (line 353) | class StratifiedRaySampler(nn.Module): method __init__ (line 358) | def __init__(self, depth_mode: str = "linear"): method sample (line 366) | def sample( class ImportanceRaySampler (line 405) | class ImportanceRaySampler(nn.Module): method __init__ (line 410) | def __init__( method sample (line 433) | def sample(self, t0: torch.Tensor, t1: torch.Tensor, n_samples: int) -... class MeshDecoderOutput (line 467) | class MeshDecoderOutput(BaseOutput): class MeshDecoder (line 485) | class MeshDecoder(nn.Module): method __init__ (line 490) | def __init__(self): method forward (line 498) | def forward(self, field: torch.Tensor, min_point: torch.Tensor, size: ... class MLPNeRFModelOutput (line 602) | class MLPNeRFModelOutput(BaseOutput): class MLPNeRSTFModel (line 609) | class MLPNeRSTFModel(ModelMixin, ConfigMixin): method __init__ (line 611) | def __init__( method map_indices_to_keys (line 648) | def map_indices_to_keys(self, output): method forward (line 662) | def forward(self, *, position, direction, ts, nerf_level="coarse", ren... class ChannelsProj (line 708) | class ChannelsProj(nn.Module): method __init__ (line 709) | def __init__( method forward (line 723) | def forward(self, x: torch.Tensor) -> torch.Tensor: class ShapEParamsProjModel (line 734) | class ShapEParamsProjModel(ModelMixin, ConfigMixin): method __init__ (line 742) | def __init__( method forward (line 772) | def forward(self, x: torch.Tensor): class ShapERenderer (line 784) | class ShapERenderer(ModelMixin, ConfigMixin): method __init__ (line 786) | def __init__( method render_rays (line 826) | def render_rays(self, rays, sampler, n_samples, prev_model_out=None, r... method decode_to_image (line 905) | def decode_to_image( method decode_to_mesh (line 950) | def decode_to_mesh( FILE: diffusers/src/diffusers/pipelines/stable_cascade/pipeline_stable_cascade.py class StableCascadeDecoderPipeline (line 50) | class StableCascadeDecoderPipeline(DiffusionPipeline): method __init__ (line 84) | def __init__( method prepare_latents (line 103) | def prepare_latents( method encode_prompt (line 124) | def encode_prompt( method check_inputs (line 231) | def check_inputs( method guidance_scale (line 273) | def guidance_scale(self): method do_classifier_free_guidance (line 277) | def do_classifier_free_guidance(self): method num_timesteps (line 281) | def num_timesteps(self): method __call__ (line 286) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_cascade/pipeline_stable_cascade_combined.py class StableCascadeCombinedPipeline (line 45) | class StableCascadeCombinedPipeline(DiffusionPipeline): method __init__ (line 76) | def __init__( method enable_xformers_memory_efficient_attention (line 121) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_model_cpu_offload (line 124) | def enable_model_cpu_offload(self, gpu_id: Optional[int] = None, devic... method enable_sequential_cpu_offload (line 134) | def enable_sequential_cpu_offload(self, gpu_id: Optional[int] = None, ... method progress_bar (line 144) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 148) | def set_progress_bar_config(self, **kwargs): method __call__ (line 154) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_cascade/pipeline_stable_cascade_prior.py class StableCascadePriorPipelineOutput (line 52) | class StableCascadePriorPipelineOutput(BaseOutput): class StableCascadePriorPipeline (line 72) | class StableCascadePriorPipeline(DiffusionPipeline): method __init__ (line 104) | def __init__( method prepare_latents (line 125) | def prepare_latents( method encode_prompt (line 145) | def encode_prompt( method encode_image (line 252) | def encode_image(self, images, device, dtype, batch_size, num_images_p... method check_inputs (line 266) | def check_inputs( method guidance_scale (line 344) | def guidance_scale(self): method do_classifier_free_guidance (line 348) | def do_classifier_free_guidance(self): method num_timesteps (line 352) | def num_timesteps(self): method get_timestep_ratio_conditioning (line 355) | def get_timestep_ratio_conditioning(self, t, alphas_cumprod): method __call__ (line 367) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/clip_image_project_model.py class CLIPImageProjection (line 21) | class CLIPImageProjection(ModelMixin, ConfigMixin): method __init__ (line 23) | def __init__(self, hidden_size: int = 768): method forward (line 28) | def forward(self, x): FILE: diffusers/src/diffusers/pipelines/stable_diffusion/convert_from_ckpt.py function shave_segments (line 69) | def shave_segments(path, n_shave_prefix_segments=1): function renew_resnet_paths (line 79) | def renew_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_vae_resnet_paths (line 101) | def renew_vae_resnet_paths(old_list, n_shave_prefix_segments=0): function renew_attention_paths (line 117) | def renew_attention_paths(old_list, n_shave_prefix_segments=0): function renew_vae_attention_paths (line 138) | def renew_vae_attention_paths(old_list, n_shave_prefix_segments=0): function assign_to_checkpoint (line 168) | def assign_to_checkpoint( function conv_attn_to_linear (line 223) | def conv_attn_to_linear(checkpoint): function create_unet_diffusers_config (line 235) | def create_unet_diffusers_config(original_config, image_size: int, contr... function create_vae_diffusers_config (line 344) | def create_vae_diffusers_config(original_config, image_size: int): function create_diffusers_schedular (line 368) | def create_diffusers_schedular(original_config): function create_ldm_bert_config (line 378) | def create_ldm_bert_config(original_config): function convert_ldm_unet_checkpoint (line 388) | def convert_ldm_unet_checkpoint( function convert_ldm_vae_checkpoint (line 636) | def convert_ldm_vae_checkpoint(checkpoint, config): function convert_ldm_bert_checkpoint (line 743) | def convert_ldm_bert_checkpoint(checkpoint, config): function convert_ldm_clip_checkpoint (line 793) | def convert_ldm_clip_checkpoint(checkpoint, local_files_only=False, text... function convert_paint_by_example_checkpoint (line 857) | def convert_paint_by_example_checkpoint(checkpoint, local_files_only=Fal... function convert_open_clip_checkpoint (line 924) | def convert_open_clip_checkpoint( function stable_unclip_image_encoder (line 1006) | def stable_unclip_image_encoder(original_config, local_files_only=False): function stable_unclip_image_noising_components (line 1043) | def stable_unclip_image_noising_components( function convert_controlnet_checkpoint (line 1088) | def convert_controlnet_checkpoint( function download_from_original_stable_diffusion_ckpt (line 1138) | def download_from_original_stable_diffusion_ckpt( function download_controlnet_from_original_ckpt (line 1815) | def download_controlnet_from_original_ckpt( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_flax_stable_diffusion.py class FlaxStableDiffusionPipeline (line 81) | class FlaxStableDiffusionPipeline(FlaxDiffusionPipeline): method __init__ (line 109) | def __init__( method prepare_inputs (line 167) | def prepare_inputs(self, prompt: Union[str, List[str]]): method _get_has_nsfw_concepts (line 180) | def _get_has_nsfw_concepts(self, features, params): method _run_safety_checker (line 184) | def _run_safety_checker(self, images, safety_model_params, jit=False): method _generate (line 214) | def _generate( method __call__ (line 311) | def __call__( function _p_generate (line 439) | def _p_generate( function _p_get_has_nsfw_concepts (line 465) | def _p_get_has_nsfw_concepts(pipe, features, params): function unshard (line 469) | def unshard(x: jnp.ndarray): FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_flax_stable_diffusion_img2img.py class FlaxStableDiffusionImg2ImgPipeline (line 105) | class FlaxStableDiffusionImg2ImgPipeline(FlaxDiffusionPipeline): method __init__ (line 133) | def __init__( method prepare_inputs (line 170) | def prepare_inputs(self, prompt: Union[str, List[str]], image: Union[I... method _get_has_nsfw_concepts (line 191) | def _get_has_nsfw_concepts(self, features, params): method _run_safety_checker (line 195) | def _run_safety_checker(self, images, safety_model_params, jit=False): method get_timestep_start (line 225) | def get_timestep_start(self, num_inference_steps, strength): method _generate (line 233) | def _generate( method __call__ (line 338) | def __call__( function _p_generate (line 485) | def _p_generate( function _p_get_has_nsfw_concepts (line 515) | def _p_get_has_nsfw_concepts(pipe, features, params): function unshard (line 519) | def unshard(x: jnp.ndarray): function preprocess (line 526) | def preprocess(image, dtype): FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_flax_stable_diffusion_inpaint.py class FlaxStableDiffusionInpaintPipeline (line 102) | class FlaxStableDiffusionInpaintPipeline(FlaxDiffusionPipeline): method __init__ (line 136) | def __init__( method prepare_inputs (line 194) | def prepare_inputs( method _get_has_nsfw_concepts (line 233) | def _get_has_nsfw_concepts(self, features, params): method _run_safety_checker (line 237) | def _run_safety_checker(self, images, safety_model_params, jit=False): method _generate (line 267) | def _generate( method __call__ (line 395) | def __call__( function _p_generate (line 532) | def _p_generate( function _p_get_has_nsfw_concepts (line 562) | def _p_get_has_nsfw_concepts(pipe, features, params): function unshard (line 566) | def unshard(x: jnp.ndarray): function preprocess_image (line 573) | def preprocess_image(image, dtype): function preprocess_mask (line 582) | def preprocess_mask(mask, dtype): FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_onnx_stable_diffusion.py class OnnxStableDiffusionPipeline (line 33) | class OnnxStableDiffusionPipeline(DiffusionPipeline): method __init__ (line 46) | def __init__( method _encode_prompt (line 115) | def _encode_prompt( method check_inputs (line 218) | def check_inputs( method __call__ (line 265) | def __call__( class StableDiffusionOnnxPipeline (line 464) | class StableDiffusionOnnxPipeline(OnnxStableDiffusionPipeline): method __init__ (line 465) | def __init__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_onnx_stable_diffusion_img2img.py function preprocess (line 35) | def preprocess(image): class OnnxStableDiffusionImg2ImgPipeline (line 58) | class OnnxStableDiffusionImg2ImgPipeline(DiffusionPipeline): method __init__ (line 98) | def __init__( method _encode_prompt (line 168) | def _encode_prompt( method check_inputs (line 271) | def check_inputs( method __call__ (line 313) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_onnx_stable_diffusion_inpaint.py function prepare_mask_and_masked_image (line 38) | def prepare_mask_and_masked_image(image, mask, latents_shape): class OnnxStableDiffusionInpaintPipeline (line 56) | class OnnxStableDiffusionInpaintPipeline(DiffusionPipeline): method __init__ (line 96) | def __init__( method _encode_prompt (line 167) | def _encode_prompt( method check_inputs (line 271) | def check_inputs( method __call__ (line 319) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_onnx_stable_diffusion_upscale.py function preprocess (line 34) | def preprocess(image): class OnnxStableDiffusionUpscalePipeline (line 56) | class OnnxStableDiffusionUpscalePipeline(DiffusionPipeline): method __init__ (line 69) | def __init__( method check_inputs (line 145) | def check_inputs( method prepare_latents (line 230) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method decode_latents (line 239) | def decode_latents(self, latents): method _encode_prompt (line 246) | def _encode_prompt( method __call__ (line 349) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_output.py class StableDiffusionPipelineOutput (line 11) | class StableDiffusionPipelineOutput(BaseOutput): class FlaxStableDiffusionPipelineOutput (line 32) | class FlaxStableDiffusionPipelineOutput(BaseOutput): FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion.py function rescale_noise_cfg (line 59) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 73) | def retrieve_timesteps( class StableDiffusionPipeline (line 132) | class StableDiffusionPipeline( method __init__ (line 178) | def __init__( method _encode_prompt (line 270) | def _encode_prompt( method encode_prompt (line 302) | def encode_prompt( method encode_image (line 484) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 508) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 559) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 573) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 584) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 601) | def check_inputs( method prepare_latents (line 670) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 693) | def get_guidance_scale_embedding( method guidance_scale (line 724) | def guidance_scale(self): method guidance_rescale (line 728) | def guidance_rescale(self): method clip_skip (line 732) | def clip_skip(self): method do_classifier_free_guidance (line 739) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 743) | def cross_attention_kwargs(self): method num_timesteps (line 747) | def num_timesteps(self): method interrupt (line 751) | def interrupt(self): method __call__ (line 756) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_depth2img.py function retrieve_latents (line 40) | def retrieve_latents( function preprocess (line 54) | def preprocess(image): class StableDiffusionDepth2ImgPipeline (line 77) | class StableDiffusionDepth2ImgPipeline(DiffusionPipeline, TextualInversi... method __init__ (line 106) | def __init__( method _encode_prompt (line 152) | def _encode_prompt( method encode_prompt (line 185) | def encode_prompt( method run_safety_checker (line 368) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 383) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 395) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 412) | def check_inputs( method get_timesteps (line 464) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 476) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method prepare_depth_map (line 534) | def prepare_depth_map(self, image, depth_map, batch_size, do_classifie... method guidance_scale (line 586) | def guidance_scale(self): method clip_skip (line 590) | def clip_skip(self): method do_classifier_free_guidance (line 597) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 601) | def cross_attention_kwargs(self): method num_timesteps (line 605) | def num_timesteps(self): method __call__ (line 609) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_image_variation.py class StableDiffusionImageVariationPipeline (line 37) | class StableDiffusionImageVariationPipeline(DiffusionPipeline, StableDif... method __init__ (line 72) | def __init__( method _encode_image (line 133) | def _encode_image(self, image, device, num_images_per_prompt, do_class... method run_safety_checker (line 159) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 174) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 186) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 203) | def check_inputs(self, image, height, width, callback_steps): method prepare_latents (line 226) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 249) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_img2img.py function retrieve_latents (line 77) | def retrieve_latents( function preprocess (line 90) | def preprocess(image): function retrieve_timesteps (line 114) | def retrieve_timesteps( class StableDiffusionImg2ImgPipeline (line 173) | class StableDiffusionImg2ImgPipeline( method __init__ (line 219) | def __init__( method _encode_prompt (line 312) | def _encode_prompt( method encode_prompt (line 345) | def encode_prompt( method encode_image (line 528) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 553) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 605) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 620) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 632) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 649) | def check_inputs( method get_timesteps (line 717) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 728) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method get_guidance_scale_embedding (line 787) | def get_guidance_scale_embedding( method guidance_scale (line 818) | def guidance_scale(self): method clip_skip (line 822) | def clip_skip(self): method do_classifier_free_guidance (line 829) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 833) | def cross_attention_kwargs(self): method num_timesteps (line 837) | def num_timesteps(self): method interrupt (line 841) | def interrupt(self): method __call__ (line 846) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_inpaint.py function retrieve_latents (line 41) | def retrieve_latents( function retrieve_timesteps (line 55) | def retrieve_timesteps( class StableDiffusionInpaintPipeline (line 114) | class StableDiffusionInpaintPipeline( method __init__ (line 160) | def __init__( method _encode_prompt (line 261) | def _encode_prompt( method encode_prompt (line 294) | def encode_prompt( method encode_image (line 477) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 502) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 554) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 569) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 586) | def check_inputs( method prepare_latents (line 676) | def prepare_latents( method _encode_vae_image (line 739) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method prepare_mask_latents (line 753) | def prepare_mask_latents( method get_timesteps (line 799) | def get_timesteps(self, num_inference_steps, strength, device): method get_guidance_scale_embedding (line 811) | def get_guidance_scale_embedding( method guidance_scale (line 842) | def guidance_scale(self): method clip_skip (line 846) | def clip_skip(self): method do_classifier_free_guidance (line 853) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 857) | def cross_attention_kwargs(self): method num_timesteps (line 861) | def num_timesteps(self): method interrupt (line 865) | def interrupt(self): method __call__ (line 869) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_instruct_pix2pix.py function preprocess (line 39) | def preprocess(image): function retrieve_latents (line 63) | def retrieve_latents( class StableDiffusionInstructPix2PixPipeline (line 76) | class StableDiffusionInstructPix2PixPipeline( method __init__ (line 116) | def __init__( method __call__ (line 161) | def __call__( method _encode_prompt (line 493) | def _encode_prompt( method encode_image (line 643) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 667) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 727) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 742) | def prepare_extra_step_kwargs(self, generator, eta): method decode_latents (line 760) | def decode_latents(self, latents): method check_inputs (line 771) | def check_inputs( method prepare_latents (line 837) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method prepare_image_latents (line 859) | def prepare_image_latents( method guidance_scale (line 901) | def guidance_scale(self): method image_guidance_scale (line 905) | def image_guidance_scale(self): method num_timesteps (line 909) | def num_timesteps(self): method do_classifier_free_guidance (line 916) | def do_classifier_free_guidance(self): FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_latent_upscale.py function preprocess (line 37) | def preprocess(image): class StableDiffusionLatentUpscalePipeline (line 63) | class StableDiffusionLatentUpscalePipeline(DiffusionPipeline, StableDiff... method __init__ (line 88) | def __init__( method _encode_prompt (line 108) | def _encode_prompt(self, prompt, device, do_classifier_free_guidance, ... method decode_latents (line 199) | def decode_latents(self, latents): method check_inputs (line 210) | def check_inputs(self, prompt, image, callback_steps): method prepare_latents (line 248) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 262) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_upscale.py function preprocess (line 44) | def preprocess(image): class StableDiffusionUpscalePipeline (line 70) | class StableDiffusionUpscalePipeline( method __init__ (line 106) | def __init__( method run_safety_checker (line 154) | def run_safety_checker(self, image, device, dtype): method _encode_prompt (line 172) | def _encode_prompt( method encode_prompt (line 205) | def encode_prompt( method prepare_extra_step_kwargs (line 388) | def prepare_extra_step_kwargs(self, generator, eta): method decode_latents (line 406) | def decode_latents(self, latents): method check_inputs (line 417) | def check_inputs( method prepare_latents (line 502) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method upcast_vae (line 515) | def upcast_vae(self): method __call__ (line 535) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_stable_unclip.py class StableUnCLIPPipeline (line 61) | class StableUnCLIPPipeline(DiffusionPipeline, StableDiffusionMixin, Text... method __init__ (line 121) | def __init__( method _encode_prior_prompt (line 159) | def _encode_prior_prompt( method _encode_prompt (line 253) | def _encode_prompt( method encode_prompt (line 286) | def encode_prompt( method decode_latents (line 469) | def decode_latents(self, latents): method prepare_prior_extra_step_kwargs (line 481) | def prepare_prior_extra_step_kwargs(self, generator, eta): method prepare_extra_step_kwargs (line 499) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 516) | def check_inputs( method prepare_latents (line 577) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method noise_image_embeddings (line 588) | def noise_image_embeddings( method __call__ (line 636) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/pipeline_stable_unclip_img2img.py class StableUnCLIPImg2ImgPipeline (line 72) | class StableUnCLIPImg2ImgPipeline( method __init__ (line 128) | def __init__( method _encode_prompt (line 162) | def _encode_prompt( method _encode_image (line 194) | def _encode_image( method encode_prompt (line 250) | def encode_prompt( method decode_latents (line 433) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 445) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 462) | def check_inputs( method prepare_latents (line 546) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method noise_image_embeddings (line 569) | def noise_image_embeddings( method __call__ (line 617) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/safety_checker.py function cosine_distance (line 26) | def cosine_distance(image_embeds, text_embeds): class StableDiffusionSafetyChecker (line 32) | class StableDiffusionSafetyChecker(PreTrainedModel): method __init__ (line 38) | def __init__(self, config: CLIPConfig): method forward (line 51) | def forward(self, clip_input, images): method forward_onnx (line 103) | def forward_onnx(self, clip_input: torch.Tensor, images: torch.Tensor): FILE: diffusers/src/diffusers/pipelines/stable_diffusion/safety_checker_flax.py function jax_cosine_distance (line 25) | def jax_cosine_distance(emb_1, emb_2, eps=1e-12): class FlaxStableDiffusionSafetyCheckerModule (line 31) | class FlaxStableDiffusionSafetyCheckerModule(nn.Module): method setup (line 35) | def setup(self): method __call__ (line 47) | def __call__(self, clip_input): class FlaxStableDiffusionSafetyChecker (line 71) | class FlaxStableDiffusionSafetyChecker(FlaxPreTrainedModel): method __init__ (line 76) | def __init__( method init_weights (line 90) | def init_weights(self, rng: jax.Array, input_shape: Tuple, params: Fro... method __call__ (line 101) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion/stable_unclip_image_normalizer.py class StableUnCLIPImageNormalizer (line 24) | class StableUnCLIPImageNormalizer(ModelMixin, ConfigMixin): method __init__ (line 33) | def __init__( method to (line 42) | def to( method scale (line 51) | def scale(self, embeds): method unscale (line 55) | def unscale(self, embeds): FILE: diffusers/src/diffusers/pipelines/stable_diffusion_3/pipeline_output.py class StableDiffusion3PipelineOutput (line 11) | class StableDiffusion3PipelineOutput(BaseOutput): FILE: diffusers/src/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3.py function retrieve_timesteps (line 69) | def retrieve_timesteps( class StableDiffusion3Pipeline (line 128) | class StableDiffusion3Pipeline(DiffusionPipeline, SD3LoraLoaderMixin, Fr... method __init__ (line 166) | def __init__( method _get_t5_prompt_embeds (line 204) | def _get_t5_prompt_embeds( method _get_clip_prompt_embeds (line 255) | def _get_clip_prompt_embeds( method encode_prompt (line 310) | def encode_prompt( method check_inputs (line 474) | def check_inputs( method prepare_latents (line 560) | def prepare_latents( method guidance_scale (line 592) | def guidance_scale(self): method clip_skip (line 596) | def clip_skip(self): method do_classifier_free_guidance (line 603) | def do_classifier_free_guidance(self): method joint_attention_kwargs (line 607) | def joint_attention_kwargs(self): method num_timesteps (line 611) | def num_timesteps(self): method interrupt (line 615) | def interrupt(self): method __call__ (line 620) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3_img2img.py function retrieve_latents (line 75) | def retrieve_latents( function retrieve_timesteps (line 89) | def retrieve_timesteps( class StableDiffusion3Img2ImgPipeline (line 148) | class StableDiffusion3Img2ImgPipeline(DiffusionPipeline): method __init__ (line 186) | def __init__( method _get_t5_prompt_embeds (line 219) | def _get_t5_prompt_embeds( method _get_clip_prompt_embeds (line 271) | def _get_clip_prompt_embeds( method encode_prompt (line 327) | def encode_prompt( method check_inputs (line 491) | def check_inputs( method get_timesteps (line 576) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 587) | def prepare_latents(self, image, timestep, batch_size, num_images_per_... method guidance_scale (line 638) | def guidance_scale(self): method clip_skip (line 642) | def clip_skip(self): method do_classifier_free_guidance (line 649) | def do_classifier_free_guidance(self): method num_timesteps (line 653) | def num_timesteps(self): method interrupt (line 657) | def interrupt(self): method __call__ (line 662) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3_instructpix2pix.py function retrieve_latents (line 72) | def retrieve_latents( class StableDiffusion3InstructPix2PixPipeline (line 83) | class StableDiffusion3InstructPix2PixPipeline(DiffusionPipeline, SD3Lora... method __init__ (line 121) | def __init__( method _get_t5_prompt_embeds (line 159) | def _get_t5_prompt_embeds( method _get_clip_prompt_embeds (line 210) | def _get_clip_prompt_embeds( method encode_prompt (line 265) | def encode_prompt( method check_inputs (line 429) | def check_inputs( method prepare_latents (line 515) | def prepare_latents( method prepare_image_latents (line 546) | def prepare_image_latents( method guidance_scale (line 597) | def guidance_scale(self): method image_guidance_scale (line 600) | def image_guidance_scale(self): method clip_skip (line 604) | def clip_skip(self): method do_classifier_free_guidance (line 611) | def do_classifier_free_guidance(self): method joint_attention_kwargs (line 615) | def joint_attention_kwargs(self): method num_timesteps (line 619) | def num_timesteps(self): method interrupt (line 623) | def interrupt(self): method __call__ (line 628) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_attend_and_excite/pipeline_stable_diffusion_attend_and_excite.py class AttentionStore (line 82) | class AttentionStore: method get_empty_store (line 84) | def get_empty_store(): method __call__ (line 87) | def __call__(self, attn, is_cross: bool, place_in_unet: str): method between_steps (line 97) | def between_steps(self): method get_average_attention (line 101) | def get_average_attention(self): method aggregate_attention (line 105) | def aggregate_attention(self, from_where: List[str]) -> torch.Tensor: method reset (line 117) | def reset(self): method __init__ (line 122) | def __init__(self, attn_res): class AttendExciteAttnProcessor (line 135) | class AttendExciteAttnProcessor: method __init__ (line 136) | def __init__(self, attnstore, place_in_unet): method __call__ (line 141) | def __call__(self, attn: Attention, hidden_states, encoder_hidden_stat... class StableDiffusionAttendAndExcitePipeline (line 173) | class StableDiffusionAttendAndExcitePipeline(DiffusionPipeline, StableDi... method __init__ (line 207) | def __init__( method _encode_prompt (line 250) | def _encode_prompt( method encode_prompt (line 283) | def encode_prompt( method run_safety_checker (line 466) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 481) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 493) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 510) | def check_inputs( method prepare_latents (line 584) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _compute_max_attention_per_index (line 607) | def _compute_max_attention_per_index( method _aggregate_and_get_max_attention_per_token (line 629) | def _aggregate_and_get_max_attention_per_token( method _compute_loss (line 644) | def _compute_loss(max_attention_per_index: List[torch.Tensor]) -> torc... method _update_latent (line 651) | def _update_latent(latents: torch.Tensor, loss: torch.Tensor, step_siz... method _perform_iterative_refinement_step (line 657) | def _perform_iterative_refinement_step( method register_attention_control (line 711) | def register_attention_control(self): method get_indices (line 730) | def get_indices(self, prompt: str) -> Dict[str, int]: method __call__ (line 738) | def __call__( class GaussianSmoothing (line 1035) | class GaussianSmoothing(torch.nn.Module): method __init__ (line 1048) | def __init__( method forward (line 1089) | def forward(self, input): FILE: diffusers/src/diffusers/pipelines/stable_diffusion_diffedit/pipeline_stable_diffusion_diffedit.py class DiffEditInversionPipelineOutput (line 51) | class DiffEditInversionPipelineOutput(BaseOutput): function auto_corr_loss (line 139) | def auto_corr_loss(hidden_states, generator=None): function kl_divergence (line 155) | def kl_divergence(hidden_states): function preprocess (line 160) | def preprocess(image): function preprocess_mask (line 183) | def preprocess_mask(mask, batch_size: int = 1): class StableDiffusionDiffEditPipeline (line 236) | class StableDiffusionDiffEditPipeline( method __init__ (line 281) | def __init__( method _encode_prompt (line 375) | def _encode_prompt( method encode_prompt (line 408) | def encode_prompt( method run_safety_checker (line 591) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 606) | def prepare_extra_step_kwargs(self, generator, eta): method decode_latents (line 624) | def decode_latents(self, latents): method check_inputs (line 635) | def check_inputs( method check_source_inputs (line 683) | def check_source_inputs( method get_timesteps (line 718) | def get_timesteps(self, num_inference_steps, strength, device): method get_inverse_timesteps (line 727) | def get_inverse_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 741) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method prepare_image_latents (line 763) | def prepare_image_latents(self, image, batch_size, dtype, device, gene... method get_epsilon (line 812) | def get_epsilon(self, model_output: torch.Tensor, sample: torch.Tensor... method generate_mask (line 831) | def generate_mask( method invert (line 1050) | def invert( method __call__ (line 1288) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_gligen/pipeline_stable_diffusion_gligen.py class StableDiffusionGLIGENPipeline (line 102) | class StableDiffusionGLIGENPipeline(DiffusionPipeline, StableDiffusionMi... method __init__ (line 133) | def __init__( method _encode_prompt (line 176) | def _encode_prompt( method encode_prompt (line 209) | def encode_prompt( method run_safety_checker (line 392) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 406) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 423) | def check_inputs( method prepare_latents (line 479) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method enable_fuser (line 501) | def enable_fuser(self, enabled=True): method draw_inpaint_mask_from_boxes (line 506) | def draw_inpaint_mask_from_boxes(self, boxes, size): method crop (line 514) | def crop(self, im, new_width, new_height): method target_size_center_crop (line 522) | def target_size_center_crop(self, im, new_hw): method __call__ (line 530) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_gligen/pipeline_stable_diffusion_gligen_text_image.py class StableDiffusionGLIGENTextImagePipeline (line 148) | class StableDiffusionGLIGENTextImagePipeline(DiffusionPipeline, StableDi... method __init__ (line 185) | def __init__( method encode_prompt (line 234) | def encode_prompt( method run_safety_checker (line 417) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 432) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 450) | def check_inputs( method prepare_latents (line 503) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method enable_fuser (line 525) | def enable_fuser(self, enabled=True): method draw_inpaint_mask_from_boxes (line 530) | def draw_inpaint_mask_from_boxes(self, boxes, size): method crop (line 542) | def crop(self, im, new_width, new_height): method target_size_center_crop (line 553) | def target_size_center_crop(self, im, new_hw): method complete_mask (line 562) | def complete_mask(self, has_mask, max_objs, device): method get_clip_feature (line 578) | def get_clip_feature(self, input, normalize_constant, device, is_image... method get_cross_attention_kwargs_with_grounded (line 602) | def get_cross_attention_kwargs_with_grounded( method get_cross_attention_kwargs_without_grounded (line 666) | def get_cross_attention_kwargs_without_grounded(self, hidden_size, rep... method __call__ (line 691) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_k_diffusion/pipeline_stable_diffusion_k_diffusion.py class ModelWrapper (line 36) | class ModelWrapper: method __init__ (line 37) | def __init__(self, model, alphas_cumprod): method apply_model (line 41) | def apply_model(self, *args, **kwargs): class StableDiffusionKDiffusionPipeline (line 50) | class StableDiffusionKDiffusionPipeline( method __init__ (line 95) | def __init__( method set_scheduler (line 136) | def set_scheduler(self, scheduler_type: str): method _encode_prompt (line 150) | def _encode_prompt( method encode_prompt (line 183) | def encode_prompt( method run_safety_checker (line 366) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 381) | def decode_latents(self, latents): method check_inputs (line 392) | def check_inputs( method prepare_latents (line 444) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 462) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_k_diffusion/pipeline_stable_diffusion_xl_k_diffusion.py class ModelWrapper (line 78) | class ModelWrapper: method __init__ (line 79) | def __init__(self, model, alphas_cumprod): method apply_model (line 83) | def apply_model(self, *args, **kwargs): class StableDiffusionXLKDiffusionPipeline (line 92) | class StableDiffusionXLKDiffusionPipeline( method __init__ (line 150) | def __init__( method set_scheduler (line 187) | def set_scheduler(self, scheduler_type: str): method encode_prompt (line 201) | def encode_prompt( method check_inputs (line 435) | def check_inputs( method prepare_latents (line 499) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 519) | def _get_add_time_ids( method upcast_vae (line 538) | def upcast_vae(self): method guidance_scale (line 559) | def guidance_scale(self): method clip_skip (line 563) | def clip_skip(self): method do_classifier_free_guidance (line 570) | def do_classifier_free_guidance(self): method __call__ (line 575) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_ldm3d/pipeline_stable_diffusion_ldm3d.py function rescale_noise_cfg (line 63) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 78) | def retrieve_timesteps( class LDM3DPipelineOutput (line 138) | class LDM3DPipelineOutput(BaseOutput): class StableDiffusionLDM3DPipeline (line 159) | class StableDiffusionLDM3DPipeline( method __init__ (line 205) | def __init__( method _encode_prompt (line 250) | def _encode_prompt( method encode_prompt (line 283) | def encode_prompt( method encode_image (line 466) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 491) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 542) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 558) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 576) | def check_inputs( method prepare_latents (line 645) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 668) | def get_guidance_scale_embedding( method guidance_scale (line 699) | def guidance_scale(self): method guidance_rescale (line 703) | def guidance_rescale(self): method clip_skip (line 707) | def clip_skip(self): method do_classifier_free_guidance (line 714) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 718) | def cross_attention_kwargs(self): method num_timesteps (line 722) | def num_timesteps(self): method interrupt (line 726) | def interrupt(self): method __call__ (line 731) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_panorama/pipeline_stable_diffusion_panorama.py function rescale_noise_cfg (line 63) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 78) | def retrieve_timesteps( class StableDiffusionPanoramaPipeline (line 137) | class StableDiffusionPanoramaPipeline( method __init__ (line 177) | def __init__( method _encode_prompt (line 222) | def _encode_prompt( method encode_prompt (line 255) | def encode_prompt( method encode_image (line 438) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 463) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 515) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 530) | def decode_latents(self, latents): method decode_latents_with_padding (line 541) | def decode_latents_with_padding(self, latents: torch.Tensor, padding: ... method prepare_extra_step_kwargs (line 568) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 586) | def check_inputs( method prepare_latents (line 656) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method get_guidance_scale_embedding (line 679) | def get_guidance_scale_embedding( method get_views (line 709) | def get_views( method guidance_scale (line 752) | def guidance_scale(self): method guidance_rescale (line 756) | def guidance_rescale(self): method cross_attention_kwargs (line 760) | def cross_attention_kwargs(self): method clip_skip (line 764) | def clip_skip(self): method do_classifier_free_guidance (line 768) | def do_classifier_free_guidance(self): method num_timesteps (line 772) | def num_timesteps(self): method interrupt (line 776) | def interrupt(self): method __call__ (line 781) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_safe/__init__.py class SafetyConfig (line 21) | class SafetyConfig(object): FILE: diffusers/src/diffusers/pipelines/stable_diffusion_safe/pipeline_output.py class StableDiffusionSafePipelineOutput (line 13) | class StableDiffusionSafePipelineOutput(BaseOutput): FILE: diffusers/src/diffusers/pipelines/stable_diffusion_safe/pipeline_stable_diffusion_safe.py class StableDiffusionPipelineSafe (line 25) | class StableDiffusionPipelineSafe(DiffusionPipeline, StableDiffusionMixi... method __init__ (line 58) | def __init__( method safety_concept (line 156) | def safety_concept(self): method safety_concept (line 166) | def safety_concept(self, concept): method _encode_prompt (line 176) | def _encode_prompt( method run_safety_checker (line 311) | def run_safety_checker(self, image, device, dtype, enable_safety_guida... method decode_latents (line 335) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 347) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 365) | def check_inputs( method prepare_latents (line 418) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method perform_safety_guidance (line 440) | def perform_safety_guidance( method encode_image (line 483) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method __call__ (line 508) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_safe/safety_checker.py function cosine_distance (line 25) | def cosine_distance(image_embeds, text_embeds): class SafeStableDiffusionSafetyChecker (line 31) | class SafeStableDiffusionSafetyChecker(PreTrainedModel): method __init__ (line 36) | def __init__(self, config: CLIPConfig): method forward (line 49) | def forward(self, clip_input, images): method forward_onnx (line 88) | def forward_onnx(self, clip_input: torch.Tensor, images: torch.Tensor): FILE: diffusers/src/diffusers/pipelines/stable_diffusion_sag/pipeline_stable_diffusion_sag.py class CrossAttnStoreProcessor (line 61) | class CrossAttnStoreProcessor: method __init__ (line 62) | def __init__(self): method __call__ (line 65) | def __call__( class StableDiffusionSAGPipeline (line 101) | class StableDiffusionSAGPipeline(DiffusionPipeline, StableDiffusionMixin... method __init__ (line 136) | def __init__( method _encode_prompt (line 165) | def _encode_prompt( method encode_prompt (line 198) | def encode_prompt( method encode_image (line 381) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 405) | def prepare_ip_adapter_image_embeds( method run_safety_checker (line 440) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 455) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 467) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 485) | def check_inputs( method prepare_latents (line 538) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 562) | def __call__( method sag_masking (line 866) | def sag_masking(self, original_latents, attn_map, map_size, t, eps): method pred_x0 (line 896) | def pred_x0(self, sample, model_output, timestep): method pred_epsilon (line 916) | def pred_epsilon(self, sample, model_output, timestep): function gaussian_blur_2d (line 936) | def gaussian_blur_2d(img, kernel_size, sigma): FILE: diffusers/src/diffusers/pipelines/stable_diffusion_xl/pipeline_flax_stable_diffusion_xl.py class FlaxStableDiffusionXLPipeline (line 42) | class FlaxStableDiffusionXLPipeline(FlaxDiffusionPipeline): method __init__ (line 43) | def __init__( method prepare_inputs (line 70) | def prepare_inputs(self, prompt: Union[str, List[str]]): method __call__ (line 88) | def __call__( method get_embeddings (line 147) | def get_embeddings(self, prompt_ids: jnp.array, params): method _get_add_time_ids (line 164) | def _get_add_time_ids(self, original_size, crops_coords_top_left, targ... method _generate (line 169) | def _generate( function _p_generate (line 284) | def _p_generate( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_xl/pipeline_output.py class StableDiffusionXLPipelineOutput (line 11) | class StableDiffusionXLPipelineOutput(BaseOutput): class FlaxStableDiffusionXLPipelineOutput (line 28) | class FlaxStableDiffusionXLPipelineOutput(BaseOutput): FILE: diffusers/src/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl.py function rescale_noise_cfg (line 91) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 106) | def retrieve_timesteps( class StableDiffusionXLPipeline (line 165) | class StableDiffusionXLPipeline( method __init__ (line 237) | def __init__( method encode_prompt (line 277) | def encode_prompt( method encode_image (line 512) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 537) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 589) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 606) | def check_inputs( method prepare_latents (line 703) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 725) | def _get_add_time_ids( method upcast_vae (line 743) | def upcast_vae(self): method get_guidance_scale_embedding (line 764) | def get_guidance_scale_embedding( method guidance_scale (line 795) | def guidance_scale(self): method guidance_rescale (line 799) | def guidance_rescale(self): method clip_skip (line 803) | def clip_skip(self): method do_classifier_free_guidance (line 810) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 814) | def cross_attention_kwargs(self): method denoising_end (line 818) | def denoising_end(self): method num_timesteps (line 822) | def num_timesteps(self): method interrupt (line 826) | def interrupt(self): method __call__ (line 831) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl_img2img.py function rescale_noise_cfg (line 94) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_latents (line 109) | def retrieve_latents( function retrieve_timesteps (line 123) | def retrieve_timesteps( class StableDiffusionXLImg2ImgPipeline (line 182) | class StableDiffusionXLImg2ImgPipeline( method __init__ (line 257) | def __init__( method encode_prompt (line 298) | def encode_prompt( method prepare_extra_step_kwargs (line 533) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 550) | def check_inputs( method get_timesteps (line 640) | def get_timesteps(self, num_inference_steps, strength, device, denoisi... method prepare_latents (line 676) | def prepare_latents( method encode_image (line 756) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 781) | def prepare_ip_adapter_image_embeds( method _get_add_time_ids (line 832) | def _get_add_time_ids( method upcast_vae (line 884) | def upcast_vae(self): method get_guidance_scale_embedding (line 904) | def get_guidance_scale_embedding( method guidance_scale (line 935) | def guidance_scale(self): method guidance_rescale (line 939) | def guidance_rescale(self): method clip_skip (line 943) | def clip_skip(self): method do_classifier_free_guidance (line 950) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 954) | def cross_attention_kwargs(self): method denoising_end (line 958) | def denoising_end(self): method denoising_start (line 962) | def denoising_start(self): method num_timesteps (line 966) | def num_timesteps(self): method interrupt (line 970) | def interrupt(self): method __call__ (line 975) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl_inpaint.py function rescale_noise_cfg (line 105) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function mask_pil_to_torch (line 119) | def mask_pil_to_torch(mask, height, width): function retrieve_latents (line 136) | def retrieve_latents( function retrieve_timesteps (line 150) | def retrieve_timesteps( class StableDiffusionXLInpaintPipeline (line 209) | class StableDiffusionXLInpaintPipeline( method __init__ (line 287) | def __init__( method encode_image (line 331) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 356) | def prepare_ip_adapter_image_embeds( method encode_prompt (line 408) | def encode_prompt( method prepare_extra_step_kwargs (line 643) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 660) | def check_inputs( method prepare_latents (line 764) | def prepare_latents( method _encode_vae_image (line 830) | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Gene... method prepare_mask_latents (line 853) | def prepare_mask_latents( method get_timesteps (line 907) | def get_timesteps(self, num_inference_steps, strength, device, denoisi... method _get_add_time_ids (line 944) | def _get_add_time_ids( method upcast_vae (line 996) | def upcast_vae(self): method get_guidance_scale_embedding (line 1016) | def get_guidance_scale_embedding( method guidance_scale (line 1047) | def guidance_scale(self): method guidance_rescale (line 1051) | def guidance_rescale(self): method clip_skip (line 1055) | def clip_skip(self): method do_classifier_free_guidance (line 1062) | def do_classifier_free_guidance(self): method cross_attention_kwargs (line 1066) | def cross_attention_kwargs(self): method denoising_end (line 1070) | def denoising_end(self): method denoising_start (line 1074) | def denoising_start(self): method num_timesteps (line 1078) | def num_timesteps(self): method interrupt (line 1082) | def interrupt(self): method __call__ (line 1087) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl_instruct_pix2pix.py function retrieve_latents (line 93) | def retrieve_latents( function rescale_noise_cfg (line 106) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class StableDiffusionXLInstructPix2PixPipeline (line 120) | class StableDiffusionXLInstructPix2PixPipeline( method __init__ (line 179) | def __init__( method encode_prompt (line 216) | def encode_prompt( method prepare_extra_step_kwargs (line 422) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 439) | def check_inputs( method prepare_latents (line 488) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method prepare_image_latents (line 510) | def prepare_image_latents( method _get_add_time_ids (line 568) | def _get_add_time_ids( method upcast_vae (line 587) | def upcast_vae(self): method __call__ (line 609) | def __call__( FILE: diffusers/src/diffusers/pipelines/stable_diffusion_xl/watermark.py class StableDiffusionXLWatermarker (line 17) | class StableDiffusionXLWatermarker: method __init__ (line 18) | def __init__(self): method apply_watermark (line 24) | def apply_watermark(self, images: torch.Tensor): FILE: diffusers/src/diffusers/pipelines/stable_video_diffusion/pipeline_stable_video_diffusion.py function _append_dims (line 57) | def _append_dims(x, target_dims): function retrieve_timesteps (line 66) | def retrieve_timesteps( class StableVideoDiffusionPipelineOutput (line 126) | class StableVideoDiffusionPipelineOutput(BaseOutput): class StableVideoDiffusionPipeline (line 139) | class StableVideoDiffusionPipeline(DiffusionPipeline): method __init__ (line 163) | def __init__( method _encode_image (line 183) | def _encode_image( method _encode_vae_image (line 231) | def _encode_vae_image( method _get_add_time_ids (line 254) | def _get_add_time_ids( method decode_latents (line 282) | def decode_latents(self, latents: torch.Tensor, num_frames: int, decod... method check_inputs (line 311) | def check_inputs(self, image, height, width): method prepare_latents (line 325) | def prepare_latents( method guidance_scale (line 360) | def guidance_scale(self): method do_classifier_free_guidance (line 367) | def do_classifier_free_guidance(self): method num_timesteps (line 373) | def num_timesteps(self): method __call__ (line 378) | def __call__( function _resize_with_antialiasing (line 622) | def _resize_with_antialiasing(input, size, interpolation="bicubic", alig... function _compute_padding (line 651) | def _compute_padding(kernel_size): function _filter2d (line 674) | def _filter2d(input, kernel): function _gaussian (line 697) | def _gaussian(window_size: int, sigma): function _gaussian_blur2d (line 713) | def _gaussian_blur2d(input, kernel_size, sigma): FILE: diffusers/src/diffusers/pipelines/t2i_adapter/pipeline_stable_diffusion_adapter.py class StableDiffusionAdapterPipelineOutput (line 45) | class StableDiffusionAdapterPipelineOutput(BaseOutput): function _preprocess_adapter_image (line 94) | def _preprocess_adapter_image(image, height, width): function retrieve_timesteps (line 122) | def retrieve_timesteps( class StableDiffusionAdapterPipeline (line 181) | class StableDiffusionAdapterPipeline(DiffusionPipeline, StableDiffusionM... method __init__ (line 219) | def __init__( method _encode_prompt (line 267) | def _encode_prompt( method encode_prompt (line 300) | def encode_prompt( method run_safety_checker (line 483) | def run_safety_checker(self, image, device, dtype): method decode_latents (line 498) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 510) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 527) | def check_inputs( method prepare_latents (line 587) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _default_height_width (line 609) | def _default_height_width(self, height, width, image): method get_guidance_scale_embedding (line 637) | def get_guidance_scale_embedding( method guidance_scale (line 668) | def guidance_scale(self): method do_classifier_free_guidance (line 675) | def do_classifier_free_guidance(self): method __call__ (line 680) | def __call__( FILE: diffusers/src/diffusers/pipelines/t2i_adapter/pipeline_stable_diffusion_xl_adapter.py function _preprocess_adapter_image (line 95) | def _preprocess_adapter_image(image, height, width): function rescale_noise_cfg (line 123) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): function retrieve_timesteps (line 138) | def retrieve_timesteps( class StableDiffusionXLAdapterPipeline (line 197) | class StableDiffusionXLAdapterPipeline( method __init__ (line 256) | def __init__( method encode_prompt (line 290) | def encode_prompt( method encode_image (line 525) | def encode_image(self, image, device, num_images_per_prompt, output_hi... method prepare_ip_adapter_image_embeds (line 550) | def prepare_ip_adapter_image_embeds( method prepare_extra_step_kwargs (line 602) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 620) | def check_inputs( method prepare_latents (line 717) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method _get_add_time_ids (line 740) | def _get_add_time_ids( method upcast_vae (line 759) | def upcast_vae(self): method _default_height_width (line 779) | def _default_height_width(self, height, width, image): method get_guidance_scale_embedding (line 807) | def get_guidance_scale_embedding( method guidance_scale (line 838) | def guidance_scale(self): method do_classifier_free_guidance (line 845) | def do_classifier_free_guidance(self): method __call__ (line 850) | def __call__( FILE: diffusers/src/diffusers/pipelines/text_to_video_synthesis/pipeline_output.py class TextToVideoSDPipelineOutput (line 14) | class TextToVideoSDPipelineOutput(BaseOutput): FILE: diffusers/src/diffusers/pipelines/text_to_video_synthesis/pipeline_text_to_video_synth.py class TextToVideoSDPipeline (line 61) | class TextToVideoSDPipeline(DiffusionPipeline, StableDiffusionMixin, Tex... method __init__ (line 89) | def __init__( method _encode_prompt (line 110) | def _encode_prompt( method encode_prompt (line 143) | def encode_prompt( method decode_latents (line 325) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 338) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 356) | def check_inputs( method prepare_latents (line 408) | def prepare_latents( method __call__ (line 435) | def __call__( FILE: diffusers/src/diffusers/pipelines/text_to_video_synthesis/pipeline_text_to_video_synth_img2img.py function retrieve_latents (line 83) | def retrieve_latents( class VideoToVideoSDPipeline (line 96) | class VideoToVideoSDPipeline(DiffusionPipeline, StableDiffusionMixin, Te... method __init__ (line 124) | def __init__( method _encode_prompt (line 145) | def _encode_prompt( method encode_prompt (line 178) | def encode_prompt( method decode_latents (line 361) | def decode_latents(self, latents): method prepare_extra_step_kwargs (line 374) | def prepare_extra_step_kwargs(self, generator, eta): method check_inputs (line 391) | def check_inputs( method get_timesteps (line 443) | def get_timesteps(self, num_inference_steps, strength, device): method prepare_latents (line 454) | def prepare_latents(self, video, timestep, batch_size, dtype, device, ... method __call__ (line 500) | def __call__( FILE: diffusers/src/diffusers/pipelines/text_to_video_synthesis/pipeline_text_to_video_zero.py function rearrange_0 (line 27) | def rearrange_0(tensor, f): function rearrange_1 (line 33) | def rearrange_1(tensor): function rearrange_3 (line 38) | def rearrange_3(tensor, f): function rearrange_4 (line 43) | def rearrange_4(tensor): class CrossFrameAttnProcessor (line 48) | class CrossFrameAttnProcessor: method __init__ (line 58) | def __init__(self, batch_size=2): method __call__ (line 61) | def __call__(self, attn, hidden_states, encoder_hidden_states=None, at... class CrossFrameAttnProcessor2_0 (line 107) | class CrossFrameAttnProcessor2_0: method __init__ (line 117) | def __init__(self, batch_size=2): method __call__ (line 122) | def __call__(self, attn, hidden_states, encoder_hidden_states=None, at... class TextToVideoPipelineOutput (line 183) | class TextToVideoPipelineOutput(BaseOutput): function coords_grid (line 200) | def coords_grid(batch, ht, wd, device): function warp_single_latent (line 207) | def warp_single_latent(latent, reference_flow): function create_motion_field (line 234) | def create_motion_field(motion_field_strength_x, motion_field_strength_y... function create_motion_field_and_warp_latents (line 257) | def create_motion_field_and_warp_latents(motion_field_strength_x, motion... class TextToVideoZeroPipeline (line 284) | class TextToVideoZeroPipeline(DiffusionPipeline, StableDiffusionMixin, T... method __init__ (line 311) | def __init__( method forward_loop (line 344) | def forward_loop(self, x_t0, t0, t1, generator): method backward_loop (line 368) | def backward_loop( method check_inputs (line 444) | def check_inputs( method prepare_latents (line 497) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method __call__ (line 520) | def __call__( method run_safety_checker (line 758) | def run_safety_checker(self, image, device, dtype): method prepare_extra_step_kwargs (line 773) | def prepare_extra_step_kwargs(self, generator, eta): method encode_prompt (line 791) | def encode_prompt( method decode_latents (line 973) | def decode_latents(self, latents): FILE: diffusers/src/diffusers/pipelines/text_to_video_synthesis/pipeline_text_to_video_zero_sdxl.py function rearrange_0 (line 51) | def rearrange_0(tensor, f): function rearrange_1 (line 58) | def rearrange_1(tensor): function rearrange_3 (line 64) | def rearrange_3(tensor, f): function rearrange_4 (line 70) | def rearrange_4(tensor): class CrossFrameAttnProcessor (line 76) | class CrossFrameAttnProcessor: method __init__ (line 86) | def __init__(self, batch_size=2): method __call__ (line 89) | def __call__(self, attn, hidden_states, encoder_hidden_states=None, at... class CrossFrameAttnProcessor2_0 (line 136) | class CrossFrameAttnProcessor2_0: method __init__ (line 146) | def __init__(self, batch_size=2): method __call__ (line 151) | def __call__(self, attn, hidden_states, encoder_hidden_states=None, at... class TextToVideoSDXLPipelineOutput (line 212) | class TextToVideoSDXLPipelineOutput(BaseOutput): function coords_grid (line 226) | def coords_grid(batch, ht, wd, device): function warp_single_latent (line 234) | def warp_single_latent(latent, reference_flow): function create_motion_field (line 262) | def create_motion_field(motion_field_strength_x, motion_field_strength_y... function create_motion_field_and_warp_latents (line 286) | def create_motion_field_and_warp_latents(motion_field_strength_x, motion... function rescale_noise_cfg (line 314) | def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): class TextToVideoZeroSDXLPipeline (line 328) | class TextToVideoZeroSDXLPipeline( method __init__ (line 375) | def __init__( method prepare_extra_step_kwargs (line 415) | def prepare_extra_step_kwargs(self, generator, eta): method upcast_vae (line 433) | def upcast_vae(self): method _get_add_time_ids (line 454) | def _get_add_time_ids( method prepare_latents (line 473) | def prepare_latents(self, batch_size, num_channels_latents, height, wi... method check_inputs (line 495) | def check_inputs( method encode_prompt (line 575) | def encode_prompt( method forward_loop (line 810) | def forward_loop(self, x_t0, t0, t1, generator): method backward_loop (line 834) | def backward_loop( method __call__ (line 920) | def __call__( FILE: diffusers/src/diffusers/pipelines/unclip/pipeline_unclip.py class UnCLIPPipeline (line 34) | class UnCLIPPipeline(DiffusionPipeline): method __init__ (line 81) | def __init__( method prepare_latents (line 109) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method _encode_prompt (line 120) | def _encode_prompt( method __call__ (line 212) | def __call__( FILE: diffusers/src/diffusers/pipelines/unclip/pipeline_unclip_image_variation.py class UnCLIPImageVariationPipeline (line 39) | class UnCLIPImageVariationPipeline(DiffusionPipeline): method __init__ (line 82) | def __init__( method prepare_latents (line 111) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method _encode_prompt (line 122) | def _encode_prompt(self, prompt, device, num_images_per_prompt, do_cla... method _encode_image (line 185) | def _encode_image(self, image, device, num_images_per_prompt, image_em... method __call__ (line 200) | def __call__( FILE: diffusers/src/diffusers/pipelines/unclip/text_proj.py class UnCLIPTextProjModel (line 22) | class UnCLIPTextProjModel(ModelMixin, ConfigMixin): method __init__ (line 31) | def __init__( method forward (line 55) | def forward(self, *, image_embeddings, prompt_embeds, text_encoder_hid... FILE: diffusers/src/diffusers/pipelines/unidiffuser/modeling_text_decoder.py class UniDiffuserTextDecoder (line 14) | class UniDiffuserTextDecoder(ModelMixin, ConfigMixin, ModuleUtilsMixin): method __init__ (line 67) | def __init__( method forward (line 131) | def forward( method get_dummy_token (line 163) | def get_dummy_token(self, batch_size: int, device: torch.device) -> to... method encode (line 166) | def encode(self, prefix): method generate_captions (line 170) | def generate_captions(self, features, eos_token_id, device): method generate_beam (line 202) | def generate_beam( FILE: diffusers/src/diffusers/pipelines/unidiffuser/modeling_uvit.py function _no_grad_trunc_normal_ (line 20) | def _no_grad_trunc_normal_(tensor, mean, std, a, b): function trunc_normal_ (line 57) | def trunc_normal_(tensor, mean=0.0, std=1.0, a=-2.0, b=2.0): class PatchEmbed (line 76) | class PatchEmbed(nn.Module): method __init__ (line 79) | def __init__( method forward (line 110) | def forward(self, latent): class SkipBlock (line 122) | class SkipBlock(nn.Module): method __init__ (line 123) | def __init__(self, dim: int): method forward (line 131) | def forward(self, x, skip): class UTransformerBlock (line 141) | class UTransformerBlock(nn.Module): method __init__ (line 175) | def __init__( method forward (line 252) | def forward( class UniDiffuserBlock (line 334) | class UniDiffuserBlock(nn.Module): method __init__ (line 370) | def __init__( method forward (line 447) | def forward( class UTransformer2DModel (line 531) | class UTransformer2DModel(ModelMixin, ConfigMixin): method __init__ (line 590) | def __init__( method forward (line 728) | def forward( class UniDiffuserModel (line 833) | class UniDiffuserModel(ModelMixin, ConfigMixin): method __init__ (line 898) | def __init__( method no_weight_decay (line 1035) | def no_weight_decay(self): method forward (line 1038) | def forward( FILE: diffusers/src/diffusers/pipelines/unidiffuser/pipeline_unidiffuser.py class ImageTextPipelineOutput (line 34) | class ImageTextPipelineOutput(BaseOutput): class UniDiffuserPipeline (line 51) | class UniDiffuserPipeline(DiffusionPipeline): method __init__ (line 88) | def __init__( method prepare_extra_step_kwargs (line 139) | def prepare_extra_step_kwargs(self, generator, eta): method _infer_mode (line 156) | def _infer_mode(self, prompt, prompt_embeds, image, latents, prompt_la... method enable_vae_slicing (line 215) | def enable_vae_slicing(self): method disable_vae_slicing (line 223) | def disable_vae_slicing(self): method enable_vae_tiling (line 231) | def enable_vae_tiling(self): method disable_vae_tiling (line 240) | def disable_vae_tiling(self): method set_text_mode (line 248) | def set_text_mode(self): method set_image_mode (line 252) | def set_image_mode(self): method set_text_to_image_mode (line 256) | def set_text_to_image_mode(self): method set_image_to_text_mode (line 260) | def set_image_to_text_mode(self): method set_joint_mode (line 264) | def set_joint_mode(self): method reset_mode (line 268) | def reset_mode(self): method _infer_batch_size (line 272) | def _infer_batch_size( method _encode_prompt (line 349) | def _encode_prompt( method encode_prompt (line 382) | def encode_prompt( method encode_image_vae_latents (line 566) | def encode_image_vae_latents( method encode_image_clip_latents (line 626) | def encode_image_clip_latents( method prepare_text_latents (line 682) | def prepare_text_latents( method prepare_image_vae_latents (line 706) | def prepare_image_vae_latents( method prepare_image_clip_latents (line 741) | def prepare_image_clip_latents( method decode_text_latents (line 763) | def decode_text_latents(self, text_latents, device): method _split (line 774) | def _split(self, x, height, width): method _combine (line 790) | def _combine(self, img_vae, img_clip): method _split_joint (line 799) | def _split_joint(self, x, height, width): method _combine_joint (line 818) | def _combine_joint(self, img_vae, img_clip, text): method _get_noise_pred (line 829) | def _get_noise_pred( method check_latents_shape (line 951) | def check_latents_shape(self, latents_name, latents, expected_shape): method check_inputs (line 967) | def check_inputs( method __call__ (line 1081) | def __call__( FILE: diffusers/src/diffusers/pipelines/wuerstchen/modeling_paella_vq_model.py class MixingResidualBlock (line 28) | class MixingResidualBlock(nn.Module): method __init__ (line 33) | def __init__(self, inp_channels, embed_dim): method forward (line 49) | def forward(self, x): class PaellaVQModel (line 58) | class PaellaVQModel(ModelMixin, ConfigMixin): method __init__ (line 77) | def __init__( method encode (line 133) | def encode(self, x: torch.Tensor, return_dict: bool = True) -> VQEncod... method decode (line 143) | def decode( method forward (line 158) | def forward(self, sample: torch.Tensor, return_dict: bool = True) -> U... FILE: diffusers/src/diffusers/pipelines/wuerstchen/modeling_wuerstchen_common.py class WuerstchenLayerNorm (line 7) | class WuerstchenLayerNorm(nn.LayerNorm): method __init__ (line 8) | def __init__(self, *args, **kwargs): method forward (line 11) | def forward(self, x): class TimestepBlock (line 17) | class TimestepBlock(nn.Module): method __init__ (line 18) | def __init__(self, c, c_timestep): method forward (line 23) | def forward(self, x, t): class ResBlock (line 28) | class ResBlock(nn.Module): method __init__ (line 29) | def __init__(self, c, c_skip=0, kernel_size=3, dropout=0.0): method forward (line 38) | def forward(self, x, x_skip=None): class GlobalResponseNorm (line 48) | class GlobalResponseNorm(nn.Module): method __init__ (line 49) | def __init__(self, dim): method forward (line 54) | def forward(self, x): class AttnBlock (line 60) | class AttnBlock(nn.Module): method __init__ (line 61) | def __init__(self, c, c_cond, nhead, self_attn=True, dropout=0.0): method forward (line 69) | def forward(self, x, kv): FILE: diffusers/src/diffusers/pipelines/wuerstchen/modeling_wuerstchen_diffnext.py class WuerstchenDiffNeXt (line 27) | class WuerstchenDiffNeXt(ModelMixin, ConfigMixin): method __init__ (line 29) | def __init__( method _init_weights (line 124) | def _init_weights(self, m): method gen_r_embedding (line 146) | def gen_r_embedding(self, r, max_positions=10000): method gen_c_embeddings (line 157) | def gen_c_embeddings(self, clip): method _down_encode (line 162) | def _down_encode(self, x, r_embed, effnet, clip=None): method _up_decode (line 186) | def _up_decode(self, level_outputs, r_embed, effnet, clip=None): method forward (line 214) | def forward(self, x, r, effnet, clip=None, x_cat=None, eps=1e-3, retur... class ResBlockStageB (line 235) | class ResBlockStageB(nn.Module): method __init__ (line 236) | def __init__(self, c, c_skip=0, kernel_size=3, dropout=0.0): method forward (line 248) | def forward(self, x, x_skip=None): FILE: diffusers/src/diffusers/pipelines/wuerstchen/modeling_wuerstchen_prior.py class WuerstchenPrior (line 36) | class WuerstchenPrior(ModelMixin, ConfigMixin, UNet2DConditionLoadersMix... method __init__ (line 41) | def __init__(self, c_in=16, c=1280, c_cond=1024, c_r=64, depth=16, nhe... method attn_processors (line 67) | def attn_processors(self) -> Dict[str, AttentionProcessor]: method set_attn_processor (line 91) | def set_attn_processor(self, processor: Union[AttentionProcessor, Dict... method set_default_attn_processor (line 126) | def set_default_attn_processor(self): method _set_gradient_checkpointing (line 141) | def _set_gradient_checkpointing(self, module, value=False): method gen_r_embedding (line 144) | def gen_r_embedding(self, r, max_positions=10000): method forward (line 155) | def forward(self, x, r, c): FILE: diffusers/src/diffusers/pipelines/wuerstchen/pipeline_wuerstchen.py class WuerstchenDecoderPipeline (line 51) | class WuerstchenDecoderPipeline(DiffusionPipeline): method __init__ (line 83) | def __init__( method prepare_latents (line 103) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method encode_prompt (line 114) | def encode_prompt( method guidance_scale (line 197) | def guidance_scale(self): method do_classifier_free_guidance (line 201) | def do_classifier_free_guidance(self): method num_timesteps (line 205) | def num_timesteps(self): method __call__ (line 210) | def __call__( FILE: diffusers/src/diffusers/pipelines/wuerstchen/pipeline_wuerstchen_combined.py class WuerstchenCombinedPipeline (line 43) | class WuerstchenCombinedPipeline(DiffusionPipeline): method __init__ (line 73) | def __init__( method enable_xformers_memory_efficient_attention (line 112) | def enable_xformers_memory_efficient_attention(self, attention_op: Opt... method enable_model_cpu_offload (line 115) | def enable_model_cpu_offload(self, gpu_id: Optional[int] = None, devic... method enable_sequential_cpu_offload (line 125) | def enable_sequential_cpu_offload(self, gpu_id: Optional[int] = None, ... method progress_bar (line 135) | def progress_bar(self, iterable=None, total=None): method set_progress_bar_config (line 139) | def set_progress_bar_config(self, **kwargs): method __call__ (line 145) | def __call__( FILE: diffusers/src/diffusers/pipelines/wuerstchen/pipeline_wuerstchen_prior.py class WuerstchenPriorPipelineOutput (line 52) | class WuerstchenPriorPipelineOutput(BaseOutput): class WuerstchenPriorPipeline (line 65) | class WuerstchenPriorPipeline(DiffusionPipeline, LoraLoaderMixin): method __init__ (line 99) | def __init__( method prepare_latents (line 121) | def prepare_latents(self, shape, dtype, device, generator, latents, sc... method encode_prompt (line 132) | def encode_prompt( method check_inputs (line 227) | def check_inputs( method guidance_scale (line 269) | def guidance_scale(self): method do_classifier_free_guidance (line 273) | def do_classifier_free_guidance(self): method num_timesteps (line 277) | def num_timesteps(self): method __call__ (line 282) | def __call__( FILE: diffusers/src/diffusers/schedulers/deprecated/scheduling_karras_ve.py class KarrasVeOutput (line 29) | class KarrasVeOutput(BaseOutput): class KarrasVeScheduler (line 49) | class KarrasVeScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 82) | def __init__( method scale_model_input (line 99) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 116) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method add_noise_to_input (line 138) | def add_noise_to_input( method step (line 164) | def step( method step_correct (line 203) | def step_correct( method add_noise (line 242) | def add_noise(self, original_samples, noise, timesteps): FILE: diffusers/src/diffusers/schedulers/deprecated/scheduling_sde_vp.py class ScoreSdeVpScheduler (line 27) | class ScoreSdeVpScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 46) | def __init__(self, num_train_timesteps=2000, beta_min=0.1, beta_max=20... method set_timesteps (line 51) | def set_timesteps(self, num_inference_steps, device: Union[str, torch.... method step_pred (line 63) | def step_pred(self, score, x, t, generator=None): method __len__ (line 108) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_amused.py function gumbel_noise (line 12) | def gumbel_noise(t, generator=None): function mask_by_random_topk (line 18) | def mask_by_random_topk(mask_len, probs, temperature=1.0, generator=None): class AmusedSchedulerOutput (line 27) | class AmusedSchedulerOutput(BaseOutput): class AmusedScheduler (line 44) | class AmusedScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 50) | def __init__( method set_timesteps (line 58) | def set_timesteps( method step (line 71) | def step( method add_noise (line 140) | def add_noise(self, sample, timesteps, generator=None): FILE: diffusers/src/diffusers/schedulers/scheduling_consistency_decoder.py function betas_for_alpha_bar (line 14) | def betas_for_alpha_bar( class ConsistencyDecoderSchedulerOutput (line 59) | class ConsistencyDecoderSchedulerOutput(BaseOutput): class ConsistencyDecoderScheduler (line 72) | class ConsistencyDecoderScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 76) | def __init__( method set_timesteps (line 97) | def set_timesteps( method init_noise_sigma (line 113) | def init_noise_sigma(self): method scale_model_input (line 116) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method step (line 133) | def step( FILE: diffusers/src/diffusers/schedulers/scheduling_consistency_models.py class CMStochasticIterativeSchedulerOutput (line 31) | class CMStochasticIterativeSchedulerOutput(BaseOutput): class CMStochasticIterativeScheduler (line 44) | class CMStochasticIterativeScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 77) | def __init__( method step_index (line 105) | def step_index(self): method begin_index (line 112) | def begin_index(self): method set_begin_index (line 119) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 129) | def scale_model_input(self, sample: torch.Tensor, timestep: Union[floa... method sigma_to_t (line 154) | def sigma_to_t(self, sigmas: Union[float, np.ndarray]): method set_timesteps (line 173) | def set_timesteps( method _convert_to_karras (line 248) | def _convert_to_karras(self, ramp): method get_scalings (line 260) | def get_scalings(self, sigma): method get_scalings_for_boundary_condition (line 267) | def get_scalings_for_boundary_condition(self, sigma): method index_for_timestep (line 295) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 310) | def _init_step_index(self, timestep): method step (line 318) | def step( method add_noise (line 412) | def add_noise( method __len__ (line 445) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_ddim.py class DDIMSchedulerOutput (line 33) | class DDIMSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 51) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 95) | def rescale_zero_terminal_snr(betas): class DDIMScheduler (line 131) | class DDIMScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 185) | def __init__( method scale_model_input (line 236) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method _get_variance (line 253) | def _get_variance(self, timestep, prev_timestep): method _threshold_sample (line 264) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method set_timesteps (line 297) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method step (line 342) | def step( method add_noise (line 471) | def add_noise( method get_velocity (line 498) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 517) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_ddim_flax.py class DDIMSchedulerState (line 36) | class DDIMSchedulerState: method create (line 46) | def create( class FlaxDDIMSchedulerOutput (line 62) | class FlaxDDIMSchedulerOutput(FlaxSchedulerOutput): class FlaxDDIMScheduler (line 66) | class FlaxDDIMScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 110) | def has_state(self): method __init__ (line 114) | def __init__( method create_state (line 130) | def create_state(self, common: Optional[CommonSchedulerState] = None) ... method scale_model_input (line 154) | def scale_model_input( method set_timesteps (line 168) | def set_timesteps( method _get_variance (line 190) | def _get_variance(self, state: DDIMSchedulerState, timestep, prev_time... method step (line 202) | def step( method add_noise (line 295) | def add_noise( method get_velocity (line 304) | def get_velocity( method __len__ (line 313) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_ddim_inverse.py class DDIMSchedulerOutput (line 31) | class DDIMSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 49) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 94) | def rescale_zero_terminal_snr(betas): class DDIMInverseScheduler (line 130) | class DDIMInverseScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 177) | def __init__( method scale_model_input (line 234) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 251) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method step (line 289) | def step( method __len__ (line 373) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_ddim_parallel.py class DDIMParallelSchedulerOutput (line 33) | class DDIMParallelSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 51) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 96) | def rescale_zero_terminal_snr(betas): class DDIMParallelScheduler (line 132) | class DDIMParallelScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 192) | def __init__( method scale_model_input (line 244) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method _get_variance (line 261) | def _get_variance(self, timestep, prev_timestep=None): method _batch_get_variance (line 274) | def _batch_get_variance(self, t, prev_t): method _threshold_sample (line 286) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method set_timesteps (line 320) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method step (line 365) | def step( method batch_step_no_noise (line 487) | def batch_step_no_noise( method add_noise (line 596) | def add_noise( method get_velocity (line 623) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 642) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_ddpm.py class DDPMSchedulerOutput (line 31) | class DDPMSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 48) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 93) | def rescale_zero_terminal_snr(betas): class DDPMScheduler (line 129) | class DDPMScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 181) | def __init__( method scale_model_input (line 234) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 251) | def set_timesteps( method _get_variance (line 326) | def _get_variance(self, t, predicted_variance=None, variance_type=None): method _threshold_sample (line 366) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method step (line 399) | def step( method add_noise (line 499) | def add_noise( method get_velocity (line 525) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 544) | def __len__(self): method previous_timestep (line 547) | def previous_timestep(self, timestep): FILE: diffusers/src/diffusers/schedulers/scheduling_ddpm_flax.py class DDPMSchedulerState (line 36) | class DDPMSchedulerState: method create (line 45) | def create(cls, common: CommonSchedulerState, init_noise_sigma: jnp.nd... class FlaxDDPMSchedulerOutput (line 50) | class FlaxDDPMSchedulerOutput(FlaxSchedulerOutput): class FlaxDDPMScheduler (line 54) | class FlaxDDPMScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 92) | def has_state(self): method __init__ (line 96) | def __init__( method create_state (line 110) | def create_state(self, common: Optional[CommonSchedulerState] = None) ... method scale_model_input (line 125) | def scale_model_input( method set_timesteps (line 139) | def set_timesteps( method _get_variance (line 162) | def _get_variance(self, state: DDPMSchedulerState, t, predicted_varian... method step (line 195) | def step( method add_noise (line 284) | def add_noise( method get_velocity (line 293) | def get_velocity( method __len__ (line 302) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_ddpm_parallel.py class DDPMParallelSchedulerOutput (line 32) | class DDPMParallelSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 50) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 95) | def rescale_zero_terminal_snr(betas): class DDPMParallelScheduler (line 131) | class DDPMParallelScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 189) | def __init__( method scale_model_input (line 243) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 261) | def set_timesteps( method _get_variance (line 337) | def _get_variance(self, t, predicted_variance=None, variance_type=None): method _threshold_sample (line 378) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method step (line 411) | def step( method batch_step_no_noise (line 507) | def batch_step_no_noise( method add_noise (line 588) | def add_noise( method get_velocity (line 615) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 634) | def __len__(self): method previous_timestep (line 638) | def previous_timestep(self, timestep): FILE: diffusers/src/diffusers/schedulers/scheduling_ddpm_wuerstchen.py class DDPMWuerstchenSchedulerOutput (line 31) | class DDPMWuerstchenSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 44) | def betas_for_alpha_bar( class DDPMWuerstchenScheduler (line 88) | class DDPMWuerstchenScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 106) | def __init__( method _alpha_cumprod (line 118) | def _alpha_cumprod(self, t, device): method scale_model_input (line 128) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 142) | def set_timesteps( method step (line 164) | def step( method add_noise (line 210) | def add_noise( method __len__ (line 224) | def __len__(self): method previous_timestep (line 227) | def previous_timestep(self, timestep): FILE: diffusers/src/diffusers/schedulers/scheduling_deis_multistep.py function betas_for_alpha_bar (line 30) | def betas_for_alpha_bar( class DEISMultistepScheduler (line 74) | class DEISMultistepScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 125) | def __init__( method step_index (line 192) | def step_index(self): method begin_index (line 199) | def begin_index(self): method set_begin_index (line 206) | def set_begin_index(self, begin_index: int = 0): method set_timesteps (line 216) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method _threshold_sample (line 279) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method _sigma_to_t (line 313) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 337) | def _sigma_to_alpha_sigma_t(self, sigma): method _convert_to_karras (line 344) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method convert_model_output (line 369) | def convert_model_output( method deis_first_order_update (line 426) | def deis_first_order_update( method multistep_deis_second_order_update (line 484) | def multistep_deis_second_order_update( method multistep_deis_third_order_update (line 553) | def multistep_deis_third_order_update( method index_for_timestep (line 642) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 662) | def _init_step_index(self, timestep): method step (line 674) | def step( method scale_model_input (line 739) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method add_noise (line 755) | def add_noise( method __len__ (line 789) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_dpmsolver_multistep.py function betas_for_alpha_bar (line 30) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 75) | def rescale_zero_terminal_snr(betas): class DPMSolverMultistepScheduler (line 111) | class DPMSolverMultistepScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 192) | def __init__( method step_index (line 283) | def step_index(self): method begin_index (line 290) | def begin_index(self): method set_begin_index (line 296) | def set_begin_index(self, begin_index: int = 0): method set_timesteps (line 306) | def set_timesteps( method _threshold_sample (line 411) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method _sigma_to_t (line 445) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 468) | def _sigma_to_alpha_sigma_t(self, sigma): method _convert_to_karras (line 475) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method _convert_to_lu (line 500) | def _convert_to_lu(self, in_lambdas: torch.Tensor, num_inference_steps... method convert_model_output (line 513) | def convert_model_output( method dpm_solver_first_order_update (line 612) | def dpm_solver_first_order_update( method multistep_dpm_solver_second_order_update (line 681) | def multistep_dpm_solver_second_order_update( method multistep_dpm_solver_third_order_update (line 804) | def multistep_dpm_solver_third_order_update( method index_for_timestep (line 889) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 908) | def _init_step_index(self, timestep): method step (line 920) | def step( method scale_model_input (line 1009) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method add_noise (line 1024) | def add_noise( method __len__ (line 1058) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_dpmsolver_multistep_flax.py class DPMSolverMultistepSchedulerState (line 35) | class DPMSolverMultistepSchedulerState: method create (line 53) | def create( class FlaxDPMSolverMultistepSchedulerOutput (line 73) | class FlaxDPMSolverMultistepSchedulerOutput(FlaxSchedulerOutput): class FlaxDPMSolverMultistepScheduler (line 77) | class FlaxDPMSolverMultistepScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 150) | def has_state(self): method __init__ (line 154) | def __init__( method create_state (line 174) | def create_state(self, common: Optional[CommonSchedulerState] = None) ... method set_timesteps (line 203) | def set_timesteps( method convert_model_output (line 257) | def convert_model_output( method dpm_solver_first_order_update (line 327) | def dpm_solver_first_order_update( method multistep_dpm_solver_second_order_update (line 362) | def multistep_dpm_solver_second_order_update( method multistep_dpm_solver_third_order_update (line 422) | def multistep_dpm_solver_third_order_update( method step (line 478) | def step( method scale_model_input (line 615) | def scale_model_input( method add_noise (line 633) | def add_noise( method __len__ (line 642) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_dpmsolver_multistep_inverse.py function betas_for_alpha_bar (line 30) | def betas_for_alpha_bar( class DPMSolverMultistepInverseScheduler (line 74) | class DPMSolverMultistepInverseScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 144) | def __init__( method step_index (line 218) | def step_index(self): method set_timesteps (line 224) | def set_timesteps(self, num_inference_steps: int = None, device: Union... method _threshold_sample (line 298) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method _sigma_to_t (line 332) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 356) | def _sigma_to_alpha_sigma_t(self, sigma): method _convert_to_karras (line 363) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method convert_model_output (line 389) | def convert_model_output( method dpm_solver_first_order_update (line 489) | def dpm_solver_first_order_update( method multistep_dpm_solver_second_order_update (line 559) | def multistep_dpm_solver_second_order_update( method multistep_dpm_solver_third_order_update (line 683) | def multistep_dpm_solver_third_order_update( method _init_step_index (line 768) | def _init_step_index(self, timestep): method step (line 787) | def step( method scale_model_input (line 870) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method add_noise (line 885) | def add_noise( method __len__ (line 920) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_dpmsolver_sde.py class BatchedBrownianTree (line 26) | class BatchedBrownianTree: method __init__ (line 29) | def __init__(self, x, t0, t1, seed=None, **kwargs): method sort (line 44) | def sort(a, b): method __call__ (line 47) | def __call__(self, t0, t1): class BrownianTreeNoiseSampler (line 53) | class BrownianTreeNoiseSampler: method __init__ (line 68) | def __init__(self, x, sigma_min, sigma_max, seed=None, transform=lambd... method __call__ (line 73) | def __call__(self, sigma, sigma_next): function betas_for_alpha_bar (line 79) | def betas_for_alpha_bar( class DPMSolverSDEScheduler (line 123) | class DPMSolverSDEScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 163) | def __init__( method index_for_timestep (line 202) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 217) | def _init_step_index(self, timestep): method init_noise_sigma (line 226) | def init_noise_sigma(self): method step_index (line 234) | def step_index(self): method begin_index (line 241) | def begin_index(self): method set_begin_index (line 248) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 258) | def scale_model_input( method set_timesteps (line 285) | def set_timesteps( method _second_order_timesteps (line 358) | def _second_order_timesteps(self, sigmas, log_sigmas): method _sigma_to_t (line 374) | def _sigma_to_t(self, sigma, log_sigmas): method _convert_to_karras (line 398) | def _convert_to_karras(self, in_sigmas: torch.Tensor) -> torch.Tensor: method state_in_first_order (line 412) | def state_in_first_order(self): method step (line 415) | def step( method add_noise (line 527) | def add_noise( method __len__ (line 560) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_dpmsolver_singlestep.py function betas_for_alpha_bar (line 32) | def betas_for_alpha_bar( class DPMSolverSinglestepScheduler (line 76) | class DPMSolverSinglestepScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 140) | def __init__( method get_order_list (line 216) | def get_order_list(self, num_inference_steps: int) -> List[int]: method step_index (line 253) | def step_index(self): method begin_index (line 260) | def begin_index(self): method set_begin_index (line 267) | def set_begin_index(self, begin_index: int = 0): method set_timesteps (line 277) | def set_timesteps( method _threshold_sample (line 364) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method _sigma_to_t (line 398) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 422) | def _sigma_to_alpha_sigma_t(self, sigma): method _convert_to_karras (line 429) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method convert_model_output (line 454) | def convert_model_output( method dpm_solver_first_order_update (line 543) | def dpm_solver_first_order_update( method singlestep_dpm_solver_second_order_update (line 599) | def singlestep_dpm_solver_second_order_update( method singlestep_dpm_solver_third_order_update (line 693) | def singlestep_dpm_solver_third_order_update( method singlestep_dpm_solver_update (line 797) | def singlestep_dpm_solver_update( method index_for_timestep (line 860) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 880) | def _init_step_index(self, timestep): method step (line 892) | def step( method scale_model_input (line 953) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method add_noise (line 969) | def add_noise( method __len__ (line 1003) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_edm_dpmsolver_multistep.py class EDMDPMSolverMultistepScheduler (line 28) | class EDMDPMSolverMultistepScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 92) | def __init__( method init_noise_sigma (line 148) | def init_noise_sigma(self): method step_index (line 153) | def step_index(self): method begin_index (line 160) | def begin_index(self): method set_begin_index (line 167) | def set_begin_index(self, begin_index: int = 0): method precondition_inputs (line 178) | def precondition_inputs(self, sample, sigma): method precondition_noise (line 184) | def precondition_noise(self, sigma): method precondition_outputs (line 193) | def precondition_outputs(self, sample, model_output, sigma): method scale_model_input (line 209) | def scale_model_input(self, sample: torch.Tensor, timestep: Union[floa... method set_timesteps (line 233) | def set_timesteps(self, num_inference_steps: int = None, device: Union... method _compute_karras_sigmas (line 277) | def _compute_karras_sigmas(self, ramp, sigma_min=None, sigma_max=None)... method _compute_exponential_sigmas (line 289) | def _compute_exponential_sigmas(self, ramp, sigma_min=None, sigma_max=... method _threshold_sample (line 300) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method _sigma_to_t (line 334) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 357) | def _sigma_to_alpha_sigma_t(self, sigma): method convert_model_output (line 363) | def convert_model_output( method dpm_solver_first_order_update (line 398) | def dpm_solver_first_order_update( method multistep_dpm_solver_second_order_update (line 436) | def multistep_dpm_solver_second_order_update( method multistep_dpm_solver_third_order_update (line 507) | def multistep_dpm_solver_third_order_update( method index_for_timestep (line 562) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 582) | def _init_step_index(self, timestep): method step (line 594) | def step( method add_noise (line 673) | def add_noise( method __len__ (line 706) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_edm_euler.py class EDMEulerSchedulerOutput (line 32) | class EDMEulerSchedulerOutput(BaseOutput): class EDMEulerScheduler (line 49) | class EDMEulerScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 86) | def __init__( method init_noise_sigma (line 119) | def init_noise_sigma(self): method step_index (line 124) | def step_index(self): method begin_index (line 131) | def begin_index(self): method set_begin_index (line 138) | def set_begin_index(self, begin_index: int = 0): method precondition_inputs (line 148) | def precondition_inputs(self, sample, sigma): method precondition_noise (line 153) | def precondition_noise(self, sigma): method precondition_outputs (line 161) | def precondition_outputs(self, sample, model_output, sigma): method scale_model_input (line 176) | def scale_model_input(self, sample: torch.Tensor, timestep: Union[floa... method set_timesteps (line 200) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method _compute_karras_sigmas (line 227) | def _compute_karras_sigmas(self, ramp, sigma_min=None, sigma_max=None)... method _compute_exponential_sigmas (line 238) | def _compute_exponential_sigmas(self, ramp, sigma_min=None, sigma_max=... method index_for_timestep (line 249) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 264) | def _init_step_index(self, timestep): method step (line 272) | def step( method add_noise (line 368) | def add_noise( method __len__ (line 401) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_euler_ancestral_discrete.py class EulerAncestralDiscreteSchedulerOutput (line 33) | class EulerAncestralDiscreteSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 51) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 96) | def rescale_zero_terminal_snr(betas): class EulerAncestralDiscreteScheduler (line 132) | class EulerAncestralDiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 170) | def __init__( method init_noise_sigma (line 221) | def init_noise_sigma(self): method step_index (line 229) | def step_index(self): method begin_index (line 236) | def begin_index(self): method set_begin_index (line 243) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 253) | def scale_model_input(self, sample: torch.Tensor, timestep: Union[floa... method set_timesteps (line 277) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method index_for_timestep (line 322) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 337) | def _init_step_index(self, timestep): method step (line 345) | def step( method add_noise (line 445) | def add_noise( method __len__ (line 478) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_euler_discrete.py class EulerDiscreteSchedulerOutput (line 33) | class EulerDiscreteSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 51) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 96) | def rescale_zero_terminal_snr(betas): class EulerDiscreteScheduler (line 132) | class EulerDiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 179) | def __init__( method init_noise_sigma (line 244) | def init_noise_sigma(self): method step_index (line 253) | def step_index(self): method begin_index (line 260) | def begin_index(self): method set_begin_index (line 267) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 277) | def scale_model_input(self, sample: torch.Tensor, timestep: Union[floa... method set_timesteps (line 301) | def set_timesteps( method _sigma_to_t (line 422) | def _sigma_to_t(self, sigma, log_sigmas): method _convert_to_karras (line 446) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method index_for_timestep (line 471) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 485) | def _init_step_index(self, timestep): method step (line 493) | def step( method add_noise (line 601) | def add_noise( method get_velocity (line 634) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 671) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_euler_discrete_flax.py class EulerDiscreteSchedulerState (line 32) | class EulerDiscreteSchedulerState: method create (line 42) | def create( class FlaxEulerDiscreteSchedulerOutput (line 49) | class FlaxEulerDiscreteSchedulerOutput(FlaxSchedulerOutput): class FlaxEulerDiscreteScheduler (line 53) | class FlaxEulerDiscreteScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 87) | def has_state(self): method __init__ (line 91) | def __init__( method create_state (line 104) | def create_state(self, common: Optional[CommonSchedulerState] = None) ... method scale_model_input (line 126) | def scale_model_input(self, state: EulerDiscreteSchedulerState, sample... method set_timesteps (line 148) | def set_timesteps( method step (line 189) | def step( method add_noise (line 250) | def add_noise( method __len__ (line 264) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_flow_match_euler_discrete.py class FlowMatchEulerDiscreteSchedulerOutput (line 31) | class FlowMatchEulerDiscreteSchedulerOutput(BaseOutput): class FlowMatchEulerDiscreteScheduler (line 44) | class FlowMatchEulerDiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 65) | def __init__( method step_index (line 86) | def step_index(self): method begin_index (line 93) | def begin_index(self): method set_begin_index (line 100) | def set_begin_index(self, begin_index: int = 0): method scale_noise (line 110) | def scale_noise( method _sigma_to_t (line 137) | def _sigma_to_t(self, sigma): method set_timesteps (line 140) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method index_for_timestep (line 167) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 181) | def _init_step_index(self, timestep): method step (line 189) | def step( method __len__ (line 286) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_heun_discrete.py function betas_for_alpha_bar (line 26) | def betas_for_alpha_bar( class HeunDiscreteScheduler (line 70) | class HeunDiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 111) | def __init__( method index_for_timestep (line 152) | def index_for_timestep(self, timestep, schedule_timesteps=None): method init_noise_sigma (line 167) | def init_noise_sigma(self): method step_index (line 175) | def step_index(self): method begin_index (line 182) | def begin_index(self): method set_begin_index (line 189) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 199) | def scale_model_input( method set_timesteps (line 225) | def set_timesteps( method _sigma_to_t (line 308) | def _sigma_to_t(self, sigma, log_sigmas): method _convert_to_karras (line 332) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method state_in_first_order (line 358) | def state_in_first_order(self): method _init_step_index (line 362) | def _init_step_index(self, timestep): method step (line 370) | def step( method add_noise (line 470) | def add_noise( method __len__ (line 503) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_ipndm.py class IPNDMScheduler (line 25) | class IPNDMScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 42) | def __init__( method step_index (line 62) | def step_index(self): method begin_index (line 69) | def begin_index(self): method set_begin_index (line 76) | def set_begin_index(self, begin_index: int = 0): method set_timesteps (line 86) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method index_for_timestep (line 115) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 130) | def _init_step_index(self, timestep): method step (line 138) | def step( method scale_model_input (line 196) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method _get_prev_sample (line 211) | def _get_prev_sample(self, sample, timestep_index, prev_timestep_index... method __len__ (line 223) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_k_dpm_2_ancestral_discrete.py function betas_for_alpha_bar (line 27) | def betas_for_alpha_bar( class KDPM2AncestralDiscreteScheduler (line 71) | class KDPM2AncestralDiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 109) | def __init__( method init_noise_sigma (line 144) | def init_noise_sigma(self): method step_index (line 152) | def step_index(self): method begin_index (line 159) | def begin_index(self): method set_begin_index (line 166) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 176) | def scale_model_input( method set_timesteps (line 206) | def set_timesteps( method _sigma_to_t (line 300) | def _sigma_to_t(self, sigma, log_sigmas): method _convert_to_karras (line 324) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method state_in_first_order (line 350) | def state_in_first_order(self): method index_for_timestep (line 354) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 369) | def _init_step_index(self, timestep): method step (line 377) | def step( method add_noise (line 478) | def add_noise( method __len__ (line 511) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_k_dpm_2_discrete.py function betas_for_alpha_bar (line 26) | def betas_for_alpha_bar( class KDPM2DiscreteScheduler (line 70) | class KDPM2DiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 108) | def __init__( method init_noise_sigma (line 144) | def init_noise_sigma(self): method step_index (line 152) | def step_index(self): method begin_index (line 159) | def begin_index(self): method set_begin_index (line 166) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 176) | def scale_model_input( method set_timesteps (line 206) | def set_timesteps( method state_in_first_order (line 285) | def state_in_first_order(self): method index_for_timestep (line 289) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 304) | def _init_step_index(self, timestep): method _sigma_to_t (line 313) | def _sigma_to_t(self, sigma, log_sigmas): method _convert_to_karras (line 337) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method step (line 362) | def step( method add_noise (line 453) | def add_noise( method __len__ (line 486) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_karras_ve_flax.py class KarrasVeSchedulerState (line 30) | class KarrasVeSchedulerState: method create (line 37) | def create(cls): class FlaxKarrasVeOutput (line 42) | class FlaxKarrasVeOutput(BaseOutput): class FlaxKarrasVeScheduler (line 60) | class FlaxKarrasVeScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 92) | def has_state(self): method __init__ (line 96) | def __init__( method create_state (line 107) | def create_state(self): method set_timesteps (line 110) | def set_timesteps( method add_noise_to_input (line 138) | def add_noise_to_input( method step (line 164) | def step( method step_correct (line 200) | def step_correct( method add_noise (line 237) | def add_noise(self, state: KarrasVeSchedulerState, original_samples, n... FILE: diffusers/src/diffusers/schedulers/scheduling_lcm.py class LCMSchedulerOutput (line 35) | class LCMSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 53) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 98) | def rescale_zero_terminal_snr(betas: torch.Tensor) -> torch.Tensor: class LCMScheduler (line 134) | class LCMScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 196) | def __init__( method index_for_timestep (line 254) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 269) | def _init_step_index(self, timestep): method step_index (line 278) | def step_index(self): method begin_index (line 282) | def begin_index(self): method set_begin_index (line 289) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 299) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method _threshold_sample (line 316) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method set_timesteps (line 349) | def set_timesteps( method get_scalings_for_boundary_condition_discrete (line 490) | def get_scalings_for_boundary_condition_discrete(self, timestep): method step (line 498) | def step( method add_noise (line 595) | def add_noise( method get_velocity (line 622) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 641) | def __len__(self): method previous_timestep (line 645) | def previous_timestep(self, timestep): FILE: diffusers/src/diffusers/schedulers/scheduling_lms_discrete.py class LMSDiscreteSchedulerOutput (line 30) | class LMSDiscreteSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 48) | def betas_for_alpha_bar( class LMSDiscreteScheduler (line 92) | class LMSDiscreteScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 129) | def __init__( method init_noise_sigma (line 173) | def init_noise_sigma(self): method step_index (line 181) | def step_index(self): method begin_index (line 188) | def begin_index(self): method set_begin_index (line 195) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 205) | def scale_model_input(self, sample: torch.Tensor, timestep: Union[floa... method get_lms_coefficient (line 229) | def get_lms_coefficient(self, order, t, current_order): method set_timesteps (line 251) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method index_for_timestep (line 304) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 319) | def _init_step_index(self, timestep): method _sigma_to_t (line 328) | def _sigma_to_t(self, sigma, log_sigmas): method _convert_to_karras (line 352) | def _convert_to_karras(self, in_sigmas: torch.Tensor) -> torch.Tensor: method step (line 365) | def step( method add_noise (line 443) | def add_noise( method __len__ (line 476) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_lms_discrete_flax.py class LMSDiscreteSchedulerState (line 33) | class LMSDiscreteSchedulerState: method create (line 46) | def create( class FlaxLMSSchedulerOutput (line 53) | class FlaxLMSSchedulerOutput(FlaxSchedulerOutput): class FlaxLMSDiscreteScheduler (line 57) | class FlaxLMSDiscreteScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 90) | def has_state(self): method __init__ (line 94) | def __init__( method create_state (line 106) | def create_state(self, common: Optional[CommonSchedulerState] = None) ... method scale_model_input (line 123) | def scale_model_input(self, state: LMSDiscreteSchedulerState, sample: ... method get_lms_coefficient (line 145) | def get_lms_coefficient(self, state: LMSDiscreteSchedulerState, order,... method set_timesteps (line 167) | def set_timesteps( method step (line 203) | def step( method add_noise (line 268) | def add_noise( method __len__ (line 282) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_pndm.py function betas_for_alpha_bar (line 28) | def betas_for_alpha_bar( class PNDMScheduler (line 72) | class PNDMScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 114) | def __init__( method set_timesteps (line 166) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method step (line 226) | def step( method step_prk (line 259) | def step_prk( method step_plms (line 319) | def step_plms( method scale_model_input (line 390) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method _get_prev_sample (line 405) | def _get_prev_sample(self, sample, timestep, prev_timestep, model_outp... method add_noise (line 449) | def add_noise( method __len__ (line 475) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_pndm_flax.py class PNDMSchedulerState (line 35) | class PNDMSchedulerState: method create (line 53) | def create( class FlaxPNDMSchedulerOutput (line 69) | class FlaxPNDMSchedulerOutput(FlaxSchedulerOutput): class FlaxPNDMScheduler (line 73) | class FlaxPNDMScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 117) | def has_state(self): method __init__ (line 121) | def __init__( method create_state (line 141) | def create_state(self, common: Optional[CommonSchedulerState] = None) ... method set_timesteps (line 165) | def set_timesteps(self, state: PNDMSchedulerState, num_inference_steps... method scale_model_input (line 220) | def scale_model_input( method step (line 237) | def step( method step_prk (line 292) | def step_prk( method step_plms (line 360) | def step_plms( method _get_prev_sample (line 454) | def _get_prev_sample(self, state: PNDMSchedulerState, sample, timestep... method add_noise (line 499) | def add_noise( method __len__ (line 508) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_repaint.py class RePaintSchedulerOutput (line 29) | class RePaintSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 47) | def betas_for_alpha_bar( class RePaintScheduler (line 91) | class RePaintScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 121) | def __init__( method scale_model_input (line 163) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 180) | def set_timesteps( method _get_variance (line 227) | def _get_variance(self, t): method step (line 246) | def step( method undo_step (line 335) | def undo_step(self, sample, timestep, generator=None): method add_noise (line 352) | def add_noise( method __len__ (line 360) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_sasolver.py function betas_for_alpha_bar (line 31) | def betas_for_alpha_bar( class SASolverScheduler (line 75) | class SASolverScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 142) | def __init__( method step_index (line 218) | def step_index(self): method begin_index (line 225) | def begin_index(self): method set_begin_index (line 232) | def set_begin_index(self, begin_index: int = 0): method set_timesteps (line 242) | def set_timesteps(self, num_inference_steps: int = None, device: Union... method _threshold_sample (line 308) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method _sigma_to_t (line 342) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 366) | def _sigma_to_alpha_sigma_t(self, sigma): method _convert_to_karras (line 373) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method convert_model_output (line 398) | def convert_model_output( method get_coefficients_exponential_negative (line 490) | def get_coefficients_exponential_negative(self, order, interval_start,... method get_coefficients_exponential_positive (line 514) | def get_coefficients_exponential_positive(self, order, interval_start,... method lagrange_polynomial_coefficient (line 558) | def lagrange_polynomial_coefficient(self, order, lambda_list): method get_coefficients_fn (line 667) | def get_coefficients_fn(self, order, interval_start, interval_end, lam... method stochastic_adams_bashforth_update (line 687) | def stochastic_adams_bashforth_update( method stochastic_adams_moulton_update (line 814) | def stochastic_adams_moulton_update( method index_for_timestep (line 948) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 968) | def _init_step_index(self, timestep): method step (line 980) | def step( method scale_model_input (line 1082) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method add_noise (line 1098) | def add_noise( method __len__ (line 1124) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_sde_ve.py class SdeVeOutput (line 30) | class SdeVeOutput(BaseOutput): class ScoreSdeVeScheduler (line 46) | class ScoreSdeVeScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 72) | def __init__( method scale_model_input (line 89) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 106) | def set_timesteps( method set_sigmas (line 125) | def set_sigmas( method get_adjacent_sigma (line 153) | def get_adjacent_sigma(self, timesteps, t): method step_pred (line 160) | def step_pred( method step_correct (line 228) | def step_correct( method add_noise (line 283) | def add_noise( method __len__ (line 300) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_sde_ve_flax.py class ScoreSdeVeSchedulerState (line 30) | class ScoreSdeVeSchedulerState: method create (line 37) | def create(cls): class FlaxSdeVeOutput (line 42) | class FlaxSdeVeOutput(FlaxSchedulerOutput): class FlaxScoreSdeVeScheduler (line 60) | class FlaxScoreSdeVeScheduler(FlaxSchedulerMixin, ConfigMixin): method has_state (line 85) | def has_state(self): method __init__ (line 89) | def __init__( method create_state (line 100) | def create_state(self): method set_timesteps (line 110) | def set_timesteps( method set_sigmas (line 129) | def set_sigmas( method get_adjacent_sigma (line 164) | def get_adjacent_sigma(self, state, timesteps, t): method step_pred (line 167) | def step_pred( method step_correct (line 227) | def step_correct( method __len__ (line 279) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_tcd.py class TCDSchedulerOutput (line 35) | class TCDSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 52) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 97) | def rescale_zero_terminal_snr(betas: torch.Tensor) -> torch.Tensor: class TCDScheduler (line 133) | class TCDScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 197) | def __init__( method index_for_timestep (line 255) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 270) | def _init_step_index(self, timestep): method step_index (line 279) | def step_index(self): method begin_index (line 283) | def begin_index(self): method set_begin_index (line 290) | def set_begin_index(self, begin_index: int = 0): method scale_model_input (line 300) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method _get_variance (line 318) | def _get_variance(self, timestep, prev_timestep): method _threshold_sample (line 329) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method set_timesteps (line 362) | def set_timesteps( method step (line 525) | def step( method add_noise (line 632) | def add_noise( method get_velocity (line 659) | def get_velocity(self, sample: torch.Tensor, noise: torch.Tensor, time... method __len__ (line 678) | def __len__(self): method previous_timestep (line 682) | def previous_timestep(self, timestep): FILE: diffusers/src/diffusers/schedulers/scheduling_unclip.py class UnCLIPSchedulerOutput (line 30) | class UnCLIPSchedulerOutput(BaseOutput): function betas_for_alpha_bar (line 48) | def betas_for_alpha_bar( class UnCLIPScheduler (line 92) | class UnCLIPScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 122) | def __init__( method scale_model_input (line 149) | def scale_model_input(self, sample: torch.Tensor, timestep: Optional[i... method set_timesteps (line 163) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method _get_variance (line 180) | def _get_variance(self, t, prev_timestep=None, predicted_variance=None... method step (line 216) | def step( method add_noise (line 328) | def add_noise( FILE: diffusers/src/diffusers/schedulers/scheduling_unipc_multistep.py function betas_for_alpha_bar (line 30) | def betas_for_alpha_bar( function rescale_zero_terminal_snr (line 75) | def rescale_zero_terminal_snr(betas): class UniPCMultistepScheduler (line 111) | class UniPCMultistepScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 180) | def __init__( method step_index (line 258) | def step_index(self): method begin_index (line 265) | def begin_index(self): method set_begin_index (line 272) | def set_begin_index(self, begin_index: int = 0): method set_timesteps (line 282) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method _threshold_sample (line 363) | def _threshold_sample(self, sample: torch.Tensor) -> torch.Tensor: method _sigma_to_t (line 397) | def _sigma_to_t(self, sigma, log_sigmas): method _sigma_to_alpha_sigma_t (line 421) | def _sigma_to_alpha_sigma_t(self, sigma): method _convert_to_karras (line 428) | def _convert_to_karras(self, in_sigmas: torch.Tensor, num_inference_st... method convert_model_output (line 453) | def convert_model_output( method multistep_uni_p_bh_update (line 523) | def multistep_uni_p_bh_update( method multistep_uni_c_bh_update (line 652) | def multistep_uni_c_bh_update( method index_for_timestep (line 790) | def index_for_timestep(self, timestep, schedule_timesteps=None): method _init_step_index (line 810) | def _init_step_index(self, timestep): method step (line 822) | def step( method scale_model_input (line 903) | def scale_model_input(self, sample: torch.Tensor, *args, **kwargs) -> ... method add_noise (line 919) | def add_noise( method __len__ (line 953) | def __len__(self): FILE: diffusers/src/diffusers/schedulers/scheduling_utils.py class KarrasDiffusionSchedulers (line 33) | class KarrasDiffusionSchedulers(Enum): class SchedulerOutput (line 61) | class SchedulerOutput(BaseOutput): class SchedulerMixin (line 74) | class SchedulerMixin(PushToHubMixin): method from_pretrained (line 96) | def from_pretrained( method save_pretrained (line 161) | def save_pretrained(self, save_directory: Union[str, os.PathLike], pus... method compatibles (line 179) | def compatibles(self): method _get_compatibles (line 189) | def _get_compatibles(cls): FILE: diffusers/src/diffusers/schedulers/scheduling_utils_flax.py class FlaxKarrasDiffusionSchedulers (line 35) | class FlaxKarrasDiffusionSchedulers(Enum): class FlaxSchedulerOutput (line 45) | class FlaxSchedulerOutput(BaseOutput): class FlaxSchedulerMixin (line 58) | class FlaxSchedulerMixin(PushToHubMixin): method from_pretrained (line 75) | def from_pretrained( method save_pretrained (line 154) | def save_pretrained(self, save_directory: Union[str, os.PathLike], pus... method compatibles (line 172) | def compatibles(self): method _get_compatibles (line 182) | def _get_compatibles(cls): function broadcast_to_shape_from_left (line 191) | def broadcast_to_shape_from_left(x: jnp.ndarray, shape: Tuple[int]) -> j... function betas_for_alpha_bar (line 196) | def betas_for_alpha_bar(num_diffusion_timesteps: int, max_beta=0.999, dt... class CommonSchedulerState (line 226) | class CommonSchedulerState: method create (line 232) | def create(cls, scheduler): function get_sqrt_alpha_prod (line 266) | def get_sqrt_alpha_prod( function add_noise_common (line 282) | def add_noise_common( function get_velocity_common (line 290) | def get_velocity_common(state: CommonSchedulerState, sample: jnp.ndarray... FILE: diffusers/src/diffusers/schedulers/scheduling_vq_diffusion.py class VQDiffusionSchedulerOutput (line 28) | class VQDiffusionSchedulerOutput(BaseOutput): function index_to_log_onehot (line 41) | def index_to_log_onehot(x: torch.LongTensor, num_classes: int) -> torch.... function gumbel_noised (line 62) | def gumbel_noised(logits: torch.Tensor, generator: Optional[torch.Genera... function alpha_schedules (line 72) | def alpha_schedules(num_diffusion_timesteps: int, alpha_cum_start=0.9999... function gamma_schedules (line 88) | def gamma_schedules(num_diffusion_timesteps: int, gamma_cum_start=0.0000... class VQDiffusionScheduler (line 106) | class VQDiffusionScheduler(SchedulerMixin, ConfigMixin): method __init__ (line 132) | def __init__( method set_timesteps (line 178) | def set_timesteps(self, num_inference_steps: int, device: Union[str, t... method step (line 200) | def step( method q_posterior (line 245) | def q_posterior(self, log_p_x_0, x_t, t): method log_Q_t_transitioning_to_known_class (line 356) | def log_Q_t_transitioning_to_known_class( method apply_cumulative_transitions (line 455) | def apply_cumulative_transitions(self, q, t): FILE: diffusers/src/diffusers/training_utils.py function set_seed (line 35) | def set_seed(seed: int): function compute_snr (line 51) | def compute_snr(noise_scheduler, timesteps): function resolve_interpolation_mode (line 77) | def resolve_interpolation_mode(interpolation_type: str): function compute_dream_and_update_latents (line 121) | def compute_dream_and_update_latents( function unet_lora_state_dict (line 175) | def unet_lora_state_dict(unet: UNet2DConditionModel) -> Dict[str, torch.... function cast_training_params (line 194) | def cast_training_params(model: Union[torch.nn.Module, List[torch.nn.Mod... function _set_state_dict_into_text_encoder (line 204) | def _set_state_dict_into_text_encoder( class EMAModel (line 224) | class EMAModel: method __init__ (line 229) | def __init__( method from_pretrained (line 311) | def from_pretrained(cls, path, model_cls) -> "EMAModel": method save_pretrained (line 320) | def save_pretrained(self, path, **kwargs): method get_decay (line 335) | def get_decay(self, optimization_step: int) -> float: method step (line 355) | def step(self, parameters: Iterable[torch.nn.Parameter]): method copy_to (line 392) | def copy_to(self, parameters: Iterable[torch.nn.Parameter]) -> None: method to (line 405) | def to(self, device=None, dtype=None) -> None: method state_dict (line 417) | def state_dict(self) -> dict: method store (line 436) | def store(self, parameters: Iterable[torch.nn.Parameter]) -> None: method restore (line 445) | def restore(self, parameters: Iterable[torch.nn.Parameter]) -> None: method load_state_dict (line 463) | def load_state_dict(self, state_dict: dict) -> None: FILE: diffusers/src/diffusers/utils/__init__.py function check_min_version (line 124) | def check_min_version(min_version): FILE: diffusers/src/diffusers/utils/accelerate_utils.py function apply_forward_hook (line 27) | def apply_forward_hook(method): FILE: diffusers/src/diffusers/utils/deprecation_utils.py function deprecate (line 8) | def deprecate(*args, take_from: Optional[Union[Dict, Any]] = None, stand... FILE: diffusers/src/diffusers/utils/doc_utils.py function replace_example_docstring (line 21) | def replace_example_docstring(example_docstring): FILE: diffusers/src/diffusers/utils/dummy_flax_and_transformers_objects.py class FlaxStableDiffusionControlNetPipeline (line 5) | class FlaxStableDiffusionControlNetPipeline(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class FlaxStableDiffusionImg2ImgPipeline (line 20) | class FlaxStableDiffusionImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): class FlaxStableDiffusionInpaintPipeline (line 35) | class FlaxStableDiffusionInpaintPipeline(metaclass=DummyObject): method __init__ (line 38) | def __init__(self, *args, **kwargs): method from_config (line 42) | def from_config(cls, *args, **kwargs): method from_pretrained (line 46) | def from_pretrained(cls, *args, **kwargs): class FlaxStableDiffusionPipeline (line 50) | class FlaxStableDiffusionPipeline(metaclass=DummyObject): method __init__ (line 53) | def __init__(self, *args, **kwargs): method from_config (line 57) | def from_config(cls, *args, **kwargs): method from_pretrained (line 61) | def from_pretrained(cls, *args, **kwargs): class FlaxStableDiffusionXLPipeline (line 65) | class FlaxStableDiffusionXLPipeline(metaclass=DummyObject): method __init__ (line 68) | def __init__(self, *args, **kwargs): method from_config (line 72) | def from_config(cls, *args, **kwargs): method from_pretrained (line 76) | def from_pretrained(cls, *args, **kwargs): FILE: diffusers/src/diffusers/utils/dummy_flax_objects.py class FlaxControlNetModel (line 5) | class FlaxControlNetModel(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class FlaxModelMixin (line 20) | class FlaxModelMixin(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): class FlaxUNet2DConditionModel (line 35) | class FlaxUNet2DConditionModel(metaclass=DummyObject): method __init__ (line 38) | def __init__(self, *args, **kwargs): method from_config (line 42) | def from_config(cls, *args, **kwargs): method from_pretrained (line 46) | def from_pretrained(cls, *args, **kwargs): class FlaxAutoencoderKL (line 50) | class FlaxAutoencoderKL(metaclass=DummyObject): method __init__ (line 53) | def __init__(self, *args, **kwargs): method from_config (line 57) | def from_config(cls, *args, **kwargs): method from_pretrained (line 61) | def from_pretrained(cls, *args, **kwargs): class FlaxDiffusionPipeline (line 65) | class FlaxDiffusionPipeline(metaclass=DummyObject): method __init__ (line 68) | def __init__(self, *args, **kwargs): method from_config (line 72) | def from_config(cls, *args, **kwargs): method from_pretrained (line 76) | def from_pretrained(cls, *args, **kwargs): class FlaxDDIMScheduler (line 80) | class FlaxDDIMScheduler(metaclass=DummyObject): method __init__ (line 83) | def __init__(self, *args, **kwargs): method from_config (line 87) | def from_config(cls, *args, **kwargs): method from_pretrained (line 91) | def from_pretrained(cls, *args, **kwargs): class FlaxDDPMScheduler (line 95) | class FlaxDDPMScheduler(metaclass=DummyObject): method __init__ (line 98) | def __init__(self, *args, **kwargs): method from_config (line 102) | def from_config(cls, *args, **kwargs): method from_pretrained (line 106) | def from_pretrained(cls, *args, **kwargs): class FlaxDPMSolverMultistepScheduler (line 110) | class FlaxDPMSolverMultistepScheduler(metaclass=DummyObject): method __init__ (line 113) | def __init__(self, *args, **kwargs): method from_config (line 117) | def from_config(cls, *args, **kwargs): method from_pretrained (line 121) | def from_pretrained(cls, *args, **kwargs): class FlaxEulerDiscreteScheduler (line 125) | class FlaxEulerDiscreteScheduler(metaclass=DummyObject): method __init__ (line 128) | def __init__(self, *args, **kwargs): method from_config (line 132) | def from_config(cls, *args, **kwargs): method from_pretrained (line 136) | def from_pretrained(cls, *args, **kwargs): class FlaxKarrasVeScheduler (line 140) | class FlaxKarrasVeScheduler(metaclass=DummyObject): method __init__ (line 143) | def __init__(self, *args, **kwargs): method from_config (line 147) | def from_config(cls, *args, **kwargs): method from_pretrained (line 151) | def from_pretrained(cls, *args, **kwargs): class FlaxLMSDiscreteScheduler (line 155) | class FlaxLMSDiscreteScheduler(metaclass=DummyObject): method __init__ (line 158) | def __init__(self, *args, **kwargs): method from_config (line 162) | def from_config(cls, *args, **kwargs): method from_pretrained (line 166) | def from_pretrained(cls, *args, **kwargs): class FlaxPNDMScheduler (line 170) | class FlaxPNDMScheduler(metaclass=DummyObject): method __init__ (line 173) | def __init__(self, *args, **kwargs): method from_config (line 177) | def from_config(cls, *args, **kwargs): method from_pretrained (line 181) | def from_pretrained(cls, *args, **kwargs): class FlaxSchedulerMixin (line 185) | class FlaxSchedulerMixin(metaclass=DummyObject): method __init__ (line 188) | def __init__(self, *args, **kwargs): method from_config (line 192) | def from_config(cls, *args, **kwargs): method from_pretrained (line 196) | def from_pretrained(cls, *args, **kwargs): class FlaxScoreSdeVeScheduler (line 200) | class FlaxScoreSdeVeScheduler(metaclass=DummyObject): method __init__ (line 203) | def __init__(self, *args, **kwargs): method from_config (line 207) | def from_config(cls, *args, **kwargs): method from_pretrained (line 211) | def from_pretrained(cls, *args, **kwargs): FILE: diffusers/src/diffusers/utils/dummy_note_seq_objects.py class MidiProcessor (line 5) | class MidiProcessor(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): FILE: diffusers/src/diffusers/utils/dummy_onnx_objects.py class OnnxRuntimeModel (line 5) | class OnnxRuntimeModel(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): FILE: diffusers/src/diffusers/utils/dummy_pt_objects.py class AsymmetricAutoencoderKL (line 5) | class AsymmetricAutoencoderKL(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class AutoencoderKL (line 20) | class AutoencoderKL(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): class AutoencoderKLTemporalDecoder (line 35) | class AutoencoderKLTemporalDecoder(metaclass=DummyObject): method __init__ (line 38) | def __init__(self, *args, **kwargs): method from_config (line 42) | def from_config(cls, *args, **kwargs): method from_pretrained (line 46) | def from_pretrained(cls, *args, **kwargs): class AutoencoderTiny (line 50) | class AutoencoderTiny(metaclass=DummyObject): method __init__ (line 53) | def __init__(self, *args, **kwargs): method from_config (line 57) | def from_config(cls, *args, **kwargs): method from_pretrained (line 61) | def from_pretrained(cls, *args, **kwargs): class ConsistencyDecoderVAE (line 65) | class ConsistencyDecoderVAE(metaclass=DummyObject): method __init__ (line 68) | def __init__(self, *args, **kwargs): method from_config (line 72) | def from_config(cls, *args, **kwargs): method from_pretrained (line 76) | def from_pretrained(cls, *args, **kwargs): class ControlNetModel (line 80) | class ControlNetModel(metaclass=DummyObject): method __init__ (line 83) | def __init__(self, *args, **kwargs): method from_config (line 87) | def from_config(cls, *args, **kwargs): method from_pretrained (line 91) | def from_pretrained(cls, *args, **kwargs): class ControlNetXSAdapter (line 95) | class ControlNetXSAdapter(metaclass=DummyObject): method __init__ (line 98) | def __init__(self, *args, **kwargs): method from_config (line 102) | def from_config(cls, *args, **kwargs): method from_pretrained (line 106) | def from_pretrained(cls, *args, **kwargs): class DiTTransformer2DModel (line 110) | class DiTTransformer2DModel(metaclass=DummyObject): method __init__ (line 113) | def __init__(self, *args, **kwargs): method from_config (line 117) | def from_config(cls, *args, **kwargs): method from_pretrained (line 121) | def from_pretrained(cls, *args, **kwargs): class HunyuanDiT2DModel (line 125) | class HunyuanDiT2DModel(metaclass=DummyObject): method __init__ (line 128) | def __init__(self, *args, **kwargs): method from_config (line 132) | def from_config(cls, *args, **kwargs): method from_pretrained (line 136) | def from_pretrained(cls, *args, **kwargs): class I2VGenXLUNet (line 140) | class I2VGenXLUNet(metaclass=DummyObject): method __init__ (line 143) | def __init__(self, *args, **kwargs): method from_config (line 147) | def from_config(cls, *args, **kwargs): method from_pretrained (line 151) | def from_pretrained(cls, *args, **kwargs): class Kandinsky3UNet (line 155) | class Kandinsky3UNet(metaclass=DummyObject): method __init__ (line 158) | def __init__(self, *args, **kwargs): method from_config (line 162) | def from_config(cls, *args, **kwargs): method from_pretrained (line 166) | def from_pretrained(cls, *args, **kwargs): class ModelMixin (line 170) | class ModelMixin(metaclass=DummyObject): method __init__ (line 173) | def __init__(self, *args, **kwargs): method from_config (line 177) | def from_config(cls, *args, **kwargs): method from_pretrained (line 181) | def from_pretrained(cls, *args, **kwargs): class MotionAdapter (line 185) | class MotionAdapter(metaclass=DummyObject): method __init__ (line 188) | def __init__(self, *args, **kwargs): method from_config (line 192) | def from_config(cls, *args, **kwargs): method from_pretrained (line 196) | def from_pretrained(cls, *args, **kwargs): class MultiAdapter (line 200) | class MultiAdapter(metaclass=DummyObject): method __init__ (line 203) | def __init__(self, *args, **kwargs): method from_config (line 207) | def from_config(cls, *args, **kwargs): method from_pretrained (line 211) | def from_pretrained(cls, *args, **kwargs): class PixArtTransformer2DModel (line 215) | class PixArtTransformer2DModel(metaclass=DummyObject): method __init__ (line 218) | def __init__(self, *args, **kwargs): method from_config (line 222) | def from_config(cls, *args, **kwargs): method from_pretrained (line 226) | def from_pretrained(cls, *args, **kwargs): class PriorTransformer (line 230) | class PriorTransformer(metaclass=DummyObject): method __init__ (line 233) | def __init__(self, *args, **kwargs): method from_config (line 237) | def from_config(cls, *args, **kwargs): method from_pretrained (line 241) | def from_pretrained(cls, *args, **kwargs): class SD3Transformer2DModel (line 245) | class SD3Transformer2DModel(metaclass=DummyObject): method __init__ (line 248) | def __init__(self, *args, **kwargs): method from_config (line 252) | def from_config(cls, *args, **kwargs): method from_pretrained (line 256) | def from_pretrained(cls, *args, **kwargs): class T2IAdapter (line 260) | class T2IAdapter(metaclass=DummyObject): method __init__ (line 263) | def __init__(self, *args, **kwargs): method from_config (line 267) | def from_config(cls, *args, **kwargs): method from_pretrained (line 271) | def from_pretrained(cls, *args, **kwargs): class T5FilmDecoder (line 275) | class T5FilmDecoder(metaclass=DummyObject): method __init__ (line 278) | def __init__(self, *args, **kwargs): method from_config (line 282) | def from_config(cls, *args, **kwargs): method from_pretrained (line 286) | def from_pretrained(cls, *args, **kwargs): class Transformer2DModel (line 290) | class Transformer2DModel(metaclass=DummyObject): method __init__ (line 293) | def __init__(self, *args, **kwargs): method from_config (line 297) | def from_config(cls, *args, **kwargs): method from_pretrained (line 301) | def from_pretrained(cls, *args, **kwargs): class UNet1DModel (line 305) | class UNet1DModel(metaclass=DummyObject): method __init__ (line 308) | def __init__(self, *args, **kwargs): method from_config (line 312) | def from_config(cls, *args, **kwargs): method from_pretrained (line 316) | def from_pretrained(cls, *args, **kwargs): class UNet2DConditionModel (line 320) | class UNet2DConditionModel(metaclass=DummyObject): method __init__ (line 323) | def __init__(self, *args, **kwargs): method from_config (line 327) | def from_config(cls, *args, **kwargs): method from_pretrained (line 331) | def from_pretrained(cls, *args, **kwargs): class UNet2DModel (line 335) | class UNet2DModel(metaclass=DummyObject): method __init__ (line 338) | def __init__(self, *args, **kwargs): method from_config (line 342) | def from_config(cls, *args, **kwargs): method from_pretrained (line 346) | def from_pretrained(cls, *args, **kwargs): class UNet3DConditionModel (line 350) | class UNet3DConditionModel(metaclass=DummyObject): method __init__ (line 353) | def __init__(self, *args, **kwargs): method from_config (line 357) | def from_config(cls, *args, **kwargs): method from_pretrained (line 361) | def from_pretrained(cls, *args, **kwargs): class UNetControlNetXSModel (line 365) | class UNetControlNetXSModel(metaclass=DummyObject): method __init__ (line 368) | def __init__(self, *args, **kwargs): method from_config (line 372) | def from_config(cls, *args, **kwargs): method from_pretrained (line 376) | def from_pretrained(cls, *args, **kwargs): class UNetMotionModel (line 380) | class UNetMotionModel(metaclass=DummyObject): method __init__ (line 383) | def __init__(self, *args, **kwargs): method from_config (line 387) | def from_config(cls, *args, **kwargs): method from_pretrained (line 391) | def from_pretrained(cls, *args, **kwargs): class UNetSpatioTemporalConditionModel (line 395) | class UNetSpatioTemporalConditionModel(metaclass=DummyObject): method __init__ (line 398) | def __init__(self, *args, **kwargs): method from_config (line 402) | def from_config(cls, *args, **kwargs): method from_pretrained (line 406) | def from_pretrained(cls, *args, **kwargs): class UVit2DModel (line 410) | class UVit2DModel(metaclass=DummyObject): method __init__ (line 413) | def __init__(self, *args, **kwargs): method from_config (line 417) | def from_config(cls, *args, **kwargs): method from_pretrained (line 421) | def from_pretrained(cls, *args, **kwargs): class VQModel (line 425) | class VQModel(metaclass=DummyObject): method __init__ (line 428) | def __init__(self, *args, **kwargs): method from_config (line 432) | def from_config(cls, *args, **kwargs): method from_pretrained (line 436) | def from_pretrained(cls, *args, **kwargs): function get_constant_schedule (line 440) | def get_constant_schedule(*args, **kwargs): function get_constant_schedule_with_warmup (line 444) | def get_constant_schedule_with_warmup(*args, **kwargs): function get_cosine_schedule_with_warmup (line 448) | def get_cosine_schedule_with_warmup(*args, **kwargs): function get_cosine_with_hard_restarts_schedule_with_warmup (line 452) | def get_cosine_with_hard_restarts_schedule_with_warmup(*args, **kwargs): function get_linear_schedule_with_warmup (line 456) | def get_linear_schedule_with_warmup(*args, **kwargs): function get_polynomial_decay_schedule_with_warmup (line 460) | def get_polynomial_decay_schedule_with_warmup(*args, **kwargs): function get_scheduler (line 464) | def get_scheduler(*args, **kwargs): class AudioPipelineOutput (line 468) | class AudioPipelineOutput(metaclass=DummyObject): method __init__ (line 471) | def __init__(self, *args, **kwargs): method from_config (line 475) | def from_config(cls, *args, **kwargs): method from_pretrained (line 479) | def from_pretrained(cls, *args, **kwargs): class AutoPipelineForImage2Image (line 483) | class AutoPipelineForImage2Image(metaclass=DummyObject): method __init__ (line 486) | def __init__(self, *args, **kwargs): method from_config (line 490) | def from_config(cls, *args, **kwargs): method from_pretrained (line 494) | def from_pretrained(cls, *args, **kwargs): class AutoPipelineForInpainting (line 498) | class AutoPipelineForInpainting(metaclass=DummyObject): method __init__ (line 501) | def __init__(self, *args, **kwargs): method from_config (line 505) | def from_config(cls, *args, **kwargs): method from_pretrained (line 509) | def from_pretrained(cls, *args, **kwargs): class AutoPipelineForText2Image (line 513) | class AutoPipelineForText2Image(metaclass=DummyObject): method __init__ (line 516) | def __init__(self, *args, **kwargs): method from_config (line 520) | def from_config(cls, *args, **kwargs): method from_pretrained (line 524) | def from_pretrained(cls, *args, **kwargs): class BlipDiffusionControlNetPipeline (line 528) | class BlipDiffusionControlNetPipeline(metaclass=DummyObject): method __init__ (line 531) | def __init__(self, *args, **kwargs): method from_config (line 535) | def from_config(cls, *args, **kwargs): method from_pretrained (line 539) | def from_pretrained(cls, *args, **kwargs): class BlipDiffusionPipeline (line 543) | class BlipDiffusionPipeline(metaclass=DummyObject): method __init__ (line 546) | def __init__(self, *args, **kwargs): method from_config (line 550) | def from_config(cls, *args, **kwargs): method from_pretrained (line 554) | def from_pretrained(cls, *args, **kwargs): class CLIPImageProjection (line 558) | class CLIPImageProjection(metaclass=DummyObject): method __init__ (line 561) | def __init__(self, *args, **kwargs): method from_config (line 565) | def from_config(cls, *args, **kwargs): method from_pretrained (line 569) | def from_pretrained(cls, *args, **kwargs): class ConsistencyModelPipeline (line 573) | class ConsistencyModelPipeline(metaclass=DummyObject): method __init__ (line 576) | def __init__(self, *args, **kwargs): method from_config (line 580) | def from_config(cls, *args, **kwargs): method from_pretrained (line 584) | def from_pretrained(cls, *args, **kwargs): class DanceDiffusionPipeline (line 588) | class DanceDiffusionPipeline(metaclass=DummyObject): method __init__ (line 591) | def __init__(self, *args, **kwargs): method from_config (line 595) | def from_config(cls, *args, **kwargs): method from_pretrained (line 599) | def from_pretrained(cls, *args, **kwargs): class DDIMPipeline (line 603) | class DDIMPipeline(metaclass=DummyObject): method __init__ (line 606) | def __init__(self, *args, **kwargs): method from_config (line 610) | def from_config(cls, *args, **kwargs): method from_pretrained (line 614) | def from_pretrained(cls, *args, **kwargs): class DDPMPipeline (line 618) | class DDPMPipeline(metaclass=DummyObject): method __init__ (line 621) | def __init__(self, *args, **kwargs): method from_config (line 625) | def from_config(cls, *args, **kwargs): method from_pretrained (line 629) | def from_pretrained(cls, *args, **kwargs): class DiffusionPipeline (line 633) | class DiffusionPipeline(metaclass=DummyObject): method __init__ (line 636) | def __init__(self, *args, **kwargs): method from_config (line 640) | def from_config(cls, *args, **kwargs): method from_pretrained (line 644) | def from_pretrained(cls, *args, **kwargs): class DiTPipeline (line 648) | class DiTPipeline(metaclass=DummyObject): method __init__ (line 651) | def __init__(self, *args, **kwargs): method from_config (line 655) | def from_config(cls, *args, **kwargs): method from_pretrained (line 659) | def from_pretrained(cls, *args, **kwargs): class ImagePipelineOutput (line 663) | class ImagePipelineOutput(metaclass=DummyObject): method __init__ (line 666) | def __init__(self, *args, **kwargs): method from_config (line 670) | def from_config(cls, *args, **kwargs): method from_pretrained (line 674) | def from_pretrained(cls, *args, **kwargs): class KarrasVePipeline (line 678) | class KarrasVePipeline(metaclass=DummyObject): method __init__ (line 681) | def __init__(self, *args, **kwargs): method from_config (line 685) | def from_config(cls, *args, **kwargs): method from_pretrained (line 689) | def from_pretrained(cls, *args, **kwargs): class LDMPipeline (line 693) | class LDMPipeline(metaclass=DummyObject): method __init__ (line 696) | def __init__(self, *args, **kwargs): method from_config (line 700) | def from_config(cls, *args, **kwargs): method from_pretrained (line 704) | def from_pretrained(cls, *args, **kwargs): class LDMSuperResolutionPipeline (line 708) | class LDMSuperResolutionPipeline(metaclass=DummyObject): method __init__ (line 711) | def __init__(self, *args, **kwargs): method from_config (line 715) | def from_config(cls, *args, **kwargs): method from_pretrained (line 719) | def from_pretrained(cls, *args, **kwargs): class PNDMPipeline (line 723) | class PNDMPipeline(metaclass=DummyObject): method __init__ (line 726) | def __init__(self, *args, **kwargs): method from_config (line 730) | def from_config(cls, *args, **kwargs): method from_pretrained (line 734) | def from_pretrained(cls, *args, **kwargs): class RePaintPipeline (line 738) | class RePaintPipeline(metaclass=DummyObject): method __init__ (line 741) | def __init__(self, *args, **kwargs): method from_config (line 745) | def from_config(cls, *args, **kwargs): method from_pretrained (line 749) | def from_pretrained(cls, *args, **kwargs): class ScoreSdeVePipeline (line 753) | class ScoreSdeVePipeline(metaclass=DummyObject): method __init__ (line 756) | def __init__(self, *args, **kwargs): method from_config (line 760) | def from_config(cls, *args, **kwargs): method from_pretrained (line 764) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionMixin (line 768) | class StableDiffusionMixin(metaclass=DummyObject): method __init__ (line 771) | def __init__(self, *args, **kwargs): method from_config (line 775) | def from_config(cls, *args, **kwargs): method from_pretrained (line 779) | def from_pretrained(cls, *args, **kwargs): class AmusedScheduler (line 783) | class AmusedScheduler(metaclass=DummyObject): method __init__ (line 786) | def __init__(self, *args, **kwargs): method from_config (line 790) | def from_config(cls, *args, **kwargs): method from_pretrained (line 794) | def from_pretrained(cls, *args, **kwargs): class CMStochasticIterativeScheduler (line 798) | class CMStochasticIterativeScheduler(metaclass=DummyObject): method __init__ (line 801) | def __init__(self, *args, **kwargs): method from_config (line 805) | def from_config(cls, *args, **kwargs): method from_pretrained (line 809) | def from_pretrained(cls, *args, **kwargs): class DDIMInverseScheduler (line 813) | class DDIMInverseScheduler(metaclass=DummyObject): method __init__ (line 816) | def __init__(self, *args, **kwargs): method from_config (line 820) | def from_config(cls, *args, **kwargs): method from_pretrained (line 824) | def from_pretrained(cls, *args, **kwargs): class DDIMParallelScheduler (line 828) | class DDIMParallelScheduler(metaclass=DummyObject): method __init__ (line 831) | def __init__(self, *args, **kwargs): method from_config (line 835) | def from_config(cls, *args, **kwargs): method from_pretrained (line 839) | def from_pretrained(cls, *args, **kwargs): class DDIMScheduler (line 843) | class DDIMScheduler(metaclass=DummyObject): method __init__ (line 846) | def __init__(self, *args, **kwargs): method from_config (line 850) | def from_config(cls, *args, **kwargs): method from_pretrained (line 854) | def from_pretrained(cls, *args, **kwargs): class DDPMParallelScheduler (line 858) | class DDPMParallelScheduler(metaclass=DummyObject): method __init__ (line 861) | def __init__(self, *args, **kwargs): method from_config (line 865) | def from_config(cls, *args, **kwargs): method from_pretrained (line 869) | def from_pretrained(cls, *args, **kwargs): class DDPMScheduler (line 873) | class DDPMScheduler(metaclass=DummyObject): method __init__ (line 876) | def __init__(self, *args, **kwargs): method from_config (line 880) | def from_config(cls, *args, **kwargs): method from_pretrained (line 884) | def from_pretrained(cls, *args, **kwargs): class DDPMWuerstchenScheduler (line 888) | class DDPMWuerstchenScheduler(metaclass=DummyObject): method __init__ (line 891) | def __init__(self, *args, **kwargs): method from_config (line 895) | def from_config(cls, *args, **kwargs): method from_pretrained (line 899) | def from_pretrained(cls, *args, **kwargs): class DEISMultistepScheduler (line 903) | class DEISMultistepScheduler(metaclass=DummyObject): method __init__ (line 906) | def __init__(self, *args, **kwargs): method from_config (line 910) | def from_config(cls, *args, **kwargs): method from_pretrained (line 914) | def from_pretrained(cls, *args, **kwargs): class DPMSolverMultistepInverseScheduler (line 918) | class DPMSolverMultistepInverseScheduler(metaclass=DummyObject): method __init__ (line 921) | def __init__(self, *args, **kwargs): method from_config (line 925) | def from_config(cls, *args, **kwargs): method from_pretrained (line 929) | def from_pretrained(cls, *args, **kwargs): class DPMSolverMultistepScheduler (line 933) | class DPMSolverMultistepScheduler(metaclass=DummyObject): method __init__ (line 936) | def __init__(self, *args, **kwargs): method from_config (line 940) | def from_config(cls, *args, **kwargs): method from_pretrained (line 944) | def from_pretrained(cls, *args, **kwargs): class DPMSolverSinglestepScheduler (line 948) | class DPMSolverSinglestepScheduler(metaclass=DummyObject): method __init__ (line 951) | def __init__(self, *args, **kwargs): method from_config (line 955) | def from_config(cls, *args, **kwargs): method from_pretrained (line 959) | def from_pretrained(cls, *args, **kwargs): class EDMDPMSolverMultistepScheduler (line 963) | class EDMDPMSolverMultistepScheduler(metaclass=DummyObject): method __init__ (line 966) | def __init__(self, *args, **kwargs): method from_config (line 970) | def from_config(cls, *args, **kwargs): method from_pretrained (line 974) | def from_pretrained(cls, *args, **kwargs): class EDMEulerScheduler (line 978) | class EDMEulerScheduler(metaclass=DummyObject): method __init__ (line 981) | def __init__(self, *args, **kwargs): method from_config (line 985) | def from_config(cls, *args, **kwargs): method from_pretrained (line 989) | def from_pretrained(cls, *args, **kwargs): class EulerAncestralDiscreteScheduler (line 993) | class EulerAncestralDiscreteScheduler(metaclass=DummyObject): method __init__ (line 996) | def __init__(self, *args, **kwargs): method from_config (line 1000) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1004) | def from_pretrained(cls, *args, **kwargs): class EulerDiscreteScheduler (line 1008) | class EulerDiscreteScheduler(metaclass=DummyObject): method __init__ (line 1011) | def __init__(self, *args, **kwargs): method from_config (line 1015) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1019) | def from_pretrained(cls, *args, **kwargs): class FlowMatchEulerDiscreteScheduler (line 1023) | class FlowMatchEulerDiscreteScheduler(metaclass=DummyObject): method __init__ (line 1026) | def __init__(self, *args, **kwargs): method from_config (line 1030) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1034) | def from_pretrained(cls, *args, **kwargs): class HeunDiscreteScheduler (line 1038) | class HeunDiscreteScheduler(metaclass=DummyObject): method __init__ (line 1041) | def __init__(self, *args, **kwargs): method from_config (line 1045) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1049) | def from_pretrained(cls, *args, **kwargs): class IPNDMScheduler (line 1053) | class IPNDMScheduler(metaclass=DummyObject): method __init__ (line 1056) | def __init__(self, *args, **kwargs): method from_config (line 1060) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1064) | def from_pretrained(cls, *args, **kwargs): class KarrasVeScheduler (line 1068) | class KarrasVeScheduler(metaclass=DummyObject): method __init__ (line 1071) | def __init__(self, *args, **kwargs): method from_config (line 1075) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1079) | def from_pretrained(cls, *args, **kwargs): class KDPM2AncestralDiscreteScheduler (line 1083) | class KDPM2AncestralDiscreteScheduler(metaclass=DummyObject): method __init__ (line 1086) | def __init__(self, *args, **kwargs): method from_config (line 1090) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1094) | def from_pretrained(cls, *args, **kwargs): class KDPM2DiscreteScheduler (line 1098) | class KDPM2DiscreteScheduler(metaclass=DummyObject): method __init__ (line 1101) | def __init__(self, *args, **kwargs): method from_config (line 1105) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1109) | def from_pretrained(cls, *args, **kwargs): class LCMScheduler (line 1113) | class LCMScheduler(metaclass=DummyObject): method __init__ (line 1116) | def __init__(self, *args, **kwargs): method from_config (line 1120) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1124) | def from_pretrained(cls, *args, **kwargs): class PNDMScheduler (line 1128) | class PNDMScheduler(metaclass=DummyObject): method __init__ (line 1131) | def __init__(self, *args, **kwargs): method from_config (line 1135) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1139) | def from_pretrained(cls, *args, **kwargs): class RePaintScheduler (line 1143) | class RePaintScheduler(metaclass=DummyObject): method __init__ (line 1146) | def __init__(self, *args, **kwargs): method from_config (line 1150) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1154) | def from_pretrained(cls, *args, **kwargs): class SASolverScheduler (line 1158) | class SASolverScheduler(metaclass=DummyObject): method __init__ (line 1161) | def __init__(self, *args, **kwargs): method from_config (line 1165) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1169) | def from_pretrained(cls, *args, **kwargs): class SchedulerMixin (line 1173) | class SchedulerMixin(metaclass=DummyObject): method __init__ (line 1176) | def __init__(self, *args, **kwargs): method from_config (line 1180) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1184) | def from_pretrained(cls, *args, **kwargs): class ScoreSdeVeScheduler (line 1188) | class ScoreSdeVeScheduler(metaclass=DummyObject): method __init__ (line 1191) | def __init__(self, *args, **kwargs): method from_config (line 1195) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1199) | def from_pretrained(cls, *args, **kwargs): class TCDScheduler (line 1203) | class TCDScheduler(metaclass=DummyObject): method __init__ (line 1206) | def __init__(self, *args, **kwargs): method from_config (line 1210) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1214) | def from_pretrained(cls, *args, **kwargs): class UnCLIPScheduler (line 1218) | class UnCLIPScheduler(metaclass=DummyObject): method __init__ (line 1221) | def __init__(self, *args, **kwargs): method from_config (line 1225) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1229) | def from_pretrained(cls, *args, **kwargs): class UniPCMultistepScheduler (line 1233) | class UniPCMultistepScheduler(metaclass=DummyObject): method __init__ (line 1236) | def __init__(self, *args, **kwargs): method from_config (line 1240) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1244) | def from_pretrained(cls, *args, **kwargs): class VQDiffusionScheduler (line 1248) | class VQDiffusionScheduler(metaclass=DummyObject): method __init__ (line 1251) | def __init__(self, *args, **kwargs): method from_config (line 1255) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1259) | def from_pretrained(cls, *args, **kwargs): class EMAModel (line 1263) | class EMAModel(metaclass=DummyObject): method __init__ (line 1266) | def __init__(self, *args, **kwargs): method from_config (line 1270) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1274) | def from_pretrained(cls, *args, **kwargs): FILE: diffusers/src/diffusers/utils/dummy_torch_and_librosa_objects.py class AudioDiffusionPipeline (line 5) | class AudioDiffusionPipeline(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class Mel (line 20) | class Mel(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): FILE: diffusers/src/diffusers/utils/dummy_torch_and_scipy_objects.py class LMSDiscreteScheduler (line 5) | class LMSDiscreteScheduler(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): FILE: diffusers/src/diffusers/utils/dummy_torch_and_torchsde_objects.py class DPMSolverSDEScheduler (line 5) | class DPMSolverSDEScheduler(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): FILE: diffusers/src/diffusers/utils/dummy_torch_and_transformers_and_k_diffusion_objects.py class StableDiffusionKDiffusionPipeline (line 5) | class StableDiffusionKDiffusionPipeline(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLKDiffusionPipeline (line 20) | class StableDiffusionXLKDiffusionPipeline(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): FILE: diffusers/src/diffusers/utils/dummy_torch_and_transformers_and_onnx_objects.py class OnnxStableDiffusionImg2ImgPipeline (line 5) | class OnnxStableDiffusionImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class OnnxStableDiffusionInpaintPipeline (line 20) | class OnnxStableDiffusionInpaintPipeline(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): class OnnxStableDiffusionInpaintPipelineLegacy (line 35) | class OnnxStableDiffusionInpaintPipelineLegacy(metaclass=DummyObject): method __init__ (line 38) | def __init__(self, *args, **kwargs): method from_config (line 42) | def from_config(cls, *args, **kwargs): method from_pretrained (line 46) | def from_pretrained(cls, *args, **kwargs): class OnnxStableDiffusionPipeline (line 50) | class OnnxStableDiffusionPipeline(metaclass=DummyObject): method __init__ (line 53) | def __init__(self, *args, **kwargs): method from_config (line 57) | def from_config(cls, *args, **kwargs): method from_pretrained (line 61) | def from_pretrained(cls, *args, **kwargs): class OnnxStableDiffusionUpscalePipeline (line 65) | class OnnxStableDiffusionUpscalePipeline(metaclass=DummyObject): method __init__ (line 68) | def __init__(self, *args, **kwargs): method from_config (line 72) | def from_config(cls, *args, **kwargs): method from_pretrained (line 76) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionOnnxPipeline (line 80) | class StableDiffusionOnnxPipeline(metaclass=DummyObject): method __init__ (line 83) | def __init__(self, *args, **kwargs): method from_config (line 87) | def from_config(cls, *args, **kwargs): method from_pretrained (line 91) | def from_pretrained(cls, *args, **kwargs): FILE: diffusers/src/diffusers/utils/dummy_torch_and_transformers_objects.py class AltDiffusionImg2ImgPipeline (line 5) | class AltDiffusionImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): class AltDiffusionPipeline (line 20) | class AltDiffusionPipeline(metaclass=DummyObject): method __init__ (line 23) | def __init__(self, *args, **kwargs): method from_config (line 27) | def from_config(cls, *args, **kwargs): method from_pretrained (line 31) | def from_pretrained(cls, *args, **kwargs): class AmusedImg2ImgPipeline (line 35) | class AmusedImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 38) | def __init__(self, *args, **kwargs): method from_config (line 42) | def from_config(cls, *args, **kwargs): method from_pretrained (line 46) | def from_pretrained(cls, *args, **kwargs): class AmusedInpaintPipeline (line 50) | class AmusedInpaintPipeline(metaclass=DummyObject): method __init__ (line 53) | def __init__(self, *args, **kwargs): method from_config (line 57) | def from_config(cls, *args, **kwargs): method from_pretrained (line 61) | def from_pretrained(cls, *args, **kwargs): class AmusedPipeline (line 65) | class AmusedPipeline(metaclass=DummyObject): method __init__ (line 68) | def __init__(self, *args, **kwargs): method from_config (line 72) | def from_config(cls, *args, **kwargs): method from_pretrained (line 76) | def from_pretrained(cls, *args, **kwargs): class AnimateDiffPipeline (line 80) | class AnimateDiffPipeline(metaclass=DummyObject): method __init__ (line 83) | def __init__(self, *args, **kwargs): method from_config (line 87) | def from_config(cls, *args, **kwargs): method from_pretrained (line 91) | def from_pretrained(cls, *args, **kwargs): class AnimateDiffSDXLPipeline (line 95) | class AnimateDiffSDXLPipeline(metaclass=DummyObject): method __init__ (line 98) | def __init__(self, *args, **kwargs): method from_config (line 102) | def from_config(cls, *args, **kwargs): method from_pretrained (line 106) | def from_pretrained(cls, *args, **kwargs): class AnimateDiffVideoToVideoPipeline (line 110) | class AnimateDiffVideoToVideoPipeline(metaclass=DummyObject): method __init__ (line 113) | def __init__(self, *args, **kwargs): method from_config (line 117) | def from_config(cls, *args, **kwargs): method from_pretrained (line 121) | def from_pretrained(cls, *args, **kwargs): class AudioLDM2Pipeline (line 125) | class AudioLDM2Pipeline(metaclass=DummyObject): method __init__ (line 128) | def __init__(self, *args, **kwargs): method from_config (line 132) | def from_config(cls, *args, **kwargs): method from_pretrained (line 136) | def from_pretrained(cls, *args, **kwargs): class AudioLDM2ProjectionModel (line 140) | class AudioLDM2ProjectionModel(metaclass=DummyObject): method __init__ (line 143) | def __init__(self, *args, **kwargs): method from_config (line 147) | def from_config(cls, *args, **kwargs): method from_pretrained (line 151) | def from_pretrained(cls, *args, **kwargs): class AudioLDM2UNet2DConditionModel (line 155) | class AudioLDM2UNet2DConditionModel(metaclass=DummyObject): method __init__ (line 158) | def __init__(self, *args, **kwargs): method from_config (line 162) | def from_config(cls, *args, **kwargs): method from_pretrained (line 166) | def from_pretrained(cls, *args, **kwargs): class AudioLDMPipeline (line 170) | class AudioLDMPipeline(metaclass=DummyObject): method __init__ (line 173) | def __init__(self, *args, **kwargs): method from_config (line 177) | def from_config(cls, *args, **kwargs): method from_pretrained (line 181) | def from_pretrained(cls, *args, **kwargs): class CLIPImageProjection (line 185) | class CLIPImageProjection(metaclass=DummyObject): method __init__ (line 188) | def __init__(self, *args, **kwargs): method from_config (line 192) | def from_config(cls, *args, **kwargs): method from_pretrained (line 196) | def from_pretrained(cls, *args, **kwargs): class CycleDiffusionPipeline (line 200) | class CycleDiffusionPipeline(metaclass=DummyObject): method __init__ (line 203) | def __init__(self, *args, **kwargs): method from_config (line 207) | def from_config(cls, *args, **kwargs): method from_pretrained (line 211) | def from_pretrained(cls, *args, **kwargs): class HunyuanDiTPipeline (line 215) | class HunyuanDiTPipeline(metaclass=DummyObject): method __init__ (line 218) | def __init__(self, *args, **kwargs): method from_config (line 222) | def from_config(cls, *args, **kwargs): method from_pretrained (line 226) | def from_pretrained(cls, *args, **kwargs): class I2VGenXLPipeline (line 230) | class I2VGenXLPipeline(metaclass=DummyObject): method __init__ (line 233) | def __init__(self, *args, **kwargs): method from_config (line 237) | def from_config(cls, *args, **kwargs): method from_pretrained (line 241) | def from_pretrained(cls, *args, **kwargs): class IFImg2ImgPipeline (line 245) | class IFImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 248) | def __init__(self, *args, **kwargs): method from_config (line 252) | def from_config(cls, *args, **kwargs): method from_pretrained (line 256) | def from_pretrained(cls, *args, **kwargs): class IFImg2ImgSuperResolutionPipeline (line 260) | class IFImg2ImgSuperResolutionPipeline(metaclass=DummyObject): method __init__ (line 263) | def __init__(self, *args, **kwargs): method from_config (line 267) | def from_config(cls, *args, **kwargs): method from_pretrained (line 271) | def from_pretrained(cls, *args, **kwargs): class IFInpaintingPipeline (line 275) | class IFInpaintingPipeline(metaclass=DummyObject): method __init__ (line 278) | def __init__(self, *args, **kwargs): method from_config (line 282) | def from_config(cls, *args, **kwargs): method from_pretrained (line 286) | def from_pretrained(cls, *args, **kwargs): class IFInpaintingSuperResolutionPipeline (line 290) | class IFInpaintingSuperResolutionPipeline(metaclass=DummyObject): method __init__ (line 293) | def __init__(self, *args, **kwargs): method from_config (line 297) | def from_config(cls, *args, **kwargs): method from_pretrained (line 301) | def from_pretrained(cls, *args, **kwargs): class IFPipeline (line 305) | class IFPipeline(metaclass=DummyObject): method __init__ (line 308) | def __init__(self, *args, **kwargs): method from_config (line 312) | def from_config(cls, *args, **kwargs): method from_pretrained (line 316) | def from_pretrained(cls, *args, **kwargs): class IFSuperResolutionPipeline (line 320) | class IFSuperResolutionPipeline(metaclass=DummyObject): method __init__ (line 323) | def __init__(self, *args, **kwargs): method from_config (line 327) | def from_config(cls, *args, **kwargs): method from_pretrained (line 331) | def from_pretrained(cls, *args, **kwargs): class ImageTextPipelineOutput (line 335) | class ImageTextPipelineOutput(metaclass=DummyObject): method __init__ (line 338) | def __init__(self, *args, **kwargs): method from_config (line 342) | def from_config(cls, *args, **kwargs): method from_pretrained (line 346) | def from_pretrained(cls, *args, **kwargs): class Kandinsky3Img2ImgPipeline (line 350) | class Kandinsky3Img2ImgPipeline(metaclass=DummyObject): method __init__ (line 353) | def __init__(self, *args, **kwargs): method from_config (line 357) | def from_config(cls, *args, **kwargs): method from_pretrained (line 361) | def from_pretrained(cls, *args, **kwargs): class Kandinsky3Pipeline (line 365) | class Kandinsky3Pipeline(metaclass=DummyObject): method __init__ (line 368) | def __init__(self, *args, **kwargs): method from_config (line 372) | def from_config(cls, *args, **kwargs): method from_pretrained (line 376) | def from_pretrained(cls, *args, **kwargs): class KandinskyCombinedPipeline (line 380) | class KandinskyCombinedPipeline(metaclass=DummyObject): method __init__ (line 383) | def __init__(self, *args, **kwargs): method from_config (line 387) | def from_config(cls, *args, **kwargs): method from_pretrained (line 391) | def from_pretrained(cls, *args, **kwargs): class KandinskyImg2ImgCombinedPipeline (line 395) | class KandinskyImg2ImgCombinedPipeline(metaclass=DummyObject): method __init__ (line 398) | def __init__(self, *args, **kwargs): method from_config (line 402) | def from_config(cls, *args, **kwargs): method from_pretrained (line 406) | def from_pretrained(cls, *args, **kwargs): class KandinskyImg2ImgPipeline (line 410) | class KandinskyImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 413) | def __init__(self, *args, **kwargs): method from_config (line 417) | def from_config(cls, *args, **kwargs): method from_pretrained (line 421) | def from_pretrained(cls, *args, **kwargs): class KandinskyInpaintCombinedPipeline (line 425) | class KandinskyInpaintCombinedPipeline(metaclass=DummyObject): method __init__ (line 428) | def __init__(self, *args, **kwargs): method from_config (line 432) | def from_config(cls, *args, **kwargs): method from_pretrained (line 436) | def from_pretrained(cls, *args, **kwargs): class KandinskyInpaintPipeline (line 440) | class KandinskyInpaintPipeline(metaclass=DummyObject): method __init__ (line 443) | def __init__(self, *args, **kwargs): method from_config (line 447) | def from_config(cls, *args, **kwargs): method from_pretrained (line 451) | def from_pretrained(cls, *args, **kwargs): class KandinskyPipeline (line 455) | class KandinskyPipeline(metaclass=DummyObject): method __init__ (line 458) | def __init__(self, *args, **kwargs): method from_config (line 462) | def from_config(cls, *args, **kwargs): method from_pretrained (line 466) | def from_pretrained(cls, *args, **kwargs): class KandinskyPriorPipeline (line 470) | class KandinskyPriorPipeline(metaclass=DummyObject): method __init__ (line 473) | def __init__(self, *args, **kwargs): method from_config (line 477) | def from_config(cls, *args, **kwargs): method from_pretrained (line 481) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22CombinedPipeline (line 485) | class KandinskyV22CombinedPipeline(metaclass=DummyObject): method __init__ (line 488) | def __init__(self, *args, **kwargs): method from_config (line 492) | def from_config(cls, *args, **kwargs): method from_pretrained (line 496) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22ControlnetImg2ImgPipeline (line 500) | class KandinskyV22ControlnetImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 503) | def __init__(self, *args, **kwargs): method from_config (line 507) | def from_config(cls, *args, **kwargs): method from_pretrained (line 511) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22ControlnetPipeline (line 515) | class KandinskyV22ControlnetPipeline(metaclass=DummyObject): method __init__ (line 518) | def __init__(self, *args, **kwargs): method from_config (line 522) | def from_config(cls, *args, **kwargs): method from_pretrained (line 526) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22Img2ImgCombinedPipeline (line 530) | class KandinskyV22Img2ImgCombinedPipeline(metaclass=DummyObject): method __init__ (line 533) | def __init__(self, *args, **kwargs): method from_config (line 537) | def from_config(cls, *args, **kwargs): method from_pretrained (line 541) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22Img2ImgPipeline (line 545) | class KandinskyV22Img2ImgPipeline(metaclass=DummyObject): method __init__ (line 548) | def __init__(self, *args, **kwargs): method from_config (line 552) | def from_config(cls, *args, **kwargs): method from_pretrained (line 556) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22InpaintCombinedPipeline (line 560) | class KandinskyV22InpaintCombinedPipeline(metaclass=DummyObject): method __init__ (line 563) | def __init__(self, *args, **kwargs): method from_config (line 567) | def from_config(cls, *args, **kwargs): method from_pretrained (line 571) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22InpaintPipeline (line 575) | class KandinskyV22InpaintPipeline(metaclass=DummyObject): method __init__ (line 578) | def __init__(self, *args, **kwargs): method from_config (line 582) | def from_config(cls, *args, **kwargs): method from_pretrained (line 586) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22Pipeline (line 590) | class KandinskyV22Pipeline(metaclass=DummyObject): method __init__ (line 593) | def __init__(self, *args, **kwargs): method from_config (line 597) | def from_config(cls, *args, **kwargs): method from_pretrained (line 601) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22PriorEmb2EmbPipeline (line 605) | class KandinskyV22PriorEmb2EmbPipeline(metaclass=DummyObject): method __init__ (line 608) | def __init__(self, *args, **kwargs): method from_config (line 612) | def from_config(cls, *args, **kwargs): method from_pretrained (line 616) | def from_pretrained(cls, *args, **kwargs): class KandinskyV22PriorPipeline (line 620) | class KandinskyV22PriorPipeline(metaclass=DummyObject): method __init__ (line 623) | def __init__(self, *args, **kwargs): method from_config (line 627) | def from_config(cls, *args, **kwargs): method from_pretrained (line 631) | def from_pretrained(cls, *args, **kwargs): class LatentConsistencyModelImg2ImgPipeline (line 635) | class LatentConsistencyModelImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 638) | def __init__(self, *args, **kwargs): method from_config (line 642) | def from_config(cls, *args, **kwargs): method from_pretrained (line 646) | def from_pretrained(cls, *args, **kwargs): class LatentConsistencyModelPipeline (line 650) | class LatentConsistencyModelPipeline(metaclass=DummyObject): method __init__ (line 653) | def __init__(self, *args, **kwargs): method from_config (line 657) | def from_config(cls, *args, **kwargs): method from_pretrained (line 661) | def from_pretrained(cls, *args, **kwargs): class LDMTextToImagePipeline (line 665) | class LDMTextToImagePipeline(metaclass=DummyObject): method __init__ (line 668) | def __init__(self, *args, **kwargs): method from_config (line 672) | def from_config(cls, *args, **kwargs): method from_pretrained (line 676) | def from_pretrained(cls, *args, **kwargs): class LEditsPPPipelineStableDiffusion (line 680) | class LEditsPPPipelineStableDiffusion(metaclass=DummyObject): method __init__ (line 683) | def __init__(self, *args, **kwargs): method from_config (line 687) | def from_config(cls, *args, **kwargs): method from_pretrained (line 691) | def from_pretrained(cls, *args, **kwargs): class LEditsPPPipelineStableDiffusionXL (line 695) | class LEditsPPPipelineStableDiffusionXL(metaclass=DummyObject): method __init__ (line 698) | def __init__(self, *args, **kwargs): method from_config (line 702) | def from_config(cls, *args, **kwargs): method from_pretrained (line 706) | def from_pretrained(cls, *args, **kwargs): class MarigoldDepthPipeline (line 710) | class MarigoldDepthPipeline(metaclass=DummyObject): method __init__ (line 713) | def __init__(self, *args, **kwargs): method from_config (line 717) | def from_config(cls, *args, **kwargs): method from_pretrained (line 721) | def from_pretrained(cls, *args, **kwargs): class MarigoldNormalsPipeline (line 725) | class MarigoldNormalsPipeline(metaclass=DummyObject): method __init__ (line 728) | def __init__(self, *args, **kwargs): method from_config (line 732) | def from_config(cls, *args, **kwargs): method from_pretrained (line 736) | def from_pretrained(cls, *args, **kwargs): class MusicLDMPipeline (line 740) | class MusicLDMPipeline(metaclass=DummyObject): method __init__ (line 743) | def __init__(self, *args, **kwargs): method from_config (line 747) | def from_config(cls, *args, **kwargs): method from_pretrained (line 751) | def from_pretrained(cls, *args, **kwargs): class PaintByExamplePipeline (line 755) | class PaintByExamplePipeline(metaclass=DummyObject): method __init__ (line 758) | def __init__(self, *args, **kwargs): method from_config (line 762) | def from_config(cls, *args, **kwargs): method from_pretrained (line 766) | def from_pretrained(cls, *args, **kwargs): class PIAPipeline (line 770) | class PIAPipeline(metaclass=DummyObject): method __init__ (line 773) | def __init__(self, *args, **kwargs): method from_config (line 777) | def from_config(cls, *args, **kwargs): method from_pretrained (line 781) | def from_pretrained(cls, *args, **kwargs): class PixArtAlphaPipeline (line 785) | class PixArtAlphaPipeline(metaclass=DummyObject): method __init__ (line 788) | def __init__(self, *args, **kwargs): method from_config (line 792) | def from_config(cls, *args, **kwargs): method from_pretrained (line 796) | def from_pretrained(cls, *args, **kwargs): class PixArtSigmaPipeline (line 800) | class PixArtSigmaPipeline(metaclass=DummyObject): method __init__ (line 803) | def __init__(self, *args, **kwargs): method from_config (line 807) | def from_config(cls, *args, **kwargs): method from_pretrained (line 811) | def from_pretrained(cls, *args, **kwargs): class SemanticStableDiffusionPipeline (line 815) | class SemanticStableDiffusionPipeline(metaclass=DummyObject): method __init__ (line 818) | def __init__(self, *args, **kwargs): method from_config (line 822) | def from_config(cls, *args, **kwargs): method from_pretrained (line 826) | def from_pretrained(cls, *args, **kwargs): class ShapEImg2ImgPipeline (line 830) | class ShapEImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 833) | def __init__(self, *args, **kwargs): method from_config (line 837) | def from_config(cls, *args, **kwargs): method from_pretrained (line 841) | def from_pretrained(cls, *args, **kwargs): class ShapEPipeline (line 845) | class ShapEPipeline(metaclass=DummyObject): method __init__ (line 848) | def __init__(self, *args, **kwargs): method from_config (line 852) | def from_config(cls, *args, **kwargs): method from_pretrained (line 856) | def from_pretrained(cls, *args, **kwargs): class StableCascadeCombinedPipeline (line 860) | class StableCascadeCombinedPipeline(metaclass=DummyObject): method __init__ (line 863) | def __init__(self, *args, **kwargs): method from_config (line 867) | def from_config(cls, *args, **kwargs): method from_pretrained (line 871) | def from_pretrained(cls, *args, **kwargs): class StableCascadeDecoderPipeline (line 875) | class StableCascadeDecoderPipeline(metaclass=DummyObject): method __init__ (line 878) | def __init__(self, *args, **kwargs): method from_config (line 882) | def from_config(cls, *args, **kwargs): method from_pretrained (line 886) | def from_pretrained(cls, *args, **kwargs): class StableCascadePriorPipeline (line 890) | class StableCascadePriorPipeline(metaclass=DummyObject): method __init__ (line 893) | def __init__(self, *args, **kwargs): method from_config (line 897) | def from_config(cls, *args, **kwargs): method from_pretrained (line 901) | def from_pretrained(cls, *args, **kwargs): class StableDiffusion3Img2ImgPipeline (line 905) | class StableDiffusion3Img2ImgPipeline(metaclass=DummyObject): method __init__ (line 908) | def __init__(self, *args, **kwargs): method from_config (line 912) | def from_config(cls, *args, **kwargs): method from_pretrained (line 916) | def from_pretrained(cls, *args, **kwargs): class StableDiffusion3Pipeline (line 920) | class StableDiffusion3Pipeline(metaclass=DummyObject): method __init__ (line 923) | def __init__(self, *args, **kwargs): method from_config (line 927) | def from_config(cls, *args, **kwargs): method from_pretrained (line 931) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionAdapterPipeline (line 935) | class StableDiffusionAdapterPipeline(metaclass=DummyObject): method __init__ (line 938) | def __init__(self, *args, **kwargs): method from_config (line 942) | def from_config(cls, *args, **kwargs): method from_pretrained (line 946) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionAttendAndExcitePipeline (line 950) | class StableDiffusionAttendAndExcitePipeline(metaclass=DummyObject): method __init__ (line 953) | def __init__(self, *args, **kwargs): method from_config (line 957) | def from_config(cls, *args, **kwargs): method from_pretrained (line 961) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionControlNetImg2ImgPipeline (line 965) | class StableDiffusionControlNetImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 968) | def __init__(self, *args, **kwargs): method from_config (line 972) | def from_config(cls, *args, **kwargs): method from_pretrained (line 976) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionControlNetInpaintPipeline (line 980) | class StableDiffusionControlNetInpaintPipeline(metaclass=DummyObject): method __init__ (line 983) | def __init__(self, *args, **kwargs): method from_config (line 987) | def from_config(cls, *args, **kwargs): method from_pretrained (line 991) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionControlNetPipeline (line 995) | class StableDiffusionControlNetPipeline(metaclass=DummyObject): method __init__ (line 998) | def __init__(self, *args, **kwargs): method from_config (line 1002) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1006) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionControlNetXSPipeline (line 1010) | class StableDiffusionControlNetXSPipeline(metaclass=DummyObject): method __init__ (line 1013) | def __init__(self, *args, **kwargs): method from_config (line 1017) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1021) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionDepth2ImgPipeline (line 1025) | class StableDiffusionDepth2ImgPipeline(metaclass=DummyObject): method __init__ (line 1028) | def __init__(self, *args, **kwargs): method from_config (line 1032) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1036) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionDiffEditPipeline (line 1040) | class StableDiffusionDiffEditPipeline(metaclass=DummyObject): method __init__ (line 1043) | def __init__(self, *args, **kwargs): method from_config (line 1047) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1051) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionGLIGENPipeline (line 1055) | class StableDiffusionGLIGENPipeline(metaclass=DummyObject): method __init__ (line 1058) | def __init__(self, *args, **kwargs): method from_config (line 1062) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1066) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionGLIGENTextImagePipeline (line 1070) | class StableDiffusionGLIGENTextImagePipeline(metaclass=DummyObject): method __init__ (line 1073) | def __init__(self, *args, **kwargs): method from_config (line 1077) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1081) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionImageVariationPipeline (line 1085) | class StableDiffusionImageVariationPipeline(metaclass=DummyObject): method __init__ (line 1088) | def __init__(self, *args, **kwargs): method from_config (line 1092) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1096) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionImg2ImgPipeline (line 1100) | class StableDiffusionImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 1103) | def __init__(self, *args, **kwargs): method from_config (line 1107) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1111) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionInpaintPipeline (line 1115) | class StableDiffusionInpaintPipeline(metaclass=DummyObject): method __init__ (line 1118) | def __init__(self, *args, **kwargs): method from_config (line 1122) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1126) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionInpaintPipelineLegacy (line 1130) | class StableDiffusionInpaintPipelineLegacy(metaclass=DummyObject): method __init__ (line 1133) | def __init__(self, *args, **kwargs): method from_config (line 1137) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1141) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionInstructPix2PixPipeline (line 1145) | class StableDiffusionInstructPix2PixPipeline(metaclass=DummyObject): method __init__ (line 1148) | def __init__(self, *args, **kwargs): method from_config (line 1152) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1156) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionLatentUpscalePipeline (line 1160) | class StableDiffusionLatentUpscalePipeline(metaclass=DummyObject): method __init__ (line 1163) | def __init__(self, *args, **kwargs): method from_config (line 1167) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1171) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionLDM3DPipeline (line 1175) | class StableDiffusionLDM3DPipeline(metaclass=DummyObject): method __init__ (line 1178) | def __init__(self, *args, **kwargs): method from_config (line 1182) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1186) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionModelEditingPipeline (line 1190) | class StableDiffusionModelEditingPipeline(metaclass=DummyObject): method __init__ (line 1193) | def __init__(self, *args, **kwargs): method from_config (line 1197) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1201) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionPanoramaPipeline (line 1205) | class StableDiffusionPanoramaPipeline(metaclass=DummyObject): method __init__ (line 1208) | def __init__(self, *args, **kwargs): method from_config (line 1212) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1216) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionParadigmsPipeline (line 1220) | class StableDiffusionParadigmsPipeline(metaclass=DummyObject): method __init__ (line 1223) | def __init__(self, *args, **kwargs): method from_config (line 1227) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1231) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionPipeline (line 1235) | class StableDiffusionPipeline(metaclass=DummyObject): method __init__ (line 1238) | def __init__(self, *args, **kwargs): method from_config (line 1242) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1246) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionPipelineSafe (line 1250) | class StableDiffusionPipelineSafe(metaclass=DummyObject): method __init__ (line 1253) | def __init__(self, *args, **kwargs): method from_config (line 1257) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1261) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionPix2PixZeroPipeline (line 1265) | class StableDiffusionPix2PixZeroPipeline(metaclass=DummyObject): method __init__ (line 1268) | def __init__(self, *args, **kwargs): method from_config (line 1272) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1276) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionSAGPipeline (line 1280) | class StableDiffusionSAGPipeline(metaclass=DummyObject): method __init__ (line 1283) | def __init__(self, *args, **kwargs): method from_config (line 1287) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1291) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionUpscalePipeline (line 1295) | class StableDiffusionUpscalePipeline(metaclass=DummyObject): method __init__ (line 1298) | def __init__(self, *args, **kwargs): method from_config (line 1302) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1306) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLAdapterPipeline (line 1310) | class StableDiffusionXLAdapterPipeline(metaclass=DummyObject): method __init__ (line 1313) | def __init__(self, *args, **kwargs): method from_config (line 1317) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1321) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLControlNetImg2ImgPipeline (line 1325) | class StableDiffusionXLControlNetImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 1328) | def __init__(self, *args, **kwargs): method from_config (line 1332) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1336) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLControlNetInpaintPipeline (line 1340) | class StableDiffusionXLControlNetInpaintPipeline(metaclass=DummyObject): method __init__ (line 1343) | def __init__(self, *args, **kwargs): method from_config (line 1347) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1351) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLControlNetPipeline (line 1355) | class StableDiffusionXLControlNetPipeline(metaclass=DummyObject): method __init__ (line 1358) | def __init__(self, *args, **kwargs): method from_config (line 1362) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1366) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLControlNetXSPipeline (line 1370) | class StableDiffusionXLControlNetXSPipeline(metaclass=DummyObject): method __init__ (line 1373) | def __init__(self, *args, **kwargs): method from_config (line 1377) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1381) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLImg2ImgPipeline (line 1385) | class StableDiffusionXLImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 1388) | def __init__(self, *args, **kwargs): method from_config (line 1392) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1396) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLInpaintPipeline (line 1400) | class StableDiffusionXLInpaintPipeline(metaclass=DummyObject): method __init__ (line 1403) | def __init__(self, *args, **kwargs): method from_config (line 1407) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1411) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLInstructPix2PixPipeline (line 1415) | class StableDiffusionXLInstructPix2PixPipeline(metaclass=DummyObject): method __init__ (line 1418) | def __init__(self, *args, **kwargs): method from_config (line 1422) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1426) | def from_pretrained(cls, *args, **kwargs): class StableDiffusionXLPipeline (line 1430) | class StableDiffusionXLPipeline(metaclass=DummyObject): method __init__ (line 1433) | def __init__(self, *args, **kwargs): method from_config (line 1437) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1441) | def from_pretrained(cls, *args, **kwargs): class StableUnCLIPImg2ImgPipeline (line 1445) | class StableUnCLIPImg2ImgPipeline(metaclass=DummyObject): method __init__ (line 1448) | def __init__(self, *args, **kwargs): method from_config (line 1452) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1456) | def from_pretrained(cls, *args, **kwargs): class StableUnCLIPPipeline (line 1460) | class StableUnCLIPPipeline(metaclass=DummyObject): method __init__ (line 1463) | def __init__(self, *args, **kwargs): method from_config (line 1467) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1471) | def from_pretrained(cls, *args, **kwargs): class StableVideoDiffusionPipeline (line 1475) | class StableVideoDiffusionPipeline(metaclass=DummyObject): method __init__ (line 1478) | def __init__(self, *args, **kwargs): method from_config (line 1482) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1486) | def from_pretrained(cls, *args, **kwargs): class TextToVideoSDPipeline (line 1490) | class TextToVideoSDPipeline(metaclass=DummyObject): method __init__ (line 1493) | def __init__(self, *args, **kwargs): method from_config (line 1497) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1501) | def from_pretrained(cls, *args, **kwargs): class TextToVideoZeroPipeline (line 1505) | class TextToVideoZeroPipeline(metaclass=DummyObject): method __init__ (line 1508) | def __init__(self, *args, **kwargs): method from_config (line 1512) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1516) | def from_pretrained(cls, *args, **kwargs): class TextToVideoZeroSDXLPipeline (line 1520) | class TextToVideoZeroSDXLPipeline(metaclass=DummyObject): method __init__ (line 1523) | def __init__(self, *args, **kwargs): method from_config (line 1527) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1531) | def from_pretrained(cls, *args, **kwargs): class UnCLIPImageVariationPipeline (line 1535) | class UnCLIPImageVariationPipeline(metaclass=DummyObject): method __init__ (line 1538) | def __init__(self, *args, **kwargs): method from_config (line 1542) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1546) | def from_pretrained(cls, *args, **kwargs): class UnCLIPPipeline (line 1550) | class UnCLIPPipeline(metaclass=DummyObject): method __init__ (line 1553) | def __init__(self, *args, **kwargs): method from_config (line 1557) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1561) | def from_pretrained(cls, *args, **kwargs): class UniDiffuserModel (line 1565) | class UniDiffuserModel(metaclass=DummyObject): method __init__ (line 1568) | def __init__(self, *args, **kwargs): method from_config (line 1572) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1576) | def from_pretrained(cls, *args, **kwargs): class UniDiffuserPipeline (line 1580) | class UniDiffuserPipeline(metaclass=DummyObject): method __init__ (line 1583) | def __init__(self, *args, **kwargs): method from_config (line 1587) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1591) | def from_pretrained(cls, *args, **kwargs): class UniDiffuserTextDecoder (line 1595) | class UniDiffuserTextDecoder(metaclass=DummyObject): method __init__ (line 1598) | def __init__(self, *args, **kwargs): method from_config (line 1602) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1606) | def from_pretrained(cls, *args, **kwargs): class VersatileDiffusionDualGuidedPipeline (line 1610) | class VersatileDiffusionDualGuidedPipeline(metaclass=DummyObject): method __init__ (line 1613) | def __init__(self, *args, **kwargs): method from_config (line 1617) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1621) | def from_pretrained(cls, *args, **kwargs): class VersatileDiffusionImageVariationPipeline (line 1625) | class VersatileDiffusionImageVariationPipeline(metaclass=DummyObject): method __init__ (line 1628) | def __init__(self, *args, **kwargs): method from_config (line 1632) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1636) | def from_pretrained(cls, *args, **kwargs): class VersatileDiffusionPipeline (line 1640) | class VersatileDiffusionPipeline(metaclass=DummyObject): method __init__ (line 1643) | def __init__(self, *args, **kwargs): method from_config (line 1647) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1651) | def from_pretrained(cls, *args, **kwargs): class VersatileDiffusionTextToImagePipeline (line 1655) | class VersatileDiffusionTextToImagePipeline(metaclass=DummyObject): method __init__ (line 1658) | def __init__(self, *args, **kwargs): method from_config (line 1662) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1666) | def from_pretrained(cls, *args, **kwargs): class VideoToVideoSDPipeline (line 1670) | class VideoToVideoSDPipeline(metaclass=DummyObject): method __init__ (line 1673) | def __init__(self, *args, **kwargs): method from_config (line 1677) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1681) | def from_pretrained(cls, *args, **kwargs): class VQDiffusionPipeline (line 1685) | class VQDiffusionPipeline(metaclass=DummyObject): method __init__ (line 1688) | def __init__(self, *args, **kwargs): method from_config (line 1692) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1696) | def from_pretrained(cls, *args, **kwargs): class WuerstchenCombinedPipeline (line 1700) | class WuerstchenCombinedPipeline(metaclass=DummyObject): method __init__ (line 1703) | def __init__(self, *args, **kwargs): method from_config (line 1707) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1711) | def from_pretrained(cls, *args, **kwargs): class WuerstchenDecoderPipeline (line 1715) | class WuerstchenDecoderPipeline(metaclass=DummyObject): method __init__ (line 1718) | def __init__(self, *args, **kwargs): method from_config (line 1722) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1726) | def from_pretrained(cls, *args, **kwargs): class WuerstchenPriorPipeline (line 1730) | class WuerstchenPriorPipeline(metaclass=DummyObject): method __init__ (line 1733) | def __init__(self, *args, **kwargs): method from_config (line 1737) | def from_config(cls, *args, **kwargs): method from_pretrained (line 1741) | def from_pretrained(cls, *args, **kwargs): FILE: diffusers/src/diffusers/utils/dummy_transformers_and_torch_and_note_seq_objects.py class SpectrogramDiffusionPipeline (line 5) | class SpectrogramDiffusionPipeline(metaclass=DummyObject): method __init__ (line 8) | def __init__(self, *args, **kwargs): method from_config (line 12) | def from_config(cls, *args, **kwargs): method from_pretrained (line 16) | def from_pretrained(cls, *args, **kwargs): FILE: diffusers/src/diffusers/utils/dynamic_modules_utils.py function get_diffusers_versions (line 42) | def get_diffusers_versions(): function init_hf_modules (line 48) | def init_hf_modules(): function create_dynamic_module (line 63) | def create_dynamic_module(name: Union[str, os.PathLike]): function get_relative_imports (line 78) | def get_relative_imports(module_file): function get_relative_import_files (line 96) | def get_relative_import_files(module_file): function check_imports (line 125) | def check_imports(filename): function get_class_in_module (line 157) | def get_class_in_module(class_name, module_path): function find_pipeline_class (line 169) | def find_pipeline_class(loaded_module): function get_cached_module_file (line 197) | def get_cached_module_file( function get_class_from_dynamic_module (line 379) | def get_class_from_dynamic_module( FILE: diffusers/src/diffusers/utils/export_utils.py function buffered_writer (line 25) | def buffered_writer(raw_f): function export_to_gif (line 31) | def export_to_gif(image: List[PIL.Image.Image], output_gif_path: str = N... function export_to_ply (line 46) | def export_to_ply(mesh, output_ply_path: str = None): function export_to_obj (line 98) | def export_to_obj(mesh, output_obj_path: str = None): function export_to_video (line 118) | def export_to_video( FILE: diffusers/src/diffusers/utils/hub_utils.py function http_user_agent (line 75) | def http_user_agent(user_agent: Union[Dict, str, None] = None) -> str: function load_or_create_model_card (line 99) | def load_or_create_model_card( function populate_model_card (line 169) | def populate_model_card(model_card: ModelCard, tags: Union[str, List[str... function extract_commit_hash (line 185) | def extract_commit_hash(resolved_file: Optional[str], commit_hash: Optio... function move_cache (line 209) | def move_cache(old_cache_dir: Optional[str] = None, new_cache_dir: Optio... function _add_variant (line 271) | def _add_variant(weights_name: str, variant: Optional[str] = None) -> str: function _get_model_file (line 281) | def _get_model_file( function _check_if_shards_exist_locally (line 404) | def _check_if_shards_exist_locally(local_dir, subfolder, original_shard_... function _get_checkpoint_shard_files (line 415) | def _get_checkpoint_shard_files( class PushToHubMixin (line 502) | class PushToHubMixin: method _upload_folder (line 507) | def _upload_folder( method push_to_hub (line 531) | def push_to_hub( FILE: diffusers/src/diffusers/utils/import_utils.py function is_timm_available (line 313) | def is_timm_available(): function is_torch_available (line 338) | def is_torch_available(): function is_torch_xla_available (line 342) | def is_torch_xla_available(): function is_torch_npu_available (line 346) | def is_torch_npu_available(): function is_flax_available (line 350) | def is_flax_available(): function is_transformers_available (line 354) | def is_transformers_available(): function is_inflect_available (line 358) | def is_inflect_available(): function is_unidecode_available (line 362) | def is_unidecode_available(): function is_onnx_available (line 366) | def is_onnx_available(): function is_opencv_available (line 370) | def is_opencv_available(): function is_scipy_available (line 374) | def is_scipy_available(): function is_librosa_available (line 378) | def is_librosa_available(): function is_xformers_available (line 382) | def is_xformers_available(): function is_accelerate_available (line 386) | def is_accelerate_available(): function is_k_diffusion_available (line 390) | def is_k_diffusion_available(): function is_note_seq_available (line 394) | def is_note_seq_available(): function is_wandb_available (line 398) | def is_wandb_available(): function is_tensorboard_available (line 402) | def is_tensorboard_available(): function is_compel_available (line 406) | def is_compel_available(): function is_ftfy_available (line 410) | def is_ftfy_available(): function is_bs4_available (line 414) | def is_bs4_available(): function is_torchsde_available (line 418) | def is_torchsde_available(): function is_invisible_watermark_available (line 422) | def is_invisible_watermark_available(): function is_peft_available (line 426) | def is_peft_available(): function is_torchvision_available (line 430) | def is_torchvision_available(): function is_matplotlib_available (line 434) | def is_matplotlib_available(): function is_safetensors_available (line 438) | def is_safetensors_available(): function is_bitsandbytes_available (line 442) | def is_bitsandbytes_available(): function is_notebook (line 446) | def is_notebook(): function is_google_colab (line 450) | def is_google_colab(): function requires_backends (line 603) | def requires_backends(obj, backends): class DummyObject (line 634) | class DummyObject(type): method __getattr__ (line 640) | def __getattr__(cls, key): function compare_versions (line 647) | def compare_versions(library_or_version: Union[str, Version], operation:... function is_torch_version (line 667) | def is_torch_version(operation: str, version: str): function is_transformers_version (line 679) | def is_transformers_version(operation: str, version: str): function is_accelerate_version (line 693) | def is_accelerate_version(operation: str, version: str): function is_peft_version (line 707) | def is_peft_version(operation: str, version: str): function is_k_diffusion_version (line 721) | def is_k_diffusion_version(operation: str, version: str): function get_objects_from_module (line 735) | def get_objects_from_module(module): class OptionalDependencyNotAvailable (line 755) | class OptionalDependencyNotAvailable(BaseException): class _LazyModule (line 759) | class _LazyModule(ModuleType): method __init__ (line 766) | def __init__(self, name, module_file, import_structure, module_spec=No... method __dir__ (line 783) | def __dir__(self): method __getattr__ (line 792) | def __getattr__(self, name: str) -> Any: method _get_module (line 806) | def _get_module(self, module_name: str): method __reduce__ (line 815) | def __reduce__(self): FILE: diffusers/src/diffusers/utils/loading_utils.py function load_image (line 9) | def load_image( FILE: diffusers/src/diffusers/utils/logging.py function _get_default_logging_level (line 52) | def _get_default_logging_level() -> int: function _get_library_name (line 69) | def _get_library_name() -> str: function _get_library_root_logger (line 73) | def _get_library_root_logger() -> logging.Logger: function _configure_library_root_logger (line 77) | def _configure_library_root_logger() -> None: function _reset_library_root_logger (line 96) | def _reset_library_root_logger() -> None: function get_log_levels_dict (line 109) | def get_log_levels_dict() -> Dict[str, int]: function get_logger (line 113) | def get_logger(name: Optional[str] = None) -> logging.Logger: function get_verbosity (line 127) | def get_verbosity() -> int: function set_verbosity (line 147) | def set_verbosity(verbosity: int) -> None: function set_verbosity_info (line 166) | def set_verbosity_info() -> None: function set_verbosity_warning (line 171) | def set_verbosity_warning() -> None: function set_verbosity_debug (line 176) | def set_verbosity_debug() -> None: function set_verbosity_error (line 181) | def set_verbosity_error() -> None: function disable_default_handler (line 186) | def disable_default_handler() -> None: function enable_default_handler (line 195) | def enable_default_handler() -> None: function add_handler (line 204) | def add_handler(handler: logging.Handler) -> None: function remove_handler (line 213) | def remove_handler(handler: logging.Handler) -> None: function disable_propagation (line 222) | def disable_propagation() -> None: function enable_propagation (line 231) | def enable_propagation() -> None: function enable_explicit_format (line 241) | def enable_explicit_format() -> None: function reset_format (line 256) | def reset_format() -> None: function warning_advice (line 268) | def warning_advice(self, *args, **kwargs) -> None: class EmptyTqdm (line 282) | class EmptyTqdm: method __init__ (line 285) | def __init__(self, *args, **kwargs): # pylint: disable=unused-argument method __iter__ (line 288) | def __iter__(self): method __getattr__ (line 291) | def __getattr__(self, _): method __enter__ (line 299) | def __enter__(self): method __exit__ (line 302) | def __exit__(self, type_, value, traceback): class _tqdm_cls (line 306) | class _tqdm_cls: method __call__ (line 307) | def __call__(self, *args, **kwargs): method set_lock (line 313) | def set_lock(self, *args, **kwargs): method get_lock (line 318) | def get_lock(self): function is_progress_bar_enabled (line 326) | def is_progress_bar_enabled() -> bool: function enable_progress_bar (line 332) | def enable_progress_bar() -> None: function disable_progress_bar (line 338) | def disable_progress_bar() -> None: FILE: diffusers/src/diffusers/utils/outputs.py function is_tensor (line 27) | def is_tensor(x) -> bool: class BaseOutput (line 40) | class BaseOutput(OrderedDict): method __init_subclass__ (line 54) | def __init_subclass__(cls) -> None: method __post_init__ (line 76) | def __post_init__(self) -> None: method __delitem__ (line 95) | def __delitem__(self, *args, **kwargs): method setdefault (line 98) | def setdefault(self, *args, **kwargs): method pop (line 101) | def pop(self, *args, **kwargs): method update (line 104) | def update(self, *args, **kwargs): method __getitem__ (line 107) | def __getitem__(self, k: Any) -> Any: method __setattr__ (line 114) | def __setattr__(self, name: Any, value: Any) -> None: method __setitem__ (line 120) | def __setitem__(self, key, value): method __reduce__ (line 126) | def __reduce__(self): method to_tuple (line 133) | def to_tuple(self) -> Tuple[Any, ...]: FILE: diffusers/src/diffusers/utils/peft_utils.py function recurse_remove_peft_layers (line 31) | def recurse_remove_peft_layers(model): function scale_lora_layers (line 103) | def scale_lora_layers(model, weight): function unscale_lora_layers (line 123) | def unscale_lora_layers(model, weight: Optional[float] = None): function get_peft_kwargs (line 150) | def get_peft_kwargs(rank_dict, network_alpha_dict, peft_state_dict, is_u... function get_adapter_name (line 195) | def get_adapter_name(model): function set_adapter_layers (line 204) | def set_adapter_layers(model, enabled=True): function delete_adapter_layers (line 216) | def delete_adapter_layers(model, adapter_name): function set_weights_and_activate_adapters (line 238) | def set_weights_and_activate_adapters(model, adapter_names, weights): function check_peft_version (line 278) | def check_peft_version(min_version: str) -> None: FILE: diffusers/src/diffusers/utils/pil_utils.py function pt_to_pil (line 27) | def pt_to_pil(images): function numpy_to_pil (line 37) | def numpy_to_pil(images): function make_image_grid (line 53) | def make_image_grid(images: List[PIL.Image.Image], rows: int, cols: int,... FILE: diffusers/src/diffusers/utils/state_dict_utils.py class StateDictType (line 26) | class StateDictType(enum.Enum): function convert_state_dict (line 141) | def convert_state_dict(state_dict, mapping): function convert_state_dict_to_peft (line 174) | def convert_state_dict_to_peft(state_dict, original_type=None, **kwargs): function convert_state_dict_to_diffusers (line 201) | def convert_state_dict_to_diffusers(state_dict, original_type=None, **kw... function convert_unet_state_dict_to_peft (line 248) | def convert_unet_state_dict_to_peft(state_dict): function convert_all_state_dict_to_peft (line 256) | def convert_all_state_dict_to_peft(state_dict): function convert_state_dict_to_kohya (line 275) | def convert_state_dict_to_kohya(state_dict, original_type=None, **kwargs): FILE: diffusers/src/diffusers/utils/testing_utils.py function torch_all_close (line 94) | def torch_all_close(a, b, *args, **kwargs): function numpy_cosine_similarity_distance (line 102) | def numpy_cosine_similarity_distance(a, b): function print_tensor_test (line 109) | def print_tensor_test( function get_tests_dir (line 135) | def get_tests_dir(append_path=None): function str_to_bool (line 158) | def str_to_bool(value) -> int: function parse_flag_from_env (line 172) | def parse_flag_from_env(key, default=False): function floats_tensor (line 192) | def floats_tensor(shape, scale=1.0, rng=None, name=None): function slow (line 208) | def slow(test_case): function nightly (line 218) | def nightly(test_case): function require_torch (line 228) | def require_torch(test_case): function require_torch_2 (line 235) | def require_torch_2(test_case): function require_torch_gpu (line 244) | def require_torch_gpu(test_case): function require_torch_accelerator (line 252) | def require_torch_accelerator(test_case): function require_torch_multi_gpu (line 259) | def require_torch_multi_gpu(test_case): function require_torch_accelerator_with_fp16 (line 273) | def require_torch_accelerator_with_fp16(test_case): function require_torch_accelerator_with_fp64 (line 280) | def require_torch_accelerator_with_fp64(test_case): function require_torch_accelerator_with_training (line 287) | def require_torch_accelerator_with_training(test_case): function skip_mps (line 295) | def skip_mps(test_case): function require_flax (line 300) | def require_flax(test_case): function require_compel (line 307) | def require_compel(test_case): function require_onnxruntime (line 315) | def require_onnxruntime(test_case): function require_note_seq (line 322) | def require_note_seq(test_case): function require_torchsde (line 329) | def require_torchsde(test_case): function require_peft_backend (line 336) | def require_peft_backend(test_case): function require_timm (line 344) | def require_timm(test_case): function require_peft_version_greater (line 351) | def require_peft_version_greater(peft_version): function require_accelerate_version_greater (line 368) | def require_accelerate_version_greater(accelerate_version): function deprecate_after_peft_backend (line 380) | def deprecate_after_peft_backend(test_case): function get_python_version (line 387) | def get_python_version(): function require_python39_or_higher (line 393) | def require_python39_or_higher(test_case): function load_numpy (line 401) | def load_numpy(arry: Union[str, np.ndarray], local_path: Optional[str] =... function load_pt (line 427) | def load_pt(url: str): function load_image (line 434) | def load_image(image: Union[str, PIL.Image.Image]) -> PIL.Image.Image: function preprocess_image (line 465) | def preprocess_image(image: PIL.Image, batch_size: int): function export_to_gif (line 475) | def export_to_gif(image: List[PIL.Image.Image], output_gif_path: str = N... function buffered_writer (line 491) | def buffered_writer(raw_f): function export_to_ply (line 497) | def export_to_ply(mesh, output_ply_path: str = None): function export_to_obj (line 549) | def export_to_obj(mesh, output_obj_path: str = None): function export_to_video (line 569) | def export_to_video(video_frames: List[np.ndarray], output_video_path: s... function load_hf_numpy (line 586) | def load_hf_numpy(path) -> np.ndarray: function pytest_addoption_shared (line 601) | def pytest_addoption_shared(parser): function pytest_terminal_summary_main (line 620) | def pytest_terminal_summary_main(tr, id): function is_flaky (line 750) | def is_flaky(max_attempts: int = 5, wait_before_retry: Optional[float] =... function run_test_in_subprocess (line 787) | def run_test_in_subprocess(test_case, target_func, inputs=None, timeout=... class CaptureLogger (line 830) | class CaptureLogger: method __init__ (line 851) | def __init__(self, logger): method __enter__ (line 857) | def __enter__(self): method __exit__ (line 861) | def __exit__(self, *exc): method __repr__ (line 865) | def __repr__(self): function enable_full_determinism (line 869) | def enable_full_determinism(): function disable_full_determinism (line 887) | def disable_full_determinism(): function _is_torch_fp16_available (line 894) | def _is_torch_fp16_available(device): function _is_torch_fp64_available (line 916) | def _is_torch_fp64_available(device): function _device_agnostic_dispatch (line 950) | def _device_agnostic_dispatch(device: str, dispatch_table: Dict[str, Cal... function backend_manual_seed (line 965) | def backend_manual_seed(device: str, seed: int): function backend_empty_cache (line 969) | def backend_empty_cache(device: str): function backend_device_count (line 973) | def backend_device_count(device: str): function backend_supports_training (line 979) | def backend_supports_training(device: str): function update_mapping_from_spec (line 992) | def update_mapping_from_spec(device_fn_dict: Dict[str, Callable], attrib... FILE: diffusers/src/diffusers/utils/torch_utils.py function maybe_allow_in_graph (line 34) | def maybe_allow_in_graph(cls): function randn_tensor (line 38) | def randn_tensor( function is_compiled_module (line 86) | def is_compiled_module(module) -> bool: function fourier_filter (line 93) | def fourier_filter(x_in: "torch.Tensor", threshold: int, scale: int) -> ... function apply_freeu (line 124) | def apply_freeu( FILE: diffusers/src/diffusers/utils/versions.py function _compare_versions (line 37) | def _compare_versions(op, got_ver, want_ver, requirement, pkg, hint): function require_version (line 49) | def require_version(requirement: str, hint: Optional[str] = None) -> None: function require_version_core (line 114) | def require_version_core(requirement): FILE: diffusers/src/diffusers/video_processor.py class VideoProcessor (line 25) | class VideoProcessor(VaeImageProcessor): method preprocess_video (line 28) | def preprocess_video(self, video, height: Optional[int] = None, width:... method postprocess_video (line 89) | def postprocess_video( FILE: diffusers/tests/conftest.py function pytest_addoption (line 33) | def pytest_addoption(parser): function pytest_terminal_summary (line 39) | def pytest_terminal_summary(terminalreporter): FILE: diffusers/tests/fixtures/custom_pipeline/pipeline.py class CustomLocalPipeline (line 24) | class CustomLocalPipeline(DiffusionPipeline): method __init__ (line 36) | def __init__(self, unet, scheduler): method __call__ (line 41) | def __call__( FILE: diffusers/tests/fixtures/custom_pipeline/what_ever.py class CustomLocalPipeline (line 24) | class CustomLocalPipeline(DiffusionPipeline): method __init__ (line 36) | def __init__(self, unet, scheduler): method __call__ (line 41) | def __call__( FILE: diffusers/tests/lora/test_lora_layers_sd.py class StableDiffusionLoRATests (line 54) | class StableDiffusionLoRATests(PeftLoraLoaderMixinTests, unittest.TestCa... method setUp (line 84) | def setUp(self): method tearDown (line 89) | def tearDown(self): method test_integration_move_lora_cpu (line 98) | def test_integration_move_lora_cpu(self): method test_integration_move_lora_dora_cpu (line 154) | def test_integration_move_lora_dora_cpu(self): class LoraIntegrationTests (line 205) | class LoraIntegrationTests(unittest.TestCase): method setUp (line 206) | def setUp(self): method tearDown (line 211) | def tearDown(self): method test_integration_logits_with_scale (line 216) | def test_integration_logits_with_scale(self): method test_integration_logits_no_scale (line 248) | def test_integration_logits_no_scale(self): method test_dreambooth_old_format (line 275) | def test_dreambooth_old_format(self): method test_dreambooth_text_encoder_new_format (line 299) | def test_dreambooth_text_encoder_new_format(self): method test_a1111 (line 322) | def test_a1111(self): method test_lycoris (line 345) | def test_lycoris(self): method test_a1111_with_model_cpu_offload (line 368) | def test_a1111_with_model_cpu_offload(self): method test_a1111_with_sequential_cpu_offload (line 390) | def test_a1111_with_sequential_cpu_offload(self): method test_kohya_sd_v15_with_higher_dimensions (line 412) | def test_kohya_sd_v15_with_higher_dimensions(self): method test_vanilla_funetuning (line 435) | def test_vanilla_funetuning(self): method test_unload_kohya_lora (line 458) | def test_unload_kohya_lora(self): method test_load_unload_load_kohya_lora (line 493) | def test_load_unload_load_kohya_lora(self): method test_not_empty_state_dict (line 541) | def test_not_empty_state_dict(self): method test_load_unload_load_state_dict (line 555) | def test_load_unload_load_state_dict(self): method test_sdv1_5_lcm_lora (line 575) | def test_sdv1_5_lcm_lora(self): method test_sdv1_5_lcm_lora_img2img (line 603) | def test_sdv1_5_lcm_lora_img2img(self): method test_sd_load_civitai_empty_network_alpha (line 640) | def test_sd_load_civitai_empty_network_alpha(self): FILE: diffusers/tests/lora/test_lora_layers_sd3.py class SD3LoRATests (line 43) | class SD3LoRATests(unittest.TestCase): method get_dummy_components (line 46) | def get_dummy_components(self): method get_dummy_inputs (line 115) | def get_dummy_inputs(self, device, seed=0): method get_lora_config_for_transformer (line 130) | def get_lora_config_for_transformer(self): method test_simple_inference_with_transformer_lora_save_load (line 140) | def test_simple_inference_with_transformer_lora_save_load(self): method test_simple_inference_with_transformer_lora_and_scale (line 176) | def test_simple_inference_with_transformer_lora_and_scale(self): FILE: diffusers/tests/lora/test_lora_layers_sdxl.py class StableDiffusionXLLoRATests (line 56) | class StableDiffusionXLLoRATests(PeftLoraLoaderMixinTests, unittest.Test... method setUp (line 93) | def setUp(self): method tearDown (line 98) | def tearDown(self): class LoraSDXLIntegrationTests (line 107) | class LoraSDXLIntegrationTests(unittest.TestCase): method setUp (line 108) | def setUp(self): method tearDown (line 113) | def tearDown(self): method test_sdxl_0_9_lora_one (line 118) | def test_sdxl_0_9_lora_one(self): method test_sdxl_0_9_lora_two (line 139) | def test_sdxl_0_9_lora_two(self): method test_sdxl_0_9_lora_three (line 161) | def test_sdxl_0_9_lora_three(self): method test_sdxl_1_0_lora (line 183) | def test_sdxl_1_0_lora(self): method test_sdxl_1_0_blockwise_lora (line 205) | def test_sdxl_1_0_blockwise_lora(self): method test_sdxl_lcm_lora (line 235) | def test_sdxl_lcm_lora(self): method test_sdxl_1_0_lora_fusion (line 265) | def test_sdxl_1_0_lora_fusion(self): method test_sdxl_1_0_lora_unfusion (line 293) | def test_sdxl_1_0_lora_unfusion(self): method test_sdxl_1_0_lora_unfusion_effectivity (line 321) | def test_sdxl_1_0_lora_unfusion_effectivity(self): method test_sdxl_1_0_lora_fusion_efficiency (line 357) | def test_sdxl_1_0_lora_fusion_efficiency(self): method test_sdxl_1_0_last_ben (line 402) | def test_sdxl_1_0_last_ben(self): method test_sdxl_1_0_fuse_unfuse_all (line 422) | def test_sdxl_1_0_fuse_unfuse_all(self): method test_sdxl_1_0_lora_with_sequential_cpu_offloading (line 473) | def test_sdxl_1_0_lora_with_sequential_cpu_offloading(self): method test_controlnet_canny_lora (line 496) | def test_controlnet_canny_lora(self): method test_sdxl_t2i_adapter_canny_lora (line 523) | def test_sdxl_t2i_adapter_canny_lora(self): method test_sequential_fuse_unfuse (line 552) | def test_sequential_fuse_unfuse(self): method test_integration_logits_multi_adapter (line 596) | def test_integration_logits_multi_adapter(self): method test_integration_logits_for_dora_lora (line 674) | def test_integration_logits_for_dora_lora(self): FILE: diffusers/tests/lora/utils.py function state_dicts_almost_equal (line 46) | def state_dicts_almost_equal(sd1, sd2): function check_if_lora_correctly_set (line 58) | def check_if_lora_correctly_set(model) -> bool: class PeftLoraLoaderMixinTests (line 69) | class PeftLoraLoaderMixinTests: method get_dummy_components (line 77) | def get_dummy_components(self, scheduler_cls=None, use_dora=False): method get_dummy_inputs (line 138) | def get_dummy_inputs(self, with_generator=True): method get_dummy_tokens (line 160) | def get_dummy_tokens(self): method test_simple_inference (line 169) | def test_simple_inference(self): method test_simple_inference_with_text_lora (line 183) | def test_simple_inference_with_text_lora(self): method test_simple_inference_with_text_lora_and_scale (line 212) | def test_simple_inference_with_text_lora_and_scale(self): method test_simple_inference_with_text_lora_fused (line 257) | def test_simple_inference_with_text_lora_fused(self): method test_simple_inference_with_text_lora_unloaded (line 295) | def test_simple_inference_with_text_lora_unloaded(self): method test_simple_inference_with_text_lora_save_load (line 337) | def test_simple_inference_with_text_lora_save_load(self): method test_simple_inference_save_pretrained (line 398) | def test_simple_inference_save_pretrained(self): method test_simple_inference_with_text_unet_lora_save_load (line 447) | def test_simple_inference_with_text_unet_lora_save_load(self): method test_simple_inference_with_text_unet_lora_and_scale (line 515) | def test_simple_inference_with_text_unet_lora_and_scale(self): method test_simple_inference_with_text_lora_unet_fused (line 567) | def test_simple_inference_with_text_lora_unet_fused(self): method test_simple_inference_with_text_unet_lora_unloaded (line 609) | def test_simple_inference_with_text_unet_lora_unloaded(self): method test_simple_inference_with_text_unet_lora_unfused (line 654) | def test_simple_inference_with_text_unet_lora_unfused(self): method test_simple_inference_with_text_unet_multi_adapter (line 700) | def test_simple_inference_with_text_unet_multi_adapter(self): method test_simple_inference_with_text_unet_block_scale (line 766) | def test_simple_inference_with_text_unet_block_scale(self): method test_simple_inference_with_text_unet_multi_adapter_block_lora (line 821) | def test_simple_inference_with_text_unet_multi_adapter_block_lora(self): method test_simple_inference_with_text_unet_block_scale_for_all_dict_options (line 893) | def test_simple_inference_with_text_unet_block_scale_for_all_dict_opti... method test_simple_inference_with_text_unet_multi_adapter_delete_adapter (line 978) | def test_simple_inference_with_text_unet_multi_adapter_delete_adapter(... method test_simple_inference_with_text_unet_multi_adapter_weighted (line 1066) | def test_simple_inference_with_text_unet_multi_adapter_weighted(self): method test_lora_fuse_nan (line 1141) | def test_lora_fuse_nan(self): method test_get_adapters (line 1173) | def test_get_adapters(self): method test_get_list_adapters (line 1200) | def test_get_list_adapters(self): method test_simple_inference_with_text_lora_unet_fused_multi (line 1238) | def test_simple_inference_with_text_lora_unet_fused_multi(self): method test_simple_inference_with_dora (line 1298) | def test_simple_inference_with_dora(self): method test_simple_inference_with_text_unet_lora_unfused_torch_compile (line 1329) | def test_simple_inference_with_text_unet_lora_unfused_torch_compile(se... method test_modify_padding_mode (line 1362) | def test_modify_padding_mode(self): FILE: diffusers/tests/models/autoencoders/test_models_vae.py function get_autoencoder_kl_config (line 55) | def get_autoencoder_kl_config(block_out_channels=None, norm_num_groups=N... function get_asym_autoencoder_kl_config (line 70) | def get_asym_autoencoder_kl_config(block_out_channels=None, norm_num_gro... function get_autoencoder_tiny_config (line 91) | def get_autoencoder_tiny_config(block_out_channels=None): function get_consistency_vae_config (line 104) | def get_consistency_vae_config(block_out_channels=None, norm_num_groups=... class AutoencoderKLTests (line 131) | class AutoencoderKLTests(ModelTesterMixin, UNetTesterMixin, unittest.Tes... method dummy_input (line 137) | def dummy_input(self): method input_shape (line 147) | def input_shape(self): method output_shape (line 151) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 154) | def prepare_init_args_and_inputs_for_common(self): method test_forward_signature (line 159) | def test_forward_signature(self): method test_training (line 162) | def test_training(self): method test_gradient_checkpointing (line 166) | def test_gradient_checkpointing(self): method test_from_pretrained_hub (line 206) | def test_from_pretrained_hub(self): method test_output_pretrained (line 216) | def test_output_pretrained(self): class AsymmetricAutoencoderKLTests (line 289) | class AsymmetricAutoencoderKLTests(ModelTesterMixin, UNetTesterMixin, un... method dummy_input (line 295) | def dummy_input(self): method input_shape (line 306) | def input_shape(self): method output_shape (line 310) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 313) | def prepare_init_args_and_inputs_for_common(self): method test_forward_signature (line 318) | def test_forward_signature(self): method test_forward_with_norm_groups (line 321) | def test_forward_with_norm_groups(self): class AutoencoderTinyTests (line 325) | class AutoencoderTinyTests(ModelTesterMixin, unittest.TestCase): method dummy_input (line 331) | def dummy_input(self): method input_shape (line 341) | def input_shape(self): method output_shape (line 345) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 348) | def prepare_init_args_and_inputs_for_common(self): method test_outputs_equivalence (line 353) | def test_outputs_equivalence(self): class ConsistencyDecoderVAETests (line 357) | class ConsistencyDecoderVAETests(ModelTesterMixin, unittest.TestCase): method inputs_dict (line 363) | def inputs_dict(self, seed=None): method input_shape (line 372) | def input_shape(self): method output_shape (line 376) | def output_shape(self): method init_dict (line 380) | def init_dict(self): method prepare_init_args_and_inputs_for_common (line 383) | def prepare_init_args_and_inputs_for_common(self): method test_training (line 387) | def test_training(self): method test_ema_training (line 391) | def test_ema_training(self): class AutoencoderKLTemporalDecoderFastTests (line 395) | class AutoencoderKLTemporalDecoderFastTests(ModelTesterMixin, unittest.T... method dummy_input (line 401) | def dummy_input(self): method input_shape (line 412) | def input_shape(self): method output_shape (line 416) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 419) | def prepare_init_args_and_inputs_for_common(self): method test_forward_signature (line 431) | def test_forward_signature(self): method test_training (line 434) | def test_training(self): method test_gradient_checkpointing (line 438) | def test_gradient_checkpointing(self): class AutoencoderTinyIntegrationTests (line 483) | class AutoencoderTinyIntegrationTests(unittest.TestCase): method tearDown (line 484) | def tearDown(self): method get_file_format (line 490) | def get_file_format(self, seed, shape): method get_sd_image (line 493) | def get_sd_image(self, seed=0, shape=(4, 3, 512, 512), fp16=False): method get_sd_vae_model (line 498) | def get_sd_vae_model(self, model_id="hf-internal-testing/taesd-diffuse... method test_tae_tiling (line 514) | def test_tae_tiling(self, in_shape, out_shape): method test_stable_diffusion (line 522) | def test_stable_diffusion(self): method test_tae_roundtrip (line 537) | def test_tae_roundtrip(self, enable_tiling): class AutoencoderKLIntegrationTests (line 560) | class AutoencoderKLIntegrationTests(unittest.TestCase): method get_file_format (line 561) | def get_file_format(self, seed, shape): method tearDown (line 564) | def tearDown(self): method get_sd_image (line 570) | def get_sd_image(self, seed=0, shape=(4, 3, 512, 512), fp16=False): method get_sd_vae_model (line 575) | def get_sd_vae_model(self, model_id="CompVis/stable-diffusion-v1-4", f... method get_generator (line 589) | def get_generator(self, seed=0): method test_stable_diffusion (line 611) | def test_stable_diffusion(self, seed, expected_slice, expected_slice_m... method test_stable_diffusion_fp16 (line 635) | def test_stable_diffusion_fp16(self, seed, expected_slice): method test_stable_diffusion_mode (line 666) | def test_stable_diffusion_mode(self, seed, expected_slice, expected_sl... method test_stable_diffusion_decode (line 690) | def test_stable_diffusion_decode(self, seed, expected_slice): method test_stable_diffusion_decode_fp16 (line 713) | def test_stable_diffusion_decode_fp16(self, seed, expected_slice): method test_stable_diffusion_decode_xformers_vs_2_0_fp16 (line 733) | def test_stable_diffusion_decode_xformers_vs_2_0_fp16(self, seed): method test_stable_diffusion_decode_xformers_vs_2_0 (line 754) | def test_stable_diffusion_decode_xformers_vs_2_0(self, seed): method test_stable_diffusion_encode_sample (line 777) | def test_stable_diffusion_encode_sample(self, seed, expected_slice): class AsymmetricAutoencoderKLIntegrationTests (line 796) | class AsymmetricAutoencoderKLIntegrationTests(unittest.TestCase): method get_file_format (line 797) | def get_file_format(self, seed, shape): method tearDown (line 800) | def tearDown(self): method get_sd_image (line 806) | def get_sd_image(self, seed=0, shape=(4, 3, 512, 512), fp16=False): method get_sd_vae_model (line 811) | def get_sd_vae_model(self, model_id="cross-attention/asymmetric-autoen... method get_generator (line 824) | def get_generator(self, seed=0): method test_stable_diffusion (line 846) | def test_stable_diffusion(self, seed, expected_slice, expected_slice_m... method test_stable_diffusion_mode (line 877) | def test_stable_diffusion_mode(self, seed, expected_slice, expected_sl... method test_stable_diffusion_decode (line 901) | def test_stable_diffusion_decode(self, seed, expected_slice): method test_stable_diffusion_decode_xformers_vs_2_0 (line 921) | def test_stable_diffusion_decode_xformers_vs_2_0(self, seed): method test_stable_diffusion_encode_sample (line 944) | def test_stable_diffusion_encode_sample(self, seed, expected_slice): class ConsistencyDecoderVAEIntegrationTests (line 963) | class ConsistencyDecoderVAEIntegrationTests(unittest.TestCase): method setUp (line 964) | def setUp(self): method tearDown (line 970) | def tearDown(self): method test_encode_decode (line 977) | def test_encode_decode(self): method test_sd (line 998) | def test_sd(self): method test_encode_decode_f16 (line 1015) | def test_encode_decode_f16(self): method test_sd_f16 (line 1043) | def test_sd_f16(self): method test_vae_tiling (line 1070) | def test_vae_tiling(self): FILE: diffusers/tests/models/autoencoders/test_models_vae_flax.py class FlaxAutoencoderKLTests (line 15) | class FlaxAutoencoderKLTests(FlaxModelTesterMixin, unittest.TestCase): method dummy_input (line 19) | def dummy_input(self): method prepare_init_args_and_inputs_for_common (line 29) | def prepare_init_args_and_inputs_for_common(self): FILE: diffusers/tests/models/autoencoders/test_models_vq.py class VQModelTests (line 34) | class VQModelTests(ModelTesterMixin, UNetTesterMixin, unittest.TestCase): method dummy_input (line 39) | def dummy_input(self, sizes=(32, 32)): method input_shape (line 48) | def input_shape(self): method output_shape (line 52) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 55) | def prepare_init_args_and_inputs_for_common(self): method test_forward_signature (line 68) | def test_forward_signature(self): method test_training (line 71) | def test_training(self): method test_from_pretrained_hub (line 74) | def test_from_pretrained_hub(self): method test_output_pretrained (line 84) | def test_output_pretrained(self): method test_loss_pretrained (line 102) | def test_loss_pretrained(self): FILE: diffusers/tests/models/test_activations.py class ActivationsTests (line 9) | class ActivationsTests(unittest.TestCase): method test_swish (line 10) | def test_swish(self): method test_silu (line 20) | def test_silu(self): method test_mish (line 30) | def test_mish(self): method test_gelu (line 40) | def test_gelu(self): FILE: diffusers/tests/models/test_attention_processor.py class AttnAddedKVProcessorTests (line 11) | class AttnAddedKVProcessorTests(unittest.TestCase): method get_constructor_arguments (line 12) | def get_constructor_arguments(self, only_cross_attention: bool = False): method get_forward_arguments (line 32) | def get_forward_arguments(self, query_dim, added_kv_proj_dim): method test_only_cross_attention (line 45) | def test_only_cross_attention(self): class DeprecatedAttentionBlockTests (line 81) | class DeprecatedAttentionBlockTests(unittest.TestCase): method test_conversion_when_using_device_map (line 82) | def test_conversion_when_using_device_map(self): FILE: diffusers/tests/models/test_layers_utils.py class EmbeddingsTests (line 34) | class EmbeddingsTests(unittest.TestCase): method test_timestep_embeddings (line 35) | def test_timestep_embeddings(self): method test_timestep_defaults (line 58) | def test_timestep_defaults(self): method test_timestep_flip_sin_cos (line 69) | def test_timestep_flip_sin_cos(self): method test_timestep_downscale_freq_shift (line 80) | def test_timestep_downscale_freq_shift(self): method test_sinoid_embeddings_hardcoded (line 93) | def test_sinoid_embeddings_hardcoded(self): class Upsample2DBlockTests (line 121) | class Upsample2DBlockTests(unittest.TestCase): method test_upsample_default (line 122) | def test_upsample_default(self): method test_upsample_with_conv (line 134) | def test_upsample_with_conv(self): method test_upsample_with_conv_out_dim (line 146) | def test_upsample_with_conv_out_dim(self): method test_upsample_with_transpose (line 158) | def test_upsample_with_transpose(self): class Downsample2DBlockTests (line 171) | class Downsample2DBlockTests(unittest.TestCase): method test_downsample_default (line 172) | def test_downsample_default(self): method test_downsample_with_conv (line 186) | def test_downsample_with_conv(self): method test_downsample_with_conv_pad1 (line 201) | def test_downsample_with_conv_pad1(self): method test_downsample_with_conv_out_dim (line 213) | def test_downsample_with_conv_out_dim(self): class ResnetBlock2DTests (line 226) | class ResnetBlock2DTests(unittest.TestCase): method test_resnet_default (line 227) | def test_resnet_default(self): method test_restnet_with_use_in_shortcut (line 242) | def test_restnet_with_use_in_shortcut(self): method test_resnet_up (line 257) | def test_resnet_up(self): method test_resnet_down (line 272) | def test_resnet_down(self): method test_restnet_with_kernel_fir (line 287) | def test_restnet_with_kernel_fir(self): method test_restnet_with_kernel_sde_vp (line 302) | def test_restnet_with_kernel_sde_vp(self): class Transformer2DModelTests (line 318) | class Transformer2DModelTests(unittest.TestCase): method test_spatial_transformer_default (line 319) | def test_spatial_transformer_default(self): method test_spatial_transformer_cross_attention_dim (line 342) | def test_spatial_transformer_cross_attention_dim(self): method test_spatial_transformer_timestep (line 365) | def test_spatial_transformer_timestep(self): method test_spatial_transformer_dropout (line 402) | def test_spatial_transformer_dropout(self): method test_spatial_transformer_discrete (line 430) | def test_spatial_transformer_discrete(self): method test_spatial_transformer_default_norm_layers (line 458) | def test_spatial_transformer_default_norm_layers(self): method test_spatial_transformer_ada_norm_layers (line 464) | def test_spatial_transformer_ada_norm_layers(self): method test_spatial_transformer_default_ff_layers (line 475) | def test_spatial_transformer_default_ff_layers(self): method test_spatial_transformer_geglu_approx_ff_layers (line 498) | def test_spatial_transformer_geglu_approx_ff_layers(self): method test_spatial_transformer_attention_bias (line 521) | def test_spatial_transformer_attention_bias(self): FILE: diffusers/tests/models/test_modeling_common.py function _test_from_save_pretrained_dynamo (line 59) | def _test_from_save_pretrained_dynamo(in_queue, out_queue, timeout): class ModelUtilsTest (line 82) | class ModelUtilsTest(unittest.TestCase): method tearDown (line 83) | def tearDown(self): method test_accelerate_loading_error_message (line 86) | def test_accelerate_loading_error_message(self): method test_cached_files_are_used_when_no_internet (line 93) | def test_cached_files_are_used_when_no_internet(self): method test_one_request_upon_cached (line 117) | def test_one_request_upon_cached(self): method test_weight_overwrite (line 152) | def test_weight_overwrite(self): class UNetTesterMixin (line 177) | class UNetTesterMixin: method test_forward_signature (line 178) | def test_forward_signature(self): method test_forward_with_norm_groups (line 189) | def test_forward_with_norm_groups(self): class ModelTesterMixin (line 210) | class ModelTesterMixin: method check_device_map_is_respected (line 216) | def check_device_map_is_respected(self, model, device_map): method test_from_save_pretrained (line 230) | def test_from_save_pretrained(self, expected_max_diff=5e-5): method test_getattr_is_correct (line 269) | def test_getattr_is_correct(self): method test_set_torch_npu_flash_attn_processor_determinism (line 316) | def test_set_torch_npu_flash_attn_processor_determinism(self): method test_set_xformers_attn_processor_for_determinism (line 363) | def test_set_xformers_attn_processor_for_determinism(self): method test_set_attn_processor_for_determinism (line 407) | def test_set_attn_processor_for_determinism(self): method test_from_save_pretrained_variant (line 459) | def test_from_save_pretrained_variant(self, expected_max_diff=5e-5): method test_from_save_pretrained_dynamo (line 512) | def test_from_save_pretrained_dynamo(self): method test_from_save_pretrained_dtype (line 517) | def test_from_save_pretrained_dtype(self): method test_determinism (line 535) | def test_determinism(self, expected_max_diff=1e-5): method test_output (line 566) | def test_output(self, expected_output_shape=None): method test_model_from_pretrained (line 589) | def test_model_from_pretrained(self): method test_training (line 624) | def test_training(self): method test_ema_training (line 641) | def test_ema_training(self): method test_outputs_equivalence (line 660) | def test_outputs_equivalence(self): method test_enable_disable_gradient_checkpointing (line 712) | def test_enable_disable_gradient_checkpointing(self): method test_deprecated_kwargs (line 730) | def test_deprecated_kwargs(self): method test_cpu_offload (line 751) | def test_cpu_offload(self): method test_disk_offload_without_safetensors (line 781) | def test_disk_offload_without_safetensors(self): method test_disk_offload_with_safetensors (line 815) | def test_disk_offload_with_safetensors(self): method test_model_parallelism (line 843) | def test_model_parallelism(self): method test_sharded_checkpoints (line 874) | def test_sharded_checkpoints(self): class ModelPushToHubTester (line 910) | class ModelPushToHubTester(unittest.TestCase): method test_push_to_hub (line 915) | def test_push_to_hub(self): method test_push_to_hub_in_organization (line 946) | def test_push_to_hub_in_organization(self): method test_push_to_hub_library_name (line 981) | def test_push_to_hub_library_name(self): FILE: diffusers/tests/models/test_modeling_common_flax.py class FlaxModelTesterMixin (line 12) | class FlaxModelTesterMixin: method test_output (line 13) | def test_output(self): method test_forward_with_norm_groups (line 29) | def test_forward_with_norm_groups(self): method test_deprecated_kwargs (line 48) | def test_deprecated_kwargs(self): FILE: diffusers/tests/models/transformers/test_models_dit_transformer2d.py class DiTTransformer2DModelTests (line 34) | class DiTTransformer2DModelTests(ModelTesterMixin, unittest.TestCase): method dummy_input (line 39) | def dummy_input(self): method input_shape (line 53) | def input_shape(self): method output_shape (line 57) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 60) | def prepare_init_args_and_inputs_for_common(self): method test_output (line 77) | def test_output(self): method test_correct_class_remapping_from_dict_config (line 82) | def test_correct_class_remapping_from_dict_config(self): method test_correct_class_remapping_from_pretrained_config (line 87) | def test_correct_class_remapping_from_pretrained_config(self): method test_correct_class_remapping (line 93) | def test_correct_class_remapping(self): FILE: diffusers/tests/models/transformers/test_models_pixart_transformer2d.py class PixArtTransformer2DModelTests (line 34) | class PixArtTransformer2DModelTests(ModelTesterMixin, unittest.TestCase): method dummy_input (line 41) | def dummy_input(self): method input_shape (line 61) | def input_shape(self): method output_shape (line 65) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 68) | def prepare_init_args_and_inputs_for_common(self): method test_output (line 90) | def test_output(self): method test_correct_class_remapping_from_dict_config (line 95) | def test_correct_class_remapping_from_dict_config(self): method test_correct_class_remapping_from_pretrained_config (line 100) | def test_correct_class_remapping_from_pretrained_config(self): method test_correct_class_remapping (line 106) | def test_correct_class_remapping(self): FILE: diffusers/tests/models/transformers/test_models_prior.py class PriorTransformerTests (line 39) | class PriorTransformerTests(ModelTesterMixin, unittest.TestCase): method dummy_input (line 44) | def dummy_input(self): method get_dummy_seed_input (line 61) | def get_dummy_seed_input(self, seed=0): method input_shape (line 80) | def input_shape(self): method output_shape (line 84) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 87) | def prepare_init_args_and_inputs_for_common(self): method test_from_pretrained_hub (line 99) | def test_from_pretrained_hub(self): method test_forward_signature (line 111) | def test_forward_signature(self): method test_output_pretrained (line 122) | def test_output_pretrained(self): class PriorTransformerIntegrationTests (line 144) | class PriorTransformerIntegrationTests(unittest.TestCase): method get_dummy_seed_input (line 145) | def get_dummy_seed_input(self, batch_size=1, embedding_dim=768, num_em... method tearDown (line 163) | def tearDown(self): method test_kandinsky_prior (line 177) | def test_kandinsky_prior(self, seed, expected_slice): FILE: diffusers/tests/models/transformers/test_models_transformer_sd3.py class SD3TransformerTests (line 32) | class SD3TransformerTests(ModelTesterMixin, unittest.TestCase): method dummy_input (line 37) | def dummy_input(self): method input_shape (line 57) | def input_shape(self): method output_shape (line 61) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 64) | def prepare_init_args_and_inputs_for_common(self): FILE: diffusers/tests/models/unets/test_models_unet_1d.py class UNet1DModelTests (line 31) | class UNet1DModelTests(ModelTesterMixin, UNetTesterMixin, unittest.TestC... method dummy_input (line 36) | def dummy_input(self): method input_shape (line 47) | def input_shape(self): method output_shape (line 51) | def output_shape(self): method test_ema_training (line 54) | def test_ema_training(self): method test_training (line 57) | def test_training(self): method test_determinism (line 60) | def test_determinism(self): method test_outputs_equivalence (line 63) | def test_outputs_equivalence(self): method test_from_save_pretrained (line 66) | def test_from_save_pretrained(self): method test_from_save_pretrained_variant (line 69) | def test_from_save_pretrained_variant(self): method test_model_from_pretrained (line 72) | def test_model_from_pretrained(self): method test_output (line 75) | def test_output(self): method prepare_init_args_and_inputs_for_common (line 78) | def prepare_init_args_and_inputs_for_common(self): method test_from_pretrained_hub (line 96) | def test_from_pretrained_hub(self): method test_output_pretrained (line 108) | def test_output_pretrained(self): method test_forward_with_norm_groups (line 129) | def test_forward_with_norm_groups(self): method test_unet_1d_maestro (line 134) | def test_unet_1d_maestro(self): class UNetRLModelTests (line 153) | class UNetRLModelTests(ModelTesterMixin, UNetTesterMixin, unittest.TestC... method dummy_input (line 158) | def dummy_input(self): method input_shape (line 169) | def input_shape(self): method output_shape (line 173) | def output_shape(self): method test_determinism (line 176) | def test_determinism(self): method test_outputs_equivalence (line 179) | def test_outputs_equivalence(self): method test_from_save_pretrained (line 182) | def test_from_save_pretrained(self): method test_from_save_pretrained_variant (line 185) | def test_from_save_pretrained_variant(self): method test_model_from_pretrained (line 188) | def test_model_from_pretrained(self): method test_output (line 191) | def test_output(self): method test_ema_training (line 208) | def test_ema_training(self): method test_training (line 211) | def test_training(self): method prepare_init_args_and_inputs_for_common (line 214) | def prepare_init_args_and_inputs_for_common(self): method test_from_pretrained_hub (line 234) | def test_from_pretrained_hub(self): method test_output_pretrained (line 246) | def test_output_pretrained(self): method test_forward_with_norm_groups (line 268) | def test_forward_with_norm_groups(self): FILE: diffusers/tests/models/unets/test_models_unet_2d.py class Unet2DModelTests (line 41) | class Unet2DModelTests(ModelTesterMixin, UNetTesterMixin, unittest.TestC... method dummy_input (line 46) | def dummy_input(self): method input_shape (line 57) | def input_shape(self): method output_shape (line 61) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 64) | def prepare_init_args_and_inputs_for_common(self): method test_mid_block_attn_groups (line 79) | def test_mid_block_attn_groups(self): class UNetLDMModelTests (line 109) | class UNetLDMModelTests(ModelTesterMixin, UNetTesterMixin, unittest.Test... method dummy_input (line 114) | def dummy_input(self): method input_shape (line 125) | def input_shape(self): method output_shape (line 129) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 132) | def prepare_init_args_and_inputs_for_common(self): method test_from_pretrained_hub (line 146) | def test_from_pretrained_hub(self): method test_from_pretrained_accelerate (line 158) | def test_from_pretrained_accelerate(self): method test_from_pretrained_accelerate_wont_change_results (line 166) | def test_from_pretrained_accelerate_wont_change_results(self): method test_output_pretrained (line 198) | def test_output_pretrained(self): class NCSNppModelTests (line 224) | class NCSNppModelTests(ModelTesterMixin, UNetTesterMixin, unittest.TestC... method dummy_input (line 229) | def dummy_input(self, sizes=(32, 32)): method input_shape (line 239) | def input_shape(self): method output_shape (line 243) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 246) | def prepare_init_args_and_inputs_for_common(self): method test_from_pretrained_hub (line 273) | def test_from_pretrained_hub(self): method test_output_pretrained_ve_mid (line 287) | def test_output_pretrained_ve_mid(self): method test_output_pretrained_ve_large (line 308) | def test_output_pretrained_ve_large(self): method test_forward_with_norm_groups (line 329) | def test_forward_with_norm_groups(self): FILE: diffusers/tests/models/unets/test_models_unet_2d_condition.py function get_unet_lora_config (line 67) | def get_unet_lora_config(): function check_if_lora_correctly_set (line 79) | def check_if_lora_correctly_set(model) -> bool: function create_ip_adapter_state_dict (line 89) | def create_ip_adapter_state_dict(model): function create_ip_adapter_plus_state_dict (line 144) | def create_ip_adapter_plus_state_dict(model): function create_ip_adapter_faceid_state_dict (line 245) | def create_ip_adapter_faceid_state_dict(model): function create_custom_diffusion_layers (line 303) | def create_custom_diffusion_layers(model, mock_weights: bool = True): class UNet2DConditionModelTests (line 352) | class UNet2DConditionModelTests(ModelTesterMixin, UNetTesterMixin, unitt... method dummy_input (line 359) | def dummy_input(self): method input_shape (line 371) | def input_shape(self): method output_shape (line 375) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 378) | def prepare_init_args_and_inputs_for_common(self): method test_xformers_enable_works (line 398) | def test_xformers_enable_works(self): method test_gradient_checkpointing (line 410) | def test_gradient_checkpointing(self): method test_model_with_attention_head_dim_tuple (line 450) | def test_model_with_attention_head_dim_tuple(self): method test_model_with_use_linear_projection (line 470) | def test_model_with_use_linear_projection(self): method test_model_with_cross_attention_dim_tuple (line 489) | def test_model_with_cross_attention_dim_tuple(self): method test_model_with_simple_projection (line 508) | def test_model_with_simple_projection(self): method test_model_with_class_embeddings_concat (line 532) | def test_model_with_class_embeddings_concat(self): method test_model_attention_slicing (line 557) | def test_model_attention_slicing(self): method test_model_sliceable_head_dim (line 582) | def test_model_sliceable_head_dim(self): method test_gradient_checkpointing_is_applied (line 601) | def test_gradient_checkpointing_is_applied(self): method test_special_attn_proc (line 638) | def test_special_attn_proc(self): method test_model_xattn_mask (line 705) | def test_model_xattn_mask(self, mask_dtype): method test_model_xattn_padding (line 743) | def test_model_xattn_padding(self): method test_custom_diffusion_processors (line 766) | def test_custom_diffusion_processors(self): method test_custom_diffusion_save_load (line 793) | def test_custom_diffusion_save_load(self): method test_custom_diffusion_xformers_on_off (line 833) | def test_custom_diffusion_xformers_on_off(self): method test_pickle (line 859) | def test_pickle(self): method test_asymmetrical_unet (line 876) | def test_asymmetrical_unet(self): method test_ip_adapter (line 892) | def test_ip_adapter(self): method test_ip_adapter_plus (line 965) | def test_ip_adapter_plus(self): method test_load_sharded_checkpoint_from_hub (line 1039) | def test_load_sharded_checkpoint_from_hub(self): method test_load_sharded_checkpoint_from_hub_local (line 1048) | def test_load_sharded_checkpoint_from_hub_local(self): method test_lora (line 1058) | def test_lora(self): method test_lora_serialization (line 1081) | def test_lora_serialization(self): class UNet2DConditionModelIntegrationTests (line 1118) | class UNet2DConditionModelIntegrationTests(unittest.TestCase): method get_file_format (line 1119) | def get_file_format(self, seed, shape): method tearDown (line 1122) | def tearDown(self): method get_latents (line 1128) | def get_latents(self, seed=0, shape=(4, 4, 64, 64), fp16=False): method get_unet_model (line 1133) | def get_unet_model(self, fp16=False, model_id="CompVis/stable-diffusio... method test_set_attention_slice_auto (line 1145) | def test_set_attention_slice_auto(self): method test_set_attention_slice_max (line 1165) | def test_set_attention_slice_max(self): method test_set_attention_slice_int (line 1185) | def test_set_attention_slice_int(self): method test_set_attention_slice_list (line 1205) | def test_set_attention_slice_list(self): method get_encoder_hidden_states (line 1226) | def get_encoder_hidden_states(self, seed=0, shape=(4, 77, 768), fp16=F... method test_compvis_sd_v1_4 (line 1242) | def test_compvis_sd_v1_4(self, seed, timestep, expected_slice): method test_compvis_sd_v1_4_fp16 (line 1270) | def test_compvis_sd_v1_4_fp16(self, seed, timestep, expected_slice): method test_compvis_sd_v1_5 (line 1299) | def test_compvis_sd_v1_5(self, seed, timestep, expected_slice): method test_compvis_sd_v1_5_fp16 (line 1327) | def test_compvis_sd_v1_5_fp16(self, seed, timestep, expected_slice): method test_compvis_sd_inpaint (line 1356) | def test_compvis_sd_inpaint(self, seed, timestep, expected_slice): method test_compvis_sd_inpaint_fp16 (line 1384) | def test_compvis_sd_inpaint_fp16(self, seed, timestep, expected_slice): method test_stabilityai_sd_v2_fp16 (line 1412) | def test_stabilityai_sd_v2_fp16(self, seed, timestep, expected_slice): FILE: diffusers/tests/models/unets/test_models_unet_2d_flax.py class FlaxUNet2DConditionModelIntegrationTests (line 18) | class FlaxUNet2DConditionModelIntegrationTests(unittest.TestCase): method get_file_format (line 19) | def get_file_format(self, seed, shape): method tearDown (line 22) | def tearDown(self): method get_latents (line 27) | def get_latents(self, seed=0, shape=(4, 4, 64, 64), fp16=False): method get_unet_model (line 32) | def get_unet_model(self, fp16=False, model_id="CompVis/stable-diffusio... method get_encoder_hidden_states (line 41) | def get_encoder_hidden_states(self, seed=0, shape=(4, 77, 768), fp16=F... method test_compvis_sd_v1_4_flax_vs_torch_fp16 (line 56) | def test_compvis_sd_v1_4_flax_vs_torch_fp16(self, seed, timestep, expe... method test_stabilityai_sd_v2_flax_vs_torch_fp16 (line 86) | def test_stabilityai_sd_v2_flax_vs_torch_fp16(self, seed, timestep, ex... FILE: diffusers/tests/models/unets/test_models_unet_3d_condition.py class UNet3DConditionModelTests (line 35) | class UNet3DConditionModelTests(ModelTesterMixin, UNetTesterMixin, unitt... method dummy_input (line 40) | def dummy_input(self): method input_shape (line 53) | def input_shape(self): method output_shape (line 57) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 60) | def prepare_init_args_and_inputs_for_common(self): method test_xformers_enable_works (line 83) | def test_xformers_enable_works(self): method test_forward_with_norm_groups (line 95) | def test_forward_with_norm_groups(self): method test_determinism (line 115) | def test_determinism(self): method test_model_attention_slicing (line 141) | def test_model_attention_slicing(self): method test_feed_forward_chunking (line 166) | def test_feed_forward_chunking(self): FILE: diffusers/tests/models/unets/test_models_unet_controlnetxs.py class UNetControlNetXSModelTests (line 35) | class UNetControlNetXSModelTests(ModelTesterMixin, UNetTesterMixin, unit... method dummy_input (line 40) | def dummy_input(self): method input_shape (line 61) | def input_shape(self): method output_shape (line 65) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 68) | def prepare_init_args_and_inputs_for_common(self): method get_dummy_unet (line 87) | def get_dummy_unet(self): method get_dummy_controlnet_from_unet (line 102) | def get_dummy_controlnet_from_unet(self, unet, **kwargs): method test_from_unet (line 107) | def test_from_unet(self): method test_freeze_unet (line 188) | def test_freeze_unet(self): method test_gradient_checkpointing_is_applied (line 271) | def test_gradient_checkpointing_is_applied(self): method test_forward_no_control (line 305) | def test_forward_no_control(self): method test_time_embedding_mixing (line 325) | def test_time_embedding_mixing(self): method test_forward_with_norm_groups (line 347) | def test_forward_with_norm_groups(self): FILE: diffusers/tests/models/unets/test_models_unet_motion.py class UNetMotionModelTests (line 41) | class UNetMotionModelTests(ModelTesterMixin, UNetTesterMixin, unittest.T... method dummy_input (line 46) | def dummy_input(self): method input_shape (line 59) | def input_shape(self): method output_shape (line 63) | def output_shape(self): method prepare_init_args_and_inputs_for_common (line 66) | def prepare_init_args_and_inputs_for_common(self): method test_from_unet2d (line 82) | def test_from_unet2d(self): method test_freeze_unet2d (line 93) | def test_freeze_unet2d(self): method test_loading_motion_adapter (line 105) | def test_loading_motion_adapter(self): method test_saving_motion_modules (line 124) | def test_saving_motion_modules(self): method test_xformers_enable_works (line 152) | def test_xformers_enable_works(self): method test_gradient_checkpointing_is_applied (line 163) | def test_gradient_checkpointing_is_applied(self): method test_feed_forward_chunking (line 196) | def test_feed_forward_chunking(self): method test_pickle (line 215) | def test_pickle(self): method test_from_save_pretrained (line 228) | def test_from_save_pretrained(self, expected_max_diff=5e-5): method test_from_save_pretrained_variant (line 255) | def test_from_save_pretrained_variant(self, expected_max_diff=5e-5): method test_forward_with_norm_groups (line 290) | def test_forward_with_norm_groups(self): FILE: diffusers/tests/models/unets/test_models_unet_spatiotemporal.py class UNetSpatioTemporalConditionModelTests (line 41) | class UNetSpatioTemporalConditionModelTests(ModelTesterMixin, UNetTester... method dummy_input (line 46) | def dummy_input(self): method input_shape (line 64) | def input_shape(self): method output_shape (line 68) | def output_shape(self): method fps (line 72) | def fps(self): method motion_bucket_id (line 76) | def motion_bucket_id(self): method noise_aug_strength (line 80) | def noise_aug_strength(self): method addition_time_embed_dim (line 84) | def addition_time_embed_dim(self): method prepare_init_args_and_inputs_for_common (line 87) | def prepare_init_args_and_inputs_for_common(self): method _get_add_time_ids (line 110) | def _get_add_time_ids(self, do_classifier_free_guidance=True): method test_forward_with_norm_groups (line 129) | def test_forward_with_norm_groups(self): method test_model_attention_slicing (line 133) | def test_model_attention_slicing(self): method test_model_with_use_linear_projection (line 137) | def test_model_with_use_linear_projection(self): method test_model_with_simple_projection (line 141) | def test_model_with_simple_projection(self): method test_model_with_class_embeddings_concat (line 145) | def test_model_with_class_embeddings_concat(self): method test_xformers_enable_works (line 152) | def test_xformers_enable_works(self): method test_gradient_checkpointing (line 164) | def test_gradient_checkpointing(self): method test_model_with_num_attention_heads_tuple (line 204) | def test_model_with_num_attention_heads_tuple(self): method test_model_with_cross_attention_dim_tuple (line 222) | def test_model_with_cross_attention_dim_tuple(self): method test_gradient_checkpointing_is_applied (line 241) | def test_gradient_checkpointing_is_applied(self): method test_pickle (line 277) | def test_pickle(self): FILE: diffusers/tests/models/unets/test_unet_2d_blocks.py class DownBlock2DTests (line 23) | class DownBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method test_output (line 27) | def test_output(self): class ResnetDownsampleBlock2DTests (line 32) | class ResnetDownsampleBlock2DTests(UNetBlockTesterMixin, unittest.TestCa... method test_output (line 36) | def test_output(self): class AttnDownBlock2DTests (line 41) | class AttnDownBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method test_output (line 45) | def test_output(self): class CrossAttnDownBlock2DTests (line 50) | class CrossAttnDownBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method prepare_init_args_and_inputs_for_common (line 54) | def prepare_init_args_and_inputs_for_common(self): method test_output (line 59) | def test_output(self): class SimpleCrossAttnDownBlock2DTests (line 64) | class SimpleCrossAttnDownBlock2DTests(UNetBlockTesterMixin, unittest.Tes... method dummy_input (line 69) | def dummy_input(self): method prepare_init_args_and_inputs_for_common (line 72) | def prepare_init_args_and_inputs_for_common(self): method test_output (line 78) | def test_output(self): class SkipDownBlock2DTests (line 83) | class SkipDownBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method dummy_input (line 88) | def dummy_input(self): method test_output (line 91) | def test_output(self): class AttnSkipDownBlock2DTests (line 96) | class AttnSkipDownBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method dummy_input (line 101) | def dummy_input(self): method test_output (line 104) | def test_output(self): class DownEncoderBlock2DTests (line 109) | class DownEncoderBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method dummy_input (line 114) | def dummy_input(self): method prepare_init_args_and_inputs_for_common (line 117) | def prepare_init_args_and_inputs_for_common(self): method test_output (line 125) | def test_output(self): class AttnDownEncoderBlock2DTests (line 130) | class AttnDownEncoderBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method dummy_input (line 135) | def dummy_input(self): method prepare_init_args_and_inputs_for_common (line 138) | def prepare_init_args_and_inputs_for_common(self): method test_output (line 146) | def test_output(self): class UNetMidBlock2DTests (line 151) | class UNetMidBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method prepare_init_args_and_inputs_for_common (line 155) | def prepare_init_args_and_inputs_for_common(self): method test_output (line 163) | def test_output(self): class UNetMidBlock2DCrossAttnTests (line 168) | class UNetMidBlock2DCrossAttnTests(UNetBlockTesterMixin, unittest.TestCa... method prepare_init_args_and_inputs_for_common (line 172) | def prepare_init_args_and_inputs_for_common(self): method test_output (line 177) | def test_output(self): class UNetMidBlock2DSimpleCrossAttnTests (line 182) | class UNetMidBlock2DSimpleCrossAttnTests(UNetBlockTesterMixin, unittest.... method dummy_input (line 187) | def dummy_input(self): method prepare_init_args_and_inputs_for_common (line 190) | def prepare_init_args_and_inputs_for_common(self): method test_output (line 195) | def test_output(self): class UpBlock2DTests (line 200) | class UpBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method dummy_input (line 205) | def dummy_input(self): method test_output (line 208) | def test_output(self): class ResnetUpsampleBlock2DTests (line 213) | class ResnetUpsampleBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method dummy_input (line 218) | def dummy_input(self): method test_output (line 221) | def test_output(self): class CrossAttnUpBlock2DTests (line 226) | class CrossAttnUpBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method dummy_input (line 231) | def dummy_input(self): method prepare_init_args_and_inputs_for_common (line 234) | def prepare_init_args_and_inputs_for_common(self): method test_output (line 239) | def test_output(self): class SimpleCrossAttnUpBlock2DTests (line 244) | class SimpleCrossAttnUpBlock2DTests(UNetBlockTesterMixin, unittest.TestC... method dummy_input (line 249) | def dummy_input(self): method prepare_init_args_and_inputs_for_common (line 252) | def prepare_init_args_and_inputs_for_common(self): method test_output (line 257) | def test_output(self): class AttnUpBlock2DTests (line 262) | class AttnUpBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method dummy_input (line 267) | def dummy_input(self): method test_output (line 271) | def test_output(self): class SkipUpBlock2DTests (line 276) | class SkipUpBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method dummy_input (line 281) | def dummy_input(self): method test_output (line 284) | def test_output(self): class AttnSkipUpBlock2DTests (line 289) | class AttnSkipUpBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method dummy_input (line 294) | def dummy_input(self): method test_output (line 297) | def test_output(self): class UpDecoderBlock2DTests (line 302) | class UpDecoderBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method dummy_input (line 307) | def dummy_input(self): method prepare_init_args_and_inputs_for_common (line 310) | def prepare_init_args_and_inputs_for_common(self): method test_output (line 316) | def test_output(self): class AttnUpDecoderBlock2DTests (line 321) | class AttnUpDecoderBlock2DTests(UNetBlockTesterMixin, unittest.TestCase): method dummy_input (line 326) | def dummy_input(self): method prepare_init_args_and_inputs_for_common (line 329) | def prepare_init_args_and_inputs_for_common(self): method test_output (line 335) | def test_output(self): FILE: diffusers/tests/models/unets/test_unet_blocks_common.py class UNetBlockTesterMixin (line 30) | class UNetBlockTesterMixin: method dummy_input (line 32) | def dummy_input(self): method output_shape (line 36) | def output_shape(self): method get_dummy_input (line 46) | def get_dummy_input( method prepare_init_args_and_inputs_for_common (line 79) | def prepare_init_args_and_inputs_for_common(self): method test_output (line 94) | def test_output(self, expected_slice): method test_training (line 113) | def test_training(self): FILE: diffusers/tests/others/test_check_copies.py class CopyCheckTester (line 48) | class CopyCheckTester(unittest.TestCase): method setUp (line 49) | def setUp(self): method tearDown (line 58) | def tearDown(self): method check_copy_consistency (line 62) | def check_copy_consistency(self, comment, class_name, class_code, over... method test_find_code_in_diffusers (line 77) | def test_find_code_in_diffusers(self): method test_is_copy_consistent (line 81) | def test_is_copy_consistent(self): FILE: diffusers/tests/others/test_check_dummies.py class CheckDummiesTester (line 31) | class CheckDummiesTester(unittest.TestCase): method test_find_backend (line 32) | def test_find_backend(self): method test_read_init (line 52) | def test_read_init(self): method test_create_dummy_object (line 68) | def test_create_dummy_object(self): method test_create_dummy_files (line 95) | def test_create_dummy_files(self): FILE: diffusers/tests/others/test_config.py class SampleObject (line 32) | class SampleObject(ConfigMixin): method __init__ (line 36) | def __init__( class SampleObject2 (line 47) | class SampleObject2(ConfigMixin): method __init__ (line 51) | def __init__( class SampleObject3 (line 62) | class SampleObject3(ConfigMixin): method __init__ (line 66) | def __init__( class SampleObject4 (line 78) | class SampleObject4(ConfigMixin): method __init__ (line 82) | def __init__( class ConfigTester (line 94) | class ConfigTester(unittest.TestCase): method test_load_not_from_mixin (line 95) | def test_load_not_from_mixin(self): method test_register_to_config (line 99) | def test_register_to_config(self): method test_save_load (line 135) | def test_save_load(self): method test_load_ddim_from_pndm (line 159) | def test_load_ddim_from_pndm(self): method test_load_euler_from_pndm (line 173) | def test_load_euler_from_pndm(self): method test_load_euler_ancestral_from_pndm (line 187) | def test_load_euler_ancestral_from_pndm(self): method test_load_pndm (line 201) | def test_load_pndm(self): method test_overwrite_config_on_load (line 215) | def test_overwrite_config_on_load(self): method test_load_dpmsolver (line 240) | def test_load_dpmsolver(self): method test_use_default_values (line 254) | def test_use_default_values(self): FILE: diffusers/tests/others/test_dependencies.py class DependencyTester (line 20) | class DependencyTester(unittest.TestCase): method test_diffusers_import (line 21) | def test_diffusers_import(self): method test_backend_registration (line 27) | def test_backend_registration(self): method test_pipeline_imports (line 42) | def test_pipeline_imports(self): FILE: diffusers/tests/others/test_ema.py class EMAModelTests (line 29) | class EMAModelTests(unittest.TestCase): method get_models (line 38) | def get_models(self, decay=0.9999): method get_dummy_inputs (line 44) | def get_dummy_inputs(self): method simulate_backprop (line 54) | def simulate_backprop(self, unet): method test_optimization_steps_updated (line 62) | def test_optimization_steps_updated(self): method test_shadow_params_not_updated (line 73) | def test_shadow_params_not_updated(self): method test_shadow_params_updated (line 90) | def test_shadow_params_updated(self): method test_consecutive_shadow_params_updated (line 111) | def test_consecutive_shadow_params_updated(self): method test_zero_decay (line 129) | def test_zero_decay(self): method test_serialization (line 146) | def test_serialization(self): FILE: diffusers/tests/others/test_hub_utils.py class CreateModelCardTest (line 22) | class CreateModelCardTest(unittest.TestCase): method test_generate_model_card_with_library_name (line 23) | def test_generate_model_card_with_library_name(self): FILE: diffusers/tests/others/test_image_processor.py class ImageProcessorTest (line 25) | class ImageProcessorTest(unittest.TestCase): method dummy_sample (line 27) | def dummy_sample(self): method dummy_mask (line 38) | def dummy_mask(self): method to_np (line 48) | def to_np(self, image): method test_vae_image_processor_pt (line 55) | def test_vae_image_processor_pt(self): method test_vae_image_processor_np (line 72) | def test_vae_image_processor_np(self): method test_vae_image_processor_pil (line 85) | def test_vae_image_processor_pil(self): method test_preprocess_input_3d (line 100) | def test_preprocess_input_3d(self): method test_preprocess_input_list (line 130) | def test_preprocess_input_list(self): method test_preprocess_input_mask_3d (line 162) | def test_preprocess_input_mask_3d(self): method test_preprocess_input_mask_list (line 215) | def test_preprocess_input_mask_list(self): method test_preprocess_input_mask_3d_batch (line 254) | def test_preprocess_input_mask_3d_batch(self): method test_vae_image_processor_resize_pt (line 289) | def test_vae_image_processor_resize_pt(self): method test_vae_image_processor_resize_np (line 300) | def test_vae_image_processor_resize_np(self): FILE: diffusers/tests/others/test_outputs.py class CustomOutput (line 14) | class CustomOutput(BaseOutput): class ConfigTester (line 18) | class ConfigTester(unittest.TestCase): method test_outputs_single_attribute (line 19) | def test_outputs_single_attribute(self): method test_outputs_dict_init (line 41) | def test_outputs_dict_init(self): method test_outputs_serialization (line 64) | def test_outputs_serialization(self): method test_torch_pytree (line 75) | def test_torch_pytree(self): FILE: diffusers/tests/others/test_training.py class TrainingTests (line 28) | class TrainingTests(unittest.TestCase): method get_model_optimizer (line 29) | def get_model_optimizer(self, resolution=32): method test_training_step_equality (line 36) | def test_training_step_equality(self): FILE: diffusers/tests/others/test_utils.py class DeprecateTester (line 34) | class DeprecateTester(unittest.TestCase): method test_deprecate_function_arg (line 38) | def test_deprecate_function_arg(self): method test_deprecate_function_arg_tuple (line 51) | def test_deprecate_function_arg_tuple(self): method test_deprecate_function_args (line 64) | def test_deprecate_function_args(self): method test_deprecate_function_incorrect_arg (line 85) | def test_deprecate_function_incorrect_arg(self): method test_deprecate_arg_no_kwarg (line 95) | def test_deprecate_arg_no_kwarg(self): method test_deprecate_args_no_kwarg (line 104) | def test_deprecate_args_no_kwarg(self): method test_deprecate_class_obj (line 119) | def test_deprecate_class_obj(self): method test_deprecate_class_objs (line 132) | def test_deprecate_class_objs(self): method test_deprecate_incorrect_version (line 160) | def test_deprecate_incorrect_version(self): method test_deprecate_incorrect_no_standard_warn (line 172) | def test_deprecate_incorrect_no_standard_warn(self): method test_deprecate_stacklevel (line 178) | def test_deprecate_stacklevel(self): function parse_flag_from_env (line 185) | def parse_flag_from_env(key, default=False): function is_staging_test (line 204) | def is_staging_test(test_case): FILE: diffusers/tests/others/test_video_processor.py class VideoProcessorTest (line 30) | class VideoProcessorTest(unittest.TestCase): method get_dummy_sample (line 31) | def get_dummy_sample(self, input_type): method to_np (line 76) | def to_np(self, video): method test_video_processor_pil (line 132) | def test_video_processor_pil(self, input_type): method test_video_processor_np (line 144) | def test_video_processor_np(self, input_type): method test_video_processor_pt (line 158) | def test_video_processor_pt(self, input_type): FILE: diffusers/tests/pipelines/amused/test_amused.py class AmusedPipelineFastTests (line 33) | class AmusedPipelineFastTests(PipelineTesterMixin, unittest.TestCase): method get_dummy_components (line 38) | def get_dummy_components(self): method get_dummy_inputs (line 109) | def get_dummy_inputs(self, device, seed=0): method test_inference_batch_consistent (line 124) | def test_inference_batch_consistent(self, batch_sizes=[2]): method test_inference_batch_single_identical (line 128) | def test_inference_batch_single_identical(self): class AmusedPipelineSlowTests (line 134) | class AmusedPipelineSlowTests(unittest.TestCase): method test_amused_256 (line 135) | def test_amused_256(self): method test_amused_256_fp16 (line 147) | def test_amused_256_fp16(self): method test_amused_512 (line 159) | def test_amused_512(self): method test_amused_512_fp16 (line 171) | def test_amused_512_fp16(self): FILE: diffusers/tests/pipelines/amused/test_amused_img2img.py class AmusedImg2ImgPipelineFastTests (line 34) | class AmusedImg2ImgPipelineFastTests(PipelineTesterMixin, unittest.TestC... method get_dummy_components (line 42) | def get_dummy_components(self): method get_dummy_inputs (line 113) | def get_dummy_inputs(self, device, seed=0): method test_inference_batch_consistent (line 128) | def test_inference_batch_consistent(self, batch_sizes=[2]): method test_inference_batch_single_identical (line 132) | def test_inference_batch_single_identical(self): class AmusedImg2ImgPipelineSlowTests (line 138) | class AmusedImg2ImgPipelineSlowTests(unittest.TestCase): method test_amused_256 (line 139) | def test_amused_256(self): method test_amused_256_fp16 (line 164) | def test_amused_256_fp16(self): method test_amused_512 (line 189) | def test_amused_512(self): method test_amused_512_fp16 (line 213) | def test_amused_512_fp16(self): FILE: diffusers/tests/pipelines/amused/test_amused_inpaint.py class AmusedInpaintPipelineFastTests (line 34) | class AmusedInpaintPipelineFastTests(PipelineTesterMixin, unittest.TestC... method get_dummy_components (line 42) | def get_dummy_components(self): method get_dummy_inputs (line 113) | def get_dummy_inputs(self, device, seed=0): method test_inference_batch_consistent (line 132) | def test_inference_batch_consistent(self, batch_sizes=[2]): method test_inference_batch_single_identical (line 136) | def test_inference_batch_single_identical(self): class AmusedInpaintPipelineSlowTests (line 142) | class AmusedInpaintPipelineSlowTests(unittest.TestCase): method test_amused_256 (line 143) | def test_amused_256(self): method test_amused_256_fp16 (line 176) | def test_amused_256_fp16(self): method test_amused_512 (line 209) | def test_amused_512(self): method test_amused_512_fp16 (line 242) | def test_amused_512_fp16(self): FILE: diffusers/tests/pipelines/animatediff/test_animatediff.py function to_np (line 29) | def to_np(tensor): class AnimateDiffPipelineFastTests (line 36) | class AnimateDiffPipelineFastTests( method get_dummy_components (line 53) | def get_dummy_components(self): method get_dummy_inputs (line 114) | def get_dummy_inputs(self, device, seed=0): method test_motion_unet_loading (line 129) | def test_motion_unet_loading(self): method test_attention_slicing_forward_pass (line 136) | def test_attention_slicing_forward_pass(self): method test_ip_adapter_single (line 139) | def test_ip_adapter_single(self): method test_dict_tuple_outputs_equivalent (line 175) | def test_dict_tuple_outputs_equivalent(self): method test_inference_batch_single_identical (line 181) | def test_inference_batch_single_identical( method test_to_device (line 237) | def test_to_device(self): method test_to_dtype (line 261) | def test_to_dtype(self): method test_prompt_embeds (line 274) | def test_prompt_embeds(self): method test_free_init (line 285) | def test_free_init(self): method test_xformers_attention_forwardGenerator_pass (line 324) | def test_xformers_attention_forwardGenerator_pass(self): method test_vae_slicing (line 349) | def test_vae_slicing(self): class AnimateDiffPipelineSlowTests (line 355) | class AnimateDiffPipelineSlowTests(unittest.TestCase): method setUp (line 356) | def setUp(self): method tearDown (line 362) | def tearDown(self): method test_animatediff (line 368) | def test_animatediff(self): FILE: diffusers/tests/pipelines/animatediff/test_animatediff_sdxl.py function to_np (line 28) | def to_np(tensor): class AnimateDiffPipelineSDXLFastTests (line 35) | class AnimateDiffPipelineSDXLFastTests( method get_dummy_components (line 57) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_inputs (line 135) | def get_dummy_inputs(self, device, seed=0): method test_motion_unet_loading (line 150) | def test_motion_unet_loading(self): method test_attention_slicing_forward_pass (line 157) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 160) | def test_inference_batch_single_identical( method test_to_device (line 216) | def test_to_device(self): method test_to_dtype (line 240) | def test_to_dtype(self): method test_prompt_embeds (line 253) | def test_prompt_embeds(self): method test_save_load_optional_components (line 277) | def test_save_load_optional_components(self): method test_xformers_attention_forwardGenerator_pass (line 284) | def test_xformers_attention_forwardGenerator_pass(self): FILE: diffusers/tests/pipelines/animatediff/test_animatediff_video2video.py function to_np (line 24) | def to_np(tensor): class AnimateDiffVideoToVideoPipelineFastTests (line 31) | class AnimateDiffVideoToVideoPipelineFastTests( method get_dummy_components (line 48) | def get_dummy_components(self): method get_dummy_inputs (line 109) | def get_dummy_inputs(self, device, seed=0): method test_motion_unet_loading (line 130) | def test_motion_unet_loading(self): method test_attention_slicing_forward_pass (line 137) | def test_attention_slicing_forward_pass(self): method test_ip_adapter_single (line 140) | def test_ip_adapter_single(self): method test_inference_batch_single_identical (line 168) | def test_inference_batch_single_identical( method test_to_device (line 224) | def test_to_device(self): method test_to_dtype (line 248) | def test_to_dtype(self): method test_prompt_embeds (line 261) | def test_prompt_embeds(self): method test_latent_inputs (line 272) | def test_latent_inputs(self): method test_xformers_attention_forwardGenerator_pass (line 287) | def test_xformers_attention_forwardGenerator_pass(self): method test_free_init (line 312) | def test_free_init(self): FILE: diffusers/tests/pipelines/audioldm/test_audioldm.py class AudioLDMPipelineFastTests (line 49) | class AudioLDMPipelineFastTests(PipelineTesterMixin, unittest.TestCase): method get_dummy_components (line 66) | def get_dummy_components(self): method get_dummy_inputs (line 138) | def get_dummy_inputs(self, device, seed=0): method test_audioldm_ddim (line 151) | def test_audioldm_ddim(self): method test_audioldm_prompt_embeds (line 173) | def test_audioldm_prompt_embeds(self): method test_audioldm_negative_prompt_embeds (line 214) | def test_audioldm_negative_prompt_embeds(self): method test_audioldm_negative_prompt (line 261) | def test_audioldm_negative_prompt(self): method test_audioldm_num_waveforms_per_prompt (line 284) | def test_audioldm_num_waveforms_per_prompt(self): method test_audioldm_audio_length_in_s (line 319) | def test_audioldm_audio_length_in_s(self): method test_audioldm_vocoder_model_in_dim (line 340) | def test_audioldm_vocoder_model_in_dim(self): method test_attention_slicing_forward_pass (line 360) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 363) | def test_inference_batch_single_identical(self): method test_xformers_attention_forwardGenerator_pass (line 370) | def test_xformers_attention_forwardGenerator_pass(self): class AudioLDMPipelineSlowTests (line 375) | class AudioLDMPipelineSlowTests(unittest.TestCase): method setUp (line 376) | def setUp(self): method tearDown (line 381) | def tearDown(self): method get_inputs (line 386) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_audioldm (line 399) | def test_audioldm(self): class AudioLDMPipelineNightlyTests (line 420) | class AudioLDMPipelineNightlyTests(unittest.TestCase): method setUp (line 421) | def setUp(self): method tearDown (line 426) | def tearDown(self): method get_inputs (line 431) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_audioldm_lms (line 444) | def test_audioldm_lms(self): FILE: diffusers/tests/pipelines/audioldm2/test_audioldm2.py class AudioLDM2PipelineFastTests (line 56) | class AudioLDM2PipelineFastTests(PipelineTesterMixin, unittest.TestCase): method get_dummy_components (line 73) | def get_dummy_components(self): method get_dummy_inputs (line 193) | def get_dummy_inputs(self, device, seed=0): method test_audioldm2_ddim (line 206) | def test_audioldm2_ddim(self): method test_audioldm2_prompt_embeds (line 228) | def test_audioldm2_prompt_embeds(self): method test_audioldm2_negative_prompt_embeds (line 283) | def test_audioldm2_negative_prompt_embeds(self): method test_audioldm2_negative_prompt (line 346) | def test_audioldm2_negative_prompt(self): method test_audioldm2_num_waveforms_per_prompt (line 369) | def test_audioldm2_num_waveforms_per_prompt(self): method test_audioldm2_audio_length_in_s (line 404) | def test_audioldm2_audio_length_in_s(self): method test_audioldm2_vocoder_model_in_dim (line 425) | def test_audioldm2_vocoder_model_in_dim(self): method test_attention_slicing_forward_pass (line 445) | def test_attention_slicing_forward_pass(self): method test_xformers_attention_forwardGenerator_pass (line 449) | def test_xformers_attention_forwardGenerator_pass(self): method test_dict_tuple_outputs_equivalent (line 452) | def test_dict_tuple_outputs_equivalent(self): method test_inference_batch_single_identical (line 456) | def test_inference_batch_single_identical(self): method test_save_load_local (line 460) | def test_save_load_local(self): method test_save_load_optional_components (line 464) | def test_save_load_optional_components(self): method test_to_dtype (line 468) | def test_to_dtype(self): method test_sequential_cpu_offload_forward_pass (line 490) | def test_sequential_cpu_offload_forward_pass(self): class AudioLDM2PipelineSlowTests (line 495) | class AudioLDM2PipelineSlowTests(unittest.TestCase): method setUp (line 496) | def setUp(self): method tearDown (line 501) | def tearDown(self): method get_inputs (line 506) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method get_inputs_tts (line 519) | def get_inputs_tts(self, device, generator_device="cpu", dtype=torch.f... method test_audioldm2 (line 533) | def test_audioldm2(self): method test_audioldm2_lms (line 551) | def test_audioldm2_lms(self): method test_audioldm2_large (line 571) | def test_audioldm2_large(self): method test_audioldm2_tts (line 590) | def test_audioldm2_tts(self): FILE: diffusers/tests/pipelines/blipdiffusion/test_blipdiffusion.py class BlipDiffusionPipelineFastTests (line 36) | class BlipDiffusionPipelineFastTests(PipelineTesterMixin, unittest.TestC... method get_dummy_components (line 63) | def get_dummy_components(self): method get_dummy_inputs (line 155) | def get_dummy_inputs(self, device, seed=0): method test_blipdiffusion (line 178) | def test_blipdiffusion(self): FILE: diffusers/tests/pipelines/consistency_models/test_consistency_models.py class ConsistencyModelPipelineFastTests (line 29) | class ConsistencyModelPipelineFastTests(PipelineTesterMixin, unittest.Te... method dummy_uncond_unet (line 48) | def dummy_uncond_unet(self): method dummy_cond_unet (line 56) | def dummy_cond_unet(self): method get_dummy_components (line 63) | def get_dummy_components(self, class_cond=False): method get_dummy_inputs (line 83) | def get_dummy_inputs(self, device, seed=0): method test_consistency_model_pipeline_multistep (line 99) | def test_consistency_model_pipeline_multistep(self): method test_consistency_model_pipeline_multistep_class_cond (line 115) | def test_consistency_model_pipeline_multistep_class_cond(self): method test_consistency_model_pipeline_onestep (line 132) | def test_consistency_model_pipeline_onestep(self): method test_consistency_model_pipeline_onestep_class_cond (line 150) | def test_consistency_model_pipeline_onestep_class_cond(self): class ConsistencyModelPipelineSlowTests (line 172) | class ConsistencyModelPipelineSlowTests(unittest.TestCase): method setUp (line 173) | def setUp(self): method tearDown (line 178) | def tearDown(self): method get_inputs (line 183) | def get_inputs(self, seed=0, get_fixed_latents=False, device="cpu", dt... method get_fixed_latents (line 200) | def get_fixed_latents(self, seed=0, device="cpu", dtype=torch.float32,... method test_consistency_model_cd_multistep (line 207) | def test_consistency_model_cd_multistep(self): method test_consistency_model_cd_onestep (line 228) | def test_consistency_model_cd_onestep(self): method test_consistency_model_cd_multistep_flash_attn (line 252) | def test_consistency_model_cd_multistep_flash_attn(self): method test_consistency_model_cd_onestep_flash_attn (line 276) | def test_consistency_model_cd_onestep_flash_attn(self): FILE: diffusers/tests/pipelines/controlnet/test_controlnet.py function _test_stable_diffusion_compile (line 68) | def _test_stable_diffusion_compile(in_queue, out_queue, timeout): class ControlNetPipelineFastTests (line 113) | class ControlNetPipelineFastTests( method get_dummy_components (line 126) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_inputs (line 196) | def get_dummy_inputs(self, device, seed=0): method test_attention_slicing_forward_pass (line 220) | def test_attention_slicing_forward_pass(self): method test_ip_adapter_single (line 223) | def test_ip_adapter_single(self): method test_xformers_attention_forwardGenerator_pass (line 233) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 236) | def test_inference_batch_single_identical(self): method test_controlnet_lcm (line 239) | def test_controlnet_lcm(self): method test_controlnet_lcm_custom_timesteps (line 261) | def test_controlnet_lcm_custom_timesteps(self): class StableDiffusionMultiControlNetPipelineFastTests (line 286) | class StableDiffusionMultiControlNetPipelineFastTests( method get_dummy_components (line 294) | def get_dummy_components(self): method get_dummy_inputs (line 385) | def get_dummy_inputs(self, device, seed=0): method test_control_guidance_switch (line 417) | def test_control_guidance_switch(self): method test_attention_slicing_forward_pass (line 450) | def test_attention_slicing_forward_pass(self): method test_xformers_attention_forwardGenerator_pass (line 457) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 460) | def test_inference_batch_single_identical(self): method test_ip_adapter_single (line 463) | def test_ip_adapter_single(self): method test_save_pretrained_raise_not_implemented_exception (line 469) | def test_save_pretrained_raise_not_implemented_exception(self): method test_inference_multiple_prompt_input (line 481) | def test_inference_multiple_prompt_input(self): class StableDiffusionMultiControlNetOneModelPipelineFastTests (line 518) | class StableDiffusionMultiControlNetOneModelPipelineFastTests( method get_dummy_components (line 526) | def get_dummy_components(self): method get_dummy_inputs (line 605) | def get_dummy_inputs(self, device, seed=0): method test_control_guidance_switch (line 632) | def test_control_guidance_switch(self): method test_attention_slicing_forward_pass (line 669) | def test_attention_slicing_forward_pass(self): method test_xformers_attention_forwardGenerator_pass (line 676) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 679) | def test_inference_batch_single_identical(self): method test_ip_adapter_single (line 682) | def test_ip_adapter_single(self): method test_save_pretrained_raise_not_implemented_exception (line 688) | def test_save_pretrained_raise_not_implemented_exception(self): class ControlNetPipelineSlowTests (line 703) | class ControlNetPipelineSlowTests(unittest.TestCase): method setUp (line 704) | def setUp(self): method tearDown (line 709) | def tearDown(self): method test_canny (line 714) | def test_canny(self): method test_depth (line 741) | def test_depth(self): method test_hed (line 768) | def test_hed(self): method test_mlsd (line 795) | def test_mlsd(self): method test_normal (line 822) | def test_normal(self): method test_openpose (line 849) | def test_openpose(self): method test_scribble (line 876) | def test_scribble(self): method test_seg (line 903) | def test_seg(self): method test_sequential_cpu_offloading (line 930) | def test_sequential_cpu_offloading(self): method test_canny_guess_mode (line 960) | def test_canny_guess_mode(self): method test_canny_guess_mode_euler (line 992) | def test_canny_guess_mode_euler(self): method test_stable_diffusion_compile (line 1031) | def test_stable_diffusion_compile(self): method test_v11_shuffle_global_pool_conditions (line 1034) | def test_v11_shuffle_global_pool_conditions(self): class StableDiffusionMultiControlNetPipelineSlowTests (line 1068) | class StableDiffusionMultiControlNetPipelineSlowTests(unittest.TestCase): method setUp (line 1069) | def setUp(self): method tearDown (line 1074) | def tearDown(self): method test_pose_and_canny (line 1079) | def test_pose_and_canny(self): FILE: diffusers/tests/pipelines/controlnet/test_controlnet_blip_diffusion.py class BlipDiffusionControlNetPipelineFastTests (line 42) | class BlipDiffusionControlNetPipelineFastTests(PipelineTesterMixin, unit... method get_dummy_components (line 71) | def get_dummy_components(self): method get_dummy_inputs (line 173) | def get_dummy_inputs(self, device, seed=0): method test_dict_tuple_outputs_equivalent (line 199) | def test_dict_tuple_outputs_equivalent(self): method test_blipdiffusion_controlnet (line 205) | def test_blipdiffusion_controlnet(self): FILE: diffusers/tests/pipelines/controlnet/test_controlnet_img2img.py class ControlNetImg2ImgPipelineFastTests (line 64) | class ControlNetImg2ImgPipelineFastTests( method get_dummy_components (line 77) | def get_dummy_components(self): method get_dummy_inputs (line 146) | def get_dummy_inputs(self, device, seed=0): method test_attention_slicing_forward_pass (line 173) | def test_attention_slicing_forward_pass(self): method test_ip_adapter_single (line 176) | def test_ip_adapter_single(self): method test_xformers_attention_forwardGenerator_pass (line 186) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 189) | def test_inference_batch_single_identical(self): class StableDiffusionMultiControlNetPipelineFastTests (line 193) | class StableDiffusionMultiControlNetPipelineFastTests( method get_dummy_components (line 201) | def get_dummy_components(self): method get_dummy_inputs (line 292) | def get_dummy_inputs(self, device, seed=0): method test_control_guidance_switch (line 328) | def test_control_guidance_switch(self): method test_attention_slicing_forward_pass (line 361) | def test_attention_slicing_forward_pass(self): method test_xformers_attention_forwardGenerator_pass (line 368) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 371) | def test_inference_batch_single_identical(self): method test_ip_adapter_single (line 374) | def test_ip_adapter_single(self): method test_save_pretrained_raise_not_implemented_exception (line 380) | def test_save_pretrained_raise_not_implemented_exception(self): class ControlNetImg2ImgPipelineSlowTests (line 395) | class ControlNetImg2ImgPipelineSlowTests(unittest.TestCase): method setUp (line 396) | def setUp(self): method tearDown (line 401) | def tearDown(self): method test_canny (line 406) | def test_canny(self): FILE: diffusers/tests/pipelines/controlnet/test_controlnet_inpaint.py class ControlNetInpaintPipelineFastTests (line 60) | class ControlNetInpaintPipelineFastTests( method get_dummy_components (line 69) | def get_dummy_components(self): method get_dummy_inputs (line 135) | def get_dummy_inputs(self, device, seed=0): method test_attention_slicing_forward_pass (line 166) | def test_attention_slicing_forward_pass(self): method test_xformers_attention_forwardGenerator_pass (line 173) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 176) | def test_inference_batch_single_identical(self): class ControlNetSimpleInpaintPipelineFastTests (line 180) | class ControlNetSimpleInpaintPipelineFastTests(ControlNetInpaintPipeline... method get_dummy_components (line 186) | def get_dummy_components(self): class MultiControlNetInpaintPipelineFastTests (line 253) | class MultiControlNetInpaintPipelineFastTests( method get_dummy_components (line 260) | def get_dummy_components(self): method get_dummy_inputs (line 347) | def get_dummy_inputs(self, device, seed=0): method test_control_guidance_switch (line 386) | def test_control_guidance_switch(self): method test_attention_slicing_forward_pass (line 419) | def test_attention_slicing_forward_pass(self): method test_xformers_attention_forwardGenerator_pass (line 426) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 429) | def test_inference_batch_single_identical(self): method test_save_pretrained_raise_not_implemented_exception (line 432) | def test_save_pretrained_raise_not_implemented_exception(self): class ControlNetInpaintPipelineSlowTests (line 447) | class ControlNetInpaintPipelineSlowTests(unittest.TestCase): method setUp (line 448) | def setUp(self): method tearDown (line 453) | def tearDown(self): method test_canny (line 458) | def test_canny(self): method test_inpaint (line 503) | def test_inpaint(self): FILE: diffusers/tests/pipelines/controlnet/test_controlnet_inpaint_sdxl.py class ControlNetPipelineSDXLFastTests (line 59) | class ControlNetPipelineSDXLFastTests( method get_dummy_components (line 76) | def get_dummy_components(self): method get_dummy_inputs (line 187) | def get_dummy_inputs(self, device, seed=0, img_res=64): method test_attention_slicing_forward_pass (line 228) | def test_attention_slicing_forward_pass(self): method test_dict_tuple_outputs_equivalent (line 231) | def test_dict_tuple_outputs_equivalent(self): method test_xformers_attention_forwardGenerator_pass (line 241) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 244) | def test_inference_batch_single_identical(self): method test_stable_diffusion_xl_offloads (line 248) | def test_stable_diffusion_xl_offloads(self): method test_stable_diffusion_xl_multi_prompts (line 276) | def test_stable_diffusion_xl_multi_prompts(self): method test_controlnet_sdxl_guess (line 329) | def test_controlnet_sdxl_guess(self): method test_save_load_optional_components (line 350) | def test_save_load_optional_components(self): method test_float16_inference (line 353) | def test_float16_inference(self): FILE: diffusers/tests/pipelines/controlnet/test_controlnet_sdxl.py class StableDiffusionXLControlNetPipelineFastTests (line 64) | class StableDiffusionXLControlNetPipelineFastTests( method get_dummy_components (line 78) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_inputs (line 166) | def get_dummy_inputs(self, device, seed=0): method test_attention_slicing_forward_pass (line 190) | def test_attention_slicing_forward_pass(self): method test_ip_adapter_single (line 193) | def test_ip_adapter_single(self, from_ssd1b=False, expected_pipe_slice... method test_xformers_attention_forwardGenerator_pass (line 206) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 209) | def test_inference_batch_single_identical(self): method test_save_load_optional_components (line 212) | def test_save_load_optional_components(self): method test_stable_diffusion_xl_offloads (line 216) | def test_stable_diffusion_xl_offloads(self): method test_stable_diffusion_xl_multi_prompts (line 244) | def test_stable_diffusion_xl_multi_prompts(self): method test_stable_diffusion_xl_prompt_embeds (line 298) | def test_stable_diffusion_xl_prompt_embeds(self): method test_controlnet_sdxl_guess (line 336) | def test_controlnet_sdxl_guess(self): method test_controlnet_sdxl_lcm (line 358) | def test_controlnet_sdxl_lcm(self): method test_controlnet_sdxl_two_mixture_of_denoiser_fast (line 380) | def test_controlnet_sdxl_two_mixture_of_denoiser_fast(self): class StableDiffusionXLMultiControlNetPipelineFastTests (line 483) | class StableDiffusionXLMultiControlNetPipelineFastTests( method get_dummy_components (line 491) | def get_dummy_components(self): method get_dummy_inputs (line 606) | def get_dummy_inputs(self, device, seed=0): method test_control_guidance_switch (line 638) | def test_control_guidance_switch(self): method test_attention_slicing_forward_pass (line 671) | def test_attention_slicing_forward_pass(self): method test_xformers_attention_forwardGenerator_pass (line 678) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 681) | def test_inference_batch_single_identical(self): method test_save_load_optional_components (line 684) | def test_save_load_optional_components(self): class StableDiffusionXLMultiControlNetOneModelPipelineFastTests (line 688) | class StableDiffusionXLMultiControlNetOneModelPipelineFastTests( method get_dummy_components (line 696) | def get_dummy_components(self): method get_dummy_inputs (line 793) | def get_dummy_inputs(self, device, seed=0): method test_control_guidance_switch (line 819) | def test_control_guidance_switch(self): method test_attention_slicing_forward_pass (line 856) | def test_attention_slicing_forward_pass(self): method test_xformers_attention_forwardGenerator_pass (line 863) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 866) | def test_inference_batch_single_identical(self): method test_save_load_optional_components (line 869) | def test_save_load_optional_components(self): method test_negative_conditions (line 872) | def test_negative_conditions(self): class ControlNetSDXLPipelineSlowTests (line 894) | class ControlNetSDXLPipelineSlowTests(unittest.TestCase): method setUp (line 895) | def setUp(self): method tearDown (line 900) | def tearDown(self): method test_canny (line 905) | def test_canny(self): method test_depth (line 928) | def test_depth(self): class StableDiffusionSSD1BControlNetPipelineFastTests (line 952) | class StableDiffusionSSD1BControlNetPipelineFastTests(StableDiffusionXLC... method test_controlnet_sdxl_guess (line 953) | def test_controlnet_sdxl_guess(self): method test_ip_adapter_single (line 975) | def test_ip_adapter_single(self): method test_controlnet_sdxl_lcm (line 981) | def test_controlnet_sdxl_lcm(self): method test_conditioning_channels (line 1001) | def test_conditioning_channels(self): method get_dummy_components (line 1026) | def get_dummy_components(self, time_cond_proj_dim=None): FILE: diffusers/tests/pipelines/controlnet/test_controlnet_sdxl_img2img.py class ControlNetPipelineSDXLImg2ImgFastTests (line 50) | class ControlNetPipelineSDXLImg2ImgFastTests( method get_dummy_components (line 67) | def get_dummy_components(self, skip_first_text_encoder=False): method get_dummy_inputs (line 154) | def get_dummy_inputs(self, device, seed=0): method test_ip_adapter_single (line 178) | def test_ip_adapter_single(self): method test_stable_diffusion_xl_controlnet_img2img (line 184) | def test_stable_diffusion_xl_controlnet_img2img(self): method test_stable_diffusion_xl_controlnet_img2img_guess (line 202) | def test_stable_diffusion_xl_controlnet_img2img_guess(self): method test_attention_slicing_forward_pass (line 226) | def test_attention_slicing_forward_pass(self): method test_xformers_attention_forwardGenerator_pass (line 233) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 236) | def test_inference_batch_single_identical(self): method test_save_load_optional_components (line 240) | def test_save_load_optional_components(self): method test_stable_diffusion_xl_offloads (line 244) | def test_stable_diffusion_xl_offloads(self): method test_stable_diffusion_xl_multi_prompts (line 272) | def test_stable_diffusion_xl_multi_prompts(self): method test_stable_diffusion_xl_prompt_embeds (line 326) | def test_stable_diffusion_xl_prompt_embeds(self): FILE: diffusers/tests/pipelines/controlnet/test_flax_controlnet.py class FlaxControlNetPipelineIntegrationTests (line 33) | class FlaxControlNetPipelineIntegrationTests(unittest.TestCase): method tearDown (line 34) | def tearDown(self): method test_canny (line 39) | def test_canny(self): method test_pose (line 84) | def test_pose(self): FILE: diffusers/tests/pipelines/controlnet_xs/test_controlnetxs.py function to_np (line 72) | def to_np(tensor): function _test_stable_diffusion_compile (line 80) | def _test_stable_diffusion_compile(in_queue, out_queue, timeout): class ControlNetXSPipelineFastTests (line 126) | class ControlNetXSPipelineFastTests( method get_dummy_components (line 141) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_inputs (line 208) | def get_dummy_inputs(self, device, seed=0): method test_xformers_attention_forwardGenerator_pass (line 236) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 239) | def test_inference_batch_single_identical(self): method test_controlnet_lcm (line 242) | def test_controlnet_lcm(self): method test_to_dtype (line 262) | def test_to_dtype(self): method test_multi_vae (line 275) | def test_multi_vae(self): method test_to_device (line 310) | def test_to_device(self): class ControlNetXSPipelineSlowTests (line 337) | class ControlNetXSPipelineSlowTests(unittest.TestCase): method tearDown (line 338) | def tearDown(self): method test_canny (line 343) | def test_canny(self): method test_depth (line 369) | def test_depth(self): method test_stable_diffusion_compile (line 397) | def test_stable_diffusion_compile(self): FILE: diffusers/tests/pipelines/controlnet_xs/test_controlnetxs_sdxl.py class StableDiffusionXLControlNetXSPipelineFastTests (line 60) | class StableDiffusionXLControlNetXSPipelineFastTests( method get_dummy_components (line 75) | def get_dummy_components(self): method get_dummy_inputs (line 155) | def get_dummy_inputs(self, device, seed=0): method test_attention_slicing_forward_pass (line 180) | def test_attention_slicing_forward_pass(self): method test_xformers_attention_forwardGenerator_pass (line 188) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 192) | def test_inference_batch_single_identical(self): method test_stable_diffusion_xl_offloads (line 197) | def test_stable_diffusion_xl_offloads(self): method test_stable_diffusion_xl_multi_prompts (line 226) | def test_stable_diffusion_xl_multi_prompts(self): method test_stable_diffusion_xl_prompt_embeds (line 280) | def test_stable_diffusion_xl_prompt_embeds(self): method test_save_load_optional_components (line 319) | def test_save_load_optional_components(self): method test_to_dtype (line 323) | def test_to_dtype(self): method test_multi_vae (line 336) | def test_multi_vae(self): class StableDiffusionXLControlNetXSPipelineSlowTests (line 373) | class StableDiffusionXLControlNetXSPipelineSlowTests(unittest.TestCase): method tearDown (line 374) | def tearDown(self): method test_canny (line 379) | def test_canny(self): method test_depth (line 403) | def test_depth(self): FILE: diffusers/tests/pipelines/dance_diffusion/test_dance_diffusion.py class DanceDiffusionPipelineFastTests (line 32) | class DanceDiffusionPipelineFastTests(PipelineTesterMixin, unittest.Test... method get_dummy_components (line 45) | def get_dummy_components(self): method get_dummy_inputs (line 69) | def get_dummy_inputs(self, device, seed=0): method test_dance_diffusion (line 81) | def test_dance_diffusion(self): method test_save_load_local (line 99) | def test_save_load_local(self): method test_dict_tuple_outputs_equivalent (line 103) | def test_dict_tuple_outputs_equivalent(self): method test_save_load_optional_components (line 107) | def test_save_load_optional_components(self): method test_attention_slicing_forward_pass (line 111) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 114) | def test_inference_batch_single_identical(self): class PipelineIntegrationTests (line 120) | class PipelineIntegrationTests(unittest.TestCase): method setUp (line 121) | def setUp(self): method tearDown (line 127) | def tearDown(self): method test_dance_diffusion (line 133) | def test_dance_diffusion(self): method test_dance_diffusion_fp16 (line 151) | def test_dance_diffusion_fp16(self): FILE: diffusers/tests/pipelines/ddim/test_ddim.py class DDIMPipelineFastTests (line 31) | class DDIMPipelineFastTests(PipelineTesterMixin, unittest.TestCase): method get_dummy_components (line 42) | def get_dummy_components(self): method get_dummy_inputs (line 58) | def get_dummy_inputs(self, device, seed=0): method test_inference (line 71) | def test_inference(self): method test_dict_tuple_outputs_equivalent (line 88) | def test_dict_tuple_outputs_equivalent(self): method test_save_load_local (line 91) | def test_save_load_local(self): method test_save_load_optional_components (line 94) | def test_save_load_optional_components(self): method test_inference_batch_single_identical (line 97) | def test_inference_batch_single_identical(self): class DDIMPipelineIntegrationTests (line 103) | class DDIMPipelineIntegrationTests(unittest.TestCase): method test_inference_cifar10 (line 104) | def test_inference_cifar10(self): method test_inference_ema_bedroom (line 124) | def test_inference_ema_bedroom(self): FILE: diffusers/tests/pipelines/ddpm/test_ddpm.py class DDPMPipelineFastTests (line 28) | class DDPMPipelineFastTests(unittest.TestCase): method dummy_uncond_unet (line 30) | def dummy_uncond_unet(self): method test_fast_inference (line 44) | def test_fast_inference(self): method test_inference_predict_sample (line 68) | def test_inference_predict_sample(self): class DDPMPipelineIntegrationTests (line 92) | class DDPMPipelineIntegrationTests(unittest.TestCase): method test_inference_cifar10 (line 93) | def test_inference_cifar10(self): FILE: diffusers/tests/pipelines/deepfloyd_if/__init__.py class IFPipelineTesterMixin (line 18) | class IFPipelineTesterMixin: method _get_dummy_components (line 19) | def _get_dummy_components(self): method _get_superresolution_dummy_components (line 76) | def _get_superresolution_dummy_components(self): method _test_save_load_optional_components (line 152) | def _test_save_load_optional_components(self): method _test_save_load_local (line 251) | def _test_save_load_local(self): FILE: diffusers/tests/pipelines/deepfloyd_if/test_if.py class IFPipelineFastTests (line 34) | class IFPipelineFastTests(PipelineTesterMixin, IFPipelineTesterMixin, un... method get_dummy_components (line 40) | def get_dummy_components(self): method get_dummy_inputs (line 43) | def get_dummy_inputs(self, device, seed=0): method test_save_load_optional_components (line 58) | def test_save_load_optional_components(self): method test_save_load_float16 (line 62) | def test_save_load_float16(self): method test_attention_slicing_forward_pass (line 66) | def test_attention_slicing_forward_pass(self): method test_save_load_local (line 69) | def test_save_load_local(self): method test_inference_batch_single_identical (line 72) | def test_inference_batch_single_identical(self): method test_xformers_attention_forwardGenerator_pass (line 81) | def test_xformers_attention_forwardGenerator_pass(self): class IFPipelineSlowTests (line 87) | class IFPipelineSlowTests(unittest.TestCase): method setUp (line 88) | def setUp(self): method tearDown (line 94) | def tearDown(self): method test_if_text_to_image (line 100) | def test_if_text_to_image(self): FILE: diffusers/tests/pipelines/deepfloyd_if/test_if_img2img.py class IFImg2ImgPipelineFastTests (line 36) | class IFImg2ImgPipelineFastTests(PipelineTesterMixin, IFPipelineTesterMi... method get_dummy_components (line 42) | def get_dummy_components(self): method get_dummy_inputs (line 45) | def get_dummy_inputs(self, device, seed=0): method test_save_load_optional_components (line 63) | def test_save_load_optional_components(self): method test_xformers_attention_forwardGenerator_pass (line 70) | def test_xformers_attention_forwardGenerator_pass(self): method test_save_load_float16 (line 74) | def test_save_load_float16(self): method test_float16_inference (line 79) | def test_float16_inference(self): method test_attention_slicing_forward_pass (line 82) | def test_attention_slicing_forward_pass(self): method test_save_load_local (line 85) | def test_save_load_local(self): method test_inference_batch_single_identical (line 88) | def test_inference_batch_single_identical(self): class IFImg2ImgPipelineSlowTests (line 96) | class IFImg2ImgPipelineSlowTests(unittest.TestCase): method setUp (line 97) | def setUp(self): method tearDown (line 103) | def tearDown(self): method test_if_img2img (line 109) | def test_if_img2img(self): FILE: diffusers/tests/pipelines/deepfloyd_if/test_if_img2img_superresolution.py class IFImg2ImgSuperResolutionPipelineFastTests (line 36) | class IFImg2ImgSuperResolutionPipelineFastTests(PipelineTesterMixin, IFP... method get_dummy_components (line 42) | def get_dummy_components(self): method get_dummy_inputs (line 45) | def get_dummy_inputs(self, device, seed=0): method test_xformers_attention_forwardGenerator_pass (line 69) | def test_xformers_attention_forwardGenerator_pass(self): method test_save_load_optional_components (line 72) | def test_save_load_optional_components(self): method test_save_load_float16 (line 76) | def test_save_load_float16(self): method test_attention_slicing_forward_pass (line 80) | def test_attention_slicing_forward_pass(self): method test_save_load_local (line 83) | def test_save_load_local(self): method test_inference_batch_single_identical (line 86) | def test_inference_batch_single_identical(self): class IFImg2ImgSuperResolutionPipelineSlowTests (line 94) | class IFImg2ImgSuperResolutionPipelineSlowTests(unittest.TestCase): method setUp (line 95) | def setUp(self): method tearDown (line 101) | def tearDown(self): method test_if_img2img_superresolution (line 107) | def test_if_img2img_superresolution(self): FILE: diffusers/tests/pipelines/deepfloyd_if/test_if_inpainting.py class IFInpaintingPipelineFastTests (line 36) | class IFInpaintingPipelineFastTests(PipelineTesterMixin, IFPipelineTeste... method get_dummy_components (line 42) | def get_dummy_components(self): method get_dummy_inputs (line 45) | def get_dummy_inputs(self, device, seed=0): method test_xformers_attention_forwardGenerator_pass (line 69) | def test_xformers_attention_forwardGenerator_pass(self): method test_save_load_optional_components (line 72) | def test_save_load_optional_components(self): method test_save_load_float16 (line 76) | def test_save_load_float16(self): method test_attention_slicing_forward_pass (line 80) | def test_attention_slicing_forward_pass(self): method test_save_load_local (line 83) | def test_save_load_local(self): method test_inference_batch_single_identical (line 86) | def test_inference_batch_single_identical(self): class IFInpaintingPipelineSlowTests (line 94) | class IFInpaintingPipelineSlowTests(unittest.TestCase): method setUp (line 95) | def setUp(self): method tearDown (line 101) | def tearDown(self): method test_if_inpainting (line 107) | def test_if_inpainting(self): FILE: diffusers/tests/pipelines/deepfloyd_if/test_if_inpainting_superresolution.py class IFInpaintingSuperResolutionPipelineFastTests (line 36) | class IFInpaintingSuperResolutionPipelineFastTests(PipelineTesterMixin, ... method get_dummy_components (line 42) | def get_dummy_components(self): method get_dummy_inputs (line 45) | def get_dummy_inputs(self, device, seed=0): method test_xformers_attention_forwardGenerator_pass (line 71) | def test_xformers_attention_forwardGenerator_pass(self): method test_save_load_optional_components (line 74) | def test_save_load_optional_components(self): method test_save_load_float16 (line 78) | def test_save_load_float16(self): method test_attention_slicing_forward_pass (line 82) | def test_attention_slicing_forward_pass(self): method test_save_load_local (line 85) | def test_save_load_local(self): method test_inference_batch_single_identical (line 88) | def test_inference_batch_single_identical(self): class IFInpaintingSuperResolutionPipelineSlowTests (line 96) | class IFInpaintingSuperResolutionPipelineSlowTests(unittest.TestCase): method setUp (line 97) | def setUp(self): method tearDown (line 103) | def tearDown(self): method test_if_inpainting_superresolution (line 109) | def test_if_inpainting_superresolution(self): FILE: diffusers/tests/pipelines/deepfloyd_if/test_if_superresolution.py class IFSuperResolutionPipelineFastTests (line 33) | class IFSuperResolutionPipelineFastTests(PipelineTesterMixin, IFPipeline... method get_dummy_components (line 39) | def get_dummy_components(self): method get_dummy_inputs (line 42) | def get_dummy_inputs(self, device, seed=0): method test_xformers_attention_forwardGenerator_pass (line 64) | def test_xformers_attention_forwardGenerator_pass(self): method test_save_load_optional_components (line 67) | def test_save_load_optional_components(self): method test_save_load_float16 (line 71) | def test_save_load_float16(self): method test_attention_slicing_forward_pass (line 75) | def test_attention_slicing_forward_pass(self): method test_save_load_local (line 78) | def test_save_load_local(self): method test_inference_batch_single_identical (line 81) | def test_inference_batch_single_identical(self): class IFSuperResolutionPipelineSlowTests (line 89) | class IFSuperResolutionPipelineSlowTests(unittest.TestCase): method setUp (line 90) | def setUp(self): method tearDown (line 96) | def tearDown(self): method test_if_superresolution (line 102) | def test_if_superresolution(self): FILE: diffusers/tests/pipelines/dit/test_dit.py class DiTPipelineFastTests (line 36) | class DiTPipelineFastTests(PipelineTesterMixin, unittest.TestCase): method get_dummy_components (line 47) | def get_dummy_components(self): method get_dummy_inputs (line 68) | def get_dummy_inputs(self, device, seed=0): method test_inference (line 81) | def test_inference(self): method test_inference_batch_single_identical (line 98) | def test_inference_batch_single_identical(self): method test_xformers_attention_forwardGenerator_pass (line 105) | def test_xformers_attention_forwardGenerator_pass(self): class DiTPipelineIntegrationTests (line 111) | class DiTPipelineIntegrationTests(unittest.TestCase): method setUp (line 112) | def setUp(self): method tearDown (line 117) | def tearDown(self): method test_dit_256 (line 122) | def test_dit_256(self): method test_dit_512 (line 139) | def test_dit_512(self): FILE: diffusers/tests/pipelines/hunyuan_dit/test_hunyuan_dit.py class HunyuanDiTPipelineFastTests (line 45) | class HunyuanDiTPipelineFastTests(PipelineTesterMixin, unittest.TestCase): method get_dummy_components (line 54) | def get_dummy_components(self): method get_dummy_inputs (line 91) | def get_dummy_inputs(self, device, seed=0): method test_inference (line 106) | def test_inference(self): method test_sequential_cpu_offload_forward_pass (line 125) | def test_sequential_cpu_offload_forward_pass(self): method test_sequential_offload_forward_pass_twice (line 129) | def test_sequential_offload_forward_pass_twice(self): method test_inference_batch_single_identical (line 133) | def test_inference_batch_single_identical(self): method test_save_load_optional_components (line 138) | def test_save_load_optional_components(self): method test_feed_forward_chunking (line 231) | def test_feed_forward_chunking(self): method test_fused_qkv_projections (line 251) | def test_fused_qkv_projections(self): class HunyuanDiTPipelineIntegrationTests (line 288) | class HunyuanDiTPipelineIntegrationTests(unittest.TestCase): method setUp (line 291) | def setUp(self): method tearDown (line 296) | def tearDown(self): method test_hunyuan_dit_1024 (line 301) | def test_hunyuan_dit_1024(self): FILE: diffusers/tests/pipelines/i2vgen_xl/test_i2vgenxl.py class I2VGenXLPipelineFastTests (line 56) | class I2VGenXLPipelineFastTests(SDFunctionTesterMixin, PipelineTesterMix... method get_dummy_components (line 63) | def get_dummy_components(self): method get_dummy_inputs (line 142) | def get_dummy_inputs(self, device, seed=0): method test_text_to_video_default_case (line 162) | def test_text_to_video_default_case(self): method test_save_load_local (line 180) | def test_save_load_local(self): method test_sequential_cpu_offload_forward_pass (line 183) | def test_sequential_cpu_offload_forward_pass(self): method test_dict_tuple_outputs_equivalent (line 186) | def test_dict_tuple_outputs_equivalent(self): method test_save_load_optional_components (line 189) | def test_save_load_optional_components(self): method test_attention_slicing_forward_pass (line 193) | def test_attention_slicing_forward_pass(self): method test_xformers_attention_forwardGenerator_pass (line 200) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 203) | def test_inference_batch_single_identical(self): method test_model_cpu_offload_forward_pass (line 206) | def test_model_cpu_offload_forward_pass(self): method test_num_videos_per_prompt (line 209) | def test_num_videos_per_prompt(self): class I2VGenXLPipelineSlowTests (line 231) | class I2VGenXLPipelineSlowTests(unittest.TestCase): method setUp (line 232) | def setUp(self): method tearDown (line 238) | def tearDown(self): method test_i2vgen_xl (line 244) | def test_i2vgen_xl(self): FILE: diffusers/tests/pipelines/ip_adapters/test_ip_adapter_stable_diffusion.py class IPAdapterNightlyTestsMixin (line 50) | class IPAdapterNightlyTestsMixin(unittest.TestCase): method setUp (line 53) | def setUp(self): method tearDown (line 59) | def tearDown(self): method get_image_encoder (line 65) | def get_image_encoder(self, repo_id, subfolder): method get_image_processor (line 71) | def get_image_processor(self, repo_id): method get_dummy_inputs (line 75) | def get_dummy_inputs( class IPAdapterSDIntegrationTests (line 169) | class IPAdapterSDIntegrationTests(IPAdapterNightlyTestsMixin): method test_text_to_image (line 170) | def test_text_to_image(self): method test_image_to_image (line 200) | def test_image_to_image(self): method test_inpainting (line 232) | def test_inpainting(self): method test_text_to_image_model_cpu_offload (line 260) | def test_text_to_image_model_cpu_offload(self): method test_text_to_image_full_face (line 287) | def test_text_to_image_full_face(self): method test_unload (line 304) | def test_unload(self): method test_multi (line 323) | def test_multi(self): method test_text_to_image_face_id (line 344) | def test_text_to_image_face_id(self): class IPAdapterSDXLIntegrationTests (line 376) | class IPAdapterSDXLIntegrationTests(IPAdapterNightlyTestsMixin): method test_text_to_image_sdxl (line 377) | def test_text_to_image_sdxl(self): method test_image_to_image_sdxl (line 437) | def test_image_to_image_sdxl(self): method test_inpainting_sdxl (line 506) | def test_inpainting_sdxl(self): method test_ip_adapter_single_mask (line 557) | def test_ip_adapter_single_mask(self): method test_ip_adapter_multiple_masks (line 585) | def test_ip_adapter_multiple_masks(self): method test_instant_style_multiple_masks (line 612) | def test_instant_style_multiple_masks(self): method test_ip_adapter_multiple_masks_one_adapter (line 654) | def test_ip_adapter_multiple_masks_one_adapter(self): FILE: diffusers/tests/pipelines/kandinsky/test_kandinsky.py class Dummies (line 41) | class Dummies: method text_embedder_hidden_size (line 43) | def text_embedder_hidden_size(self): method time_input_dim (line 47) | def time_input_dim(self): method block_out_channels_0 (line 51) | def block_out_channels_0(self): method time_embed_dim (line 55) | def time_embed_dim(self): method cross_attention_dim (line 59) | def cross_attention_dim(self): method dummy_tokenizer (line 63) | def dummy_tokenizer(self): method dummy_text_encoder (line 68) | def dummy_text_encoder(self): method dummy_unet (line 86) | def dummy_unet(self): method dummy_movq_kwargs (line 111) | def dummy_movq_kwargs(self): method dummy_movq (line 130) | def dummy_movq(self): method get_dummy_components (line 135) | def get_dummy_components(self): method get_dummy_inputs (line 162) | def get_dummy_inputs(self, device, seed=0): class KandinskyPipelineFastTests (line 183) | class KandinskyPipelineFastTests(PipelineTesterMixin, unittest.TestCase): method get_dummy_components (line 207) | def get_dummy_components(self): method get_dummy_inputs (line 211) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky (line 215) | def test_kandinsky(self): method test_offloads (line 248) | def test_offloads(self): class KandinskyPipelineIntegrationTests (line 277) | class KandinskyPipelineIntegrationTests(unittest.TestCase): method setUp (line 278) | def setUp(self): method tearDown (line 284) | def tearDown(self): method test_kandinsky_text2img (line 290) | def test_kandinsky_text2img(self): FILE: diffusers/tests/pipelines/kandinsky/test_kandinsky_combined.py class KandinskyPipelineCombinedFastTests (line 33) | class KandinskyPipelineCombinedFastTests(PipelineTesterMixin, unittest.T... method get_dummy_components (line 55) | def get_dummy_components(self): method get_dummy_inputs (line 63) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky (line 74) | def test_kandinsky(self): method test_offloads (line 107) | def test_offloads(self): method test_inference_batch_single_identical (line 133) | def test_inference_batch_single_identical(self): method test_float16_inference (line 136) | def test_float16_inference(self): method test_dict_tuple_outputs_equivalent (line 139) | def test_dict_tuple_outputs_equivalent(self): class KandinskyPipelineImg2ImgCombinedFastTests (line 143) | class KandinskyPipelineImg2ImgCombinedFastTests(PipelineTesterMixin, uni... method get_dummy_components (line 163) | def get_dummy_components(self): method get_dummy_inputs (line 171) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky (line 180) | def test_kandinsky(self): method test_offloads (line 213) | def test_offloads(self): method test_inference_batch_single_identical (line 239) | def test_inference_batch_single_identical(self): method test_float16_inference (line 242) | def test_float16_inference(self): method test_dict_tuple_outputs_equivalent (line 245) | def test_dict_tuple_outputs_equivalent(self): method test_save_load_optional_components (line 248) | def test_save_load_optional_components(self): class KandinskyPipelineInpaintCombinedFastTests (line 252) | class KandinskyPipelineInpaintCombinedFastTests(PipelineTesterMixin, uni... method get_dummy_components (line 272) | def get_dummy_components(self): method get_dummy_inputs (line 280) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky (line 289) | def test_kandinsky(self): method test_offloads (line 322) | def test_offloads(self): method test_inference_batch_single_identical (line 348) | def test_inference_batch_single_identical(self): method test_float16_inference (line 351) | def test_float16_inference(self): method test_dict_tuple_outputs_equivalent (line 354) | def test_dict_tuple_outputs_equivalent(self): method test_save_load_optional_components (line 357) | def test_save_load_optional_components(self): method test_save_load_local (line 360) | def test_save_load_local(self): FILE: diffusers/tests/pipelines/kandinsky/test_kandinsky_img2img.py class Dummies (line 51) | class Dummies: method text_embedder_hidden_size (line 53) | def text_embedder_hidden_size(self): method time_input_dim (line 57) | def time_input_dim(self): method block_out_channels_0 (line 61) | def block_out_channels_0(self): method time_embed_dim (line 65) | def time_embed_dim(self): method cross_attention_dim (line 69) | def cross_attention_dim(self): method dummy_tokenizer (line 73) | def dummy_tokenizer(self): method dummy_text_encoder (line 78) | def dummy_text_encoder(self): method dummy_unet (line 96) | def dummy_unet(self): method dummy_movq_kwargs (line 121) | def dummy_movq_kwargs(self): method dummy_movq (line 140) | def dummy_movq(self): method get_dummy_components (line 145) | def get_dummy_components(self): method get_dummy_inputs (line 175) | def get_dummy_inputs(self, device, seed=0): class KandinskyImg2ImgPipelineFastTests (line 203) | class KandinskyImg2ImgPipelineFastTests(PipelineTesterMixin, unittest.Te... method get_dummy_components (line 229) | def get_dummy_components(self): method get_dummy_inputs (line 233) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky_img2img (line 237) | def test_kandinsky_img2img(self): method test_offloads (line 269) | def test_offloads(self): method test_dict_tuple_outputs_equivalent (line 295) | def test_dict_tuple_outputs_equivalent(self): class KandinskyImg2ImgPipelineIntegrationTests (line 301) | class KandinskyImg2ImgPipelineIntegrationTests(unittest.TestCase): method setUp (line 302) | def setUp(self): method tearDown (line 308) | def tearDown(self): method test_kandinsky_img2img (line 314) | def test_kandinsky_img2img(self): class KandinskyImg2ImgPipelineNightlyTests (line 367) | class KandinskyImg2ImgPipelineNightlyTests(unittest.TestCase): method setUp (line 368) | def setUp(self): method tearDown (line 374) | def tearDown(self): method test_kandinsky_img2img_ddpm (line 380) | def test_kandinsky_img2img_ddpm(self): FILE: diffusers/tests/pipelines/kandinsky/test_kandinsky_inpaint.py class Dummies (line 43) | class Dummies: method text_embedder_hidden_size (line 45) | def text_embedder_hidden_size(self): method time_input_dim (line 49) | def time_input_dim(self): method block_out_channels_0 (line 53) | def block_out_channels_0(self): method time_embed_dim (line 57) | def time_embed_dim(self): method cross_attention_dim (line 61) | def cross_attention_dim(self): method dummy_tokenizer (line 65) | def dummy_tokenizer(self): method dummy_text_encoder (line 70) | def dummy_text_encoder(self): method dummy_unet (line 88) | def dummy_unet(self): method dummy_movq_kwargs (line 113) | def dummy_movq_kwargs(self): method dummy_movq (line 132) | def dummy_movq(self): method get_dummy_components (line 137) | def get_dummy_components(self): method get_dummy_inputs (line 165) | def get_dummy_inputs(self, device, seed=0): class KandinskyInpaintPipelineFastTests (line 196) | class KandinskyInpaintPipelineFastTests(PipelineTesterMixin, unittest.Te... method get_dummy_components (line 223) | def get_dummy_components(self): method get_dummy_inputs (line 227) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky_inpaint (line 231) | def test_kandinsky_inpaint(self): method test_inference_batch_single_identical (line 263) | def test_inference_batch_single_identical(self): method test_offloads (line 267) | def test_offloads(self): method test_float16_inference (line 293) | def test_float16_inference(self): class KandinskyInpaintPipelineIntegrationTests (line 299) | class KandinskyInpaintPipelineIntegrationTests(unittest.TestCase): method setUp (line 300) | def setUp(self): method tearDown (line 306) | def tearDown(self): method test_kandinsky_inpaint (line 312) | def test_kandinsky_inpaint(self): FILE: diffusers/tests/pipelines/kandinsky/test_kandinsky_prior.py class Dummies (line 39) | class Dummies: method text_embedder_hidden_size (line 41) | def text_embedder_hidden_size(self): method time_input_dim (line 45) | def time_input_dim(self): method block_out_channels_0 (line 49) | def block_out_channels_0(self): method time_embed_dim (line 53) | def time_embed_dim(self): method cross_attention_dim (line 57) | def cross_attention_dim(self): method dummy_tokenizer (line 61) | def dummy_tokenizer(self): method dummy_text_encoder (line 66) | def dummy_text_encoder(self): method dummy_prior (line 83) | def dummy_prior(self): method dummy_image_encoder (line 99) | def dummy_image_encoder(self): method dummy_image_processor (line 116) | def dummy_image_processor(self): method get_dummy_components (line 130) | def get_dummy_components(self): method get_dummy_inputs (line 156) | def get_dummy_inputs(self, device, seed=0): class KandinskyPriorPipelineFastTests (line 171) | class KandinskyPriorPipelineFastTests(PipelineTesterMixin, unittest.Test... method get_dummy_components (line 187) | def get_dummy_components(self): method get_dummy_inputs (line 191) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky_prior (line 195) | def test_kandinsky_prior(self): method test_inference_batch_single_identical (line 226) | def test_inference_batch_single_identical(self): method test_attention_slicing_forward_pass (line 230) | def test_attention_slicing_forward_pass(self): FILE: diffusers/tests/pipelines/kandinsky2_2/test_kandinsky.py class Dummies (line 39) | class Dummies: method text_embedder_hidden_size (line 41) | def text_embedder_hidden_size(self): method time_input_dim (line 45) | def time_input_dim(self): method block_out_channels_0 (line 49) | def block_out_channels_0(self): method time_embed_dim (line 53) | def time_embed_dim(self): method cross_attention_dim (line 57) | def cross_attention_dim(self): method dummy_unet (line 61) | def dummy_unet(self): method dummy_movq_kwargs (line 86) | def dummy_movq_kwargs(self): method dummy_movq (line 105) | def dummy_movq(self): method get_dummy_components (line 110) | def get_dummy_components(self): method get_dummy_inputs (line 133) | def get_dummy_inputs(self, device, seed=0): class KandinskyV22PipelineFastTests (line 155) | class KandinskyV22PipelineFastTests(PipelineTesterMixin, unittest.TestCa... method get_dummy_inputs (line 178) | def get_dummy_inputs(self, device, seed=0): method get_dummy_components (line 182) | def get_dummy_components(self): method test_kandinsky (line 186) | def test_kandinsky(self): method test_float16_inference (line 219) | def test_float16_inference(self): class KandinskyV22PipelineIntegrationTests (line 225) | class KandinskyV22PipelineIntegrationTests(unittest.TestCase): method setUp (line 226) | def setUp(self): method tearDown (line 232) | def tearDown(self): method test_kandinsky_text2img (line 238) | def test_kandinsky_text2img(self): FILE: diffusers/tests/pipelines/kandinsky2_2/test_kandinsky_combined.py class KandinskyV22PipelineCombinedFastTests (line 37) | class KandinskyV22PipelineCombinedFastTests(PipelineTesterMixin, unittes... method get_dummy_components (line 60) | def get_dummy_components(self): method get_dummy_inputs (line 68) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky (line 79) | def test_kandinsky(self): method test_offloads (line 112) | def test_offloads(self): method test_inference_batch_single_identical (line 138) | def test_inference_batch_single_identical(self): method test_float16_inference (line 141) | def test_float16_inference(self): method test_dict_tuple_outputs_equivalent (line 144) | def test_dict_tuple_outputs_equivalent(self): method test_model_cpu_offload_forward_pass (line 147) | def test_model_cpu_offload_forward_pass(self): method test_save_load_local (line 150) | def test_save_load_local(self): method test_save_load_optional_components (line 153) | def test_save_load_optional_components(self): method test_callback_inputs (line 156) | def test_callback_inputs(self): method test_callback_cfg (line 159) | def test_callback_cfg(self): class KandinskyV22PipelineImg2ImgCombinedFastTests (line 163) | class KandinskyV22PipelineImg2ImgCombinedFastTests(PipelineTesterMixin, ... method get_dummy_components (line 184) | def get_dummy_components(self): method get_dummy_inputs (line 192) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky (line 201) | def test_kandinsky(self): method test_offloads (line 234) | def test_offloads(self): method test_inference_batch_single_identical (line 260) | def test_inference_batch_single_identical(self): method test_float16_inference (line 263) | def test_float16_inference(self): method test_dict_tuple_outputs_equivalent (line 266) | def test_dict_tuple_outputs_equivalent(self): method test_model_cpu_offload_forward_pass (line 269) | def test_model_cpu_offload_forward_pass(self): method test_save_load_optional_components (line 272) | def test_save_load_optional_components(self): method save_load_local (line 275) | def save_load_local(self): method test_callback_inputs (line 278) | def test_callback_inputs(self): method test_callback_cfg (line 281) | def test_callback_cfg(self): class KandinskyV22PipelineInpaintCombinedFastTests (line 285) | class KandinskyV22PipelineInpaintCombinedFastTests(PipelineTesterMixin, ... method get_dummy_components (line 305) | def get_dummy_components(self): method get_dummy_inputs (line 313) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky (line 322) | def test_kandinsky(self): method test_offloads (line 355) | def test_offloads(self): method test_inference_batch_single_identical (line 381) | def test_inference_batch_single_identical(self): method test_float16_inference (line 384) | def test_float16_inference(self): method test_dict_tuple_outputs_equivalent (line 387) | def test_dict_tuple_outputs_equivalent(self): method test_model_cpu_offload_forward_pass (line 390) | def test_model_cpu_offload_forward_pass(self): method test_save_load_local (line 393) | def test_save_load_local(self): method test_save_load_optional_components (line 396) | def test_save_load_optional_components(self): method test_sequential_cpu_offload_forward_pass (line 399) | def test_sequential_cpu_offload_forward_pass(self): method test_callback_inputs (line 402) | def test_callback_inputs(self): method test_callback_cfg (line 405) | def test_callback_cfg(self): FILE: diffusers/tests/pipelines/kandinsky2_2/test_kandinsky_controlnet.py class KandinskyV22ControlnetPipelineFastTests (line 46) | class KandinskyV22ControlnetPipelineFastTests(PipelineTesterMixin, unitt... method text_embedder_hidden_size (line 66) | def text_embedder_hidden_size(self): method time_input_dim (line 70) | def time_input_dim(self): method block_out_channels_0 (line 74) | def block_out_channels_0(self): method time_embed_dim (line 78) | def time_embed_dim(self): method cross_attention_dim (line 82) | def cross_attention_dim(self): method dummy_unet (line 86) | def dummy_unet(self): method dummy_movq_kwargs (line 111) | def dummy_movq_kwargs(self): method dummy_movq (line 132) | def dummy_movq(self): method get_dummy_components (line 137) | def get_dummy_components(self): method get_dummy_inputs (line 160) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky_controlnet (line 186) | def test_kandinsky_controlnet(self): method test_float16_inference (line 221) | def test_float16_inference(self): method test_inference_batch_single_identical (line 224) | def test_inference_batch_single_identical(self): class KandinskyV22ControlnetPipelineIntegrationTests (line 230) | class KandinskyV22ControlnetPipelineIntegrationTests(unittest.TestCase): method setUp (line 231) | def setUp(self): method tearDown (line 237) | def tearDown(self): method test_kandinsky_controlnet (line 243) | def test_kandinsky_controlnet(self): FILE: diffusers/tests/pipelines/kandinsky2_2/test_kandinsky_controlnet_img2img.py class KandinskyV22ControlnetImg2ImgPipelineFastTests (line 47) | class KandinskyV22ControlnetImg2ImgPipelineFastTests(PipelineTesterMixin... method text_embedder_hidden_size (line 67) | def text_embedder_hidden_size(self): method time_input_dim (line 71) | def time_input_dim(self): method block_out_channels_0 (line 75) | def block_out_channels_0(self): method time_embed_dim (line 79) | def time_embed_dim(self): method cross_attention_dim (line 83) | def cross_attention_dim(self): method dummy_unet (line 87) | def dummy_unet(self): method dummy_movq_kwargs (line 112) | def dummy_movq_kwargs(self): method dummy_movq (line 133) | def dummy_movq(self): method get_dummy_components (line 138) | def get_dummy_components(self): method get_dummy_inputs (line 164) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky_controlnet_img2img (line 195) | def test_kandinsky_controlnet_img2img(self): method test_inference_batch_single_identical (line 228) | def test_inference_batch_single_identical(self): method test_float16_inference (line 231) | def test_float16_inference(self): class KandinskyV22ControlnetImg2ImgPipelineIntegrationTests (line 237) | class KandinskyV22ControlnetImg2ImgPipelineIntegrationTests(unittest.Tes... method setUp (line 238) | def setUp(self): method tearDown (line 244) | def tearDown(self): method test_kandinsky_controlnet_img2img (line 250) | def test_kandinsky_controlnet_img2img(self): FILE: diffusers/tests/pipelines/kandinsky2_2/test_kandinsky_img2img.py class Dummies (line 47) | class Dummies: method text_embedder_hidden_size (line 49) | def text_embedder_hidden_size(self): method time_input_dim (line 53) | def time_input_dim(self): method block_out_channels_0 (line 57) | def block_out_channels_0(self): method time_embed_dim (line 61) | def time_embed_dim(self): method cross_attention_dim (line 65) | def cross_attention_dim(self): method dummy_unet (line 69) | def dummy_unet(self): method dummy_movq_kwargs (line 94) | def dummy_movq_kwargs(self): method dummy_movq (line 113) | def dummy_movq(self): method get_dummy_components (line 118) | def get_dummy_components(self): method get_dummy_inputs (line 144) | def get_dummy_inputs(self, device, seed=0): class KandinskyV22Img2ImgPipelineFastTests (line 173) | class KandinskyV22Img2ImgPipelineFastTests(PipelineTesterMixin, unittest... method get_dummy_components (line 197) | def get_dummy_components(self): method get_dummy_inputs (line 201) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky_img2img (line 205) | def test_kandinsky_img2img(self): method test_float16_inference (line 236) | def test_float16_inference(self): class KandinskyV22Img2ImgPipelineIntegrationTests (line 242) | class KandinskyV22Img2ImgPipelineIntegrationTests(unittest.TestCase): method setUp (line 243) | def setUp(self): method tearDown (line 249) | def tearDown(self): method test_kandinsky_img2img (line 255) | def test_kandinsky_img2img(self): FILE: diffusers/tests/pipelines/kandinsky2_2/test_kandinsky_inpaint.py class Dummies (line 49) | class Dummies: method text_embedder_hidden_size (line 51) | def text_embedder_hidden_size(self): method time_input_dim (line 55) | def time_input_dim(self): method block_out_channels_0 (line 59) | def block_out_channels_0(self): method time_embed_dim (line 63) | def time_embed_dim(self): method cross_attention_dim (line 67) | def cross_attention_dim(self): method dummy_unet (line 71) | def dummy_unet(self): method dummy_movq_kwargs (line 96) | def dummy_movq_kwargs(self): method dummy_movq (line 115) | def dummy_movq(self): method get_dummy_components (line 120) | def get_dummy_components(self): method get_dummy_inputs (line 144) | def get_dummy_inputs(self, device, seed=0): class KandinskyV22InpaintPipelineFastTests (line 176) | class KandinskyV22InpaintPipelineFastTests(PipelineTesterMixin, unittest... method get_dummy_components (line 201) | def get_dummy_components(self): method get_dummy_inputs (line 205) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky_inpaint (line 209) | def test_kandinsky_inpaint(self): method test_inference_batch_single_identical (line 243) | def test_inference_batch_single_identical(self): method test_float16_inference (line 246) | def test_float16_inference(self): method test_model_cpu_offload_forward_pass (line 250) | def test_model_cpu_offload_forward_pass(self): method test_save_load_optional_components (line 253) | def test_save_load_optional_components(self): method test_sequential_cpu_offload_forward_pass (line 256) | def test_sequential_cpu_offload_forward_pass(self): method test_callback_inputs (line 260) | def test_callback_inputs(self): class KandinskyV22InpaintPipelineIntegrationTests (line 296) | class KandinskyV22InpaintPipelineIntegrationTests(unittest.TestCase): method setUp (line 297) | def setUp(self): method tearDown (line 303) | def tearDown(self): method test_kandinsky_inpaint (line 309) | def test_kandinsky_inpaint(self): FILE: diffusers/tests/pipelines/kandinsky2_2/test_kandinsky_prior.py class Dummies (line 40) | class Dummies: method text_embedder_hidden_size (line 42) | def text_embedder_hidden_size(self): method time_input_dim (line 46) | def time_input_dim(self): method block_out_channels_0 (line 50) | def block_out_channels_0(self): method time_embed_dim (line 54) | def time_embed_dim(self): method cross_attention_dim (line 58) | def cross_attention_dim(self): method dummy_tokenizer (line 62) | def dummy_tokenizer(self): method dummy_text_encoder (line 67) | def dummy_text_encoder(self): method dummy_prior (line 84) | def dummy_prior(self): method dummy_image_encoder (line 100) | def dummy_image_encoder(self): method dummy_image_processor (line 117) | def dummy_image_processor(self): method get_dummy_components (line 131) | def get_dummy_components(self): method get_dummy_inputs (line 157) | def get_dummy_inputs(self, device, seed=0): class KandinskyV22PriorPipelineFastTests (line 172) | class KandinskyV22PriorPipelineFastTests(PipelineTesterMixin, unittest.T... method get_dummy_components (line 189) | def get_dummy_components(self): method get_dummy_inputs (line 193) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky_prior (line 197) | def test_kandinsky_prior(self): method test_inference_batch_single_identical (line 228) | def test_inference_batch_single_identical(self): method test_attention_slicing_forward_pass (line 232) | def test_attention_slicing_forward_pass(self): method test_callback_inputs (line 242) | def test_callback_inputs(self): FILE: diffusers/tests/pipelines/kandinsky2_2/test_kandinsky_prior_emb2emb.py class KandinskyV22PriorEmb2EmbPipelineFastTests (line 41) | class KandinskyV22PriorEmb2EmbPipelineFastTests(PipelineTesterMixin, uni... method text_embedder_hidden_size (line 58) | def text_embedder_hidden_size(self): method time_input_dim (line 62) | def time_input_dim(self): method block_out_channels_0 (line 66) | def block_out_channels_0(self): method time_embed_dim (line 70) | def time_embed_dim(self): method cross_attention_dim (line 74) | def cross_attention_dim(self): method dummy_tokenizer (line 78) | def dummy_tokenizer(self): method dummy_text_encoder (line 83) | def dummy_text_encoder(self): method dummy_prior (line 100) | def dummy_prior(self): method dummy_image_encoder (line 116) | def dummy_image_encoder(self): method dummy_image_processor (line 133) | def dummy_image_processor(self): method get_dummy_components (line 147) | def get_dummy_components(self): method get_dummy_inputs (line 173) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky_prior_emb2emb (line 194) | def test_kandinsky_prior_emb2emb(self): method test_inference_batch_single_identical (line 236) | def test_inference_batch_single_identical(self): method test_attention_slicing_forward_pass (line 240) | def test_attention_slicing_forward_pass(self): FILE: diffusers/tests/pipelines/kandinsky3/test_kandinsky3.py class Kandinsky3PipelineFastTests (line 52) | class Kandinsky3PipelineFastTests(PipelineTesterMixin, unittest.TestCase): method dummy_movq_kwargs (line 62) | def dummy_movq_kwargs(self): method dummy_movq (line 81) | def dummy_movq(self): method get_dummy_components (line 86) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_inputs (line 123) | def get_dummy_inputs(self, device, seed=0): method test_kandinsky3 (line 139) | def test_kandinsky3(self): method test_float16_inference (line 162) | def test_float16_inference(self): method test_inference_batch_single_identical (line 165) | def test_inference_batch_single_identical(self): class Kandinsky3PipelineIntegrationTests (line 171) | class Kandinsky3PipelineIntegrationTests(unittest.TestCase): method setUp (line 172) | def setUp(self): method tearDown (line 178) | def tearDown(self): method test_kandinskyV3 (line 184) | def test_kandinskyV3(self): method test_kandinskyV3_img2img (line 210) | def test_kandinskyV3_img2img(self): FILE: diffusers/tests/pipelines/kandinsky3/test_kandinsky3_img2img.py class Kandinsky3Img2ImgPipelineFastTests (line 55) | class Kandinsky3Img2ImgPipelineFastTests(PipelineTesterMixin, unittest.T... method dummy_movq_kwargs (line 74) | def dummy_movq_kwargs(self): method dummy_movq (line 93) | def dummy_movq(self): method get_dummy_components (line 98) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_inputs (line 135) | def get_dummy_inputs(self, device, seed=0): method test_dict_tuple_outputs_equivalent (line 156) | def test_dict_tuple_outputs_equivalent(self): method test_kandinsky3_img2img (line 162) | def test_kandinsky3_img2img(self): method test_float16_inference (line 187) | def test_float16_inference(self): method test_inference_batch_single_identical (line 190) | def test_inference_batch_single_identical(self): class Kandinsky3Img2ImgPipelineIntegrationTests (line 196) | class Kandinsky3Img2ImgPipelineIntegrationTests(unittest.TestCase): method setUp (line 197) | def setUp(self): method tearDown (line 203) | def tearDown(self): method test_kandinskyV3_img2img (line 209) | def test_kandinskyV3_img2img(self): FILE: diffusers/tests/pipelines/latent_consistency_models/test_latent_consistency_models.py class LatentConsistencyModelPipelineFastTests (line 29) | class LatentConsistencyModelPipelineFastTests( method get_dummy_components (line 38) | def get_dummy_components(self): method get_dummy_inputs (line 97) | def get_dummy_inputs(self, device, seed=0): method test_ip_adapter_single (line 111) | def test_ip_adapter_single(self): method test_lcm_onestep (line 117) | def test_lcm_onestep(self): method test_lcm_multistep (line 135) | def test_lcm_multistep(self): method test_lcm_custom_timesteps (line 152) | def test_lcm_custom_timesteps(self): method test_inference_batch_single_identical (line 171) | def test_inference_batch_single_identical(self): method test_callback_cfg (line 175) | def test_callback_cfg(self): method test_callback_inputs (line 179) | def test_callback_inputs(self): class LatentConsistencyModelPipelineSlowTests (line 219) | class LatentConsistencyModelPipelineSlowTests(unittest.TestCase): method setUp (line 220) | def setUp(self): method get_inputs (line 224) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_lcm_onestep (line 238) | def test_lcm_onestep(self): method test_lcm_multistep (line 253) | def test_lcm_multistep(self): FILE: diffusers/tests/pipelines/latent_consistency_models/test_latent_consistency_models_img2img.py class LatentConsistencyModelImg2ImgPipelineFastTests (line 36) | class LatentConsistencyModelImg2ImgPipelineFastTests( method get_dummy_components (line 46) | def get_dummy_components(self): method get_dummy_inputs (line 105) | def get_dummy_inputs(self, device, seed=0): method test_ip_adapter_single (line 122) | def test_ip_adapter_single(self): method test_lcm_onestep (line 128) | def test_lcm_onestep(self): method test_lcm_multistep (line 146) | def test_lcm_multistep(self): method test_lcm_custom_timesteps (line 163) | def test_lcm_custom_timesteps(self): method test_inference_batch_single_identical (line 182) | def test_inference_batch_single_identical(self): method test_callback_inputs (line 186) | def test_callback_inputs(self): class LatentConsistencyModelImg2ImgPipelineSlowTests (line 226) | class LatentConsistencyModelImg2ImgPipelineSlowTests(unittest.TestCase): method setUp (line 227) | def setUp(self): method get_inputs (line 231) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_lcm_onestep (line 252) | def test_lcm_onestep(self): method test_lcm_multistep (line 269) | def test_lcm_multistep(self): FILE: diffusers/tests/pipelines/latent_diffusion/test_latent_diffusion.py class LDMTextToImagePipelineFastTests (line 39) | class LDMTextToImagePipelineFastTests(PipelineTesterMixin, unittest.Test... method get_dummy_components (line 54) | def get_dummy_components(self): method get_dummy_inputs (line 106) | def get_dummy_inputs(self, device, seed=0): method test_inference_text2img (line 120) | def test_inference_text2img(self): class LDMTextToImagePipelineSlowTests (line 140) | class LDMTextToImagePipelineSlowTests(unittest.TestCase): method setUp (line 141) | def setUp(self): method tearDown (line 146) | def tearDown(self): method get_inputs (line 151) | def get_inputs(self, device, dtype=torch.float32, seed=0): method test_ldm_default_ddim (line 165) | def test_ldm_default_ddim(self): class LDMTextToImagePipelineNightlyTests (line 181) | class LDMTextToImagePipelineNightlyTests(unittest.TestCase): method setUp (line 182) | def setUp(self): method tearDown (line 187) | def tearDown(self): method get_inputs (line 192) | def get_inputs(self, device, dtype=torch.float32, seed=0): method test_ldm_default_ddim (line 206) | def test_ldm_default_ddim(self): FILE: diffusers/tests/pipelines/latent_diffusion/test_latent_diffusion_superresolution.py class LDMSuperResolutionPipelineFastTests (line 37) | class LDMSuperResolutionPipelineFastTests(unittest.TestCase): method dummy_image (line 39) | def dummy_image(self): method dummy_uncond_unet (line 48) | def dummy_uncond_unet(self): method dummy_vq_model (line 62) | def dummy_vq_model(self): method test_inference_superresolution (line 74) | def test_inference_superresolution(self): method test_inference_superresolution_fp16 (line 97) | def test_inference_superresolution_fp16(self): class LDMSuperResolutionPipelineIntegrationTests (line 119) | class LDMSuperResolutionPipelineIntegrationTests(unittest.TestCase): method test_inference_superresolution (line 120) | def test_inference_superresolution(self): FILE: diffusers/tests/pipelines/ledits_pp/test_ledits_pp_stable_diffusion.py class LEditsPPPipelineStableDiffusionFastTests (line 46) | class LEditsPPPipelineStableDiffusionFastTests(unittest.TestCase): method get_dummy_components (line 49) | def get_dummy_components(self): method get_dummy_inputs (line 97) | def get_dummy_inputs(self, device, seed=0): method get_dummy_inversion_inputs (line 110) | def get_dummy_inversion_inputs(self, device, seed=0): method test_ledits_pp_inversion (line 131) | def test_ledits_pp_inversion(self): method test_ledits_pp_inversion_batch (line 153) | def test_ledits_pp_inversion_batch(self): method test_ledits_pp_warmup_steps (line 179) | def test_ledits_pp_warmup_steps(self): class LEditsPPPipelineStableDiffusionSlowTests (line 206) | class LEditsPPPipelineStableDiffusionSlowTests(unittest.TestCase): method setUp (line 207) | def setUp(self): method tearDown (line 212) | def tearDown(self): method setUpClass (line 218) | def setUpClass(cls): method test_ledits_pp_editing (line 225) | def test_ledits_pp_editing(self): FILE: diffusers/tests/pipelines/ledits_pp/test_ledits_pp_stable_diffusion_xl.py class LEditsPPPipelineStableDiffusionXLFastTests (line 55) | class LEditsPPPipelineStableDiffusionXLFastTests(unittest.TestCase): method get_dummy_components (line 58) | def get_dummy_components(self, skip_first_text_encoder=False, time_con... method get_dummy_inputs (line 147) | def get_dummy_inputs(self, device, seed=0): method get_dummy_inversion_inputs (line 160) | def get_dummy_inversion_inputs(self, device, seed=0): method test_ledits_pp_inversion (line 181) | def test_ledits_pp_inversion(self): method test_ledits_pp_inversion_batch (line 202) | def test_ledits_pp_inversion_batch(self): method test_ledits_pp_warmup_steps (line 229) | def test_ledits_pp_warmup_steps(self): class LEditsPPPipelineStableDiffusionXLSlowTests (line 257) | class LEditsPPPipelineStableDiffusionXLSlowTests(unittest.TestCase): method setUpClass (line 259) | def setUpClass(cls): method test_ledits_pp_edit (line 266) | def test_ledits_pp_edit(self): FILE: diffusers/tests/pipelines/marigold/test_marigold_depth.py class MarigoldDepthPipelineFastTests (line 48) | class MarigoldDepthPipelineFastTests(PipelineTesterMixin, unittest.TestC... method get_dummy_components (line 64) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_tiny_autoencoder (line 123) | def get_dummy_tiny_autoencoder(self): method get_dummy_inputs (line 126) | def get_dummy_inputs(self, device, seed=0): method _test_marigold_depth (line 142) | def _test_marigold_depth( method test_marigold_depth_dummy_defaults (line 175) | def test_marigold_depth_dummy_defaults(self): method test_marigold_depth_dummy_G0_S1_P32_E1_B1_M1 (line 180) | def test_marigold_depth_dummy_G0_S1_P32_E1_B1_M1(self): method test_marigold_depth_dummy_G0_S1_P16_E1_B1_M1 (line 191) | def test_marigold_depth_dummy_G0_S1_P16_E1_B1_M1(self): method test_marigold_depth_dummy_G2024_S1_P32_E1_B1_M1 (line 202) | def test_marigold_depth_dummy_G2024_S1_P32_E1_B1_M1(self): method test_marigold_depth_dummy_G0_S2_P32_E1_B1_M1 (line 213) | def test_marigold_depth_dummy_G0_S2_P32_E1_B1_M1(self): method test_marigold_depth_dummy_G0_S1_P64_E1_B1_M1 (line 224) | def test_marigold_depth_dummy_G0_S1_P64_E1_B1_M1(self): method test_marigold_depth_dummy_G0_S1_P32_E3_B1_M1 (line 235) | def test_marigold_depth_dummy_G0_S1_P32_E3_B1_M1(self): method test_marigold_depth_dummy_G0_S1_P32_E4_B2_M1 (line 247) | def test_marigold_depth_dummy_G0_S1_P32_E4_B2_M1(self): method test_marigold_depth_dummy_G0_S1_P16_E1_B1_M0 (line 259) | def test_marigold_depth_dummy_G0_S1_P16_E1_B1_M0(self): method test_marigold_depth_dummy_no_num_inference_steps (line 270) | def test_marigold_depth_dummy_no_num_inference_steps(self): method test_marigold_depth_dummy_no_processing_resolution (line 278) | def test_marigold_depth_dummy_no_processing_resolution(self): class MarigoldDepthPipelineIntegrationTests (line 289) | class MarigoldDepthPipelineIntegrationTests(unittest.TestCase): method setUp (line 290) | def setUp(self): method tearDown (line 295) | def tearDown(self): method _test_marigold_depth (line 300) | def _test_marigold_depth( method test_marigold_depth_einstein_f32_cpu_G0_S1_P32_E1_B1_M1 (line 342) | def test_marigold_depth_einstein_f32_cpu_G0_S1_P32_E1_B1_M1(self): method test_marigold_depth_einstein_f32_cuda_G0_S1_P768_E1_B1_M1 (line 355) | def test_marigold_depth_einstein_f32_cuda_G0_S1_P768_E1_B1_M1(self): method test_marigold_depth_einstein_f16_cuda_G0_S1_P768_E1_B1_M1 (line 368) | def test_marigold_depth_einstein_f16_cuda_G0_S1_P768_E1_B1_M1(self): method test_marigold_depth_einstein_f16_cuda_G2024_S1_P768_E1_B1_M1 (line 381) | def test_marigold_depth_einstein_f16_cuda_G2024_S1_P768_E1_B1_M1(self): method test_marigold_depth_einstein_f16_cuda_G0_S2_P768_E1_B1_M1 (line 394) | def test_marigold_depth_einstein_f16_cuda_G0_S2_P768_E1_B1_M1(self): method test_marigold_depth_einstein_f16_cuda_G0_S1_P512_E1_B1_M1 (line 407) | def test_marigold_depth_einstein_f16_cuda_G0_S1_P512_E1_B1_M1(self): method test_marigold_depth_einstein_f16_cuda_G0_S1_P768_E3_B1_M1 (line 420) | def test_marigold_depth_einstein_f16_cuda_G0_S1_P768_E3_B1_M1(self): method test_marigold_depth_einstein_f16_cuda_G0_S1_P768_E4_B2_M1 (line 434) | def test_marigold_depth_einstein_f16_cuda_G0_S1_P768_E4_B2_M1(self): method test_marigold_depth_einstein_f16_cuda_G0_S1_P512_E1_B1_M0 (line 448) | def test_marigold_depth_einstein_f16_cuda_G0_S1_P512_E1_B1_M0(self): FILE: diffusers/tests/pipelines/marigold/test_marigold_normals.py class MarigoldNormalsPipelineFastTests (line 48) | class MarigoldNormalsPipelineFastTests(PipelineTesterMixin, unittest.Tes... method get_dummy_components (line 64) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_tiny_autoencoder (line 123) | def get_dummy_tiny_autoencoder(self): method get_dummy_inputs (line 126) | def get_dummy_inputs(self, device, seed=0): method _test_marigold_normals (line 142) | def _test_marigold_normals( method test_marigold_depth_dummy_defaults (line 175) | def test_marigold_depth_dummy_defaults(self): method test_marigold_depth_dummy_G0_S1_P32_E1_B1_M1 (line 180) | def test_marigold_depth_dummy_G0_S1_P32_E1_B1_M1(self): method test_marigold_depth_dummy_G0_S1_P16_E1_B1_M1 (line 191) | def test_marigold_depth_dummy_G0_S1_P16_E1_B1_M1(self): method test_marigold_depth_dummy_G2024_S1_P32_E1_B1_M1 (line 202) | def test_marigold_depth_dummy_G2024_S1_P32_E1_B1_M1(self): method test_marigold_depth_dummy_G0_S2_P32_E1_B1_M1 (line 213) | def test_marigold_depth_dummy_G0_S2_P32_E1_B1_M1(self): method test_marigold_depth_dummy_G0_S1_P64_E1_B1_M1 (line 224) | def test_marigold_depth_dummy_G0_S1_P64_E1_B1_M1(self): method test_marigold_depth_dummy_G0_S1_P32_E3_B1_M1 (line 235) | def test_marigold_depth_dummy_G0_S1_P32_E3_B1_M1(self): method test_marigold_depth_dummy_G0_S1_P32_E4_B2_M1 (line 247) | def test_marigold_depth_dummy_G0_S1_P32_E4_B2_M1(self): method test_marigold_depth_dummy_G0_S1_P16_E1_B1_M0 (line 259) | def test_marigold_depth_dummy_G0_S1_P16_E1_B1_M0(self): method test_marigold_depth_dummy_no_num_inference_steps (line 270) | def test_marigold_depth_dummy_no_num_inference_steps(self): method test_marigold_depth_dummy_no_processing_resolution (line 278) | def test_marigold_depth_dummy_no_processing_resolution(self): class MarigoldNormalsPipelineIntegrationTests (line 289) | class MarigoldNormalsPipelineIntegrationTests(unittest.TestCase): method setUp (line 290) | def setUp(self): method tearDown (line 295) | def tearDown(self): method _test_marigold_normals (line 300) | def _test_marigold_normals( method test_marigold_normals_einstein_f32_cpu_G0_S1_P32_E1_B1_M1 (line 342) | def test_marigold_normals_einstein_f32_cpu_G0_S1_P32_E1_B1_M1(self): method test_marigold_normals_einstein_f32_cuda_G0_S1_P768_E1_B1_M1 (line 355) | def test_marigold_normals_einstein_f32_cuda_G0_S1_P768_E1_B1_M1(self): method test_marigold_normals_einstein_f16_cuda_G0_S1_P768_E1_B1_M1 (line 368) | def test_marigold_normals_einstein_f16_cuda_G0_S1_P768_E1_B1_M1(self): method test_marigold_normals_einstein_f16_cuda_G2024_S1_P768_E1_B1_M1 (line 381) | def test_marigold_normals_einstein_f16_cuda_G2024_S1_P768_E1_B1_M1(self): method test_marigold_normals_einstein_f16_cuda_G0_S2_P768_E1_B1_M1 (line 394) | def test_marigold_normals_einstein_f16_cuda_G0_S2_P768_E1_B1_M1(self): method test_marigold_normals_einstein_f16_cuda_G0_S1_P512_E1_B1_M1 (line 407) | def test_marigold_normals_einstein_f16_cuda_G0_S1_P512_E1_B1_M1(self): method test_marigold_normals_einstein_f16_cuda_G0_S1_P768_E3_B1_M1 (line 420) | def test_marigold_normals_einstein_f16_cuda_G0_S1_P768_E3_B1_M1(self): method test_marigold_normals_einstein_f16_cuda_G0_S1_P768_E4_B2_M1 (line 434) | def test_marigold_normals_einstein_f16_cuda_G0_S1_P768_E4_B2_M1(self): method test_marigold_normals_einstein_f16_cuda_G0_S1_P512_E1_B1_M0 (line 448) | def test_marigold_normals_einstein_f16_cuda_G0_S1_P512_E1_B1_M0(self): FILE: diffusers/tests/pipelines/musicldm/test_musicldm.py class MusicLDMPipelineFastTests (line 51) | class MusicLDMPipelineFastTests(PipelineTesterMixin, unittest.TestCase): method get_dummy_components (line 68) | def get_dummy_components(self): method get_dummy_inputs (line 159) | def get_dummy_inputs(self, device, seed=0): method test_musicldm_ddim (line 172) | def test_musicldm_ddim(self): method test_musicldm_prompt_embeds (line 194) | def test_musicldm_prompt_embeds(self): method test_musicldm_negative_prompt_embeds (line 230) | def test_musicldm_negative_prompt_embeds(self): method test_musicldm_negative_prompt (line 273) | def test_musicldm_negative_prompt(self): method test_musicldm_num_waveforms_per_prompt (line 296) | def test_musicldm_num_waveforms_per_prompt(self): method test_musicldm_audio_length_in_s (line 331) | def test_musicldm_audio_length_in_s(self): method test_musicldm_vocoder_model_in_dim (line 352) | def test_musicldm_vocoder_model_in_dim(self): method test_attention_slicing_forward_pass (line 372) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 375) | def test_inference_batch_single_identical(self): method test_xformers_attention_forwardGenerator_pass (line 382) | def test_xformers_attention_forwardGenerator_pass(self): method test_to_dtype (line 385) | def test_to_dtype(self): class MusicLDMPipelineNightlyTests (line 410) | class MusicLDMPipelineNightlyTests(unittest.TestCase): method setUp (line 411) | def setUp(self): method tearDown (line 416) | def tearDown(self): method get_inputs (line 421) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_musicldm (line 434) | def test_musicldm(self): method test_musicldm_lms (line 454) | def test_musicldm_lms(self): FILE: diffusers/tests/pipelines/paint_by_example/test_paint_by_example.py class PaintByExamplePipelineFastTests (line 43) | class PaintByExamplePipelineFastTests(PipelineTesterMixin, unittest.Test... method get_dummy_components (line 49) | def get_dummy_components(self): method convert_to_pt (line 95) | def convert_to_pt(self, image): method get_dummy_inputs (line 101) | def get_dummy_inputs(self, device="cpu", seed=0): method test_paint_by_example_inpaint (line 124) | def test_paint_by_example_inpaint(self): method test_paint_by_example_image_tensor (line 143) | def test_paint_by_example_image_tensor(self): method test_inference_batch_single_identical (line 170) | def test_inference_batch_single_identical(self): class PaintByExamplePipelineIntegrationTests (line 176) | class PaintByExamplePipelineIntegrationTests(unittest.TestCase): method setUp (line 177) | def setUp(self): method tearDown (line 183) | def tearDown(self): method test_paint_by_example (line 189) | def test_paint_by_example(self): FILE: diffusers/tests/pipelines/pia/test_pia.py function to_np (line 23) | def to_np(tensor): class PIAPipelineFastTests (line 30) | class PIAPipelineFastTests(IPAdapterTesterMixin, PipelineTesterMixin, Pi... method get_dummy_components (line 56) | def get_dummy_components(self): method get_dummy_inputs (line 118) | def get_dummy_inputs(self, device, seed=0): method test_motion_unet_loading (line 135) | def test_motion_unet_loading(self): method test_ip_adapter_single (line 141) | def test_ip_adapter_single(self): method test_dict_tuple_outputs_equivalent (line 178) | def test_dict_tuple_outputs_equivalent(self): method test_attention_slicing_forward_pass (line 185) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 188) | def test_inference_batch_single_identical( method test_to_device (line 244) | def test_to_device(self): method test_to_dtype (line 268) | def test_to_dtype(self): method test_prompt_embeds (line 281) | def test_prompt_embeds(self): method test_free_init (line 292) | def test_free_init(self): method test_xformers_attention_forwardGenerator_pass (line 331) | def test_xformers_attention_forwardGenerator_pass(self): FILE: diffusers/tests/pipelines/pixart_alpha/test_pixart.py class PixArtAlphaPipelineFastTests (line 45) | class PixArtAlphaPipelineFastTests(PipelineTesterMixin, unittest.TestCase): method get_dummy_components (line 54) | def get_dummy_components(self): method get_dummy_inputs (line 90) | def get_dummy_inputs(self, device, seed=0): method test_sequential_cpu_offload_forward_pass (line 105) | def test_sequential_cpu_offload_forward_pass(self): method test_save_load_optional_components (line 109) | def test_save_load_optional_components(self): method test_inference (line 184) | def test_inference(self): method test_inference_non_square_images (line 201) | def test_inference_non_square_images(self): method test_inference_with_embeddings_and_multiple_images (line 218) | def test_inference_with_embeddings_and_multiple_images(self): method test_inference_with_multiple_images_per_prompt (line 290) | def test_inference_with_multiple_images_per_prompt(self): method test_raises_warning_for_mask_feature (line 308) | def test_raises_warning_for_mask_feature(self): method test_inference_batch_single_identical (line 324) | def test_inference_batch_single_identical(self): class PixArtAlphaPipelineIntegrationTests (line 330) | class PixArtAlphaPipelineIntegrationTests(unittest.TestCase): method setUp (line 335) | def setUp(self): method tearDown (line 340) | def tearDown(self): method test_pixart_1024 (line 345) | def test_pixart_1024(self): method test_pixart_512 (line 360) | def test_pixart_512(self): method test_pixart_1024_without_resolution_binning (line 376) | def test_pixart_1024_without_resolution_binning(self): method test_pixart_512_without_resolution_binning (line 410) | def test_pixart_512_without_resolution_binning(self): FILE: diffusers/tests/pipelines/pixart_sigma/test_pixart.py class PixArtSigmaPipelineFastTests (line 45) | class PixArtSigmaPipelineFastTests(PipelineTesterMixin, unittest.TestCase): method get_dummy_components (line 54) | def get_dummy_components(self): method get_dummy_inputs (line 90) | def get_dummy_inputs(self, device, seed=0): method test_sequential_cpu_offload_forward_pass (line 105) | def test_sequential_cpu_offload_forward_pass(self): method test_save_load_optional_components (line 109) | def test_save_load_optional_components(self): method test_inference (line 184) | def test_inference(self): method test_inference_non_square_images (line 201) | def test_inference_non_square_images(self): method test_inference_with_embeddings_and_multiple_images (line 218) | def test_inference_with_embeddings_and_multiple_images(self): method test_inference_with_multiple_images_per_prompt (line 290) | def test_inference_with_multiple_images_per_prompt(self): method test_inference_batch_single_identical (line 308) | def test_inference_batch_single_identical(self): class PixArtSigmaPipelineIntegrationTests (line 314) | class PixArtSigmaPipelineIntegrationTests(unittest.TestCase): method setUp (line 319) | def setUp(self): method tearDown (line 324) | def tearDown(self): method test_pixart_1024 (line 329) | def test_pixart_1024(self): method test_pixart_512 (line 344) | def test_pixart_512(self): method test_pixart_1024_without_resolution_binning (line 365) | def test_pixart_1024_without_resolution_binning(self): method test_pixart_512_without_resolution_binning (line 399) | def test_pixart_512_without_resolution_binning(self): FILE: diffusers/tests/pipelines/pndm/test_pndm.py class PNDMPipelineFastTests (line 28) | class PNDMPipelineFastTests(unittest.TestCase): method dummy_uncond_unet (line 30) | def dummy_uncond_unet(self): method test_inference (line 43) | def test_inference(self): class PNDMPipelineIntegrationTests (line 69) | class PNDMPipelineIntegrationTests(unittest.TestCase): method test_inference_cifar10 (line 70) | def test_inference_cifar10(self): FILE: diffusers/tests/pipelines/semantic_stable_diffusion/test_semantic_diffusion.py class SafeDiffusionPipelineFastTests (line 39) | class SafeDiffusionPipelineFastTests(unittest.TestCase): method setUp (line 40) | def setUp(self): method tearDown (line 46) | def tearDown(self): method dummy_image (line 53) | def dummy_image(self): method dummy_cond_unet (line 62) | def dummy_cond_unet(self): method dummy_vae (line 77) | def dummy_vae(self): method dummy_text_encoder (line 90) | def dummy_text_encoder(self): method dummy_extractor (line 106) | def dummy_extractor(self): method test_semantic_diffusion_ddim (line 120) | def test_semantic_diffusion_ddim(self): method test_semantic_diffusion_pndm (line 173) | def test_semantic_diffusion_pndm(self): method test_semantic_diffusion_no_safety_checker (line 219) | def test_semantic_diffusion_no_safety_checker(self): method test_semantic_diffusion_fp16 (line 241) | def test_semantic_diffusion_fp16(self): class SemanticDiffusionPipelineIntegrationTests (line 275) | class SemanticDiffusionPipelineIntegrationTests(unittest.TestCase): method setUp (line 276) | def setUp(self): method tearDown (line 282) | def tearDown(self): method test_positive_guidance (line 288) | def test_positive_guidance(self): method test_negative_guidance (line 371) | def test_negative_guidance(self): method test_multi_cond_guidance (line 454) | def test_multi_cond_guidance(self): method test_guidance_fp16 (line 537) | def test_guidance_fp16(self): FILE: diffusers/tests/pipelines/shap_e/test_shap_e.py class ShapEPipelineFastTests (line 29) | class ShapEPipelineFastTests(PipelineTesterMixin, unittest.TestCase): method text_embedder_hidden_size (line 46) | def text_embedder_hidden_size(self): method time_input_dim (line 50) | def time_input_dim(self): method time_embed_dim (line 54) | def time_embed_dim(self): method renderer_dim (line 58) | def renderer_dim(self): method dummy_tokenizer (line 62) | def dummy_tokenizer(self): method dummy_text_encoder (line 67) | def dummy_text_encoder(self): method dummy_prior (line 84) | def dummy_prior(self): method dummy_renderer (line 107) | def dummy_renderer(self): method get_dummy_components (line 129) | def get_dummy_components(self): method get_dummy_inputs (line 153) | def get_dummy_inputs(self, device, seed=0): method test_shap_e (line 167) | def test_shap_e(self): method test_inference_batch_consistent (line 187) | def test_inference_batch_consistent(self): method test_inference_batch_single_identical (line 191) | def test_inference_batch_single_identical(self): method test_num_images_per_prompt (line 194) | def test_num_images_per_prompt(self): method test_float16_inference (line 213) | def test_float16_inference(self): method test_save_load_local (line 216) | def test_save_load_local(self): method test_sequential_cpu_offload_forward_pass (line 220) | def test_sequential_cpu_offload_forward_pass(self): class ShapEPipelineIntegrationTests (line 226) | class ShapEPipelineIntegrationTests(unittest.TestCase): method setUp (line 227) | def setUp(self): method tearDown (line 233) | def tearDown(self): method test_shap_e (line 239) | def test_shap_e(self): FILE: diffusers/tests/pipelines/shap_e/test_shap_e_img2img.py class ShapEImg2ImgPipelineFastTests (line 37) | class ShapEImg2ImgPipelineFastTests(PipelineTesterMixin, unittest.TestCa... method text_embedder_hidden_size (line 54) | def text_embedder_hidden_size(self): method time_input_dim (line 58) | def time_input_dim(self): method time_embed_dim (line 62) | def time_embed_dim(self): method renderer_dim (line 66) | def renderer_dim(self): method dummy_image_encoder (line 70) | def dummy_image_encoder(self): method dummy_image_processor (line 87) | def dummy_image_processor(self): method dummy_prior (line 102) | def dummy_prior(self): method dummy_renderer (line 126) | def dummy_renderer(self): method get_dummy_components (line 148) | def get_dummy_components(self): method get_dummy_inputs (line 172) | def get_dummy_inputs(self, device, seed=0): method test_shap_e (line 188) | def test_shap_e(self): method test_inference_batch_consistent (line 210) | def test_inference_batch_consistent(self): method test_inference_batch_single_identical (line 214) | def test_inference_batch_single_identical(self): method test_num_images_per_prompt (line 220) | def test_num_images_per_prompt(self): method test_float16_inference (line 239) | def test_float16_inference(self): method test_save_load_local (line 242) | def test_save_load_local(self): method test_sequential_cpu_offload_forward_pass (line 246) | def test_sequential_cpu_offload_forward_pass(self): class ShapEImg2ImgPipelineIntegrationTests (line 252) | class ShapEImg2ImgPipelineIntegrationTests(unittest.TestCase): method setUp (line 253) | def setUp(self): method tearDown (line 259) | def tearDown(self): method test_shap_e_img2img (line 265) | def test_shap_e_img2img(self): FILE: diffusers/tests/pipelines/stable_cascade/test_stable_cascade_combined.py class StableCascadeCombinedPipelineFastTests (line 33) | class StableCascadeCombinedPipelineFastTests(PipelineTesterMixin, unitte... method text_embedder_hidden_size (line 53) | def text_embedder_hidden_size(self): method dummy_prior (line 57) | def dummy_prior(self): method dummy_tokenizer (line 76) | def dummy_tokenizer(self): method dummy_text_encoder (line 81) | def dummy_text_encoder(self): method dummy_vqgan (line 98) | def dummy_vqgan(self): method dummy_decoder (line 109) | def dummy_decoder(self): method get_dummy_components (line 135) | def get_dummy_components(self): method get_dummy_inputs (line 162) | def get_dummy_inputs(self, device, seed=0): method test_stable_cascade (line 180) | def test_stable_cascade(self): method test_offloads (line 209) | def test_offloads(self): method test_inference_batch_single_identical (line 235) | def test_inference_batch_single_identical(self): method test_float16_inference (line 239) | def test_float16_inference(self): method test_callback_inputs (line 243) | def test_callback_inputs(self): method test_stable_cascade_combined_prompt_embeds (line 246) | def test_stable_cascade_combined_prompt_embeds(self): FILE: diffusers/tests/pipelines/stable_cascade/test_stable_cascade_decoder.py class StableCascadeDecoderPipelineFastTests (line 44) | class StableCascadeDecoderPipelineFastTests(PipelineTesterMixin, unittes... method text_embedder_hidden_size (line 61) | def text_embedder_hidden_size(self): method time_input_dim (line 65) | def time_input_dim(self): method block_out_channels_0 (line 69) | def block_out_channels_0(self): method time_embed_dim (line 73) | def time_embed_dim(self): method dummy_tokenizer (line 77) | def dummy_tokenizer(self): method dummy_text_encoder (line 82) | def dummy_text_encoder(self): method dummy_vqgan (line 99) | def dummy_vqgan(self): method dummy_decoder (line 110) | def dummy_decoder(self): method get_dummy_components (line 135) | def get_dummy_components(self): method get_dummy_inputs (line 154) | def get_dummy_inputs(self, device, seed=0): method test_wuerstchen_decoder (line 169) | def test_wuerstchen_decoder(self): method test_inference_batch_single_identical (line 194) | def test_inference_batch_single_identical(self): method test_attention_slicing_forward_pass (line 198) | def test_attention_slicing_forward_pass(self): method test_float16_inference (line 208) | def test_float16_inference(self): method test_stable_cascade_decoder_prompt_embeds (line 211) | def test_stable_cascade_decoder_prompt_embeds(self): method test_stable_cascade_decoder_single_prompt_multiple_image_embeddings (line 250) | def test_stable_cascade_decoder_single_prompt_multiple_image_embedding... method test_stable_cascade_decoder_single_prompt_multiple_image_embeddings_with_guidance (line 280) | def test_stable_cascade_decoder_single_prompt_multiple_image_embedding... class StableCascadeDecoderPipelineIntegrationTests (line 313) | class StableCascadeDecoderPipelineIntegrationTests(unittest.TestCase): method setUp (line 314) | def setUp(self): method tearDown (line 320) | def tearDown(self): method test_stable_cascade_decoder (line 326) | def test_stable_cascade_decoder(self): FILE: diffusers/tests/pipelines/stable_cascade/test_stable_cascade_prior.py function create_prior_lora_layers (line 52) | def create_prior_lora_layers(unet: nn.Module): class StableCascadePriorPipelineFastTests (line 65) | class StableCascadePriorPipelineFastTests(PipelineTesterMixin, unittest.... method text_embedder_hidden_size (line 83) | def text_embedder_hidden_size(self): method time_input_dim (line 87) | def time_input_dim(self): method block_out_channels_0 (line 91) | def block_out_channels_0(self): method time_embed_dim (line 95) | def time_embed_dim(self): method dummy_tokenizer (line 99) | def dummy_tokenizer(self): method dummy_text_encoder (line 104) | def dummy_text_encoder(self): method dummy_prior (line 121) | def dummy_prior(self): method get_dummy_components (line 140) | def get_dummy_components(self): method get_dummy_inputs (line 158) | def get_dummy_inputs(self, device, seed=0): method test_wuerstchen_prior (line 172) | def test_wuerstchen_prior(self): method test_inference_batch_single_identical (line 210) | def test_inference_batch_single_identical(self): method test_attention_slicing_forward_pass (line 214) | def test_attention_slicing_forward_pass(self): method test_float16_inference (line 224) | def test_float16_inference(self): method check_if_lora_correctly_set (line 227) | def check_if_lora_correctly_set(self, model) -> bool: method get_lora_components (line 236) | def get_lora_components(self): method test_inference_with_prior_lora (line 254) | def test_inference_with_prior_lora(self): method test_stable_cascade_decoder_prompt_embeds (line 277) | def test_stable_cascade_decoder_prompt_embeds(self): class StableCascadePriorPipelineIntegrationTests (line 315) | class StableCascadePriorPipelineIntegrationTests(unittest.TestCase): method setUp (line 316) | def setUp(self): method tearDown (line 322) | def tearDown(self): method test_stable_cascade_prior (line 328) | def test_stable_cascade_prior(self): FILE: diffusers/tests/pipelines/stable_diffusion/test_onnx_stable_diffusion.py class OnnxStableDiffusionPipelineFastTests (line 39) | class OnnxStableDiffusionPipelineFastTests(OnnxPipelineTesterMixin, unit... method get_dummy_inputs (line 42) | def get_dummy_inputs(self, seed=0): method test_pipeline_default_ddim (line 53) | def test_pipeline_default_ddim(self): method test_pipeline_pndm (line 66) | def test_pipeline_pndm(self): method test_pipeline_lms (line 80) | def test_pipeline_lms(self): method test_pipeline_euler (line 94) | def test_pipeline_euler(self): method test_pipeline_euler_ancestral (line 108) | def test_pipeline_euler_ancestral(self): method test_pipeline_dpm_multistep (line 122) | def test_pipeline_dpm_multistep(self): method test_stable_diffusion_prompt_embeds (line 136) | def test_stable_diffusion_prompt_embeds(self): method test_stable_diffusion_negative_prompt_embeds (line 169) | def test_stable_diffusion_negative_prompt_embeds(self): class OnnxStableDiffusionPipelineIntegrationTests (line 210) | class OnnxStableDiffusionPipelineIntegrationTests(unittest.TestCase): method gpu_provider (line 212) | def gpu_provider(self): method gpu_options (line 222) | def gpu_options(self): method test_inference_default_pndm (line 227) | def test_inference_default_pndm(self): method test_inference_ddim (line 251) | def test_inference_ddim(self): method test_inference_k_lms (line 277) | def test_inference_k_lms(self): method test_intermediate_state (line 303) | def test_intermediate_state(self): method test_stable_diffusion_no_safety_checker (line 353) | def test_stable_diffusion_no_safety_checker(self): FILE: diffusers/tests/pipelines/stable_diffusion/test_onnx_stable_diffusion_img2img.py class OnnxStableDiffusionImg2ImgPipelineFastTests (line 45) | class OnnxStableDiffusionImg2ImgPipelineFastTests(OnnxPipelineTesterMixi... method get_dummy_inputs (line 48) | def get_dummy_inputs(self, seed=0): method test_pipeline_default_ddim (line 62) | def test_pipeline_default_ddim(self): method test_pipeline_pndm (line 74) | def test_pipeline_pndm(self): method test_pipeline_lms (line 88) | def test_pipeline_lms(self): method test_pipeline_euler (line 105) | def test_pipeline_euler(self): method test_pipeline_euler_ancestral (line 119) | def test_pipeline_euler_ancestral(self): method test_pipeline_dpm_multistep (line 133) | def test_pipeline_dpm_multistep(self): class OnnxStableDiffusionImg2ImgPipelineIntegrationTests (line 151) | class OnnxStableDiffusionImg2ImgPipelineIntegrationTests(unittest.TestCa... method gpu_provider (line 153) | def gpu_provider(self): method gpu_options (line 163) | def gpu_options(self): method test_inference_default_pndm (line 168) | def test_inference_default_pndm(self): method test_inference_k_lms (line 206) | def test_inference_k_lms(self): FILE: diffusers/tests/pipelines/stable_diffusion/test_onnx_stable_diffusion_inpaint.py class OnnxStableDiffusionPipelineFastTests (line 36) | class OnnxStableDiffusionPipelineFastTests(OnnxPipelineTesterMixin, unit... class OnnxStableDiffusionInpaintPipelineIntegrationTests (line 44) | class OnnxStableDiffusionInpaintPipelineIntegrationTests(unittest.TestCa... method gpu_provider (line 46) | def gpu_provider(self): method gpu_options (line 56) | def gpu_options(self): method test_inference_default_pndm (line 61) | def test_inference_default_pndm(self): method test_inference_k_lms (line 100) | def test_inference_k_lms(self): FILE: diffusers/tests/pipelines/stable_diffusion/test_onnx_stable_diffusion_upscale.py class OnnxStableDiffusionUpscalePipelineFastTests (line 45) | class OnnxStableDiffusionUpscalePipelineFastTests(OnnxPipelineTesterMixi... method get_dummy_inputs (line 49) | def get_dummy_inputs(self, seed=0): method test_pipeline_default_ddpm (line 62) | def test_pipeline_default_ddpm(self): method test_pipeline_pndm (line 75) | def test_pipeline_pndm(self): method test_pipeline_dpm_multistep (line 88) | def test_pipeline_dpm_multistep(self): method test_pipeline_euler (line 104) | def test_pipeline_euler(self): method test_pipeline_euler_ancestral (line 119) | def test_pipeline_euler_ancestral(self): class OnnxStableDiffusionUpscalePipelineIntegrationTests (line 139) | class OnnxStableDiffusionUpscalePipelineIntegrationTests(unittest.TestCa... method gpu_provider (line 141) | def gpu_provider(self): method gpu_options (line 151) | def gpu_options(self): method test_inference_default_ddpm (line 156) | def test_inference_default_ddpm(self): method test_inference_k_lms (line 190) | def test_inference_k_lms(self): FILE: diffusers/tests/pipelines/stable_diffusion/test_stable_diffusion.py function _test_stable_diffusion_compile (line 81) | def _test_stable_diffusion_compile(in_queue, out_queue, timeout): class StableDiffusionPipelineFastTests (line 113) | class StableDiffusionPipelineFastTests( method get_dummy_components (line 127) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_inputs (line 187) | def get_dummy_inputs(self, device, seed=0): method test_stable_diffusion_ddim (line 201) | def test_stable_diffusion_ddim(self): method test_stable_diffusion_lcm (line 220) | def test_stable_diffusion_lcm(self): method test_stable_diffusion_lcm_custom_timesteps (line 240) | def test_stable_diffusion_lcm_custom_timesteps(self): method test_stable_diffusion_ays (line 262) | def test_stable_diffusion_ays(self): method test_stable_diffusion_prompt_embeds (line 300) | def test_stable_diffusion_prompt_embeds(self): method test_stable_diffusion_negative_prompt_embeds (line 336) | def test_stable_diffusion_negative_prompt_embeds(self): method test_stable_diffusion_prompt_embeds_no_text_encoder_or_tokenizer (line 376) | def test_stable_diffusion_prompt_embeds_no_text_encoder_or_tokenizer(s... method test_stable_diffusion_prompt_embeds_with_plain_negative_prompt_list (line 415) | def test_stable_diffusion_prompt_embeds_with_plain_negative_prompt_lis... method test_stable_diffusion_ddim_factor_8 (line 454) | def test_stable_diffusion_ddim_factor_8(self): method test_stable_diffusion_pndm (line 473) | def test_stable_diffusion_pndm(self): method test_stable_diffusion_no_safety_checker (line 491) | def test_stable_diffusion_no_safety_checker(self): method test_stable_diffusion_k_lms (line 512) | def test_stable_diffusion_k_lms(self): method test_stable_diffusion_k_euler_ancestral (line 531) | def test_stable_diffusion_k_euler_ancestral(self): method test_stable_diffusion_k_euler (line 550) | def test_stable_diffusion_k_euler(self): method test_stable_diffusion_vae_slicing (line 569) | def test_stable_diffusion_vae_slicing(self): method test_stable_diffusion_vae_tiling (line 592) | def test_stable_diffusion_vae_tiling(self): method test_stable_diffusion_negative_prompt (line 621) | def test_stable_diffusion_negative_prompt(self): method test_stable_diffusion_long_prompt (line 641) | def test_stable_diffusion_long_prompt(self): method test_stable_diffusion_height_width_opt (line 687) | def test_stable_diffusion_height_width_opt(self): method test_attention_slicing_forward_pass (line 711) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 714) | def test_inference_batch_single_identical(self): method test_freeu_enabled (line 719) | def test_freeu_enabled(self): method test_freeu_disabled (line 735) | def test_freeu_disabled(self): method test_fused_qkv_projections (line 760) | def test_fused_qkv_projections(self): method test_pipeline_interrupt (line 791) | def test_pipeline_interrupt(self): class StableDiffusionPipelineSlowTests (line 846) | class StableDiffusionPipelineSlowTests(unittest.TestCase): method setUp (line 847) | def setUp(self): method get_inputs (line 851) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_stable_diffusion_1_1_pndm (line 865) | def test_stable_diffusion_1_1_pndm(self): method test_stable_diffusion_v1_4_with_freeu (line 878) | def test_stable_diffusion_v1_4_with_freeu(self): method test_stable_diffusion_1_4_pndm (line 892) | def test_stable_diffusion_1_4_pndm(self): method test_stable_diffusion_ddim (line 905) | def test_stable_diffusion_ddim(self): method test_stable_diffusion_lms (line 919) | def test_stable_diffusion_lms(self): method test_stable_diffusion_dpm (line 933) | def test_stable_diffusion_dpm(self): method test_stable_diffusion_attention_slicing (line 950) | def test_stable_diffusion_attention_slicing(self): method test_stable_diffusion_vae_slicing (line 979) | def test_stable_diffusion_vae_slicing(self): method test_stable_diffusion_vae_tiling (line 1012) | def test_stable_diffusion_vae_tiling(self): method test_stable_diffusion_fp16_vs_autocast (line 1060) | def test_stable_diffusion_fp16_vs_autocast(self): method test_stable_diffusion_intermediate_state (line 1080) | def test_stable_diffusion_intermediate_state(self): method test_stable_diffusion_low_cpu_mem_usage (line 1118) | def test_stable_diffusion_low_cpu_mem_usage(self): method test_stable_diffusion_pipeline_with_sequential_cpu_offloading (line 1132) | def test_stable_diffusion_pipeline_with_sequential_cpu_offloading(self): method test_stable_diffusion_pipeline_with_model_offloading (line 1149) | def test_stable_diffusion_pipeline_with_model_offloading(self): method test_stable_diffusion_textual_inversion (line 1210) | def test_stable_diffusion_textual_inversion(self): method test_stable_diffusion_textual_inversion_with_model_cpu_offload (line 1235) | def test_stable_diffusion_textual_inversion_with_model_cpu_offload(self): method test_stable_diffusion_textual_inversion_with_sequential_cpu_offload (line 1260) | def test_stable_diffusion_textual_inversion_with_sequential_cpu_offloa... method test_stable_diffusion_compile (line 1287) | def test_stable_diffusion_compile(self): method test_stable_diffusion_lcm (line 1296) | def test_stable_diffusion_lcm(self): class StableDiffusionPipelineCkptTests (line 1322) | class StableDiffusionPipelineCkptTests(unittest.TestCase): method setUp (line 1323) | def setUp(self): method tearDown (line 1328) | def tearDown(self): method test_download_from_hub (line 1333) | def test_download_from_hub(self): method test_download_local (line 1348) | def test_download_local(self): class StableDiffusionPipelineNightlyTests (line 1365) | class StableDiffusionPipelineNightlyTests(unittest.TestCase): method setUp (line 1366) | def setUp(self): method tearDown (line 1371) | def tearDown(self): method get_inputs (line 1376) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_stable_diffusion_1_4_pndm (line 1390) | def test_stable_diffusion_1_4_pndm(self): method test_stable_diffusion_1_5_pndm (line 1404) | def test_stable_diffusion_1_5_pndm(self): method test_stable_diffusion_ddim (line 1418) | def test_stable_diffusion_ddim(self): method test_stable_diffusion_lms (line 1433) | def test_stable_diffusion_lms(self): method test_stable_diffusion_euler (line 1448) | def test_stable_diffusion_euler(self): class StableDiffusionPipelineDeviceMapTests (line 1468) | class StableDiffusionPipelineDeviceMapTests(unittest.TestCase): method tearDown (line 1469) | def tearDown(self): method get_inputs (line 1474) | def get_inputs(self, generator_device="cpu", seed=0): method get_pipeline_output_without_device_map (line 1485) | def get_pipeline_output_without_device_map(self): method test_forward_pass_balanced_device_map (line 1497) | def test_forward_pass_balanced_device_map(self): method test_components_put_in_right_devices (line 1510) | def test_components_put_in_right_devices(self): method test_max_memory (line 1517) | def test_max_memory(self): method test_reset_device_map (line 1533) | def test_reset_device_map(self): method test_reset_device_map_to (line 1545) | def test_reset_device_map_to(self): method test_reset_device_map_enable_model_cpu_offload (line 1557) | def test_reset_device_map_enable_model_cpu_offload(self): method test_reset_device_map_enable_sequential_cpu_offload (line 1569) | def test_reset_device_map_enable_sequential_cpu_offload(self): FILE: diffusers/tests/pipelines/stable_diffusion/test_stable_diffusion_img2img.py function _test_img2img_compile (line 70) | def _test_img2img_compile(in_queue, out_queue, timeout): class StableDiffusionImg2ImgPipelineFastTests (line 101) | class StableDiffusionImg2ImgPipelineFastTests( method get_dummy_components (line 116) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_tiny_autoencoder (line 166) | def get_dummy_tiny_autoencoder(self): method get_dummy_inputs (line 169) | def get_dummy_inputs(self, device, seed=0): method test_stable_diffusion_img2img_default_case (line 186) | def test_stable_diffusion_img2img_default_case(self): method test_stable_diffusion_img2img_default_case_lcm (line 202) | def test_stable_diffusion_img2img_default_case_lcm(self): method test_stable_diffusion_img2img_default_case_lcm_custom_timesteps (line 219) | def test_stable_diffusion_img2img_default_case_lcm_custom_timesteps(se... method test_stable_diffusion_img2img_negative_prompt (line 238) | def test_stable_diffusion_img2img_negative_prompt(self): method test_ip_adapter_single (line 256) | def test_ip_adapter_single(self): method test_stable_diffusion_img2img_multiple_init_images (line 262) | def test_stable_diffusion_img2img_multiple_init_images(self): method test_stable_diffusion_img2img_k_lms (line 280) | def test_stable_diffusion_img2img_k_lms(self): method test_stable_diffusion_img2img_tiny_autoencoder (line 299) | def test_stable_diffusion_img2img_tiny_autoencoder(self): method test_save_load_local (line 317) | def test_save_load_local(self): method test_dict_tuple_outputs_equivalent (line 321) | def test_dict_tuple_outputs_equivalent(self): method test_save_load_optional_components (line 325) | def test_save_load_optional_components(self): method test_attention_slicing_forward_pass (line 329) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 332) | def test_inference_batch_single_identical(self): method test_float16_inference (line 335) | def test_float16_inference(self): method test_pipeline_interrupt (line 338) | def test_pipeline_interrupt(self): class StableDiffusionImg2ImgPipelineSlowTests (line 397) | class StableDiffusionImg2ImgPipelineSlowTests(unittest.TestCase): method setUp (line 398) | def setUp(self): method tearDown (line 403) | def tearDown(self): method get_inputs (line 408) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_stable_diffusion_img2img_default (line 425) | def test_stable_diffusion_img2img_default(self): method test_stable_diffusion_img2img_k_lms (line 440) | def test_stable_diffusion_img2img_k_lms(self): method test_stable_diffusion_img2img_ddim (line 456) | def test_stable_diffusion_img2img_ddim(self): method test_stable_diffusion_img2img_intermediate_state (line 472) | def test_stable_diffusion_img2img_intermediate_state(self): method test_stable_diffusion_pipeline_with_sequential_cpu_offloading (line 508) | def test_stable_diffusion_pipeline_with_sequential_cpu_offloading(self): method test_stable_diffusion_pipeline_with_model_offloading (line 527) | def test_stable_diffusion_pipeline_with_model_offloading(self): method test_img2img_2nd_order (line 568) | def test_img2img_2nd_order(self): method test_stable_diffusion_img2img_pipeline_multiple_of_8 (line 595) | def test_stable_diffusion_img2img_pipeline_multiple_of_8(self): method test_img2img_safety_checker_works (line 632) | def test_img2img_safety_checker_works(self): method test_img2img_compile (line 648) | def test_img2img_compile(self): class StableDiffusionImg2ImgPipelineNightlyTests (line 660) | class StableDiffusionImg2ImgPipelineNightlyTests(unittest.TestCase): method setUp (line 661) | def setUp(self): method tearDown (line 666) | def tearDown(self): method get_inputs (line 671) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_img2img_pndm (line 688) | def test_img2img_pndm(self): method test_img2img_ddim (line 703) | def test_img2img_ddim(self): method test_img2img_lms (line 719) | def test_img2img_lms(self): method test_img2img_dpm (line 735) | def test_img2img_dpm(self): FILE: diffusers/tests/pipelines/stable_diffusion/test_stable_diffusion_inpaint.py function _test_inpaint_compile (line 70) | def _test_inpaint_compile(in_queue, out_queue, timeout): class StableDiffusionInpaintPipelineFastTests (line 103) | class StableDiffusionInpaintPipelineFastTests( method get_dummy_components (line 118) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_inputs (line 168) | def get_dummy_inputs(self, device, seed=0, img_res=64, output_pil=True): method test_stable_diffusion_inpaint (line 204) | def test_stable_diffusion_inpaint(self): method test_stable_diffusion_inpaint_lcm (line 220) | def test_stable_diffusion_inpaint_lcm(self): method test_stable_diffusion_inpaint_lcm_custom_timesteps (line 237) | def test_stable_diffusion_inpaint_lcm_custom_timesteps(self): method test_stable_diffusion_inpaint_image_tensor (line 256) | def test_stable_diffusion_inpaint_image_tensor(self): method test_inference_batch_single_identical (line 276) | def test_inference_batch_single_identical(self): method test_stable_diffusion_inpaint_strength_zero_test (line 279) | def test_stable_diffusion_inpaint_strength_zero_test(self): method test_stable_diffusion_inpaint_mask_latents (line 293) | def test_stable_diffusion_inpaint_mask_latents(self): method test_pipeline_interrupt (line 330) | def test_pipeline_interrupt(self): method test_ip_adapter_single (line 388) | def test_ip_adapter_single(self, from_simple=False, expected_pipe_slic... class StableDiffusionSimpleInpaintPipelineFastTests (line 398) | class StableDiffusionSimpleInpaintPipelineFastTests(StableDiffusionInpai... method get_dummy_components (line 405) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_inputs_2images (line 455) | def get_dummy_inputs_2images(self, device, seed=0, img_res=64): method test_ip_adapter_single (line 484) | def test_ip_adapter_single(self): method test_stable_diffusion_inpaint (line 490) | def test_stable_diffusion_inpaint(self): method test_stable_diffusion_inpaint_lcm (line 506) | def test_stable_diffusion_inpaint_lcm(self): method test_stable_diffusion_inpaint_lcm_custom_timesteps (line 523) | def test_stable_diffusion_inpaint_lcm_custom_timesteps(self): method test_stable_diffusion_inpaint_2_images (line 542) | def test_stable_diffusion_inpaint_2_images(self): method test_stable_diffusion_inpaint_euler (line 573) | def test_stable_diffusion_inpaint_euler(self): class StableDiffusionInpaintPipelineSlowTests (line 599) | class StableDiffusionInpaintPipelineSlowTests(unittest.TestCase): method setUp (line 600) | def setUp(self): method tearDown (line 603) | def tearDown(self): method get_inputs (line 608) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_stable_diffusion_inpaint_ddim (line 629) | def test_stable_diffusion_inpaint_ddim(self): method test_stable_diffusion_inpaint_fp16 (line 646) | def test_stable_diffusion_inpaint_fp16(self): method test_stable_diffusion_inpaint_pndm (line 663) | def test_stable_diffusion_inpaint_pndm(self): method test_stable_diffusion_inpaint_k_lms (line 681) | def test_stable_diffusion_inpaint_k_lms(self): method test_stable_diffusion_inpaint_with_sequential_cpu_offloading (line 699) | def test_stable_diffusion_inpaint_with_sequential_cpu_offloading(self): method test_inpaint_compile (line 720) | def test_inpaint_compile(self): method test_stable_diffusion_inpaint_pil_input_resolution_test (line 729) | def test_stable_diffusion_inpaint_pil_input_resolution_test(self): method test_stable_diffusion_inpaint_strength_test (line 748) | def test_stable_diffusion_inpaint_strength_test(self): method test_stable_diffusion_simple_inpaint_ddim (line 769) | def test_stable_diffusion_simple_inpaint_ddim(self): class StableDiffusionInpaintPipelineAsymmetricAutoencoderKLSlowTests (line 788) | class StableDiffusionInpaintPipelineAsymmetricAutoencoderKLSlowTests(uni... method setUp (line 789) | def setUp(self): method tearDown (line 792) | def tearDown(self): method get_inputs (line 797) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_stable_diffusion_inpaint_ddim (line 818) | def test_stable_diffusion_inpaint_ddim(self): method test_stable_diffusion_inpaint_fp16 (line 838) | def test_stable_diffusion_inpaint_fp16(self): method test_stable_diffusion_inpaint_pndm (line 860) | def test_stable_diffusion_inpaint_pndm(self): method test_stable_diffusion_inpaint_k_lms (line 881) | def test_stable_diffusion_inpaint_k_lms(self): method test_stable_diffusion_inpaint_with_sequential_cpu_offloading (line 900) | def test_stable_diffusion_inpaint_with_sequential_cpu_offloading(self): method test_inpaint_compile (line 925) | def test_inpaint_compile(self): method test_stable_diffusion_inpaint_pil_input_resolution_test (line 928) | def test_stable_diffusion_inpaint_pil_input_resolution_test(self): method test_stable_diffusion_inpaint_strength_test (line 951) | def test_stable_diffusion_inpaint_strength_test(self): method test_stable_diffusion_simple_inpaint_ddim (line 974) | def test_stable_diffusion_simple_inpaint_ddim(self): method test_download_local (line 992) | def test_download_local(self): class StableDiffusionInpaintPipelineNightlyTests (line 1012) | class StableDiffusionInpaintPipelineNightlyTests(unittest.TestCase): method setUp (line 1013) | def setUp(self): method tearDown (line 1018) | def tearDown(self): method get_inputs (line 1023) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_inpaint_ddim (line 1044) | def test_inpaint_ddim(self): method test_inpaint_pndm (line 1059) | def test_inpaint_pndm(self): method test_inpaint_lms (line 1075) | def test_inpaint_lms(self): method test_inpaint_dpm (line 1091) | def test_inpaint_dpm(self): FILE: diffusers/tests/pipelines/stable_diffusion/test_stable_diffusion_instruction_pix2pix.py class StableDiffusionInstructPix2PixPipelineFastTests (line 60) | class StableDiffusionInstructPix2PixPipelineFastTests( method get_dummy_components (line 70) | def get_dummy_components(self): method get_dummy_inputs (line 119) | def get_dummy_inputs(self, device, seed=0): method test_stable_diffusion_pix2pix_default_case (line 138) | def test_stable_diffusion_pix2pix_default_case(self): method test_stable_diffusion_pix2pix_negative_prompt (line 153) | def test_stable_diffusion_pix2pix_negative_prompt(self): method test_stable_diffusion_pix2pix_multiple_init_images (line 171) | def test_stable_diffusion_pix2pix_multiple_init_images(self): method test_stable_diffusion_pix2pix_euler (line 195) | def test_stable_diffusion_pix2pix_euler(self): method test_inference_batch_single_identical (line 217) | def test_inference_batch_single_identical(self): method test_latents_input (line 221) | def test_latents_input(self): method test_callback_cfg (line 243) | def test_callback_cfg(self): class StableDiffusionInstructPix2PixPipelineSlowTests (line 273) | class StableDiffusionInstructPix2PixPipelineSlowTests(unittest.TestCase): method setUp (line 274) | def setUp(self): method tearDown (line 279) | def tearDown(self): method get_inputs (line 284) | def get_inputs(self, seed=0): method test_stable_diffusion_pix2pix_default (line 300) | def test_stable_diffusion_pix2pix_default(self): method test_stable_diffusion_pix2pix_k_lms (line 317) | def test_stable_diffusion_pix2pix_k_lms(self): method test_stable_diffusion_pix2pix_ddim (line 335) | def test_stable_diffusion_pix2pix_ddim(self): method test_stable_diffusion_pix2pix_intermediate_state (line 353) | def test_stable_diffusion_pix2pix_intermediate_state(self): method test_stable_diffusion_pipeline_with_sequential_cpu_offloading (line 389) | def test_stable_diffusion_pipeline_with_sequential_cpu_offloading(self): method test_stable_diffusion_pix2pix_pipeline_multiple_of_8 (line 408) | def test_stable_diffusion_pix2pix_pipeline_multiple_of_8(self): FILE: diffusers/tests/pipelines/stable_diffusion_2/test_stable_diffusion.py class StableDiffusion2PipelineFastTests (line 66) | class StableDiffusion2PipelineFastTests( method get_dummy_components (line 80) | def get_dummy_components(self): method get_dummy_inputs (line 142) | def get_dummy_inputs(self, device, seed=0): method test_stable_diffusion_ddim (line 158) | def test_stable_diffusion_ddim(self): method test_stable_diffusion_pndm (line 174) | def test_stable_diffusion_pndm(self): method test_stable_diffusion_k_lms (line 191) | def test_stable_diffusion_k_lms(self): method test_stable_diffusion_k_euler_ancestral (line 208) | def test_stable_diffusion_k_euler_ancestral(self): method test_stable_diffusion_k_euler (line 225) | def test_stable_diffusion_k_euler(self): method test_stable_diffusion_unflawed (line 242) | def test_stable_diffusion_unflawed(self): method test_stable_diffusion_long_prompt (line 263) | def test_stable_diffusion_long_prompt(self): method test_attention_slicing_forward_pass (line 308) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 311) | def test_inference_batch_single_identical(self): class StableDiffusion2PipelineSlowTests (line 318) | class StableDiffusion2PipelineSlowTests(unittest.TestCase): method tearDown (line 319) | def tearDown(self): method get_inputs (line 324) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_stable_diffusion_default_ddim (line 343) | def test_stable_diffusion_default_ddim(self): method test_stable_diffusion_pndm (line 356) | def test_stable_diffusion_pndm(self): method test_stable_diffusion_k_lms (line 370) | def test_stable_diffusion_k_lms(self): method test_stable_diffusion_attention_slicing (line 385) | def test_stable_diffusion_attention_slicing(self): method test_stable_diffusion_text2img_intermediate_state (line 416) | def test_stable_diffusion_text2img_intermediate_state(self): method test_stable_diffusion_pipeline_with_sequential_cpu_offloading (line 457) | def test_stable_diffusion_pipeline_with_sequential_cpu_offloading(self): method test_stable_diffusion_pipeline_with_model_offloading (line 477) | def test_stable_diffusion_pipeline_with_model_offloading(self): class StableDiffusion2PipelineNightlyTests (line 538) | class StableDiffusion2PipelineNightlyTests(unittest.TestCase): method tearDown (line 539) | def tearDown(self): method get_inputs (line 544) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_stable_diffusion_2_0_default_ddim (line 563) | def test_stable_diffusion_2_0_default_ddim(self): method test_stable_diffusion_2_1_default_pndm (line 577) | def test_stable_diffusion_2_1_default_pndm(self): method test_stable_diffusion_ddim (line 591) | def test_stable_diffusion_ddim(self): method test_stable_diffusion_lms (line 606) | def test_stable_diffusion_lms(self): method test_stable_diffusion_euler (line 621) | def test_stable_diffusion_euler(self): method test_stable_diffusion_dpm (line 636) | def test_stable_diffusion_dpm(self): FILE: diffusers/tests/pipelines/stable_diffusion_2/test_stable_diffusion_attend_and_excite.py class StableDiffusionAttendAndExcitePipelineFastTests (line 51) | class StableDiffusionAttendAndExcitePipelineFastTests( method setUpClass (line 69) | def setUpClass(cls): method tearDownClass (line 74) | def tearDownClass(cls): method get_dummy_components (line 78) | def get_dummy_components(self): method get_dummy_inputs (line 140) | def get_dummy_inputs(self, device, seed=0): method test_dict_tuple_outputs_equivalent (line 157) | def test_dict_tuple_outputs_equivalent(self): method test_inference (line 163) | def test_inference(self): method test_sequential_cpu_offload_forward_pass (line 182) | def test_sequential_cpu_offload_forward_pass(self): method test_inference_batch_consistent (line 185) | def test_inference_batch_consistent(self): method test_inference_batch_single_identical (line 189) | def test_inference_batch_single_identical(self): method test_pt_np_pil_outputs_equivalent (line 192) | def test_pt_np_pil_outputs_equivalent(self): method test_save_load_local (line 195) | def test_save_load_local(self): method test_save_load_optional_components (line 198) | def test_save_load_optional_components(self): method test_karras_schedulers_shape (line 201) | def test_karras_schedulers_shape(self): class StableDiffusionAttendAndExcitePipelineIntegrationTests (line 207) | class StableDiffusionAttendAndExcitePipelineIntegrationTests(unittest.Te... method setUpClass (line 212) | def setUpClass(cls): method tearDownClass (line 217) | def tearDownClass(cls): method setUp (line 221) | def setUp(self): method tearDown (line 226) | def tearDown(self): method test_attend_and_excite_fp16 (line 231) | def test_attend_and_excite_fp16(self): FILE: diffusers/tests/pipelines/stable_diffusion_2/test_stable_diffusion_depth.py class StableDiffusionDepth2ImgPipelineFastTests (line 69) | class StableDiffusionDepth2ImgPipelineFastTests( method get_dummy_components (line 81) | def get_dummy_components(self): method get_dummy_inputs (line 163) | def get_dummy_inputs(self, device, seed=0): method test_save_load_local (line 181) | def test_save_load_local(self): method test_save_load_float16 (line 203) | def test_save_load_float16(self): method test_float16_inference (line 235) | def test_float16_inference(self): method test_cpu_offload_forward_pass (line 258) | def test_cpu_offload_forward_pass(self): method test_dict_tuple_outputs_equivalent (line 274) | def test_dict_tuple_outputs_equivalent(self): method test_progress_bar (line 286) | def test_progress_bar(self): method test_stable_diffusion_depth2img_default_case (line 289) | def test_stable_diffusion_depth2img_default_case(self): method test_stable_diffusion_depth2img_negative_prompt (line 308) | def test_stable_diffusion_depth2img_negative_prompt(self): method test_stable_diffusion_depth2img_multiple_init_images (line 329) | def test_stable_diffusion_depth2img_multiple_init_images(self): method test_stable_diffusion_depth2img_pil (line 351) | def test_stable_diffusion_depth2img_pil(self): method test_attention_slicing_forward_pass (line 371) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 374) | def test_inference_batch_single_identical(self): class StableDiffusionDepth2ImgPipelineSlowTests (line 380) | class StableDiffusionDepth2ImgPipelineSlowTests(unittest.TestCase): method setUp (line 381) | def setUp(self): method tearDown (line 386) | def tearDown(self): method get_inputs (line 391) | def get_inputs(self, device="cpu", dtype=torch.float32, seed=0): method test_stable_diffusion_depth2img_pipeline_default (line 407) | def test_stable_diffusion_depth2img_pipeline_default(self): method test_stable_diffusion_depth2img_pipeline_k_lms (line 424) | def test_stable_diffusion_depth2img_pipeline_k_lms(self): method test_stable_diffusion_depth2img_pipeline_ddim (line 443) | def test_stable_diffusion_depth2img_pipeline_ddim(self): method test_stable_diffusion_depth2img_intermediate_state (line 461) | def test_stable_diffusion_depth2img_intermediate_state(self): method test_stable_diffusion_pipeline_with_sequential_cpu_offloading (line 501) | def test_stable_diffusion_pipeline_with_sequential_cpu_offloading(self): class StableDiffusionImg2ImgPipelineNightlyTests (line 523) | class StableDiffusionImg2ImgPipelineNightlyTests(unittest.TestCase): method setUp (line 524) | def setUp(self): method tearDown (line 529) | def tearDown(self): method get_inputs (line 534) | def get_inputs(self, device="cpu", dtype=torch.float32, seed=0): method test_depth2img_pndm (line 550) | def test_depth2img_pndm(self): method test_depth2img_ddim (line 565) | def test_depth2img_ddim(self): method test_img2img_lms (line 581) | def test_img2img_lms(self): method test_img2img_dpm (line 597) | def test_img2img_dpm(self): FILE: diffusers/tests/pipelines/stable_diffusion_2/test_stable_diffusion_diffedit.py class StableDiffusionDiffEditPipelineFastTests (line 52) | class StableDiffusionDiffEditPipelineFastTests( method get_dummy_components (line 63) | def get_dummy_components(self): method get_dummy_inputs (line 133) | def get_dummy_inputs(self, device, seed=0): method get_dummy_mask_inputs (line 153) | def get_dummy_mask_inputs(self, device, seed=0): method get_dummy_inversion_inputs (line 175) | def get_dummy_inversion_inputs(self, device, seed=0): method test_save_load_optional_components (line 195) | def test_save_load_optional_components(self): method test_mask (line 230) | def test_mask(self): method test_inversion (line 248) | def test_inversion(self): method test_inference_batch_single_identical (line 267) | def test_inference_batch_single_identical(self): method test_inversion_dpm (line 270) | def test_inversion_dpm(self): class StableDiffusionDiffEditPipelineIntegrationTests (line 297) | class StableDiffusionDiffEditPipelineIntegrationTests(unittest.TestCase): method setUp (line 298) | def setUp(self): method tearDown (line 303) | def tearDown(self): method setUpClass (line 309) | def setUpClass(cls): method test_stable_diffusion_diffedit_full (line 317) | def test_stable_diffusion_diffedit_full(self): class StableDiffusionDiffEditPipelineNightlyTests (line 374) | class StableDiffusionDiffEditPipelineNightlyTests(unittest.TestCase): method setUp (line 375) | def setUp(self): method tearDown (line 380) | def tearDown(self): method setUpClass (line 386) | def setUpClass(cls): method test_stable_diffusion_diffedit_dpm (line 395) | def test_stable_diffusion_diffedit_dpm(self): FILE: diffusers/tests/pipelines/stable_diffusion_2/test_stable_diffusion_flax.py class FlaxStableDiffusion2PipelineIntegrationTests (line 33) | class FlaxStableDiffusion2PipelineIntegrationTests(unittest.TestCase): method tearDown (line 34) | def tearDown(self): method test_stable_diffusion_flax (line 39) | def test_stable_diffusion_flax(self): class FlaxStableDiffusion2PipelineNightlyTests (line 71) | class FlaxStableDiffusion2PipelineNightlyTests(unittest.TestCase): method tearDown (line 72) | def tearDown(self): method test_stable_diffusion_dpm_flax (line 77) | def test_stable_diffusion_dpm_flax(self): FILE: diffusers/tests/pipelines/stable_diffusion_2/test_stable_diffusion_flax_inpaint.py class FlaxStableDiffusionInpaintPipelineIntegrationTests (line 33) | class FlaxStableDiffusionInpaintPipelineIntegrationTests(unittest.TestCa... method tearDown (line 34) | def tearDown(self): method test_stable_diffusion_inpaint_pipeline (line 39) | def test_stable_diffusion_inpaint_pipeline(self): FILE: diffusers/tests/pipelines/stable_diffusion_2/test_stable_diffusion_inpaint.py class StableDiffusion2InpaintPipelineFastTests (line 47) | class StableDiffusion2InpaintPipelineFastTests( method get_dummy_components (line 59) | def get_dummy_components(self): method get_dummy_inputs (line 115) | def get_dummy_inputs(self, device, seed=0): method test_stable_diffusion_inpaint (line 136) | def test_stable_diffusion_inpaint(self): method test_inference_batch_single_identical (line 152) | def test_inference_batch_single_identical(self): class StableDiffusionInpaintPipelineIntegrationTests (line 158) | class StableDiffusionInpaintPipelineIntegrationTests(unittest.TestCase): method setUp (line 159) | def setUp(self): method tearDown (line 165) | def tearDown(self): method test_stable_diffusion_inpaint_pipeline (line 171) | def test_stable_diffusion_inpaint_pipeline(self): method test_stable_diffusion_inpaint_pipeline_fp16 (line 205) | def test_stable_diffusion_inpaint_pipeline_fp16(self): method test_stable_diffusion_pipeline_with_sequential_cpu_offloading (line 243) | def test_stable_diffusion_pipeline_with_sequential_cpu_offloading(self): FILE: diffusers/tests/pipelines/stable_diffusion_2/test_stable_diffusion_latent_upscale.py function check_same_shape (line 50) | def check_same_shape(tensor_list): class StableDiffusionLatentUpscalePipelineFastTests (line 55) | class StableDiffusionLatentUpscalePipelineFastTests( method dummy_image (line 74) | def dummy_image(self): method get_dummy_components (line 82) | def get_dummy_components(self): method get_dummy_inputs (line 148) | def get_dummy_inputs(self, device, seed=0): method test_inference (line 162) | def test_inference(self): method test_attention_slicing_forward_pass (line 181) | def test_attention_slicing_forward_pass(self): method test_sequential_cpu_offload_forward_pass (line 184) | def test_sequential_cpu_offload_forward_pass(self): method test_dict_tuple_outputs_equivalent (line 187) | def test_dict_tuple_outputs_equivalent(self): method test_inference_batch_single_identical (line 190) | def test_inference_batch_single_identical(self): method test_pt_np_pil_outputs_equivalent (line 193) | def test_pt_np_pil_outputs_equivalent(self): method test_save_load_local (line 196) | def test_save_load_local(self): method test_save_load_optional_components (line 199) | def test_save_load_optional_components(self): method test_karras_schedulers_shape (line 202) | def test_karras_schedulers_shape(self): method test_float16_inference (line 239) | def test_float16_inference(self): class StableDiffusionLatentUpscalePipelineIntegrationTests (line 245) | class StableDiffusionLatentUpscalePipelineIntegrationTests(unittest.Test... method setUp (line 246) | def setUp(self): method tearDown (line 251) | def tearDown(self): method test_latent_upscaler_fp16 (line 256) | def test_latent_upscaler_fp16(self): method test_latent_upscaler_fp16_image (line 285) | def test_latent_upscaler_fp16_image(self): FILE: diffusers/tests/pipelines/stable_diffusion_2/test_stable_diffusion_upscale.py class StableDiffusionUpscalePipelineFastTests (line 41) | class StableDiffusionUpscalePipelineFastTests(unittest.TestCase): method setUp (line 42) | def setUp(self): method tearDown (line 48) | def tearDown(self): method dummy_image (line 55) | def dummy_image(self): method dummy_cond_unet_upscale (line 64) | def dummy_cond_unet_upscale(self): method dummy_vae (line 84) | def dummy_vae(self): method dummy_text_encoder (line 97) | def dummy_text_encoder(self): method test_stable_diffusion_upscale (line 115) | def test_stable_diffusion_upscale(self): method test_stable_diffusion_upscale_batch (line 176) | def test_stable_diffusion_upscale_batch(self): method test_stable_diffusion_upscale_prompt_embeds (line 227) | def test_stable_diffusion_upscale_prompt_embeds(self): method test_stable_diffusion_upscale_fp16 (line 293) | def test_stable_diffusion_upscale_fp16(self): method test_stable_diffusion_upscale_from_save_pretrained (line 335) | def test_stable_diffusion_upscale_from_save_pretrained(self): class StableDiffusionUpscalePipelineIntegrationTests (line 384) | class StableDiffusionUpscalePipelineIntegrationTests(unittest.TestCase): method setUp (line 385) | def setUp(self): method tearDown (line 391) | def tearDown(self): method test_stable_diffusion_upscale_pipeline (line 397) | def test_stable_diffusion_upscale_pipeline(self): method test_stable_diffusion_upscale_pipeline_fp16 (line 427) | def test_stable_diffusion_upscale_pipeline_fp16(self): method test_stable_diffusion_pipeline_with_sequential_cpu_offloading (line 460) | def test_stable_diffusion_pipeline_with_sequential_cpu_offloading(self): FILE: diffusers/tests/pipelines/stable_diffusion_2/test_stable_diffusion_v_pred.py class StableDiffusion2VPredictionPipelineFastTests (line 46) | class StableDiffusion2VPredictionPipelineFastTests(unittest.TestCase): method setUp (line 47) | def setUp(self): method tearDown (line 53) | def tearDown(self): method dummy_cond_unet (line 60) | def dummy_cond_unet(self): method dummy_vae (line 78) | def dummy_vae(self): method dummy_text_encoder (line 92) | def dummy_text_encoder(self): method test_stable_diffusion_v_pred_ddim (line 110) | def test_stable_diffusion_v_pred_ddim(self): method test_stable_diffusion_v_pred_k_euler (line 166) | def test_stable_diffusion_v_pred_k_euler(self): method test_stable_diffusion_v_pred_fp16 (line 217) | def test_stable_diffusion_v_pred_fp16(self): class StableDiffusion2VPredictionPipelineIntegrationTests (line 261) | class StableDiffusion2VPredictionPipelineIntegrationTests(unittest.TestC... method setUp (line 262) | def setUp(self): method tearDown (line 268) | def tearDown(self): method test_stable_diffusion_v_pred_default (line 274) | def test_stable_diffusion_v_pred_default(self): method test_stable_diffusion_v_pred_upcast_attention (line 292) | def test_stable_diffusion_v_pred_upcast_attention(self): method test_stable_diffusion_v_pred_euler (line 312) | def test_stable_diffusion_v_pred_euler(self): method test_stable_diffusion_v_pred_dpm (line 332) | def test_stable_diffusion_v_pred_dpm(self): method test_stable_diffusion_attention_slicing_v_pred (line 358) | def test_stable_diffusion_attention_slicing_v_pred(self): method test_stable_diffusion_text2img_pipeline_v_pred_default (line 392) | def test_stable_diffusion_text2img_pipeline_v_pred_default(self): method test_stable_diffusion_text2img_pipeline_unflawed (line 413) | def test_stable_diffusion_text2img_pipeline_unflawed(self): method test_stable_diffusion_text2img_pipeline_v_pred_fp16 (line 443) | def test_stable_diffusion_text2img_pipeline_v_pred_fp16(self): method test_download_local (line 463) | def test_download_local(self): method test_stable_diffusion_text2img_intermediate_state_v_pred (line 474) | def test_stable_diffusion_text2img_intermediate_state_v_pred(self): method test_stable_diffusion_low_cpu_mem_usage_v_pred (line 517) | def test_stable_diffusion_low_cpu_mem_usage_v_pred(self): method test_stable_diffusion_pipeline_with_sequential_cpu_offloading_v_pred (line 531) | def test_stable_diffusion_pipeline_with_sequential_cpu_offloading_v_pr... FILE: diffusers/tests/pipelines/stable_diffusion_3/test_pipeline_stable_diffusion_3.py class StableDiffusion3PipelineFastTests (line 19) | class StableDiffusion3PipelineFastTests(unittest.TestCase, PipelineTeste... method get_dummy_components (line 34) | def get_dummy_components(self): method get_dummy_inputs (line 103) | def get_dummy_inputs(self, device, seed=0): method test_stable_diffusion_3_different_prompts (line 118) | def test_stable_diffusion_3_different_prompts(self): method test_stable_diffusion_3_different_negative_prompts (line 134) | def test_stable_diffusion_3_different_negative_prompts(self): method test_stable_diffusion_3_prompt_embeds (line 150) | def test_stable_diffusion_3_prompt_embeds(self): method test_fused_qkv_projections (line 183) | def test_fused_qkv_projections(self): class StableDiffusion3PipelineSlowTests (line 217) | class StableDiffusion3PipelineSlowTests(unittest.TestCase): method setUp (line 221) | def setUp(self): method tearDown (line 226) | def tearDown(self): method get_inputs (line 231) | def get_inputs(self, device, seed=0): method test_sd3_inference (line 245) | def test_sd3_inference(self): FILE: diffusers/tests/pipelines/stable_diffusion_3/test_pipeline_stable_diffusion_3_img2img.py class StableDiffusion3Img2ImgPipelineFastTests (line 32) | class StableDiffusion3Img2ImgPipelineFastTests(PipelineLatentTesterMixin... method get_dummy_components (line 40) | def get_dummy_components(self): method get_dummy_inputs (line 109) | def get_dummy_inputs(self, device, seed=0): method test_stable_diffusion_3_img2img_different_prompts (line 128) | def test_stable_diffusion_3_img2img_different_prompts(self): method test_stable_diffusion_3_img2img_different_negative_prompts (line 144) | def test_stable_diffusion_3_img2img_different_negative_prompts(self): method test_stable_diffusion_3_img2img_prompt_embeds (line 160) | def test_stable_diffusion_3_img2img_prompt_embeds(self): method test_multi_vae (line 193) | def test_multi_vae(self): class StableDiffusion3Img2ImgPipelineSlowTests (line 199) | class StableDiffusion3Img2ImgPipelineSlowTests(unittest.TestCase): method setUp (line 203) | def setUp(self): method tearDown (line 208) | def tearDown(self): method get_inputs (line 213) | def get_inputs(self, device, seed=0): method test_sd3_img2img_inference (line 232) | def test_sd3_img2img_inference(self): FILE: diffusers/tests/pipelines/stable_diffusion_adapter/test_stable_diffusion_adapter.py class AdapterTests (line 55) | class AdapterTests: method get_dummy_components (line 60) | def get_dummy_components(self, adapter_type, time_cond_proj_dim=None): method get_dummy_components_with_full_downscaling (line 144) | def get_dummy_components_with_full_downscaling(self, adapter_type): method get_dummy_inputs (line 231) | def get_dummy_inputs(self, device, seed=0, height=64, width=64, num_im... method test_attention_slicing_forward_pass (line 253) | def test_attention_slicing_forward_pass(self): method test_xformers_attention_forwardGenerator_pass (line 260) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_single_identical (line 263) | def test_inference_batch_single_identical(self): method test_multiple_image_dimensions (line 278) | def test_multiple_image_dimensions(self, dim): method test_adapter_lcm (line 297) | def test_adapter_lcm(self): method test_adapter_lcm_custom_timesteps (line 317) | def test_adapter_lcm_custom_timesteps(self): class StableDiffusionFullAdapterPipelineFastTests (line 340) | class StableDiffusionFullAdapterPipelineFastTests( method get_dummy_components (line 343) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_components_with_full_downscaling (line 346) | def get_dummy_components_with_full_downscaling(self): method test_stable_diffusion_adapter_default_case (line 349) | def test_stable_diffusion_adapter_default_case(self): class StableDiffusionLightAdapterPipelineFastTests (line 365) | class StableDiffusionLightAdapterPipelineFastTests(AdapterTests, Pipelin... method get_dummy_components (line 366) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_components_with_full_downscaling (line 369) | def get_dummy_components_with_full_downscaling(self): method test_stable_diffusion_adapter_default_case (line 372) | def test_stable_diffusion_adapter_default_case(self): class StableDiffusionMultiAdapterPipelineFastTests (line 388) | class StableDiffusionMultiAdapterPipelineFastTests(AdapterTests, Pipelin... method get_dummy_components (line 389) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_components_with_full_downscaling (line 392) | def get_dummy_components_with_full_downscaling(self): method get_dummy_inputs (line 395) | def get_dummy_inputs(self, device, height=64, width=64, seed=0): method test_stable_diffusion_adapter_default_case (line 400) | def test_stable_diffusion_adapter_default_case(self): method test_inference_batch_consistent (line 415) | def test_inference_batch_consistent( method test_num_images_per_prompt (line 479) | def test_num_images_per_prompt(self): method test_inference_batch_single_identical (line 508) | def test_inference_batch_single_identical( class StableDiffusionAdapterPipelineSlowTests (line 597) | class StableDiffusionAdapterPipelineSlowTests(unittest.TestCase): method setUp (line 598) | def setUp(self): method tearDown (line 603) | def tearDown(self): method test_stable_diffusion_adapter_color (line 608) | def test_stable_diffusion_adapter_color(self): method test_stable_diffusion_adapter_depth (line 636) | def test_stable_diffusion_adapter_depth(self): method test_stable_diffusion_adapter_depth_sd_v14 (line 664) | def test_stable_diffusion_adapter_depth_sd_v14(self): method test_stable_diffusion_adapter_depth_sd_v15 (line 690) | def test_stable_diffusion_adapter_depth_sd_v15(self): method test_stable_diffusion_adapter_keypose_sd_v14 (line 716) | def test_stable_diffusion_adapter_keypose_sd_v14(self): method test_stable_diffusion_adapter_openpose_sd_v14 (line 742) | def test_stable_diffusion_adapter_openpose_sd_v14(self): method test_stable_diffusion_adapter_seg_sd_v14 (line 768) | def test_stable_diffusion_adapter_seg_sd_v14(self): method test_stable_diffusion_adapter_zoedepth_sd_v15 (line 796) | def test_stable_diffusion_adapter_zoedepth_sd_v15(self): method test_stable_diffusion_adapter_canny_sd_v14 (line 820) | def test_stable_diffusion_adapter_canny_sd_v14(self): method test_stable_diffusion_adapter_canny_sd_v15 (line 847) | def test_stable_diffusion_adapter_canny_sd_v15(self): method test_stable_diffusion_adapter_sketch_sd14 (line 875) | def test_stable_diffusion_adapter_sketch_sd14(self): method test_stable_diffusion_adapter_sketch_sd15 (line 905) | def test_stable_diffusion_adapter_sketch_sd15(self): method test_stable_diffusion_adapter_pipeline_with_sequential_cpu_offloading (line 935) | def test_stable_diffusion_adapter_pipeline_with_sequential_cpu_offload... FILE: diffusers/tests/pipelines/stable_diffusion_gligen/test_stable_diffusion_gligen.py class GligenPipelineFastTests (line 47) | class GligenPipelineFastTests( method get_dummy_components (line 60) | def get_dummy_components(self): method get_dummy_inputs (line 117) | def get_dummy_inputs(self, device, seed=0): method test_stable_diffusion_gligen_default_case (line 133) | def test_stable_diffusion_gligen_default_case(self): method test_stable_diffusion_gligen_k_euler_ancestral (line 149) | def test_stable_diffusion_gligen_k_euler_ancestral(self): method test_attention_slicing_forward_pass (line 167) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 170) | def test_inference_batch_single_identical(self): FILE: diffusers/tests/pipelines/stable_diffusion_gligen_text_image/test_stable_diffusion_gligen_text_image.py class GligenTextImagePipelineFastTests (line 56) | class GligenTextImagePipelineFastTests( method get_dummy_components (line 69) | def get_dummy_components(self): method get_dummy_inputs (line 142) | def get_dummy_inputs(self, device, seed=0): method test_dict_tuple_outputs_equivalent (line 163) | def test_dict_tuple_outputs_equivalent(self): method test_stable_diffusion_gligen_text_image_default_case (line 169) | def test_stable_diffusion_gligen_text_image_default_case(self): method test_stable_diffusion_gligen_k_euler_ancestral (line 185) | def test_stable_diffusion_gligen_k_euler_ancestral(self): method test_attention_slicing_forward_pass (line 203) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 206) | def test_inference_batch_single_identical(self): FILE: diffusers/tests/pipelines/stable_diffusion_image_variation/test_stable_diffusion_image_variation.py class StableDiffusionImageVariationPipelineFastTests (line 51) | class StableDiffusionImageVariationPipelineFastTests( method get_dummy_components (line 61) | def get_dummy_components(self): method get_dummy_inputs (line 107) | def get_dummy_inputs(self, device, seed=0): method test_stable_diffusion_img_variation_default_case (line 124) | def test_stable_diffusion_img_variation_default_case(self): method test_stable_diffusion_img_variation_multiple_images (line 140) | def test_stable_diffusion_img_variation_multiple_images(self): method test_inference_batch_single_identical (line 160) | def test_inference_batch_single_identical(self): class StableDiffusionImageVariationPipelineSlowTests (line 166) | class StableDiffusionImageVariationPipelineSlowTests(unittest.TestCase): method setUp (line 167) | def setUp(self): method tearDown (line 172) | def tearDown(self): method get_inputs (line 177) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_stable_diffusion_img_variation_pipeline_default (line 195) | def test_stable_diffusion_img_variation_pipeline_default(self): method test_stable_diffusion_img_variation_intermediate_state (line 213) | def test_stable_diffusion_img_variation_intermediate_state(self): method test_stable_diffusion_pipeline_with_sequential_cpu_offloading (line 258) | def test_stable_diffusion_pipeline_with_sequential_cpu_offloading(self): class StableDiffusionImageVariationPipelineNightlyTests (line 280) | class StableDiffusionImageVariationPipelineNightlyTests(unittest.TestCase): method setUp (line 281) | def setUp(self): method tearDown (line 286) | def tearDown(self): method get_inputs (line 291) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_img_variation_pndm (line 309) | def test_img_variation_pndm(self): method test_img_variation_dpm (line 324) | def test_img_variation_dpm(self): FILE: diffusers/tests/pipelines/stable_diffusion_k_diffusion/test_stable_diffusion_k_diffusion.py class StableDiffusionPipelineIntegrationTests (line 31) | class StableDiffusionPipelineIntegrationTests(unittest.TestCase): method setUp (line 32) | def setUp(self): method tearDown (line 38) | def tearDown(self): method test_stable_diffusion_1 (line 44) | def test_stable_diffusion_1(self): method test_stable_diffusion_2 (line 64) | def test_stable_diffusion_2(self): method test_stable_diffusion_karras_sigmas (line 84) | def test_stable_diffusion_karras_sigmas(self): method test_stable_diffusion_noise_sampler_seed (line 113) | def test_stable_diffusion_noise_sampler_seed(self): FILE: diffusers/tests/pipelines/stable_diffusion_ldm3d/test_stable_diffusion_ldm3d.py class StableDiffusionLDM3DPipelineFastTests (line 39) | class StableDiffusionLDM3DPipelineFastTests(unittest.TestCase): method get_dummy_components (line 45) | def get_dummy_components(self): method get_dummy_inputs (line 100) | def get_dummy_inputs(self, device, seed=0): method test_stable_diffusion_ddim (line 114) | def test_stable_diffusion_ddim(self): method test_stable_diffusion_prompt_embeds (line 140) | def test_stable_diffusion_prompt_embeds(self): method test_stable_diffusion_negative_prompt (line 180) | def test_stable_diffusion_negative_prompt(self): class StableDiffusionLDM3DPipelineSlowTests (line 209) | class StableDiffusionLDM3DPipelineSlowTests(unittest.TestCase): method setUp (line 210) | def setUp(self): method tearDown (line 215) | def tearDown(self): method get_inputs (line 220) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_ldm3d_stable_diffusion (line 234) | def test_ldm3d_stable_diffusion(self): class StableDiffusionPipelineNightlyTests (line 260) | class StableDiffusionPipelineNightlyTests(unittest.TestCase): method setUp (line 261) | def setUp(self): method tearDown (line 266) | def tearDown(self): method get_inputs (line 271) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_ldm3d (line 285) | def test_ldm3d(self): method test_ldm3d_v2 (line 302) | def test_ldm3d_v2(self): FILE: diffusers/tests/pipelines/stable_diffusion_panorama/test_stable_diffusion_panorama.py class StableDiffusionPanoramaPipelineFastTests (line 47) | class StableDiffusionPanoramaPipelineFastTests( method get_dummy_components (line 60) | def get_dummy_components(self): method get_dummy_inputs (line 109) | def get_dummy_inputs(self, device, seed=0): method test_stable_diffusion_panorama_default_case (line 123) | def test_stable_diffusion_panorama_default_case(self): method test_stable_diffusion_panorama_circular_padding_case (line 139) | def test_stable_diffusion_panorama_circular_padding_case(self): method test_inference_batch_consistent (line 156) | def test_inference_batch_consistent(self): method test_inference_batch_single_identical (line 160) | def test_inference_batch_single_identical(self): method test_float16_inference (line 163) | def test_float16_inference(self): method test_stable_diffusion_panorama_negative_prompt (line 166) | def test_stable_diffusion_panorama_negative_prompt(self): method test_stable_diffusion_panorama_views_batch (line 185) | def test_stable_diffusion_panorama_views_batch(self): method test_stable_diffusion_panorama_views_batch_circular_padding (line 203) | def test_stable_diffusion_panorama_views_batch_circular_padding(self): method test_stable_diffusion_panorama_euler (line 221) | def test_stable_diffusion_panorama_euler(self): method test_stable_diffusion_panorama_pndm (line 241) | def test_stable_diffusion_panorama_pndm(self): class StableDiffusionPanoramaNightlyTests (line 264) | class StableDiffusionPanoramaNightlyTests(unittest.TestCase): method setUp (line 265) | def setUp(self): method tearDown (line 270) | def tearDown(self): method get_inputs (line 275) | def get_inputs(self, seed=0): method test_stable_diffusion_panorama_default (line 286) | def test_stable_diffusion_panorama_default(self): method test_stable_diffusion_panorama_k_lms (line 316) | def test_stable_diffusion_panorama_k_lms(self): method test_stable_diffusion_panorama_intermediate_state (line 349) | def test_stable_diffusion_panorama_intermediate_state(self): method test_stable_diffusion_panorama_pipeline_with_sequential_cpu_offloading (line 410) | def test_stable_diffusion_panorama_pipeline_with_sequential_cpu_offloa... FILE: diffusers/tests/pipelines/stable_diffusion_safe/test_safe_diffusion.py class SafeDiffusionPipelineFastTests (line 30) | class SafeDiffusionPipelineFastTests(unittest.TestCase): method setUp (line 31) | def setUp(self): method tearDown (line 37) | def tearDown(self): method dummy_image (line 44) | def dummy_image(self): method dummy_cond_unet (line 53) | def dummy_cond_unet(self): method dummy_vae (line 68) | def dummy_vae(self): method dummy_text_encoder (line 81) | def dummy_text_encoder(self): method dummy_extractor (line 97) | def dummy_extractor(self): method test_safe_diffusion_ddim (line 111) | def test_safe_diffusion_ddim(self): method test_stable_diffusion_pndm (line 164) | def test_stable_diffusion_pndm(self): method test_stable_diffusion_no_safety_checker (line 210) | def test_stable_diffusion_no_safety_checker(self): method test_stable_diffusion_fp16 (line 232) | def test_stable_diffusion_fp16(self): class SafeDiffusionPipelineIntegrationTests (line 266) | class SafeDiffusionPipelineIntegrationTests(unittest.TestCase): method setUp (line 267) | def setUp(self): method tearDown (line 273) | def tearDown(self): method test_harm_safe_stable_diffusion (line 279) | def test_harm_safe_stable_diffusion(self): method test_nudity_safe_stable_diffusion (line 340) | def test_nudity_safe_stable_diffusion(self): method test_nudity_safetychecker_safe_stable_diffusion (line 394) | def test_nudity_safetychecker_safe_stable_diffusion(self): FILE: diffusers/tests/pipelines/stable_diffusion_sag/test_stable_diffusion_sag.py class StableDiffusionSAGPipelineFastTests (line 46) | class StableDiffusionSAGPipelineFastTests( method get_dummy_components (line 59) | def get_dummy_components(self): method get_dummy_inputs (line 116) | def get_dummy_inputs(self, device, seed=0): method test_inference_batch_single_identical (line 131) | def test_inference_batch_single_identical(self): method test_xformers_attention_forwardGenerator_pass (line 135) | def test_xformers_attention_forwardGenerator_pass(self): method test_pipeline_different_schedulers (line 138) | def test_pipeline_different_schedulers(self): class StableDiffusionPipelineIntegrationTests (line 159) | class StableDiffusionPipelineIntegrationTests(unittest.TestCase): method setUp (line 160) | def setUp(self): method tearDown (line 166) | def tearDown(self): method test_stable_diffusion_1 (line 172) | def test_stable_diffusion_1(self): method test_stable_diffusion_2 (line 192) | def test_stable_diffusion_2(self): method test_stable_diffusion_2_non_square (line 212) | def test_stable_diffusion_2_non_square(self): FILE: diffusers/tests/pipelines/stable_diffusion_xl/test_stable_diffusion_xl.py class StableDiffusionXLPipelineFastTests (line 64) | class StableDiffusionXLPipelineFastTests( method get_dummy_components (line 79) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_inputs (line 151) | def get_dummy_inputs(self, device, seed=0): method test_stable_diffusion_xl_euler (line 165) | def test_stable_diffusion_xl_euler(self): method test_stable_diffusion_xl_euler_lcm (line 181) | def test_stable_diffusion_xl_euler_lcm(self): method test_stable_diffusion_xl_euler_lcm_custom_timesteps (line 198) | def test_stable_diffusion_xl_euler_lcm_custom_timesteps(self): method test_stable_diffusion_ays (line 217) | def test_stable_diffusion_ays(self): method test_stable_diffusion_xl_prompt_embeds (line 255) | def test_stable_diffusion_xl_prompt_embeds(self): method test_stable_diffusion_xl_negative_prompt_embeds (line 293) | def test_stable_diffusion_xl_negative_prompt_embeds(self): method test_ip_adapter_single (line 333) | def test_ip_adapter_single(self): method test_attention_slicing_forward_pass (line 339) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 342) | def test_inference_batch_single_identical(self): method test_save_load_optional_components (line 345) | def test_save_load_optional_components(self): method test_stable_diffusion_xl_offloads (line 349) | def test_stable_diffusion_xl_offloads(self): method test_stable_diffusion_xl_img2img_prompt_embeds_only (line 377) | def test_stable_diffusion_xl_img2img_prompt_embeds_only(self): method test_stable_diffusion_two_xl_mixture_of_denoiser_fast (line 413) | def test_stable_diffusion_two_xl_mixture_of_denoiser_fast(self): method test_stable_diffusion_two_xl_mixture_of_denoiser (line 514) | def test_stable_diffusion_two_xl_mixture_of_denoiser(self): method test_stable_diffusion_three_xl_mixture_of_denoiser (line 798) | def test_stable_diffusion_three_xl_mixture_of_denoiser(self): method test_stable_diffusion_xl_multi_prompts (line 901) | def test_stable_diffusion_xl_multi_prompts(self): method test_stable_diffusion_xl_negative_conditions (line 954) | def test_stable_diffusion_xl_negative_conditions(self): method test_stable_diffusion_xl_save_from_pretrained (line 975) | def test_stable_diffusion_xl_save_from_pretrained(self): method test_pipeline_interrupt (line 997) | def test_pipeline_interrupt(self): class StableDiffusionXLPipelineIntegrationTests (line 1051) | class StableDiffusionXLPipelineIntegrationTests(unittest.TestCase): method setUp (line 1052) | def setUp(self): method tearDown (line 1057) | def tearDown(self): method test_stable_diffusion_lcm (line 1062) | def test_stable_diffusion_lcm(self): FILE: diffusers/tests/pipelines/stable_diffusion_xl/test_stable_diffusion_xl_adapter.py class StableDiffusionXLAdapterPipelineFastTests (line 53) | class StableDiffusionXLAdapterPipelineFastTests( method get_dummy_components (line 60) | def get_dummy_components(self, adapter_type="full_adapter_xl", time_co... method get_dummy_components_with_full_downscaling (line 164) | def get_dummy_components_with_full_downscaling(self, adapter_type="ful... method get_dummy_inputs (line 271) | def get_dummy_inputs(self, device, seed=0, height=64, width=64, num_im... method test_ip_adapter_single (line 293) | def test_ip_adapter_single(self, from_multi=False, expected_pipe_slice... method test_stable_diffusion_adapter_default_case (line 302) | def test_stable_diffusion_adapter_default_case(self): method test_multiple_image_dimensions (line 327) | def test_multiple_image_dimensions(self, dim): method test_total_downscale_factor (line 347) | def test_total_downscale_factor(self, adapter_type): method test_save_load_optional_components (line 377) | def test_save_load_optional_components(self): method test_adapter_sdxl_lcm (line 380) | def test_adapter_sdxl_lcm(self): method test_adapter_sdxl_lcm_custom_timesteps (line 400) | def test_adapter_sdxl_lcm_custom_timesteps(self): class StableDiffusionXLMultiAdapterPipelineFastTests (line 423) | class StableDiffusionXLMultiAdapterPipelineFastTests( method get_dummy_components (line 426) | def get_dummy_components(self, time_cond_proj_dim=None): method get_dummy_components_with_full_downscaling (line 429) | def get_dummy_components_with_full_downscaling(self): method get_dummy_inputs (line 432) | def get_dummy_inputs(self, device, seed=0, height=64, width=64): method test_stable_diffusion_adapter_default_case (line 437) | def test_stable_diffusion_adapter_default_case(self): method test_ip_adapter_single (line 454) | def test_ip_adapter_single(self): method test_inference_batch_consistent (line 460) | def test_inference_batch_consistent( method test_num_images_per_prompt (line 518) | def test_num_images_per_prompt(self): method test_inference_batch_single_identical (line 547) | def test_inference_batch_single_identical( method test_adapter_sdxl_lcm (line 630) | def test_adapter_sdxl_lcm(self): method test_adapter_sdxl_lcm_custom_timesteps (line 653) | def test_adapter_sdxl_lcm_custom_timesteps(self): FILE: diffusers/tests/pipelines/stable_diffusion_xl/test_stable_diffusion_xl_img2img.py class StableDiffusionXLImg2ImgPipelineFastTests (line 63) | class StableDiffusionXLImg2ImgPipelineFastTests( method get_dummy_components (line 76) | def get_dummy_components(self, skip_first_text_encoder=False, time_con... method get_dummy_tiny_autoencoder (line 173) | def get_dummy_tiny_autoencoder(self): method test_components_function (line 176) | def test_components_function(self): method get_dummy_inputs (line 184) | def get_dummy_inputs(self, device, seed=0): method test_stable_diffusion_xl_img2img_euler (line 202) | def test_stable_diffusion_xl_img2img_euler(self): method test_stable_diffusion_xl_img2img_euler_lcm (line 219) | def test_stable_diffusion_xl_img2img_euler_lcm(self): method test_stable_diffusion_xl_img2img_euler_lcm_custom_timesteps (line 237) | def test_stable_diffusion_xl_img2img_euler_lcm_custom_timesteps(self): method test_attention_slicing_forward_pass (line 257) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 260) | def test_inference_batch_single_identical(self): method test_save_load_optional_components (line 264) | def test_save_load_optional_components(self): method test_stable_diffusion_xl_img2img_negative_prompt_embeds (line 267) | def test_stable_diffusion_xl_img2img_negative_prompt_embeds(self): method test_ip_adapter_single (line 309) | def test_ip_adapter_single(self): method test_stable_diffusion_xl_img2img_tiny_autoencoder (line 315) | def test_stable_diffusion_xl_img2img_tiny_autoencoder(self): method test_stable_diffusion_xl_offloads (line 333) | def test_stable_diffusion_xl_offloads(self): method test_stable_diffusion_xl_multi_prompts (line 362) | def test_stable_diffusion_xl_multi_prompts(self): method test_stable_diffusion_xl_img2img_negative_conditions (line 427) | def test_stable_diffusion_xl_img2img_negative_conditions(self): method test_pipeline_interrupt (line 455) | def test_pipeline_interrupt(self): class StableDiffusionXLImg2ImgRefinerOnlyPipelineFastTests (line 514) | class StableDiffusionXLImg2ImgRefinerOnlyPipelineFastTests( method get_dummy_components (line 524) | def get_dummy_components(self): method test_components_function (line 592) | def test_components_function(self): method get_dummy_inputs (line 600) | def get_dummy_inputs(self, device, seed=0): method test_stable_diffusion_xl_img2img_euler (line 618) | def test_stable_diffusion_xl_img2img_euler(self): method test_stable_diffusion_xl_offloads (line 636) | def test_stable_diffusion_xl_offloads(self): method test_stable_diffusion_xl_img2img_negative_conditions (line 665) | def test_stable_diffusion_xl_img2img_negative_conditions(self): method test_stable_diffusion_xl_img2img_negative_prompt_embeds (line 693) | def test_stable_diffusion_xl_img2img_negative_prompt_embeds(self): method test_stable_diffusion_xl_img2img_prompt_embeds_only (line 735) | def test_stable_diffusion_xl_img2img_prompt_embeds_only(self): method test_attention_slicing_forward_pass (line 771) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 774) | def test_inference_batch_single_identical(self): method test_save_load_optional_components (line 777) | def test_save_load_optional_components(self): FILE: diffusers/tests/pipelines/stable_diffusion_xl/test_stable_diffusion_xl_inpaint.py class StableDiffusionXLInpaintPipelineFastTests (line 57) | class StableDiffusionXLInpaintPipelineFastTests( method get_dummy_components (line 75) | def get_dummy_components(self, skip_first_text_encoder=False, time_con... method get_dummy_inputs (line 172) | def get_dummy_inputs(self, device, seed=0): method get_dummy_inputs_2images (line 197) | def get_dummy_inputs_2images(self, device, seed=0, img_res=64): method test_ip_adapter_single (line 226) | def test_ip_adapter_single(self): method test_components_function (line 232) | def test_components_function(self): method test_stable_diffusion_xl_inpaint_euler (line 240) | def test_stable_diffusion_xl_inpaint_euler(self): method test_stable_diffusion_xl_inpaint_euler_lcm (line 257) | def test_stable_diffusion_xl_inpaint_euler_lcm(self): method test_stable_diffusion_xl_inpaint_euler_lcm_custom_timesteps (line 275) | def test_stable_diffusion_xl_inpaint_euler_lcm_custom_timesteps(self): method test_attention_slicing_forward_pass (line 295) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 298) | def test_inference_batch_single_identical(self): method test_save_load_optional_components (line 302) | def test_save_load_optional_components(self): method test_stable_diffusion_xl_inpaint_negative_prompt_embeds (line 305) | def test_stable_diffusion_xl_inpaint_negative_prompt_embeds(self): method test_stable_diffusion_xl_offloads (line 346) | def test_stable_diffusion_xl_offloads(self): method test_stable_diffusion_xl_refiner (line 374) | def test_stable_diffusion_xl_refiner(self): method test_stable_diffusion_two_xl_mixture_of_denoiser_fast (line 392) | def test_stable_diffusion_two_xl_mixture_of_denoiser_fast(self): method test_stable_diffusion_two_xl_mixture_of_denoiser (line 452) | def test_stable_diffusion_two_xl_mixture_of_denoiser(self): method test_stable_diffusion_three_xl_mixture_of_denoiser (line 519) | def test_stable_diffusion_three_xl_mixture_of_denoiser(self): method test_stable_diffusion_xl_multi_prompts (line 609) | def test_stable_diffusion_xl_multi_prompts(self): method test_stable_diffusion_xl_img2img_negative_conditions (line 668) | def test_stable_diffusion_xl_img2img_negative_conditions(self): method test_stable_diffusion_xl_inpaint_mask_latents (line 695) | def test_stable_diffusion_xl_inpaint_mask_latents(self): method test_stable_diffusion_xl_inpaint_2_images (line 727) | def test_stable_diffusion_xl_inpaint_2_images(self): method test_pipeline_interrupt (line 758) | def test_pipeline_interrupt(self): FILE: diffusers/tests/pipelines/stable_diffusion_xl/test_stable_diffusion_xl_instruction_pix2pix.py class StableDiffusionXLInstructPix2PixPipelineFastTests (line 50) | class StableDiffusionXLInstructPix2PixPipelineFastTests( method get_dummy_components (line 63) | def get_dummy_components(self): method get_dummy_inputs (line 132) | def get_dummy_inputs(self, device, seed=0): method test_components_function (line 150) | def test_components_function(self): method test_inference_batch_single_identical (line 157) | def test_inference_batch_single_identical(self): method test_attention_slicing_forward_pass (line 160) | def test_attention_slicing_forward_pass(self): method test_latents_input (line 164) | def test_latents_input(self): method test_cfg (line 185) | def test_cfg(self): method test_save_load_optional_components (line 188) | def test_save_load_optional_components(self): FILE: diffusers/tests/pipelines/stable_diffusion_xl/test_stable_diffusion_xl_k_diffusion.py class StableDiffusionXLKPipelineIntegrationTests (line 31) | class StableDiffusionXLKPipelineIntegrationTests(unittest.TestCase): method setUp (line 34) | def setUp(self): method tearDown (line 40) | def tearDown(self): method test_stable_diffusion_xl (line 46) | def test_stable_diffusion_xl(self): method test_stable_diffusion_karras_sigmas (line 78) | def test_stable_diffusion_karras_sigmas(self): method test_stable_diffusion_noise_sampler_seed (line 111) | def test_stable_diffusion_noise_sampler_seed(self): FILE: diffusers/tests/pipelines/stable_unclip/test_stable_unclip.py class StableUnCLIPPipelineFastTests (line 30) | class StableUnCLIPPipelineFastTests( method get_dummy_components (line 42) | def get_dummy_components(self): method get_dummy_inputs (line 161) | def get_dummy_inputs(self, device, seed=0): method test_attention_slicing_forward_pass (line 177) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 184) | def test_inference_batch_single_identical(self): class StableUnCLIPPipelineIntegrationTests (line 190) | class StableUnCLIPPipelineIntegrationTests(unittest.TestCase): method setUp (line 191) | def setUp(self): method tearDown (line 197) | def tearDown(self): method test_stable_unclip (line 203) | def test_stable_unclip(self): method test_stable_unclip_pipeline_with_sequential_cpu_offloading (line 224) | def test_stable_unclip_pipeline_with_sequential_cpu_offloading(self): FILE: diffusers/tests/pipelines/stable_unclip/test_stable_unclip_img2img.py class StableUnCLIPImg2ImgPipelineFastTests (line 43) | class StableUnCLIPImg2ImgPipelineFastTests( method get_dummy_components (line 54) | def get_dummy_components(self): method get_dummy_inputs (line 149) | def get_dummy_inputs(self, device, seed=0, pil_image=True): method test_image_embeds_none (line 172) | def test_image_embeds_none(self): method test_attention_slicing_forward_pass (line 191) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 198) | def test_inference_batch_single_identical(self): method test_xformers_attention_forwardGenerator_pass (line 205) | def test_xformers_attention_forwardGenerator_pass(self): class StableUnCLIPImg2ImgPipelineIntegrationTests (line 211) | class StableUnCLIPImg2ImgPipelineIntegrationTests(unittest.TestCase): method setUp (line 212) | def setUp(self): method tearDown (line 218) | def tearDown(self): method test_stable_unclip_l_img2img (line 224) | def test_stable_unclip_l_img2img(self): method test_stable_unclip_h_img2img (line 251) | def test_stable_unclip_h_img2img(self): method test_stable_unclip_img2img_pipeline_with_sequential_cpu_offloading (line 278) | def test_stable_unclip_img2img_pipeline_with_sequential_cpu_offloading... FILE: diffusers/tests/pipelines/stable_video_diffusion/test_stable_video_diffusion.py function to_np (line 39) | def to_np(tensor): class StableVideoDiffusionPipelineFastTests (line 46) | class StableVideoDiffusionPipelineFastTests(PipelineTesterMixin, unittes... method get_dummy_components (line 59) | def get_dummy_components(self): method get_dummy_inputs (line 125) | def get_dummy_inputs(self, device, seed=0): method test_attention_slicing_forward_pass (line 146) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 150) | def test_inference_batch_single_identical( method test_inference_batch_consistent (line 187) | def test_inference_batch_consistent(self): method test_np_output_type (line 190) | def test_np_output_type(self): method test_dict_tuple_outputs_equivalent (line 207) | def test_dict_tuple_outputs_equivalent(self, expected_max_difference=1... method test_float16_inference (line 225) | def test_float16_inference(self, expected_max_diff=5e-2): method test_save_load_float16 (line 254) | def test_save_load_float16(self, expected_max_diff=1e-2): method test_save_load_optional_components (line 293) | def test_save_load_optional_components(self, expected_max_difference=1... method test_save_load_local (line 334) | def test_save_load_local(self, expected_max_difference=9e-4): method test_to_device (line 370) | def test_to_device(self): method test_to_dtype (line 393) | def test_to_dtype(self): method test_sequential_cpu_offload_forward_pass (line 409) | def test_sequential_cpu_offload_forward_pass(self, expected_max_diff=1... method test_model_cpu_offload_forward_pass (line 434) | def test_model_cpu_offload_forward_pass(self, expected_max_diff=2e-4): method test_xformers_attention_forwardGenerator_pass (line 469) | def test_xformers_attention_forwardGenerator_pass(self): method test_disable_cfg (line 499) | def test_disable_cfg(self): class StableVideoDiffusionPipelineSlowTests (line 518) | class StableVideoDiffusionPipelineSlowTests(unittest.TestCase): method setUp (line 519) | def setUp(self): method tearDown (line 525) | def tearDown(self): method test_sd_video (line 531) | def test_sd_video(self): FILE: diffusers/tests/pipelines/test_pipeline_utils.py class IsSafetensorsCompatibleTests (line 6) | class IsSafetensorsCompatibleTests(unittest.TestCase): method test_all_is_compatible (line 7) | def test_all_is_compatible(self): method test_diffusers_model_is_compatible (line 20) | def test_diffusers_model_is_compatible(self): method test_diffusers_model_is_not_compatible (line 27) | def test_diffusers_model_is_not_compatible(self): method test_transformer_model_is_compatible (line 40) | def test_transformer_model_is_compatible(self): method test_transformer_model_is_not_compatible (line 47) | def test_transformer_model_is_not_compatible(self): method test_all_is_compatible_variant (line 60) | def test_all_is_compatible_variant(self): method test_diffusers_model_is_compatible_variant (line 74) | def test_diffusers_model_is_compatible_variant(self): method test_diffusers_model_is_compatible_variant_partial (line 82) | def test_diffusers_model_is_compatible_variant_partial(self): method test_diffusers_model_is_not_compatible_variant (line 91) | def test_diffusers_model_is_not_compatible_variant(self): method test_transformer_model_is_compatible_variant (line 105) | def test_transformer_model_is_compatible_variant(self): method test_transformer_model_is_compatible_variant_partial (line 113) | def test_transformer_model_is_compatible_variant_partial(self): method test_transformer_model_is_not_compatible_variant (line 122) | def test_transformer_model_is_not_compatible_variant(self): FILE: diffusers/tests/pipelines/test_pipelines.py function _test_from_save_pretrained_dynamo (line 91) | def _test_from_save_pretrained_dynamo(in_queue, out_queue, timeout): class CustomEncoder (line 136) | class CustomEncoder(ModelMixin, ConfigMixin): method __init__ (line 137) | def __init__(self): class CustomPipeline (line 142) | class CustomPipeline(DiffusionPipeline): method __init__ (line 143) | def __init__(self, encoder: CustomEncoder, scheduler: DDIMScheduler): class DownloadTests (line 148) | class DownloadTests(unittest.TestCase): method test_one_request_upon_cached (line 149) | def test_one_request_upon_cached(self): method test_less_downloads_passed_object (line 177) | def test_less_downloads_passed_object(self): method test_less_downloads_passed_object_calls (line 194) | def test_less_downloads_passed_object_calls(self): method test_download_only_pytorch (line 226) | def test_download_only_pytorch(self): method test_force_safetensors_error (line 242) | def test_force_safetensors_error(self): method test_download_safetensors (line 253) | def test_download_safetensors(self): method test_download_safetensors_index (line 269) | def test_download_safetensors_index(self): method test_download_bin_index (line 293) | def test_download_bin_index(self): method test_download_no_openvino_by_default (line 317) | def test_download_no_openvino_by_default(self): method test_download_no_onnx_by_default (line 331) | def test_download_no_onnx_by_default(self): method test_download_onnx_by_default_for_onnx_pipelines (line 347) | def test_download_onnx_by_default_for_onnx_pipelines(self): method test_download_no_safety_checker (line 362) | def test_download_no_safety_checker(self): method test_load_no_safety_checker_explicit_locally (line 378) | def test_load_no_safety_checker_explicit_locally(self): method test_load_no_safety_checker_default_locally (line 398) | def test_load_no_safety_checker_default_locally(self): method test_cached_files_are_used_when_no_internet (line 417) | def test_cached_files_are_used_when_no_internet(self): method test_local_files_only_are_used_when_no_internet (line 444) | def test_local_files_only_are_used_when_no_internet(self): method test_download_from_variant_folder (line 480) | def test_download_from_variant_folder(self): method test_download_variant_all (line 499) | def test_download_variant_all(self): method test_download_variant_partly (line 524) | def test_download_variant_partly(self): method test_download_broken_variant (line 552) | def test_download_broken_variant(self): method test_local_save_load_index (line 584) | def test_local_save_load_index(self): method test_text_inversion_download (line 611) | def test_text_inversion_download(self): method test_text_inversion_multi_tokens (line 798) | def test_text_inversion_multi_tokens(self): method test_download_ignore_files (line 846) | def test_download_ignore_files(self): method test_get_pipeline_class_from_flax (line 859) | def test_get_pipeline_class_from_flax(self): class CustomPipelineTests (line 868) | class CustomPipelineTests(unittest.TestCase): method test_load_custom_pipeline (line 869) | def test_load_custom_pipeline(self): method test_load_custom_github (line 878) | def test_load_custom_github(self): method test_run_custom_pipeline (line 903) | def test_run_custom_pipeline(self): method test_remote_components (line 915) | def test_remote_components(self): method test_remote_auto_custom_pipe (line 948) | def test_remote_auto_custom_pipe(self): method test_local_custom_pipeline_repo (line 967) | def test_local_custom_pipeline_repo(self): method test_local_custom_pipeline_file (line 980) | def test_local_custom_pipeline_file(self): method test_custom_model_and_pipeline (line 994) | def test_custom_model_and_pipeline(self): method test_download_from_git (line 1015) | def test_download_from_git(self): method test_save_pipeline_change_config (line 1040) | def test_save_pipeline_change_config(self): class PipelineFastTests (line 1060) | class PipelineFastTests(unittest.TestCase): method setUp (line 1061) | def setUp(self): method tearDown (line 1067) | def tearDown(self): method dummy_image (line 1073) | def dummy_image(self): method dummy_uncond_unet (line 1081) | def dummy_uncond_unet(self, sample_size=32): method dummy_cond_unet (line 1094) | def dummy_cond_unet(self, sample_size=32): method dummy_vae (line 1109) | def dummy_vae(self): method dummy_text_encoder (line 1122) | def dummy_text_encoder(self): method dummy_extractor (line 1138) | def dummy_extractor(self): method test_uncond_unet_components (line 1160) | def test_uncond_unet_components(self, scheduler_fn=DDPMScheduler, pipe... method test_stable_diffusion_components (line 1174) | def test_stable_diffusion_components(self): method test_pipe_false_offload_warn (line 1229) | def test_pipe_false_offload_warn(self): method test_set_scheduler (line 1264) | def test_set_scheduler(self): method test_set_component_to_none (line 1296) | def test_set_component_to_none(self): method test_optional_components_is_none (line 1336) | def test_optional_components_is_none(self): method test_set_scheduler_consistency (line 1359) | def test_set_scheduler_consistency(self): method test_save_safe_serialization (line 1403) | def test_save_safe_serialization(self): method test_no_pytorch_download_when_doing_safetensors (line 1430) | def test_no_pytorch_download_when_doing_safetensors(self): method test_no_safetensors_download_when_doing_pytorch (line 1449) | def test_no_safetensors_download_when_doing_pytorch(self): method test_optional_components (line 1471) | def test_optional_components(self): method test_name_or_path (line 1575) | def test_name_or_path(self): method test_error_no_variant_available (line 1587) | def test_error_no_variant_available(self): method test_pipe_to (line 1597) | def test_pipe_to(self): method test_pipe_same_device_id_offload (line 1658) | def test_pipe_same_device_id_offload(self): class PipelineSlowTests (line 1683) | class PipelineSlowTests(unittest.TestCase): method setUp (line 1684) | def setUp(self): method tearDown (line 1690) | def tearDown(self): method test_smart_download (line 1696) | def test_smart_download(self): method test_warning_unused_kwargs (line 1717) | def test_warning_unused_kwargs(self): method test_from_save_pretrained (line 1734) | def test_from_save_pretrained(self): method test_from_save_pretrained_dynamo (line 1770) | def test_from_save_pretrained_dynamo(self): method test_from_pretrained_hub (line 1773) | def test_from_pretrained_hub(self): method test_from_pretrained_hub_pass_model (line 1794) | def test_from_pretrained_hub_pass_model(self): method test_output_format (line 1817) | def test_output_format(self): method test_from_flax_from_pt (line 1840) | def test_from_flax_from_pt(self): method test_weighted_prompts_compel (line 1881) | def test_weighted_prompts_compel(self): class PipelineNightlyTests (line 1914) | class PipelineNightlyTests(unittest.TestCase): method setUp (line 1915) | def setUp(self): method tearDown (line 1921) | def tearDown(self): method test_ddpm_ddim_equality_batched (line 1927) | def test_ddpm_ddim_equality_batched(self): FILE: diffusers/tests/pipelines/test_pipelines_auto.py class AutoPipelineFastTest (line 51) | class AutoPipelineFastTest(unittest.TestCase): method dummy_image_encoder (line 53) | def dummy_image_encoder(self): method test_from_pipe_consistent (line 66) | def test_from_pipe_consistent(self): method test_from_pipe_override (line 78) | def test_from_pipe_override(self): method test_from_pipe_consistent_sdxl (line 89) | def test_from_pipe_consistent_sdxl(self): method test_kwargs_local_files_only (line 103) | def test_kwargs_local_files_only(self): method test_from_pipe_controlnet_text2img (line 126) | def test_from_pipe_controlnet_text2img(self): method test_from_pipe_controlnet_img2img (line 138) | def test_from_pipe_controlnet_img2img(self): method test_from_pipe_controlnet_inpaint (line 150) | def test_from_pipe_controlnet_inpaint(self): method test_from_pipe_controlnet_new_task (line 162) | def test_from_pipe_controlnet_new_task(self): method test_from_pipe_optional_components (line 222) | def test_from_pipe_optional_components(self): class AutoPipelineIntegrationTest (line 238) | class AutoPipelineIntegrationTest(unittest.TestCase): method test_pipe_auto (line 239) | def test_pipe_auto(self): method test_from_pipe_consistent (line 300) | def test_from_pipe_consistent(self): method test_controlnet (line 319) | def test_controlnet(self): FILE: diffusers/tests/pipelines/test_pipelines_combined.py function state_dicts_almost_equal (line 31) | def state_dicts_almost_equal(sd1, sd2): class CombinedPipelineFastTest (line 43) | class CombinedPipelineFastTest(unittest.TestCase): method modelcard_has_connected_pipeline (line 44) | def modelcard_has_connected_pipeline(self, model_id): method test_correct_modelcard_format (line 51) | def test_correct_modelcard_format(self): method test_load_connected_checkpoint_when_specified (line 58) | def test_load_connected_checkpoint_when_specified(self): method test_load_connected_checkpoint_default (line 82) | def test_load_connected_checkpoint_default(self): method test_load_connected_checkpoint_with_passed_obj (line 111) | def test_load_connected_checkpoint_with_passed_obj(self): FILE: diffusers/tests/pipelines/test_pipelines_common.py function to_np (line 58) | def to_np(tensor): function check_same_shape (line 65) | def check_same_shape(tensor_list): class SDFunctionTesterMixin (line 70) | class SDFunctionTesterMixin: method test_vae_slicing (line 76) | def test_vae_slicing(self, image_count=4): method test_vae_tiling (line 101) | def test_vae_tiling(self): method test_freeu_enabled (line 134) | def test_freeu_enabled(self): method test_freeu_disabled (line 157) | def test_freeu_disabled(self): method test_fused_qkv_projections (line 186) | def test_fused_qkv_projections(self): class IPAdapterTesterMixin (line 221) | class IPAdapterTesterMixin: method test_pipeline_signature (line 227) | def test_pipeline_signature(self): method _get_dummy_image_embeds (line 242) | def _get_dummy_image_embeds(self, cross_attention_dim: int = 32): method _get_dummy_faceid_image_embeds (line 245) | def _get_dummy_faceid_image_embeds(self, cross_attention_dim: int = 32): method _get_dummy_masks (line 248) | def _get_dummy_masks(self, input_size: int = 64): method _modify_inputs_for_ip_adapter_test (line 253) | def _modify_inputs_for_ip_adapter_test(self, inputs: Dict[str, Any]): method test_ip_adapter_single (line 262) | def test_ip_adapter_single(self, expected_max_diff: float = 1e-4, expe... method test_ip_adapter_multi (line 310) | def test_ip_adapter_multi(self, expected_max_diff: float = 1e-4): method test_ip_adapter_cfg (line 351) | def test_ip_adapter_cfg(self, expected_max_diff: float = 1e-4): method test_ip_adapter_masks (line 380) | def test_ip_adapter_masks(self, expected_max_diff: float = 1e-4): method test_ip_adapter_faceid (line 425) | def test_ip_adapter_faceid(self, expected_max_diff: float = 1e-4): class PipelineLatentTesterMixin (line 466) | class PipelineLatentTesterMixin: method image_params (line 474) | def image_params(self) -> frozenset: method image_latents_params (line 481) | def image_latents_params(self) -> frozenset: method get_dummy_inputs_by_type (line 487) | def get_dummy_inputs_by_type(self, device, seed=0, input_image_type="p... method test_pt_np_pil_outputs_equivalent (line 524) | def test_pt_np_pil_outputs_equivalent(self, expected_max_diff=1e-4): method _test_pt_np_pil_outputs_equivalent (line 527) | def _test_pt_np_pil_outputs_equivalent(self, expected_max_diff=1e-4, i... method test_pt_np_pil_inputs_equivalent (line 551) | def test_pt_np_pil_inputs_equivalent(self): method test_latents_input (line 569) | def test_latents_input(self): method test_multi_vae (line 594) | def test_multi_vae(self): class PipelineFromPipeTesterMixin (line 624) | class PipelineFromPipeTesterMixin: method original_pipeline_class (line 626) | def original_pipeline_class(self): method get_dummy_inputs_pipe (line 634) | def get_dummy_inputs_pipe(self, device, seed=0): method get_dummy_inputs_for_pipe_original (line 640) | def get_dummy_inputs_for_pipe_original(self, device, seed=0): method test_from_pipe_consistent_config (line 647) | def test_from_pipe_consistent_config(self): method test_from_pipe_consistent_forward_pass (line 684) | def test_from_pipe_consistent_forward_pass(self, expected_max_diff=1e-3): method test_from_pipe_consistent_forward_pass_cpu_offload (line 754) | def test_from_pipe_consistent_forward_pass_cpu_offload(self, expected_... class PipelineKarrasSchedulerTesterMixin (line 804) | class PipelineKarrasSchedulerTesterMixin: method test_karras_schedulers_shape (line 811) | def test_karras_schedulers_shape( class PipelineTesterMixin (line 846) | class PipelineTesterMixin: method get_generator (line 872) | def get_generator(self, seed): method pipeline_class (line 878) | def pipeline_class(self) -> Union[Callable, DiffusionPipeline]: method get_dummy_components (line 884) | def get_dummy_components(self): method get_dummy_inputs (line 890) | def get_dummy_inputs(self, device, seed=0): method params (line 897) | def params(self) -> frozenset: method batch_params (line 912) | def batch_params(self) -> frozenset: method callback_cfg_params (line 926) | def callback_cfg_params(self) -> frozenset: method setUp (line 939) | def setUp(self): method tearDown (line 945) | def tearDown(self): method test_save_load_local (line 951) | def test_save_load_local(self, expected_max_difference=5e-4): method test_pipeline_call_signature (line 990) | def test_pipeline_call_signature(self): method test_inference_batch_consistent (line 1029) | def test_inference_batch_consistent(self, batch_sizes=[2]): method _test_inference_batch_consistent (line 1032) | def _test_inference_batch_consistent( method test_inference_batch_single_identical (line 1081) | def test_inference_batch_single_identical(self, batch_size=3, expected... method _test_inference_batch_single_identical (line 1084) | def _test_inference_batch_single_identical( method test_dict_tuple_outputs_equivalent (line 1139) | def test_dict_tuple_outputs_equivalent(self, expected_slice=None, expe... method test_components_function (line 1167) | def test_components_function(self): method test_float16_inference (line 1177) | def test_float16_inference(self, expected_max_diff=5e-2): method test_save_load_float16 (line 1214) | def test_save_load_float16(self, expected_max_diff=1e-2): method test_save_load_optional_components (line 1253) | def test_save_load_optional_components(self, expected_max_difference=1... method test_to_device (line 1295) | def test_to_device(self): method test_to_dtype (line 1314) | def test_to_dtype(self): method test_attention_slicing_forward_pass (line 1326) | def test_attention_slicing_forward_pass(self, expected_max_diff=1e-3): method _test_attention_slicing_forward_pass (line 1329) | def _test_attention_slicing_forward_pass( method test_sequential_cpu_offload_forward_pass (line 1362) | def test_sequential_cpu_offload_forward_pass(self, expected_max_diff=1... method test_model_cpu_offload_forward_pass (line 1425) | def test_model_cpu_offload_forward_pass(self, expected_max_diff=2e-4): method test_cpu_offload_forward_pass_twice (line 1482) | def test_cpu_offload_forward_pass_twice(self, expected_max_diff=2e-4): method test_sequential_offload_forward_pass_twice (line 1539) | def test_sequential_offload_forward_pass_twice(self, expected_max_diff... method test_xformers_attention_forwardGenerator_pass (line 1604) | def test_xformers_attention_forwardGenerator_pass(self): method _test_xformers_attention_forwardGenerator_pass (line 1607) | def _test_xformers_attention_forwardGenerator_pass( method test_progress_bar (line 1641) | def test_progress_bar(self): method test_num_images_per_prompt (line 1663) | def test_num_images_per_prompt(self): method test_cfg (line 1689) | def test_cfg(self): method test_callback_inputs (line 1710) | def test_callback_inputs(self): method test_callback_cfg (line 1772) | def test_callback_cfg(self): method test_StableDiffusionMixin_component (line 1811) | def test_StableDiffusionMixin_component(self): class PipelinePushToHubTester (line 1828) | class PipelinePushToHubTester(unittest.TestCase): method get_pipeline_components (line 1833) | def get_pipeline_components(self): method test_push_to_hub (line 1898) | def test_push_to_hub(self): method test_push_to_hub_in_organization (line 1922) | def test_push_to_hub_in_organization(self): method test_push_to_hub_library_name (line 1950) | def test_push_to_hub_library_name(self): class SDXLOptionalComponentsTesterMixin (line 1965) | class SDXLOptionalComponentsTesterMixin: method encode_prompt (line 1966) | def encode_prompt( method _test_save_load_optional_components (line 2041) | def _test_save_load_optional_components(self, expected_max_difference=... function assert_mean_pixel_difference (line 2109) | def assert_mean_pixel_difference(image, expected_image, expected_max_dif... FILE: diffusers/tests/pipelines/test_pipelines_flax.py class DownloadTests (line 36) | class DownloadTests(unittest.TestCase): method test_download_only_pytorch (line 37) | def test_download_only_pytorch(self): class FlaxPipelineTests (line 54) | class FlaxPipelineTests(unittest.TestCase): method test_dummy_all_tpus (line 55) | def test_dummy_all_tpus(self): method test_stable_diffusion_v1_4 (line 87) | def test_stable_diffusion_v1_4(self): method test_stable_diffusion_v1_4_bfloat_16 (line 116) | def test_stable_diffusion_v1_4_bfloat_16(self): method test_stable_diffusion_v1_4_bfloat_16_with_safety (line 145) | def test_stable_diffusion_v1_4_bfloat_16_with_safety(self): method test_stable_diffusion_v1_4_bfloat_16_ddim (line 174) | def test_stable_diffusion_v1_4_bfloat_16_ddim(self): method test_jax_memory_efficient_attention (line 218) | def test_jax_memory_efficient_attention(self): FILE: diffusers/tests/pipelines/test_pipelines_onnx_common.py class OnnxPipelineTesterMixin (line 5) | class OnnxPipelineTesterMixin: FILE: diffusers/tests/pipelines/text_to_video_synthesis/test_text_to_video.py class TextToVideoSDPipelineFastTests (line 48) | class TextToVideoSDPipelineFastTests(PipelineTesterMixin, SDFunctionTest... method get_dummy_components (line 64) | def get_dummy_components(self): method get_dummy_inputs (line 122) | def get_dummy_inputs(self, device, seed=0): method test_dict_tuple_outputs_equivalent (line 136) | def test_dict_tuple_outputs_equivalent(self): method test_text_to_video_default_case (line 142) | def test_text_to_video_default_case(self): method test_attention_slicing_forward_pass (line 161) | def test_attention_slicing_forward_pass(self): method test_xformers_attention_forwardGenerator_pass (line 168) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_consistent (line 173) | def test_inference_batch_consistent(self): method test_inference_batch_single_identical (line 178) | def test_inference_batch_single_identical(self): method test_num_images_per_prompt (line 182) | def test_num_images_per_prompt(self): method test_progress_bar (line 185) | def test_progress_bar(self): class TextToVideoSDPipelineSlowTests (line 192) | class TextToVideoSDPipelineSlowTests(unittest.TestCase): method setUp (line 193) | def setUp(self): method tearDown (line 199) | def tearDown(self): method test_two_step_model (line 205) | def test_two_step_model(self): method test_two_step_model_with_freeu (line 219) | def test_two_step_model_with_freeu(self): FILE: diffusers/tests/pipelines/text_to_video_synthesis/test_text_to_video_zero.py class TextToVideoZeroPipelineSlowTests (line 29) | class TextToVideoZeroPipelineSlowTests(unittest.TestCase): method setUp (line 30) | def setUp(self): method tearDown (line 36) | def tearDown(self): method test_full_model (line 42) | def test_full_model(self): FILE: diffusers/tests/pipelines/text_to_video_synthesis/test_text_to_video_zero_sdxl.py function to_np (line 39) | def to_np(tensor): class TextToVideoZeroSDXLPipelineFastTests (line 46) | class TextToVideoZeroSDXLPipelineFastTests(PipelineTesterMixin, Pipeline... method get_dummy_components (line 54) | def get_dummy_components(self, seed=0): method get_dummy_inputs (line 135) | def get_dummy_inputs(self, device, seed=0): method get_generator (line 153) | def get_generator(self, device, seed=0): method test_text_to_video_zero_sdxl (line 160) | def test_text_to_video_zero_sdxl(self): method test_attention_slicing_forward_pass (line 180) | def test_attention_slicing_forward_pass(self): method test_cfg (line 183) | def test_cfg(self): method test_dict_tuple_outputs_equivalent (line 203) | def test_dict_tuple_outputs_equivalent(self, expected_max_difference=1... method test_float16_inference (line 216) | def test_float16_inference(self, expected_max_diff=5e-2): method test_inference_batch_consistent (line 248) | def test_inference_batch_consistent(self): method test_inference_batch_single_identical (line 254) | def test_inference_batch_single_identical(self): method test_model_cpu_offload_forward_pass (line 261) | def test_model_cpu_offload_forward_pass(self, expected_max_diff=2e-4): method test_pipeline_call_signature (line 278) | def test_pipeline_call_signature(self): method test_progress_bar (line 281) | def test_progress_bar(self): method test_save_load_float16 (line 304) | def test_save_load_float16(self, expected_max_diff=1e-2): method test_save_load_local (line 340) | def test_save_load_local(self): method test_save_load_optional_components (line 346) | def test_save_load_optional_components(self): method test_sequential_cpu_offload_forward_pass (line 352) | def test_sequential_cpu_offload_forward_pass(self): method test_to_device (line 356) | def test_to_device(self): method test_xformers_attention_forwardGenerator_pass (line 378) | def test_xformers_attention_forwardGenerator_pass(self): class TextToVideoZeroSDXLPipelineSlowTests (line 384) | class TextToVideoZeroSDXLPipelineSlowTests(unittest.TestCase): method setUp (line 385) | def setUp(self): method tearDown (line 391) | def tearDown(self): method test_full_model (line 397) | def test_full_model(self): FILE: diffusers/tests/pipelines/text_to_video_synthesis/test_video_to_video.py class VideoToVideoSDPipelineFastTests (line 51) | class VideoToVideoSDPipelineFastTests(PipelineTesterMixin, unittest.Test... method get_dummy_components (line 70) | def get_dummy_components(self): method get_dummy_inputs (line 132) | def get_dummy_inputs(self, device, seed=0): method test_text_to_video_default_case (line 150) | def test_text_to_video_default_case(self): method test_save_load_optional_components (line 168) | def test_save_load_optional_components(self): method test_dict_tuple_outputs_equivalent (line 172) | def test_dict_tuple_outputs_equivalent(self): method test_save_load_local (line 176) | def test_save_load_local(self): method test_xformers_attention_forwardGenerator_pass (line 183) | def test_xformers_attention_forwardGenerator_pass(self): method test_inference_batch_consistent (line 188) | def test_inference_batch_consistent(self): method test_inference_batch_single_identical (line 193) | def test_inference_batch_single_identical(self): method test_num_images_per_prompt (line 197) | def test_num_images_per_prompt(self): method test_progress_bar (line 200) | def test_progress_bar(self): class VideoToVideoSDPipelineSlowTests (line 206) | class VideoToVideoSDPipelineSlowTests(unittest.TestCase): method test_two_step_model (line 207) | def test_two_step_model(self): FILE: diffusers/tests/pipelines/unclip/test_unclip.py class UnCLIPPipelineFastTests (line 41) | class UnCLIPPipelineFastTests(PipelineTesterMixin, unittest.TestCase): method text_embedder_hidden_size (line 63) | def text_embedder_hidden_size(self): method time_input_dim (line 67) | def time_input_dim(self): method block_out_channels_0 (line 71) | def block_out_channels_0(self): method time_embed_dim (line 75) | def time_embed_dim(self): method cross_attention_dim (line 79) | def cross_attention_dim(self): method dummy_tokenizer (line 83) | def dummy_tokenizer(self): method dummy_text_encoder (line 88) | def dummy_text_encoder(self): method dummy_prior (line 105) | def dummy_prior(self): method dummy_text_proj (line 119) | def dummy_text_proj(self): method dummy_decoder (line 132) | def dummy_decoder(self): method dummy_super_res_kwargs (line 156) | def dummy_super_res_kwargs(self): method dummy_super_res_first (line 168) | def dummy_super_res_first(self): method dummy_super_res_last (line 175) | def dummy_super_res_last(self): method get_dummy_components (line 182) | def get_dummy_components(self): method get_dummy_inputs (line 225) | def get_dummy_inputs(self, device, seed=0): method test_unclip (line 240) | def test_unclip(self): method test_unclip_passed_text_embed (line 280) | def test_unclip_passed_text_embed(self): method test_attention_slicing_forward_pass (line 367) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 375) | def test_inference_batch_single_identical(self): method test_inference_batch_consistent (line 386) | def test_inference_batch_consistent(self): method test_dict_tuple_outputs_equivalent (line 406) | def test_dict_tuple_outputs_equivalent(self): method test_save_load_local (line 410) | def test_save_load_local(self): method test_save_load_optional_components (line 414) | def test_save_load_optional_components(self): method test_float16_inference (line 418) | def test_float16_inference(self): class UnCLIPPipelineCPUIntegrationTests (line 423) | class UnCLIPPipelineCPUIntegrationTests(unittest.TestCase): method setUp (line 424) | def setUp(self): method tearDown (line 430) | def tearDown(self): method test_unclip_karlo_cpu_fp32 (line 436) | def test_unclip_karlo_cpu_fp32(self): class UnCLIPPipelineIntegrationTests (line 461) | class UnCLIPPipelineIntegrationTests(unittest.TestCase): method setUp (line 462) | def setUp(self): method tearDown (line 468) | def tearDown(self): method test_unclip_karlo (line 474) | def test_unclip_karlo(self): method test_unclip_pipeline_with_sequential_cpu_offloading (line 497) | def test_unclip_pipeline_with_sequential_cpu_offloading(self): FILE: diffusers/tests/pipelines/unclip/test_unclip_image_variation.py class UnCLIPImageVariationPipelineFastTests (line 57) | class UnCLIPImageVariationPipelineFastTests(PipelineTesterMixin, unittes... method text_embedder_hidden_size (line 71) | def text_embedder_hidden_size(self): method time_input_dim (line 75) | def time_input_dim(self): method block_out_channels_0 (line 79) | def block_out_channels_0(self): method time_embed_dim (line 83) | def time_embed_dim(self): method cross_attention_dim (line 87) | def cross_attention_dim(self): method dummy_tokenizer (line 91) | def dummy_tokenizer(self): method dummy_text_encoder (line 96) | def dummy_text_encoder(self): method dummy_image_encoder (line 113) | def dummy_image_encoder(self): method dummy_text_proj (line 127) | def dummy_text_proj(self): method dummy_decoder (line 140) | def dummy_decoder(self): method dummy_super_res_kwargs (line 164) | def dummy_super_res_kwargs(self): method dummy_super_res_first (line 176) | def dummy_super_res_first(self): method dummy_super_res_last (line 183) | def dummy_super_res_last(self): method get_dummy_components (line 190) | def get_dummy_components(self): method get_dummy_inputs (line 227) | def get_dummy_inputs(self, device, seed=0, pil_image=True): method test_unclip_image_variation_input_tensor (line 248) | def test_unclip_image_variation_input_tensor(self): method test_unclip_image_variation_input_image (line 292) | def test_unclip_image_variation_input_image(self): method test_unclip_image_variation_input_list_images (line 324) | def test_unclip_image_variation_input_list_images(self): method test_unclip_passed_image_embed (line 376) | def test_unclip_passed_image_embed(self): method test_attention_slicing_forward_pass (line 437) | def test_attention_slicing_forward_pass(self): method test_inference_batch_single_identical (line 451) | def test_inference_batch_single_identical(self): method test_inference_batch_consistent (line 460) | def test_inference_batch_consistent(self): method test_dict_tuple_outputs_equivalent (line 479) | def test_dict_tuple_outputs_equivalent(self): method test_save_load_local (line 484) | def test_save_load_local(self): method test_save_load_optional_components (line 488) | def test_save_load_optional_components(self): method test_float16_inference (line 492) | def test_float16_inference(self): class UnCLIPImageVariationPipelineIntegrationTests (line 498) | class UnCLIPImageVariationPipelineIntegrationTests(unittest.TestCase): method setUp (line 499) | def setUp(self): method tearDown (line 505) | def tearDown(self): method test_unclip_image_variation_karlo (line 511) | def test_unclip_image_variation_karlo(self): FILE: diffusers/tests/pipelines/unidiffuser/test_unidiffuser.py function _test_unidiffuser_compile (line 48) | def _test_unidiffuser_compile(in_queue, out_queue, timeout): class UniDiffuserPipelineFastTests (line 79) | class UniDiffuserPipelineFastTests( method get_dummy_components (line 89) | def get_dummy_components(self): method get_dummy_inputs (line 147) | def get_dummy_inputs(self, device, seed=0): method get_fixed_latents (line 165) | def get_fixed_latents(self, device, seed=0): method get_dummy_inputs_with_latents (line 181) | def get_dummy_inputs_with_latents(self, device, seed=0): method test_dict_tuple_outputs_equivalent (line 209) | def test_dict_tuple_outputs_equivalent(self): method test_unidiffuser_default_joint_v0 (line 215) | def test_unidiffuser_default_joint_v0(self): method test_unidiffuser_default_joint_no_cfg_v0 (line 243) | def test_unidiffuser_default_joint_no_cfg_v0(self): method test_unidiffuser_default_text2img_v0 (line 273) | def test_unidiffuser_default_text2img_v0(self): method test_unidiffuser_default_image_0 (line 294) | def test_unidiffuser_default_image_0(self): method test_unidiffuser_default_text_v0 (line 316) | def test_unidiffuser_default_text_v0(self): method test_unidiffuser_default_img2text_v0 (line 336) | def test_unidiffuser_default_img2text_v0(self): method test_unidiffuser_default_joint_v1 (line 355) | def test_unidiffuser_default_joint_v1(self): method test_unidiffuser_default_text2img_v1 (line 383) | def test_unidiffuser_default_text2img_v1(self): method test_unidiffuser_default_img2text_v1 (line 403) | def test_unidiffuser_default_img2text_v1(self): method test_unidiffuser_text2img_multiple_images (line 421) | def test_unidiffuser_text2img_multiple_images(self): method test_unidiffuser_img2text_multiple_prompts (line 440) | def test_unidiffuser_img2text_multiple_prompts(self): method test_unidiffuser_text2img_multiple_images_with_latents (line 460) | def test_unidiffuser_text2img_multiple_images_with_latents(self): method test_unidiffuser_img2text_multiple_prompts_with_latents (line 479) | def test_unidiffuser_img2text_multiple_prompts_with_latents(self): method test_inference_batch_single_identical (line 499) | def test_inference_batch_single_identical(self): method test_unidiffuser_default_joint_v1_cuda_fp16 (line 503) | def test_unidiffuser_default_joint_v1_cuda_fp16(self): method test_unidiffuser_default_text2img_v1_cuda_fp16 (line 533) | def test_unidiffuser_default_text2img_v1_cuda_fp16(self): method test_unidiffuser_default_img2text_v1_cuda_fp16 (line 558) | def test_unidiffuser_default_img2text_v1_cuda_fp16(self): class UniDiffuserPipelineSlowTests (line 582) | class UniDiffuserPipelineSlowTests(unittest.TestCase): method setUp (line 583) | def setUp(self): method tearDown (line 588) | def tearDown(self): method get_inputs (line 593) | def get_inputs(self, device, seed=0, generate_latents=False): method get_fixed_latents (line 612) | def get_fixed_latents(self, device, seed=0): method test_unidiffuser_default_joint_v1 (line 634) | def test_unidiffuser_default_joint_v1(self): method test_unidiffuser_default_text2img_v1 (line 657) | def test_unidiffuser_default_text2img_v1(self): method test_unidiffuser_default_img2text_v1 (line 673) | def test_unidiffuser_default_img2text_v1(self): method test_unidiffuser_compile (line 689) | def test_unidiffuser_compile(self, seed=0): class UniDiffuserPipelineNightlyTests (line 703) | class UniDiffuserPipelineNightlyTests(unittest.TestCase): method setUp (line 704) | def setUp(self): method tearDown (line 709) | def tearDown(self): method get_inputs (line 714) | def get_inputs(self, device, seed=0, generate_latents=False): method get_fixed_latents (line 733) | def get_fixed_latents(self, device, seed=0): method test_unidiffuser_default_joint_v1_fp16 (line 755) | def test_unidiffuser_default_joint_v1_fp16(self): method test_unidiffuser_default_text2img_v1_fp16 (line 778) | def test_unidiffuser_default_text2img_v1_fp16(self): method test_unidiffuser_default_img2text_v1_fp16 (line 794) | def test_unidiffuser_default_img2text_v1_fp16(self): FILE: diffusers/tests/pipelines/wuerstchen/test_wuerstchen_combined.py class WuerstchenCombinedPipelineFastTests (line 32) | class WuerstchenCombinedPipelineFastTests(PipelineTesterMixin, unittest.... method text_embedder_hidden_size (line 52) | def text_embedder_hidden_size(self): method dummy_prior (line 56) | def dummy_prior(self): method dummy_tokenizer (line 64) | def dummy_tokenizer(self): method dummy_prior_text_encoder (line 69) | def dummy_prior_text_encoder(self): method dummy_text_encoder (line 85) | def dummy_text_encoder(self): method dummy_vqgan (line 102) | def dummy_vqgan(self): method dummy_decoder (line 113) | def dummy_decoder(self): method get_dummy_components (line 129) | def get_dummy_components(self): method get_dummy_inputs (line 154) | def get_dummy_inputs(self, device, seed=0): method test_wuerstchen (line 172) | def test_wuerstchen(self): method test_offloads (line 202) | def test_offloads(self): method test_inference_batch_single_identical (line 228) | def test_inference_batch_single_identical(self): method test_float16_inference (line 232) | def test_float16_inference(self): method test_callback_inputs (line 235) | def test_callback_inputs(self): method test_callback_cfg (line 238) | def test_callback_cfg(self): FILE: diffusers/tests/pipelines/wuerstchen/test_wuerstchen_decoder.py class WuerstchenDecoderPipelineFastTests (line 32) | class WuerstchenDecoderPipelineFastTests(PipelineTesterMixin, unittest.T... method text_embedder_hidden_size (line 49) | def text_embedder_hidden_size(self): method time_input_dim (line 53) | def time_input_dim(self): method block_out_channels_0 (line 57) | def block_out_channels_0(self): method time_embed_dim (line 61) | def time_embed_dim(self): method dummy_tokenizer (line 65) | def dummy_tokenizer(self): method dummy_text_encoder (line 70) | def dummy_text_encoder(self): method dummy_vqgan (line 87) | def dummy_vqgan(self): method dummy_decoder (line 98) | def dummy_decoder(self): method get_dummy_components (line 114) | def get_dummy_components(self): method get_dummy_inputs (line 133) | def get_dummy_inputs(self, device, seed=0): method test_wuerstchen_decoder (line 148) | def test_wuerstchen_decoder(self): method test_inference_batch_single_identical (line 173) | def test_inference_batch_single_identical(self): method test_attention_slicing_forward_pass (line 177) | def test_attention_slicing_forward_pass(self): method test_float16_inference (line 187) | def test_float16_inference(self): FILE: diffusers/tests/pipelines/wuerstchen/test_wuerstchen_prior.py function create_prior_lora_layers (line 45) | def create_prior_lora_layers(unet: nn.Module): class WuerstchenPriorPipelineFastTests (line 58) | class WuerstchenPriorPipelineFastTests(PipelineTesterMixin, unittest.Tes... method text_embedder_hidden_size (line 76) | def text_embedder_hidden_size(self): method time_input_dim (line 80) | def time_input_dim(self): method block_out_channels_0 (line 84) | def block_out_channels_0(self): method time_embed_dim (line 88) | def time_embed_dim(self): method dummy_tokenizer (line 92) | def dummy_tokenizer(self): method dummy_text_encoder (line 97) | def dummy_text_encoder(self): method dummy_prior (line 113) | def dummy_prior(self): method get_dummy_components (line 128) | def get_dummy_components(self): method get_dummy_inputs (line 144) | def get_dummy_inputs(self, device, seed=0): method test_wuerstchen_prior (line 158) | def test_wuerstchen_prior(self): method test_inference_batch_single_identical (line 195) | def test_inference_batch_single_identical(self): method test_attention_slicing_forward_pass (line 201) | def test_attention_slicing_forward_pass(self): method test_float16_inference (line 211) | def test_float16_inference(self): method test_callback_inputs (line 215) | def test_callback_inputs(self): method check_if_lora_correctly_set (line 249) | def check_if_lora_correctly_set(self, model) -> bool: method get_lora_components (line 258) | def get_lora_components(self): method test_inference_with_prior_lora (line 275) | def test_inference_with_prior_lora(self): FILE: diffusers/tests/schedulers/test_scheduler_consistency_model.py class CMStochasticIterativeSchedulerTest (line 8) | class CMStochasticIterativeSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 12) | def get_scheduler_config(self, **kwargs): method test_step_shape (line 26) | def test_step_shape(self): method test_timesteps (line 46) | def test_timesteps(self): method test_clip_denoised (line 50) | def test_clip_denoised(self): method test_full_loop_no_noise_onestep (line 54) | def test_full_loop_no_noise_onestep(self): method test_full_loop_no_noise_multistep (line 86) | def test_full_loop_no_noise_multistep(self): method test_full_loop_with_noise (line 118) | def test_full_loop_with_noise(self): method test_custom_timesteps_increasing_order (line 157) | def test_custom_timesteps_increasing_order(self): method test_custom_timesteps_passing_both_num_inference_steps_and_timesteps (line 167) | def test_custom_timesteps_passing_both_num_inference_steps_and_timeste... method test_custom_timesteps_too_large (line 178) | def test_custom_timesteps_too_large(self): FILE: diffusers/tests/schedulers/test_scheduler_ddim.py class DDIMSchedulerTest (line 8) | class DDIMSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 12) | def get_scheduler_config(self, **kwargs): method full_loop (line 24) | def full_loop(self, **config): method test_timesteps (line 42) | def test_timesteps(self): method test_steps_offset (line 46) | def test_steps_offset(self): method test_betas (line 56) | def test_betas(self): method test_schedules (line 60) | def test_schedules(self): method test_prediction_type (line 64) | def test_prediction_type(self): method test_clip_sample (line 68) | def test_clip_sample(self): method test_timestep_spacing (line 72) | def test_timestep_spacing(self): method test_rescale_betas_zero_snr (line 76) | def test_rescale_betas_zero_snr(self): method test_thresholding (line 80) | def test_thresholding(self): method test_time_indices (line 90) | def test_time_indices(self): method test_inference_steps (line 94) | def test_inference_steps(self): method test_eta (line 98) | def test_eta(self): method test_variance (line 102) | def test_variance(self): method test_full_loop_no_noise (line 114) | def test_full_loop_no_noise(self): method test_full_loop_with_v_prediction (line 123) | def test_full_loop_with_v_prediction(self): method test_full_loop_with_set_alpha_to_one (line 132) | def test_full_loop_with_set_alpha_to_one(self): method test_full_loop_with_no_set_alpha_to_one (line 141) | def test_full_loop_with_no_set_alpha_to_one(self): method test_full_loop_with_noise (line 150) | def test_full_loop_with_noise(self): FILE: diffusers/tests/schedulers/test_scheduler_ddim_inverse.py class DDIMInverseSchedulerTest (line 8) | class DDIMInverseSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 12) | def get_scheduler_config(self, **kwargs): method full_loop (line 24) | def full_loop(self, **config): method test_timesteps (line 42) | def test_timesteps(self): method test_steps_offset (line 46) | def test_steps_offset(self): method test_betas (line 56) | def test_betas(self): method test_schedules (line 60) | def test_schedules(self): method test_prediction_type (line 64) | def test_prediction_type(self): method test_clip_sample (line 68) | def test_clip_sample(self): method test_timestep_spacing (line 72) | def test_timestep_spacing(self): method test_rescale_betas_zero_snr (line 76) | def test_rescale_betas_zero_snr(self): method test_thresholding (line 80) | def test_thresholding(self): method test_time_indices (line 90) | def test_time_indices(self): method test_inference_steps (line 94) | def test_inference_steps(self): method test_add_noise_device (line 98) | def test_add_noise_device(self): method test_full_loop_no_noise (line 101) | def test_full_loop_no_noise(self): method test_full_loop_with_v_prediction (line 110) | def test_full_loop_with_v_prediction(self): method test_full_loop_with_set_alpha_to_one (line 119) | def test_full_loop_with_set_alpha_to_one(self): method test_full_loop_with_no_set_alpha_to_one (line 128) | def test_full_loop_with_no_set_alpha_to_one(self): FILE: diffusers/tests/schedulers/test_scheduler_ddim_parallel.py class DDIMParallelSchedulerTest (line 22) | class DDIMParallelSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 26) | def get_scheduler_config(self, **kwargs): method full_loop (line 38) | def full_loop(self, **config): method test_timesteps (line 56) | def test_timesteps(self): method test_steps_offset (line 60) | def test_steps_offset(self): method test_betas (line 70) | def test_betas(self): method test_schedules (line 74) | def test_schedules(self): method test_prediction_type (line 78) | def test_prediction_type(self): method test_clip_sample (line 82) | def test_clip_sample(self): method test_timestep_spacing (line 86) | def test_timestep_spacing(self): method test_rescale_betas_zero_snr (line 90) | def test_rescale_betas_zero_snr(self): method test_thresholding (line 94) | def test_thresholding(self): method test_time_indices (line 104) | def test_time_indices(self): method test_inference_steps (line 108) | def test_inference_steps(self): method test_eta (line 112) | def test_eta(self): method test_variance (line 116) | def test_variance(self): method test_batch_step_no_noise (line 128) | def test_batch_step_no_noise(self): method test_full_loop_no_noise (line 154) | def test_full_loop_no_noise(self): method test_full_loop_with_v_prediction (line 163) | def test_full_loop_with_v_prediction(self): method test_full_loop_with_set_alpha_to_one (line 172) | def test_full_loop_with_set_alpha_to_one(self): method test_full_loop_with_no_set_alpha_to_one (line 181) | def test_full_loop_with_no_set_alpha_to_one(self): method test_full_loop_with_noise (line 190) | def test_full_loop_with_noise(self): FILE: diffusers/tests/schedulers/test_scheduler_ddpm.py class DDPMSchedulerTest (line 8) | class DDPMSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 11) | def get_scheduler_config(self, **kwargs): method test_timesteps (line 24) | def test_timesteps(self): method test_betas (line 28) | def test_betas(self): method test_schedules (line 32) | def test_schedules(self): method test_variance_type (line 36) | def test_variance_type(self): method test_clip_sample (line 40) | def test_clip_sample(self): method test_thresholding (line 44) | def test_thresholding(self): method test_prediction_type (line 54) | def test_prediction_type(self): method test_time_indices (line 58) | def test_time_indices(self): method test_variance (line 62) | def test_variance(self): method test_rescale_betas_zero_snr (line 71) | def test_rescale_betas_zero_snr(self): method test_full_loop_no_noise (line 75) | def test_full_loop_no_noise(self): method test_full_loop_with_v_prediction (line 106) | def test_full_loop_with_v_prediction(self): method test_custom_timesteps (line 137) | def test_custom_timesteps(self): method test_custom_timesteps_increasing_order (line 159) | def test_custom_timesteps_increasing_order(self): method test_custom_timesteps_passing_both_num_inference_steps_and_timesteps (line 169) | def test_custom_timesteps_passing_both_num_inference_steps_and_timeste... method test_custom_timesteps_too_large (line 180) | def test_custom_timesteps_too_large(self): method test_full_loop_with_noise (line 193) | def test_full_loop_with_noise(self): FILE: diffusers/tests/schedulers/test_scheduler_ddpm_parallel.py class DDPMParallelSchedulerTest (line 22) | class DDPMParallelSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 25) | def get_scheduler_config(self, **kwargs): method test_timesteps (line 38) | def test_timesteps(self): method test_betas (line 42) | def test_betas(self): method test_schedules (line 46) | def test_schedules(self): method test_variance_type (line 50) | def test_variance_type(self): method test_clip_sample (line 54) | def test_clip_sample(self): method test_thresholding (line 58) | def test_thresholding(self): method test_prediction_type (line 68) | def test_prediction_type(self): method test_time_indices (line 72) | def test_time_indices(self): method test_variance (line 76) | def test_variance(self): method test_rescale_betas_zero_snr (line 85) | def test_rescale_betas_zero_snr(self): method test_batch_step_no_noise (line 89) | def test_batch_step_no_noise(self): method test_full_loop_no_noise (line 114) | def test_full_loop_no_noise(self): method test_full_loop_with_v_prediction (line 140) | def test_full_loop_with_v_prediction(self): method test_custom_timesteps (line 166) | def test_custom_timesteps(self): method test_custom_timesteps_increasing_order (line 188) | def test_custom_timesteps_increasing_order(self): method test_custom_timesteps_passing_both_num_inference_steps_and_timesteps (line 198) | def test_custom_timesteps_passing_both_num_inference_steps_and_timeste... method test_custom_timesteps_too_large (line 209) | def test_custom_timesteps_too_large(self): method test_full_loop_with_noise (line 222) | def test_full_loop_with_noise(self): FILE: diffusers/tests/schedulers/test_scheduler_deis.py class DEISMultistepSchedulerTest (line 15) | class DEISMultistepSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 19) | def get_scheduler_config(self, **kwargs): method check_over_configs (line 31) | def check_over_configs(self, time_step=0, **config): method test_from_save_pretrained (line 60) | def test_from_save_pretrained(self): method check_over_forward (line 63) | def check_over_forward(self, time_step=0, **forward_kwargs): method full_loop (line 92) | def full_loop(self, scheduler=None, **config): method test_step_shape (line 113) | def test_step_shape(self): method test_switch (line 143) | def test_switch(self): method test_timesteps (line 162) | def test_timesteps(self): method test_thresholding (line 166) | def test_thresholding(self): method test_prediction_type (line 181) | def test_prediction_type(self): method test_solver_order_and_type (line 185) | def test_solver_order_and_type(self): method test_lower_order_final (line 204) | def test_lower_order_final(self): method test_inference_steps (line 208) | def test_inference_steps(self): method test_full_loop_no_noise (line 212) | def test_full_loop_no_noise(self): method test_full_loop_with_v_prediction (line 218) | def test_full_loop_with_v_prediction(self): method test_fp16_support (line 224) | def test_fp16_support(self): method test_full_loop_with_noise (line 240) | def test_full_loop_with_noise(self): FILE: diffusers/tests/schedulers/test_scheduler_dpm_multi.py class DPMSolverMultistepSchedulerTest (line 15) | class DPMSolverMultistepSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 19) | def get_scheduler_config(self, **kwargs): method check_over_configs (line 41) | def check_over_configs(self, time_step=0, **config): method test_from_save_pretrained (line 70) | def test_from_save_pretrained(self): method check_over_forward (line 73) | def check_over_forward(self, time_step=0, **forward_kwargs): method full_loop (line 103) | def full_loop(self, scheduler=None, **config): method full_loop_custom_timesteps (line 122) | def full_loop_custom_timesteps(self, **config): method test_step_shape (line 144) | def test_step_shape(self): method test_timesteps (line 174) | def test_timesteps(self): method test_thresholding (line 178) | def test_thresholding(self): method test_prediction_type (line 193) | def test_prediction_type(self): method test_solver_order_and_type (line 197) | def test_solver_order_and_type(self): method test_lower_order_final (line 220) | def test_lower_order_final(self): method test_euler_at_final (line 224) | def test_euler_at_final(self): method test_lambda_min_clipped (line 228) | def test_lambda_min_clipped(self): method test_variance_type (line 232) | def test_variance_type(self): method test_inference_steps (line 236) | def test_inference_steps(self): method test_rescale_betas_zero_snr (line 240) | def test_rescale_betas_zero_snr(self): method test_full_loop_no_noise (line 244) | def test_full_loop_no_noise(self): method test_full_loop_with_noise (line 250) | def test_full_loop_with_noise(self): method test_full_loop_no_noise_thres (line 277) | def test_full_loop_no_noise_thres(self): method test_full_loop_with_v_prediction (line 283) | def test_full_loop_with_v_prediction(self): method test_full_loop_with_karras_and_v_prediction (line 289) | def test_full_loop_with_karras_and_v_prediction(self): method test_full_loop_with_lu_and_v_prediction (line 295) | def test_full_loop_with_lu_and_v_prediction(self): method test_switch (line 301) | def test_switch(self): method test_fp16_support (line 320) | def test_fp16_support(self): method test_duplicated_timesteps (line 336) | def test_duplicated_timesteps(self): method test_custom_timesteps (line 344) | def test_custom_timesteps(self): FILE: diffusers/tests/schedulers/test_scheduler_dpm_multi_inverse.py class DPMSolverMultistepSchedulerTest (line 10) | class DPMSolverMultistepSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 14) | def get_scheduler_config(self, **kwargs): method check_over_configs (line 34) | def check_over_configs(self, time_step=0, **config): method test_from_save_pretrained (line 63) | def test_from_save_pretrained(self): method check_over_forward (line 66) | def check_over_forward(self, time_step=0, **forward_kwargs): method full_loop (line 95) | def full_loop(self, scheduler=None, **config): method test_step_shape (line 112) | def test_step_shape(self): method test_timesteps (line 142) | def test_timesteps(self): method test_thresholding (line 146) | def test_thresholding(self): method test_prediction_type (line 161) | def test_prediction_type(self): method test_solver_order_and_type (line 165) | def test_solver_order_and_type(self): method test_lower_order_final (line 184) | def test_lower_order_final(self): method test_lambda_min_clipped (line 188) | def test_lambda_min_clipped(self): method test_variance_type (line 192) | def test_variance_type(self): method test_timestep_spacing (line 196) | def test_timestep_spacing(self): method test_inference_steps (line 200) | def test_inference_steps(self): method test_full_loop_no_noise (line 204) | def test_full_loop_no_noise(self): method test_full_loop_no_noise_thres (line 210) | def test_full_loop_no_noise_thres(self): method test_full_loop_with_v_prediction (line 216) | def test_full_loop_with_v_prediction(self): method test_full_loop_with_karras_and_v_prediction (line 222) | def test_full_loop_with_karras_and_v_prediction(self): method test_switch (line 228) | def test_switch(self): method test_fp16_support (line 245) | def test_fp16_support(self): method test_unique_timesteps (line 261) | def test_unique_timesteps(self, **config): FILE: diffusers/tests/schedulers/test_scheduler_dpm_sde.py class DPMSolverSDESchedulerTest (line 10) | class DPMSolverSDESchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 14) | def get_scheduler_config(self, **kwargs): method test_timesteps (line 26) | def test_timesteps(self): method test_betas (line 30) | def test_betas(self): method test_schedules (line 34) | def test_schedules(self): method test_prediction_type (line 38) | def test_prediction_type(self): method test_full_loop_no_noise (line 42) | def test_full_loop_no_noise(self): method test_full_loop_with_v_prediction (line 74) | def test_full_loop_with_v_prediction(self): method test_full_loop_device (line 106) | def test_full_loop_device(self): method test_full_loop_device_karras_sigmas (line 137) | def test_full_loop_device_karras_sigmas(self): FILE: diffusers/tests/schedulers/test_scheduler_dpm_single.py class DPMSolverSinglestepSchedulerTest (line 15) | class DPMSolverSinglestepSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 19) | def get_scheduler_config(self, **kwargs): method check_over_configs (line 39) | def check_over_configs(self, time_step=0, **config): method test_from_save_pretrained (line 68) | def test_from_save_pretrained(self): method check_over_forward (line 71) | def check_over_forward(self, time_step=0, **forward_kwargs): method full_loop (line 100) | def full_loop(self, scheduler=None, **config): method full_loop_custom_timesteps (line 117) | def full_loop_custom_timesteps(self, **config): method test_full_uneven_loop (line 138) | def test_full_uneven_loop(self): method test_timesteps (line 154) | def test_timesteps(self): method test_switch (line 158) | def test_switch(self): method test_thresholding (line 177) | def test_thresholding(self): method test_prediction_type (line 192) | def test_prediction_type(self): method test_solver_order_and_type (line 196) | def test_solver_order_and_type(self): method test_lower_order_final (line 215) | def test_lower_order_final(self): method test_lambda_min_clipped (line 219) | def test_lambda_min_clipped(self): method test_variance_type (line 223) | def test_variance_type(self): method test_inference_steps (line 227) | def test_inference_steps(self): method test_full_loop_no_noise (line 231) | def test_full_loop_no_noise(self): method test_full_loop_with_karras (line 237) | def test_full_loop_with_karras(self): method test_full_loop_with_v_prediction (line 243) | def test_full_loop_with_v_prediction(self): method test_full_loop_with_karras_and_v_prediction (line 249) | def test_full_loop_with_karras_and_v_prediction(self): method test_fp16_support (line 255) | def test_fp16_support(self): method test_step_shape (line 271) | def test_step_shape(self): method test_full_loop_with_noise (line 301) | def test_full_loop_with_noise(self): method test_custom_timesteps (line 328) | def test_custom_timesteps(self): FILE: diffusers/tests/schedulers/test_scheduler_edm_dpmsolver_multistep.py class EDMDPMSolverMultistepSchedulerTest (line 13) | class EDMDPMSolverMultistepSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 17) | def get_scheduler_config(self, **kwargs): method check_over_configs (line 37) | def check_over_configs(self, time_step=0, **config): method test_from_save_pretrained (line 66) | def test_from_save_pretrained(self): method check_over_forward (line 69) | def check_over_forward(self, time_step=0, **forward_kwargs): method full_loop (line 99) | def full_loop(self, scheduler=None, **config): method test_step_shape (line 116) | def test_step_shape(self): method test_timesteps (line 146) | def test_timesteps(self): method test_thresholding (line 150) | def test_thresholding(self): method test_prediction_type (line 165) | def test_prediction_type(self): method test_solver_order_and_type (line 171) | def test_solver_order_and_type(self): method test_lower_order_final (line 196) | def test_lower_order_final(self): method test_euler_at_final (line 200) | def test_euler_at_final(self): method test_inference_steps (line 204) | def test_inference_steps(self): method test_full_loop_no_noise (line 208) | def test_full_loop_no_noise(self): method test_full_loop_with_noise (line 214) | def test_full_loop_with_noise(self): method test_full_loop_no_noise_thres (line 241) | def test_full_loop_no_noise_thres(self): method test_full_loop_with_v_prediction (line 247) | def test_full_loop_with_v_prediction(self): method test_duplicated_timesteps (line 253) | def test_duplicated_timesteps(self, **config): method test_trained_betas (line 261) | def test_trained_betas(self): FILE: diffusers/tests/schedulers/test_scheduler_edm_euler.py class EDMEulerSchedulerTest (line 13) | class EDMEulerSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 17) | def get_scheduler_config(self, **kwargs): method test_timesteps (line 27) | def test_timesteps(self): method test_prediction_type (line 31) | def test_prediction_type(self): method test_full_loop_no_noise (line 35) | def test_full_loop_no_noise(self, num_inference_steps=10, seed=0): method test_full_loop_device (line 59) | def test_full_loop_device(self, num_inference_steps=10, seed=0): method test_from_save_pretrained (line 84) | def test_from_save_pretrained(self): method test_step_shape (line 122) | def test_step_shape(self): method test_scheduler_outputs_equivalence (line 144) | def test_scheduler_outputs_equivalence(self): method test_trained_betas (line 205) | def test_trained_betas(self): FILE: diffusers/tests/schedulers/test_scheduler_euler.py class EulerDiscreteSchedulerTest (line 9) | class EulerDiscreteSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 13) | def get_scheduler_config(self, **kwargs): method test_timesteps (line 24) | def test_timesteps(self): method test_betas (line 28) | def test_betas(self): method test_schedules (line 32) | def test_schedules(self): method test_prediction_type (line 36) | def test_prediction_type(self): method test_timestep_type (line 40) | def test_timestep_type(self): method test_karras_sigmas (line 45) | def test_karras_sigmas(self): method test_rescale_betas_zero_snr (line 48) | def test_rescale_betas_zero_snr(self): method full_loop (line 52) | def full_loop(self, **config): method full_loop_custom_timesteps (line 75) | def full_loop_custom_timesteps(self, **config): method full_loop_custom_sigmas (line 102) | def full_loop_custom_sigmas(self, **config): method test_full_loop_no_noise (line 129) | def test_full_loop_no_noise(self): method test_full_loop_with_v_prediction (line 138) | def test_full_loop_with_v_prediction(self): method test_full_loop_device (line 147) | def test_full_loop_device(self): method test_full_loop_device_karras_sigmas (line 174) | def test_full_loop_device_karras_sigmas(self): method test_full_loop_with_noise (line 201) | def test_full_loop_with_noise(self): method test_custom_timesteps (line 234) | def test_custom_timesteps(self): method test_custom_sigmas (line 252) | def test_custom_sigmas(self): FILE: diffusers/tests/schedulers/test_scheduler_euler_ancestral.py class EulerAncestralDiscreteSchedulerTest (line 9) | class EulerAncestralDiscreteSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 13) | def get_scheduler_config(self, **kwargs): method test_timesteps (line 24) | def test_timesteps(self): method test_betas (line 28) | def test_betas(self): method test_schedules (line 32) | def test_schedules(self): method test_prediction_type (line 36) | def test_prediction_type(self): method test_rescale_betas_zero_snr (line 40) | def test_rescale_betas_zero_snr(self): method test_full_loop_no_noise (line 44) | def test_full_loop_no_noise(self): method test_full_loop_with_v_prediction (line 71) | def test_full_loop_with_v_prediction(self): method test_full_loop_device (line 98) | def test_full_loop_device(self): method test_full_loop_with_noise (line 124) | def test_full_loop_with_noise(self): FILE: diffusers/tests/schedulers/test_scheduler_flax.py class FlaxSchedulerCommonTest (line 34) | class FlaxSchedulerCommonTest(unittest.TestCase): method dummy_sample (line 39) | def dummy_sample(self): method dummy_sample_deter (line 51) | def dummy_sample_deter(self): method get_scheduler_config (line 63) | def get_scheduler_config(self): method dummy_model (line 66) | def dummy_model(self): method check_over_configs (line 72) | def check_over_configs(self, time_step=0, **config): method check_over_forward (line 100) | def check_over_forward(self, time_step=0, **forward_kwargs): method test_from_save_pretrained (line 129) | def test_from_save_pretrained(self): method test_step_shape (line 157) | def test_step_shape(self): method test_scheduler_outputs_equivalence (line 181) | def test_scheduler_outputs_equivalence(self): method test_deprecated_kwargs (line 232) | def test_deprecated_kwargs(self): class FlaxDDPMSchedulerTest (line 255) | class FlaxDDPMSchedulerTest(FlaxSchedulerCommonTest): method get_scheduler_config (line 258) | def get_scheduler_config(self, **kwargs): method test_timesteps (line 271) | def test_timesteps(self): method test_betas (line 275) | def test_betas(self): method test_schedules (line 279) | def test_schedules(self): method test_variance_type (line 283) | def test_variance_type(self): method test_clip_sample (line 287) | def test_clip_sample(self): method test_time_indices (line 291) | def test_time_indices(self): method test_variance (line 295) | def test_variance(self): method test_full_loop_no_noise (line 305) | def test_full_loop_no_noise(self): class FlaxDDIMSchedulerTest (line 346) | class FlaxDDIMSchedulerTest(FlaxSchedulerCommonTest): method get_scheduler_config (line 350) | def get_scheduler_config(self, **kwargs): method full_loop (line 361) | def full_loop(self, **config): method check_over_configs (line 384) | def check_over_configs(self, time_step=0, **config): method test_from_save_pretrained (line 412) | def test_from_save_pretrained(self): method check_over_forward (line 440) | def check_over_forward(self, time_step=0, **forward_kwargs): method test_scheduler_outputs_equivalence (line 469) | def test_scheduler_outputs_equivalence(self): method test_step_shape (line 520) | def test_step_shape(self): method test_timesteps (line 544) | def test_timesteps(self): method test_steps_offset (line 548) | def test_steps_offset(self): method test_betas (line 559) | def test_betas(self): method test_schedules (line 563) | def test_schedules(self): method test_time_indices (line 567) | def test_time_indices(self): method test_inference_steps (line 571) | def test_inference_steps(self): method test_variance (line 575) | def test_variance(self): method test_full_loop_no_noise (line 588) | def test_full_loop_no_noise(self): method test_full_loop_with_set_alpha_to_one (line 597) | def test_full_loop_with_set_alpha_to_one(self): method test_full_loop_with_no_set_alpha_to_one (line 610) | def test_full_loop_with_no_set_alpha_to_one(self): method test_prediction_type (line 625) | def test_prediction_type(self): class FlaxPNDMSchedulerTest (line 631) | class FlaxPNDMSchedulerTest(FlaxSchedulerCommonTest): method get_scheduler_config (line 635) | def get_scheduler_config(self, **kwargs): method check_over_configs (line 646) | def check_over_configs(self, time_step=0, **config): method test_from_save_pretrained (line 678) | def test_from_save_pretrained(self): method test_scheduler_outputs_equivalence (line 681) | def test_scheduler_outputs_equivalence(self): method check_over_forward (line 732) | def check_over_forward(self, time_step=0, **forward_kwargs): method full_loop (line 767) | def full_loop(self, **config): method test_step_shape (line 788) | def test_step_shape(self): method test_timesteps (line 822) | def test_timesteps(self): method test_steps_offset (line 826) | def test_steps_offset(self): method test_betas (line 840) | def test_betas(self): method test_schedules (line 844) | def test_schedules(self): method test_time_indices (line 848) | def test_time_indices(self): method test_inference_steps (line 852) | def test_inference_steps(self): method test_pow_of_3_inference_steps (line 856) | def test_pow_of_3_inference_steps(self): method test_inference_plms_no_past_residuals (line 874) | def test_inference_plms_no_past_residuals(self): method test_full_loop_no_noise (line 883) | def test_full_loop_no_noise(self): method test_full_loop_with_set_alpha_to_one (line 895) | def test_full_loop_with_set_alpha_to_one(self): method test_full_loop_with_no_set_alpha_to_one (line 908) | def test_full_loop_with_no_set_alpha_to_one(self): FILE: diffusers/tests/schedulers/test_scheduler_heun.py class HeunDiscreteSchedulerTest (line 9) | class HeunDiscreteSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 13) | def get_scheduler_config(self, **kwargs): method test_timesteps (line 24) | def test_timesteps(self): method test_betas (line 28) | def test_betas(self): method test_schedules (line 32) | def test_schedules(self): method test_clip_sample (line 36) | def test_clip_sample(self): method test_prediction_type (line 40) | def test_prediction_type(self): method full_loop (line 44) | def full_loop(self, **config): method full_loop_custom_timesteps (line 66) | def full_loop_custom_timesteps(self, **config): method test_full_loop_no_noise (line 93) | def test_full_loop_no_noise(self): method test_full_loop_with_v_prediction (line 106) | def test_full_loop_with_v_prediction(self): method test_full_loop_device (line 119) | def test_full_loop_device(self): method test_full_loop_device_karras_sigmas (line 152) | def test_full_loop_device_karras_sigmas(self): method test_full_loop_with_noise (line 177) | def test_full_loop_with_noise(self): method test_custom_timesteps (line 208) | def test_custom_timesteps(self): FILE: diffusers/tests/schedulers/test_scheduler_ipndm.py class IPNDMSchedulerTest (line 10) | class IPNDMSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 14) | def get_scheduler_config(self, **kwargs): method check_over_configs (line 19) | def check_over_configs(self, time_step=0, **config): method test_from_save_pretrained (line 53) | def test_from_save_pretrained(self): method check_over_forward (line 56) | def check_over_forward(self, time_step=0, **forward_kwargs): method full_loop (line 93) | def full_loop(self, **config): method test_step_shape (line 115) | def test_step_shape(self): method test_timesteps (line 151) | def test_timesteps(self): method test_inference_steps (line 155) | def test_inference_steps(self): method test_full_loop_no_noise (line 159) | def test_full_loop_no_noise(self): FILE: diffusers/tests/schedulers/test_scheduler_kdpm2_ancestral.py class KDPM2AncestralDiscreteSchedulerTest (line 9) | class KDPM2AncestralDiscreteSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 13) | def get_scheduler_config(self, **kwargs): method test_timesteps (line 24) | def test_timesteps(self): method test_betas (line 28) | def test_betas(self): method test_schedules (line 32) | def test_schedules(self): method test_full_loop_no_noise (line 36) | def test_full_loop_no_noise(self): method test_prediction_type (line 65) | def test_prediction_type(self): method test_full_loop_with_v_prediction (line 69) | def test_full_loop_with_v_prediction(self): method test_full_loop_device (line 98) | def test_full_loop_device(self): method test_full_loop_with_noise (line 125) | def test_full_loop_with_noise(self): FILE: diffusers/tests/schedulers/test_scheduler_kdpm2_discrete.py class KDPM2DiscreteSchedulerTest (line 9) | class KDPM2DiscreteSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 13) | def get_scheduler_config(self, **kwargs): method test_timesteps (line 24) | def test_timesteps(self): method test_betas (line 28) | def test_betas(self): method test_schedules (line 32) | def test_schedules(self): method test_prediction_type (line 36) | def test_prediction_type(self): method test_full_loop_with_v_prediction (line 40) | def test_full_loop_with_v_prediction(self): method test_full_loop_no_noise (line 70) | def test_full_loop_no_noise(self): method test_full_loop_device (line 102) | def test_full_loop_device(self): method test_full_loop_with_noise (line 134) | def test_full_loop_with_noise(self): FILE: diffusers/tests/schedulers/test_scheduler_lcm.py class LCMSchedulerTest (line 12) | class LCMSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 16) | def get_scheduler_config(self, **kwargs): method default_valid_timestep (line 29) | def default_valid_timestep(self): method test_timesteps (line 40) | def test_timesteps(self): method test_betas (line 45) | def test_betas(self): method test_schedules (line 49) | def test_schedules(self): method test_prediction_type (line 53) | def test_prediction_type(self): method test_clip_sample (line 57) | def test_clip_sample(self): method test_thresholding (line 61) | def test_thresholding(self): method test_time_indices (line 72) | def test_time_indices(self): method test_inference_steps (line 85) | def test_inference_steps(self): method test_add_noise_device (line 92) | def test_add_noise_device(self, num_inference_steps=10): method test_from_save_pretrained (line 108) | def test_from_save_pretrained(self): method test_step_shape (line 137) | def test_step_shape(self): method test_scheduler_outputs_equivalence (line 160) | def test_scheduler_outputs_equivalence(self): method full_loop (line 209) | def full_loop(self, num_inference_steps=10, seed=0, **config): method test_full_loop_onestep (line 226) | def test_full_loop_onestep(self): method test_full_loop_multistep (line 236) | def test_full_loop_multistep(self): method test_custom_timesteps (line 246) | def test_custom_timesteps(self): method test_custom_timesteps_increasing_order (line 268) | def test_custom_timesteps_increasing_order(self): method test_custom_timesteps_passing_both_num_inference_steps_and_timesteps (line 278) | def test_custom_timesteps_passing_both_num_inference_steps_and_timeste... method test_custom_timesteps_too_large (line 289) | def test_custom_timesteps_too_large(self): FILE: diffusers/tests/schedulers/test_scheduler_lms.py class LMSDiscreteSchedulerTest (line 9) | class LMSDiscreteSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 13) | def get_scheduler_config(self, **kwargs): method test_timesteps (line 24) | def test_timesteps(self): method test_betas (line 28) | def test_betas(self): method test_schedules (line 32) | def test_schedules(self): method test_prediction_type (line 36) | def test_prediction_type(self): method test_time_indices (line 40) | def test_time_indices(self): method test_full_loop_no_noise (line 44) | def test_full_loop_no_noise(self): method test_full_loop_with_v_prediction (line 68) | def test_full_loop_with_v_prediction(self): method test_full_loop_device (line 92) | def test_full_loop_device(self): method test_full_loop_device_karras_sigmas (line 117) | def test_full_loop_device_karras_sigmas(self): method test_full_loop_with_noise (line 142) | def test_full_loop_with_noise(self): FILE: diffusers/tests/schedulers/test_scheduler_pndm.py class PNDMSchedulerTest (line 10) | class PNDMSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 14) | def get_scheduler_config(self, **kwargs): method check_over_configs (line 25) | def check_over_configs(self, time_step=0, **config): method test_from_save_pretrained (line 56) | def test_from_save_pretrained(self): method check_over_forward (line 59) | def check_over_forward(self, time_step=0, **forward_kwargs): method full_loop (line 93) | def full_loop(self, **config): method test_step_shape (line 113) | def test_step_shape(self): method test_timesteps (line 146) | def test_timesteps(self): method test_steps_offset (line 150) | def test_steps_offset(self): method test_betas (line 165) | def test_betas(self): method test_schedules (line 169) | def test_schedules(self): method test_prediction_type (line 173) | def test_prediction_type(self): method test_time_indices (line 177) | def test_time_indices(self): method test_inference_steps (line 181) | def test_inference_steps(self): method test_pow_of_3_inference_steps (line 185) | def test_pow_of_3_inference_steps(self): method test_inference_plms_no_past_residuals (line 202) | def test_inference_plms_no_past_residuals(self): method test_full_loop_no_noise (line 210) | def test_full_loop_no_noise(self): method test_full_loop_with_v_prediction (line 218) | def test_full_loop_with_v_prediction(self): method test_full_loop_with_set_alpha_to_one (line 226) | def test_full_loop_with_set_alpha_to_one(self): method test_full_loop_with_no_set_alpha_to_one (line 235) | def test_full_loop_with_no_set_alpha_to_one(self): FILE: diffusers/tests/schedulers/test_scheduler_sasolver.py class SASolverSchedulerTest (line 10) | class SASolverSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 15) | def get_scheduler_config(self, **kwargs): method test_step_shape (line 26) | def test_step_shape(self): method test_timesteps (line 61) | def test_timesteps(self): method test_betas (line 65) | def test_betas(self): method test_schedules (line 69) | def test_schedules(self): method test_prediction_type (line 73) | def test_prediction_type(self): method test_full_loop_no_noise (line 77) | def test_full_loop_no_noise(self): method test_full_loop_with_v_prediction (line 109) | def test_full_loop_with_v_prediction(self): method test_full_loop_device (line 141) | def test_full_loop_device(self): method test_full_loop_device_karras_sigmas (line 172) | def test_full_loop_device_karras_sigmas(self): FILE: diffusers/tests/schedulers/test_scheduler_score_sde_ve.py class ScoreSdeVeSchedulerTest (line 10) | class ScoreSdeVeSchedulerTest(unittest.TestCase): method dummy_sample (line 16) | def dummy_sample(self): method dummy_sample_deter (line 27) | def dummy_sample_deter(self): method dummy_model (line 41) | def dummy_model(self): method get_scheduler_config (line 47) | def get_scheduler_config(self, **kwargs): method check_over_configs (line 59) | def check_over_configs(self, time_step=0, **config): method check_over_forward (line 89) | def check_over_forward(self, time_step=0, **forward_kwargs): method test_timesteps (line 120) | def test_timesteps(self): method test_sigmas (line 124) | def test_sigmas(self): method test_time_indices (line 128) | def test_time_indices(self): method test_full_loop_no_noise (line 132) | def test_full_loop_no_noise(self): method test_step_shape (line 168) | def test_step_shape(self): FILE: diffusers/tests/schedulers/test_scheduler_tcd.py class TCDSchedulerTest (line 8) | class TCDSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 12) | def get_scheduler_config(self, **kwargs): method default_num_inference_steps (line 25) | def default_num_inference_steps(self): method default_valid_timestep (line 29) | def default_valid_timestep(self): method test_timesteps (line 40) | def test_timesteps(self): method test_betas (line 45) | def test_betas(self): method test_schedules (line 49) | def test_schedules(self): method test_prediction_type (line 53) | def test_prediction_type(self): method test_clip_sample (line 57) | def test_clip_sample(self): method test_thresholding (line 61) | def test_thresholding(self): method test_time_indices (line 72) | def test_time_indices(self): method test_inference_steps (line 85) | def test_inference_steps(self): method full_loop (line 90) | def full_loop(self, num_inference_steps=10, seed=0, **config): method test_full_loop_onestep_deter (line 108) | def test_full_loop_onestep_deter(self): method test_full_loop_multistep_deter (line 117) | def test_full_loop_multistep_deter(self): method test_custom_timesteps (line 126) | def test_custom_timesteps(self): method test_custom_timesteps_increasing_order (line 148) | def test_custom_timesteps_increasing_order(self): method test_custom_timesteps_passing_both_num_inference_steps_and_timesteps (line 158) | def test_custom_timesteps_passing_both_num_inference_steps_and_timeste... method test_custom_timesteps_too_large (line 169) | def test_custom_timesteps_too_large(self): FILE: diffusers/tests/schedulers/test_scheduler_unclip.py class UnCLIPSchedulerTest (line 9) | class UnCLIPSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 12) | def get_scheduler_config(self, **kwargs): method test_timesteps (line 24) | def test_timesteps(self): method test_variance_type (line 28) | def test_variance_type(self): method test_clip_sample (line 32) | def test_clip_sample(self): method test_clip_sample_range (line 36) | def test_clip_sample_range(self): method test_prediction_type (line 40) | def test_prediction_type(self): method test_time_indices (line 44) | def test_time_indices(self): method test_variance_fixed_small_log (line 52) | def test_variance_fixed_small_log(self): method test_variance_learned_range (line 61) | def test_variance_learned_range(self): method test_full_loop (line 72) | def test_full_loop(self): method test_full_loop_skip_timesteps (line 98) | def test_full_loop_skip_timesteps(self): method test_trained_betas (line 133) | def test_trained_betas(self): method test_add_noise_device (line 136) | def test_add_noise_device(self): FILE: diffusers/tests/schedulers/test_scheduler_unipc.py class UniPCMultistepSchedulerTest (line 15) | class UniPCMultistepSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 19) | def get_scheduler_config(self, **kwargs): method check_over_configs (line 33) | def check_over_configs(self, time_step=0, **config): method check_over_forward (line 62) | def check_over_forward(self, time_step=0, **forward_kwargs): method full_loop (line 91) | def full_loop(self, scheduler=None, **config): method test_step_shape (line 112) | def test_step_shape(self): method test_switch (line 142) | def test_switch(self): method test_timesteps (line 161) | def test_timesteps(self): method test_thresholding (line 165) | def test_thresholding(self): method test_prediction_type (line 179) | def test_prediction_type(self): method test_rescale_betas_zero_snr (line 183) | def test_rescale_betas_zero_snr(self): method test_solver_order_and_type (line 187) | def test_solver_order_and_type(self): method test_lower_order_final (line 203) | def test_lower_order_final(self): method test_inference_steps (line 207) | def test_inference_steps(self): method test_full_loop_no_noise (line 211) | def test_full_loop_no_noise(self): method test_full_loop_with_karras (line 217) | def test_full_loop_with_karras(self): method test_full_loop_with_v_prediction (line 223) | def test_full_loop_with_v_prediction(self): method test_full_loop_with_karras_and_v_prediction (line 229) | def test_full_loop_with_karras_and_v_prediction(self): method test_fp16_support (line 235) | def test_fp16_support(self): method test_full_loop_with_noise (line 260) | def test_full_loop_with_noise(self): class UniPCMultistepScheduler1DTest (line 288) | class UniPCMultistepScheduler1DTest(UniPCMultistepSchedulerTest): method dummy_sample (line 290) | def dummy_sample(self): method dummy_noise_deter (line 300) | def dummy_noise_deter(self): method dummy_sample_deter (line 314) | def dummy_sample_deter(self): method test_switch (line 327) | def test_switch(self): method test_full_loop_no_noise (line 346) | def test_full_loop_no_noise(self): method test_full_loop_with_karras (line 352) | def test_full_loop_with_karras(self): method test_full_loop_with_v_prediction (line 358) | def test_full_loop_with_v_prediction(self): method test_full_loop_with_karras_and_v_prediction (line 364) | def test_full_loop_with_karras_and_v_prediction(self): method test_full_loop_with_noise (line 370) | def test_full_loop_with_noise(self): FILE: diffusers/tests/schedulers/test_scheduler_vq_diffusion.py class VQDiffusionSchedulerTest (line 9) | class VQDiffusionSchedulerTest(SchedulerCommonTest): method get_scheduler_config (line 12) | def get_scheduler_config(self, **kwargs): method dummy_sample (line 21) | def dummy_sample(self, num_vec_classes): method dummy_sample_deter (line 31) | def dummy_sample_deter(self): method dummy_model (line 34) | def dummy_model(self, num_vec_classes): method test_timesteps (line 43) | def test_timesteps(self): method test_num_vec_classes (line 47) | def test_num_vec_classes(self): method test_time_indices (line 51) | def test_time_indices(self): method test_add_noise_device (line 55) | def test_add_noise_device(self): FILE: diffusers/tests/schedulers/test_schedulers.py class SchedulerObject (line 55) | class SchedulerObject(SchedulerMixin, ConfigMixin): method __init__ (line 59) | def __init__( class SchedulerObject2 (line 70) | class SchedulerObject2(SchedulerMixin, ConfigMixin): method __init__ (line 74) | def __init__( class SchedulerObject3 (line 85) | class SchedulerObject3(SchedulerMixin, ConfigMixin): method __init__ (line 89) | def __init__( class SchedulerBaseTests (line 101) | class SchedulerBaseTests(unittest.TestCase): method test_save_load_from_different_config (line 102) | def test_save_load_from_different_config(self): method test_save_load_compatible_schedulers (line 144) | def test_save_load_compatible_schedulers(self): method test_save_load_from_different_config_comp_schedulers (line 180) | def test_save_load_from_different_config_comp_schedulers(self): method test_default_arguments_not_in_config (line 217) | def test_default_arguments_not_in_config(self): method test_default_solver_type_after_switch (line 242) | def test_default_solver_type_after_switch(self): class SchedulerCommonTest (line 256) | class SchedulerCommonTest(unittest.TestCase): method default_num_inference_steps (line 261) | def default_num_inference_steps(self): method default_timestep (line 265) | def default_timestep(self): method default_timestep_2 (line 287) | def default_timestep_2(self): method dummy_sample (line 315) | def dummy_sample(self): method dummy_noise_deter (line 326) | def dummy_noise_deter(self): method dummy_sample_deter (line 341) | def dummy_sample_deter(self): method get_scheduler_config (line 355) | def get_scheduler_config(self): method dummy_model (line 358) | def dummy_model(self): method check_over_configs (line 370) | def check_over_configs(self, time_step=0, **config): method check_over_forward (line 431) | def check_over_forward(self, time_step=0, **forward_kwargs): method test_from_save_pretrained (line 474) | def test_from_save_pretrained(self): method test_compatibles (line 520) | def test_compatibles(self): method test_from_pretrained (line 544) | def test_from_pretrained(self): method test_step_shape (line 560) | def test_step_shape(self): method test_scheduler_outputs_equivalence (line 596) | def test_scheduler_outputs_equivalence(self): method test_scheduler_public_api (line 672) | def test_scheduler_public_api(self): method test_add_noise_device (line 706) | def test_add_noise_device(self): method test_deprecated_kwargs (line 730) | def test_deprecated_kwargs(self): method test_trained_betas (line 751) | def test_trained_betas(self): method test_getattr_is_correct (line 765) | def test_getattr_is_correct(self): class SchedulerPushToHubTester (line 811) | class SchedulerPushToHubTester(unittest.TestCase): method test_push_to_hub (line 816) | def test_push_to_hub(self): method test_push_to_hub_in_organization (line 843) | def test_push_to_hub_in_organization(self): FILE: diffusers/tests/single_file/single_file_testing_utils.py function download_single_file_checkpoint (line 15) | def download_single_file_checkpoint(repo_id, filename, tmpdir): function download_original_config (line 20) | def download_original_config(config_url, tmpdir): function download_diffusers_config (line 29) | def download_diffusers_config(repo_id, tmpdir): class SDSingleFileTesterMixin (line 48) | class SDSingleFileTesterMixin: method _compare_component_configs (line 49) | def _compare_component_configs(self, pipe, single_file_pipe): method test_single_file_components (line 89) | def test_single_file_components(self, pipe=None, single_file_pipe=None): method test_single_file_components_local_files_only (line 97) | def test_single_file_components_local_files_only(self, pipe=None, sing... method test_single_file_components_with_original_config (line 110) | def test_single_file_components_with_original_config( method test_single_file_components_with_original_config_local_files_only (line 129) | def test_single_file_components_with_original_config_local_files_only( method test_single_file_format_inference_is_same_as_pretrained (line 155) | def test_single_file_format_inference_is_same_as_pretrained(self, expe... method test_single_file_components_with_diffusers_config (line 174) | def test_single_file_components_with_diffusers_config( method test_single_file_components_with_diffusers_config_local_files_only (line 186) | def test_single_file_components_with_diffusers_config_local_files_only( class SDXLSingleFileTesterMixin (line 205) | class SDXLSingleFileTesterMixin: method _compare_component_configs (line 206) | def _compare_component_configs(self, pipe, single_file_pipe): method test_single_file_components (line 256) | def test_single_file_components(self, pipe=None, single_file_pipe=None): method test_single_file_components_local_files_only (line 267) | def test_single_file_components_local_files_only( method test_single_file_components_with_original_config (line 284) | def test_single_file_components_with_original_config( method test_single_file_components_with_original_config_local_files_only (line 305) | def test_single_file_components_with_original_config_local_files_only( method test_single_file_components_with_diffusers_config (line 333) | def test_single_file_components_with_diffusers_config( method test_single_file_components_with_diffusers_config_local_files_only (line 345) | def test_single_file_components_with_diffusers_config_local_files_only( method test_single_file_format_inference_is_same_as_pretrained (line 363) | def test_single_file_format_inference_is_same_as_pretrained(self, expe... FILE: diffusers/tests/single_file/test_model_controlnet_single_file.py class ControlNetModelSingleFileTests (line 36) | class ControlNetModelSingleFileTests(unittest.TestCase): method setUp (line 41) | def setUp(self): method tearDown (line 46) | def tearDown(self): method test_single_file_components (line 51) | def test_single_file_components(self): method test_single_file_arguments (line 63) | def test_single_file_arguments(self): FILE: diffusers/tests/single_file/test_model_sd_cascade_unet_single_file.py class StableCascadeUNetSingleFileTest (line 37) | class StableCascadeUNetSingleFileTest(unittest.TestCase): method setUp (line 38) | def setUp(self): method tearDown (line 43) | def tearDown(self): method test_single_file_components_stage_b (line 48) | def test_single_file_components_stage_b(self): method test_single_file_components_stage_b_lite (line 65) | def test_single_file_components_stage_b_lite(self): method test_single_file_components_stage_c (line 82) | def test_single_file_components_stage_c(self): method test_single_file_components_stage_c_lite (line 99) | def test_single_file_components_stage_c_lite(self): FILE: diffusers/tests/single_file/test_model_vae_single_file.py class AutoencoderKLSingleFileTests (line 39) | class AutoencoderKLSingleFileTests(unittest.TestCase): method setUp (line 48) | def setUp(self): method tearDown (line 53) | def tearDown(self): method get_file_format (line 58) | def get_file_format(self, seed, shape): method get_sd_image (line 61) | def get_sd_image(self, seed=0, shape=(4, 3, 512, 512), fp16=False): method test_single_file_inference_same_as_pretrained (line 66) | def test_single_file_inference_same_as_pretrained(self): method test_single_file_components (line 85) | def test_single_file_components(self): method test_single_file_arguments (line 97) | def test_single_file_arguments(self): FILE: diffusers/tests/single_file/test_stable_diffusion_controlnet_img2img_single_file.py class StableDiffusionControlNetPipelineSingleFileSlowTests (line 30) | class StableDiffusionControlNetPipelineSingleFileSlowTests(unittest.Test... method setUp (line 38) | def setUp(self): method tearDown (line 43) | def tearDown(self): method get_inputs (line 48) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_single_file_format_inference_is_same_as_pretrained (line 71) | def test_single_file_format_inference_is_same_as_pretrained(self): method test_single_file_components (line 93) | def test_single_file_components(self): method test_single_file_components_local_files_only (line 106) | def test_single_file_components_local_files_only(self): method test_single_file_components_with_original_config (line 120) | def test_single_file_components_with_original_config(self): method test_single_file_components_with_original_config_local_files_only (line 129) | def test_single_file_components_with_original_config_local_files_only(... method test_single_file_components_with_diffusers_config (line 152) | def test_single_file_components_with_diffusers_config(self): method test_single_file_components_with_diffusers_config_local_files_only (line 161) | def test_single_file_components_with_diffusers_config_local_files_only... FILE: diffusers/tests/single_file/test_stable_diffusion_controlnet_inpaint_single_file.py class StableDiffusionControlNetInpaintPipelineSingleFileSlowTests (line 29) | class StableDiffusionControlNetInpaintPipelineSingleFileSlowTests(unitte... method setUp (line 35) | def setUp(self): method tearDown (line 40) | def tearDown(self): method get_inputs (line 45) | def get_inputs(self): method test_single_file_format_inference_is_same_as_pretrained (line 69) | def test_single_file_format_inference_is_same_as_pretrained(self): method test_single_file_components (line 88) | def test_single_file_components(self): method test_single_file_components_local_files_only (line 101) | def test_single_file_components_local_files_only(self): method test_single_file_components_with_original_config (line 115) | def test_single_file_components_with_original_config(self): method test_single_file_components_with_original_config_local_files_only (line 124) | def test_single_file_components_with_original_config_local_files_only(... method test_single_file_components_with_diffusers_config (line 148) | def test_single_file_components_with_diffusers_config(self): method test_single_file_components_with_diffusers_config_local_files_only (line 159) | def test_single_file_components_with_diffusers_config_local_files_only... FILE: diffusers/tests/single_file/test_stable_diffusion_controlnet_single_file.py class StableDiffusionControlNetPipelineSingleFileSlowTests (line 29) | class StableDiffusionControlNetPipelineSingleFileSlowTests(unittest.Test... method setUp (line 37) | def setUp(self): method tearDown (line 42) | def tearDown(self): method get_inputs (line 47) | def get_inputs(self): method test_single_file_format_inference_is_same_as_pretrained (line 61) | def test_single_file_format_inference_is_same_as_pretrained(self): method test_single_file_components (line 83) | def test_single_file_components(self): method test_single_file_components_local_files_only (line 96) | def test_single_file_components_local_files_only(self): method test_single_file_components_with_original_config (line 110) | def test_single_file_components_with_original_config(self): method test_single_file_components_with_original_config_local_files_only (line 119) | def test_single_file_components_with_original_config_local_files_only(... method test_single_file_components_with_diffusers_config (line 140) | def test_single_file_components_with_diffusers_config(self): method test_single_file_components_with_diffusers_config_local_files_only (line 149) | def test_single_file_components_with_diffusers_config_local_files_only... FILE: diffusers/tests/single_file/test_stable_diffusion_img2img_single_file.py class StableDiffusionImg2ImgPipelineSingleFileSlowTests (line 24) | class StableDiffusionImg2ImgPipelineSingleFileSlowTests(unittest.TestCas... method setUp (line 32) | def setUp(self): method tearDown (line 37) | def tearDown(self): method get_inputs (line 42) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_single_file_format_inference_is_same_as_pretrained (line 59) | def test_single_file_format_inference_is_same_as_pretrained(self): class StableDiffusion21Img2ImgPipelineSingleFileSlowTests (line 65) | class StableDiffusion21Img2ImgPipelineSingleFileSlowTests(unittest.TestC... method setUp (line 71) | def setUp(self): method tearDown (line 76) | def tearDown(self): method get_inputs (line 81) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_single_file_format_inference_is_same_as_pretrained (line 98) | def test_single_file_format_inference_is_same_as_pretrained(self): FILE: diffusers/tests/single_file/test_stable_diffusion_inpaint_single_file.py class StableDiffusionInpaintPipelineSingleFileSlowTests (line 24) | class StableDiffusionInpaintPipelineSingleFileSlowTests(unittest.TestCas... method setUp (line 30) | def setUp(self): method tearDown (line 35) | def tearDown(self): method get_inputs (line 40) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_single_file_format_inference_is_same_as_pretrained (line 61) | def test_single_file_format_inference_is_same_as_pretrained(self): method test_single_file_loading_4_channel_unet (line 64) | def test_single_file_loading_4_channel_unet(self): class StableDiffusion21InpaintPipelineSingleFileSlowTests (line 74) | class StableDiffusion21InpaintPipelineSingleFileSlowTests(unittest.TestC... method setUp (line 82) | def setUp(self): method tearDown (line 87) | def tearDown(self): method get_inputs (line 92) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_single_file_format_inference_is_same_as_pretrained (line 113) | def test_single_file_format_inference_is_same_as_pretrained(self): FILE: diffusers/tests/single_file/test_stable_diffusion_single_file.py class StableDiffusionPipelineSingleFileSlowTests (line 26) | class StableDiffusionPipelineSingleFileSlowTests(unittest.TestCase, SDSi... method setUp (line 34) | def setUp(self): method tearDown (line 39) | def tearDown(self): method get_inputs (line 44) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_single_file_format_inference_is_same_as_pretrained (line 56) | def test_single_file_format_inference_is_same_as_pretrained(self): method test_single_file_legacy_scheduler_loading (line 59) | def test_single_file_legacy_scheduler_loading(self): method test_single_file_legacy_scaling_factor (line 76) | def test_single_file_legacy_scaling_factor(self): class StableDiffusion21PipelineSingleFileSlowTests (line 86) | class StableDiffusion21PipelineSingleFileSlowTests(unittest.TestCase, SD... method setUp (line 92) | def setUp(self): method tearDown (line 97) | def tearDown(self): method get_inputs (line 102) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_single_file_format_inference_is_same_as_pretrained (line 114) | def test_single_file_format_inference_is_same_as_pretrained(self): FILE: diffusers/tests/single_file/test_stable_diffusion_upscale_single_file.py class StableDiffusionUpscalePipelineSingleFileSlowTests (line 25) | class StableDiffusionUpscalePipelineSingleFileSlowTests(unittest.TestCas... method setUp (line 31) | def setUp(self): method tearDown (line 36) | def tearDown(self): method test_single_file_format_inference_is_same_as_pretrained (line 41) | def test_single_file_format_inference_is_same_as_pretrained(self): FILE: diffusers/tests/single_file/test_stable_diffusion_xl_adapter_single_file.py class StableDiffusionXLAdapterPipelineSingleFileSlowTests (line 32) | class StableDiffusionXLAdapterPipelineSingleFileSlowTests(unittest.TestC... method setUp (line 40) | def setUp(self): method tearDown (line 45) | def tearDown(self): method get_inputs (line 50) | def get_inputs(self): method test_single_file_format_inference_is_same_as_pretrained (line 68) | def test_single_file_format_inference_is_same_as_pretrained(self): method test_single_file_components (line 99) | def test_single_file_components(self): method test_single_file_components_local_files_only (line 111) | def test_single_file_components_local_files_only(self): method test_single_file_components_with_diffusers_config (line 130) | def test_single_file_components_with_diffusers_config(self): method test_single_file_components_with_diffusers_config_local_files_only (line 143) | def test_single_file_components_with_diffusers_config_local_files_only... method test_single_file_components_with_original_config (line 166) | def test_single_file_components_with_original_config(self): method test_single_file_components_with_original_config_local_files_only (line 181) | def test_single_file_components_with_original_config_local_files_only(... FILE: diffusers/tests/single_file/test_stable_diffusion_xl_controlnet_single_file.py class StableDiffusionXLControlNetPipelineSingleFileSlowTests (line 29) | class StableDiffusionXLControlNetPipelineSingleFileSlowTests(unittest.Te... method setUp (line 37) | def setUp(self): method tearDown (line 42) | def tearDown(self): method get_inputs (line 47) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_single_file_format_inference_is_same_as_pretrained (line 64) | def test_single_file_format_inference_is_same_as_pretrained(self): method test_single_file_components (line 89) | def test_single_file_components(self): method test_single_file_components_local_files_only (line 103) | def test_single_file_components_local_files_only(self): method test_single_file_components_with_original_config (line 124) | def test_single_file_components_with_original_config(self): method test_single_file_components_with_original_config_local_files_only (line 142) | def test_single_file_components_with_original_config_local_files_only(... method test_single_file_components_with_diffusers_config (line 165) | def test_single_file_components_with_diffusers_config(self): method test_single_file_components_with_diffusers_config_local_files_only (line 176) | def test_single_file_components_with_diffusers_config_local_files_only... FILE: diffusers/tests/single_file/test_stable_diffusion_xl_img2img_single_file.py class StableDiffusionXLImg2ImgPipelineSingleFileSlowTests (line 26) | class StableDiffusionXLImg2ImgPipelineSingleFileSlowTests(unittest.TestC... method setUp (line 34) | def setUp(self): method tearDown (line 39) | def tearDown(self): method get_inputs (line 44) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_single_file_format_inference_is_same_as_pretrained (line 61) | def test_single_file_format_inference_is_same_as_pretrained(self): class StableDiffusionXLImg2ImgRefinerPipelineSingleFileSlowTests (line 67) | class StableDiffusionXLImg2ImgRefinerPipelineSingleFileSlowTests(unittes... method test_single_file_format_inference_is_same_as_pretrained (line 77) | def test_single_file_format_inference_is_same_as_pretrained(self): FILE: diffusers/tests/single_file/test_stable_diffusion_xl_instruct_pix2pix.py class StableDiffusionXLInstructPix2PixPipeline (line 19) | class StableDiffusionXLInstructPix2PixPipeline(unittest.TestCase): method setUp (line 25) | def setUp(self): method tearDown (line 30) | def tearDown(self): method get_inputs (line 35) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_single_file_setting_cosxl_edit (line 47) | def test_single_file_setting_cosxl_edit(self): FILE: diffusers/tests/single_file/test_stable_diffusion_xl_single_file.py class StableDiffusionXLPipelineSingleFileSlowTests (line 23) | class StableDiffusionXLPipelineSingleFileSlowTests(unittest.TestCase, SD... method setUp (line 31) | def setUp(self): method tearDown (line 36) | def tearDown(self): method get_inputs (line 41) | def get_inputs(self, device, generator_device="cpu", dtype=torch.float... method test_single_file_format_inference_is_same_as_pretrained (line 53) | def test_single_file_format_inference_is_same_as_pretrained(self): FILE: diffusers/utils/check_config_docstrings.py function check_config_docstrings_have_checkpoints (line 53) | def check_config_docstrings_have_checkpoints(): FILE: diffusers/utils/check_copies.py function _should_continue (line 29) | def _should_continue(line, indent): function find_code_in_diffusers (line 33) | def find_code_in_diffusers(object_name): function get_indent (line 86) | def get_indent(code): function run_ruff (line 96) | def run_ruff(code): function stylify (line 103) | def stylify(code: str) -> str: function is_copy_consistent (line 121) | def is_copy_consistent(filename, overwrite=False): function check_copies (line 198) | def check_copies(overwrite: bool = False): FILE: diffusers/utils/check_doc_toc.py function clean_doc_toc (line 25) | def clean_doc_toc(doc_list): function check_scheduler_doc (line 70) | def check_scheduler_doc(overwrite=False): function check_pipeline_doc (line 105) | def check_pipeline_doc(overwrite=False): FILE: diffusers/utils/check_dummies.py function find_backend (line 58) | def find_backend(line): function read_init (line 67) | def read_init(): function create_dummy_object (line 105) | def create_dummy_object(name, backend_name): function create_dummy_files (line 115) | def create_dummy_files(backend_specific_objects=None): function check_dummies (line 132) | def check_dummies(overwrite=False): FILE: diffusers/utils/check_inits.py function find_backend (line 50) | def find_backend(line): function parse_init (line 59) | def parse_init(init_file): function analyze_results (line 190) | def analyze_results(import_dict_objects, type_hint_objects): function check_all_inits (line 222) | def check_all_inits(): function get_transformers_submodules (line 241) | def get_transformers_submodules(): function check_submodules (line 274) | def check_submodules(): FILE: diffusers/utils/check_repo.py function check_model_list (line 209) | def check_model_list(): function get_model_modules (line 231) | def get_model_modules(): function get_models (line 270) | def get_models(module, include_pretrained=False): function is_a_private_model (line 283) | def is_a_private_model(model): function check_models_are_in_init (line 298) | def check_models_are_in_init(): function get_model_test_files (line 315) | def get_model_test_files(): function find_tested_models (line 354) | def find_tested_models(test_file): function check_models_are_tested (line 372) | def check_models_are_tested(module, test_file): function check_all_models_are_tested (line 397) | def check_all_models_are_tested(): function get_all_auto_configured_models (line 417) | def get_all_auto_configured_models(): function ignore_unautoclassed (line 431) | def ignore_unautoclassed(model_name): function check_models_are_auto_configured (line 442) | def check_models_are_auto_configured(module, all_auto_models): function check_all_models_are_auto_configured (line 456) | def check_all_models_are_auto_configured(): function check_decorator_order (line 491) | def check_decorator_order(filename): function check_all_decorator_order (line 509) | def check_all_decorator_order(): function find_all_documented_objects (line 525) | def find_all_documented_objects(): function ignore_undocumented (line 615) | def ignore_undocumented(name): function check_all_objects_are_documented (line 653) | def check_all_objects_are_documented(): function check_model_type_doc_match (line 668) | def check_model_type_doc_match(): function is_rst_docstring (line 703) | def is_rst_docstring(docstring): function check_docstrings_are_in_md (line 716) | def check_docstrings_are_in_md(): function check_repo_quality (line 739) | def check_repo_quality(): FILE: diffusers/utils/check_table.py function _find_text_in_file (line 30) | def _find_text_in_file(filename, start_prompt, end_prompt): function camel_case_split (line 75) | def camel_case_split(identifier): function _center_text (line 81) | def _center_text(text, width): function get_model_table_from_auto_modules (line 88) | def get_model_table_from_auto_modules(): function check_model_table (line 161) | def check_model_table(overwrite=False): FILE: diffusers/utils/custom_init_isort.py function get_indent (line 59) | def get_indent(line: str) -> str: function split_code_in_indented_blocks (line 65) | def split_code_in_indented_blocks( function ignore_underscore_and_lowercase (line 132) | def ignore_underscore_and_lowercase(key: Callable[[Any], str]) -> Callab... function sort_objects (line 143) | def sort_objects(objects: List[Any], key: Optional[Callable[[Any], str]]... function sort_objects_in_import (line 177) | def sort_objects_in_import(import_statement: str) -> str: function sort_imports (line 235) | def sort_imports(file: str, check_only: bool = True): function sort_imports_in_all_inits (line 308) | def sort_imports_in_all_inits(check_only=True): FILE: diffusers/utils/fetch_latest_release_branch.py function fetch_all_branches (line 25) | def fetch_all_branches(user, repo): function main (line 49) | def main(): FILE: diffusers/utils/fetch_torch_cuda_pipeline_test_matrix.py function filter_pipelines (line 33) | def filter_pipelines(usage_dict, usage_cutoff=10000): function fetch_pipeline_objects (line 48) | def fetch_pipeline_objects(): function fetch_pipeline_modules_to_test (line 69) | def fetch_pipeline_modules_to_test(): function main (line 86) | def main(): FILE: diffusers/utils/notify_slack_about_release.py function check_pypi_for_latest_release (line 27) | def check_pypi_for_latest_release(library_name): function get_github_release_info (line 38) | def get_github_release_info(github_repo): function notify_slack (line 52) | def notify_slack(webhook_url, library_name, version, release_info): function main (line 68) | def main(): FILE: diffusers/utils/overwrite_expected_slice.py function overwrite_file (line 19) | def overwrite_file(file, class_name, test_name, correct_line, done_test): function main (line 67) | def main(correct, fail=None): FILE: diffusers/utils/release.py function update_version_in_file (line 37) | def update_version_in_file(fname, version, pattern): function update_version_in_examples (line 48) | def update_version_in_examples(version): function global_version_update (line 61) | def global_version_update(version, patch=False): function clean_main_ref_in_model_list (line 69) | def clean_main_ref_in_model_list(): function get_version (line 97) | def get_version(): function pre_release_work (line 105) | def pre_release_work(patch=False): function post_release_work (line 132) | def post_release_work(): FILE: diffusers/utils/stale.py function main (line 37) | def main(): FILE: diffusers/utils/tests_fetcher.py function checkout_commit (line 88) | def checkout_commit(repo: Repo, commit_id: str): function clean_code (line 106) | def clean_code(content: str) -> str: function keep_doc_examples_only (line 138) | def keep_doc_examples_only(content: str) -> str: function get_all_tests (line 165) | def get_all_tests() -> List[str]: function diff_is_docstring_only (line 183) | def diff_is_docstring_only(repo: Repo, branching_point: str, filename: s... function diff_contains_doc_examples (line 209) | def diff_contains_doc_examples(repo: Repo, branching_point: str, filenam... function get_diff (line 235) | def get_diff(repo: Repo, base_commit: str, commits: List[str]) -> List[s... function get_modified_python_files (line 277) | def get_modified_python_files(diff_with_last_commit: bool = False) -> Li... function get_diff_for_doctesting (line 310) | def get_diff_for_doctesting(repo: Repo, base_commit: str, commits: List[... function get_all_doctest_files (line 351) | def get_all_doctest_files() -> List[str]: function get_new_doctest_files (line 379) | def get_new_doctest_files(repo, base_commit, branching_commit) -> List[s... function get_doctest_files (line 406) | def get_doctest_files(diff_with_last_commit: bool = False) -> List[str]: function extract_imports (line 476) | def extract_imports(module_fname: str, cache: Dict[str, List[str]] = Non... function get_module_dependencies (line 560) | def get_module_dependencies(module_fname: str, cache: Dict[str, List[str... function create_reverse_dependency_tree (line 620) | def create_reverse_dependency_tree() -> List[Tuple[str, str]]: function get_tree_starting_at (line 632) | def get_tree_starting_at(module: str, edges: List[Tuple[str, str]]) -> L... function print_tree_deps_of (line 660) | def print_tree_deps_of(module, all_edges=None): function init_test_examples_dependencies (line 693) | def init_test_examples_dependencies() -> Tuple[Dict[str, List[str]], Lis... function create_reverse_dependency_map (line 729) | def create_reverse_dependency_map() -> Dict[str, List[str]]: function create_module_to_test_map (line 781) | def create_module_to_test_map(reverse_map: Dict[str, List[str]] = None) ... function check_imports_all_exist (line 812) | def check_imports_all_exist(): function _print_list (line 828) | def _print_list(l) -> str: function update_test_map_with_core_pipelines (line 835) | def update_test_map_with_core_pipelines(json_output_file: str): function create_json_map (line 867) | def create_json_map(test_files_to_run: List[str], json_output_file: Opti... function infer_tests_to_run (line 909) | def infer_tests_to_run( function filter_tests (line 999) | def filter_tests(output_file: str, filters: List[str]): function parse_commit_message (line 1026) | def parse_commit_message(commit_message: str) -> Dict[str, bool]: FILE: diffusers/utils/update_metadata.py function get_supported_pipeline_table (line 46) | def get_supported_pipeline_table() -> dict: function update_metadata (line 75) | def update_metadata(commit_sha: str): FILE: traning/train_sd3_pix2pix.py function load_text_encoders (line 64) | def load_text_encoders(class_one, class_two, class_three): function tokenize_prompt (line 77) | def tokenize_prompt(tokenizer, prompt, max_sequence_length=77): function _encode_prompt_with_t5 (line 89) | def _encode_prompt_with_t5( function _encode_prompt_with_clip (line 125) | def _encode_prompt_with_clip( function encode_prompt (line 160) | def encode_prompt( function parse_args (line 229) | def parse_args(): function combine_rgb_and_mask_to_rgba (line 641) | def combine_rgb_and_mask_to_rgba(rgb_image, mask_image): function convert_to_np (line 659) | def convert_to_np(image, resolution): function import_model_class_from_model_name_or_path (line 678) | def import_model_class_from_model_name_or_path( function main (line 697) | def main(): FILE: traning/train_sdxl_pix2pix.py function log_validation (line 75) | def log_validation( function import_model_class_from_model_name_or_path (line 125) | def import_model_class_from_model_name_or_path( function parse_args (line 145) | def parse_args(): function convert_to_np (line 508) | def convert_to_np(image, resolution): function download_image (line 526) | def download_image(path_or_url): function main (line 540) | def main():