SYMBOL INDEX (190 symbols across 24 files) FILE: animesr/archs/discriminator_arch.py function get_conv_layer (line 9) | def get_conv_layer(input_nc, ndf, kernel_size, stride, padding, bias=Tru... class UNetDiscriminatorSN (line 16) | class UNetDiscriminatorSN(nn.Module): method __init__ (line 19) | def __init__(self, num_in_ch, num_feat=64, skip_connection=True): method forward (line 40) | def forward(self, x): class PatchDiscriminator (line 72) | class PatchDiscriminator(nn.Module): method __init__ (line 79) | def __init__(self, method _get_norm_layer (line 141) | def _get_norm_layer(self, norm_type='batch'): method forward (line 155) | def forward(self, x): class MultiScaleDiscriminator (line 160) | class MultiScaleDiscriminator(nn.Module): method __init__ (line 171) | def __init__(self, method forward (line 204) | def forward(self, x): FILE: animesr/archs/simple_degradation_arch.py class SimpleDegradationArch (line 8) | class SimpleDegradationArch(nn.Module): method __init__ (line 13) | def __init__(self, num_in_ch=3, num_out_ch=3, num_feat=64, downscale=2): method forward (line 37) | def forward(self, x): FILE: animesr/archs/vsr_arch.py class RightAlignMSConvResidualBlocks (line 9) | class RightAlignMSConvResidualBlocks(nn.Module): method __init__ (line 12) | def __init__(self, num_in_ch=3, num_state_ch=64, num_out_ch=64, num_bl... method up (line 47) | def up(self, x, scale=2): method forward (line 55) | def forward(self, x): class MSRSWVSR (line 79) | class MSRSWVSR(nn.Module): method __init__ (line 85) | def __init__(self, num_feat=64, num_block=(5, 3, 2), netscale=4): method cell (line 96) | def cell(self, x, fb, state): method forward (line 108) | def forward(self, x): FILE: animesr/data/data_utils.py function random_crop (line 5) | def random_crop(imgs, patch_size, top=None, left=None): FILE: animesr/data/ffmpeg_anime_dataset.py class FFMPEGAnimeDataset (line 19) | class FFMPEGAnimeDataset(data.Dataset): method __init__ (line 22) | def __init__(self, opt): method get_gt_clip (line 63) | def get_gt_clip(self, index): method add_ffmpeg_compression (line 114) | def add_ffmpeg_compression(self, img_lqs, width, height): method __getitem__ (line 175) | def __getitem__(self, index): method __len__ (line 209) | def __len__(self): FILE: animesr/data/ffmpeg_anime_lbo_dataset.py class FFMPEGAnimeLBODataset (line 16) | class FFMPEGAnimeLBODataset(FFMPEGAnimeDataset): method __init__ (line 19) | def __init__(self, opt): method reload_degradation_model (line 35) | def reload_degradation_model(self): method custom_resize (line 49) | def custom_resize(self, x, scale=2): method __getitem__ (line 66) | def __getitem__(self, index): method __len__ (line 108) | def __len__(self): FILE: animesr/data/paired_image_dataset.py class CustomPairedImageDataset (line 12) | class CustomPairedImageDataset(data.Dataset): method __init__ (line 33) | def __init__(self, opt): method __getitem__ (line 58) | def __getitem__(self, index): method __len__ (line 91) | def __len__(self): FILE: animesr/models/degradation_gan_model.py class DegradationGANModel (line 8) | class DegradationGANModel(SRGANModel): method feed_data (line 11) | def feed_data(self, data): method optimize_parameters (line 17) | def optimize_parameters(self, current_iter): FILE: animesr/models/degradation_model.py class DegradationModel (line 9) | class DegradationModel(SRModel): method init_training_settings (line 12) | def init_training_settings(self): method feed_data (line 24) | def feed_data(self, data): method optimize_parameters (line 30) | def optimize_parameters(self, current_iter): FILE: animesr/models/video_recurrent_gan_model.py class VideoRecurrentGANCustomModel (line 11) | class VideoRecurrentGANCustomModel(VideoRecurrentCustomModel): method init_training_settings (line 16) | def init_training_settings(self): method setup_optimizers (line 70) | def setup_optimizers(self): method optimize_parameters (line 103) | def optimize_parameters(self, current_iter): method save (line 172) | def save(self, epoch, current_iter): FILE: animesr/models/video_recurrent_model.py class VideoRecurrentCustomModel (line 16) | class VideoRecurrentCustomModel(VideoBaseModel): method __init__ (line 18) | def __init__(self, opt): method feed_data (line 34) | def feed_data(self, data): method setup_optimizers (line 82) | def setup_optimizers(self): method optimize_parameters_base (line 112) | def optimize_parameters_base(self, current_iter): method optimize_parameters (line 127) | def optimize_parameters(self, current_iter): method dist_validation (line 163) | def dist_validation(self, dataloader, current_iter, tb_logger, save_img): method test (line 231) | def test(self): FILE: animesr/utils/inference_base.py function get_base_argument_parser (line 8) | def get_base_argument_parser() -> argparse.ArgumentParser: function get_inference_model (line 45) | def get_inference_model(args, device) -> MSRSWVSR: FILE: animesr/utils/shot_detector.py function compute_downscale_factor (line 44) | def compute_downscale_factor(frame_width): class ShotDetector (line 58) | class ShotDetector(object): method __init__ (line 64) | def __init__(self, threshold=30.0, min_shot_len=15): method add_cut (line 83) | def add_cut(self, cut): method process_frame (line 90) | def process_frame(self, frame_num, frame_img): method detect_shots (line 167) | def detect_shots(self, frame_source, frame_skip=0, show_progress=True,... FILE: animesr/utils/video_util.py function get_video_fps (line 11) | def get_video_fps(video_path, ret_type='float'): function get_video_num_frames (line 46) | def get_video_num_frames(video_path): function get_video_bitrate (line 64) | def get_video_bitrate(video_path): function get_video_resolution (line 85) | def get_video_resolution(video_path): function video2frames (line 113) | def video2frames(video_path, out_dir, force=False, high_quality=True, ss... function frames2video (line 152) | def frames2video(frames_dir, out_path, fps=25, filter='*', suffix=None): FILE: predict.py class ModelOutput (line 16) | class ModelOutput(BaseModel): class Predictor (line 21) | class Predictor(BasePredictor): method predict (line 23) | def predict( FILE: scripts/anime_videos_preprocessing.py function main (line 22) | def main(args): function run_step1 (line 89) | def run_step1(opt): function worker1 (line 127) | def worker1(opt, video_name, video_path, frame_path, meta_path): function run_step2 (line 166) | def run_step2(opt): function worker2 (line 192) | def worker2(opt, video_name): function run_step3 (line 211) | def run_step3(opt): function read_img (line 244) | def read_img(img_path, device, downscale_factor=1): function worker3 (line 255) | def worker3(opt, video_name, device): function run_step4 (line 306) | def run_step4(opt): function worker4 (line 333) | def worker4(opt, video_name): function run_step5 (line 373) | def run_step5(opt): function worker5 (line 404) | def worker5(opt, video_name, device): function filter_frozen_shots (line 466) | def filter_frozen_shots(shots, flows): function generate_clips (line 492) | def generate_clips(shots, flows, filter_frames, hyperiqa, max_length=500): function run_step6 (line 561) | def run_step6(opt): function worker6 (line 589) | def worker6(opt, video_name): FILE: scripts/inference_animesr_frames.py function read_img (line 21) | def read_img(path, require_mod_crop=True, mod_scale=4, input_rescaling_f... class IOConsumer (line 47) | class IOConsumer(threading.Thread): method __init__ (line 52) | def __init__(self, args: argparse.Namespace, que, qid): method run (line 58) | def run(self): function main (line 79) | def main(): FILE: scripts/inference_animesr_video.py function get_video_meta_info (line 20) | def get_video_meta_info(video_path): function get_sub_video (line 39) | def get_sub_video(args, num_process, process_idx): class Reader (line 62) | class Reader: method __init__ (line 65) | def __init__(self, args, total_workers=1, worker_idx=0, device=torch.d... method get_resolution (line 105) | def get_resolution(self): method get_fps (line 108) | def get_fps(self): method get_audio (line 117) | def get_audio(self): method __len__ (line 120) | def __len__(self): method get_frame_from_stream (line 124) | def get_frame_from_stream(self): method get_frame_from_list (line 132) | def get_frame_from_list(self): method get_frame (line 139) | def get_frame(self): method close (line 157) | def close(self): class Writer (line 164) | class Writer: method __init__ (line 167) | def __init__(self, args, audio, height, width, video_save_path, fps): method write_frame (line 195) | def write_frame(self, frame): method close (line 200) | def close(self): function inference_video (line 206) | def inference_video(args, video_save_path, device=None, total_workers=1,... function run (line 277) | def run(args): function main (line 332) | def main(): FILE: scripts/metrics/MANIQA/inference_MANIQA.py function parse_args (line 12) | def parse_args(): function main (line 25) | def main(): FILE: scripts/metrics/MANIQA/models/model_attentionIQA2.py class ChannelAttn (line 10) | class ChannelAttn(nn.Module): method __init__ (line 12) | def __init__(self, dim, drop=0.1): method forward (line 22) | def forward(self, x): class SaveOutput (line 38) | class SaveOutput: method __init__ (line 40) | def __init__(self): method __call__ (line 43) | def __call__(self, module, module_in, module_out): method clear (line 46) | def clear(self): class AttentionIQA (line 50) | class AttentionIQA(nn.Module): method __init__ (line 52) | def __init__(self, method extract_feature (line 105) | def extract_feature(self, save_output): method forward (line 113) | def forward(self, x): FILE: scripts/metrics/MANIQA/models/swin.py function get_attn_decoder_mask (line 14) | def get_attn_decoder_mask(seq): function get_attn_pad_mask (line 23) | def get_attn_pad_mask(seq_q, seq_k, i_pad): class DecoderWindowAttention (line 31) | class DecoderWindowAttention(nn.Module): method __init__ (line 45) | def __init__(self, dim, window_size, num_heads, qk_scale=None, attn_dr... method forward (line 83) | def forward(self, q, k, v, mask=None, attn_mask=None): method extra_repr (line 119) | def extra_repr(self) -> str: method flops (line 122) | def flops(self, N): class DecoderLayer (line 139) | class DecoderLayer(nn.Module): method __init__ (line 141) | def __init__(self, method partition (line 196) | def partition(self, inputs, B, H, W, C, shift_size=0): method reverse (line 208) | def reverse(self, inputs, B, H, W, C, shift_size=0): method forward (line 220) | def forward(self, mask_dec_inputs, enc_outputs, self_attn_mask, dec_en... class SwinDecoder (line 262) | class SwinDecoder(nn.Module): method __init__ (line 264) | def __init__(self, method forward (line 294) | def forward(self, y_embed, enc_outputs): class Mlp (line 328) | class Mlp(nn.Module): method __init__ (line 330) | def __init__(self, in_features, hidden_features=None, out_features=Non... method forward (line 339) | def forward(self, x): function window_partition (line 348) | def window_partition(x, window_size): function window_reverse (line 363) | def window_reverse(windows, window_size, H, W): class WindowAttention (line 380) | class WindowAttention(nn.Module): method __init__ (line 394) | def __init__(self, dim, window_size, num_heads, qkv_bias=True, qk_scal... method forward (line 428) | def forward(self, x, mask=None): method extra_repr (line 461) | def extra_repr(self) -> str: method flops (line 464) | def flops(self, N): class SwinBlock (line 478) | class SwinBlock(nn.Module): method __init__ (line 497) | def __init__(self, method forward (line 562) | def forward(self, x): method extra_repr (line 601) | def extra_repr(self) -> str: method flops (line 605) | def flops(self): class BasicLayer (line 620) | class BasicLayer(nn.Module): method __init__ (line 640) | def __init__(self, method forward (line 686) | def forward(self, x): method extra_repr (line 697) | def extra_repr(self) -> str: method flops (line 700) | def flops(self): class SwinTransformer (line 709) | class SwinTransformer(nn.Module): method __init__ (line 711) | def __init__(self, method forward (line 762) | def forward(self, x): FILE: scripts/metrics/MANIQA/pipal_data.py class NTIRE2022 (line 7) | class NTIRE2022(torch.utils.data.Dataset): method __init__ (line 9) | def __init__(self, ref_path, dis_path, transform): method __len__ (line 22) | def __len__(self): method __getitem__ (line 25) | def __getitem__(self, idx): FILE: scripts/metrics/MANIQA/utils.py function crop_image (line 5) | def crop_image(top, left, patch_size, img=None): class RandCrop (line 10) | class RandCrop(object): method __init__ (line 12) | def __init__(self, patch_size, num_crop): method __call__ (line 16) | def __call__(self, sample): class Normalize (line 41) | class Normalize(object): method __init__ (line 43) | def __init__(self, mean, var): method __call__ (line 47) | def __call__(self, sample): class RandHorizontalFlip (line 59) | class RandHorizontalFlip(object): method __init__ (line 61) | def __init__(self): method __call__ (line 64) | def __call__(self, sample): class ToTensor (line 77) | class ToTensor(object): method __init__ (line 79) | def __init__(self): method __call__ (line 82) | def __call__(self, sample): FILE: setup.py function readme (line 12) | def readme(): function get_git_hash (line 18) | def get_git_hash(): function get_hash (line 43) | def get_hash(): function write_version_py (line 58) | def write_version_py(): function get_version (line 75) | def get_version(): function get_requirements (line 81) | def get_requirements(filename='requirements.txt'):