SYMBOL INDEX (2762 symbols across 348 files) FILE: camera_pose_annotation/camera_tracking/camera_tracking.py function image_stream (line 41) | def image_stream( function save_full_reconstruction (line 103) | def save_full_reconstruction( function parse_args (line 145) | def parse_args(): function main (line 177) | def main(): FILE: camera_pose_annotation/camera_tracking/inference_batch.py function process_single_row (line 21) | def process_single_row(row, index, args, worker_id=0): function worker (line 42) | def worker(task_queue, args, worker_id, pbar): function parse_args (line 56) | def parse_args(): function main (line 77) | def main(): FILE: camera_pose_annotation/cvd_opt/cvd_opt.py function save_depth (line 39) | def save_depth(path, depths): function gradient_loss (line 52) | def gradient_loss(gt, pred, u): function si_loss (line 76) | def si_loss(gt, pred): function sobel_fg_alpha (line 87) | def sobel_fg_alpha(disp, mode="sobel", beta=10.0): function consistency_loss (line 101) | def consistency_loss( function parse_args (line 245) | def parse_args(): FILE: camera_pose_annotation/cvd_opt/geometry_utils.py function to_homogeneous (line 30) | def to_homogeneous(input_tensor: Tensor, dim: int = 0) -> Tensor: class BackprojectDepth (line 37) | class BackprojectDepth(nn.Module): method __init__ (line 43) | def __init__(self, height: int, width: int): method forward (line 70) | def forward(self, depth_b1hw: Tensor, invK_b44: Tensor) -> Tensor: class Project3D (line 80) | class Project3D(jit.ScriptModule): method __init__ (line 83) | def __init__(self, eps: float = 1e-8): method forward (line 89) | def forward( class NormalGenerator (line 110) | class NormalGenerator(nn.Module): method __init__ (line 113) | def __init__( method forward (line 131) | def forward(self, depth_b1hw: Tensor, invK_b44: Tensor) -> Tensor: function get_camera_rays (line 158) | def get_camera_rays( function pose_distance (line 191) | def pose_distance(pose_b44): function qvec2rotmat (line 206) | def qvec2rotmat(qvec): function rotx (line 227) | def rotx(t): function roty (line 234) | def roty(t): function rotz (line 241) | def rotz(t): FILE: camera_pose_annotation/cvd_opt/inference_batch.py function process_single_row (line 16) | def process_single_row(row, index, args, worker_id=0): function worker (line 37) | def worker(task_queue, args, worker_id, pbar): function parse_args (line 49) | def parse_args(): function main (line 70) | def main(): FILE: camera_pose_annotation/cvd_opt/preprocess/core/corr.py class CorrBlock (line 30) | class CorrBlock: method __init__ (line 33) | def __init__(self, fmap1, fmap2, num_levels=4, radius=4): method __call__ (line 49) | def __call__(self, coords): method corr (line 73) | def corr(cls, fmap1, fmap2): class AlternateCorrBlock (line 84) | class AlternateCorrBlock: method __init__ (line 87) | def __init__(self, fmap1, fmap2, num_levels=4, radius=4): method __call__ (line 97) | def __call__(self, coords): FILE: camera_pose_annotation/cvd_opt/preprocess/core/datasets.py class FlowDataset (line 31) | class FlowDataset(data.Dataset): method __init__ (line 34) | def __init__(self, aug_params=None, sparse=False): method __getitem__ (line 49) | def __getitem__(self, index): method __rmul__ (line 107) | def __rmul__(self, v): method __len__ (line 112) | def __len__(self): class MpiSintel (line 116) | class MpiSintel(FlowDataset): method __init__ (line 119) | def __init__( class FlyingChairs (line 143) | class FlyingChairs(FlowDataset): method __init__ (line 146) | def __init__( class FlyingThings3D (line 168) | class FlyingThings3D(FlowDataset): method __init__ (line 171) | def __init__( class KITTI (line 199) | class KITTI(FlowDataset): method __init__ (line 202) | def __init__(self, aug_params=None, split='training', root='datasets/K... class HD1K (line 220) | class HD1K(FlowDataset): method __init__ (line 223) | def __init__(self, aug_params=None, root='datasets/HD1k'): function fetch_dataloader (line 248) | def fetch_dataloader(args, TRAIN_DS='C+T+K+S+H'): FILE: camera_pose_annotation/cvd_opt/preprocess/core/extractor.py class ResidualBlock (line 22) | class ResidualBlock(nn.Module): method __init__ (line 25) | def __init__(self, in_planes, planes, norm_fn='group', stride=1): method forward (line 68) | def forward(self, x): class BottleneckBlock (line 79) | class BottleneckBlock(nn.Module): method __init__ (line 82) | def __init__(self, in_planes, planes, norm_fn='group', stride=1): method forward (line 130) | def forward(self, x): class BasicEncoder (line 142) | class BasicEncoder(nn.Module): method __init__ (line 145) | def __init__(self, output_dim=128, norm_fn='batch', dropout=0.0): method _make_layer (line 185) | def _make_layer(self, dim, stride=1): method forward (line 193) | def forward(self, x): class SmallEncoder (line 220) | class SmallEncoder(nn.Module): method __init__ (line 223) | def __init__(self, output_dim=128, norm_fn='batch', dropout=0.0): method _make_layer (line 262) | def _make_layer(self, dim, stride=1): method forward (line 270) | def forward(self, x): FILE: camera_pose_annotation/cvd_opt/preprocess/core/raft.py class autocast (line 34) | class autocast: # pylint: disable=invalid-name method __init__ (line 36) | def __init__(self, enabled): method __enter__ (line 39) | def __enter__(self): method __exit__ (line 42) | def __exit__(self, *args): class RAFT (line 46) | class RAFT(nn.Module): method __init__ (line 49) | def __init__(self, args): method freeze_bn (line 90) | def freeze_bn(self): method initialize_flow (line 95) | def initialize_flow(self, img): method upsample_flow (line 105) | def upsample_flow(self, flow, mask): method forward (line 119) | def forward( FILE: camera_pose_annotation/cvd_opt/preprocess/core/update.py class FlowHead (line 23) | class FlowHead(nn.Module): method __init__ (line 25) | def __init__(self, input_dim=128, hidden_dim=256): method forward (line 31) | def forward(self, x): class ConvGRU (line 35) | class ConvGRU(nn.Module): method __init__ (line 38) | def __init__(self, hidden_dim=128, input_dim=192 + 128): method forward (line 44) | def forward(self, h, x): class SepConvGRU (line 55) | class SepConvGRU(nn.Module): method __init__ (line 58) | def __init__(self, hidden_dim=128, input_dim=192 + 128): method forward (line 80) | def forward(self, h, x): class SmallMotionEncoder (line 98) | class SmallMotionEncoder(nn.Module): method __init__ (line 101) | def __init__(self, args): method forward (line 109) | def forward(self, flow, corr): class BasicMotionEncoder (line 118) | class BasicMotionEncoder(nn.Module): method __init__ (line 121) | def __init__(self, args): method forward (line 130) | def forward(self, flow, corr): class SmallUpdateBlock (line 141) | class SmallUpdateBlock(nn.Module): method __init__ (line 144) | def __init__(self, args, hidden_dim=96): method forward (line 150) | def forward(self, net, inp, corr, flow): class BasicUpdateBlock (line 159) | class BasicUpdateBlock(nn.Module): method __init__ (line 162) | def __init__(self, args, hidden_dim=128, input_dim=128): method forward (line 175) | def forward(self, net, inp, corr, flow, upsample=True): FILE: camera_pose_annotation/cvd_opt/preprocess/core/utils/augmentor.py class FlowAugmentor (line 30) | class FlowAugmentor: method __init__ (line 33) | def __init__(self, crop_size, min_scale=-0.2, max_scale=0.5, do_flip=T... method color_transform (line 55) | def color_transform(self, img1, img2): method eraser_transform (line 73) | def eraser_transform(self, img1, img2, bounds=[50, 100]): # pylint: d... method spatial_transform (line 88) | def spatial_transform(self, img1, img2, flow): method __call__ (line 140) | def __call__(self, img1, img2, flow): class SparseFlowAugmentor (line 152) | class SparseFlowAugmentor: method __init__ (line 155) | def __init__(self, crop_size, min_scale=-0.2, max_scale=0.5, do_flip=F... method color_transform (line 176) | def color_transform(self, img1, img2): method eraser_transform (line 184) | def eraser_transform(self, img1, img2): method resize_sparse_flow_map (line 197) | def resize_sparse_flow_map(self, flow, valid, fx=1.0, fy=1.0): method spatial_transform (line 232) | def spatial_transform(self, img1, img2, flow, valid): method __call__ (line 280) | def __call__(self, img1, img2, flow, valid): FILE: camera_pose_annotation/cvd_opt/preprocess/core/utils/flow_viz.py function make_colorwheel (line 24) | def make_colorwheel(): function flow_uv_to_colors (line 76) | def flow_uv_to_colors(u, v, convert_to_bgr=False): function flow_to_image (line 115) | def flow_to_image(flow_uv, clip_flow=None, convert_to_bgr=False): FILE: camera_pose_annotation/cvd_opt/preprocess/core/utils/frame_utils.py function readFlow (line 35) | def readFlow(fn): function readPFM (line 57) | def readPFM(file): function writeFlow (line 89) | def writeFlow(filename, uv, v=None): function readFlowKITTI (line 121) | def readFlowKITTI(filename): function readDispKITTI (line 129) | def readDispKITTI(filename): function writeFlowKITTI (line 136) | def writeFlowKITTI(filename, uv): function read_gen (line 143) | def read_gen(file_name, pil=False): FILE: camera_pose_annotation/cvd_opt/preprocess/core/utils/utils.py class InputPadder (line 26) | class InputPadder: method __init__ (line 29) | def __init__(self, dims, mode='sintel'): method pad (line 43) | def pad(self, *inputs): method unpad (line 46) | def unpad(self, x): function forward_interpolate (line 52) | def forward_interpolate(flow): function bilinear_sampler (line 86) | def bilinear_sampler(img, coords, mode='bilinear', mask=False): function coords_grid (line 104) | def coords_grid(batch, ht, wd): function upflow8 (line 110) | def upflow8(flow, mode='bilinear'): FILE: camera_pose_annotation/cvd_opt/preprocess/inference_batch.py function process_single_row (line 16) | def process_single_row(row, index, args, worker_id=0): function worker (line 36) | def worker(task_queue, args, worker_id, pbar): function parse_args (line 48) | def parse_args(): function main (line 69) | def main(): FILE: camera_pose_annotation/cvd_opt/preprocess/preprocess_flow.py function warp_flow (line 32) | def warp_flow(img, flow): function resize_flow (line 44) | def resize_flow(flow, img_h, img_w): function parse_args (line 54) | def parse_args(): FILE: camera_pose_annotation/depth_estimation/Depth-Anything/depth_anything_v2/dinov2.py function named_apply (line 26) | def named_apply(fn: Callable, module: nn.Module, name="", depth_first=Tr... class BlockChunk (line 37) | class BlockChunk(nn.ModuleList): method forward (line 38) | def forward(self, x): class DinoVisionTransformer (line 44) | class DinoVisionTransformer(nn.Module): method __init__ (line 45) | def __init__( method init_weights (line 172) | def init_weights(self): method interpolate_pos_encoding (line 179) | def interpolate_pos_encoding(self, x, w, h): method prepare_tokens_with_masks (line 212) | def prepare_tokens_with_masks(self, x, masks=None): method forward_features_list (line 233) | def forward_features_list(self, x_list, masks_list): method forward_features (line 253) | def forward_features(self, x, masks=None): method _get_intermediate_layers_not_chunked (line 271) | def _get_intermediate_layers_not_chunked(self, x, n=1): method _get_intermediate_layers_chunked (line 283) | def _get_intermediate_layers_chunked(self, x, n=1): method get_intermediate_layers (line 297) | def get_intermediate_layers( method forward (line 323) | def forward(self, *args, is_training=False, **kwargs): function init_weights_vit_timm (line 331) | def init_weights_vit_timm(module: nn.Module, name: str = ""): function vit_small (line 339) | def vit_small(patch_size=16, num_register_tokens=0, **kwargs): function vit_base (line 353) | def vit_base(patch_size=16, num_register_tokens=0, **kwargs): function vit_large (line 367) | def vit_large(patch_size=16, num_register_tokens=0, **kwargs): function vit_giant2 (line 381) | def vit_giant2(patch_size=16, num_register_tokens=0, **kwargs): function DINOv2 (line 398) | def DINOv2(model_name): FILE: camera_pose_annotation/depth_estimation/Depth-Anything/depth_anything_v2/dinov2_layers/attention.py class Attention (line 29) | class Attention(nn.Module): method __init__ (line 30) | def __init__( method forward (line 49) | def forward(self, x: Tensor) -> Tensor: class MemEffAttention (line 65) | class MemEffAttention(Attention): method forward (line 66) | def forward(self, x: Tensor, attn_bias=None) -> Tensor: FILE: camera_pose_annotation/depth_estimation/Depth-Anything/depth_anything_v2/dinov2_layers/block.py class Block (line 36) | class Block(nn.Module): method __init__ (line 37) | def __init__( method forward (line 82) | def forward(self, x: Tensor) -> Tensor: function drop_add_residual_stochastic_depth (line 110) | def drop_add_residual_stochastic_depth( function get_branges_scales (line 134) | def get_branges_scales(x, sample_drop_ratio=0.0): function add_residual (line 142) | def add_residual(x, brange, residual, residual_scale_factor, scaling_vec... function get_attn_bias_and_cat (line 157) | def get_attn_bias_and_cat(x_list, branges=None): function drop_add_residual_stochastic_depth_list (line 181) | def drop_add_residual_stochastic_depth_list( class NestedTensorBlock (line 204) | class NestedTensorBlock(Block): method forward_nested (line 205) | def forward_nested(self, x_list: List[Tensor]) -> List[Tensor]: method forward (line 245) | def forward(self, x_or_x_list): FILE: camera_pose_annotation/depth_estimation/Depth-Anything/depth_anything_v2/dinov2_layers/drop_path.py function drop_path (line 15) | def drop_path(x, drop_prob: float = 0.0, training: bool = False): class DropPath (line 27) | class DropPath(nn.Module): method __init__ (line 30) | def __init__(self, drop_prob=None): method forward (line 34) | def forward(self, x): FILE: camera_pose_annotation/depth_estimation/Depth-Anything/depth_anything_v2/dinov2_layers/layer_scale.py class LayerScale (line 16) | class LayerScale(nn.Module): method __init__ (line 17) | def __init__( method forward (line 27) | def forward(self, x: Tensor) -> Tensor: FILE: camera_pose_annotation/depth_estimation/Depth-Anything/depth_anything_v2/dinov2_layers/mlp.py class Mlp (line 17) | class Mlp(nn.Module): method __init__ (line 18) | def __init__( method forward (line 35) | def forward(self, x: Tensor) -> Tensor: FILE: camera_pose_annotation/depth_estimation/Depth-Anything/depth_anything_v2/dinov2_layers/patch_embed.py function make_2tuple (line 17) | def make_2tuple(x): class PatchEmbed (line 26) | class PatchEmbed(nn.Module): method __init__ (line 38) | def __init__( method forward (line 69) | def forward(self, x: Tensor) -> Tensor: method flops (line 84) | def flops(self) -> float: FILE: camera_pose_annotation/depth_estimation/Depth-Anything/depth_anything_v2/dinov2_layers/swiglu_ffn.py class SwiGLUFFN (line 13) | class SwiGLUFFN(nn.Module): method __init__ (line 14) | def __init__( method forward (line 29) | def forward(self, x: Tensor) -> Tensor: class SwiGLUFFNFused (line 45) | class SwiGLUFFNFused(SwiGLU): method __init__ (line 46) | def __init__( FILE: camera_pose_annotation/depth_estimation/Depth-Anything/depth_anything_v2/dpt.py function _make_fusion_block (line 12) | def _make_fusion_block(features, use_bn, size=None): class ConvBlock (line 24) | class ConvBlock(nn.Module): method __init__ (line 25) | def __init__(self, in_feature, out_feature): method forward (line 34) | def forward(self, x): class DPTHead (line 38) | class DPTHead(nn.Module): method __init__ (line 39) | def __init__( method forward (line 117) | def forward(self, out_features, patch_h, patch_w): class DepthAnythingV2 (line 153) | class DepthAnythingV2(nn.Module): method __init__ (line 154) | def __init__( method forward (line 176) | def forward(self, x): method infer_image (line 187) | def infer_image(self, raw_image, input_size=518): method image2tensor (line 196) | def image2tensor(self, raw_image, input_size=518): FILE: camera_pose_annotation/depth_estimation/Depth-Anything/depth_anything_v2/util/blocks.py function _make_scratch (line 4) | def _make_scratch(in_shape, out_shape, groups=1, expand=False): class ResidualConvUnit (line 29) | class ResidualConvUnit(nn.Module): method __init__ (line 33) | def __init__(self, features, activation, bn): method forward (line 57) | def forward(self, x): class FeatureFusionBlock (line 83) | class FeatureFusionBlock(nn.Module): method __init__ (line 87) | def __init__( method forward (line 123) | def forward(self, *xs, size=None): FILE: camera_pose_annotation/depth_estimation/Depth-Anything/depth_anything_v2/util/transform.py class Resize (line 5) | class Resize(object): method __init__ (line 9) | def __init__( method constrain_to_multiple_of (line 51) | def constrain_to_multiple_of(self, x, min_val=0, max_val=None): method get_size (line 62) | def get_size(self, width, height): method __call__ (line 109) | def __call__(self, sample): class NormalizeImage (line 125) | class NormalizeImage(object): method __init__ (line 129) | def __init__(self, mean, std): method __call__ (line 133) | def __call__(self, sample): class PrepareForNet (line 139) | class PrepareForNet(object): method __init__ (line 143) | def __init__(self): method __call__ (line 146) | def __call__(self, sample): FILE: camera_pose_annotation/depth_estimation/Depth-Anything/inference.py function parse_args (line 32) | def parse_args(): FILE: camera_pose_annotation/depth_estimation/Depth-Anything/inference_batch.py class ImageDataset (line 41) | class ImageDataset(Dataset): method __init__ (line 44) | def __init__(self, img_list, input_size): method __len__ (line 63) | def __len__(self): method image2tensor (line 66) | def image2tensor(self, raw_image): method __getitem__ (line 77) | def __getitem__(self, idx): function parse_args (line 103) | def parse_args(): function collate_fn (line 122) | def collate_fn(batch): function main (line 133) | def main(): FILE: camera_pose_annotation/depth_estimation/UniDepth/inference.py function parse_args (line 20) | def parse_args(): function main (line 29) | def main(): FILE: camera_pose_annotation/depth_estimation/UniDepth/inference_batch.py class ImageDataset (line 21) | class ImageDataset(Dataset): method __init__ (line 24) | def __init__(self, img_list, input_size): method __len__ (line 28) | def __len__(self): method __getitem__ (line 31) | def __getitem__(self, idx): function collate_fn (line 66) | def collate_fn(batch): function parse_args (line 77) | def parse_args(): function main (line 93) | def main(): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/_2d3ds.py class _2D3DS (line 9) | class _2D3DS(SequenceDataset): method __init__ (line 18) | def __init__( method preprocess (line 51) | def preprocess(self, results): method pre_pipeline (line 62) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/_4dor.py class _4DOR (line 6) | class _4DOR(SequenceDataset): method __init__ (line 16) | def __init__( method pre_pipeline (line 46) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/a2d2.py class A2D2 (line 12) | class A2D2(ImageDataset): method __init__ (line 20) | def __init__( method load_dataset (line 47) | def load_dataset(self): method pre_pipeline (line 74) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/adt.py class ADT (line 8) | class ADT(SequenceDataset): method __init__ (line 17) | def __init__( method preprocess (line 47) | def preprocess(self, results): method pre_pipeline (line 62) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/aimotive.py class aiMotive (line 6) | class aiMotive(SequenceDataset): method __init__ (line 16) | def __init__( method pre_pipeline (line 46) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/argoverse.py class Argoverse (line 12) | class Argoverse(ImageDataset): method __init__ (line 21) | def __init__( method load_dataset (line 50) | def load_dataset(self): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/argoverse2.py class Argoverse2 (line 6) | class Argoverse2(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/arkit.py class ARKit (line 6) | class ARKit(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/ase.py class ASE (line 8) | class ASE(SequenceDataset): method __init__ (line 17) | def __init__( method preprocess (line 47) | def preprocess(self, results): method pre_pipeline (line 61) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/base_dataset.py class BaseDataset (line 19) | class BaseDataset(Dataset): method __init__ (line 23) | def __init__( method __len__ (line 95) | def __len__(self): method pack_batch (line 98) | def pack_batch(self, results): method unpack_batch (line 118) | def unpack_batch(self, results): method _augmentation_space (line 136) | def _augmentation_space(self): method augment (line 151) | def augment(self, results): method prepare_depth_eval (line 156) | def prepare_depth_eval(self, inputs, preds): method prepare_points_eval (line 169) | def prepare_points_eval(self, inputs, preds): method add_points (line 180) | def add_points(self, inputs): method accumulate_metrics (line 187) | def accumulate_metrics( method accumulate_metrics_depth (line 219) | def accumulate_metrics_depth(self, gts, preds, masks): method accumulate_metrics_3d (line 236) | def accumulate_metrics_3d(self, gts, preds, masks): method get_evaluation (line 258) | def get_evaluation(self, metrics=None): method replicate (line 273) | def replicate(self, results): method log_load_dataset (line 279) | def log_load_dataset(self): method pre_pipeline (line 284) | def pre_pipeline(self, results): method eval_mask (line 302) | def eval_mask(self, valid_mask): method chunk (line 305) | def chunk(self, dataset, chunk_dim=1, pct=1.0): method preprocess (line 314) | def preprocess(self, results): method postprocess (line 318) | def postprocess(self, results): method get_mapper (line 322) | def get_mapper(self): method get_intrinsics (line 326) | def get_intrinsics(self, idx, image_name): method get_extrinsics (line 330) | def get_extrinsics(self, idx, image_name): method load_dataset (line 334) | def load_dataset(self): method get_single_item (line 338) | def get_single_item(self, idx, sample=None, mapper=None): method __getitem__ (line 342) | def __getitem__(self, idx): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/bdd.py class BDD (line 12) | class BDD(ImageDataset): method __init__ (line 21) | def __init__( method load_dataset (line 47) | def load_dataset(self): method pre_pipeline (line 76) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/bedlam.py class BEDLAM (line 6) | class BEDLAM(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/behave.py class Behave (line 6) | class Behave(SequenceDataset): method __init__ (line 16) | def __init__( method pre_pipeline (line 46) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/blendedmvg.py class BlendedMVG (line 6) | class BlendedMVG(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/cityscape.py class Cityscape (line 12) | class Cityscape(ImageDataset): method __init__ (line 21) | def __init__( method load_dataset (line 50) | def load_dataset(self): method pre_pipeline (line 75) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/ddad.py class DDAD (line 12) | class DDAD(ImageDataset): method __init__ (line 21) | def __init__( method load_dataset (line 47) | def load_dataset(self): method get_mapper (line 72) | def get_mapper(self): method pre_pipeline (line 80) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/deep360.py class Deep360 (line 9) | class Deep360(SequenceDataset): method __init__ (line 18) | def __init__( method pre_pipeline (line 51) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/dense.py class DENSE (line 11) | class DENSE(ImageDataset): method __init__ (line 28) | def __init__( method load_dataset (line 56) | def load_dataset(self): method get_intrinsics (line 78) | def get_intrinsics(self, idx, image_name): method get_mapper (line 81) | def get_mapper(self): method pre_pipeline (line 87) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/diml.py class DIML (line 12) | class DIML(ImageDataset): method __init__ (line 21) | def __init__( method load_dataset (line 49) | def load_dataset(self): method pre_pipeline (line 75) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/diode.py class DiodeIndoor (line 12) | class DiodeIndoor(ImageDataset): method __init__ (line 23) | def __init__( method load_dataset (line 50) | def load_dataset(self): method get_intrinsics (line 75) | def get_intrinsics(self, *args, **kwargs): method get_mapper (line 78) | def get_mapper(self): method pre_pipeline (line 84) | def pre_pipeline(self, results): class DiodeIndoor_F (line 91) | class DiodeIndoor_F(SequenceDataset): method __init__ (line 100) | def __init__( method pre_pipeline (line 132) | def pre_pipeline(self, results): class DiodeOutdoor (line 139) | class DiodeOutdoor(ImageDataset): method __init__ (line 151) | def __init__( method load_dataset (line 187) | def load_dataset(self): class Diode (line 210) | class Diode(ImageDataset): method __init__ (line 222) | def __init__( method load_dataset (line 256) | def load_dataset(self): method get_intrinsics (line 277) | def get_intrinsics(self, *args, **kwargs): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/dl3dv.py class DL3DV (line 6) | class DL3DV(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/driving_stereo.py class DrivingStereo (line 12) | class DrivingStereo(ImageDataset): method __init__ (line 21) | def __init__( method load_dataset (line 50) | def load_dataset(self): method pre_pipeline (line 78) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/dtu_rmvd.py class DTURMVD (line 16) | class DTURMVD(SequenceDataset): method __init__ (line 26) | def __init__( method pre_pipeline (line 57) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/dummy.py class Dummy (line 6) | class Dummy(Dataset): method __init__ (line 10) | def __init__(self, *args, **kwargs): method get_single_item (line 14) | def get_single_item(self, idx): method __getitem__ (line 25) | def __getitem__(self, idx): method __len__ (line 32) | def __len__(self): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/dynamic_replica.py class DynReplica (line 6) | class DynReplica(SequenceDataset): method __init__ (line 16) | def __init__( method pre_pipeline (line 46) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/eden.py class EDEN (line 6) | class EDEN(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/eth3d.py class ETH3D (line 13) | class ETH3D(ImageDataset): method __init__ (line 22) | def __init__( method load_dataset (line 48) | def load_dataset(self): method pre_pipeline (line 70) | def pre_pipeline(self, results): class ETH3D_F (line 77) | class ETH3D_F(SequenceDataset): method __init__ (line 86) | def __init__( method pre_pipeline (line 118) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/eth3d_rmvd.py class ETH3DRMVD (line 16) | class ETH3DRMVD(SequenceDataset): method __init__ (line 26) | def __init__( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/facedepth.py class FaceDepth (line 6) | class FaceDepth(SequenceDataset): method __init__ (line 16) | def __init__( method pre_pipeline (line 46) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/flsea.py class FLSea (line 11) | class FLSea(ImageDataset): method __init__ (line 34) | def __init__( method load_dataset (line 63) | def load_dataset(self): method get_intrinsics (line 87) | def get_intrinsics(self, idx, image_name): method get_mapper (line 90) | def get_mapper(self): method pre_pipeline (line 96) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/futurehouse.py class FutureHouse (line 9) | class FutureHouse(SequenceDataset): method __init__ (line 18) | def __init__( method pre_pipeline (line 51) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/gibson.py class Gibson (line 9) | class Gibson(SequenceDataset): method __init__ (line 18) | def __init__( method pre_pipeline (line 51) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/hammer.py class HAMMER (line 12) | class HAMMER(ImageDataset): method __init__ (line 21) | def __init__( method load_dataset (line 50) | def load_dataset(self): method pre_pipeline (line 72) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/hm3d.py class HM3D (line 6) | class HM3D(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/hoi4d.py class HOI4D (line 6) | class HOI4D(SequenceDataset): method __init__ (line 16) | def __init__( method pre_pipeline (line 46) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/hrwsi.py class HRWSI (line 11) | class HRWSI(ImageDataset): method __init__ (line 19) | def __init__( method load_dataset (line 45) | def load_dataset(self): method pre_pipeline (line 73) | def pre_pipeline(self, results): method get_mapper (line 79) | def get_mapper(self): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/hypersim.py class HyperSim (line 12) | class HyperSim(ImageDataset): method __init__ (line 21) | def __init__( method load_dataset (line 48) | def load_dataset(self): method get_mapper (line 84) | def get_mapper(self): method pre_pipeline (line 92) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/ibims.py class IBims (line 13) | class IBims(ImageDataset): method __init__ (line 22) | def __init__( method load_dataset (line 51) | def load_dataset(self): method pre_pipeline (line 73) | def pre_pipeline(self, results): class IBims_F (line 80) | class IBims_F(SequenceDataset): method __init__ (line 89) | def __init__( method pre_pipeline (line 121) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/image_dataset.py class ImageDataset (line 23) | class ImageDataset(BaseDataset): method __init__ (line 24) | def __init__( method get_single_item (line 49) | def get_single_item(self, idx, sample=None, mapper=None): method preprocess (line 132) | def preprocess(self, results): method postprocess (line 151) | def postprocess(self, results): method __getitem__ (line 161) | def __getitem__(self, idx): method get_intrinsics (line 173) | def get_intrinsics(self, idx, image_name): method get_extrinsics (line 180) | def get_extrinsics(self, idx, image_name): method get_mapper (line 187) | def get_mapper(self): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/ken_burns.py class KenBurns (line 12) | class KenBurns(ImageDataset): method __init__ (line 21) | def __init__( method load_dataset (line 48) | def load_dataset(self): method get_mapper (line 82) | def get_mapper(self): method pre_pipeline (line 90) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/kitti.py class KITTI (line 13) | class KITTI(ImageDataset): method __init__ (line 61) | def __init__( method load_dataset (line 95) | def load_dataset(self): method get_intrinsics (line 124) | def get_intrinsics(self, idx, image_name): method preprocess (line 127) | def preprocess(self, results): method eval_mask (line 147) | def eval_mask(self, valid_mask, info={}): method get_mapper (line 165) | def get_mapper(self): method pre_pipeline (line 171) | def pre_pipeline(self, results): class KITTIBenchmark (line 181) | class KITTIBenchmark(ImageDataset): method __init__ (line 190) | def __init__( method load_dataset (line 226) | def load_dataset(self): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/kitti360.py class KITTI360 (line 8) | class KITTI360(SequenceDataset): method __init__ (line 17) | def __init__( method preprocess (line 49) | def preprocess(self, results): method pre_pipeline (line 61) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/kitti_multi.py class KITTIMulti (line 16) | class KITTIMulti(SequenceDataset): method __init__ (line 26) | def __init__( method __len__ (line 69) | def __len__(self): method preprocess (line 74) | def preprocess(self, results): method eval_mask (line 86) | def eval_mask(self, valid_mask, info={}): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/kitti_rmvd.py class KITTIRMVD (line 14) | class KITTIRMVD(SequenceDataset): method __init__ (line 24) | def __init__( method eval_mask (line 57) | def eval_mask(self, valid_mask, info={}): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/lyft.py class Lyft (line 12) | class Lyft(ImageDataset): method __init__ (line 21) | def __init__( method load_dataset (line 47) | def load_dataset(self): method pre_pipeline (line 81) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/mapillary.py class Mapillary (line 12) | class Mapillary(ImageDataset): method __init__ (line 21) | def __init__( method load_dataset (line 50) | def load_dataset(self): method pre_pipeline (line 78) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/matrix_city.py class MatrixCity (line 6) | class MatrixCity(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/matterport3d.py class Matterport3D (line 9) | class Matterport3D(SequenceDataset): method __init__ (line 18) | def __init__( method pre_pipeline (line 51) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/megadepth.py class MegaDepth (line 10) | class MegaDepth(ImageDataset): method __init__ (line 18) | def __init__( method load_dataset (line 44) | def load_dataset(self): method pre_pipeline (line 72) | def pre_pipeline(self, results): method get_mapper (line 79) | def get_mapper(self): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/megadepth_s.py class MegaDepthS (line 6) | class MegaDepthS(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/midair.py class MidAir (line 6) | class MidAir(SequenceDataset): method __init__ (line 16) | def __init__( method pre_pipeline (line 46) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/mip.py class MIP (line 6) | class MIP(SequenceDataset): method __init__ (line 16) | def __init__( method pre_pipeline (line 46) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/ms2.py class MS2 (line 6) | class MS2(SequenceDataset): method __init__ (line 16) | def __init__( method pre_pipeline (line 46) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/mvimgnet.py class MVImgNet (line 91) | class MVImgNet(SequenceDataset): method __init__ (line 102) | def __init__( method pre_pipeline (line 132) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/mvsynth.py class MVSynth (line 6) | class MVSynth(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/nerds360.py class NeRDS360 (line 6) | class NeRDS360(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/niantic_mapfree.py class NianticMapFree (line 6) | class NianticMapFree(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/nuscenes.py class Nuscenes (line 12) | class Nuscenes(ImageDataset): method __init__ (line 22) | def __init__( method load_dataset (line 49) | def load_dataset(self): method get_mapper (line 77) | def get_mapper(self): method pre_pipeline (line 85) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/nyuv2.py class NYUv2Depth (line 13) | class NYUv2Depth(ImageDataset): method __init__ (line 32) | def __init__( method load_dataset (line 64) | def load_dataset(self): method pre_pipeline (line 89) | def pre_pipeline(self, results): method get_intrinsics (line 94) | def get_intrinsics(self, idx, image_name): method eval_mask (line 97) | def eval_mask(self, valid_mask, info={}): method get_mapper (line 102) | def get_mapper(self): method pre_pipeline (line 108) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/oasis.py class OASISv2 (line 11) | class OASISv2(ImageDataset): method __init__ (line 19) | def __init__( method load_dataset (line 46) | def load_dataset(self): method pre_pipeline (line 70) | def pre_pipeline(self, results): method get_mapper (line 76) | def get_mapper(self): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/pipelines/formating.py class Collect (line 7) | class Collect(object): method __init__ (line 8) | def __init__( method __call__ (line 41) | def __call__(self, results): method __repr__ (line 55) | def __repr__(self): class AnnotationMask (line 61) | class AnnotationMask(object): method __init__ (line 62) | def __init__(self, min_value, max_value, custom_fn=lambda x: x): method __call__ (line 67) | def __call__(self, results): method __repr__ (line 91) | def __repr__(self): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/pipelines/transforms.py class PanoCrop (line 15) | class PanoCrop: method __init__ (line 16) | def __init__(self, crop_v=0.1): method _crop_data (line 19) | def _crop_data(self, results, crop_size): method __call__ (line 50) | def __call__(self, results): class PanoRoll (line 58) | class PanoRoll: method __init__ (line 59) | def __init__(self, roll=[-0.5, 0.5]): method __call__ (line 62) | def __call__(self, results): class RandomFlip (line 76) | class RandomFlip: method __init__ (line 90) | def __init__(self, direction="horizontal", prob=0.5, **kwargs): method __call__ (line 101) | def __call__(self, results): method __repr__ (line 177) | def __repr__(self): class Crop (line 184) | class Crop: method __init__ (line 185) | def __init__( method _get_crop_size (line 210) | def _get_crop_size(self, image_shape): method _crop_data (line 231) | def _crop_data(self, results, crop_size): method __call__ (line 258) | def __call__(self, results): method __repr__ (line 265) | def __repr__(self): class KittiCrop (line 272) | class KittiCrop: method __init__ (line 273) | def __init__(self, crop_size): method _crop_data (line 276) | def _crop_data(self, results, crop_size): method __call__ (line 315) | def __call__(self, results): method __repr__ (line 330) | def __repr__(self): class RandomMasking (line 336) | class RandomMasking: method __init__ (line 337) | def __init__( method __call__ (line 357) | def __call__(self, results): method _nop (line 387) | def _nop(self, results, down_size, device): method random_sampling (line 394) | def random_sampling(self, results, mask_ratio, down_size, validity_mas... method inverse_sampling (line 400) | def inverse_sampling(self, results, mask_ratio, down_size, validity_ma... method guided_sampling (line 469) | def guided_sampling(self, results, mask_ratio, down_size, validity_mas... method step (line 487) | def step(self): method is_warmup (line 497) | def is_warmup(self): class Rotate (line 501) | class Rotate: method __init__ (line 502) | def __init__( method _rotate (line 525) | def _rotate(self, results, angle, center=None, fill_val=0.0): method __call__ (line 556) | def __call__(self, results): class RandomColor (line 570) | class RandomColor: method __init__ (line 571) | def __init__(self, level, prob=0.5): method _adjust_color_img (line 576) | def _adjust_color_img(self, results, factor=1.0): method __call__ (line 580) | def __call__(self, results): class RandomSaturation (line 592) | class RandomSaturation: method __init__ (line 593) | def __init__(self, level, prob=0.5): method _adjust_saturation_img (line 598) | def _adjust_saturation_img(self, results, factor=1.0): method __call__ (line 603) | def __call__(self, results): class RandomSharpness (line 615) | class RandomSharpness: method __init__ (line 616) | def __init__(self, level, prob=0.5): method _adjust_sharpeness_img (line 621) | def _adjust_sharpeness_img(self, results, factor=1.0): method __call__ (line 626) | def __call__(self, results): method __repr__ (line 637) | def __repr__(self): class RandomSolarize (line 644) | class RandomSolarize: method __init__ (line 645) | def __init__(self, level, prob=0.5): method _adjust_solarize_img (line 650) | def _adjust_solarize_img(self, results, factor=255.0): method __call__ (line 654) | def __call__(self, results): class RandomPosterize (line 666) | class RandomPosterize: method __init__ (line 667) | def __init__(self, level, prob=0.5): method _posterize_img (line 672) | def _posterize_img(self, results, factor=1.0): method __call__ (line 676) | def __call__(self, results): class RandomEqualize (line 688) | class RandomEqualize: method __init__ (line 689) | def __init__(self, prob=0.5): method _imequalize (line 693) | def _imequalize(self, results): method __call__ (line 697) | def __call__(self, results): class RandomBrightness (line 704) | class RandomBrightness: method __init__ (line 705) | def __init__(self, level, prob=0.5): method _adjust_brightness_img (line 710) | def _adjust_brightness_img(self, results, factor=1.0): method __call__ (line 714) | def __call__(self, results, level=None): class RandomContrast (line 726) | class RandomContrast: method __init__ (line 727) | def __init__(self, level, prob=0.5): method _adjust_contrast_img (line 732) | def _adjust_contrast_img(self, results, factor=1.0): method __call__ (line 736) | def __call__(self, results, level=None): class RandomGamma (line 748) | class RandomGamma: method __init__ (line 749) | def __init__(self, level, prob=0.5): method __call__ (line 754) | def __call__(self, results, level=None): class RandomInvert (line 764) | class RandomInvert: method __init__ (line 765) | def __init__(self, prob=0.5): method __call__ (line 768) | def __call__(self, results): class RandomAutoContrast (line 777) | class RandomAutoContrast: method __init__ (line 778) | def __init__(self, prob=0.5): method _autocontrast_img (line 781) | def _autocontrast_img(self, results): method __call__ (line 786) | def __call__(self, results): class RandomShear (line 793) | class RandomShear(object): method __init__ (line 794) | def __init__( method _shear_img (line 805) | def _shear_img(self, results, magnitude): method _shear_masks (line 818) | def _shear_masks(self, results, magnitude): method _shear_gt (line 831) | def _shear_gt( method __call__ (line 848) | def __call__(self, results): class RandomTranslate (line 866) | class RandomTranslate(object): method __init__ (line 867) | def __init__( method _translate_img (line 877) | def _translate_img(self, results, magnitude): method _translate_mask (line 899) | def _translate_mask(self, results, magnitude): method _translate_gt (line 913) | def _translate_gt( method __call__ (line 931) | def __call__(self, results): method __repr__ (line 955) | def __repr__(self): class RandomColorJitter (line 963) | class RandomColorJitter: method __init__ (line 964) | def __init__(self, level, prob=0.9): method _adjust_contrast_img (line 975) | def _adjust_contrast_img(self, results, factor=1.0): method _adjust_sharpness_img (line 982) | def _adjust_sharpness_img(self, results, factor=1.0): method _adjust_brightness_img (line 989) | def _adjust_brightness_img(self, results, factor=1.0): method _adjust_saturation_img (line 996) | def _adjust_saturation_img(self, results, factor=1.0): method _adjust_color_img (line 1003) | def _adjust_color_img(self, results, factor=1.0): method __call__ (line 1010) | def __call__(self, results): class RandomGrayscale (line 1028) | class RandomGrayscale: method __init__ (line 1029) | def __init__(self, prob=0.1, num_output_channels=3): method __call__ (line 1034) | def __call__(self, results): function masked_nearest_interpolation (line 1046) | def masked_nearest_interpolation(input, mask, target_size): class ContextCrop (line 1131) | class ContextCrop: method __init__ (line 1132) | def __init__( method _transform_img (line 1156) | def _transform_img(self, results, shapes): method _transform_masks (line 1167) | def _transform_masks(self, results, shapes): method _transform_gt (line 1175) | def _transform_gt(self, results, shapes): method crop (line 1182) | def crop(img, height, width, top, left) -> torch.Tensor: method test_closest_shape (line 1195) | def test_closest_shape(self, image_shape): method _get_crop_shapes (line 1209) | def _get_crop_shapes(self, image_shape, ctx=None): method __call__ (line 1246) | def __call__(self, results): class RandomFiller (line 1358) | class RandomFiller: method __init__ (line 1359) | def __init__(self, *args, **kwargs): method _transform (line 1362) | def _transform(self, results): method __call__ (line 1388) | def __call__(self, results): class GaussianBlur (line 1404) | class GaussianBlur: method __init__ (line 1405) | def __init__(self, kernel_size, sigma=(0.1, 2.0), prob=0.9): method apply (line 1412) | def apply(self, x, kernel): method _create_kernel (line 1420) | def _create_kernel(self, sigma): method __call__ (line 1434) | def __call__(self, results): class Compose (line 1445) | class Compose: method __init__ (line 1446) | def __init__(self, transforms): method __call__ (line 1449) | def __call__(self, results): method __setattr__ (line 1454) | def __setattr__(self, name: str, value) -> None: method __repr__ (line 1459) | def __repr__(self): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/point_odyssey.py class PointOdyssey (line 6) | class PointOdyssey(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/proteus.py class Proteus (line 6) | class Proteus(SequenceDataset): method __init__ (line 16) | def __init__( method pre_pipeline (line 46) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/samplers copy.py function _get_numpy_dtype (line 13) | def _get_numpy_dtype(size: int) -> Any: function _get_torch_dtype (line 17) | def _get_torch_dtype(size: int) -> Any: function _generate_randperm_indices (line 21) | def _generate_randperm_indices(*, size: int, generator: torch.Generator): function _shuffle_tensor_slice (line 38) | def _shuffle_tensor_slice( function _new_shuffle_tensor_slice (line 63) | def _new_shuffle_tensor_slice( function _make_seed (line 77) | def _make_seed(seed: int, start: int, iter_count: int) -> int: class ShardedInfiniteSampler (line 82) | class ShardedInfiniteSampler(Sampler): method __init__ (line 83) | def __init__( method __iter__ (line 108) | def __iter__(self): method _iterator (line 121) | def _iterator(self): method _shuffled_iterator (line 128) | def _shuffled_iterator(self): class DistributedSamplerNoDuplicate (line 152) | class DistributedSamplerNoDuplicate(torch.utils.data.DistributedSampler): method __init__ (line 155) | def __init__(self, *args, **kwargs): class DatasetFromSampler (line 164) | class DatasetFromSampler(torch.utils.data.Dataset): method __init__ (line 171) | def __init__(self, sampler: Sampler): method __getitem__ (line 176) | def __getitem__(self, index: int): method __len__ (line 189) | def __len__(self) -> int: class DistributedSamplerWrapper (line 197) | class DistributedSamplerWrapper(torch.utils.data.DistributedSampler): method __init__ (line 212) | def __init__( method __iter__ (line 238) | def __iter__(self): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/samplers.py class DistributedSamplerNoDuplicate (line 4) | class DistributedSamplerNoDuplicate(torch.utils.data.DistributedSampler): method __init__ (line 7) | def __init__(self, *args, **kwargs): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/scannet.py class ScanNet (line 6) | class ScanNet(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/scannetpp.py class ScanNetpp (line 6) | class ScanNetpp(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): class ScanNetpp_F (line 52) | class ScanNetpp_F(SequenceDataset): method __init__ (line 62) | def __init__( method pre_pipeline (line 94) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/sequence_dataset.py class SequenceDataset (line 21) | class SequenceDataset(BaseDataset): method __init__ (line 36) | def __init__( method load_dataset (line 73) | def load_dataset(self): method get_random_idxs (line 113) | def get_random_idxs(self, num_samples_sequence): method get_test_idxs (line 155) | def get_test_idxs(self, num_samples_sequence, keyframe_idx): method get_single_sequence (line 183) | def get_single_sequence(self, idx): method preprocess (line 244) | def preprocess(self, results): method postprocess (line 269) | def postprocess(self, results): method __getitem__ (line 279) | def __getitem__(self, idx): method log_load_dataset (line 291) | def log_load_dataset(self): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/sintel copy.py class Sintel (line 6) | class Sintel(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/sintel.py class Sintel (line 6) | class Sintel(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/sunrgbd.py class SUNRGBD (line 12) | class SUNRGBD(ImageDataset): method __init__ (line 21) | def __init__( method load_dataset (line 50) | def load_dataset(self): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/synscapes.py class Synscapes (line 6) | class Synscapes(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/tartanair.py class TartanAir (line 6) | class TartanAir(SequenceDataset): method __init__ (line 16) | def __init__( method pre_pipeline (line 46) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/taskonomy.py class Taskonomy (line 12) | class Taskonomy(ImageDataset): method __init__ (line 21) | def __init__( method load_dataset (line 47) | def load_dataset(self): method get_mapper (line 80) | def get_mapper(self): method pre_pipeline (line 87) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/tat_rmvd.py class TATRMVD (line 17) | class TATRMVD(SequenceDataset): method __init__ (line 27) | def __init__( method pre_pipeline (line 58) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/theo.py class Theo (line 8) | class Theo(SequenceDataset): method __init__ (line 18) | def __init__( method preprocess (line 48) | def preprocess(self, results): method pre_pipeline (line 61) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/unrealstereo4k.py class UnrealStereo4K (line 6) | class UnrealStereo4K(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/urbansyn.py class UrbanSyn (line 6) | class UrbanSyn(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/utils.py class ConcatDataset (line 16) | class ConcatDataset(torch.utils.data.ConcatDataset): method __init__ (line 17) | def __init__(self, datasets, shape_constraints: dict[str, list[int]] =... method sample_shape (line 29) | def sample_shape(self): method __getitem__ (line 51) | def __getitem__(self, idxs): function _paddings (line 56) | def _paddings(image_shape, network_shape): function collate_fn (line 64) | def collate_fn(in_data: List[List[Dict[str, Any]]], is_batched: bool = T... function local_scatter (line 104) | def local_scatter(array: list[Any]): class DatasetFromList (line 123) | class DatasetFromList(torch.utils.data.Dataset): # type: ignore method __init__ (line 131) | def __init__(self, lst: List[Any], deepcopy: bool = False, serialize: ... method __len__ (line 175) | def __len__(self) -> int: method __getitem__ (line 181) | def __getitem__(self, idx: int) -> Any: function get_weights (line 194) | def get_weights( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/utils_decode.py function decode_depth (line 14) | def decode_depth(results, h5file, value, idx, depth_scale, name="depth",... function decode_numpy (line 42) | def decode_numpy(results, h5file, value, idx, name="points", **kwargs): function decode_tensor (line 54) | def decode_tensor(results, value, idx, name, **kwargs): function decode_camera (line 61) | def decode_camera(results, value, idx, name, sample, j, **kwargs): function decode_K (line 69) | def decode_K(results, value, idx, name, **kwargs): function decode_mask (line 77) | def decode_mask(results, h5file, value, idx, name, **kwargs): function decode_rgb (line 87) | def decode_rgb(results, h5file, value, idx, name="image", **kwargs): function decode_flow (line 101) | def decode_flow(results, h5file, value, idx, name, **kwargs): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/vkitti.py class VKITTI (line 6) | class VKITTI(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/void.py class VOID (line 12) | class VOID(ImageDataset): method __init__ (line 21) | def __init__( method load_dataset (line 50) | def load_dataset(self): method pre_pipeline (line 75) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/waymo.py class Waymo (line 6) | class Waymo(SequenceDataset): method __init__ (line 15) | def __init__( method pre_pipeline (line 45) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/datasets/wildrgbd.py class WildRGBD (line 6) | class WildRGBD(SequenceDataset): method __init__ (line 16) | def __init__( method pre_pipeline (line 46) | def pre_pipeline(self, results): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/layers/activation.py class SwiGLU (line 6) | class SwiGLU(nn.Module): method forward (line 7) | def forward(self, x: torch.Tensor) -> torch.Tensor: class GEGLU (line 12) | class GEGLU(nn.Module): method forward (line 13) | def forward(self, x: torch.Tensor) -> torch.Tensor: FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/layers/attention.py class SimpleAttention (line 17) | class SimpleAttention(nn.Module): method __init__ (line 18) | def __init__( method forward (line 39) | def forward( class AttentionBlock (line 81) | class AttentionBlock(nn.Module): method __init__ (line 82) | def __init__( method attn (line 109) | def attn( method forward (line 143) | def forward( class AttentionLayer (line 168) | class AttentionLayer(nn.Module): method __init__ (line 169) | def __init__( method forward (line 200) | def forward( class AttentionDecoderBlock (line 219) | class AttentionDecoderBlock(nn.Module): method __init__ (line 220) | def __init__( method cross_attn (line 253) | def cross_attn( method self_attn (line 292) | def self_attn( method forward (line 321) | def forward( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/layers/convnext.py class CvnxtBlock (line 5) | class CvnxtBlock(nn.Module): method __init__ (line 6) | def __init__( method forward (line 33) | def forward(self, x): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/layers/drop_path.py function drop_path (line 5) | def drop_path(x: torch.Tensor, drop_prob: float = 0.0, training: bool = ... class DropPath (line 19) | class DropPath(nn.Module): method __init__ (line 20) | def __init__(self, drop_prob=None): method forward (line 24) | def forward(self, x): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/layers/layer_scale.py class LayerScale (line 5) | class LayerScale(nn.Module): method __init__ (line 6) | def __init__( method forward (line 16) | def forward(self, x: torch.Tensor) -> torch.Tensor: FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/layers/mlp.py class MLP (line 9) | class MLP(nn.Module): method __init__ (line 10) | def __init__( method forward (line 29) | def forward(self, x: torch.Tensor) -> torch.Tensor: FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/layers/nystrom.py class NystromSelfAttentionConfig (line 29) | class NystromSelfAttentionConfig(AttentionConfig): class AvgPool (line 62) | class AvgPool(nn.Module): method __init__ (line 63) | def __init__(self, n: int): method forward (line 67) | def forward(self, x: torch.Tensor): class NystromAttention (line 100) | class NystromAttention(Attention): method __init__ (line 102) | def __init__( method forward (line 166) | def forward( method _triu_mask (line 286) | def _triu_mask(self, dim_1: int, dim_2: int, dim_3: int, **kwargs) -> ... FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/layers/nystrom_attention.py class NystromBlock (line 12) | class NystromBlock(AttentionBlock): method __init__ (line 13) | def __init__( method attn (line 38) | def attn( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/layers/positional_encoding.py class PositionEmbeddingSine (line 14) | class PositionEmbeddingSine(nn.Module): method __init__ (line 15) | def __init__( method forward (line 28) | def forward( method __repr__ (line 59) | def __repr__(self, _repr_indent=4): class LearnedSinusoidalPosEmb (line 72) | class LearnedSinusoidalPosEmb(nn.Module): method __init__ (line 73) | def __init__(self, dim): method forward (line 79) | def forward(self, x): function generate_fourier_features (line 87) | def generate_fourier_features(x, max_freq=64, num_bands=16): function broadcat (line 102) | def broadcat(tensors, dim=-1): function rotate_half (line 121) | def rotate_half(x): class VisionRotaryEmbedding (line 128) | class VisionRotaryEmbedding(nn.Module): method __init__ (line 129) | def __init__( method forward (line 171) | def forward(self, t, start_index=0): class VisionRotaryEmbeddingFast (line 186) | class VisionRotaryEmbeddingFast(nn.Module): method __init__ (line 187) | def __init__( method forward (line 226) | def forward(self, t): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/layers/upsample.py class ConvUpsample (line 13) | class ConvUpsample(nn.Module): method __init__ (line 14) | def __init__( method forward (line 40) | def forward(self, x: torch.Tensor): class ConvUpsampleShuffle (line 48) | class ConvUpsampleShuffle(nn.Module): method __init__ (line 49) | def __init__( method forward (line 74) | def forward(self, x: torch.Tensor): class ConvUpsampleShuffleResidual (line 82) | class ConvUpsampleShuffleResidual(nn.Module): method __init__ (line 83) | def __init__( method forward (line 129) | def forward(self, x: torch.Tensor): class ResidualConvUnit (line 137) | class ResidualConvUnit(nn.Module): method __init__ (line 138) | def __init__( method forward (line 173) | def forward(self, x): class ResUpsampleBil (line 183) | class ResUpsampleBil(nn.Module): method __init__ (line 184) | def __init__( method forward (line 219) | def forward(self, x: torch.Tensor): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/backbones/convnext.py function get_num_layer_for_convnext (line 14) | def get_num_layer_for_convnext(var_name): function get_parameter_groups (line 46) | def get_parameter_groups(model, lr, wd=1e-5, ld=0.9, skip_list=None): class Downsample (line 105) | class Downsample(nn.Module): method __init__ (line 106) | def __init__(self, in_chs, out_chs, stride=1, dilation=1): method forward (line 124) | def forward(self, x): class ConvNeXtBlock (line 130) | class ConvNeXtBlock(nn.Module): method __init__ (line 141) | def __init__( method forward (line 208) | def forward(self, x): class ConvNeXtStage (line 226) | class ConvNeXtStage(nn.Module): method __init__ (line 227) | def __init__( method forward (line 289) | def forward(self, x): class ConvNeXt (line 301) | class ConvNeXt(nn.Module): method __init__ (line 302) | def __init__( method _init_weights (line 450) | def _init_weights(self, module): method forward (line 459) | def forward(self, x, masks=None): method group_matcher (line 474) | def group_matcher(self, coarse=False): method set_grad_checkpointing (line 489) | def set_grad_checkpointing(self, enable=True): method freeze (line 493) | def freeze(self) -> None: method get_params (line 499) | def get_params(self, lr, wd, ld, *args, **kwargs): method no_weight_decay (line 503) | def no_weight_decay(self): method build (line 507) | def build(cls, config): function checkpoint_filter_fn (line 512) | def checkpoint_filter_fn(state_dict, model): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/backbones/convnext2.py function get_num_layer_for_convnext_single (line 7) | def get_num_layer_for_convnext_single(var_name, depths): function get_num_layer_for_convnext (line 26) | def get_num_layer_for_convnext(var_name): function get_parameter_groups (line 57) | def get_parameter_groups(model, lr, wd=1e-5, ld=0.9, skip_list=()): class LayerNorm (line 112) | class LayerNorm(nn.Module): method __init__ (line 119) | def __init__(self, normalized_shape, eps=1e-6, data_format="channels_l... method forward (line 129) | def forward(self, x): class GRN (line 142) | class GRN(nn.Module): method __init__ (line 145) | def __init__(self, dim): method forward (line 150) | def forward(self, x): class Block (line 156) | class Block(nn.Module): method __init__ (line 164) | def __init__(self, dim, drop_path=0.0, mult=4, use_checkpoint=False): method forward (line 179) | def forward(self, x): class ConvNeXtV2 (line 194) | class ConvNeXtV2(nn.Module): method __init__ (line 206) | def __init__( method _init_weights (line 260) | def _init_weights(self, m): method forward (line 265) | def forward(self, x): method get_params (line 275) | def get_params(self, lr, wd, ld, *args, **kwargs): method freeze (line 279) | def freeze(self) -> None: method build (line 286) | def build(cls, config): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/backbones/dinov2.py function named_apply (line 19) | def named_apply( function get_parameter_groups (line 38) | def get_parameter_groups(model, lr, wd=1e-5, ld=0.9, skip_list=()): class BlockChunk (line 108) | class BlockChunk(nn.ModuleList): method forward (line 109) | def forward(self, x): class DinoVisionTransformer (line 115) | class DinoVisionTransformer(nn.Module): method __init__ (line 116) | def __init__( method init_weights (line 260) | def init_weights(self): method interpolate_pos_encoding (line 267) | def interpolate_pos_encoding(self, x, w, h): method prepare_tokens_with_masks (line 306) | def prepare_tokens_with_masks(self, x, masks=None): method forward (line 324) | def forward(self, x, masks=None): method get_params (line 349) | def get_params(self, lr, wd, ld, *args, **kwargs): method freeze (line 353) | def freeze(self) -> None: method train (line 359) | def train(self, mode=True): function init_weights_vit_timm (line 380) | def init_weights_vit_timm(module: nn.Module, name: str = ""): function vit_small (line 388) | def vit_small(patch_size=16, num_register_tokens=0, export=False, **kwar... function vit_base (line 402) | def vit_base(patch_size=16, num_register_tokens=0, export=False, **kwargs): function vit_large (line 416) | def vit_large(patch_size=16, num_register_tokens=0, export=False, **kwar... function _make_dinov2_model_name (line 430) | def _make_dinov2_model_name(arch_name: str, patch_size: int) -> str: function _make_dinov2_model (line 435) | def _make_dinov2_model( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/backbones/metadinov2/attention.py class Attention (line 31) | class Attention(nn.Module): method __init__ (line 32) | def __init__( method forward (line 51) | def forward(self, x: torch.Tensor) -> torch.Tensor: class MemEffAttention (line 65) | class MemEffAttention(Attention): method forward (line 66) | def forward(self, x: torch.Tensor, attn_bias=None) -> torch.Tensor: FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/backbones/metadinov2/block.py class Block (line 34) | class Block(nn.Module): method __init__ (line 35) | def __init__( method forward (line 84) | def forward(self, x: torch.Tensor) -> torch.Tensor: function drop_add_residual_stochastic_depth (line 112) | def drop_add_residual_stochastic_depth( function get_branges_scales (line 138) | def get_branges_scales(x, sample_drop_ratio=0.0): function add_residual (line 146) | def add_residual(x, brange, residual, residual_scale_factor, scaling_vec... function get_attn_bias_and_cat (line 167) | def get_attn_bias_and_cat(x_list, branges=None): function drop_add_residual_stochastic_depth_list (line 197) | def drop_add_residual_stochastic_depth_list( class NestedTensorBlock (line 228) | class NestedTensorBlock(Block): method forward_nested (line 229) | def forward_nested(self, x_list: List[torch.Tensor]) -> List[torch.Ten... method forward (line 273) | def forward(self, x_or_x_list): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/backbones/metadinov2/dino_head.py class DINOHead (line 13) | class DINOHead(nn.Module): method __init__ (line 14) | def __init__( method _init_weights (line 38) | def _init_weights(self, m): method forward (line 44) | def forward(self, x): function _build_mlp (line 52) | def _build_mlp( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/backbones/metadinov2/drop_path.py function drop_path (line 15) | def drop_path(x, drop_prob: float = 0.0, training: bool = False): class DropPath (line 29) | class DropPath(nn.Module): method __init__ (line 32) | def __init__(self, drop_prob=None): method forward (line 36) | def forward(self, x): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/backbones/metadinov2/layer_scale.py class LayerScale (line 16) | class LayerScale(nn.Module): method __init__ (line 17) | def __init__( method forward (line 27) | def forward(self, x: Tensor) -> Tensor: FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/backbones/metadinov2/mlp.py class Mlp (line 17) | class Mlp(nn.Module): method __init__ (line 18) | def __init__( method forward (line 35) | def forward(self, x: Tensor) -> Tensor: FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/backbones/metadinov2/patch_embed.py function make_2tuple (line 17) | def make_2tuple(x): class PatchEmbed (line 26) | class PatchEmbed(nn.Module): method __init__ (line 38) | def __init__( method forward (line 71) | def forward(self, x: Tensor) -> Tensor: method flops (line 90) | def flops(self) -> float: FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/backbones/metadinov2/swiglu_ffn.py class SwiGLUFFN (line 13) | class SwiGLUFFN(nn.Module): method __init__ (line 14) | def __init__( method forward (line 29) | def forward(self, x: Tensor) -> Tensor: class SwiGLUFFNFused (line 45) | class SwiGLUFFNFused(SwiGLU): method __init__ (line 46) | def __init__( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/encoder.py class ModelWrap (line 7) | class ModelWrap(nn.Module): method __init__ (line 8) | def __init__(self, model) -> None: method forward (line 12) | def forward(self, x, *args, **kwargs): function convnextv2_base (line 20) | def convnextv2_base(config, **kwargs): function convnextv2_large (line 37) | def convnextv2_large(config, **kwargs): function convnextv2_large_mae (line 54) | def convnextv2_large_mae(config, **kwargs): function convnextv2_huge (line 71) | def convnextv2_huge(config, **kwargs): function convnextv2_huge_mae (line 88) | def convnextv2_huge_mae(config, **kwargs): function convnext_large_pt (line 105) | def convnext_large_pt(config, **kwargs): function convnext_large (line 127) | def convnext_large(config, **kwargs): function dinov2_vits14 (line 139) | def dinov2_vits14(config, pretrained: bool = True, **kwargs): function dinov2_vitb14 (line 158) | def dinov2_vitb14(config, pretrained: bool = True, **kwargs): function dinov2_vitl14 (line 177) | def dinov2_vitl14(config, pretrained: str = "", **kwargs): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/unidepthv1/decoder.py class ListAdapter (line 21) | class ListAdapter(nn.Module): method __init__ (line 22) | def __init__(self, input_dims: List[int], hidden_dim: int): method forward (line 33) | def forward(self, x: torch.Tensor, splits: torch.Tensor) -> torch.Tensor: class CameraHead (line 39) | class CameraHead(nn.Module): method __init__ (line 40) | def __init__( method forward (line 82) | def forward(self, features, cls_tokens, pos_embed) -> torch.Tensor: method set_shapes (line 106) | def set_shapes(self, shapes: Tuple[int, int]): class DepthHead (line 110) | class DepthHead(nn.Module): method __init__ (line 111) | def __init__( method set_original_shapes (line 189) | def set_original_shapes(self, shapes: Tuple[int, int]): method set_shapes (line 192) | def set_shapes(self, shapes: Tuple[int, int]): method forward (line 195) | def forward( class Decoder (line 303) | class Decoder(nn.Module): method __init__ (line 304) | def __init__( method _init_weights (line 316) | def _init_weights(self, m): method get_adapted_features (line 329) | def get_adapted_features(self, features_flat, splits): method run_camera (line 337) | def run_camera(self, cls_tokens, features, pos_embed, original_shapes,... method forward (line 364) | def forward(self, inputs, image_metas) -> torch.Tensor: method no_weight_decay_keywords (line 466) | def no_weight_decay_keywords(self): method build (line 469) | def build(self, config): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/unidepthv1/unidepthv1.py function _paddings (line 30) | def _paddings(image_shape, network_shape): function _shapes (line 38) | def _shapes(image_shape, network_shape): function _preprocess (line 49) | def _preprocess(rgbs, intrinsics, shapes, pads, ratio, output_shapes): function _postprocess (line 65) | def _postprocess(predictions, intrinsics, shapes, pads, ratio, original_... class UniDepthV1 (line 97) | class UniDepthV1( method __init__ (line 104) | def __init__( method forward_train (line 116) | def forward_train(self, inputs, image_metas): method forward_test (line 122) | def forward_test(self, inputs, image_metas): method forward (line 150) | def forward(self, inputs, image_metas): method encode_decode (line 156) | def encode_decode(self, inputs, image_metas): method compute_losses (line 235) | def compute_losses(self, outputs, inputs, image_metas): method infer (line 288) | def infer(self, rgbs: torch.Tensor, intrinsics=None, skip_camera=False): method load_pretrained (line 375) | def load_pretrained(self, model_file): method get_params (line 394) | def get_params(self, config): method device (line 413) | def device(self): method build (line 416) | def build(self, config): method build_losses (line 445) | def build_losses(self, config): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/unidepthv2/decoder.py function orthonormal_init (line 19) | def orthonormal_init(num_tokens, dims): class ListAdapter (line 35) | class ListAdapter(nn.Module): method __init__ (line 36) | def __init__(self, input_dims: list[int], hidden_dim: int): method forward (line 43) | def forward(self, xs: torch.Tensor) -> list[torch.Tensor]: class CameraHead (line 48) | class CameraHead(nn.Module): method __init__ (line 49) | def __init__( method fill_intrinsics (line 85) | def fill_intrinsics(self, x): method forward (line 101) | def forward(self, features, cls_tokens, pos_embed) -> torch.Tensor: method set_shapes (line 113) | def set_shapes(self, shapes: tuple[int, int]): class DepthHead (line 117) | class DepthHead(nn.Module): method __init__ (line 118) | def __init__( method set_original_shapes (line 228) | def set_original_shapes(self, shapes: tuple[int, int]): method set_shapes (line 231) | def set_shapes(self, shapes: tuple[int, int]): method embed_rays (line 234) | def embed_rays(self, rays): method condition (line 255) | def condition(self, feat, rays_embeddings): method process (line 262) | def process(self, features_list, rays_embeddings): method depth_proj (line 284) | def depth_proj(self, out_features): method confidence_proj (line 305) | def confidence_proj(self, out_features): method decode (line 315) | def decode(self, out_features): method forward (line 320) | def forward( class Decoder (line 336) | class Decoder(nn.Module): method __init__ (line 337) | def __init__( method _init_weights (line 346) | def _init_weights(self, m): method run_camera (line 361) | def run_camera(self, cls_tokens, features, pos_embed, original_shapes,... method forward (line 405) | def forward( method no_weight_decay_keywords (line 465) | def no_weight_decay_keywords(self): method build (line 468) | def build(self, config): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/unidepthv2/decoder_old.py class ListAdapter (line 13) | class ListAdapter(nn.Module): method __init__ (line 14) | def __init__(self, input_dims: list[int], hidden_dim: int): method forward (line 26) | def forward(self, x: torch.Tensor, splits: torch.Tensor) -> torch.Tensor: class CameraHead (line 32) | class CameraHead(nn.Module): method __init__ (line 33) | def __init__( method fill_intrinsics (line 55) | def fill_intrinsics(self, x): method forward (line 66) | def forward(self, features, cls_tokens, pos_embed) -> torch.Tensor: method set_shapes (line 85) | def set_shapes(self, shapes: tuple[int, int]): class GlobalHead (line 89) | class GlobalHead(nn.Module): method __init__ (line 90) | def __init__( method embed_rays (line 111) | def embed_rays(self, rays, shapes): method set_original_shapes (line 123) | def set_original_shapes(self, shapes: tuple[int, int]): method set_shapes (line 126) | def set_shapes(self, shapes: tuple[int, int]): method get_scaleshift (line 129) | def get_scaleshift(self, x): method forward (line 135) | def forward(self, features, cls_tokens, rays) -> torch.Tensor: class DepthHead (line 149) | class DepthHead(nn.Module): method __init__ (line 150) | def __init__( method set_original_shapes (line 231) | def set_original_shapes(self, shapes: tuple[int, int]): method set_shapes (line 234) | def set_shapes(self, shapes: tuple[int, int]): method embed_rays (line 237) | def embed_rays(self, rays, shapes): method project_rays (line 249) | def project_rays(self, rays, shapes): method decode_depth (line 257) | def decode_depth(self, latents_16, rays, shapes): method init_latents (line 308) | def init_latents(self, features, shapes): method forward (line 317) | def forward( class Decoder (line 346) | class Decoder(nn.Module): method __init__ (line 347) | def __init__( method _init_weights (line 355) | def _init_weights(self, m): method get_adapted_features (line 370) | def get_adapted_features(self, features_flat, splits): method run_camera (line 378) | def run_camera(self, cls_tokens, features, pos_embed, original_shapes,... method run_global (line 408) | def run_global(self, cls_tokens, features, rays): method forward (line 428) | def forward(self, inputs, image_metas) -> torch.Tensor: method no_weight_decay_keywords (line 525) | def no_weight_decay_keywords(self): method build (line 528) | def build(self, config): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/unidepthv2/export.py class UniDepthV2ONNX (line 18) | class UniDepthV2ONNX(UniDepthV2): method __init__ (line 19) | def __init__( method forward (line 27) | def forward(self, rgbs): class UniDepthV2ONNXcam (line 48) | class UniDepthV2ONNXcam(UniDepthV2): method __init__ (line 49) | def __init__( method forward (line 57) | def forward(self, rgbs, rays): function export (line 80) | def export(model, path, shape=(462, 630), with_camera=False): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/unidepthv2/unidepthv2.py function get_paddings (line 36) | def get_paddings(original_shape, aspect_ratio_range): function get_resize_factor (line 61) | def get_resize_factor(original_shape, pixels_range, shape_multiplier=14): function _postprocess (line 80) | def _postprocess(tensor, shapes, paddings, interpolation_mode="bilinear"): function _postprocess_intrinsics (line 92) | def _postprocess_intrinsics(K, resize_factors, paddings): class UniDepthV2 (line 111) | class UniDepthV2( method __init__ (line 118) | def __init__( method forward_train (line 129) | def forward_train(self, inputs, image_metas): method forward_test (line 134) | def forward_test(self, inputs, image_metas): method forward (line 162) | def forward(self, inputs, image_metas): method compute_losses (line 168) | def compute_losses(self, outputs, inputs, image_metas): method infer (line 241) | def infer( method encode_decode (line 341) | def encode_decode(self, inputs, image_metas=[]): method load_pretrained (line 381) | def load_pretrained(self, model_file): method get_params (line 396) | def get_params(self, config): method device (line 415) | def device(self): method build (line 418) | def build(self, config): method build_losses (line 462) | def build_losses(self, config): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/models/unidepthv2/unidepthv2_old.py function _check_ratio (line 33) | def _check_ratio(image_ratio, ratio_bounds): function _check_resolution (line 45) | def _check_resolution(shape_constraints, resolution_level): function _get_closes_num_pixels (line 70) | def _get_closes_num_pixels(image_shape, pixels_bounds): function _shapes (line 78) | def _shapes(image_shape, shape_constraints): function _preprocess (line 95) | def _preprocess(rgbs, intrinsics, shapes, ratio): function _postprocess (line 107) | def _postprocess(outs, ratio, original_shapes, mode="nearest-exact"): class UniDepthV2old (line 119) | class UniDepthV2old( method __init__ (line 126) | def __init__( method forward (line 134) | def forward(self, inputs, image_metas): method infer (line 198) | def infer(self, rgbs: torch.Tensor, intrinsics=None): method load_pretrained (line 284) | def load_pretrained(self, model_file): method device (line 303) | def device(self): method build (line 306) | def build(self, config): method build_losses (line 351) | def build_losses(self, config): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/extract_patches/functions/extract_patches.py class ExtractPatchesFunction (line 6) | class ExtractPatchesFunction(Function): method forward (line 8) | def forward(ctx, input, centers, h, w): method backward (line 15) | def backward(ctx, grad_output): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/extract_patches/modules/patch_extractor.py class RandomPatchExtractor (line 10) | class RandomPatchExtractor(nn.Module): method __init__ (line 11) | def __init__( method forward (line 16) | def forward( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/extract_patches/setup.py function get_extensions (line 11) | def get_extensions(): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/extract_patches/src/cpu/extract_patches_cpu.cpp function extract_patches_cpu_forward (line 6) | torch::Tensor extract_patches_cpu_forward( function extract_patches_cpu_backward (line 15) | std::vector extract_patches_cpu_backward( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/extract_patches/src/extract_patches.cpp function PYBIND11_MODULE (line 4) | PYBIND11_MODULE(TORCH_EXTENSION_NAME, m) { FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/extract_patches/test.py function extract_patches (line 5) | def extract_patches(input, centers, patch_size): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/knn/functions/knn.py class _knn_points (line 20) | class _knn_points(Function): method forward (line 27) | def forward( method backward (line 97) | def backward(ctx, grad_dists, grad_idx): function knn_points (line 113) | def knn_points( function knn_gather (line 199) | def knn_gather( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/knn/setup.py function get_extensions (line 11) | def get_extensions(): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/knn/src/knn_cpu.cpp function KNearestNeighborIdxCpu (line 13) | std::tuple KNearestNeighborIdxCpu( function KNearestNeighborBackwardCpu (line 75) | std::tuple KNearestNeighborBackwardCpu( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/knn/src/knn_ext.cpp function PYBIND11_MODULE (line 4) | PYBIND11_MODULE(TORCH_EXTENSION_NAME, m) { FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/losses/arel.py class ARel (line 7) | class ARel(nn.Module): method __init__ (line 8) | def __init__( method forward (line 24) | def forward( method build (line 40) | def build(cls, config): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/losses/confidence.py class Confidence (line 7) | class Confidence(nn.Module): method __init__ (line 8) | def __init__( method forward (line 25) | def forward( method build (line 56) | def build(cls, config): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/losses/distill.py class SelfDistill (line 9) | class SelfDistill(nn.Module): method __init__ (line 10) | def __init__(self, weight: float, output_fn: str = "sqrt", eps: float ... method forward (line 19) | def forward( method loss (line 121) | def loss( method build (line 133) | def build(cls, config): class TeacherDistill (line 141) | class TeacherDistill(nn.Module): method __init__ (line 142) | def __init__( method forward (line 161) | def forward( method build (line 215) | def build(cls, config): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/losses/dummy.py class Dummy (line 5) | class Dummy(nn.Module): method __init__ (line 6) | def __init__(self, *args, **kwargs): method forward (line 11) | def forward(self, dummy: torch.Tensor, *args, **kwargs) -> torch.Tensor: method build (line 15) | def build(cls, config): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/losses/local_ssi.py function sample_strong_edges (line 11) | def sample_strong_edges(edges_img, quantile=0.95, reshape=8): function extract_patches (line 44) | def extract_patches(tensor, sample_coords, patch_size: tuple[int, int] =... class LocalSSI (line 81) | class LocalSSI(nn.Module): method __init__ (line 82) | def __init__( method forward (line 111) | def forward( method build (line 200) | def build(cls, config): class EdgeGuidedLocalSSI (line 212) | class EdgeGuidedLocalSSI(nn.Module): method __init__ (line 213) | def __init__( method get_edge (line 254) | def get_edge(self, image, mask): method compute_sample_patch_error (line 273) | def compute_sample_patch_error( method compute_image_error (line 298) | def compute_image_error(self, input, target, mask, image_size): method forward (line 311) | def forward( method build (line 354) | def build(cls, config): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/losses/regression.py class Regression (line 7) | class Regression(nn.Module): method __init__ (line 8) | def __init__( method forward (line 32) | def forward( method build (line 52) | def build(cls, config): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/losses/silog.py class SILog (line 8) | class SILog(nn.Module): method __init__ (line 9) | def __init__( method forward (line 29) | def forward( method build (line 53) | def build(cls, config): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/losses/utils.py function masked_mean_var (line 30) | def masked_mean_var( function masked_mean (line 49) | def masked_mean(data: torch.Tensor, mask: torch.Tensor | None, dim: List... function masked_quantile (line 60) | def masked_quantile( function masked_median (line 121) | def masked_median(data: torch.Tensor, mask: torch.Tensor, dim: List[int]): function masked_median_mad (line 129) | def masked_median_mad(data: torch.Tensor, mask: torch.Tensor, dim: List[... function masked_weighted_mean_var (line 138) | def masked_weighted_mean_var( function ssi (line 161) | def ssi( function ind2sub (line 190) | def ind2sub(idx, cols): function sub2ind (line 196) | def sub2ind(r, c, cols): function l2 (line 201) | def l2(input_tensor: torch.Tensor, gamma: float = 1.0, *args, **kwargs) ... function l1 (line 205) | def l1(input_tensor: torch.Tensor, gamma: float = 1.0, *args, **kwargs) ... function charbonnier (line 209) | def charbonnier( function cauchy (line 215) | def cauchy( function geman_mcclure (line 221) | def geman_mcclure( function robust_loss (line 227) | def robust_loss( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/ops/scheduler.py class PlainCosineScheduler (line 6) | class PlainCosineScheduler(object): method __init__ (line 7) | def __init__( method get_scheduler (line 31) | def get_scheduler(self): method step (line 50) | def step(self): method __getitem__ (line 56) | def __getitem__(self, it): class CosineScheduler (line 61) | class CosineScheduler(object): method __init__ (line 62) | def __init__( method get_schedulers (line 88) | def get_schedulers(self, group): method step (line 109) | def step(self): method __getitem__ (line 117) | def __getitem__(self, it): method get (line 121) | def get(self): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/utils/camera.py function invert_pinhole (line 16) | def invert_pinhole(K): class Camera (line 30) | class Camera: method __init__ (line 37) | def __init__(self, params=None, K=None): method project (line 57) | def project(self, xyz): method unproject (line 60) | def unproject(self, uv): method get_projection_mask (line 63) | def get_projection_mask(self): method get_overlap_mask (line 66) | def get_overlap_mask(self): method reconstruct (line 69) | def reconstruct(self, depth): method resize (line 78) | def resize(self, factor): method to (line 83) | def to(self, device, non_blocking=False): method get_rays (line 88) | def get_rays(self, shapes, noisy=False): method get_pinhole_rays (line 94) | def get_pinhole_rays(self, shapes, noisy=False): method flip (line 100) | def flip(self, H, W, direction="horizontal"): method clone (line 112) | def clone(self): method crop (line 115) | def crop(self, left, top, right=None, bottom=None): method get_new_fov (line 123) | def get_new_fov(self, new_shape, original_shape): method mask_overlap_projection (line 132) | def mask_overlap_projection(self, projected): method _pad_params (line 156) | def _pad_params(self): method flatten_cameras (line 169) | def flatten_cameras(cameras): # -> list[Camera]: method _stack_or_cat_cameras (line 182) | def _stack_or_cat_cameras(cameras, func, **kwargs): method __torch_function__ (line 196) | def __torch_function__(cls, func, types, args=(), kwargs=None): method device (line 212) | def device(self): method hfov (line 217) | def hfov(self): method vfov (line 221) | def vfov(self): method max_fov (line 225) | def max_fov(self): class Pinhole (line 229) | class Pinhole(Camera): method __init__ (line 230) | def __init__(self, params=None, K=None): method project (line 239) | def project(self, pcd): method unproject (line 255) | def unproject(self, uv): method reconstruct (line 269) | def reconstruct(self, depth): class EUCM (line 276) | class EUCM(Camera): method __init__ (line 277) | def __init__(self, params): method project (line 281) | def project(self, xyz): method unproject (line 307) | def unproject(self, uv): class Spherical (line 331) | class Spherical(Camera): method __init__ (line 332) | def __init__(self, params): method resize (line 336) | def resize(self, factor): method crop (line 341) | def crop(self, left, top, right, bottom): method project (line 359) | def project(self, xyz): method unproject (line 371) | def unproject(self, uv): method reconstruct (line 388) | def reconstruct(self, depth): method get_new_fov (line 394) | def get_new_fov(self, new_shape, original_shape): method hfov (line 400) | def hfov(self): method vfov (line 404) | def vfov(self): method max_fov (line 408) | def max_fov(self): class OPENCV (line 412) | class OPENCV(Camera): method __init__ (line 413) | def __init__(self, params): method project (line 423) | def project(self, xyz): method unproject (line 496) | def unproject(self, uv, max_iters: int = 10): class Fisheye624 (line 697) | class Fisheye624(Camera): method __init__ (line 698) | def __init__(self, params): method project (line 705) | def project(self, xyz): method unproject (line 778) | def unproject(self, uv, max_iters: int = 10): class MEI (line 977) | class MEI(Camera): method __init__ (line 978) | def __init__(self, params): method unproject (line 985) | def unproject(self, uv, max_iters: int = 20): method project (line 1084) | def project(self, xyz): class BatchCamera (line 1145) | class BatchCamera(Camera): method __init__ (line 1151) | def __init__(self, params, K, original_class, cameras): method project (line 1158) | def project(self, points_3d): method unproject (line 1167) | def unproject(self, points_2d): method crop (line 1173) | def crop(self, left, top, right=None, bottom=None): method resize (line 1182) | def resize(self, ratio): method reconstruct (line 1186) | def reconstruct(self, depth): method get_projection_mask (line 1195) | def get_projection_mask(self): method to (line 1200) | def to(self, device, non_blocking=False): method reshape (line 1207) | def reshape(self, *shape): method get_new_fov (line 1223) | def get_new_fov(self, new_shape, original_shape): method squeeze (line 1229) | def squeeze(self, dim): method __getitem__ (line 1237) | def __getitem__(self, idx): method __setitem__ (line 1251) | def __setitem__(self, idx, value): method __len__ (line 1271) | def __len__(self): method from_camera (line 1275) | def from_camera(cls, camera): method is_perspective (line 1279) | def is_perspective(self): method is_spherical (line 1283) | def is_spherical(self): method is_eucm (line 1287) | def is_eucm(self): method is_fisheye (line 1291) | def is_fisheye(self): method is_pinhole (line 1295) | def is_pinhole(self): method hfov (line 1299) | def hfov(self): method vfov (line 1303) | def vfov(self): method max_fov (line 1307) | def max_fov(self): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/utils/chamfer_distance.py function _validate_chamfer_reduction_inputs (line 16) | def _validate_chamfer_reduction_inputs( function _handle_pointcloud_input (line 33) | def _handle_pointcloud_input( class ChamferDistance (line 59) | class ChamferDistance(torch.nn.Module): method forward (line 60) | def forward( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/utils/coordinate.py function coords_grid (line 4) | def coords_grid(b, h, w, homogeneous=False, device=None, noisy=False): function normalize_coords (line 23) | def normalize_coords(coords, h, w): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/utils/distributed.py function is_dist_avail_and_initialized (line 16) | def is_dist_avail_and_initialized(): function get_rank (line 24) | def get_rank(): function get_local_rank (line 30) | def get_local_rank() -> int: function get_local_size (line 41) | def get_local_size() -> int: function get_world_size (line 53) | def get_world_size(): function barrier (line 59) | def barrier(): function is_main_process (line 65) | def is_main_process(): function is_rank_zero (line 69) | def is_rank_zero(args): function get_dist_info (line 73) | def get_dist_info(): function setup_multi_processes (line 83) | def setup_multi_processes(cfg): function setup_slurm (line 126) | def setup_slurm(backend: str, port: str) -> None: function sync_tensor_across_gpus (line 153) | def sync_tensor_across_gpus(t, dim=0, cat=True): function sync_string_across_gpus (line 179) | def sync_string_across_gpus(keys: list[str], device, dim=0): function create_local_process_group (line 195) | def create_local_process_group() -> None: function _get_global_gloo_group (line 211) | def _get_global_gloo_group(): function all_gather (line 218) | def all_gather(data, group=None): function local_broadcast_process_authkey (line 234) | def local_broadcast_process_authkey(): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/utils/ema_torch.py class DummyExponentialMovingAverage (line 17) | class DummyExponentialMovingAverage: method __init__ (line 18) | def __init__(self, *args, **kwargs): method _get_parameters (line 21) | def _get_parameters(self, *args, **kwargs): method get_current_decay (line 24) | def get_current_decay(self, *args, **kwargs): method update (line 27) | def update(self, *args, **kwargs): method copy_to (line 30) | def copy_to(self, *args, **kwargs): method store (line 33) | def store(self, *args, **kwargs): method restore (line 36) | def restore(self, *args, **kwargs): method average_parameters (line 40) | def average_parameters(self, *args, **kwargs): method to (line 46) | def to(self, *args, **kwargs): method state_dict (line 49) | def state_dict(self, *args, **kwargs): method load_state_dict (line 52) | def load_state_dict(self, *args, **kwargs): class ExponentialMovingAverage (line 56) | class ExponentialMovingAverage: method __init__ (line 84) | def __init__( method _get_parameters (line 110) | def _get_parameters( method get_current_decay (line 131) | def get_current_decay(self): method update (line 138) | def update(self, parameters: Optional[Iterable[torch.nn.Parameter]] = ... method copy_to (line 164) | def copy_to( method store (line 180) | def store(self, parameters: Optional[Iterable[torch.nn.Parameter]] = N... method restore (line 192) | def restore( method average_parameters (line 218) | def average_parameters( method to (line 248) | def to(self, device=None, dtype=None) -> None: method state_dict (line 274) | def state_dict(self) -> dict: method load_state_dict (line 286) | def load_state_dict(self, state_dict: dict) -> None: FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/utils/evaluation_depth.py function chamfer_dist (line 12) | def chamfer_dist(tensor1, tensor2): function auc (line 21) | def auc(tensor1, tensor2, thresholds): function delta (line 37) | def delta(tensor1, tensor2, exponent): function tau (line 42) | def tau(tensor1, tensor2, perc): function ssi (line 47) | def ssi(tensor1, tensor2): function si (line 59) | def si(tensor1, tensor2): function arel (line 63) | def arel(tensor1, tensor2): function d_auc (line 68) | def d_auc(tensor1, tensor2): function f1_score (line 74) | def f1_score(tensor1, tensor2, thresholds): function eval_depth (line 132) | def eval_depth( function eval_3d (line 150) | def eval_3d( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/utils/geometric.py function generate_rays (line 13) | def generate_rays( function spherical_zbuffer_to_euclidean (line 57) | def spherical_zbuffer_to_euclidean(spherical_tensor: torch.Tensor) -> to... function spherical_to_euclidean (line 77) | def spherical_to_euclidean(spherical_tensor: torch.Tensor) -> torch.Tensor: function euclidean_to_spherical (line 93) | def euclidean_to_spherical(spherical_tensor: torch.Tensor) -> torch.Tensor: function euclidean_to_spherical_zbuffer (line 109) | def euclidean_to_spherical_zbuffer(euclidean_tensor: torch.Tensor) -> to... function unproject_points (line 118) | def unproject_points( function project_points (line 161) | def project_points( function downsample (line 208) | def downsample(data: torch.Tensor, downsample_factor: int = 2): function flat_interpolate (line 228) | def flat_interpolate( function dilate (line 256) | def dilate(image, kernel_size: int | tuple[int, int]): function erode (line 272) | def erode(image, kernel_size: int | tuple[int, int]): function iou (line 288) | def iou(mask1: torch.Tensor, mask2: torch.Tensor) -> torch.Tensor: FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/utils/misc.py function max_stack (line 18) | def max_stack(tensors: list[torch.Tensor]) -> torch.Tensor: function last_stack (line 24) | def last_stack(tensors: list[torch.Tensor]) -> torch.Tensor: function first_stack (line 28) | def first_stack(tensors: list[torch.Tensor]) -> torch.Tensor: function softmax_stack (line 33) | def softmax_stack( function mean_stack (line 42) | def mean_stack(tensors: list[torch.Tensor]) -> torch.Tensor: function sum_stack (line 49) | def sum_stack(tensors: list[torch.Tensor]) -> torch.Tensor: function convert_module_to_f16 (line 55) | def convert_module_to_f16(l): function convert_module_to_f32 (line 65) | def convert_module_to_f32(l): function format_seconds (line 75) | def format_seconds(seconds): function get_params (line 81) | def get_params(module, lr, wd): function get_num_layer_for_swin (line 128) | def get_num_layer_for_swin(var_name, num_max_layer, layers_per_stage): function get_params_layerdecayswin (line 146) | def get_params_layerdecayswin(module, lr, wd, ld): function log (line 173) | def log(t, eps: float = 1e-5): function l2norm (line 177) | def l2norm(t): function exists (line 181) | def exists(val): function identity (line 185) | def identity(t, *args, **kwargs): function divisible_by (line 189) | def divisible_by(numer, denom): function first (line 193) | def first(arr, d=None): function default (line 199) | def default(val, d): function maybe (line 205) | def maybe(fn): function once (line 215) | def once(fn): function _many (line 229) | def _many(fn): function load_pretrained (line 242) | def load_pretrained(state_dict, checkpoint): function load_checkpoint_swin (line 257) | def load_checkpoint_swin(model, checkpoint_model): function add_padding_metas (line 359) | def add_padding_metas(out, image_metas): function remove_padding (line 369) | def remove_padding(out, paddings): function remove_padding_metas (line 378) | def remove_padding_metas(out, image_metas): function ssi_helper (line 388) | def ssi_helper(tensor1, tensor2): function calculate_mean_values (line 398) | def calculate_mean_values(names, values): function remove_leading_dim (line 416) | def remove_leading_dim(infos): function recursive_index (line 425) | def recursive_index(infos, index): function to_cpu (line 434) | def to_cpu(infos): function recursive_to (line 443) | def recursive_to(infos, device, non_blocking, cls): function masked_mean (line 454) | def masked_mean( class ProfileMethod (line 471) | class ProfileMethod: method __init__ (line 472) | def __init__(self, model, func_name, track_statistics=True, verbose=Fa... method __enter__ (line 479) | def __enter__(self): method __exit__ (line 487) | def __exit__(self, exc_type, exc_val, exc_tb): function profile_method (line 519) | def profile_method(track_statistics=True, verbose=False): class ProfileFunction (line 531) | class ProfileFunction: method __init__ (line 532) | def __init__(self, func_name, track_statistics=True, verbose=False): method __enter__ (line 538) | def __enter__(self): method __exit__ (line 546) | def __exit__(self, exc_type, exc_val, exc_tb): function profile_function (line 574) | def profile_function(track_statistics=True, verbose=False): function squeeze_list (line 586) | def squeeze_list(nested_list, dim, current_dim=0): function match_gt (line 596) | def match_gt(tensor1, tensor2, padding1, padding2, mode: str = "bilinear"): function match_intrinsics (line 645) | def match_intrinsics(K1, tensor1, tensor2, padding1, padding2): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/utils/positional_embedding.py class PositionEmbeddingSine (line 14) | class PositionEmbeddingSine(nn.Module): method __init__ (line 15) | def __init__( method forward (line 28) | def forward( method __repr__ (line 59) | def __repr__(self, _repr_indent=4): class LearnedSinusoidalPosEmb (line 72) | class LearnedSinusoidalPosEmb(nn.Module): method __init__ (line 73) | def __init__(self, dim): method forward (line 79) | def forward(self, x): function broadcat (line 87) | def broadcat(tensors, dim=-1): function rotate_half (line 106) | def rotate_half(x): class VisionRotaryEmbedding (line 113) | class VisionRotaryEmbedding(nn.Module): method __init__ (line 114) | def __init__( method forward (line 156) | def forward(self, t, start_index=0): class VisionRotaryEmbeddingFast (line 171) | class VisionRotaryEmbeddingFast(nn.Module): method __init__ (line 172) | def __init__( method forward (line 211) | def forward(self, t): function generate_fourier_features (line 218) | def generate_fourier_features( FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/utils/sht.py function rsh_cart_0 (line 11) | def rsh_cart_0(xyz: torch.Tensor): function rsh_cart_1 (line 36) | def rsh_cart_1(xyz: torch.Tensor): function rsh_cart_2 (line 67) | def rsh_cart_2(xyz: torch.Tensor): function rsh_cart_3 (line 110) | def rsh_cart_3(xyz: torch.Tensor): function rsh_cart_4 (line 160) | def rsh_cart_4(xyz: torch.Tensor): function rsh_cart_5 (line 228) | def rsh_cart_5(xyz: torch.Tensor): function rsh_cart_6 (line 329) | def rsh_cart_6(xyz: torch.Tensor): function rsh_cart_7 (line 509) | def rsh_cart_7(xyz: torch.Tensor): function rsh_cart_8 (line 833) | def rsh_cart_8(xyz: torch.Tensor): class SphHarm (line 1414) | class SphHarm(torch.nn.Module): method __init__ (line 1415) | def __init__(self, m, n, dtype=torch.float32) -> None: method device (line 1437) | def device(self): method forward (line 1440) | def forward(self, points: torch.Tensor) -> torch.Tensor: method _gen_recurrence_mask (line 1475) | def _gen_recurrence_mask(self) -> tuple[torch.Tensor, torch.Tensor]: method _recursive (line 1528) | def _recursive(self, i: int, p_val: torch.Tensor, x: torch.Tensor) -> ... method _init_legendre (line 1539) | def _init_legendre(self): method _gen_associated_legendre (line 1558) | def _gen_associated_legendre(self, x: torch.Tensor) -> torch.Tensor: FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/utils/validation.py function original_image (line 15) | def original_image(batch, preds=None): function log_metrics (line 52) | def log_metrics(metrics_all, step): function validate (line 64) | def validate(model, test_loaders, step, context): FILE: camera_pose_annotation/depth_estimation/UniDepth/unidepth/utils/visualization.py function colorize (line 17) | def colorize( function image_grid (line 39) | def image_grid(imgs: list[np.ndarray], rows: int, cols: int) -> np.ndarray: function get_pointcloud_from_rgbd (line 57) | def get_pointcloud_from_rgbd( function save_file_ply (line 107) | def save_file_ply(xyz, rgb, pc_file): function log_train_artifacts (line 136) | def log_train_artifacts(rgbs, gts, preds, ds_name, step, infos={}): FILE: camera_pose_annotation/dynamic_mask/inference_batch.py function compress (line 26) | def compress(dyn_masks, save_path=None): function segment_sky (line 41) | def segment_sky(image): function predict_mask (line 77) | def predict_mask(predictor, row, args, device): function worker (line 185) | def worker(task_queue, progress_queue, args, id): function parse_args (line 213) | def parse_args(): function main (line 242) | def main(): FILE: camera_pose_annotation/dynamic_mask/sam2/automatic_mask_generator.py class SAM2AutomaticMaskGenerator (line 36) | class SAM2AutomaticMaskGenerator: method __init__ (line 37) | def __init__( method from_pretrained (line 153) | def from_pretrained(cls, model_id: str, **kwargs) -> "SAM2AutomaticMas... method generate (line 170) | def generate(self, image: np.ndarray) -> List[Dict[str, Any]]: method _generate_masks (line 224) | def _generate_masks(self, image: np.ndarray) -> MaskData: method _process_crop (line 251) | def _process_crop( method _process_batch (line 294) | def _process_batch( method postprocess_small_regions (line 387) | def postprocess_small_regions( method refine_with_m2m (line 437) | def refine_with_m2m(self, points, point_labels, low_res_masks, points_... FILE: camera_pose_annotation/dynamic_mask/sam2/build_sam.py function build_sam2 (line 71) | def build_sam2( function build_sam2_video_predictor (line 100) | def build_sam2_video_predictor( function _hf_download (line 144) | def _hf_download(model_id): function build_sam2_hf (line 152) | def build_sam2_hf(model_id, **kwargs): function build_sam2_video_predictor_hf (line 157) | def build_sam2_video_predictor_hf(model_id, **kwargs): function _load_checkpoint (line 164) | def _load_checkpoint(model, ckpt_path): FILE: camera_pose_annotation/dynamic_mask/sam2/modeling/backbones/hieradet.py function do_pool (line 25) | def do_pool(x: torch.Tensor, pool: nn.Module, norm: nn.Module = None) ->... class MultiScaleAttention (line 39) | class MultiScaleAttention(nn.Module): method __init__ (line 40) | def __init__( method forward (line 56) | def forward(self, x: torch.Tensor) -> torch.Tensor: class MultiScaleBlock (line 84) | class MultiScaleBlock(nn.Module): method __init__ (line 85) | def __init__( method forward (line 134) | def forward(self, x: torch.Tensor) -> torch.Tensor: class Hiera (line 169) | class Hiera(nn.Module): method __init__ (line 174) | def __init__( method _get_pos_embed (line 273) | def _get_pos_embed(self, hw: Tuple[int, int]) -> torch.Tensor: method forward (line 283) | def forward(self, x: torch.Tensor) -> List[torch.Tensor]: method get_layer_id (line 301) | def get_layer_id(self, layer_name): method get_num_layers (line 316) | def get_num_layers(self) -> int: FILE: camera_pose_annotation/dynamic_mask/sam2/modeling/backbones/image_encoder.py class ImageEncoder (line 14) | class ImageEncoder(nn.Module): method __init__ (line 15) | def __init__( method forward (line 29) | def forward(self, sample: torch.Tensor): class FpnNeck (line 45) | class FpnNeck(nn.Module): method __init__ (line 52) | def __init__( method forward (line 102) | def forward(self, xs: List[torch.Tensor]): FILE: camera_pose_annotation/dynamic_mask/sam2/modeling/backbones/utils.py function window_partition (line 16) | def window_partition(x, window_size): function window_unpartition (line 39) | def window_unpartition(windows, window_size, pad_hw, hw): class PatchEmbed (line 63) | class PatchEmbed(nn.Module): method __init__ (line 68) | def __init__( method forward (line 89) | def forward(self, x: torch.Tensor) -> torch.Tensor: FILE: camera_pose_annotation/dynamic_mask/sam2/modeling/memory_attention.py class MemoryAttentionLayer (line 17) | class MemoryAttentionLayer(nn.Module): method __init__ (line 19) | def __init__( method _forward_sa (line 58) | def _forward_sa(self, tgt, query_pos): method _forward_ca (line 66) | def _forward_ca(self, tgt, memory, query_pos, pos, num_k_exclude_rope=0): method forward (line 83) | def forward( class MemoryAttention (line 102) | class MemoryAttention(nn.Module): method __init__ (line 103) | def __init__( method forward (line 119) | def forward( FILE: camera_pose_annotation/dynamic_mask/sam2/modeling/memory_encoder.py class MaskDownSampler (line 17) | class MaskDownSampler(nn.Module): method __init__ (line 26) | def __init__( method forward (line 57) | def forward(self, x): class CXBlock (line 62) | class CXBlock(nn.Module): method __init__ (line 74) | def __init__( method forward (line 104) | def forward(self, x): class Fuser (line 120) | class Fuser(nn.Module): method __init__ (line 121) | def __init__(self, layer, num_layers, dim=None, input_projection=False): method forward (line 130) | def forward(self, x): class MemoryEncoder (line 138) | class MemoryEncoder(nn.Module): method __init__ (line 139) | def __init__( method forward (line 158) | def forward( FILE: camera_pose_annotation/dynamic_mask/sam2/modeling/position_encoding.py class PositionEmbeddingSine (line 16) | class PositionEmbeddingSine(nn.Module): method __init__ (line 22) | def __init__( method _encode_xy (line 53) | def _encode_xy(self, x, y): method encode_boxes (line 73) | def encode_boxes(self, x, y, w, h): method encode_points (line 81) | def encode_points(self, x, y, labels): method _pe (line 90) | def _pe(self, B, device, *cache_key): method forward (line 127) | def forward(self, x: torch.Tensor): class PositionEmbeddingRandom (line 133) | class PositionEmbeddingRandom(nn.Module): method __init__ (line 138) | def __init__(self, num_pos_feats: int = 64, scale: Optional[float] = N... method _pe_encoding (line 147) | def _pe_encoding(self, coords: torch.Tensor) -> torch.Tensor: method forward (line 156) | def forward(self, size: Tuple[int, int]) -> torch.Tensor: method forward_with_coords (line 169) | def forward_with_coords( function init_t_xy (line 185) | def init_t_xy(end_x: int, end_y: int): function compute_axial_cis (line 192) | def compute_axial_cis(dim: int, end_x: int, end_y: int, theta: float = 1... function reshape_for_broadcast (line 204) | def reshape_for_broadcast(freqs_cis: torch.Tensor, x: torch.Tensor): function apply_rotary_enc (line 212) | def apply_rotary_enc( FILE: camera_pose_annotation/dynamic_mask/sam2/modeling/sam/mask_decoder.py class MaskDecoder (line 15) | class MaskDecoder(nn.Module): method __init__ (line 16) | def __init__( method forward (line 110) | def forward( method predict_masks (line 168) | def predict_masks( method _get_stability_scores (line 247) | def _get_stability_scores(self, mask_logits): method _dynamic_multimask_via_stability (line 259) | def _dynamic_multimask_via_stability(self, all_mask_logits, all_iou_sc... FILE: camera_pose_annotation/dynamic_mask/sam2/modeling/sam/prompt_encoder.py class PromptEncoder (line 17) | class PromptEncoder(nn.Module): method __init__ (line 18) | def __init__( method get_dense_pe (line 68) | def get_dense_pe(self) -> torch.Tensor: method _embed_points (line 79) | def _embed_points( method _embed_boxes (line 123) | def _embed_boxes(self, boxes: torch.Tensor) -> torch.Tensor: method _embed_masks (line 134) | def _embed_masks(self, masks: torch.Tensor) -> torch.Tensor: method _get_batch_size (line 139) | def _get_batch_size( method _get_device (line 157) | def _get_device(self) -> torch.device: method forward (line 160) | def forward( FILE: camera_pose_annotation/dynamic_mask/sam2/modeling/sam/transformer.py class TwoWayTransformer (line 19) | class TwoWayTransformer(nn.Module): method __init__ (line 20) | def __init__( method forward (line 65) | def forward( class TwoWayAttentionBlock (line 112) | class TwoWayAttentionBlock(nn.Module): method __init__ (line 113) | def __init__( method forward (line 156) | def forward( class Attention (line 190) | class Attention(nn.Module): method __init__ (line 196) | def __init__( method _separate_heads (line 220) | def _separate_heads(self, x: Tensor, num_heads: int) -> Tensor: method _recombine_heads (line 225) | def _recombine_heads(self, x: Tensor) -> Tensor: method forward (line 230) | def forward(self, q: Tensor, k: Tensor, v: Tensor) -> Tensor: class RoPEAttention (line 251) | class RoPEAttention(Attention): method __init__ (line 254) | def __init__( method forward (line 275) | def forward( FILE: camera_pose_annotation/dynamic_mask/sam2/modeling/sam2_base.py class SAM2Base (line 22) | class SAM2Base(torch.nn.Module): method __init__ (line 23) | def __init__( method device (line 198) | def device(self): method forward (line 201) | def forward(self, *args, **kwargs): method _build_sam_heads (line 207) | def _build_sam_heads(self): method _forward_sam_heads (line 257) | def _forward_sam_heads( method _use_mask_as_output (line 415) | def _use_mask_as_output(self, backbone_features, high_res_features, ma... method forward_image (line 467) | def forward_image(self, img_batch: torch.Tensor): method _prepare_backbone_features (line 481) | def _prepare_backbone_features(self, backbone_out): method _prepare_memory_conditioned_features (line 497) | def _prepare_memory_conditioned_features( method _encode_new_memory (line 678) | def _encode_new_memory( method _track_step (line 728) | def _track_step( method _encode_memory_in_output (line 789) | def _encode_memory_in_output( method track_step (line 814) | def track_step( method _use_multimask (line 881) | def _use_multimask(self, is_init_cond_frame, point_inputs): method _apply_non_overlapping_constraints (line 891) | def _apply_non_overlapping_constraints(self, pred_masks): FILE: camera_pose_annotation/dynamic_mask/sam2/modeling/sam2_utils.py function select_closest_cond_frames (line 19) | def select_closest_cond_frames(frame_idx, cond_frame_outputs, max_cond_f... function get_1d_sine_pe (line 64) | def get_1d_sine_pe(pos_inds, dim, temperature=10000): function get_activation_fn (line 77) | def get_activation_fn(activation): function get_clones (line 88) | def get_clones(module, N): class DropPath (line 92) | class DropPath(nn.Module): method __init__ (line 94) | def __init__(self, drop_prob=0.0, scale_by_keep=True): method forward (line 99) | def forward(self, x): class MLP (line 112) | class MLP(nn.Module): method __init__ (line 113) | def __init__( method forward (line 131) | def forward(self, x): class LayerNorm2d (line 141) | class LayerNorm2d(nn.Module): method __init__ (line 142) | def __init__(self, num_channels: int, eps: float = 1e-6) -> None: method forward (line 148) | def forward(self, x: torch.Tensor) -> torch.Tensor: function sample_box_points (line 156) | def sample_box_points( function sample_random_points_from_errors (line 202) | def sample_random_points_from_errors(gt_masks, pred_masks, num_pt=1): function sample_one_point_from_error_center (line 252) | def sample_one_point_from_error_center(gt_masks, pred_masks, padding=True): function get_next_point (line 317) | def get_next_point(gt_masks, pred_masks, method): FILE: camera_pose_annotation/dynamic_mask/sam2/sam2_image_predictor.py class SAM2ImagePredictor (line 20) | class SAM2ImagePredictor: method __init__ (line 21) | def __init__( method from_pretrained (line 69) | def from_pretrained(cls, model_id: str, **kwargs) -> "SAM2ImagePredict... method set_image (line 86) | def set_image( method set_image_batch (line 132) | def set_image_batch( method predict_batch (line 175) | def predict_batch( method predict (line 237) | def predict( method _prep_prompts (line 305) | def _prep_prompts( method _predict (line 337) | def _predict( method get_image_embedding (line 440) | def get_image_embedding(self) -> torch.Tensor: method device (line 456) | def device(self) -> torch.device: method reset_predictor (line 459) | def reset_predictor(self) -> None: FILE: camera_pose_annotation/dynamic_mask/sam2/sam2_video_predictor.py class SAM2VideoPredictor (line 19) | class SAM2VideoPredictor(SAM2Base): method __init__ (line 22) | def __init__( method init_state (line 42) | def init_state( method from_pretrained (line 102) | def from_pretrained(cls, model_id: str, **kwargs) -> "SAM2VideoPredict... method _obj_id_to_idx (line 118) | def _obj_id_to_idx(self, inference_state, obj_id): method _obj_idx_to_id (line 152) | def _obj_idx_to_id(self, inference_state, obj_idx): method _get_obj_num (line 156) | def _get_obj_num(self, inference_state): method add_new_points_or_box (line 161) | def add_new_points_or_box( method add_new_points (line 295) | def add_new_points(self, *args, **kwargs): method add_new_mask (line 300) | def add_new_mask( method _get_orig_video_res_output (line 383) | def _get_orig_video_res_output(self, inference_state, any_res_masks): method _consolidate_temp_output_across_obj (line 405) | def _consolidate_temp_output_across_obj( method propagate_in_video_preflight (line 480) | def propagate_in_video_preflight(self, inference_state): method propagate_in_video (line 546) | def propagate_in_video( method clear_all_prompts_in_frame (line 633) | def clear_all_prompts_in_frame( method reset_state (line 676) | def reset_state(self, inference_state): method _reset_tracking_results (line 689) | def _reset_tracking_results(self, inference_state): method _get_image_feature (line 704) | def _get_image_feature(self, inference_state, frame_idx, batch_size): method _run_single_frame_inference (line 737) | def _run_single_frame_inference( method _run_memory_encoder (line 805) | def _run_memory_encoder( method _get_maskmem_pos_enc (line 841) | def _get_maskmem_pos_enc(self, inference_state, current_out): method remove_object (line 867) | def remove_object(self, inference_state, obj_id, strict=False, need_ou... method _clear_non_cond_mem_around_input (line 956) | def _clear_non_cond_mem_around_input(self, inference_state, frame_idx): class SAM2VideoPredictorVOS (line 976) | class SAM2VideoPredictorVOS(SAM2VideoPredictor): method __init__ (line 979) | def __init__(self, *args, **kwargs): method _compile_all_components (line 983) | def _compile_all_components(self): method forward_image (line 1013) | def forward_image(self, img_batch: torch.Tensor): method _forward_sam_heads (line 1036) | def _forward_sam_heads( method _encode_new_memory (line 1171) | def _encode_new_memory( FILE: camera_pose_annotation/dynamic_mask/sam2/sam2_video_predictor_legacy.py class SAM2VideoPredictor (line 18) | class SAM2VideoPredictor(SAM2Base): method __init__ (line 21) | def __init__( method init_state (line 44) | def init_state( method from_pretrained (line 114) | def from_pretrained(cls, model_id: str, **kwargs) -> "SAM2VideoPredict... method _obj_id_to_idx (line 130) | def _obj_id_to_idx(self, inference_state, obj_id): method _obj_idx_to_id (line 164) | def _obj_idx_to_id(self, inference_state, obj_idx): method _get_obj_num (line 168) | def _get_obj_num(self, inference_state): method add_new_points_or_box (line 173) | def add_new_points_or_box( method add_new_points (line 316) | def add_new_points(self, *args, **kwargs): method add_new_mask (line 321) | def add_new_mask( method _get_orig_video_res_output (line 404) | def _get_orig_video_res_output(self, inference_state, any_res_masks): method _consolidate_temp_output_across_obj (line 426) | def _consolidate_temp_output_across_obj( method _get_empty_mask_ptr (line 556) | def _get_empty_mask_ptr(self, inference_state, frame_idx): method propagate_in_video_preflight (line 593) | def propagate_in_video_preflight(self, inference_state): method propagate_in_video (line 663) | def propagate_in_video( method _add_output_per_object (line 747) | def _add_output_per_object( method clear_all_prompts_in_frame (line 777) | def clear_all_prompts_in_frame( method reset_state (line 848) | def reset_state(self, inference_state): method _reset_tracking_results (line 860) | def _reset_tracking_results(self, inference_state): method _get_image_feature (line 879) | def _get_image_feature(self, inference_state, frame_idx, batch_size): method _run_single_frame_inference (line 912) | def _run_single_frame_inference( method _run_memory_encoder (line 980) | def _run_memory_encoder( method _get_maskmem_pos_enc (line 1016) | def _get_maskmem_pos_enc(self, inference_state, current_out): method remove_object (line 1042) | def remove_object(self, inference_state, obj_id, strict=False, need_ou... method _clear_non_cond_mem_around_input (line 1155) | def _clear_non_cond_mem_around_input(self, inference_state, frame_idx): FILE: camera_pose_annotation/dynamic_mask/sam2/utils/amg.py class MaskData (line 18) | class MaskData: method __init__ (line 24) | def __init__(self, **kwargs) -> None: method __setitem__ (line 31) | def __setitem__(self, key: str, item: Any) -> None: method __delitem__ (line 37) | def __delitem__(self, key: str) -> None: method __getitem__ (line 40) | def __getitem__(self, key: str) -> Any: method items (line 43) | def items(self) -> ItemsView[str, Any]: method filter (line 46) | def filter(self, keep: torch.Tensor) -> None: method cat (line 61) | def cat(self, new_stats: "MaskData") -> None: method to_numpy (line 74) | def to_numpy(self) -> None: function is_box_near_crop_edge (line 80) | def is_box_near_crop_edge( function box_xyxy_to_xywh (line 93) | def box_xyxy_to_xywh(box_xyxy: torch.Tensor) -> torch.Tensor: function batch_iterator (line 100) | def batch_iterator(batch_size: int, *args) -> Generator[List[Any], None,... function mask_to_rle_pytorch (line 109) | def mask_to_rle_pytorch(tensor: torch.Tensor) -> List[Dict[str, Any]]: function rle_to_mask (line 140) | def rle_to_mask(rle: Dict[str, Any]) -> np.ndarray: function area_from_rle (line 154) | def area_from_rle(rle: Dict[str, Any]) -> int: function calculate_stability_score (line 158) | def calculate_stability_score( function build_point_grid (line 181) | def build_point_grid(n_per_side: int) -> np.ndarray: function build_all_layer_point_grids (line 191) | def build_all_layer_point_grids( function generate_crop_boxes (line 202) | def generate_crop_boxes( function uncrop_boxes_xyxy (line 239) | def uncrop_boxes_xyxy(boxes: torch.Tensor, crop_box: List[int]) -> torch... function uncrop_points (line 248) | def uncrop_points(points: torch.Tensor, crop_box: List[int]) -> torch.Te... function uncrop_masks (line 257) | def uncrop_masks( function remove_small_regions (line 269) | def remove_small_regions( function coco_encode_rle (line 296) | def coco_encode_rle(uncompressed_rle: Dict[str, Any]) -> Dict[str, Any]: function batched_mask_to_box (line 305) | def batched_mask_to_box(masks: torch.Tensor) -> torch.Tensor: FILE: camera_pose_annotation/dynamic_mask/sam2/utils/misc.py function get_sdpa_settings (line 17) | def get_sdpa_settings(): function get_connected_components (line 47) | def get_connected_components(mask): function mask_to_box (line 66) | def mask_to_box(masks: torch.Tensor): function _load_img_as_tensor (line 92) | def _load_img_as_tensor(img_path, image_size): class AsyncVideoFrameLoader (line 104) | class AsyncVideoFrameLoader: method __init__ (line 109) | def __init__( method __getitem__ (line 147) | def __getitem__(self, index): method __len__ (line 168) | def __len__(self): function load_video_frames (line 172) | def load_video_frames( function load_video_frames_from_jpg_images (line 213) | def load_video_frames_from_jpg_images( function load_video_frames_from_video_file (line 280) | def load_video_frames_from_video_file( function fill_holes_in_mask_scores (line 312) | def fill_holes_in_mask_scores(mask, max_area): function concat_points (line 341) | def concat_points(old_point_inputs, new_points, new_labels): FILE: camera_pose_annotation/dynamic_mask/sam2/utils/transforms.py class SAM2Transforms (line 15) | class SAM2Transforms(nn.Module): method __init__ (line 16) | def __init__( method __call__ (line 37) | def __call__(self, x): method forward_batch (line 41) | def forward_batch(self, img_list): method transform_coords (line 46) | def transform_coords( method transform_boxes (line 66) | def transform_boxes( method postprocess_masks (line 76) | def postprocess_masks(self, masks: torch.Tensor, orig_hw) -> torch.Ten... FILE: caption/LLM/inference.py function get_pose (line 16) | def get_pose(pose_dir): function get_prompt (line 54) | def get_prompt(pose_dir, prompt_dir, vqa_caption, dist_level): function process_single_row (line 78) | def process_single_row(args, row): function worker (line 108) | def worker(args, task_queue, pbar): function parse_args (line 133) | def parse_args(): function main (line 165) | def main(): FILE: caption/VQA/inference.py function encode_image (line 17) | def encode_image(image_path): function get_prompt (line 35) | def get_prompt(fig_dir, prompt_text): function process_single_row (line 63) | def process_single_row(args, row): function worker (line 86) | def worker(args, task_queue, pbar): function parse_args (line 98) | def parse_args(): function main (line 126) | def main(): FILE: caption/tagging/inference.py function parse_category_tags (line 15) | def parse_category_tags(tag_caption): function process_single_row (line 54) | def process_single_row(args, json_file): function worker (line 84) | def worker(args, task_queue, pbar): function parse_args (line 97) | def parse_args(): function main (line 122) | def main(): FILE: caption/utils/api_call.py function api_call (line 4) | def api_call(prompt_text, model, api_key, base_domain): FILE: caption/utils/combine.py function parse_text_to_json (line 12) | def parse_text_to_json(text): function vqa_parse_text_to_json (line 103) | def vqa_parse_text_to_json(text): function process_single_row (line 127) | def process_single_row(args, clip_id): function worker (line 162) | def worker(args, task_queue, pbar): function parse_args (line 174) | def parse_args(): function main (line 189) | def main(): FILE: scoring/aesthetic/inference.py function merge_scores (line 26) | def merge_scores(gathered_list: list, csv: pd.DataFrame, column): class VideoTextDataset (line 50) | class VideoTextDataset(torch.utils.data.Dataset): method __init__ (line 53) | def __init__(self, csv_path, fig_load_dir, transform=None): method __getitem__ (line 59) | def __getitem__(self, index): method __len__ (line 75) | def __len__(self): class MLP (line 79) | class MLP(nn.Module): method __init__ (line 82) | def __init__(self, input_size): method forward (line 96) | def forward(self, x): class AestheticScorer (line 100) | class AestheticScorer(nn.Module): method __init__ (line 103) | def __init__(self, input_size, device): method forward (line 111) | def forward(self, x): function parse_args (line 118) | def parse_args(): function main (line 142) | def main(): FILE: scoring/luminance/inference.py function merge_scores (line 21) | def merge_scores(gathered_list: list, csv: pd.DataFrame): class VideoDataset (line 57) | class VideoDataset(torch.utils.data.Dataset): method __init__ (line 60) | def __init__(self, csv_path, fig_load_dir): method __getitem__ (line 65) | def __getitem__(self, index): method __len__ (line 80) | def __len__(self): function parse_args (line 84) | def parse_args(): function main (line 100) | def main(): FILE: scoring/motion/inference.py function get_ffmpeg_acceleration (line 18) | def get_ffmpeg_acceleration(): function process_single_row (line 42) | def process_single_row(video_path, args, process_id): function calculate_score (line 84) | def calculate_score(row, args): function worker1 (line 93) | def worker1(task_queue, progress_queue, args, process_id): function worker2 (line 105) | def worker2(task_queue, results_queue, args): function parse_args (line 117) | def parse_args(): function main (line 139) | def main(): FILE: scoring/ocr/inference.py function process_single_row (line 18) | def process_single_row(row, args, model): function worker (line 50) | def worker(task_queue, result_queue, args, id): function parse_args (line 74) | def parse_args(): function main (line 92) | def main(): FILE: utils/convert.py function scan_recursively (line 18) | def scan_recursively(root): function get_filelist (line 33) | def get_filelist(file_path, exts=None): function split_by_capital (line 53) | def split_by_capital(name): function process_general_videos (line 65) | def process_general_videos(root, output): FILE: utils/cut.py function get_ffmpeg_acceleration (line 19) | def get_ffmpeg_acceleration(): function seconds_to_timecode (line 40) | def seconds_to_timecode(seconds: float) -> str: function build_precise_cut_cmd (line 54) | def build_precise_cut_cmd( function _build_video_filters (line 153) | def _build_video_filters(shorter_size, args, accel_type) -> list[str]: function _build_encoder_args (line 182) | def _build_encoder_args(accel_type) -> list[str]: function process_single_row (line 205) | def process_single_row(row, args, process_id): function _cleanup (line 302) | def _cleanup(path: str): function parse_args (line 315) | def parse_args(): function _worker_fn (line 356) | def _worker_fn(task: tuple, args, process_id: int) -> tuple: function save_results (line 379) | def save_results(all_results: list, csv: pd.DataFrame, args): function main (line 439) | def main(): FILE: utils/cut_fast.py function process_single_row (line 32) | def process_single_row(row, save_dir, keep_audio=False): function process_video_group (line 138) | def process_video_group(group_df, save_dir, keep_audio=False): function worker (line 165) | def worker(task_queue, results_queue, video_save_dir, keep_audio=False): function parse_args (line 185) | def parse_args(): function main (line 220) | def main(): FILE: utils/download_SpatialVID.py function main (line 5) | def main(): FILE: utils/download_YouTube.py function ytb_download (line 27) | def ytb_download(url, json_info, output_dir="ytb_videos/"): function main (line 89) | async def main(csv_path, output_dir, max_workers=10, shards=0, total=-1,... function entry (line 119) | def entry( function add_download (line 152) | def add_download(csv_path): FILE: utils/evaluation.py function load_file (line 22) | def load_file(cam_pos_file, mask_file, device): function anomaly_detection (line 43) | def anomaly_detection(cam_pos, time_steps, threshold, device): function move_distance (line 73) | def move_distance(cam_pos, time_steps, device): function quaternion_multiply (line 91) | def quaternion_multiply(q1, q2): function rotation_angle (line 111) | def rotation_angle(cam_rotate, time_steps, device): function trajectory_turns (line 131) | def trajectory_turns(cam_pos, time_steps, device, threshold=0.45): function dynamic_ratio (line 167) | def dynamic_ratio(masks): function process_single_row (line 178) | def process_single_row(row, index, args, device): function worker (line 203) | def worker(task_queue, result_queue, args, worker_id): function parse_args (line 222) | def parse_args(): FILE: utils/expand_npz.py function expand (line 9) | def expand(loaded_data): FILE: utils/extract_frames.py function extract_frames_opencv (line 18) | def extract_frames_opencv( function extract_frames_av (line 57) | def extract_frames_av( function _calc_expected_frames (line 127) | def _calc_expected_frames(num_frames, interval): function _verify_frames (line 135) | def _verify_frames(img_dir, expected_frames): function process_single_row (line 151) | def process_single_row(row, row_index, args): function worker (line 207) | def worker(task_queue, progress_queue, failed_indices, args): function parse_args (line 221) | def parse_args(): function main (line 261) | def main(): FILE: utils/filter.py function main (line 13) | def main(args): function parse_args (line 64) | def parse_args(): FILE: utils/get_clip.py function process_single_row (line 16) | def process_single_row(row, args): function worker (line 84) | def worker(task_queue, results_queue, args): function parse_args (line 96) | def parse_args(): function main (line 123) | def main(): FILE: utils/get_info.py function get_video_length (line 16) | def get_video_length(cap, method="header"): function get_video_info (line 27) | def get_video_info(args): function main (line 77) | def main(args): function parse_args (line 139) | def parse_args(): FILE: utils/get_instructions.py function filter_poses (line 17) | def filter_poses(poses_array, alpha): function poses_to_multi_instructions (line 63) | def poses_to_multi_instructions(poses_array, translation_thresh, rotatio... function process_single_row (line 134) | def process_single_row(args, row): function worker (line 172) | def worker(task_queue, args, pbar): function args_parser (line 186) | def args_parser(): function main (line 223) | def main(): FILE: utils/get_instructions_enhanced.py function filter_poses (line 16) | def filter_poses(poses_array, alpha): function poses_to_multi_instructions (line 48) | def poses_to_multi_instructions(poses_array, translation_thresh, rotatio... function calculate_relative_scale (line 115) | def calculate_relative_scale(total_distance, num_poses, f_translation, m... function voter (line 125) | def voter(args, row, interval, alpha): function collect_all_results (line 153) | def collect_all_results(args, row, param_combinations): function get_mutually_exclusive_groups (line 164) | def get_mutually_exclusive_groups(): function remove_conflicting_instructions (line 173) | def remove_conflicting_instructions(instructions, conflict_groups): function smart_instruction_selection (line 189) | def smart_instruction_selection(non_conflicting_inst): function collect_interval_based_votes (line 226) | def collect_interval_based_votes(all_results, param_combinations): function vote_for_final_instructions (line 262) | def vote_for_final_instructions(all_results, param_combinations=None): function merge_consecutive_instructions (line 305) | def merge_consecutive_instructions(instructions): function process_single_row (line 319) | def process_single_row(args, row, param_combinations): function generate_param_combinations (line 338) | def generate_param_combinations(args): function worker (line 345) | def worker(task_queue, args, param_combinations, pbar): function args_parser (line 357) | def args_parser(): function main (line 381) | def main(): FILE: utils/merge_tables.py function read_csv_file (line 11) | def read_csv_file(file_path): function merge_tables_from_files (line 16) | def merge_tables_from_files(file_list, output_file, merge_on=None): function main (line 52) | def main(): FILE: utils/normalize_intrinsics.py function possess_single_row (line 22) | def possess_single_row(row, args): function worker (line 41) | def worker(task_queue, args, pbar): function parse_args (line 55) | def parse_args(): FILE: utils/pack_clip_assets.py function load_video (line 18) | def load_video(clip_path, indexes_path, height=720, width=1280): function load_intrinsics (line 49) | def load_intrinsics(intrinsics_path, tgt_width=1024, tgt_height=576): function main (line 74) | def main(): FILE: utils/quat_to_mat.py class Pose (line 25) | class Pose: method __call__ (line 31) | def __call__(self, R=None, t=None): method invert (line 63) | def invert(self, pose, use_inverse=False): # c2w <==> w2c method compose (line 84) | def compose(self, pose_list): method compose_pair (line 100) | def compose_pair(self, pose_a, pose_b): method scale_center (line 119) | def scale_center(self, pose, scale): function quaternion_to_matrix (line 137) | def quaternion_to_matrix(quaternions, eps: float = 1e-8): function pose_from_quaternion (line 172) | def pose_from_quaternion(pose): function possess_single_row (line 197) | def possess_single_row(row, index, args): function worker (line 226) | def worker(task_queue, args, pbar): function parse_args (line 238) | def parse_args(): FILE: utils/read_depth.py function read_depth (line 6) | def read_depth(zip_file_path): FILE: utils/read_video.py function read_video_av (line 23) | def read_video_av( function _read_from_stream (line 107) | def _read_from_stream( function read_video_cv2 (line 195) | def read_video_cv2(filename, start_pts=None, end_pts=None, pts_unit="pts"): function read_video (line 247) | def read_video(video_path, backend="av"): FILE: utils/scene_detect.py function timecode_to_seconds (line 29) | def timecode_to_seconds(timecode): function seconds_to_timecode (line 35) | def seconds_to_timecode(seconds): function process_single_row (line 43) | def process_single_row( function timecode_to_frames (line 130) | def timecode_to_frames(timecode, fps): function worker (line 137) | def worker(task_queue, results_queue, args): function parse_args (line 159) | def parse_args(): function main (line 213) | def main(): FILE: viser/docs/source/conf.py function docstring (line 243) | def docstring(app, what, name, obj, options, lines): function setup (line 250) | def setup(app): FILE: viser/docs/update_example_docs.py class ExampleMetadata (line 15) | class ExampleMetadata: method from_path (line 23) | def from_path(path: pathlib.Path) -> ExampleMetadata: function get_example_paths (line 46) | def get_example_paths(examples_dir: pathlib.Path) -> Iterable[pathlib.Pa... function main (line 55) | def main( FILE: viser/examples/01_image.py function main (line 18) | def main() -> None: FILE: viser/examples/02_gui.py function main (line 12) | def main() -> None: FILE: viser/examples/03_gui_callbacks.py function main (line 14) | def main() -> None: FILE: viser/examples/04_camera_poses.py function _ (line 15) | def _(client: viser.ClientHandle) -> None: FILE: viser/examples/05_camera_commands.py function _ (line 19) | def _(client: viser.ClientHandle) -> None: FILE: viser/examples/07_record3d_visualizer.py function main (line 18) | def main( FILE: viser/examples/08_smpl_visualizer.py class SmplOutputs (line 24) | class SmplOutputs: class SmplHelper (line 31) | class SmplHelper: method __init__ (line 34) | def __init__(self, model_path: Path) -> None: method get_outputs (line 49) | def get_outputs(self, betas: np.ndarray, joint_rotmats: np.ndarray) ->... function main (line 76) | def main(model_path: Path) -> None: class GuiElements (line 120) | class GuiElements: function make_gui_elements (line 133) | def make_gui_elements( FILE: viser/examples/09_urdf_visualizer.py function create_robot_control_sliders (line 24) | def create_robot_control_sliders( function main (line 55) | def main( FILE: viser/examples/10_realsense.py function realsense_pipeline (line 20) | def realsense_pipeline(fps: int = 30): function point_cloud_arrays_from_frames (line 42) | def point_cloud_arrays_from_frames( function main (line 94) | def main(): FILE: viser/examples/11_colmap_visualizer.py function main (line 24) | def main( FILE: viser/examples/12_click_meshes.py function main (line 13) | def main() -> None: FILE: viser/examples/13_theming.py function main (line 12) | def main(): FILE: viser/examples/14_markdown.py function _ (line 29) | def _(_): function _ (line 34) | def _(_): FILE: viser/examples/15_gui_in_scene.py function _ (line 22) | def _(client: viser.ClientHandle) -> None: FILE: viser/examples/16_modal.py function main (line 10) | def main(): FILE: viser/examples/18_splines.py function main (line 13) | def main() -> None: FILE: viser/examples/19_get_renders.py function main (line 13) | def main(): FILE: viser/examples/20_scene_pointer.py function _ (line 47) | def _(client: viser.ClientHandle) -> None: FILE: viser/examples/21_set_up_direction.py function main (line 10) | def main() -> None: FILE: viser/examples/22_games.py function main (line 16) | def main() -> None: function play_connect_4 (line 28) | def play_connect_4(server: viser.ViserServer) -> None: function play_tic_tac_toe (line 88) | def play_tic_tac_toe(server: viser.ViserServer) -> None: FILE: viser/examples/23_plotly.py function create_sinusoidal_wave (line 15) | def create_sinusoidal_wave(t: float) -> go.Figure: function main (line 36) | def main() -> None: FILE: viser/examples/24_notification.py function main (line 10) | def main() -> None: FILE: viser/examples/25_smpl_visualizer_skinned.py class SmplOutputs (line 30) | class SmplOutputs: class SmplHelper (line 37) | class SmplHelper: method __init__ (line 40) | def __init__(self, model_path: Path) -> None: method get_outputs (line 55) | def get_outputs(self, betas: np.ndarray, joint_rotmats: np.ndarray) ->... function main (line 82) | def main(model_path: Path) -> None: class GuiElements (line 147) | class GuiElements: function make_gui_elements (line 160) | def make_gui_elements( FILE: viser/examples/experimental/gaussian_splats.py class SplatFile (line 18) | class SplatFile(TypedDict): function load_splat_file (line 31) | def load_splat_file(splat_path: Path, center: bool = False) -> SplatFile: function load_ply_file (line 75) | def load_ply_file(ply_file_path: Path, center: bool = False) -> SplatFile: function main (line 108) | def main(splat_paths: tuple[Path, ...]) -> None: FILE: viser/examples/quick_save.py function process_folder (line 20) | def process_folder( function main (line 145) | def main( FILE: viser/src/viser/_client_autobuild.py function _check_viser_yarn_running (line 12) | def _check_viser_yarn_running() -> bool: function ensure_client_is_built (line 28) | def ensure_client_is_built() -> None: function _install_sandboxed_node (line 91) | def _install_sandboxed_node() -> Path: function _modified_time_recursive (line 120) | def _modified_time_recursive(dir: Path) -> float: FILE: viser/src/viser/_gui_api.py function _hex_from_hls (line 80) | def _hex_from_hls(h: float, l: float, s: float) -> str: function _compute_step (line 90) | def _compute_step(x: float | None) -> float: # type: ignore function _compute_precision_digits (line 103) | def _compute_precision_digits(x: float) -> int: class _RootGuiContainer (line 120) | class _RootGuiContainer: function _apply_default_order (line 127) | def _apply_default_order(order: float | None) -> float: function get_type_hints_cached (line 142) | def get_type_hints_cached(cls: type[Any]) -> dict[str, Any]: class _FileUploadState (line 146) | class _FileUploadState(TypedDict): class GuiApi (line 156) | class GuiApi: method __init__ (line 165) | def __init__( method _handle_gui_updates (line 203) | def _handle_gui_updates( method _handle_file_transfer_start (line 262) | def _handle_file_transfer_start( method _handle_file_transfer_part (line 277) | def _handle_file_transfer_part( method _get_container_id (line 338) | def _get_container_id(self) -> str: method _set_container_id (line 342) | def _set_container_id(self, container_id: str) -> None: method reset (line 346) | def reset(self) -> None: method set_panel_label (line 350) | def set_panel_label(self, label: str | None) -> None: method configure_theme (line 358) | def configure_theme( method add_folder (line 433) | def add_folder( method add_modal (line 471) | def add_modal( method add_tab_group (line 500) | def add_tab_group( method add_markdown (line 538) | def add_markdown( method add_plotly (line 580) | def add_plotly( method add_button (line 656) | def add_button( method add_upload_button (line 704) | def add_upload_button( method add_button_group (line 761) | def add_button_group( method add_button_group (line 772) | def add_button_group( method add_button_group (line 782) | def add_button_group( method add_checkbox (line 824) | def add_checkbox( method add_text (line 864) | def add_text( method add_number (line 904) | def add_number( method add_vector2 (line 973) | def add_vector2( method add_vector3 (line 1035) | def add_vector3( method add_dropdown (line 1099) | def add_dropdown( method add_dropdown (line 1111) | def add_dropdown( method add_dropdown (line 1122) | def add_dropdown( method add_progress_bar (line 1169) | def add_progress_bar( method add_slider (line 1212) | def add_slider( method add_multi_slider (line 1292) | def add_multi_slider( method add_rgb (line 1375) | def add_rgb( method add_rgba (line 1415) | def add_rgba( method _create_gui_input (line 1454) | def _create_gui_input( FILE: viser/src/viser/_gui_handles.py function _make_unique_id (line 33) | def _make_unique_id() -> str: class GuiContainerProtocol (line 38) | class GuiContainerProtocol(Protocol): class SupportsRemoveProtocol (line 44) | class SupportsRemoveProtocol(Protocol): method remove (line 45) | def remove(self) -> None: ... class _GuiHandleState (line 49) | class _GuiHandleState(Generic[T]): class _GuiInputHandle (line 80) | class _GuiInputHandle(Generic[T]): method order (line 102) | def order(self) -> float: method value (line 107) | def value(self) -> T: method value (line 112) | def value(self, value: T | onp.ndarray) -> None: method update_timestamp (line 143) | def update_timestamp(self) -> float: method disabled (line 148) | def disabled(self) -> bool: method disabled (line 154) | def disabled(self, disabled: bool) -> None: method visible (line 164) | def visible(self) -> bool: method visible (line 170) | def visible(self, visible: bool) -> None: method __post_init__ (line 179) | def __post_init__(self) -> None: method remove (line 189) | def remove(self) -> None: class GuiInputHandle (line 206) | class GuiInputHandle(_GuiInputHandle[T], Generic[T]): method on_update (line 215) | def on_update( class GuiEvent (line 224) | class GuiEvent(Generic[TGuiHandle]): class GuiButtonHandle (line 238) | class GuiButtonHandle(_GuiInputHandle[bool]): method on_click (line 243) | def on_click( class UploadedFile (line 252) | class UploadedFile: class GuiUploadButtonHandle (line 262) | class GuiUploadButtonHandle(_GuiInputHandle[UploadedFile]): method on_upload (line 268) | def on_upload( class GuiButtonGroupHandle (line 277) | class GuiButtonGroupHandle(_GuiInputHandle[StringType], Generic[StringTy... method on_click (line 282) | def on_click( method disabled (line 290) | def disabled(self) -> bool: method disabled (line 295) | def disabled(self, disabled: bool) -> None: class GuiDropdownHandle (line 301) | class GuiDropdownHandle(GuiInputHandle[StringType], Generic[StringType]): method options (line 309) | def options(self) -> tuple[StringType, ...]: method options (line 320) | def options(self, options: Iterable[StringType]) -> None: class GuiTabGroupHandle (line 342) | class GuiTabGroupHandle: method order (line 354) | def order(self) -> float: method add_tab (line 358) | def add_tab(self, label: str, icon: IconName | None = None) -> GuiTabH... method __post_init__ (line 373) | def __post_init__(self) -> None: method remove (line 377) | def remove(self) -> None: method _sync_with_client (line 386) | def _sync_with_client(self) -> None: class GuiFolderHandle (line 401) | class GuiFolderHandle: method order (line 414) | def order(self) -> float: method __enter__ (line 418) | def __enter__(self) -> GuiFolderHandle: method __exit__ (line 423) | def __exit__(self, *args) -> None: method __post_init__ (line 429) | def __post_init__(self) -> None: method remove (line 434) | def remove(self) -> None: class GuiModalHandle (line 446) | class GuiModalHandle: method __enter__ (line 456) | def __enter__(self) -> GuiModalHandle: method __exit__ (line 461) | def __exit__(self, *args) -> None: method __post_init__ (line 467) | def __post_init__(self) -> None: method close (line 470) | def close(self) -> None: class GuiTabHandle (line 481) | class GuiTabHandle: method __enter__ (line 491) | def __enter__(self) -> GuiTabHandle: method __exit__ (line 496) | def __exit__(self, *args) -> None: method __post_init__ (line 502) | def __post_init__(self) -> None: method remove (line 505) | def remove(self) -> None: function _get_data_url (line 525) | def _get_data_url(url: str, image_root: Path | None) -> str: function _parse_markdown (line 551) | def _parse_markdown(markdown: str, image_root: Path | None) -> str: class GuiProgressBarHandle (line 563) | class GuiProgressBarHandle: method value (line 575) | def value(self) -> float: method value (line 581) | def value(self, value: float) -> None: method animated (line 592) | def animated(self) -> bool: method animated (line 597) | def animated(self, animated: bool) -> None: method order (line 607) | def order(self) -> float: method visible (line 612) | def visible(self) -> bool: method visible (line 618) | def visible(self, visible: bool) -> None: method __post_init__ (line 627) | def __post_init__(self) -> None: method remove (line 632) | def remove(self) -> None: class GuiMarkdownHandle (line 641) | class GuiMarkdownHandle: method content (line 653) | def content(self) -> str: method content (line 659) | def content(self, content: str) -> None: method order (line 669) | def order(self) -> float: method visible (line 674) | def visible(self) -> bool: method visible (line 680) | def visible(self, visible: bool) -> None: method __post_init__ (line 689) | def __post_init__(self) -> None: method remove (line 694) | def remove(self) -> None: class GuiPlotlyHandle (line 703) | class GuiPlotlyHandle: method figure (line 715) | def figure(self) -> go.Figure: method figure (line 721) | def figure(self, figure: go.Figure) -> None: method aspect (line 735) | def aspect(self) -> float: method aspect (line 741) | def aspect(self, aspect: float) -> None: method order (line 751) | def order(self) -> float: method visible (line 756) | def visible(self) -> bool: method visible (line 762) | def visible(self, visible: bool) -> None: method __post_init__ (line 771) | def __post_init__(self) -> None: method remove (line 776) | def remove(self) -> None: FILE: viser/src/viser/_icons.py function svg_from_icon (line 9) | def svg_from_icon(icon_name: IconName) -> str: FILE: viser/src/viser/_icons_enum.py class _IconStringConverter (line 9) | class _IconStringConverter(type): method __getattr__ (line 10) | def __getattr__(self, __name: str) -> IconName: class Icon (line 17) | class Icon(metaclass=_IconStringConverter): FILE: viser/src/viser/_icons_enum.pyi class Icon (line 8) | class Icon: FILE: viser/src/viser/_icons_generate_enum.py function enum_name_from_icon (line 10) | def enum_name_from_icon(name: str) -> str: FILE: viser/src/viser/_messages.py class Message (line 45) | class Message(infra.Message): method redundancy_key (line 49) | def redundancy_key(self) -> str: function tag_class (line 74) | def tag_class(tag: str) -> Callable[[T], T]: class RunJavascriptMessage (line 85) | class RunJavascriptMessage(Message): method redundancy_key (line 93) | def redundancy_key(self) -> str: class NotificationMessage (line 99) | class NotificationMessage(Message): class RemoveNotificationMessage (line 113) | class RemoveNotificationMessage(Message): class ViewerCameraMessage (line 120) | class ViewerCameraMessage(Message): class ScenePointerMessage (line 137) | class ScenePointerMessage(Message): class ScenePointerEnableMessage (line 151) | class ScenePointerEnableMessage(Message): method redundancy_key (line 158) | def redundancy_key(self) -> str: class CameraFrustumMessage (line 165) | class CameraFrustumMessage(Message): class GlbMessage (line 181) | class GlbMessage(Message): class FrameMessage (line 190) | class FrameMessage(Message): class BatchedAxesMessage (line 201) | class BatchedAxesMessage(Message): class GridMessage (line 215) | class GridMessage(Message): class LabelMessage (line 237) | class LabelMessage(Message): class Gui3DMessage (line 245) | class Gui3DMessage(Message): class PointCloudMessage (line 254) | class PointCloudMessage(Message): method __post_init__ (line 268) | def __post_init__(self): class MeshBoneMessage (line 279) | class MeshBoneMessage(Message): class MeshMessage (line 286) | class MeshMessage(Message): method __post_init__ (line 304) | def __post_init__(self): class SkinnedMeshMessage (line 311) | class SkinnedMeshMessage(MeshMessage): method __post_init__ (line 321) | def __post_init__(self): class SetBoneOrientationMessage (line 334) | class SetBoneOrientationMessage(Message): method redundancy_key (line 344) | def redundancy_key(self) -> str: class SetBonePositionMessage (line 349) | class SetBonePositionMessage(Message): method redundancy_key (line 359) | def redundancy_key(self) -> str: class TransformControlsMessage (line 364) | class TransformControlsMessage(Message): class SetCameraPositionMessage (line 387) | class SetCameraPositionMessage(Message): class SetCameraUpDirectionMessage (line 394) | class SetCameraUpDirectionMessage(Message): class SetCameraLookAtMessage (line 401) | class SetCameraLookAtMessage(Message): class SetCameraFovMessage (line 408) | class SetCameraFovMessage(Message): class SetOrientationMessage (line 415) | class SetOrientationMessage(Message): class SetPositionMessage (line 425) | class SetPositionMessage(Message): class TransformControlsUpdateMessage (line 435) | class TransformControlsUpdateMessage(Message): class BackgroundImageMessage (line 446) | class BackgroundImageMessage(Message): class ImageMessage (line 455) | class ImageMessage(Message): class RemoveSceneNodeMessage (line 466) | class RemoveSceneNodeMessage(Message): class SetSceneNodeVisibilityMessage (line 473) | class SetSceneNodeVisibilityMessage(Message): class SetSceneNodeClickableMessage (line 481) | class SetSceneNodeClickableMessage(Message): class SceneNodeClickMessage (line 489) | class SceneNodeClickMessage(Message): class ResetSceneMessage (line 501) | class ResetSceneMessage(Message): class ResetGuiMessage (line 506) | class ResetGuiMessage(Message): class GuiAddFolderMessage (line 512) | class GuiAddFolderMessage(Message): class GuiAddMarkdownMessage (line 523) | class GuiAddMarkdownMessage(Message): class GuiAddProgressBarMessage (line 533) | class GuiAddProgressBarMessage(Message): class GuiAddPlotlyMessage (line 545) | class GuiAddPlotlyMessage(Message): class GuiAddTabGroupMessage (line 556) | class GuiAddTabGroupMessage(Message): class _GuiAddInputBase (line 567) | class _GuiAddInputBase(Message): class GuiModalMessage (line 581) | class GuiModalMessage(Message): class GuiCloseModalMessage (line 588) | class GuiCloseModalMessage(Message): class GuiAddButtonMessage (line 594) | class GuiAddButtonMessage(_GuiAddInputBase): class GuiAddUploadButtonMessage (line 604) | class GuiAddUploadButtonMessage(_GuiAddInputBase): class GuiAddSliderMessage (line 612) | class GuiAddSliderMessage(_GuiAddInputBase): class GuiAddMultiSliderMessage (line 623) | class GuiAddMultiSliderMessage(_GuiAddInputBase): class GuiAddNumberMessage (line 635) | class GuiAddNumberMessage(_GuiAddInputBase): class GuiAddRgbMessage (line 645) | class GuiAddRgbMessage(_GuiAddInputBase): class GuiAddRgbaMessage (line 651) | class GuiAddRgbaMessage(_GuiAddInputBase): class GuiAddCheckboxMessage (line 657) | class GuiAddCheckboxMessage(_GuiAddInputBase): class GuiAddVector2Message (line 663) | class GuiAddVector2Message(_GuiAddInputBase): class GuiAddVector3Message (line 673) | class GuiAddVector3Message(_GuiAddInputBase): class GuiAddTextMessage (line 683) | class GuiAddTextMessage(_GuiAddInputBase): class GuiAddDropdownMessage (line 689) | class GuiAddDropdownMessage(_GuiAddInputBase): class GuiAddButtonGroupMessage (line 696) | class GuiAddButtonGroupMessage(_GuiAddInputBase): class GuiRemoveMessage (line 702) | class GuiRemoveMessage(Message): class GuiUpdateMessage (line 709) | class GuiUpdateMessage(Message): method redundancy_key (line 720) | def redundancy_key(self) -> str: class ThemeConfigurationMessage (line 731) | class ThemeConfigurationMessage(Message): class CatmullRomSplineMessage (line 744) | class CatmullRomSplineMessage(Message): class CubicBezierSplineMessage (line 758) | class CubicBezierSplineMessage(Message): class GaussianSplatsMessage (line 770) | class GaussianSplatsMessage(Message): class GetRenderRequestMessage (line 791) | class GetRenderRequestMessage(Message): class GetRenderResponseMessage (line 801) | class GetRenderResponseMessage(Message): class FileTransferStart (line 808) | class FileTransferStart(Message): method redundancy_key (line 820) | def redundancy_key(self) -> str: class FileTransferPart (line 825) | class FileTransferPart(Message): method redundancy_key (line 835) | def redundancy_key(self) -> str: class FileTransferPartAck (line 840) | class FileTransferPartAck(Message): method redundancy_key (line 849) | def redundancy_key(self) -> str: class ShareUrlRequest (line 860) | class ShareUrlRequest(Message): class ShareUrlUpdated (line 865) | class ShareUrlUpdated(Message): class ShareUrlDisconnect (line 872) | class ShareUrlDisconnect(Message): class SetGuiPanelLabelMessage (line 877) | class SetGuiPanelLabelMessage(Message): FILE: viser/src/viser/_notification_handle.py class _NotificationHandleState (line 12) | class _NotificationHandleState: class NotificationHandle (line 24) | class NotificationHandle: method _sync_with_client (line 29) | def _sync_with_client(self, first: bool = False) -> None: method title (line 43) | def title(self) -> str: method title (line 48) | def title(self, title: str) -> None: method body (line 56) | def body(self) -> str: method body (line 61) | def body(self, body: str) -> None: method loading (line 69) | def loading(self) -> bool: method loading (line 74) | def loading(self, loading: bool) -> None: method with_close_button (line 82) | def with_close_button(self) -> bool: method with_close_button (line 87) | def with_close_button(self, with_close_button: bool) -> None: method auto_close (line 95) | def auto_close(self) -> int | Literal[False]: method auto_close (line 101) | def auto_close(self, auto_close: int | Literal[False]) -> None: method color (line 109) | def color(self) -> Color | None: method color (line 114) | def color(self, color: Color | None) -> None: method remove (line 121) | def remove(self) -> None: FILE: viser/src/viser/_scene_api.py function _colors_to_uint8 (line 48) | def _colors_to_uint8(colors: onp.ndarray) -> onpt.NDArray[onp.uint8]: function _encode_rgb (line 64) | def _encode_rgb(rgb: RgbTupleOrArray) -> int: function _encode_image_binary (line 75) | def _encode_image_binary( function cast_vector (line 103) | def cast_vector(vector: TVector | onp.ndarray, length: int) -> TVector: class SceneApi (line 111) | class SceneApi: method __init__ (line 117) | def __init__( method set_up_direction (line 172) | def set_up_direction( method set_global_visibility (line 242) | def set_global_visibility(self, visible: bool) -> None: method add_glb (line 257) | def add_glb( method add_spline_catmull_rom (line 291) | def add_spline_catmull_rom( method add_spline_cubic_bezier (line 346) | def add_spline_cubic_bezier( method add_camera_frustum (line 402) | def add_camera_frustum( method add_frame (line 467) | def add_frame( method add_batched_axes (line 516) | def add_batched_axes( method add_grid (line 572) | def add_grid( method add_label (line 632) | def add_label( method add_point_cloud (line 658) | def add_point_cloud( method add_mesh_skinned (line 715) | def add_mesh_skinned( method add_mesh_simple (line 836) | def add_mesh_simple( method add_mesh_trimesh (line 902) | def add_mesh_trimesh( method _add_gaussian_splats (line 938) | def _add_gaussian_splats( method add_box (line 1007) | def add_box( method add_icosphere (line 1045) | def add_icosphere( method set_background_image (line 1087) | def set_background_image( method add_image (line 1133) | def add_image( method add_transform_controls (line 1177) | def add_transform_controls( method reset (line 1270) | def reset(self) -> None: method _get_client_handle (line 1274) | def _get_client_handle(self, client_id: ClientId) -> ClientHandle: method _handle_transform_controls_updates (line 1292) | def _handle_transform_controls_updates( method _handle_node_click_updates (line 1314) | def _handle_node_click_updates( method _handle_scene_pointer_updates (line 1334) | def _handle_scene_pointer_updates( method on_pointer_event (line 1351) | def on_pointer_event( method on_pointer_callback_removed (line 1402) | def on_pointer_callback_removed( method remove_pointer_callback (line 1418) | def remove_pointer_callback( method add_3d_gui_container (line 1447) | def add_3d_gui_container( FILE: viser/src/viser/_scene_handles.py class ScenePointerEvent (line 20) | class ScenePointerEvent: method event (line 39) | def event(self): class _SceneNodeHandleState (line 48) | class _SceneNodeHandleState: class SceneNodeHandle (line 65) | class SceneNodeHandle: method _make (line 71) | def _make( method wxyz (line 92) | def wxyz(self) -> onp.ndarray: method wxyz (line 99) | def wxyz(self, wxyz: tuple[float, float, float, float] | onp.ndarray) ... method position (line 109) | def position(self) -> onp.ndarray: method position (line 116) | def position(self, position: tuple[float, float, float] | onp.ndarray)... method visible (line 126) | def visible(self) -> bool: method visible (line 131) | def visible(self, visible: bool) -> None: method remove (line 139) | def remove(self) -> None: class SceneNodePointerEvent (line 147) | class SceneNodePointerEvent(Generic[TSceneNodeHandle]): class _ClickableSceneNodeHandle (line 170) | class _ClickableSceneNodeHandle(SceneNodeHandle): method on_click (line 171) | def on_click( class CameraFrustumHandle (line 186) | class CameraFrustumHandle(_ClickableSceneNodeHandle): class PointCloudHandle (line 191) | class PointCloudHandle(SceneNodeHandle): class BatchedAxesHandle (line 196) | class BatchedAxesHandle(_ClickableSceneNodeHandle): class FrameHandle (line 201) | class FrameHandle(_ClickableSceneNodeHandle): class MeshHandle (line 206) | class MeshHandle(_ClickableSceneNodeHandle): class GaussianSplatHandle (line 211) | class GaussianSplatHandle(_ClickableSceneNodeHandle): class MeshSkinnedHandle (line 219) | class MeshSkinnedHandle(_ClickableSceneNodeHandle): class BoneState (line 228) | class BoneState: class MeshSkinnedBoneHandle (line 237) | class MeshSkinnedBoneHandle: method wxyz (line 243) | def wxyz(self) -> onp.ndarray: method wxyz (line 250) | def wxyz(self, wxyz: tuple[float, float, float, float] | onp.ndarray) ... method position (line 262) | def position(self) -> onp.ndarray: method position (line 269) | def position(self, position: tuple[float, float, float] | onp.ndarray)... class GlbHandle (line 282) | class GlbHandle(_ClickableSceneNodeHandle): class ImageHandle (line 287) | class ImageHandle(_ClickableSceneNodeHandle): class LabelHandle (line 292) | class LabelHandle(SceneNodeHandle): class _TransformControlsState (line 297) | class _TransformControlsState: class TransformControlsHandle (line 304) | class TransformControlsHandle(_ClickableSceneNodeHandle): method update_timestamp (line 310) | def update_timestamp(self) -> float: method on_update (line 313) | def on_update( class Gui3dContainerHandle (line 322) | class Gui3dContainerHandle(SceneNodeHandle): method __enter__ (line 332) | def __enter__(self) -> Gui3dContainerHandle: method __exit__ (line 337) | def __exit__(self, *args) -> None: method __post_init__ (line 343) | def __post_init__(self) -> None: method remove (line 346) | def remove(self) -> None: FILE: viser/src/viser/_tunnel.py function _is_multiprocess_ok (line 15) | def _is_multiprocess_ok() -> bool: class ViserTunnel (line 25) | class ViserTunnel: method __init__ (line 32) | def __init__(self, share_domain: str, local_port: int) -> None: method on_disconnect (line 63) | def on_disconnect(self, callback: Callable[[], None]) -> None: method on_connect (line 73) | def on_connect(self, callback: Callable[[int], None]) -> None: method get_url (line 123) | def get_url(self) -> str | None: method get_status (line 127) | def get_status( method close (line 132) | def close(self) -> None: function _connect_job (line 150) | def _connect_job( function _make_tunnel (line 185) | async def _make_tunnel( function _simple_proxy (line 239) | async def _simple_proxy( FILE: viser/src/viser/_viser.py class _BackwardsCompatibilityShim (line 30) | class _BackwardsCompatibilityShim: method __getattr__ (line 34) | def __getattr__(self, name: str) -> Any: class _CameraHandleState (line 67) | class _CameraHandleState: class CameraHandle (line 81) | class CameraHandle: method __init__ (line 85) | def __init__(self, client: ClientHandle) -> None: method client (line 99) | def client(self) -> ClientHandle: method wxyz (line 104) | def wxyz(self) -> npt.NDArray[onp.float64]: method wxyz (line 114) | def wxyz(self, wxyz: tuple[float, float, float, float] | onp.ndarray) ... method position (line 149) | def position(self) -> npt.NDArray[onp.float64]: method position (line 160) | def position(self, position: tuple[float, float, float] | onp.ndarray)... method _update_wxyz (line 169) | def _update_wxyz(self) -> None: method fov (line 180) | def fov(self) -> float: method fov (line 187) | def fov(self, fov: float) -> None: method aspect (line 195) | def aspect(self) -> float: method update_timestamp (line 201) | def update_timestamp(self) -> float: method look_at (line 206) | def look_at(self) -> npt.NDArray[onp.float64]: method look_at (line 212) | def look_at(self, look_at: tuple[float, float, float] | onp.ndarray) -... method up_direction (line 221) | def up_direction(self) -> npt.NDArray[onp.float64]: method up_direction (line 227) | def up_direction( method on_update (line 237) | def on_update( method get_render (line 244) | def get_render( class ClientHandle (line 299) | class ClientHandle(_BackwardsCompatibilityShim if not TYPE_CHECKING else... method __init__ (line 311) | def __init__( method flush (line 332) | def flush(self) -> None: method atomic (line 337) | def atomic(self) -> ContextManager[None]: method send_file_download (line 350) | def send_file_download( method add_notification (line 391) | def add_notification( class ViserServer (line 433) | class ViserServer(_BackwardsCompatibilityShim if not TYPE_CHECKING else ... method __init__ (line 457) | def __init__( method get_host (line 579) | def get_host(self) -> str: method get_port (line 587) | def get_port(self) -> int: method request_share_url (line 596) | def request_share_url(self, verbose: bool = True) -> str | None: method disconnect_share_url (line 652) | def disconnect_share_url(self) -> None: method stop (line 661) | def stop(self) -> None: method get_clients (line 667) | def get_clients(self) -> dict[int, ClientHandle]: method on_client_connect (line 677) | def on_client_connect( method on_client_disconnect (line 697) | def on_client_disconnect( method flush (line 704) | def flush(self) -> None: method atomic (line 709) | def atomic(self) -> ContextManager[None]: method send_file_download (line 722) | def send_file_download( method _start_scene_recording (line 735) | def _start_scene_recording(self) -> RecordHandle: FILE: viser/src/viser/client/src/App.tsx type ViewerContextContents (line 53) | type ViewerContextContents = { function ViewerRoot (line 117) | function ViewerRoot() { function ViewerContents (line 207) | function ViewerContents({ children }: { children: React.ReactNode }) { function ViewerCanvas (line 285) | function ViewerCanvas({ children }: { children: React.ReactNode }) { function AdaptiveDpr (line 477) | function AdaptiveDpr() { function Viewer2DCanvas (line 502) | function Viewer2DCanvas() { function BackgroundImage (line 534) | function BackgroundImage() { function SceneContextSetter (line 644) | function SceneContextSetter() { function Root (line 653) | function Root() { function ViserLogo (line 670) | function ViserLogo() { FILE: viser/src/viser/client/src/BrowserWarning.tsx function BrowserWarning (line 5) | function BrowserWarning() { FILE: viser/src/viser/client/src/CameraControls.tsx function SynchronizedCameraControls (line 11) | function SynchronizedCameraControls() { FILE: viser/src/viser/client/src/ClickUtils.tsx function ndcFromPointerXy (line 9) | function ndcFromPointerXy( function opencvXyFromPointerXy (line 31) | function opencvXyFromPointerXy( FILE: viser/src/viser/client/src/ControlPanel/BottomPanel.tsx function BottomPanel (line 12) | function BottomPanel({ FILE: viser/src/viser/client/src/ControlPanel/ControlPanel.tsx constant ROOT_CONTAINER_ID (line 43) | const ROOT_CONTAINER_ID = "root"; function ControlPanel (line 45) | function ControlPanel(props: { function ConnectionStatus (line 154) | function ConnectionStatus() { function ShareButton (line 192) | function ShareButton() { FILE: viser/src/viser/client/src/ControlPanel/FloatingPanel.tsx function FloatingPanel (line 35) | function FloatingPanel({ FILE: viser/src/viser/client/src/ControlPanel/Generated.tsx function GeneratedGuiContainer (line 27) | function GeneratedGuiContainer({ function GuiContainer (line 58) | function GuiContainer({ containerId }: { containerId: string }) { function GeneratedInput (line 85) | function GeneratedInput(props: { guiId: string }) { function assertNeverType (line 130) | function assertNeverType(x: never): never { FILE: viser/src/viser/client/src/ControlPanel/GuiComponentContext.tsx type GuiComponentContext (line 4) | interface GuiComponentContext { FILE: viser/src/viser/client/src/ControlPanel/GuiState.tsx type GuiConfig (line 8) | type GuiConfig = Messages.GuiAddComponentMessage; function isGuiConfig (line 10) | function isGuiConfig(message: Messages.Message): message is GuiConfig { type GuiState (line 14) | interface GuiState { type GuiActions (line 37) | interface GuiActions { function computeRelativeLuminance (line 77) | function computeRelativeLuminance(color: string) { function useGuiState (line 89) | function useGuiState(initialServer: string) { type UseGui (line 187) | type UseGui = ReturnType; FILE: viser/src/viser/client/src/ControlPanel/SceneTreeTable.tsx function SceneTreeTable (line 18) | function SceneTreeTable() { function setOverrideVisibility (line 50) | function setOverrideVisibility(name: string, visible: boolean | undefine... function rerenderTable (line 63) | function rerenderTable() { FILE: viser/src/viser/client/src/ControlPanel/ServerControls.tsx function ServerControls (line 16) | function ServerControls() { FILE: viser/src/viser/client/src/ControlPanel/SidebarPanel.tsx function SidebarPanel (line 21) | function SidebarPanel({ FILE: viser/src/viser/client/src/FilePlayback.tsx function deserializeGzippedMsgpackFile (line 24) | async function deserializeGzippedMsgpackFile( type SerializedMessages (line 71) | interface SerializedMessages { function PlaybackFromFile (line 77) | function PlaybackFromFile({ fileUrl }: { fileUrl: string }) { FILE: viser/src/viser/client/src/Markdown.tsx function rehypeCodeblock (line 28) | function rehypeCodeblock(): void | Transformer { function MdxText (line 42) | function MdxText(props: React.ComponentPropsWithoutRef) { function MdxAnchor (line 46) | function MdxAnchor(props: React.ComponentPropsWithoutRef) { function MdxTitle (line 50) | function MdxTitle( function MdxList (line 57) | function MdxList( function MdxListItem (line 77) | function MdxListItem( function MdxCode (line 85) | function MdxCode( function MdxBlockquote (line 92) | function MdxBlockquote( function MdxCite (line 98) | function MdxCite( function MdxTable (line 119) | function MdxTable(props: React.ComponentPropsWithoutRef) { function MdxImage (line 123) | function MdxImage(props: React.ComponentPropsWithoutRef) { function parseMarkdown (line 148) | async function parseMarkdown(markdown: string) { function Markdown (line 165) | function Markdown(props: { children?: string }) { FILE: viser/src/viser/client/src/MessageHandler.tsx function threeColorBufferFromUint8Buffer (line 35) | function threeColorBufferFromUint8Buffer(colors: ArrayBuffer) { function useMessageHandler (line 50) | function useMessageHandler() { function useFileDownloadHandler (line 1067) | function useFileDownloadHandler() { function FrameSynchronizedMessageHandler (line 1156) | function FrameSynchronizedMessageHandler() { FILE: viser/src/viser/client/src/Modal.tsx function ViserModal (line 7) | function ViserModal() { function GeneratedModal (line 18) | function GeneratedModal({ FILE: viser/src/viser/client/src/Outlines.tsx type OutlinesProps (line 69) | type OutlinesProps = JSX.IntrinsicElements["group"] & { FILE: viser/src/viser/client/src/SceneTree.tsx type MakeObject (line 16) | type MakeObject = ( class SceneNode (line 21) | class SceneNode { method constructor (line 27) | constructor( type UseSceneTree (line 49) | type UseSceneTree = ReturnType; function SceneNodeThreeChildren (line 51) | function SceneNodeThreeChildren(props: { function SceneNodeLabel (line 109) | function SceneNodeLabel(props: { name: string }) { function SceneNodeThreeObject (line 131) | function SceneNodeThreeObject(props: { FILE: viser/src/viser/client/src/SceneTreeState.tsx type SceneTreeState (line 9) | interface SceneTreeState { type SceneTreeActions (line 14) | interface SceneTreeActions extends SceneTreeState { function useSceneTreeState (line 39) | function useSceneTreeState( FILE: viser/src/viser/client/src/SearchParamsUtils.tsx function syncSearchParamServer (line 7) | function syncSearchParamServer(server: string) { FILE: viser/src/viser/client/src/Splatting/GaussianSplats.tsx type SplatState (line 35) | interface SplatState { function useGaussianSplatStore (line 45) | function useGaussianSplatStore() { function SplatRenderContext (line 71) | function SplatRenderContext({ function SplatRenderer (line 282) | function SplatRenderer() { function mergeGaussianGroups (line 444) | function mergeGaussianGroups(groupBufferFromName: { function useGaussianMeshProps (line 484) | function useGaussianMeshProps(gaussianBuffer: Uint32Array, numGroups: nu... FILE: viser/src/viser/client/src/Splatting/SplatSortWorker.ts type SorterWorkerIncoming (line 6) | type SorterWorkerIncoming = FILE: viser/src/viser/client/src/Splatting/WasmSorter/Sorter.mjs function locateFile (line 9) | function locateFile(path){if(Module["locateFile"]){return Module["locate... function updateMemoryViews (line 9) | function updateMemoryViews(){var b=wasmMemory.buffer;Module["HEAP8"]=HEA... function preRun (line 9) | function preRun(){if(Module["preRun"]){if(typeof Module["preRun"]=="func... function initRuntime (line 9) | function initRuntime(){runtimeInitialized=true;callRuntimeCallbacks(__AT... function postRun (line 9) | function postRun(){if(Module["postRun"]){if(typeof Module["postRun"]=="f... function addOnPreRun (line 9) | function addOnPreRun(cb){__ATPRERUN__.unshift(cb)} function addOnInit (line 9) | function addOnInit(cb){__ATINIT__.unshift(cb)} function addOnPostRun (line 9) | function addOnPostRun(cb){__ATPOSTRUN__.unshift(cb)} function addRunDependency (line 9) | function addRunDependency(id){runDependencies++;Module["monitorRunDepend... function removeRunDependency (line 9) | function removeRunDependency(id){runDependencies--;Module["monitorRunDep... function abort (line 9) | function abort(what){Module["onAbort"]?.(what);what="Aborted("+what+")";... function findWasmBinary (line 9) | function findWasmBinary(){if(Module["locateFile"]){var f="Sorter.wasm";i... function getBinarySync (line 9) | function getBinarySync(file){if(file==wasmBinaryFile&&wasmBinary){return... function getBinaryPromise (line 9) | function getBinaryPromise(binaryFile){if(!wasmBinary){return new Promise... function instantiateArrayBuffer (line 9) | function instantiateArrayBuffer(binaryFile,imports,receiver){return getB... function instantiateAsync (line 9) | function instantiateAsync(binary,binaryFile,imports,callback){if(!binary... function getWasmImports (line 9) | function getWasmImports(){return{a:wasmImports}} function createWasm (line 9) | function createWasm(){var info=getWasmImports();function receiveInstance... class ExceptionInfo (line 9) | class ExceptionInfo{constructor(excPtr){this.excPtr=excPtr;this.ptr=excP... method constructor (line 9) | constructor(excPtr){this.excPtr=excPtr;this.ptr=excPtr-24} method set_type (line 9) | set_type(type){HEAPU32[this.ptr+4>>2]=type} method get_type (line 9) | get_type(){return HEAPU32[this.ptr+4>>2]} method set_destructor (line 9) | set_destructor(destructor){HEAPU32[this.ptr+8>>2]=destructor} method get_destructor (line 9) | get_destructor(){return HEAPU32[this.ptr+8>>2]} method set_caught (line 9) | set_caught(caught){caught=caught?1:0;HEAP8[this.ptr+12]=caught} method get_caught (line 9) | get_caught(){return HEAP8[this.ptr+12]!=0} method set_rethrown (line 9) | set_rethrown(rethrown){rethrown=rethrown?1:0;HEAP8[this.ptr+13]=rethrown} method get_rethrown (line 9) | get_rethrown(){return HEAP8[this.ptr+13]!=0} method init (line 9) | init(type,destructor){this.set_adjusted_ptr(0);this.set_type(type);thi... method set_adjusted_ptr (line 9) | set_adjusted_ptr(adjustedPtr){HEAPU32[this.ptr+16>>2]=adjustedPtr} method get_adjusted_ptr (line 9) | get_adjusted_ptr(){return HEAPU32[this.ptr+16>>2]} method get_exception_ptr (line 9) | get_exception_ptr(){var isPointer=___cxa_is_pointer_type(this.get_type... function onComplete (line 9) | function onComplete(typeConverters){var myTypeConverters=getTypeConverte... function sharedRegisterType (line 9) | function sharedRegisterType(rawType,registeredInstance,options={}){var n... function registerType (line 9) | function registerType(rawType,registeredInstance,options={}){if(!("argPa... function getInstanceTypeName (line 9) | function getInstanceTypeName(handle){return handle.$$.ptrType.registered... function RegisteredPointer_fromWireType (line 9) | function RegisteredPointer_fromWireType(ptr){var rawPointer=this.getPoin... method isAliasOf (line 9) | isAliasOf(other){if(!(this instanceof ClassHandle)){return false}if(!(ot... method clone (line 9) | clone(){if(!this.$$.ptr){throwInstanceAlreadyDeleted(this)}if(this.$$.pr... method delete (line 9) | delete(){if(!this.$$.ptr){throwInstanceAlreadyDeleted(this)}if(this.$$.d... method isDeleted (line 9) | isDeleted(){return!this.$$.ptr} method deleteLater (line 9) | deleteLater(){if(!this.$$.ptr){throwInstanceAlreadyDeleted(this)}if(this... function ClassHandle (line 9) | function ClassHandle(){} function RegisteredClass (line 9) | function RegisteredClass(name,constructor,instancePrototype,rawDestructo... function constNoSmartPtrRawPointerToWireType (line 9) | function constNoSmartPtrRawPointerToWireType(destructors,handle){if(hand... function genericPointerToWireType (line 9) | function genericPointerToWireType(destructors,handle){var ptr;if(handle=... function nonConstNoSmartPtrRawPointerToWireType (line 9) | function nonConstNoSmartPtrRawPointerToWireType(destructors,handle){if(h... function readPointer (line 9) | function readPointer(pointer){return this["fromWireType"](HEAPU32[pointe... method getPointee (line 9) | getPointee(ptr){if(this.rawGetPointee){ptr=this.rawGetPointee(ptr)}retur... method destructor (line 9) | destructor(ptr){this.rawDestructor?.(ptr)} function RegisteredPointer (line 9) | function RegisteredPointer(name,registeredClass,isReference,isConst,isSm... function makeDynCaller (line 9) | function makeDynCaller(){if(signature.includes("j")){return getDynCaller... function visit (line 9) | function visit(type){if(seen[type]){return}if(registeredTypes[type]){ret... function usesDestructorStack (line 9) | function usesDestructorStack(argTypes){for(var i=1;i>2];var data=H... method fromWireType (line 9) | fromWireType(value){var length=HEAPU32[value>>2];var payload=value+4;var... method toWireType (line 9) | toWireType(destructors,value){if(value instanceof ArrayBuffer){value=new... method destructorFunction (line 9) | destructorFunction(ptr){_free(ptr)} method destructorFunction (line 9) | destructorFunction(ptr){_free(ptr)} method constructor (line 9) | constructor(message){super(message);this.name="BindingError"} method constructor (line 9) | constructor(message){super(message);this.name="InternalError"} function run (line 9) | function run(){if(runDependencies>0){return}preRun();if(runDependencies>... FILE: viser/src/viser/client/src/Splatting/WasmSorter/sorter.cpp function dot_f32x4 (line 11) | __attribute__((always_inline)) inline float function min_i32x4 (line 23) | __attribute__((always_inline)) inline int32_t min_i32x4(v128_t vector) { function max_i32x4 (line 33) | __attribute__((always_inline)) inline int32_t max_i32x4(v128_t vector) { class Sorter (line 41) | class Sorter { method Sorter (line 47) | Sorter( method sort (line 72) | emscripten::val sort(const emscripten::val &Tz_cam_groups_val) { function EMSCRIPTEN_BINDINGS (line 185) | EMSCRIPTEN_BINDINGS(c) { FILE: viser/src/viser/client/src/ThreeAssets.tsx type AllPossibleThreeJSMaterials (line 28) | type AllPossibleThreeJSMaterials = function disposeNode (line 185) | function disposeNode(node: any) { function disposeMaterial (line 201) | function disposeMaterial(material: AllPossibleThreeJSMaterials) { function scaledLineSegments (line 421) | function scaledLineSegments(points: [number, number, number][], thicknes... function LineSegmentInstance (line 508) | function LineSegmentInstance(props: { function OutlinesIfHovered (line 546) | function OutlinesIfHovered( FILE: viser/src/viser/client/src/Titlebar.tsx type ArrayElement (line 22) | type ArrayElement = type TitlebarContent (line 24) | type TitlebarContent = NonNullable< function assertUnreachable (line 27) | function assertUnreachable(x: never): never { function getIcon (line 31) | function getIcon( function TitlebarButton (line 54) | function TitlebarButton( function MobileTitlebarButton (line 74) | function MobileTitlebarButton( function TitlebarImage (line 94) | function TitlebarImage( function Titlebar (line 125) | function Titlebar() { FILE: viser/src/viser/client/src/Utils.ts type DragEvents (line 2) | interface DragEvents { function isTouchEvent (line 9) | function isTouchEvent( function isMouseEvent (line 14) | function isMouseEvent( FILE: viser/src/viser/client/src/WebsocketFunctions.tsx function useThrottledMessageSender (line 7) | function useThrottledMessageSender(throttleMilliseconds: number) { function makeThrottledMessageSender (line 13) | function makeThrottledMessageSender( function isTexture (line 42) | function isTexture( FILE: viser/src/viser/client/src/WebsocketInterface.tsx function WebsocketMessageProducer (line 9) | function WebsocketMessageProducer() { FILE: viser/src/viser/client/src/WebsocketMessages.tsx type RunJavascriptMessage (line 9) | interface RunJavascriptMessage { type NotificationMessage (line 17) | interface NotificationMessage { type RemoveNotificationMessage (line 47) | interface RemoveNotificationMessage { type ViewerCameraMessage (line 56) | interface ViewerCameraMessage { type ScenePointerMessage (line 72) | interface ScenePointerMessage { type ScenePointerEnableMessage (line 83) | interface ScenePointerEnableMessage { type CameraFrustumMessage (line 94) | interface CameraFrustumMessage { type GlbMessage (line 109) | interface GlbMessage { type FrameMessage (line 119) | interface FrameMessage { type BatchedAxesMessage (line 134) | interface BatchedAxesMessage { type GridMessage (line 146) | interface GridMessage { type LabelMessage (line 165) | interface LabelMessage { type Gui3DMessage (line 174) | interface Gui3DMessage { type PointCloudMessage (line 189) | interface PointCloudMessage { type MeshBoneMessage (line 201) | interface MeshBoneMessage { type MeshMessage (line 211) | interface MeshMessage { type SkinnedMeshMessage (line 230) | interface SkinnedMeshMessage { type SetBoneOrientationMessage (line 253) | interface SetBoneOrientationMessage { type SetBonePositionMessage (line 265) | interface SetBonePositionMessage { type TransformControlsMessage (line 275) | interface TransformControlsMessage { type SetCameraPositionMessage (line 295) | interface SetCameraPositionMessage { type SetCameraUpDirectionMessage (line 303) | interface SetCameraUpDirectionMessage { type SetCameraLookAtMessage (line 311) | interface SetCameraLookAtMessage { type SetCameraFovMessage (line 319) | interface SetCameraFovMessage { type SetOrientationMessage (line 329) | interface SetOrientationMessage { type SetPositionMessage (line 340) | interface SetPositionMessage { type TransformControlsUpdateMessage (line 351) | interface TransformControlsUpdateMessage { type BackgroundImageMessage (line 361) | interface BackgroundImageMessage { type ImageMessage (line 371) | interface ImageMessage { type RemoveSceneNodeMessage (line 383) | interface RemoveSceneNodeMessage { type SetSceneNodeVisibilityMessage (line 391) | interface SetSceneNodeVisibilityMessage { type SetSceneNodeClickableMessage (line 400) | interface SetSceneNodeClickableMessage { type SceneNodeClickMessage (line 409) | interface SceneNodeClickMessage { type ResetSceneMessage (line 421) | interface ResetSceneMessage { type ResetGuiMessage (line 428) | interface ResetGuiMessage { type GuiAddFolderMessage (line 435) | interface GuiAddFolderMessage { type GuiAddMarkdownMessage (line 448) | interface GuiAddMarkdownMessage { type GuiAddProgressBarMessage (line 460) | interface GuiAddProgressBarMessage { type GuiAddPlotlyMessage (line 489) | interface GuiAddPlotlyMessage { type GuiAddTabGroupMessage (line 502) | interface GuiAddTabGroupMessage { type _GuiAddInputBase (line 516) | interface _GuiAddInputBase { type GuiAddButtonMessage (line 531) | interface GuiAddButtonMessage { type GuiAddUploadButtonMessage (line 563) | interface GuiAddUploadButtonMessage { type GuiAddSliderMessage (line 596) | interface GuiAddSliderMessage { type GuiAddMultiSliderMessage (line 616) | interface GuiAddMultiSliderMessage { type GuiAddNumberMessage (line 638) | interface GuiAddNumberMessage { type GuiAddRgbMessage (line 657) | interface GuiAddRgbMessage { type GuiAddRgbaMessage (line 672) | interface GuiAddRgbaMessage { type GuiAddCheckboxMessage (line 687) | interface GuiAddCheckboxMessage { type GuiAddVector2Message (line 702) | interface GuiAddVector2Message { type GuiAddVector3Message (line 721) | interface GuiAddVector3Message { type GuiAddTextMessage (line 740) | interface GuiAddTextMessage { type GuiAddDropdownMessage (line 755) | interface GuiAddDropdownMessage { type GuiAddButtonGroupMessage (line 771) | interface GuiAddButtonGroupMessage { type GuiModalMessage (line 787) | interface GuiModalMessage { type GuiCloseModalMessage (line 797) | interface GuiCloseModalMessage { type GuiRemoveMessage (line 805) | interface GuiRemoveMessage { type GuiUpdateMessage (line 813) | interface GuiUpdateMessage { type ThemeConfigurationMessage (line 822) | interface ThemeConfigurationMessage { type CatmullRomSplineMessage (line 863) | interface CatmullRomSplineMessage { type CubicBezierSplineMessage (line 878) | interface CubicBezierSplineMessage { type GaussianSplatsMessage (line 891) | interface GaussianSplatsMessage { type GetRenderRequestMessage (line 900) | interface GetRenderRequestMessage { type GetRenderResponseMessage (line 911) | interface GetRenderResponseMessage { type FileTransferStart (line 919) | interface FileTransferStart { type FileTransferPart (line 932) | interface FileTransferPart { type FileTransferPartAck (line 943) | interface FileTransferPartAck { type ShareUrlRequest (line 954) | interface ShareUrlRequest { type ShareUrlUpdated (line 961) | interface ShareUrlUpdated { type ShareUrlDisconnect (line 969) | interface ShareUrlDisconnect { type SetGuiPanelLabelMessage (line 976) | interface SetGuiPanelLabelMessage { type Message (line 981) | type Message = type GuiAddComponentMessage (line 1053) | type GuiAddComponentMessage = FILE: viser/src/viser/client/src/WebsocketServerWorker.ts type WsWorkerIncoming (line 5) | type WsWorkerIncoming = type WsWorkerOutgoing (line 10) | type WsWorkerOutgoing = function collectArrayBuffers (line 16) | function collectArrayBuffers(obj: any, buffers: Set) { FILE: viser/src/viser/client/src/WorldTransformUtils.ts function computeT_threeworld_world (line 7) | function computeT_threeworld_world(viewer: ViewerContextContents) { function rayToViserCoords (line 22) | function rayToViserCoords( FILE: viser/src/viser/client/src/components/Button.tsx function ButtonComponent (line 9) | function ButtonComponent({ FILE: viser/src/viser/client/src/components/ButtonGroup.tsx function ButtonGroupComponent (line 7) | function ButtonGroupComponent({ FILE: viser/src/viser/client/src/components/Checkbox.tsx function CheckboxComponent (line 7) | function CheckboxComponent({ FILE: viser/src/viser/client/src/components/Dropdown.tsx function DropdownComponent (line 7) | function DropdownComponent({ FILE: viser/src/viser/client/src/components/Folder.tsx function FolderComponent (line 10) | function FolderComponent({ FILE: viser/src/viser/client/src/components/Markdown.tsx function MarkdownComponent (line 6) | function MarkdownComponent({ FILE: viser/src/viser/client/src/components/MultiSlider.tsx function MultiSliderComponent (line 9) | function MultiSliderComponent({ FILE: viser/src/viser/client/src/components/MultiSliderPrimitive/Marks/Marks.tsx type MarksProps (line 6) | interface MarksProps { function Marks (line 16) | function Marks({ FILE: viser/src/viser/client/src/components/MultiSliderPrimitive/MultiSlider/MultiSlider.tsx type SliderProps (line 35) | interface SliderProps type SliderFactory (line 115) | type SliderFactory = Factory<{ function getClientPosition (line 286) | function getClientPosition(event: any) { FILE: viser/src/viser/client/src/components/MultiSliderPrimitive/Slider.context.ts type SliderStylesNames (line 3) | type SliderStylesNames = type SliderCssVariables (line 14) | type SliderCssVariables = { type SliderContextValue (line 22) | interface SliderContextValue { FILE: viser/src/viser/client/src/components/MultiSliderPrimitive/SliderRoot/SliderRoot.tsx type SliderRootProps (line 12) | interface SliderRootProps extends BoxProps, ElementProps<"div"> { FILE: viser/src/viser/client/src/components/MultiSliderPrimitive/Thumb/Thumb.tsx type ThumbProps (line 6) | interface ThumbProps { FILE: viser/src/viser/client/src/components/MultiSliderPrimitive/Track/Track.tsx type TrackProps (line 6) | interface TrackProps { function Track (line 20) | function Track({ FILE: viser/src/viser/client/src/components/MultiSliderPrimitive/utils/get-change-value/get-change-value.ts type GetChangeValue (line 1) | interface GetChangeValue { function getChangeValue (line 10) | function getChangeValue({ FILE: viser/src/viser/client/src/components/MultiSliderPrimitive/utils/get-client-position/get-client-position.ts function getClientPosition (line 1) | function getClientPosition(event: any) { FILE: viser/src/viser/client/src/components/MultiSliderPrimitive/utils/get-floating-value/get-gloating-value.ts function getFloatingValue (line 1) | function getFloatingValue(value: number, precision: number) { FILE: viser/src/viser/client/src/components/MultiSliderPrimitive/utils/get-position/get-position.ts type GetPosition (line 1) | interface GetPosition { function getPosition (line 7) | function getPosition({ value, min, max }: GetPosition) { FILE: viser/src/viser/client/src/components/MultiSliderPrimitive/utils/get-precision/get-precision.ts function getPrecision (line 1) | function getPrecision(step: number) { FILE: viser/src/viser/client/src/components/NumberInput.tsx function NumberInputComponent (line 7) | function NumberInputComponent({ FILE: viser/src/viser/client/src/components/PlotlyComponent.tsx function PlotlyComponent (line 84) | function PlotlyComponent({ FILE: viser/src/viser/client/src/components/ProgressBar.tsx function ProgressBarComponent (line 4) | function ProgressBarComponent({ FILE: viser/src/viser/client/src/components/Rgb.tsx function RgbComponent (line 8) | function RgbComponent({ FILE: viser/src/viser/client/src/components/Rgba.tsx function RgbaComponent (line 8) | function RgbaComponent({ FILE: viser/src/viser/client/src/components/Slider.tsx function SliderComponent (line 13) | function SliderComponent({ FILE: viser/src/viser/client/src/components/TabGroup.tsx function TabGroupComponent (line 7) | function TabGroupComponent({ FILE: viser/src/viser/client/src/components/TextInput.tsx function TextInputComponent (line 7) | function TextInputComponent(props: GuiAddTextMessage) { FILE: viser/src/viser/client/src/components/UploadButton.tsx function UploadButtonComponent (line 12) | function UploadButtonComponent(conf: GuiAddUploadButtonMessage) { function useFileUpload (line 64) | function useFileUpload({ FILE: viser/src/viser/client/src/components/Vector2.tsx function Vector2Component (line 6) | function Vector2Component({ FILE: viser/src/viser/client/src/components/Vector3.tsx function Vector3Component (line 6) | function Vector3Component({ FILE: viser/src/viser/client/src/components/common.tsx function ViserInputComponent (line 5) | function ViserInputComponent({ function LabeledInput (line 52) | function LabeledInput(props: { function VectorInput (line 86) | function VectorInput( FILE: viser/src/viser/client/src/components/utils.tsx function rgbToHex (line 3) | function rgbToHex([r, g, b]: [number, number, number]): string { function hexToRgb (line 10) | function hexToRgb(hexColor: string): [number, number, number] { function rgbaToHex (line 17) | function rgbaToHex([r, g, b, a]: [ function hexToRgba (line 30) | function hexToRgba(hexColor: string): [number, number, number, number] { FILE: viser/src/viser/extras/_record3d.py class Record3dLoader (line 17) | class Record3dLoader: method __init__ (line 23) | def __init__(self, data_dir: Path): method num_frames (line 53) | def num_frames(self) -> int: method get_frame (line 56) | def get_frame(self, index: int) -> Record3dFrame: class Record3dFrame (line 91) | class Record3dFrame: method get_point_cloud (line 100) | def get_point_cloud( FILE: viser/src/viser/extras/_record3d_customized.py class Record3dLoader_Customized (line 18) | class Record3dLoader_Customized: method __init__ (line 21) | def __init__(self, data_dir: Path, conf_threshold: float = 1.0, foregr... method num_frames (line 77) | def num_frames(self) -> int: method get_frame (line 80) | def get_frame(self, index: int) -> Record3dFrame: class Record3dFrame (line 146) | class Record3dFrame: method get_point_cloud (line 159) | def get_point_cloud( FILE: viser/src/viser/extras/_record3d_customized_megasam.py class Record3dLoader_Customized_Megasam (line 18) | class Record3dLoader_Customized_Megasam: method __init__ (line 21) | def __init__(self, npz_data: dict, conf_threshold: float = 1.0, foregr... method num_frames (line 48) | def num_frames(self) -> int: method get_frame (line 51) | def get_frame(self, index: int) -> Record3dFrame: class Record3dFrame (line 98) | class Record3dFrame: method get_point_cloud (line 111) | def get_point_cloud( FILE: viser/src/viser/extras/_urdf.py class ViserUrdf (line 16) | class ViserUrdf: method __init__ (line 27) | def __init__( method remove (line 97) | def remove(self) -> None: method update_cfg (line 106) | def update_cfg(self, configuration: onp.ndarray) -> None: method get_actuated_joint_limits (line 118) | def get_actuated_joint_limits( method get_actuated_joint_names (line 134) | def get_actuated_joint_names(self) -> Tuple[str, ...]: function _viser_name_from_frame (line 139) | def _viser_name_from_frame( FILE: viser/src/viser/extras/colmap/_colmap_utils.py class CameraModel (line 43) | class CameraModel: class Camera (line 50) | class Camera: class BaseImage (line 59) | class BaseImage: class Point3D (line 70) | class Point3D: class Image (line 79) | class Image(BaseImage): method qvec2rotmat (line 80) | def qvec2rotmat(self): function read_next_bytes (line 102) | def read_next_bytes(fid, num_bytes, format_char_sequence, endian_charact... function read_cameras_text (line 114) | def read_cameras_text(path: Union[str, Path]) -> Dict[int, Camera]: function read_cameras_binary (line 140) | def read_cameras_binary(path_to_model_file: Union[str, Path]) -> Dict[in... function read_images_text (line 173) | def read_images_text(path: Union[str, Path]) -> Dict[int, Image]: function read_images_binary (line 210) | def read_images_binary(path_to_model_file: Union[str, Path]) -> Dict[int... function read_points3D_text (line 256) | def read_points3D_text(path: Union[str, Path]): function read_points3d_binary (line 288) | def read_points3d_binary(path_to_model_file: Union[str, Path]) -> Dict[i... function qvec2rotmat (line 326) | def qvec2rotmat(qvec): FILE: viser/src/viser/infra/_async_message_buffer.py class AsyncMessageBuffer (line 13) | class AsyncMessageBuffer: method push (line 34) | def push(self, message: Message) -> None: method flush (line 59) | def flush(self) -> None: method set_done (line 63) | def set_done(self) -> None: method window_generator (line 73) | async def window_generator( FILE: viser/src/viser/infra/_infra.py class _ClientHandleState (line 31) | class _ClientHandleState: class RecordHandle (line 42) | class RecordHandle: method __init__ (line 47) | def __init__( method _insert_message (line 56) | def _insert_message(self, message: Message) -> None: method insert_sleep (line 64) | def insert_sleep(self, duration: float) -> None: method set_loop_start (line 68) | def set_loop_start(self) -> None: method end_and_serialize (line 74) | def end_and_serialize(self) -> bytes: class WebsockMessageHandler (line 89) | class WebsockMessageHandler: method __init__ (line 92) | def __init__(self, thread_executor: ThreadPoolExecutor) -> None: method start_recording (line 104) | def start_recording(self, filter: Callable[[Message], bool]) -> Record... method register_handler (line 111) | def register_handler( method unregister_handler (line 121) | def unregister_handler( method _handle_incoming_message (line 135) | def _handle_incoming_message(self, client_id: ClientId, message: Messa... method unsafe_send_message (line 142) | def unsafe_send_message(self, message: Message) -> None: ... method queue_message (line 144) | def queue_message(self, message: Message) -> None: method atomic (line 165) | def atomic(self) -> Generator[None, None, None]: class WebsockClientConnection (line 192) | class WebsockClientConnection(WebsockMessageHandler): method __init__ (line 196) | def __init__( method unsafe_send_message (line 207) | def unsafe_send_message(self, message: Message) -> None: class WebsockServer (line 212) | class WebsockServer(WebsockMessageHandler): method __init__ (line 233) | def __init__( method start (line 259) | def start(self) -> None: method stop (line 276) | def stop(self) -> None: method on_client_connect (line 283) | def on_client_connect(self, cb: Callable[[WebsockClientConnection], An... method on_client_disconnect (line 287) | def on_client_disconnect( method unsafe_send_message (line 294) | def unsafe_send_message(self, message: Message) -> None: method flush (line 302) | def flush(self) -> None: method flush_client (line 308) | def flush_client(self, client_id: int) -> None: method _background_worker (line 313) | def _background_worker(self, ready_sem: threading.Semaphore) -> None: function _message_producer (line 499) | async def _message_producer( function _message_consumer (line 524) | async def _message_consumer( function error_print_wrapper (line 537) | def error_print_wrapper(inner: Callable[[], Any]) -> Callable[[], None]: FILE: viser/src/viser/infra/_messages.py function _prepare_for_deserialization (line 21) | def _prepare_for_deserialization(value: Any, annotation: Type) -> Any: function _prepare_for_serialization (line 47) | def _prepare_for_serialization(value: Any, annotation: object) -> Any: function get_type_hints_cached (line 101) | def get_type_hints_cached(cls: Type[Any]) -> Dict[str, Any]: class Message (line 105) | class Message(abc.ABC): method as_serializable_dict (line 112) | def as_serializable_dict(self) -> Dict[str, Any]: method _from_serializable_dict (line 123) | def _from_serializable_dict(cls, mapping: Dict[str, Any]) -> Dict[str,... method deserialize (line 134) | def deserialize(cls, message: bytes) -> Message: method _subclass_from_type_string (line 155) | def _subclass_from_type_string(cls: Type[T]) -> Dict[str, Type[T]]: method get_subclasses (line 160) | def get_subclasses(cls: Type[T]) -> List[Type[T]]: method redundancy_key (line 173) | def redundancy_key(self) -> str: FILE: viser/src/viser/infra/_typescript_interface_gen.py function _get_ts_type (line 37) | def _get_ts_type(typ: Type[Any]) -> str: class TypeScriptAnnotationOverride (line 109) | class TypeScriptAnnotationOverride: function generate_typescript_interfaces (line 116) | def generate_typescript_interfaces(message_cls: Type[Message]) -> str: FILE: viser/src/viser/scripts/dev_checks.py function run_command (line 17) | def run_command(command: str, continue_on_fail: bool = False) -> bool: function run_code_checks (line 32) | def run_code_checks( function entrypoint (line 79) | def entrypoint(): FILE: viser/src/viser/theme/_titlebar.py class TitlebarButton (line 4) | class TitlebarButton(TypedDict): class TitlebarImage (line 12) | class TitlebarImage(TypedDict): class TitlebarConfig (line 21) | class TitlebarConfig(TypedDict): FILE: viser/src/viser/transforms/_base.py class MatrixLieGroup (line 9) | class MatrixLieGroup(abc.ABC): method __init__ (line 27) | def __init__( method __matmul__ (line 42) | def __matmul__(self, other: Self) -> Self: ... method __matmul__ (line 45) | def __matmul__( method __matmul__ (line 49) | def __matmul__( method identity (line 69) | def identity(cls, batch_axes: Tuple[int, ...] = ()) -> Self: method from_matrix (line 81) | def from_matrix(cls, matrix: onpt.NDArray[onp.floating]) -> Self: method as_matrix (line 94) | def as_matrix(self) -> onpt.NDArray[onp.floating]: method parameters (line 98) | def parameters(self) -> onpt.NDArray[onp.floating]: method apply (line 104) | def apply(self, target: onpt.NDArray[onp.floating]) -> onpt.NDArray[on... method multiply (line 115) | def multiply(self, other: Self) -> Self: method exp (line 124) | def exp(cls, tangent: onpt.NDArray[onp.floating]) -> Self: method log (line 135) | def log(self) -> onpt.NDArray[onp.floating]: method adjoint (line 143) | def adjoint(self) -> onpt.NDArray[onp.floating]: method inverse (line 160) | def inverse(self) -> Self: method normalize (line 168) | def normalize(self) -> Self: method get_batch_axes (line 191) | def get_batch_axes(self) -> Tuple[int, ...]: class SOBase (line 198) | class SOBase(MatrixLieGroup): class SEBase (line 205) | class SEBase(Generic[ContainedSOType], MatrixLieGroup): method from_rotation_and_translation (line 216) | def from_rotation_and_translation( method from_rotation (line 233) | def from_rotation(cls, rotation: ContainedSOType) -> Self: method from_translation (line 244) | def from_translation(cls, translation: onpt.NDArray[onp.floating]) -> ... method rotation (line 253) | def rotation(self) -> ContainedSOType: method translation (line 257) | def translation(self) -> onpt.NDArray[onp.floating]: method apply (line 264) | def apply(self, target: onpt.NDArray[onp.floating]) -> onpt.NDArray[on... method multiply (line 269) | def multiply(self, other: Self) -> Self: method inverse (line 277) | def inverse(self) -> Self: method normalize (line 286) | def normalize(self) -> Self: FILE: viser/src/viser/transforms/_se2.py class SE2 (line 20) | class SE2(_base.SEBase[SO2]): method __repr__ (line 36) | def __repr__(self) -> str: method from_xy_theta (line 42) | def from_xy_theta(x: hints.Scalar, y: hints.Scalar, theta: hints.Scala... method from_rotation_and_translation (line 55) | def from_rotation_and_translation( method rotation (line 69) | def rotation(self) -> SO2: method translation (line 73) | def translation(self) -> onpt.NDArray[onp.floating]: method identity (line 80) | def identity(cls, batch_axes: Tuple[int, ...] = ()) -> "SE2": method from_matrix (line 89) | def from_matrix(cls, matrix: onpt.NDArray[onp.floating]) -> "SE2": method parameters (line 100) | def parameters(self) -> onpt.NDArray[onp.floating]: method as_matrix (line 104) | def as_matrix(self) -> onpt.NDArray[onp.floating]: method exp (line 126) | def exp(cls, tangent: onpt.NDArray[onp.floating]) -> "SE2": method log (line 181) | def log(self) -> onpt.NDArray[onp.floating]: method adjoint (line 230) | def adjoint(self: "SE2") -> onpt.NDArray[onp.floating]: FILE: viser/src/viser/transforms/_se3.py function _skew (line 15) | def _skew(omega: onpt.NDArray[onp.floating]) -> onpt.NDArray[onp.floating]: class SE3 (line 33) | class SE3(_base.SEBase[SO3]): method __repr__ (line 49) | def __repr__(self) -> str: method from_rotation_and_translation (line 58) | def from_rotation_and_translation( method rotation (line 68) | def rotation(self) -> SO3: method translation (line 72) | def translation(self) -> onpt.NDArray[onp.floating]: method identity (line 79) | def identity(cls, batch_axes: Tuple[int, ...] = ()) -> SE3: method from_matrix (line 88) | def from_matrix(cls, matrix: onpt.NDArray[onp.floating]) -> SE3: method as_matrix (line 99) | def as_matrix(self) -> onpt.NDArray[onp.floating]: method parameters (line 107) | def parameters(self) -> onpt.NDArray[onp.floating]: method exp (line 114) | def exp(cls, tangent: onpt.NDArray[onp.floating]) -> SE3: method log (line 161) | def log(self) -> onpt.NDArray[onp.floating]: method adjoint (line 206) | def adjoint(self) -> onpt.NDArray[onp.floating]: FILE: viser/src/viser/transforms/_so2.py class SO2 (line 21) | class SO2(_base.SOBase): method __repr__ (line 36) | def __repr__(self) -> str: method from_radians (line 41) | def from_radians(theta: hints.Scalar) -> SO2: method as_radians (line 47) | def as_radians(self) -> onpt.NDArray[onp.floating]: method identity (line 56) | def identity(cls, batch_axes: Tuple[int, ...] = ()) -> SO2: method from_matrix (line 65) | def from_matrix(cls, matrix: onpt.NDArray[onp.floating]) -> SO2: method as_matrix (line 72) | def as_matrix(self) -> onpt.NDArray[onp.floating]: method parameters (line 87) | def parameters(self) -> onpt.NDArray[onp.floating]: method apply (line 93) | def apply(self, target: onpt.NDArray[onp.floating]) -> onpt.NDArray[on... method multiply (line 99) | def multiply(self, other: SO2) -> SO2: method exp (line 108) | def exp(cls, tangent: onpt.NDArray[onp.floating]) -> SO2: method log (line 115) | def log(self) -> onpt.NDArray[onp.floating]: method adjoint (line 121) | def adjoint(self) -> onpt.NDArray[onp.floating]: method inverse (line 125) | def inverse(self) -> SO2: method normalize (line 129) | def normalize(self) -> SO2: FILE: viser/src/viser/transforms/_so3.py class RollPitchYaw (line 15) | class RollPitchYaw: class SO3 (line 30) | class SO3(_base.SOBase): method __repr__ (line 44) | def __repr__(self) -> str: method from_x_radians (line 49) | def from_x_radians(theta: hints.Scalar) -> SO3: method from_y_radians (line 62) | def from_y_radians(theta: hints.Scalar) -> SO3: method from_z_radians (line 75) | def from_z_radians(theta: hints.Scalar) -> SO3: method from_rpy_radians (line 88) | def from_rpy_radians( method from_quaternion_xyzw (line 111) | def from_quaternion_xyzw(xyzw: onpt.NDArray[onp.floating]) -> SO3: method as_quaternion_xyzw (line 126) | def as_quaternion_xyzw(self) -> onpt.NDArray[onp.floating]: method as_rpy_radians (line 130) | def as_rpy_radians(self) -> RollPitchYaw: method compute_roll_radians (line 142) | def compute_roll_radians(self) -> onpt.NDArray[onp.floating]: method compute_pitch_radians (line 152) | def compute_pitch_radians(self) -> onpt.NDArray[onp.floating]: method compute_yaw_radians (line 162) | def compute_yaw_radians(self) -> onpt.NDArray[onp.floating]: method identity (line 176) | def identity(cls, batch_axes: Tuple[int, ...] = ()) -> SO3: method from_matrix (line 183) | def from_matrix(cls, matrix: onpt.NDArray[onp.floating]) -> SO3: method as_matrix (line 287) | def as_matrix(self) -> onpt.NDArray[onp.floating]: method parameters (line 307) | def parameters(self) -> onpt.NDArray[onp.floating]: method apply (line 313) | def apply(self, target: onpt.NDArray[onp.floating]) -> onpt.NDArray[on... method multiply (line 324) | def multiply(self, other: SO3) -> SO3: method exp (line 341) | def exp(cls, tangent: onpt.NDArray[onp.floating]) -> SO3: method log (line 385) | def log(self) -> onpt.NDArray[onp.floating]: method adjoint (line 420) | def adjoint(self) -> onpt.NDArray[onp.floating]: method inverse (line 424) | def inverse(self) -> SO3: method normalize (line 429) | def normalize(self) -> SO3: FILE: viser/src/viser/transforms/utils/_utils.py function get_epsilon (line 12) | def get_epsilon(dtype: onp.dtype) -> float: function register_lie_group (line 29) | def register_lie_group( function broadcast_leading_axes (line 58) | def broadcast_leading_axes(inputs: TupleOfBroadcastable) -> TupleOfBroad... FILE: viser/visualize_megasam.py function main (line 20) | def main( FILE: viser/visualize_pose.py class Pose (line 27) | class Pose: method __call__ (line 33) | def __call__(self, R=None, t=None): method invert (line 58) | def invert(self, pose, use_inverse=False): method compose (line 68) | def compose(self, pose_list): method compose_pair (line 78) | def compose_pair(self, pose_a, pose_b): method scale_center (line 89) | def scale_center(self, pose, scale): function to_hom (line 100) | def to_hom(X): function cam2world (line 108) | def cam2world(X, pose): function get_camera_mesh (line 117) | def get_camera_mesh(pose, depth=1): function merge_wireframes_plotly (line 160) | def merge_wireframes_plotly(wireframe): function merge_meshes (line 169) | def merge_meshes(vertices, faces): function unbind_np (line 181) | def unbind_np(array, axis=0): function plotly_visualize_pose (line 195) | def plotly_visualize_pose( function compute_optimal_camera_view (line 285) | def compute_optimal_camera_view(poses): function compute_multiple_camera_views (line 425) | def compute_multiple_camera_views(poses): function add_view_selector_to_html (line 508) | def add_view_selector_to_html(html_str, views): function write_html (line 605) | def write_html(poses, file, vis_depth=1, xyz_length=0.2, center_size=0.0... function plotly_visualize_pose_animated (line 841) | def plotly_visualize_pose_animated( function write_html_animated (line 942) | def write_html_animated( function quaternion_to_matrix (line 1079) | def quaternion_to_matrix(quaternions, eps: float = 1e-8): function pose_from_quaternion (line 1109) | def pose_from_quaternion(pose): function viz_poses (line 1133) | def viz_poses(i, pth, file, args):