SYMBOL INDEX (859 symbols across 91 files) FILE: client/src/audio-processor.ts function asMs (line 2) | function asMs(samples) { function asSamples (line 6) | function asSamples(mili) { class MoshiProcessor (line 10) | class MoshiProcessor extends AudioWorkletProcessor { method constructor (line 11) | constructor() { method initState (line 80) | initState() { method totalMaxBufferSamples (line 101) | totalMaxBufferSamples() { method timestamp (line 105) | timestamp() { method currentSamples (line 109) | currentSamples() { method resetStart (line 118) | resetStart() { method start (line 122) | start() { method canPlay (line 128) | canPlay() { method process (line 132) | process(inputs, outputs, parameters) { FILE: client/src/components/Button/Button.tsx type ButtonProps (line 3) | type ButtonProps = React.ButtonHTMLAttributes; FILE: client/src/components/ImageGallery/ImageGallery.tsx type ImageGalleryProps (line 56) | type ImageGalleryProps = React.InputHTMLAttributes & { type ImageItemProps (line 65) | type ImageItemProps = React.InputHTMLAttributes & { function ImageSelect (line 77) | function ImageSelect(props: ImageItemProps) { function handleShuffle (line 144) | function handleShuffle() { FILE: client/src/components/Input/Input.tsx type InputProps (line 1) | type InputProps = React.InputHTMLAttributes & { FILE: client/src/env.ts type ENV (line 1) | type ENV = { FILE: client/src/pages/Conversation/Conversation.tsx type ConversationProps (line 18) | type ConversationProps = { FILE: client/src/pages/Conversation/MediaContext.ts type MediaContextType (line 2) | type MediaContextType = { FILE: client/src/pages/Conversation/SocketContext.ts type SocketContextType (line 4) | type SocketContextType = { FILE: client/src/pages/Conversation/components/AudioVisualizer/AudioVisualizer.tsx type AudioVisualizerProps (line 3) | type AudioVisualizerProps = { FILE: client/src/pages/Conversation/components/AudioVisualizer/ClientVisualizer.tsx type AudioVisualizerProps (line 4) | type AudioVisualizerProps = { constant MAX_INTENSITY (line 10) | const MAX_INTENSITY = 255; constant COLORS (line 12) | const COLORS = [ FILE: client/src/pages/Conversation/components/AudioVisualizer/ServerVisualizer.tsx type AudioVisualizerProps (line 5) | type AudioVisualizerProps = { constant MAX_INTENSITY (line 12) | const MAX_INTENSITY = 255; FILE: client/src/pages/Conversation/components/ModelParams/ModelParams.tsx type ModelParamsProps (line 5) | type ModelParamsProps = { FILE: client/src/pages/Conversation/components/ServerAudio/ServerAudio.tsx type ServerAudioProps (line 5) | type ServerAudioProps = { FILE: client/src/pages/Conversation/components/ServerAudio/ServerAudioStats.tsx type ServerAudioStatsProps (line 3) | type ServerAudioStatsProps = { FILE: client/src/pages/Conversation/components/ServerInfo/ServerInfo.tsx function pretty_format (line 3) | function pretty_format(num: number): number { FILE: client/src/pages/Conversation/components/TextDisplay/TextDisplay.tsx type TextDisplayProps (line 4) | type TextDisplayProps = { function clamp_color (line 19) | function clamp_color(v: number) { FILE: client/src/pages/Conversation/components/TextDisplay/TextDisplayStats.tsx type TextDisplayStatsProps (line 3) | type TextDisplayStatsProps = { FILE: client/src/pages/Conversation/components/UserAudio/UserAudio.tsx type UserAudioProps (line 6) | type UserAudioProps = { FILE: client/src/pages/Conversation/components/UserAudio/UserAudioStats.tsx type UserAudioStatsProps (line 3) | type UserAudioStatsProps = { FILE: client/src/pages/Conversation/hooks/useModelParams.ts constant DEFAULT_TEXT_TEMPERATURE (line 3) | const DEFAULT_TEXT_TEMPERATURE = 0.45; constant DEFAULT_TEXT_TOPK (line 4) | const DEFAULT_TEXT_TOPK = 25; constant DEFAULT_AUDIO_TEMPERATURE (line 5) | const DEFAULT_AUDIO_TEMPERATURE = 0.7; constant DEFAULT_AUDIO_TOPK (line 6) | const DEFAULT_AUDIO_TOPK = 250; constant DEFAULT_PAD_MULT (line 7) | const DEFAULT_PAD_MULT = 0; constant DEFAULT_REPETITION_PENALTY_CONTEXT (line 8) | const DEFAULT_REPETITION_PENALTY_CONTEXT = 64; constant DEFAULT_REPETITION_PENALTY (line 9) | const DEFAULT_REPETITION_PENALTY = 1.15; constant DEFAULT_IMAGE_RESOLUTION (line 10) | const DEFAULT_IMAGE_RESOLUTION = 448; constant DEFAULT_IMAGE_URL (line 11) | const DEFAULT_IMAGE_URL = undefined; constant DEFAULT_GATE_DELAY (line 12) | const DEFAULT_GATE_DELAY = 16; constant DEFAULT_GATE_INFLUENCE (line 13) | const DEFAULT_GATE_INFLUENCE = 0.0; constant DEFAULT_DISPLAY_COLOR (line 14) | const DEFAULT_DISPLAY_COLOR = true; constant DEFAULT_CENTER_CROP (line 15) | const DEFAULT_CENTER_CROP = false; type ModelParamsValues (line 17) | type ModelParamsValues = { function importantSettingsHaveChanged (line 33) | function importantSettingsHaveChanged(params: ModelParamsValues): boolean { type useModelParamsArgs (line 47) | type useModelParamsArgs = Partial; FILE: client/src/pages/Conversation/hooks/useServerAudio.ts type AudioStats (line 7) | type AudioStats = { type useServerAudioArgs (line 16) | type useServerAudioArgs = { type WorkletStats (line 20) | type WorkletStats = { FILE: client/src/pages/Conversation/hooks/useServerInfo.ts type ServerInfo (line 42) | type ServerInfo = { FILE: client/src/pages/Conversation/hooks/useSocket.ts function sendImage (line 39) | async function sendImage() { function fetchImageBytes (line 136) | async function fetchImageBytes(imageUrl: string) { FILE: client/src/pages/Conversation/hooks/useUserAudio.ts type UserMediaStatuses (line 6) | enum UserMediaStatuses { type useUserAudioArgs (line 16) | type useUserAudioArgs = { FILE: client/src/pages/Queue/Queue.tsx type Status (line 15) | type Status = "connecting" | "in_queue" | "has_credentials" | "error" | ... function getFloatFromStorage (line 18) | function getFloatFromStorage(val: string | null) { function getIntFromStorage (line 22) | function getIntFromStorage(val: string | null) { function getBoolFromStage (line 26) | function getBoolFromStage(val: string | null) { FILE: client/src/pages/Queue/api/errors/api_error.ts class APIError (line 1) | class APIError extends Error { method constructor (line 4) | constructor(message:string, status:number) { FILE: client/src/pages/Queue/api/errors/response_error.ts class ResponseError (line 1) | class ResponseError extends Error { method constructor (line 2) | constructor(message:string) { FILE: client/src/protocol/types.ts type MessageType (line 1) | type MessageType = constant VERSIONS_MAP (line 9) | const VERSIONS_MAP = { constant MODELS_MAP (line 13) | const MODELS_MAP = { type VERSION (line 17) | type VERSION = keyof typeof VERSIONS_MAP; type MODEL (line 19) | type MODEL = keyof typeof MODELS_MAP; type WSMessage (line 21) | type WSMessage = constant CONTROL_MESSAGES_MAP (line 64) | const CONTROL_MESSAGES_MAP = { type CONTROL_MESSAGE (line 71) | type CONTROL_MESSAGE = keyof typeof CONTROL_MESSAGES_MAP; FILE: kyuteye_mlx/kyuteye_mlx/benchmark.py function main (line 9) | def main(): FILE: kyuteye_mlx/kyuteye_mlx/local_web.py class ModelInput (line 46) | class ModelInput(Enum): class ModelOutput (line 51) | class ModelOutput(Enum): class ServerMediaInput (line 58) | class ServerMediaInput(Enum): function colorize (line 67) | def colorize(text: str, color: str) -> str: function log (line 73) | def log(level: str, msg: str) -> None: function hf_hub_download (line 85) | def hf_hub_download(repo: str | None, path: str) -> str: function full_warmup (line 91) | def full_warmup( function get_model_file (line 119) | def get_model_file(args) -> str: function get_tokenizer (line 131) | def get_tokenizer(args) -> sentencepiece.SentencePieceProcessor: function get_embedder (line 139) | def get_embedder(args) -> SiglipWrapper | PixtralWrapper: function get_model (line 173) | def get_model(args, load_weights: bool = True) -> models.LmGen: function model_server (line 214) | def model_server( function handle_audio (line 266) | def handle_audio( function predict_text_and_audio (line 291) | def predict_text_and_audio( function web_server (line 305) | def web_server( function get_args_for_main (line 528) | def get_args_for_main() -> argparse.Namespace: function main (line 564) | def main() -> None: function sanity_check (line 593) | def sanity_check() -> None: FILE: kyuteye_mlx/kyuteye_mlx/mlx_vlm/models/pixtral/vision.py class VisionConfig (line 9) | class VisionConfig: method from_dict (line 25) | def from_dict(cls, params): function position_ids_in_meshgrid (line 29) | def position_ids_in_meshgrid(patch_embeds_list: list[mx.array], max_widt... function generate_block_attention_mask (line 41) | def generate_block_attention_mask(patch_embeds_list: list[mx.array], ten... function rotate_half (line 59) | def rotate_half(x): function apply_rotary_pos_emb (line 65) | def apply_rotary_pos_emb(q, k, cos, sin, unsqueeze_dim=1): class Attention (line 73) | class Attention(nn.Module): method __init__ (line 74) | def __init__( method __call__ (line 110) | def __call__(self, queries, keys, values, position_embeddings, mask=No... class MLP (line 138) | class MLP(nn.Module): method __init__ (line 139) | def __init__(self, config: VisionConfig): method __call__ (line 147) | def __call__(self, x) -> mx.array: class EncoderLayer (line 151) | class EncoderLayer(nn.Module): method __init__ (line 152) | def __init__(self, config: VisionConfig): method __call__ (line 160) | def __call__( class Encoder (line 174) | class Encoder(nn.Module): method __init__ (line 175) | def __init__(self, config: VisionConfig): class PixtralRotaryEmbedding (line 180) | class PixtralRotaryEmbedding: method __init__ (line 181) | def __init__(self, config): method __call__ (line 202) | def __call__(self, x, position_ids): class PixtralVisionModel (line 210) | class PixtralVisionModel(nn.Module): method __init__ (line 211) | def __init__(self, config: VisionConfig): method __call__ (line 225) | def __call__( FILE: kyuteye_mlx/kyuteye_mlx/mlx_vlm/models/siglip/vision.py class VisionConfig (line 10) | class VisionConfig: method from_dict (line 23) | def from_dict(cls, params): function check_array_shape (line 27) | def check_array_shape(arr): class Attention (line 43) | class Attention(nn.Module): method __init__ (line 44) | def __init__( method __call__ (line 78) | def __call__(self, x, mask=None): class MLP (line 95) | class MLP(nn.Module): method __init__ (line 96) | def __init__(self, config: VisionConfig) -> None: method __call__ (line 102) | def __call__(self, x: mx.array) -> mx.array: class EncoderLayer (line 109) | class EncoderLayer(nn.Module): method __init__ (line 110) | def __init__(self, config: VisionConfig) -> None: method __call__ (line 118) | def __call__(self, x: mx.array, mask: mx.array | None = None) -> mx.ar... class Encoder (line 125) | class Encoder(nn.Module): method __init__ (line 126) | def __init__(self, config: VisionConfig) -> None: method __call__ (line 130) | def __call__( class VisionEmbeddings (line 148) | class VisionEmbeddings(nn.Module): method __init__ (line 149) | def __init__(self, config: VisionConfig) -> None: method __call__ (line 167) | def __call__(self, x: mx.array) -> mx.array: class SigLipVisionModel (line 176) | class SigLipVisionModel(nn.Module): method __init__ (line 177) | def __init__(self, config: VisionConfig): method __call__ (line 183) | def __call__( class VisionModel (line 197) | class VisionModel(nn.Module): method __init__ (line 198) | def __init__(self, config: VisionConfig) -> None: method __call__ (line 206) | def __call__(self, x: mx.array, output_hidden_states: bool | None = No... method sanitize (line 209) | def sanitize(self, weights): FILE: kyuteye_mlx/kyuteye_mlx/models/generate.py class LmGen (line 15) | class LmGen: method __init__ (line 16) | def __init__( method zero_token (line 42) | def zero_token(self) -> int: method ungenerated_token (line 48) | def ungenerated_token(self) -> int: method nb_input_tokens (line 56) | def nb_input_tokens(self) -> int: method step (line 60) | def step( method last_audio_tokens (line 105) | def last_audio_tokens(self) -> Int32[mx.array, "1 {self.nb_input_token... method reset (line 116) | def reset(self) -> None: FILE: kyuteye_mlx/kyuteye_mlx/models/lm.py class DepFormerConfig (line 20) | class DepFormerConfig: class LmConfig (line 26) | class LmConfig: method audio_eos_token (line 36) | def audio_eos_token(self) -> int: method audio_padding_token (line 40) | def audio_padding_token(self) -> int: class DepFormerSlice (line 44) | class DepFormerSlice(nn.Module): method __init__ (line 45) | def __init__( method __call__ (line 60) | def __call__(self, _: mx.array) -> mx.array: class DepFormer (line 64) | class DepFormer(nn.Module): method __init__ (line 65) | def __init__(self, cfg: LmConfig): method __call__ (line 79) | def __call__(self, _: mx.array) -> mx.array: method sample (line 82) | def sample( class Lm (line 111) | class Lm(nn.Module): method __init__ (line 112) | def __init__(self, cfg: LmConfig): method __call__ (line 139) | def __call__( method sample (line 150) | def sample( method warmup (line 180) | def warmup(self) -> None: method reset_all_caches (line 195) | def reset_all_caches(self) -> None: function config1b_202412 (line 204) | def config1b_202412() -> LmConfig: function config1b_202412_16rvq (line 265) | def config1b_202412_16rvq() -> LmConfig: function config_v0_1 (line 326) | def config_v0_1() -> LmConfig: function config_siglip (line 392) | def config_siglip() -> LmConfig: function config_pixtral (line 398) | def config_pixtral() -> LmConfig: function config_helium_1_preview_2b (line 404) | def config_helium_1_preview_2b() -> LmConfig: FILE: kyuteye_mlx/kyuteye_mlx/models/pixtral.py class PixtralWrapper (line 10) | class PixtralWrapper(mlx.nn.Module): method __init__ (line 13) | def __init__(self) -> None: method __call__ (line 33) | def __call__(self, x: mx.array) -> mx.array: method warmup (line 48) | def warmup(self) -> None: FILE: kyuteye_mlx/kyuteye_mlx/models/siglip.py class SiglipWrapper (line 11) | class SiglipWrapper(mlx.nn.Module): method __init__ (line 14) | def __init__(self) -> None: method __call__ (line 21) | def __call__(self, x: mx.array) -> mx.array: method warmup (line 35) | def warmup(self) -> None: FILE: kyuteye_mlx/kyuteye_mlx/modules/config.py class TransformerConfig (line 6) | class TransformerConfig: method head_dim (line 34) | def head_dim(self) -> int: FILE: kyuteye_mlx/kyuteye_mlx/modules/cross_attention.py class SharedModuleType (line 17) | class SharedModuleType(type): method __call__ (line 22) | def __call__(cls, *args: Any, **kwargs: Any) -> Any: class CrossAttention (line 28) | class CrossAttention(nn.Module): method __init__ (line 29) | def __init__(self, cfg: TransformerConfig): method __call__ (line 38) | def __call__( class SharedCrossAttention (line 65) | class SharedCrossAttention(CrossAttention, metaclass=SharedModuleType): class XAGate (line 71) | class XAGate(nn.Module): method __init__ (line 72) | def __init__( method __call__ (line 98) | def __call__( class GatedCrossAttention (line 104) | class GatedCrossAttention(nn.Module): method __init__ (line 105) | def __init__(self, cfg: TransformerConfig) -> None: method __call__ (line 113) | def __call__( FILE: kyuteye_mlx/kyuteye_mlx/modules/kv_cache.py class XACache (line 13) | class XACache: method __init__ (line 14) | def __init__(self) -> None: method set (line 19) | def set(self, k: mx.array, v: mx.array) -> None: method reset (line 26) | def reset(self) -> None: method state (line 32) | def state(self) -> tuple[mx.array | None, mx.array | None]: class KVCache (line 36) | class KVCache: method __init__ (line 37) | def __init__(self, head_dim: int | tuple[int, int], n_kv_heads: int) -... method update_and_fetch (line 50) | def update_and_fetch(self, keys: mx.array, values: mx.array) -> tuple[... method reset (line 75) | def reset(self) -> None: method state (line 79) | def state(self) -> tuple[mx.array | None, mx.array | None]: class RotatingKVCache (line 83) | class RotatingKVCache: method __init__ (line 84) | def __init__( method _trim (line 107) | def _trim(self, trim_size: int, v: mx.array, append: mx.array | None =... method update_and_fetch (line 117) | def update_and_fetch(self, keys: mx.array, values: mx.array) -> tuple[... method reset (line 176) | def reset(self) -> None: method state (line 181) | def state(self) -> tuple[mx.array | None, mx.array | None]: class BaseModelArgs (line 186) | class BaseModelArgs: method from_dict (line 188) | def from_dict(cls, params: dict[str, Any]): FILE: kyuteye_mlx/kyuteye_mlx/modules/transformer.py class Attention (line 15) | class Attention(nn.Module): method __init__ (line 16) | def __init__(self, cfg: TransformerConfig) -> None: method __call__ (line 29) | def __call__( class MlpGating (line 59) | class MlpGating(nn.Module): method __init__ (line 60) | def __init__(self, cfg: TransformerConfig) -> None: method __call__ (line 70) | def __call__( class MlpNoGating (line 79) | class MlpNoGating(nn.Module): method __init__ (line 80) | def __init__(self, cfg: TransformerConfig) -> None: method __call__ (line 86) | def __call__(self, xs: mx.array) -> mx.array: class TransformerLayer (line 90) | class TransformerLayer(nn.Module): method __init__ (line 91) | def __init__(self, cfg: TransformerConfig) -> None: method __call__ (line 123) | def __call__( class ImagePrefix (line 140) | class ImagePrefix(nn.Module): method __init__ (line 141) | def __init__(self, cfg: TransformerConfig) -> None: method __call__ (line 147) | def __call__( class Transformer (line 155) | class Transformer(nn.Module): method __init__ (line 156) | def __init__(self, cfg: TransformerConfig, with_img_prefix: bool = Fal... method __call__ (line 164) | def __call__( method make_cache (line 179) | def make_cache(self) -> list[KVCache]: method make_rot_cache (line 183) | def make_rot_cache(self) -> list[RotatingKVCache]: FILE: kyuteye_mlx/kyuteye_mlx/quantize.py function quantize (line 28) | def quantize( function main (line 79) | def main(): FILE: kyuteye_mlx/kyuteye_mlx/utils/loading.py function repeat_shared_weights (line 4) | def repeat_shared_weights(weights: dict[str, mx.array], num_layers: int)... function remove_shared_weights (line 18) | def remove_shared_weights(weights: dict[str, mx.array], num_layers: int)... function split_embedder_weights (line 25) | def split_embedder_weights( FILE: kyuteye_mlx/kyuteye_mlx/utils/profiling.py function profile (line 11) | def profile(x: Callable) -> Callable: FILE: kyuteye_mlx/kyuteye_mlx/utils/sampling.py function top_p_sampling (line 12) | def top_p_sampling( function categorical_sampling (line 48) | def categorical_sampling(logits: BFloat16[mx.array, "batch vocab"], temp... class Sampler (line 53) | class Sampler: method __call__ (line 57) | def __call__( FILE: kyuteye_mlx/tests/test_siglip.py function convert_weights_for_mlx (line 9) | def convert_weights_for_mlx(weights: dict[str, torch.Tensor]) -> dict[st... function test_siglip_weights_conversion (line 22) | def test_siglip_weights_conversion() -> None: FILE: kyuteye_pt/kyuteye/config/enums.py class ImageEncoder (line 8) | class ImageEncoder(Enum): method out_dims (line 26) | def out_dims(self) -> int: method to_rust (line 48) | def to_rust(self) -> str: FILE: kyuteye_pt/kyuteye/config/kyuteye_config.py class KyuteyeConfig (line 23) | class KyuteyeConfig: method __init__ (line 30) | def __init__(self, **kwargs: Any): method __getattribute__ (line 88) | def __getattribute__(self, name: str) -> Any: method __setattr__ (line 94) | def __setattr__(self, name: str, value: Any) -> None: method moshi_constructor_kwargs (line 102) | def moshi_constructor_kwargs(self) -> Dict[str, Any]: method from_yml (line 111) | def from_yml(cls, path: Path | str) -> "KyuteyeConfig": method to_yml (line 115) | def to_yml(self, path: Optional[Path | str] = None) -> None: method print (line 140) | def print(self, flat: bool = False, only: Optional[Sequence[str]] = No... method to_dict (line 163) | def to_dict(self, flat: bool = True) -> Dict[str, Any]: function __load_yaml__ (line 173) | def __load_yaml__(path: Path | str) -> Dict: function __save_yaml__ (line 194) | def __save_yaml__(config: Dict, path: Path | str) -> None: FILE: kyuteye_pt/kyuteye/config/subconfigs.py function __is_nonstring_iterable__ (line 11) | def __is_nonstring_iterable__(arg: Any) -> bool: class LMConfig (line 16) | class LMConfig: method help (line 32) | def help(field_name: str) -> str: class ImageEncoderConfig (line 47) | class ImageEncoderConfig: method __post_init__ (line 62) | def __post_init__(self) -> None: method help (line 73) | def help(field_name: str) -> str: class MoshiConfig (line 86) | class MoshiConfig: method help (line 119) | def help(field_name: str) -> str: class FusionConfig (line 134) | class FusionConfig: method help (line 157) | def help(field_name: str) -> str: method __post_init__ (line 205) | def __post_init__(self) -> None: method crossattention_kwargs (line 262) | def crossattention_kwargs(self) -> Dict[str, Any]: FILE: kyuteye_pt/kyuteye/models/helium.py class Helium (line 12) | class Helium(torch.nn.Module): method __init__ (line 39) | def __init__( method forward (line 111) | def forward( FILE: kyuteye_pt/kyuteye/models/hf_model_configs.py class HeliumConfig (line 9) | class HeliumConfig(PretrainedConfig): method __init__ (line 14) | def __init__( class MoshiVisConfig (line 80) | class MoshiVisConfig(HeliumConfig): method __init__ (line 85) | def __init__( method total_audio_loss_weight (line 163) | def total_audio_loss_weight(self) -> float: method audio_semantic_loss_weight (line 178) | def audio_semantic_loss_weight(self) -> float: method audio_acoustic_loss_weight (line 187) | def audio_acoustic_loss_weight(self) -> float: method audio_other_semantic_loss_weight (line 196) | def audio_other_semantic_loss_weight(self) -> float: method audio_other_acoustic_loss_weight (line 201) | def audio_other_acoustic_loss_weight(self) -> float: method sparsity_loss_weight (line 206) | def sparsity_loss_weight(self) -> float: FILE: kyuteye_pt/kyuteye/models/image_projection.py class ImageProjection (line 19) | class ImageProjection(torch.nn.Module): method __init__ (line 30) | def __init__( method from_config (line 75) | def from_config( method init_proj_module (line 106) | def init_proj_module(self, num_tokens: int) -> Optional[torch.nn.Module]: method encoder_out_dim (line 117) | def encoder_out_dim(self) -> int: method to_tensor_and_normalize (line 122) | def to_tensor_and_normalize(self) -> Callable: method init_norm_module (line 126) | def init_norm_module(self, norm_type: Optional[str]) -> Optional[torch... method forward (line 134) | def forward(self, x: torch.Tensor | List[torch.Tensor]) -> Dict[str, t... method encode (line 156) | def encode( method project_extra (line 176) | def project_extra(self, logits: torch.Tensor) -> torch.Tensor: method project_xa (line 184) | def project_xa(self, logits: torch.Tensor) -> torch.Tensor: FILE: kyuteye_pt/kyuteye/models/loaders.py function get_moshi_vis (line 16) | def get_moshi_vis( FILE: kyuteye_pt/kyuteye/models/moshivis.py class MoshiVis (line 15) | class MoshiVis(StreamingModule): method __init__ (line 23) | def __init__( method cross_attention (line 147) | def cross_attention(self) -> bool: method num_audio_codebooks_in (line 152) | def num_audio_codebooks_in(self) -> int: method num_audio_codebooks_out (line 157) | def num_audio_codebooks_out(self) -> int: method num_codebooks (line 162) | def num_codebooks(self) -> int: method initial_audio_token_id (line 167) | def initial_audio_token_id(self) -> int: method initial_text_token_id (line 172) | def initial_text_token_id(self) -> int: method audio_offset (line 178) | def audio_offset(self) -> int: method forward_text (line 182) | def forward_text( method forward_depformer (line 229) | def forward_depformer( method device (line 270) | def device(self) -> torch.device: method get_initial_token (line 274) | def get_initial_token(self) -> torch.Tensor: class MoshiVisGen (line 298) | class MoshiVisGen(StreamingModule): method __init__ (line 301) | def __init__( method update_gen_kwargs (line 329) | def update_gen_kwargs( method model_dim (line 343) | def model_dim(self) -> int: method num_audio_codebooks_out (line 348) | def num_audio_codebooks_out(self) -> int: method from_config (line 353) | def from_config( method precompte_ca_kv (line 380) | def precompte_ca_kv( method step (line 399) | def step( method depformer_step (line 513) | def depformer_step( FILE: kyuteye_pt/kyuteye/modules/attention.py class KVCache (line 18) | class KVCache: method __init__ (line 35) | def __init__( method clone (line 71) | def clone(self) -> "KVCache": method current_start (line 87) | def current_start(self) -> int: method __maybe_increase_capacity__ (line 91) | def __maybe_increase_capacity__(self, required_capacity: int) -> None: method complete (line 120) | def complete( class MultiheadAttention (line 140) | class MultiheadAttention(StreamingModule): method __init__ (line 161) | def __init__( method _complete_kv (line 219) | def _complete_kv( method forward (line 244) | def forward( FILE: kyuteye_pt/kyuteye/modules/cross_attention.py class SharedModuleType (line 14) | class SharedModuleType(type): method __call__ (line 19) | def __call__(cls, *args: Any, **kwargs: Any) -> Any: class XAGate (line 25) | class XAGate(torch.nn.Module): method __init__ (line 28) | def __init__( method forward (line 61) | def forward(self, x: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: class SharedXaGate (line 70) | class SharedXaGate(XAGate, metaclass=SharedModuleType): class CrossAttention (line 76) | class CrossAttention(MultiheadAttention): method __init__ (line 79) | def __init__(self, *args: Any, **kwargs: Any) -> None: class SharedCrossAttention (line 84) | class SharedCrossAttention(CrossAttention, metaclass=SharedModuleType): class GatedCrossAttention (line 90) | class GatedCrossAttention(StreamingModule): method __init__ (line 93) | def __init__( method get_xa_scope (line 166) | def get_xa_scope( method is_active (line 248) | def is_active(self, image_tokens_mask: Optional[torch.Tensor] = None) ... method forward (line 289) | def forward( FILE: kyuteye_pt/kyuteye/modules/image_encoder.py class TrimmedFlexiViTWrapper (line 26) | class TrimmedFlexiViTWrapper(torch.nn.Module): method __init__ (line 29) | def __init__( method forward (line 36) | def forward(self, x: torch.Tensor) -> torch.Tensor: function load_paligemma_vision_encoder (line 43) | def load_paligemma_vision_encoder( class PixtralOutput (line 75) | class PixtralOutput: class PixtralWrapper (line 82) | class PixtralWrapper(torch.nn.Module): method __init__ (line 85) | def __init__( method __get_num_output_tokens__ (line 100) | def __get_num_output_tokens__(self, x: List[torch.Tensor]) -> List[int]: method split_and_pad_output (line 109) | def split_and_pad_output( method forward (line 131) | def forward(self, x: List[torch.Tensor] | torch.Tensor) -> PixtralOutput: function get_img_normalize (line 145) | def get_img_normalize( function load_image_encoder (line 160) | def load_image_encoder( FILE: kyuteye_pt/kyuteye/modules/image_transforms.py function get_minimal_transforms (line 21) | def get_minimal_transforms( class Normalize (line 55) | class Normalize: method __init__ (line 59) | def __init__(self, mean: Sequence[float], std: Sequence[float]) -> None: method __call__ (line 73) | def __call__( method to_pil_transform (line 80) | def to_pil_transform(self, mode: str = "RGB") -> T.Transform: class UnitNormalize (line 91) | class UnitNormalize(Normalize): method __init__ (line 94) | def __init__(self) -> None: class CLIPNormalize (line 101) | class CLIPNormalize(Normalize): method __init__ (line 104) | def __init__(self) -> None: class SigLIPNormalize (line 111) | class SigLIPNormalize(Normalize): method __init__ (line 114) | def __init__(self) -> None: class PixtralNormalize (line 118) | class PixtralNormalize: method __init__ (line 123) | def __init__(self) -> None: method __call__ (line 128) | def __call__( FILE: kyuteye_pt/kyuteye/modules/streaming_utils.py class StreamingModule (line 16) | class StreamingModule(torch.nn.Module): method __init__ (line 19) | def __init__(self) -> None: method empty_streaming_state (line 25) | def empty_streaming_state(self) -> bool: method has_streaming_attribute (line 29) | def has_streaming_attribute(self, key: str) -> bool: method add_streaming_attribute (line 33) | def add_streaming_attribute( method get_streaming_attribute (line 39) | def get_streaming_attribute(self, key: str, default: Any = None) -> Any: method is_streaming (line 44) | def is_streaming(self) -> bool: method get_streaming_info_as_int (line 48) | def get_streaming_info_as_int(self, attr_name: str, default: int = 0) ... method streaming_offset (line 61) | def streaming_offset(self) -> int: method streaming_offset (line 66) | def streaming_offset(self, value: int | torch.Tensor) -> None: method _apply_named_streaming (line 73) | def _apply_named_streaming(self, fn: Callable) -> None: method _set_streaming (line 78) | def _set_streaming(self, streaming: bool) -> None: method streaming (line 85) | def streaming(self) -> Iterator: method streaming_forever (line 94) | def streaming_forever(self, batch_size: Optional[int] = None) -> None: method reset_streaming (line 99) | def reset_streaming(self) -> None: method get_streaming_state (line 107) | def get_streaming_state(self) -> State: method set_streaming_state (line 120) | def set_streaming_state(self, state: State) -> None: method flush (line 139) | def flush(self, x: Optional[torch.Tensor] = None) -> Optional["Streami... FILE: kyuteye_pt/kyuteye/modules/transformer.py class TransformerLayer (line 22) | class TransformerLayer(StreamingModule): method __init__ (line 47) | def __init__( method _ff_block (line 147) | def _ff_block(self, x: torch.Tensor) -> torch.Tensor: method _maybe_cross_attend (line 165) | def _maybe_cross_attend( method _self_attend (line 184) | def _self_attend( method forward (line 197) | def forward( class Transformer (line 241) | class Transformer(StreamingModule): method __init__ (line 266) | def __init__( method set_context (line 314) | def set_context(self, context: Optional[int] = None) -> None: method forward (line 320) | def forward( FILE: kyuteye_pt/kyuteye/modules/utils.py function multi_linear (line 18) | def multi_linear( function get_activation (line 43) | def get_activation( function gating_forward_kernel (line 68) | def gating_forward_kernel( class ActivationGating (line 83) | class ActivationGating(torch.nn.Module): method __init__ (line 92) | def __init__( method forward (line 115) | def forward(self, x: torch.Tensor) -> torch.Tensor: class NoGating (line 122) | class NoGating(torch.nn.Module): method __init__ (line 127) | def __init__( method forward (line 144) | def forward(self, x: torch.Tensor) -> torch.Tensor: function make_ffn (line 149) | def make_ffn( class LayerNormF32 (line 192) | class LayerNormF32(torch.nn.LayerNorm): method forward (line 195) | def forward( function _rms_norm (line 204) | def _rms_norm( class RMSNorm (line 226) | class RMSNorm(torch.nn.Module): method __init__ (line 233) | def __init__( method forward (line 249) | def forward(self, x: torch.Tensor) -> torch.Tensor: class NormalizationLayer (line 256) | class NormalizationLayer(Enum): method create_norm_fn (line 266) | def create_norm_fn(self, dim: int, **kwargs: Any) -> torch.nn.Module: class ClampedEmbedding (line 299) | class ClampedEmbedding(torch.nn.Embedding): method __init__ (line 309) | def __init__( method forward (line 319) | def forward( # pylint: disable=arguments-renamed function create_sin_embedding (line 332) | def create_sin_embedding( function apply_rope (line 360) | def apply_rope( class RotaryEmbedding (line 411) | class RotaryEmbedding(torch.nn.Module): method __init__ (line 417) | def __init__(self, max_period: float = 10000.0) -> None: method forward (line 421) | def forward( FILE: kyuteye_pt/kyuteye/server.py function colorize (line 37) | def colorize(text: str, color: str) -> str: function make_log (line 44) | def make_log(level: str, msg: str) -> str: function log (line 57) | def log(level: str, msg: str) -> None: function seed_all (line 62) | def seed_all(seed: int) -> None: class ServerState (line 75) | class ServerState: method __init__ (line 88) | def __init__( method warmup (line 117) | def warmup(self) -> None: method handle_chat (line 136) | async def handle_chat(self, request: Any) -> Any: method extract_image (line 287) | async def extract_image(self, ws: web.WebSocketResponse) -> None: function start_server (line 314) | def start_server( function sanity_check (line 435) | def sanity_check() -> None: function main (line 439) | def main() -> None: FILE: kyuteye_pt/kyuteye/utils/dist_utils.py function is_main (line 10) | def is_main() -> bool: function print_main (line 21) | def print_main(*args: Any, rich: bool = False, **kwargs: Any) -> None: FILE: kyuteye_pt/kyuteye/utils/logging_utils.py function flatten_nested_dict (line 10) | def flatten_nested_dict(d: Dict) -> Dict: function get_git_revision_hash (line 21) | def get_git_revision_hash(verbose: bool = True) -> Tuple[str, str]: function pretty_json (line 38) | def pretty_json(config_dict: dict) -> str: FILE: kyuteye_pt/kyuteye/utils/struct_utils.py class FrozenEnumMeta (line 9) | class FrozenEnumMeta(EnumMeta): method __new__ (line 12) | def __new__(mcs, name: str, bases: Any, classdict: Any) -> type: method __setattr__ (line 18) | def __setattr__(cls, name: str, value: Any) -> None: method __delattr__ (line 28) | def __delattr__(cls, name: str) -> None: class FrozenEnum (line 39) | class FrozenEnum(Enum, metaclass=FrozenEnumMeta): FILE: kyuteye_pt/tests/hello.py function write_weights_for_analysis (line 11) | def write_weights_for_analysis(model: torch.nn.Module): function test_weights_conversion_moshi (line 29) | def test_weights_conversion_moshi(): FILE: kyuteye_rs/moshi-backend/build.rs function main (line 4) | pub fn main() -> Result<()> { FILE: kyuteye_rs/moshi-backend/src/audio.rs type Sample (line 8) | pub trait Sample { method to_i16 (line 9) | fn to_i16(&self) -> i16; method to_i16 (line 13) | fn to_i16(&self) -> i16 { method to_i16 (line 19) | fn to_i16(&self) -> i16 { method to_i16 (line 25) | fn to_i16(&self) -> i16 { function write_pcm_as_wav (line 30) | pub fn write_pcm_as_wav( function conv (line 63) | fn conv(samples: &mut Vec, data: std::borrow::Cow>(path: P) -> anyhow::... function resample (line 117) | pub(crate) fn resample(pcm_in: &[f32], sr_in: usize, sr_out: usize) -> a... function write_opus_header (line 145) | pub(crate) fn write_opus_header(w: &mut W) -> std::io... function write_opus_tags (line 159) | pub(crate) fn write_opus_tags(w: &mut W) -> std::io::... FILE: kyuteye_rs/moshi-backend/src/build.rs function main (line 8) | pub fn main() -> Result<()> { FILE: kyuteye_rs/moshi-backend/src/image_embedder.rs function load_image (line 10) | fn load_image( type ImageEncoder (line 78) | pub enum ImageEncoder { type ImageEncoderModel (line 88) | pub enum ImageEncoderModel { function init_output_proj (line 94) | fn init_output_proj(in_dims: usize, out_dims: usize, vb: VarBuilder) -> ... type ImageEmbedder (line 104) | pub struct ImageEmbedder { method new (line 124) | pub fn new( method output_proj (line 238) | pub fn output_proj(&self, img_features: Tensor, dev: &Device) -> Resul... method embed (line 255) | pub fn embed( method embed_from_tensor (line 328) | pub fn embed_from_tensor(&self, img: Tensor, dev: &Device) -> Result Self::F... function tracing_init (line 86) | fn tracing_init( function main (line 115) | async fn main() -> Result<()> { FILE: kyuteye_rs/moshi-backend/src/standalone.rs type Config (line 14) | pub struct Config { method load (line 27) | pub fn load>(p: P) -> Result { method cert_file (line 44) | pub fn cert_file(&self, name: &str) -> Result { function device (line 54) | pub(crate) fn device(cpu: bool) -> Result { function new (line 68) | pub fn new(args: &StandaloneArgs, config: &stream_both::Config) -> Resul... function handle_socket (line 179) | async fn handle_socket(socket: ws::WebSocket, sm: stream_both::Streaming... function stream_handler (line 185) | pub async fn stream_handler( function download_from_hub (line 196) | pub async fn download_from_hub(config: &mut stream_both::Config) -> Resu... function run (line 243) | pub async fn run(args: &StandaloneArgs, config: &Config) -> Result<()> { FILE: kyuteye_rs/moshi-backend/src/stream_both.rs type ForceSessionConfig (line 16) | pub struct ForceSessionConfig { type Config (line 28) | pub struct Config { method requires_model_download (line 57) | pub fn requires_model_download(&self) -> bool { function default_false (line 51) | fn default_false() -> bool { type AppState (line 68) | pub type AppState = Arc; type AppStateInner (line 69) | pub struct AppStateInner { method text (line 79) | fn text( type SessionConfigReq (line 117) | pub struct SessionConfigReq { method into_session_config (line 168) | fn into_session_config(self, force_cfg: Option<&ForceSessionConfig>) -... type SessionConfig (line 136) | pub struct SessionConfig { type SessionSummary (line 155) | struct SessionSummary<'a> { type MetaData (line 224) | pub struct MetaData { type StreamOut (line 242) | pub enum StreamOut { constant OPUS_ENCODER_FRAME_SIZE (line 252) | const OPUS_ENCODER_FRAME_SIZE: usize = 960; type MsgType (line 255) | pub enum MsgType { method from_u8 (line 269) | pub fn from_u8(v: u8) -> Result { method to_u8 (line 286) | pub fn to_u8(self) -> u8 { type ModelInput (line 302) | pub enum ModelInput { type MsgSender (line 308) | pub struct MsgSender { method new (line 319) | fn new(sender: SplitSink) -> Result { method send_colored_text (line 344) | async fn send_colored_text(&mut self, text: String, intensity: f32) ->... method send_ready (line 357) | async fn send_ready(&mut self) -> Result<()> { method send_metadata (line 367) | async fn send_metadata(&mut self, md: Box) -> Result<()> { method send_pcm (line 375) | async fn send_pcm(&mut self, pcm: Vec) -> Result<()> { type StreamingModel (line 415) | pub struct StreamingModel { method run_with_state (line 423) | fn run_with_state( method new (line 545) | pub fn new(state: &AppState, session_config: SessionConfigReq) -> Self { method run (line 560) | pub fn run( type Handle (line 697) | type Handle = tokio::task::JoinHandle>; function spawn_recv_loops (line 699) | fn spawn_recv_loops( function sender_loop (line 788) | async fn sender_loop( function handle_socket (line 814) | pub async fn handle_socket( FILE: kyuteye_rs/moshi-backend/src/utils.rs type BuildInfo (line 2) | pub struct BuildInfo { method new (line 16) | pub fn new() -> BuildInfo { type WrapJson (line 32) | pub struct WrapJson(pub anyhow::Result); function into_response (line 35) | fn into_response(self) -> axum::response::Response { function replace_env_vars (line 50) | pub fn replace_env_vars(input: &str) -> String { type WrapBincode (line 59) | pub struct WrapBincode(pub anyhow::Result); function into_response (line 62) | fn into_response(self) -> axum::response::Response { function default_static_dir (line 77) | pub fn default_static_dir() -> String { type AxumError (line 81) | pub struct AxumError(anyhow::Error); method into_response (line 84) | fn into_response(self) -> axum::response::Response { method from (line 96) | fn from(value: E) -> Self { type AxumResult (line 101) | pub type AxumResult = std::result::Result; FILE: kyuteye_rs/moshi-core/src/conv.rs type Norm (line 11) | pub enum Norm { type PadMode (line 18) | pub enum PadMode { function conv1d_weight_norm (line 27) | fn conv1d_weight_norm( type NormConv1d (line 52) | pub struct NormConv1d { method new (line 60) | pub fn new( method forward (line 102) | fn forward(&self, xs: &Tensor) -> Result { type NormConvTranspose1d (line 113) | pub struct NormConvTranspose1d { method new (line 125) | pub fn new( method forward (line 188) | fn forward(&self, xs: &Tensor) -> Result { function get_extra_padding_for_conv1d (line 211) | fn get_extra_padding_for_conv1d( function pad1d (line 224) | fn pad1d(xs: &Tensor, pad_l: usize, pad_r: usize, mode: PadMode) -> Resu... function unpad1d (line 232) | fn unpad1d(xs: &Tensor, unpad_l: usize, unpad_r: usize) -> Result { type StreamableConv1d (line 241) | pub struct StreamableConv1d { method new (line 253) | pub fn new( method forward (line 289) | fn forward(&self, xs: &Tensor) -> Result { method reset_state (line 316) | fn reset_state(&mut self) { method step (line 321) | fn step(&mut self, xs: &StreamTensor) -> Result { type StreamableConvTranspose1d (line 360) | pub struct StreamableConvTranspose1d { method new (line 370) | pub fn new( method forward (line 403) | fn forward(&self, xs: &Tensor) -> Result { method reset_state (line 421) | fn reset_state(&mut self) { method step (line 425) | fn step(&mut self, xs: &StreamTensor) -> Result { type ConvDownsample1d (line 461) | pub struct ConvDownsample1d { method new (line 466) | pub fn new( method forward (line 494) | fn forward(&self, xs: &Tensor) -> Result { method reset_state (line 500) | fn reset_state(&mut self) { method step (line 504) | fn step(&mut self, xs: &StreamTensor) -> Result { type ConvTrUpsample1d (line 510) | pub struct ConvTrUpsample1d { method new (line 515) | pub fn new( method forward (line 541) | fn forward(&self, xs: &Tensor) -> Result { method reset_state (line 547) | fn reset_state(&mut self) { method step (line 551) | fn step(&mut self, xs: &StreamTensor) -> Result { function run_conv1d (line 561) | fn run_conv1d( function run_conv_tr1d (line 612) | fn run_conv_tr1d( function conv1d (line 655) | fn conv1d() -> Result<()> { function conv_tr1d (line 669) | fn conv_tr1d() -> Result<()> { FILE: kyuteye_rs/moshi-core/src/dynamic_logits_processor.rs type GateInfluencedLogitsProcessor (line 5) | pub struct GateInfluencedLogitsProcessor { method from_sampling (line 12) | pub fn from_sampling(seed: u64, sampling: Sampling) -> Self { method from_sampling_with_scale (line 21) | pub fn from_sampling_with_scale( method new (line 34) | pub fn new(seed: u64, temperature: Option, top_p: Option) ->... method sample_argmax (line 46) | fn sample_argmax(&mut self, logits: Tensor) -> Result { method sample_multinomial (line 57) | fn sample_multinomial(&mut self, prs: &Vec) -> Result { method sample_topp (line 66) | fn sample_topp(&mut self, prs: &mut Vec, top_p: f32) -> Result, top_k: usize) -> Result<... method sample_topk_topp (line 101) | fn sample_topk_topp(&mut self, prs: &mut Vec, top_k: usize, top_p... method sample (line 119) | pub fn sample(&mut self, logits: &Tensor, gate_weight: f64) -> Result<... method sample_f (line 123) | pub fn sample_f( FILE: kyuteye_rs/moshi-core/src/lib.rs type NormType (line 21) | pub enum NormType { FILE: kyuteye_rs/moshi-core/src/lm.rs type DepFormerConfig (line 22) | pub struct DepFormerConfig { type Config (line 28) | pub struct Config { method v0_1 (line 42) | pub fn v0_1() -> Self { method v0_1_vision (line 103) | pub fn v0_1_vision() -> Self { method v0_1_vision_streaming (line 168) | pub fn v0_1_vision_streaming(num_slices: usize) -> Self { method v0_1_streaming (line 178) | pub fn v0_1_streaming(num_slices: usize) -> Self { type DepFormerSlice (line 190) | struct DepFormerSlice { method new (line 200) | fn new( type DepFormer (line 222) | pub struct DepFormer { method new (line 227) | pub fn new( method sample (line 256) | pub fn sample( method sample_cfg (line 308) | pub fn sample_cfg( type LmModel (line 359) | pub struct LmModel { method new (line 372) | pub fn new(cfg: &Config, vb: MaybeQuantizedVarBuilder) -> Result { method reset_state (line 421) | pub fn reset_state(&mut self) { method in_audio_codebooks (line 426) | pub fn in_audio_codebooks(&self) -> usize { method audio_pad_token (line 430) | pub fn audio_pad_token(&self) -> u32 { method text_start_token (line 434) | pub fn text_start_token(&self) -> u32 { method generated_audio_codebooks (line 438) | pub fn generated_audio_codebooks(&self) -> usize { method is_quantized (line 442) | pub fn is_quantized(&self) -> bool { method device (line 449) | pub fn device(&self) -> &Device { method forward (line 453) | pub fn forward( method maybe_precompute_ca_kv (line 500) | pub fn maybe_precompute_ca_kv(&self, ca_src: Option) -> Result<... method forward_ca (line 508) | pub fn forward_ca( method forward_with_gate_weight (line 518) | pub fn forward_with_gate_weight( method depformer_sample (line 572) | pub fn depformer_sample( function load_lm_model (line 590) | pub fn load_lm_model>( function load (line 614) | pub fn load>( function load_streaming (line 623) | pub fn load_streaming>( function load_streaming_both_ways (line 632) | pub fn load_streaming_both_ways>( function load_vision (line 641) | pub fn load_vision>( type ForcedAudioTokens (line 655) | pub struct ForcedAudioTokens { method new (line 662) | pub fn new(acoustic_delay: usize, audio_pad_token: u32, stream_codeboo... method forced_tokens (line 676) | pub fn forced_tokens(&self, step_idx: usize) -> &[Option] { FILE: kyuteye_rs/moshi-core/src/lm_generate.rs constant UNGENERATED (line 8) | const UNGENERATED: u32 = u32::MAX; type Config (line 11) | pub struct Config { method v0_1 (line 22) | pub fn v0_1() -> Self { method audio_pad_token (line 34) | pub fn audio_pad_token(&self) -> u32 { method audio_codebooks (line 38) | pub fn audio_codebooks(&self) -> usize { type State (line 43) | pub struct State { method new (line 55) | pub fn new( method audio_codebooks (line 81) | pub fn audio_codebooks(&self) -> usize { method audio_pad_token (line 85) | pub fn audio_pad_token(&self) -> u32 { method step_gen_no_text (line 89) | pub fn step_gen_no_text(&mut self, force_text_token: Option) -> c... method step_gen (line 93) | pub fn step_gen(&mut self, prev_text_token: u32) -> candle::Result { method step_text_prompt (line 97) | pub fn step_text_prompt(&mut self, id: u32) -> candle::Result { method step_audio_prompt_ (line 101) | pub fn step_audio_prompt_( method step_audio_prompt (line 122) | pub fn step_audio_prompt(&mut self, codes: &[u32]) -> candle::Result Option> { method audio_tokens (line 147) | pub fn audio_tokens(&self) -> Vec> { method step (line 154) | fn step( FILE: kyuteye_rs/moshi-core/src/lm_generate_multistream.rs constant UNGENERATED (line 11) | pub const UNGENERATED: u32 = u32::MAX; type Config (line 14) | pub struct Config { method v0_1 (line 25) | pub fn v0_1() -> Self { method v0_1_two_ways (line 37) | pub fn v0_1_two_ways() -> Self { method v0_1_one_way (line 49) | pub fn v0_1_one_way() -> Self { method audio_pad_token (line 61) | pub fn audio_pad_token(&self) -> u32 { method total_audio_codebooks (line 65) | pub fn total_audio_codebooks(&self) -> usize { type State (line 70) | pub struct State { method new (line 88) | pub fn new( method step_idx (line 124) | pub fn step_idx(&self) -> usize { method audio_pad_token (line 128) | fn audio_pad_token(&self) -> u32 { method config (line 132) | pub fn config(&self) -> &Config { method user_rating (line 136) | pub fn user_rating(&self) -> u32 { method set_user_rating (line 139) | pub fn set_user_rating(&mut self, grade: u32) { method apply_repetition_penalty (line 143) | fn apply_repetition_penalty(&self, logits: Tensor) -> candle::Result &[Vec] { method gate_weights (line 325) | pub fn gate_weights(&self, include_all: bool) -> &[f32] { method text_tokens (line 334) | pub fn text_tokens(&self, include_all: bool) -> &[u32] { method last_audio_tokens (line 343) | pub fn last_audio_tokens(&self) -> Option> { FILE: kyuteye_rs/moshi-core/src/mimi.rs type ResampleMethod (line 11) | pub enum ResampleMethod { type Config (line 17) | pub struct Config { method v0_1 (line 32) | pub fn v0_1(num_codebooks: Option) -> Self { type Mimi (line 93) | pub struct Mimi { method new (line 105) | pub fn new(cfg: Config, vb: VarBuilder) -> Result { method config (line 161) | pub fn config(&self) -> &Config { method encode_pre_quantize (line 165) | pub fn encode_pre_quantize(&mut self, xs: &Tensor) -> Result { method encode (line 173) | pub fn encode(&mut self, xs: &Tensor) -> Result { method encode_step (line 183) | pub fn encode_step(&mut self, xs: &StreamTensor) -> Result Result { method decode_step (line 205) | pub fn decode_step(&mut self, codes: &StreamTensor) -> Result, dev: &Device... FILE: kyuteye_rs/moshi-core/src/nn.rs type MaybeQuantizedWeight (line 9) | pub enum MaybeQuantizedWeight { method to_tensor (line 16) | fn to_tensor(&self, dev: &Device) -> Result { function matmul_dtype (line 24) | pub fn matmul_dtype(device: &candle::Device) -> DType { type MaybeQuantizedVarBuilder (line 34) | pub enum MaybeQuantizedVarBuilder<'a> { function pp (line 41) | pub fn pp(&self, s: S) -> Self { function get (line 48) | pub fn get>(&self, s: S, path: &str) -> Result>(&self, s: S, path: &str) -> Result<... function get_unquantized (line 64) | pub fn get_unquantized>(&self, s: S, path: &str) -> Resul... function contains_key (line 71) | pub fn contains_key(&self, name: &str) -> bool { function device (line 78) | pub fn device(&self) -> &Device { function dtype (line 85) | pub fn dtype(&self) -> DType { type MaybeQuantizedLinear (line 94) | pub enum MaybeQuantizedLinear { method forward (line 100) | fn forward(&self, xs: &Tensor) -> Result { type MaybeQuantizedEmbedding (line 109) | pub enum MaybeQuantizedEmbedding { method new (line 115) | pub fn new(in_vocab_size: usize, dim: usize, vb: MaybeQuantizedVarBuil... method embeddings (line 127) | pub fn embeddings(&self) -> &Tensor { method hidden_size (line 134) | pub fn hidden_size(&self) -> Result { method forward (line 144) | fn forward(&self, xs: &Tensor) -> Result { function linear (line 152) | pub fn linear( function linear_from (line 173) | pub fn linear_from( FILE: kyuteye_rs/moshi-core/src/quantization.rs type CodebookEncode (line 8) | struct CodebookEncode; method name (line 11) | fn name(&self) -> &'static str { method cpu_fwd (line 15) | fn cpu_fwd( type EuclideanCodebook (line 73) | pub struct EuclideanCodebook { method new (line 86) | pub fn new(dim: usize, codebook_size: usize, vb: VarBuilder) -> Result... method encode_very_slow (line 109) | pub fn encode_very_slow(&self, xs: &Tensor) -> Result { method encode_slow (line 125) | pub fn encode_slow(&self, xs: &Tensor) -> Result { method encode (line 136) | pub fn encode(&self, xs: &Tensor) -> Result { method decode (line 146) | pub fn decode(&self, indexes: &Tensor) -> Result { type VectorQuantization (line 160) | pub struct VectorQuantization { method new (line 167) | pub fn new( method encode (line 189) | pub fn encode(&self, xs: &Tensor) -> Result { method decode (line 194) | pub fn decode(&self, codes: &Tensor) -> Result { type ResidualVectorQuantization (line 205) | pub struct ResidualVectorQuantization { method new (line 210) | pub fn new( method encode (line 226) | pub fn encode(&self, xs: &Tensor) -> Result { method decode (line 238) | pub fn decode(&self, xs: &Tensor) -> Result { type ResidualVectorQuantizer (line 260) | pub struct ResidualVectorQuantizer { method new (line 267) | pub fn new( method encode (line 318) | pub fn encode(&self, xs: &Tensor) -> Result { method decode (line 323) | pub fn decode(&self, codes: &Tensor) -> Result { type SplitResidualVectorQuantizer (line 337) | pub struct SplitResidualVectorQuantizer { method new (line 346) | pub fn new( method encode (line 383) | pub fn encode(&self, xs: &Tensor) -> Result { method decode (line 397) | pub fn decode(&self, codes: &Tensor) -> Result { FILE: kyuteye_rs/moshi-core/src/seanet.rs type Config (line 12) | pub struct Config { type SeaNetResnetBlock (line 34) | pub struct SeaNetResnetBlock { method new (line 44) | pub fn new( method forward (line 109) | fn forward(&self, xs: &Tensor) -> Result { method reset_state (line 123) | fn reset_state(&mut self) { method step (line 133) | fn step(&mut self, xs: &StreamTensor) -> Result { type EncoderLayer (line 147) | struct EncoderLayer { type SeaNetEncoder (line 153) | pub struct SeaNetEncoder { method new (line 162) | pub fn new(cfg: &Config, vb: VarBuilder) -> Result { method forward (line 267) | fn forward(&self, xs: &Tensor) -> Result { method reset_state (line 281) | fn reset_state(&mut self) { method step (line 290) | fn step(&mut self, xs: &StreamTensor) -> Result { type DecoderLayer (line 304) | struct DecoderLayer { type SeaNetDecoder (line 310) | pub struct SeaNetDecoder { method new (line 320) | pub fn new(cfg: &Config, vb: VarBuilder) -> Result { method forward (line 423) | fn forward(&self, xs: &Tensor) -> Result { method reset_state (line 442) | fn reset_state(&mut self) { method step (line 451) | fn step(&mut self, xs: &StreamTensor) -> Result { FILE: kyuteye_rs/moshi-core/src/streaming.rs type Dim (line 7) | pub trait Dim: candle::shape::Dim + Copy {} type StreamTensor (line 11) | pub struct StreamTensor(Option); method fmt (line 14) | fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { method from (line 23) | fn from(value: Option) -> Self { method from (line 29) | fn from(value: Tensor) -> Self { method from (line 35) | fn from(_value: ()) -> Self { method empty (line 41) | pub fn empty() -> Self { method from_tensor (line 45) | pub fn from_tensor(tensor: Tensor) -> Self { method shape (line 49) | pub fn shape(&self) -> Option<&candle::Shape> { method cat2 (line 53) | pub fn cat2(&self, rhs: &Self, dim: D) -> Result { method seq_len (line 65) | pub fn seq_len(&self, dim: D) -> Result { method reset (line 72) | pub fn reset(&mut self) { method narrow (line 76) | pub fn narrow(&self, dim: D, offset: usize, len: usize) -> Res... method split (line 94) | pub fn split(&self, dim: D, lhs_len: usize) -> Result<(Self, S... method as_option (line 116) | pub fn as_option(&self) -> Option<&Tensor> { method apply (line 120) | pub fn apply(&self, m: &M) -> Result { type StreamingModule (line 128) | pub trait StreamingModule { method step (line 130) | fn step(&mut self, xs: &StreamTensor) -> Result; method reset_state (line 131) | fn reset_state(&mut self); method reset_state (line 200) | fn reset_state(&mut self) {} method step (line 202) | fn step(&mut self, xs: &StreamTensor) -> Result { type BinOp (line 135) | pub enum BinOp { type StreamingBinOp (line 143) | pub struct StreamingBinOp { method new (line 151) | pub fn new(op: BinOp, dim: candle::D) -> Self { method reset_state (line 160) | pub fn reset_state(&mut self) { method forward (line 165) | pub fn forward(&self, lhs: &Tensor, rhs: &Tensor) -> Result { method step (line 174) | pub fn step(&mut self, lhs: &StreamTensor, rhs: &StreamTensor) -> Resu... type Map (line 197) | pub struct Map(T); FILE: kyuteye_rs/moshi-core/src/transformer.rs type Config (line 17) | pub struct Config { type PositionalEmbedding (line 43) | pub enum PositionalEmbedding { type CrossAttentionGating (line 50) | pub enum CrossAttentionGating { type CaSrc (line 62) | pub enum CaSrc { type LayerScale (line 72) | pub struct LayerScale { method new (line 77) | pub fn new(d_model: usize, _init: f64, vb: MaybeQuantizedVarBuilder) -... method forward (line 84) | fn forward(&self, xs: &Tensor) -> Result { type XaGate (line 90) | pub enum XaGate { method new (line 110) | pub fn new(cfg: &Config, vb: MaybeQuantizedVarBuilder) -> Result { method forward_with_gate_weight (line 166) | pub fn forward_with_gate_weight(&self, xs: &Tensor) -> Result<(Tensor,... method forward (line 195) | fn forward(&self, xs: &Tensor) -> Result { type StreamingMultiheadCrossAttention (line 202) | pub struct StreamingMultiheadCrossAttention { method new (line 217) | pub fn new( method is_quantized (line 301) | pub fn is_quantized(&self) -> bool { method compute_kv (line 308) | pub fn compute_kv(&self, ca_src: &CaSrc) -> Result<(Tensor, Tensor)> { method forward_with_gate_weight (line 335) | pub fn forward_with_gate_weight( method forward (line 383) | pub fn forward(&self, xs: &Tensor, ca_src: &CaSrc, mask: Option<&Tenso... type RotaryEmbedding (line 390) | pub struct RotaryEmbedding { method new (line 397) | pub fn new(dim: usize, max_seq_len: usize, theta: f32, dev: &Device) -... method apply_rotary_emb (line 415) | pub fn apply_rotary_emb(&self, qk: &Tensor, seqlen_offset: usize) -> R... function get_causal_mask (line 425) | pub(crate) fn get_causal_mask( function flash_attn (line 441) | fn flash_attn( function flash_attn (line 452) | fn flash_attn(_: &Tensor, _: &Tensor, _: &Tensor, _: f32, _: bool) -> Re... type StreamingMultiheadAttention (line 457) | pub struct StreamingMultiheadAttention { method new (line 474) | pub fn new( method is_quantized (line 511) | pub fn is_quantized(&self) -> bool { method forward (line 518) | pub fn forward(&mut self, xs: &Tensor, mask: Option<&Tensor>) -> Resul... method reset_kv_cache (line 598) | pub fn reset_kv_cache(&mut self) { method set_kv_cache (line 602) | pub fn set_kv_cache(&mut self, kv_cache: candle_nn::kv_cache::KvCache) { type Mlp (line 608) | pub enum Mlp { method new (line 622) | pub fn new(cfg: &Config, vb: MaybeQuantizedVarBuilder) -> Result { method forward (line 650) | fn forward(&self, xs: &Tensor) -> Result { type RmsNorm (line 669) | pub struct RmsNorm { method new (line 675) | pub fn new(d_model: usize, eps: f32, vb: MaybeQuantizedVarBuilder) -> ... method forward (line 684) | fn forward(&self, xs: &Tensor) -> Result { type LayerNorm (line 690) | pub struct LayerNorm { method new (line 695) | pub fn new(d_model: usize, eps: f32, vb: MaybeQuantizedVarBuilder) -> ... method forward (line 709) | fn forward(&self, xs: &Tensor) -> Result { type Norm (line 715) | pub enum Norm { method new (line 721) | pub fn new(d_model: usize, cfg: &Config, vb: MaybeQuantizedVarBuilder)... method new_shortcut (line 726) | pub fn new_shortcut( method forward (line 746) | fn forward(&self, xs: &Tensor) -> Result { type StreamingTransformerLayer (line 755) | pub struct StreamingTransformerLayer { method new (line 768) | pub fn new( method forward_with_gate_weight (line 826) | pub fn forward_with_gate_weight( method forward (line 860) | pub fn forward( method reset_kv_cache (line 870) | pub fn reset_kv_cache(&mut self) { method set_kv_cache (line 874) | pub fn set_kv_cache(&mut self, kv_cache: candle_nn::kv_cache::KvCache) { type StreamingTransformer (line 880) | pub struct StreamingTransformer { method new (line 890) | pub fn new(cfg: &Config, vb: MaybeQuantizedVarBuilder) -> Result { method forward (line 921) | pub fn forward(&mut self, xs: &Tensor) -> Result { method forward_ca (line 925) | pub fn forward_ca(&mut self, xs: &Tensor, ca_src: Option<&CaSrc>) -> R... method forward_with_gate_weight (line 930) | pub fn forward_with_gate_weight( method maybe_precompute_ca_kv (line 986) | pub fn maybe_precompute_ca_kv(&self, ca_src: Option) -> Result<... method copy_state (line 1007) | pub fn copy_state(&mut self, from: &Self) -> Result<()> { method reset_state (line 1020) | fn reset_state(&mut self) { method step (line 1024) | fn step(&mut self, xs: &StreamTensor) -> Result { type ProjectedTransformer (line 1033) | pub struct ProjectedTransformer { method new (line 1043) | pub fn new( method forward (line 1076) | pub fn forward(&mut self, xs: &Tensor) -> Result> { method reset_state (line 1100) | fn reset_state(&mut self) { method step (line 1104) | fn step(&mut self, xs: &StreamTensor) -> Result { FILE: scripts/convert_ckpt_utils.py function remove_other_output_codebooks (line 31) | def remove_other_output_codebooks( class Launcher (line 52) | class Launcher: method rust_to_pt (line 54) | def rust_to_pt(self, safetensors_file: str, out_file: Optional[str] = ... method pt_to_mlx (line 126) | def pt_to_mlx(self, safetensors_file: str, out_file: Optional[str] = N... FILE: scripts/get_static_client.py function get (line 18) | def get() -> None: FILE: ssvd/generate.py function get_pipeline (line 41) | def get_pipeline( function get_captions (line 60) | def get_captions( class Launcher (line 94) | class Launcher: method __get_db_file__ (line 98) | def __get_db_file__( method __get_table_name__ (line 105) | def __get_table_name__( method __get_annot_file__ (line 112) | def __get_annot_file__( method watch (line 125) | def watch( method run (line 161) | def run( FILE: ssvd/multiturn_instruct.py function get_base_setting (line 9) | def get_base_setting() -> Tuple[str, str, str, str]: function get_location_setting (line 55) | def get_location_setting() -> Tuple[str, str, str, str]: function get_num_setting (line 70) | def get_num_setting() -> Tuple[str, str, str, str]: function get_property_setting (line 85) | def get_property_setting() -> Tuple[str, str, str, str]: function get_lead_short_setting (line 103) | def get_lead_short_setting() -> Tuple[str, str, str, str]: function get_lead_long_setting (line 122) | def get_lead_long_setting() -> Tuple[str, str, str, str]: function get_comb_start_setting (line 165) | def get_comb_start_setting() -> Tuple[str, str, str, str]: function get_tns_setting (line 251) | def get_tns_setting() -> Tuple[str, str, str, str]: function get_tbs_setting (line 292) | def get_tbs_setting() -> Tuple[str, str, str, str]: class MTCInstruct (line 335) | class MTCInstruct(Enum): method get_method (line 347) | def get_method(self, convo_len: int = -1) -> Callable: FILE: ssvd/multiturn_prompting.py function list_to_prompt (line 21) | def list_to_prompt( function postprocess_mtc (line 95) | def postprocess_mtc( class ConvoIter (line 162) | class ConvoIter: method __init__ (line 165) | def __init__( method add_to_convos (line 180) | def add_to_convos(self, uid: str, answer: str) -> None: method make_iter (line 187) | def make_iter(self, captions: Sequence[str], img_ids: Sequence[str]) -... function run_multiturn_pipeline (line 220) | def run_multiturn_pipeline( FILE: ssvd/utils.py function preprocess_pixelprose_captions (line 18) | def preprocess_pixelprose_captions(caption: str) -> Dict[str, str]: function maybe_shorten_caption (line 38) | def maybe_shorten_caption(caption: str, max_cap_len: int = 1500) -> str: function compile_pattern (line 57) | def compile_pattern(s: str) -> Pattern: function get_replace_pattern (line 63) | def get_replace_pattern() -> Pattern: function get_strings_for_logging (line 73) | def get_strings_for_logging( function sanitize_line (line 103) | def sanitize_line(s: str) -> str: function postprocess_synth_annot (line 115) | def postprocess_synth_annot(