SYMBOL INDEX (2561 symbols across 129 files) FILE: demo/demo.js method formatBytes (line 58) | formatBytes(bytes) { method sleep (line 69) | sleep(ms) { method createModelInputs (line 76) | createModelInputs(model, seqLen = config.defaultSeqLen) { method inferText (line 97) | async inferText(text) { method $ (line 148) | $(id) { method setOutput (line 155) | setOutput(id, content, type = '') { method showLoading (line 166) | showLoading(id, message = 'Loading...') { method showSuccess (line 173) | showSuccess(id, message) { method showError (line 180) | showError(id, error) { method renderStatusList (line 188) | renderStatusList(id, items) { method renderMetrics (line 203) | renderMetrics(id, metrics) { method updateRuntimeStatus (line 220) | async updateRuntimeStatus() { method updateMemoryStatus (line 240) | updateMemoryStatus() { method updateMonitorMetrics (line 260) | updateMonitorMetrics(sample) { method initOutputs (line 273) | initOutputs() { method loadModel (line 310) | async loadModel() { method testModel (line 341) | async testModel() { method testTensors (line 373) | testTensors() { method classifyText (line 401) | async classifyText() { method classifyBatch (line 425) | async classifyBatch() { method extractFeatures (line 454) | async extractFeatures() { method quantize (line 492) | quantize() { method prune (line 521) | prune() { method debug (line 544) | debug() { method benchmark (line 574) | async benchmark() { method testScheduler (line 602) | async testScheduler() { method allocateMemory (line 629) | allocateMemory() { method cleanupMemory (line 655) | cleanupMemory() { method testConcurrency (line 669) | async testConcurrency() { method startMonitor (line 704) | startMonitor() { method stopMonitor (line 718) | stopMonitor() { method simulateInferences (line 727) | simulateInferences() { method openDashboard (line 742) | openDashboard() { method closeDashboard (line 761) | closeDashboard() { method init (line 778) | async init() { method loadModels (line 812) | async loadModels() { method handleFileSelect (line 869) | handleFileSelect(e) { method loadImage (line 879) | async loadImage(source) { method loadSampleImage (line 959) | async loadSampleImage() { method handleClick (line 973) | async handleClick(e, label) { method drawPoints (line 1008) | drawPoints() { method drawMask (line 1027) | drawMask(result) { method clear (line 1063) | clear() { method downloadMask (line 1080) | downloadMask() { method reset (line 1096) | reset() { method init (line 1126) | init() { method loadModel (line 1141) | async loadModel() { method send (line 1216) | async send() { method addMessage (line 1316) | addMessage(role, content, isTyping = false) { method updateStatus (line 1333) | updateStatus(status, text) { method clear (line 1349) | clear() { function init (line 1427) | async function init() { FILE: demo/server.js constant ROOT (line 15) | const ROOT = join(__dirname, '..'); constant MIME_TYPES (line 17) | const MIME_TYPES = { constant PORT (line 29) | const PORT = process.env.PORT || 3000; FILE: dist/backends/index.js function registerAllBackends (line 25) | function registerAllBackends() { FILE: dist/backends/onnx.d.ts class ONNXRuntime (line 15) | class ONNXRuntime implements Runtime { FILE: dist/backends/onnx.js function getOrt (line 14) | async function getOrt() { function isOnnxAvailable (line 33) | async function isOnnxAvailable() { class ONNXRuntime (line 43) | class ONNXRuntime { method capabilities (line 47) | get capabilities() { method isAvailable (line 60) | async isAvailable() { method initialize (line 66) | async initialize() { method loadModel (line 87) | async loadModel(modelData, options = {}) { method run (line 149) | async run(model, inputs) { method runNamed (line 199) | async runNamed(model, namedInputs) { method unloadModel (line 246) | async unloadModel(modelId) { method dispose (line 256) | dispose() { function createONNXRuntime (line 265) | function createONNXRuntime() { FILE: dist/backends/transformers-adapter.d.ts type TransformersPipelineInstance (line 31) | interface TransformersPipelineInstance { type TransformersPipelineFactory (line 39) | type TransformersPipelineFactory = (task: string, model?: string, option... type TransformersAdapterOptions (line 43) | interface TransformersAdapterOptions { class TransformersAdapterRuntime (line 53) | class TransformersAdapterRuntime implements Runtime { FILE: dist/backends/transformers-adapter.js class TransformersAdapterRuntime (line 38) | class TransformersAdapterRuntime { method capabilities (line 40) | get capabilities() { method isAvailable (line 50) | async isAvailable() { method initialize (line 53) | async initialize() { method loadModel (line 59) | async loadModel(modelData, options = {}) { method loadPipeline (line 88) | async loadPipeline(task, model, pipelineOptions) { method run (line 107) | async run(model, inputs) { method runDirect (line 125) | async runDirect(modelId, input, options) { method dispose (line 132) | dispose() { function useTransformersBackend (line 160) | function useTransformersBackend(options) { function getTransformersAdapter (line 168) | function getTransformersAdapter() { FILE: dist/backends/wasm.d.ts class WASMRuntime (line 14) | class WASMRuntime implements Runtime { FILE: dist/backends/wasm.js class WASMRuntime (line 20) | class WASMRuntime { method capabilities (line 26) | get capabilities() { method isAvailable (line 39) | async isAvailable() { method initialize (line 58) | async initialize() { method checkSIMDSupport (line 80) | async checkSIMDSupport() { method createJSFallback (line 99) | createJSFallback(memory) { method loadModel (line 187) | async loadModel(modelData, options = {}) { method run (line 228) | async run(model, inputs) { method executeModel (line 236) | async executeModel(inputs, metadata) { method parseModelConfig (line 276) | parseModelConfig(data) { method loadWeights (line 309) | async loadWeights(_modelData, _wasmData) { method unloadModel (line 316) | unloadModel(modelId) { method ensureInitialized (line 329) | ensureInitialized() { method hasSIMDSupport (line 337) | hasSIMDSupport() { method dispose (line 343) | dispose() { function createWASMRuntime (line 355) | function createWASMRuntime() { FILE: dist/backends/webgpu.d.ts type Navigator (line 14) | interface Navigator { type GPU (line 17) | interface GPU { type GPURequestAdapterOptions (line 20) | interface GPURequestAdapterOptions { type GPUAdapter (line 23) | interface GPUAdapter { type GPUDeviceDescriptor (line 26) | interface GPUDeviceDescriptor { type GPUDevice (line 30) | interface GPUDevice { type GPULimits (line 40) | interface GPULimits { type GPUDeviceLostInfo (line 43) | interface GPUDeviceLostInfo { type GPUBuffer (line 47) | interface GPUBuffer { type GPUShaderModule (line 50) | interface GPUShaderModule { type GPUBindGroupLayout (line 52) | interface GPUBindGroupLayout { type GPUPipelineLayout (line 54) | interface GPUPipelineLayout { type GPUComputePipeline (line 56) | interface GPUComputePipeline { type GPUBufferDescriptor (line 58) | interface GPUBufferDescriptor { type GPUShaderModuleDescriptor (line 62) | interface GPUShaderModuleDescriptor { type GPUBindGroupLayoutDescriptor (line 65) | interface GPUBindGroupLayoutDescriptor { type GPUBindGroupLayoutEntry (line 68) | interface GPUBindGroupLayoutEntry { type GPUPipelineLayoutDescriptor (line 75) | interface GPUPipelineLayoutDescriptor { type GPUComputePipelineDescriptor (line 78) | interface GPUComputePipelineDescriptor { class WebGPURuntime (line 89) | class WebGPURuntime implements Runtime { FILE: dist/backends/webgpu.js class WebGPURuntime (line 32) | class WebGPURuntime { method capabilities (line 38) | get capabilities() { method isAvailable (line 51) | async isAvailable() { method initialize (line 67) | async initialize() { method loadModel (line 96) | async loadModel(modelData, options = {}) { method run (line 142) | async run(model, inputs) { method executeModel (line 151) | async executeModel(inputs, metadata) { method parseModelData (line 191) | parseModelData(data) { method uploadWeights (line 221) | async uploadWeights(_data, modelData) { method createPipelines (line 235) | async createPipelines(modelData) { method unloadModel (line 287) | unloadModel(modelId) { method ensureInitialized (line 300) | ensureInitialized() { method dispose (line 308) | dispose() { function createWebGPURuntime (line 325) | function createWebGPURuntime() { FILE: dist/backends/webnn.d.ts type MLContextType (line 15) | type MLContextType = 'default' | 'gpu' | 'cpu' | 'npu'; type MLOperandDescriptor (line 19) | interface MLOperandDescriptor { type MLContextOptions (line 26) | interface MLContextOptions { type Navigator (line 31) | interface Navigator { type MLContext (line 36) | interface MLContext { type MLGraph (line 39) | interface MLGraph { type MLGraphBuilder (line 41) | interface MLGraphBuilder { type MLOperand (line 57) | interface MLOperand { class WebNNRuntime (line 63) | class WebNNRuntime implements Runtime { FILE: dist/backends/webnn.js class WebNNRuntime (line 21) | class WebNNRuntime { method capabilities (line 27) | get capabilities() { method isAvailable (line 40) | async isAvailable() { method initialize (line 56) | async initialize() { method loadModel (line 84) | async loadModel(modelData, options = {}) { method run (line 118) | async run(model, inputs) { method executeModel (line 126) | async executeModel(inputs, metadata) { method parseModelConfig (line 146) | parseModelConfig(data) { method unloadModel (line 171) | unloadModel(modelId) { method ensureInitialized (line 177) | ensureInitialized() { method getDeviceType (line 185) | getDeviceType() { method dispose (line 191) | dispose() { function createWebNNRuntime (line 200) | function createWebNNRuntime() { FILE: dist/core/composer.d.ts type CompositionStage (line 26) | interface CompositionStage { type CompositionResult (line 46) | interface CompositionResult { type ComposedPipeline (line 59) | interface ComposedPipeline { FILE: dist/core/composer.js function compose (line 50) | function compose(stages) { function parallel (line 125) | function parallel(stages) { FILE: dist/core/device-profiler.d.ts type DeviceTier (line 22) | type DeviceTier = 'high' | 'medium' | 'low'; type DeviceProfile (line 26) | interface DeviceProfile { type ModelRecommendation (line 49) | interface ModelRecommendation { FILE: dist/core/device-profiler.js function getDeviceProfile (line 25) | async function getDeviceProfile() { function recommendQuantization (line 106) | function recommendQuantization(profile) { function recommendModelVariant (line 116) | async function recommendModelVariant() { function resetDeviceProfile (line 128) | function resetDeviceProfile() { FILE: dist/core/memory.d.ts type TrackedResource (line 15) | interface TrackedResource { class MemoryManager (line 31) | class MemoryManager { class MemoryScope (line 157) | class MemoryScope { class ModelCache (line 194) | class ModelCache { FILE: dist/core/memory.js constant DEFAULT_POOL_CONFIG (line 14) | const DEFAULT_POOL_CONFIG = { class MemoryManager (line 33) | class MemoryManager { method constructor (line 43) | constructor(config = {}) { method getInstance (line 49) | static getInstance() { method configure (line 58) | static configure(config) { method track (line 67) | track(tensor, disposer) { method trackModel (line 88) | trackModel(model, disposer) { method untrack (line 109) | untrack(id) { method release (line 120) | release(resourceOrId) { method estimateTensorSize (line 136) | estimateTensorSize(tensor) { method getBytesPerElement (line 143) | getBytesPerElement(dtype) { method captureStackTrace (line 164) | captureStackTrace() { method checkMemoryThreshold (line 175) | checkMemoryThreshold() { method gc (line 200) | gc(evict = false, maxAge = 5 * 60 * 1000) { method measureBrowserMemory (line 224) | async measureBrowserMemory() { method getDeviceMemory (line 242) | getDeviceMemory() { method getStats (line 256) | getStats() { method getResourceDetails (line 278) | getResourceDetails() { method detectLeaks (line 284) | detectLeaks(maxAge = 10 * 60 * 1000) { method on (line 297) | on(event, listener) { method off (line 308) | off(event, listener) { method emit (line 317) | emit(type, data) { method resetStats (line 338) | resetStats() { method disposeAll (line 344) | disposeAll() { method dispose (line 352) | dispose() { class MemoryScope (line 376) | class MemoryScope { method constructor (line 380) | constructor(parent) { method track (line 389) | track(resource) { method createChild (line 396) | createChild() { method keep (line 402) | keep(resource) { method dispose (line 412) | dispose() { function withMemoryScope (line 441) | async function withMemoryScope(fn) { function withMemoryScopeSync (line 453) | function withMemoryScopeSync(fn) { class ModelCache (line 468) | class ModelCache { method constructor (line 473) | constructor(options = {}) { method get (line 480) | get(key) { method set (line 491) | set(key, model) { method delete (line 509) | delete(key) { method has (line 522) | has(key) { method evictLRU (line 528) | evictLRU() { method clear (line 544) | clear() { method getStats (line 554) | getStats() { function getMemoryManager (line 569) | function getMemoryManager() { function getMemoryStats (line 575) | function getMemoryStats() { function release (line 581) | function release(resource) { function gc (line 587) | function gc() { FILE: dist/core/plugin.d.ts type PluginPipelineEntry (line 27) | interface PluginPipelineEntry { type PluginBackendEntry (line 36) | interface PluginBackendEntry { type PluginMiddleware (line 45) | interface PluginMiddleware { type EdgeFlowPlugin (line 62) | interface EdgeFlowPlugin { FILE: dist/core/plugin.js function registerPlugin (line 33) | async function registerPlugin(plugin) { function getPluginPipeline (line 64) | function getPluginPipeline(task) { function getPluginMiddleware (line 70) | function getPluginMiddleware() { function listPlugins (line 76) | function listPlugins() { function unregisterPlugin (line 85) | function unregisterPlugin(name) { FILE: dist/core/runtime.d.ts class RuntimeManager (line 17) | class RuntimeManager { class LoadedModelImpl (line 78) | class LoadedModelImpl implements LoadedModel { FILE: dist/core/runtime.js constant RUNTIME_PRIORITY (line 24) | const RUNTIME_PRIORITY = ['webgpu', 'webnn', 'wasm']; class RuntimeManager (line 37) | class RuntimeManager { method constructor (line 41) | constructor() { } method getInstance (line 45) | static getInstance() { method register (line 54) | register(type, factory) { method getRuntime (line 60) | async getRuntime(type = 'auto') { method getBestRuntime (line 94) | async getBestRuntime() { method detectAvailableRuntimes (line 125) | async detectAvailableRuntimes() { method getCapabilities (line 146) | async getCapabilities(type) { method setDefaultRuntime (line 153) | setDefaultRuntime(type) { method getDefaultRuntimeType (line 159) | getDefaultRuntimeType() { method disposeRuntime (line 165) | disposeRuntime(type) { method disposeAll (line 175) | disposeAll() { method on (line 184) | on(event, listener) { method off (line 195) | off(event, listener) { method emit (line 204) | emit(type, data) { function generateModelId (line 233) | function generateModelId() { class LoadedModelImpl (line 239) | class LoadedModelImpl { method constructor (line 245) | constructor(metadata, runtime, dispose) { method isLoaded (line 251) | get isLoaded() { method dispose (line 254) | dispose() { function loadModel (line 269) | async function loadModel(url, options = {}) { function loadModelFromBuffer (line 291) | async function loadModelFromBuffer(data, options = {}) { function runInference (line 302) | async function runInference(model, inputs) { function runInferenceNamed (line 316) | async function runInferenceNamed(model, namedInputs) { function runBatchInference (line 334) | async function runBatchInference(model, batches) { function getRuntimeManager (line 349) | function getRuntimeManager() { function registerRuntime (line 355) | function registerRuntime(type, factory) { function getBestRuntime (line 361) | async function getBestRuntime() { function getAvailableRuntimes (line 367) | async function getAvailableRuntimes() { FILE: dist/core/scheduler.d.ts class InferenceScheduler (line 18) | class InferenceScheduler { FILE: dist/core/scheduler.js class Task (line 14) | class Task { method constructor (line 27) | constructor(id, modelId, priority, executor) { method status (line 34) | get status() { method startedAt (line 37) | get startedAt() { method completedAt (line 40) | get completedAt() { method result (line 43) | get result() { method error (line 46) | get error() { method cancel (line 52) | cancel() { method wait (line 67) | wait() { method execute (line 84) | async execute() { constant PRIORITY_ORDER (line 115) | const PRIORITY_ORDER = { class PriorityQueue (line 124) | class PriorityQueue { method length (line 126) | get length() { method isEmpty (line 129) | isEmpty() { method enqueue (line 135) | enqueue(item) { method dequeue (line 152) | dequeue() { method peek (line 158) | peek() { method remove (line 164) | remove(id) { method getAll (line 175) | getAll() { method clear (line 181) | clear() { class BatchCollector (line 191) | class BatchCollector { method constructor (line 197) | constructor(maxSize, timeout, onBatch) { method add (line 202) | add(task) { method flush (line 211) | flush() { method clear (line 222) | clear() { function generateTaskId (line 238) | function generateTaskId() { constant DEFAULT_OPTIONS (line 244) | const DEFAULT_OPTIONS = { class InferenceScheduler (line 267) | class InferenceScheduler { method constructor (line 278) | constructor(options = {}) { method getCircuit (line 284) | getCircuit(modelId) { method isCircuitOpen (line 295) | isCircuitOpen(modelId) { method circuitSuccess (line 313) | circuitSuccess(modelId) { method circuitFailure (line 323) | circuitFailure(modelId) { method getQueue (line 340) | getQueue(modelId) { method getRunningSet (line 351) | getRunningSet(modelId) { method canStartTask (line 362) | canStartTask(modelId) { method processQueue (line 375) | async processQueue() { method schedule (line 441) | schedule(modelId, executor, priority = 'normal') { method scheduleWithTimeout (line 493) | scheduleWithTimeout(modelId, executor, timeout = this.options.defaultT... method scheduleAll (line 515) | async scheduleAll(tasks) { method getTask (line 522) | getTask(taskId) { method cancelTask (line 528) | cancelTask(taskId) { method cancelAllForModel (line 543) | cancelAllForModel(modelId) { method getStats (line 560) | getStats() { method on (line 597) | on(event, listener) { method off (line 608) | off(event, listener) { method emit (line 617) | emit(type, data) { method clearHistory (line 638) | clearHistory() { method dispose (line 650) | dispose() { function getScheduler (line 677) | function getScheduler() { function setScheduler (line 686) | function setScheduler(scheduler) { function configureScheduler (line 695) | function configureScheduler(options) { FILE: dist/core/tensor.d.ts class EdgeFlowTensor (line 10) | class EdgeFlowTensor implements Tensor { FILE: dist/core/tensor.js function generateTensorId (line 12) | function generateTensorId() { function getTypedArrayConstructor (line 18) | function getTypedArrayConstructor(dtype) { function calculateSize (line 41) | function calculateSize(shape) { function validateShape (line 49) | function validateShape(shape) { class EdgeFlowTensor (line 60) | class EdgeFlowTensor { method constructor (line 67) | constructor(data, shape, dtype = 'float32') { method data (line 99) | get data() { method isDisposed (line 103) | get isDisposed() { method checkDisposed (line 109) | checkDisposed() { method toFloat32Array (line 117) | toFloat32Array() { method toArray (line 131) | toArray() { method clone (line 147) | clone() { method dispose (line 156) | dispose() { method get (line 166) | get(...indices) { method set (line 187) | set(value, ...indices) { method reshape (line 208) | reshape(newShape) { method transpose (line 221) | transpose() { method toString (line 238) | toString() { function tensor (line 248) | function tensor(data, shape, dtype = 'float32') { function zeros (line 269) | function zeros(shape, dtype = 'float32') { function ones (line 278) | function ones(shape, dtype = 'float32') { function full (line 288) | function full(shape, value, dtype = 'float32') { function random (line 298) | function random(shape, dtype = 'float32') { function randn (line 309) | function randn(shape, dtype = 'float32') { function arange (line 328) | function arange(start, stop, step = 1, dtype = 'float32') { function linspace (line 343) | function linspace(start, stop, num = 50, dtype = 'float32') { function eye (line 354) | function eye(n, dtype = 'float32') { function add (line 367) | function add(a, b) { function sub (line 390) | function sub(a, b) { function mul (line 413) | function mul(a, b) { function div (line 436) | function div(a, b) { function matmul (line 459) | function matmul(a, b) { function softmax (line 485) | function softmax(t, axis = -1) { function relu (line 535) | function relu(t) { function sigmoid (line 546) | function sigmoid(t) { function tanh (line 557) | function tanh(t) { function sum (line 568) | function sum(t, axis) { function mean (line 619) | function mean(t, axis) { function argmax (line 633) | function argmax(t, axis) { function concat (line 670) | function concat(tensors, axis = 0) { FILE: dist/core/types.d.ts type DataType (line 9) | type DataType = 'float32' | 'float16' | 'int32' | 'int64' | 'uint8' | 'i... type TypedArray (line 13) | type TypedArray = Float32Array | Float64Array | Int32Array | BigInt64Arr... type Shape (line 17) | type Shape = readonly number[]; type Tensor (line 21) | interface Tensor { type RuntimeType (line 46) | type RuntimeType = 'webgpu' | 'webnn' | 'wasm' | 'auto'; type RuntimeCapabilities (line 50) | interface RuntimeCapabilities { type Runtime (line 67) | interface Runtime { type ModelFormat (line 88) | type ModelFormat = 'onnx' | 'edgeflow' | 'safetensors'; type QuantizationType (line 92) | type QuantizationType = 'float32' | 'float16' | 'int8' | 'uint8' | 'int4'; type ModelMetadata (line 96) | interface ModelMetadata { type ModelIOSpec (line 123) | interface ModelIOSpec { type ModelLoadOptions (line 136) | interface ModelLoadOptions { type LoadedModel (line 149) | interface LoadedModel { type TaskPriority (line 164) | type TaskPriority = 'low' | 'normal' | 'high' | 'critical'; type TaskStatus (line 168) | type TaskStatus = 'pending' | 'running' | 'completed' | 'failed' | 'canc... type InferenceTask (line 172) | interface InferenceTask { type SchedulerOptions (line 199) | interface SchedulerOptions { type MemoryStats (line 226) | interface MemoryStats { type MemoryPoolConfig (line 241) | interface MemoryPoolConfig { type PipelineTask (line 256) | type PipelineTask = 'text-classification' | 'token-classification' | 'qu... type PipelineConfig (line 260) | interface PipelineConfig { type PipelineOptions (line 279) | interface PipelineOptions { type TokenizerConfig (line 294) | interface TokenizerConfig { type TokenizedOutput (line 317) | interface TokenizedOutput { class EdgeFlowError (line 332) | class EdgeFlowError extends Error { type ErrorCode (line 362) | type ErrorCode = typeof ErrorCodes[keyof typeof ErrorCodes]; type EventType (line 366) | type EventType = 'model:loading' | 'model:loaded' | 'model:unloaded' | '... type EdgeFlowEvent (line 370) | interface EdgeFlowEvent { type EventListener (line 378) | type EventListener = (event: EdgeFlowEvent) => void; FILE: dist/core/types.js class EdgeFlowError (line 12) | class EdgeFlowError extends Error { method constructor (line 15) | constructor(message, code, details) { FILE: dist/core/worker.d.ts type WorkerMessageType (line 10) | type WorkerMessageType = 'init' | 'load_model' | 'run_inference' | 'disp... type WorkerMessage (line 14) | interface WorkerMessage { type LoadModelRequest (line 22) | interface LoadModelRequest { type InferenceRequest (line 32) | interface InferenceRequest { type SerializedTensor (line 39) | interface SerializedTensor { type WorkerPoolOptions (line 47) | interface WorkerPoolOptions { type WorkerHealthState (line 67) | type WorkerHealthState = 'alive' | 'dead' | 'restarting'; class InferenceWorker (line 71) | class InferenceWorker { class WorkerPool (line 138) | class WorkerPool { FILE: dist/core/worker.js function serializeTensor (line 12) | function serializeTensor(tensor) { function deserializeTensor (line 27) | async function deserializeTensor(serialized) { function deserializeTensorSync (line 36) | function deserializeTensorSync(serialized, TensorClass) { constant MAX_RESTART_ATTEMPTS (line 40) | const MAX_RESTART_ATTEMPTS = 3; constant RESTART_BASE_DELAY_MS (line 41) | const RESTART_BASE_DELAY_MS = 1000; class InferenceWorker (line 45) | class InferenceWorker { method constructor (line 54) | constructor(workerUrl) { method health (line 61) | get health() { method initWorker (line 67) | initWorker(workerUrl) { method handleCrash (line 84) | handleCrash() { method attemptRestart (line 97) | attemptRestart() { method restart (line 112) | restart() { method createWorkerBlob (line 137) | createWorkerBlob() { method handleMessage (line 244) | handleMessage(message) { method sendRequest (line 265) | async sendRequest(type, payload) { method init (line 289) | async init() { method loadModel (line 298) | async loadModel(url, options) { method runInference (line 306) | async runInference(modelId, inputs) { method dispose (line 314) | async dispose(modelId) { method terminate (line 320) | terminate() { class WorkerPool (line 335) | class WorkerPool { method constructor (line 340) | constructor(options = {}) { method getNextHealthyWorker (line 351) | getNextHealthyWorker() { method getWorkerForModel (line 368) | getWorkerForModel(modelId) { method replaceWorker (line 384) | replaceWorker(index) { method init (line 394) | async init() { method loadModel (line 400) | async loadModel(url, options) { method runInference (line 409) | async runInference(modelId, inputs) { method runBatch (line 416) | async runBatch(modelId, batchInputs) { method dispose (line 426) | async dispose(modelId) { method terminate (line 434) | terminate() { method size (line 444) | get size() { function getWorkerPool (line 455) | function getWorkerPool(options) { function runInWorker (line 464) | async function runInWorker(modelUrl, inputs, options) { function isWorkerSupported (line 474) | function isWorkerSupported() { FILE: dist/edgeflow.browser.js method "dist/core/types.js" (line 21) | "dist/core/types.js"() { function generateTensorId (line 66) | function generateTensorId() { function getTypedArrayConstructor (line 69) | function getTypedArrayConstructor(dtype) { function calculateSize (line 88) | function calculateSize(shape) { function validateShape (line 93) | function validateShape(shape) { function tensor (line 101) | function tensor(data, shape, dtype = "float32") { function zeros (line 117) | function zeros(shape, dtype = "float32") { function ones (line 123) | function ones(shape, dtype = "float32") { function full (line 130) | function full(shape, value, dtype = "float32") { function random (line 137) | function random(shape, dtype = "float32") { function randn (line 145) | function randn(shape, dtype = "float32") { function arange (line 160) | function arange(start, stop, step = 1, dtype = "float32") { function linspace (line 172) | function linspace(start, stop, num = 50, dtype = "float32") { function eye (line 180) | function eye(n, dtype = "float32") { function add (line 187) | function add(a, b) { function sub (line 207) | function sub(a, b) { function mul (line 227) | function mul(a, b) { function div (line 247) | function div(a, b) { function matmul (line 267) | function matmul(a, b) { function softmax (line 290) | function softmax(t, axis = -1) { function relu (line 334) | function relu(t) { function sigmoid (line 342) | function sigmoid(t) { function tanh (line 350) | function tanh(t) { function sum (line 358) | function sum(t, axis) { function mean (line 402) | function mean(t, axis) { function argmax (line 413) | function argmax(t, axis) { function concat (line 445) | function concat(tensors, axis = 0) { method "dist/core/tensor.js" (line 492) | "dist/core/tensor.js"() { function supportsRangeRequests (line 689) | async function supportsRangeRequests(url) { function downloadChunk (line 704) | async function downloadChunk(url, start, end, timeout) { function downloadWithResume (line 720) | async function downloadWithResume(url, options) { function downloadSimple (line 818) | async function downloadSimple(url, timeout, onProgress) { function loadModelData (line 864) | async function loadModelData(url, options = {}) { function preloadModel (line 907) | function preloadModel(url, options = {}) { function preloadModels (line 910) | function preloadModels(urls, options = {}) { function isModelCached (line 913) | async function isModelCached(url) { function getCachedModel (line 917) | async function getCachedModel(url) { function deleteCachedModel (line 920) | async function deleteCachedModel(url) { function clearModelCache (line 923) | async function clearModelCache() { function getModelCacheStats (line 926) | async function getModelCacheStats() { function getPreloadStatus (line 929) | function getPreloadStatus(url) { function cancelPreload (line 932) | function cancelPreload(url) { function getPreloadedModel (line 935) | async function getPreloadedModel(url) { method "dist/utils/model-loader.js" (line 940) | "dist/utils/model-loader.js"() { method constructor (line 1360) | constructor(id, modelId, priority, executor) { method status (line 1379) | get status() { method startedAt (line 1382) | get startedAt() { method completedAt (line 1385) | get completedAt() { method result (line 1388) | get result() { method error (line 1391) | get error() { method cancel (line 1397) | cancel() { method wait (line 1412) | wait() { method execute (line 1429) | async execute() { method constructor (line 1460) | constructor() { method length (line 1463) | get length() { method isEmpty (line 1466) | isEmpty() { method enqueue (line 1472) | enqueue(item) { method dequeue (line 1489) | dequeue() { method peek (line 1495) | peek() { method remove (line 1501) | remove(id) { method getAll (line 1512) | getAll() { method clear (line 1518) | clear() { function generateTaskId (line 1523) | function generateTaskId() { method constructor (line 1540) | constructor(options = {}) { method getCircuit (line 1556) | getCircuit(modelId) { method isCircuitOpen (line 1567) | isCircuitOpen(modelId) { method circuitSuccess (line 1585) | circuitSuccess(modelId) { method circuitFailure (line 1595) | circuitFailure(modelId) { method getQueue (line 1612) | getQueue(modelId) { method getRunningSet (line 1623) | getRunningSet(modelId) { method canStartTask (line 1634) | canStartTask(modelId) { method processQueue (line 1647) | async processQueue() { method schedule (line 1705) | schedule(modelId, executor, priority = "normal") { method scheduleWithTimeout (line 1751) | scheduleWithTimeout(modelId, executor, timeout = this.options.defaultTim... method scheduleAll (line 1771) | async scheduleAll(tasks) { method getTask (line 1778) | getTask(taskId) { method cancelTask (line 1784) | cancelTask(taskId) { method cancelAllForModel (line 1798) | cancelAllForModel(modelId) { method getStats (line 1815) | getStats() { method on (line 1852) | on(event, listener) { method off (line 1863) | off(event, listener) { method emit (line 1872) | emit(type, data) { method clearHistory (line 1892) | clearHistory() { method dispose (line 1902) | dispose() { function getScheduler (line 1921) | function getScheduler() { function setScheduler (line 1927) | function setScheduler(scheduler) { function configureScheduler (line 1933) | function configureScheduler(options) { method constructor (line 1949) | constructor(config = {}) { method getInstance (line 1963) | static getInstance() { method configure (line 1972) | static configure(config) { method track (line 1981) | track(tensor2, disposer) { method trackModel (line 2002) | trackModel(model, disposer) { method untrack (line 2023) | untrack(id) { method release (line 2034) | release(resourceOrId) { method estimateTensorSize (line 2049) | estimateTensorSize(tensor2) { method getBytesPerElement (line 2056) | getBytesPerElement(dtype) { method captureStackTrace (line 2077) | captureStackTrace() { method checkMemoryThreshold (line 2088) | checkMemoryThreshold() { method gc (line 2112) | gc(evict = false, maxAge = 5 * 60 * 1e3) { method measureBrowserMemory (line 2136) | async measureBrowserMemory() { method getDeviceMemory (line 2150) | getDeviceMemory() { method getStats (line 2162) | getStats() { method getResourceDetails (line 2184) | getResourceDetails() { method detectLeaks (line 2190) | detectLeaks(maxAge = 10 * 60 * 1e3) { method on (line 2203) | on(event, listener) { method off (line 2214) | off(event, listener) { method emit (line 2223) | emit(type, data) { method resetStats (line 2243) | resetStats() { method disposeAll (line 2249) | disposeAll() { method dispose (line 2257) | dispose() { method constructor (line 2267) | constructor(parent) { method track (line 2279) | track(resource) { method createChild (line 2286) | createChild() { method keep (line 2292) | keep(resource) { method dispose (line 2302) | dispose() { function withMemoryScope (line 2324) | async function withMemoryScope(fn) { function withMemoryScopeSync (line 2332) | function withMemoryScopeSync(fn) { method constructor (line 2341) | constructor(options = {}) { method get (line 2352) | get(key) { method set (line 2363) | set(key, model) { method delete (line 2378) | delete(key) { method has (line 2391) | has(key) { method evictLRU (line 2397) | evictLRU() { method clear (line 2413) | clear() { method getStats (line 2423) | getStats() { function getMemoryManager (line 2432) | function getMemoryManager() { function getMemoryStats (line 2435) | function getMemoryStats() { function release (line 2438) | function release(resource) { function gc (line 2441) | function gc() { method constructor (line 2451) | constructor() { method getInstance (line 2458) | static getInstance() { method register (line 2467) | register(type, factory) { method getRuntime (line 2473) | async getRuntime(type = "auto") { method getBestRuntime (line 2502) | async getBestRuntime() { method detectAvailableRuntimes (line 2529) | async detectAvailableRuntimes() { method getCapabilities (line 2549) | async getCapabilities(type) { method setDefaultRuntime (line 2556) | setDefaultRuntime(type) { method getDefaultRuntimeType (line 2562) | getDefaultRuntimeType() { method disposeRuntime (line 2568) | disposeRuntime(type) { method disposeAll (line 2578) | disposeAll() { method on (line 2587) | on(event, listener) { method off (line 2598) | off(event, listener) { method emit (line 2607) | emit(type, data) { function generateModelId (line 2628) | function generateModelId() { method constructor (line 2632) | constructor(metadata, runtime, dispose) { method isLoaded (line 2643) | get isLoaded() { method dispose (line 2646) | dispose() { function loadModel (line 2654) | async function loadModel(url, options = {}) { function loadModelFromBuffer (line 2670) | async function loadModelFromBuffer(data, options = {}) { function runInference (line 2675) | async function runInference(model, inputs) { function runInferenceNamed (line 2685) | async function runInferenceNamed(model, namedInputs) { function runBatchInference (line 2698) | async function runBatchInference(model, batches) { function getRuntimeManager (line 2705) | function getRuntimeManager() { function registerRuntime (line 2708) | function registerRuntime(type, factory) { function getBestRuntime (line 2711) | async function getBestRuntime() { function getAvailableRuntimes (line 2714) | async function getAvailableRuntimes() { function registerPlugin (line 2722) | async function registerPlugin(plugin) { function getPluginPipeline (line 2745) | function getPluginPipeline(task) { function getPluginMiddleware (line 2748) | function getPluginMiddleware() { function listPlugins (line 2751) | function listPlugins() { function unregisterPlugin (line 2757) | function unregisterPlugin(name) { function getDeviceProfile (line 2779) | async function getDeviceProfile() { function recommendQuantization (line 2840) | function recommendQuantization(profile) { function recommendModelVariant (line 2847) | async function recommendModelVariant() { function resetDeviceProfile (line 2856) | function resetDeviceProfile() { method constructor (line 2873) | constructor() { method capabilities (line 2880) | get capabilities() { method isAvailable (line 2893) | async isAvailable() { method initialize (line 2908) | async initialize() { method loadModel (line 2934) | async loadModel(modelData, options = {}) { method run (line 2972) | async run(model, inputs) { method executeModel (line 2979) | async executeModel(inputs, metadata) { method parseModelData (line 3009) | parseModelData(data) { method uploadWeights (line 3033) | async uploadWeights(_data, modelData) { method createPipelines (line 3044) | async createPipelines(modelData) { method unloadModel (line 3095) | unloadModel(modelId) { method ensureInitialized (line 3107) | ensureInitialized() { method dispose (line 3115) | dispose() { function createWebGPURuntime (line 3127) | function createWebGPURuntime() { method constructor (line 3135) | constructor() { method capabilities (line 3142) | get capabilities() { method isAvailable (line 3156) | async isAvailable() { method initialize (line 3171) | async initialize() { method loadModel (line 3196) | async loadModel(modelData, options = {}) { method run (line 3224) | async run(model, inputs) { method executeModel (line 3231) | async executeModel(inputs, metadata) { method parseModelConfig (line 3249) | parseModelConfig(data) { method unloadModel (line 3272) | unloadModel(modelId) { method ensureInitialized (line 3278) | ensureInitialized() { method getDeviceType (line 3286) | getDeviceType() { method dispose (line 3292) | dispose() { function createWebNNRuntime (line 3298) | function createWebNNRuntime() { method constructor (line 3306) | constructor() { method capabilities (line 3313) | get capabilities() { method isAvailable (line 3328) | async isAvailable() { method initialize (line 3353) | async initialize() { method checkSIMDSupport (line 3372) | async checkSIMDSupport() { method createJSFallback (line 3416) | createJSFallback(memory) { method loadModel (line 3501) | async loadModel(modelData, options = {}) { method run (line 3536) | async run(model, inputs) { method executeModel (line 3543) | async executeModel(inputs, metadata) { method parseModelConfig (line 3574) | parseModelConfig(data) { method loadWeights (line 3603) | async loadWeights(_modelData, _wasmData) { method unloadModel (line 3608) | unloadModel(modelId) { method ensureInitialized (line 3620) | ensureInitialized() { method hasSIMDSupport (line 3628) | hasSIMDSupport() { method dispose (line 3634) | dispose() { function createWASMRuntime (line 3642) | function createWASMRuntime() { function getOrt (line 3650) | async function getOrt() { function isOnnxAvailable (line 3660) | async function isOnnxAvailable() { method constructor (line 3665) | constructor() { method capabilities (line 3671) | get capabilities() { method isAvailable (line 3685) | async isAvailable() { method initialize (line 3691) | async initialize() { method loadModel (line 3707) | async loadModel(modelData, options = {}) { method run (line 3759) | async run(model, inputs) { method runNamed (line 3805) | async runNamed(model, namedInputs) { method unloadModel (line 3848) | async unloadModel(modelId) { method dispose (line 3857) | dispose() { function createONNXRuntime (line 3862) | function createONNXRuntime() { method constructor (line 3872) | constructor() { method capabilities (line 3876) | get capabilities() { method isAvailable (line 3886) | async isAvailable() { method initialize (line 3889) | async initialize() { method loadModel (line 3894) | async loadModel(modelData, options = {}) { method loadPipeline (line 3921) | async loadPipeline(task, model, pipelineOptions) { method run (line 3940) | async run(model, inputs) { method runDirect (line 3954) | async runDirect(modelId, input, options) { method dispose (line 3961) | dispose() { function useTransformersBackend (line 3971) | function useTransformersBackend(options) { function getTransformersAdapter (line 3976) | function getTransformersAdapter() { function registerAllBackends (line 3981) | function registerAllBackends() { method constructor (line 3988) | constructor(options = {}) { method get (line 4011) | get(key) { method set (line 4030) | set(key, value, size, ttl) { method has (line 4055) | has(key) { method delete (line 4068) | delete(key) { method clear (line 4083) | clear() { method getStats (line 4095) | getStats() { method evict (line 4108) | evict() { method findLRU (line 4131) | findLRU() { method findLFU (line 4145) | findLFU() { method findOldest (line 4159) | findOldest() { method findExpired (line 4173) | findExpired() { method loadFromStorage (line 4185) | async loadFromStorage() { method saveToStorage (line 4210) | async saveToStorage() { method clearStorage (line 4231) | async clearStorage() { method openDB (line 4245) | openDB() { method generateKey (line 4263) | generateKey(modelId, input) { method hashArray (line 4271) | hashArray(arr) { method constructor (line 4283) | constructor(cacheName = "edgeflow-models") { method ensureCache (line 4291) | async ensureCache() { method get (line 4303) | async get(url) { method put (line 4314) | async put(url, response) { method delete (line 4324) | async delete(url) { method clear (line 4335) | async clear() { method keys (line 4345) | async keys() { function createCache (line 4355) | function createCache(preset = "medium", options = {}) { method constructor (line 4379) | constructor(config) { method initialize (line 4397) | async initialize() { method loadModelWithCache (line 4416) | async loadModelWithCache(modelPath) { method run (line 4436) | async run(input, options) { method runBatch (line 4450) | async runBatch(inputs, options) { method task (line 4458) | get task() { method ready (line 4464) | get ready() { method dispose (line 4470) | dispose() { function registerPipeline (line 4479) | function registerPipeline(task, factory) { function getPipelineFactory (line 4482) | function getPipelineFactory(task) { method constructor (line 4514) | constructor() { method initByteEncoder (line 4545) | initByteEncoder() { method fromJSON (line 4572) | static async fromJSON(json) { method fromUrl (line 4640) | static async fromUrl(url) { method fromHuggingFace (line 4651) | static async fromHuggingFace(modelId, options) { method normalize (line 4659) | normalize(text) { method preTokenize (line 4673) | preTokenize(text) { method textToBytes (line 4681) | textToBytes(text) { method bytesToText (line 4689) | bytesToText(text) { method getPairs (line 4697) | getPairs(word) { method bpe (line 4707) | bpe(token) { method wordPiece (line 4760) | wordPiece(word) { method tokenizeWord (line 4793) | tokenizeWord(word) { method unigramTokenize (line 4814) | unigramTokenize(word) { method tokenize (line 4843) | tokenize(text) { method convertTokensToIds (line 4877) | convertTokensToIds(tokens) { method convertIdsToTokens (line 4891) | convertIdsToTokens(ids) { method postProcess (line 4897) | postProcess(ids, pairIds) { method encode (line 4945) | encode(text, options = {}) { method encodeBatch (line 4999) | encodeBatch(texts, options = {}) { method decode (line 5010) | decode(ids, skipSpecialTokens = true) { method decodeBatch (line 5028) | decodeBatch(batchIds, skipSpecialTokens = true) { method vocabSize (line 5034) | get vocabSize() { method getSpecialTokenIds (line 5040) | getSpecialTokenIds() { method getConfig (line 5054) | getConfig() { method isSpecialToken (line 5070) | isSpecialToken(token) { method getTokenId (line 5076) | getTokenId(token) { method getToken (line 5082) | getToken(id) { function createBasicTokenizer (line 5086) | function createBasicTokenizer() { function loadTokenizer (line 5090) | async function loadTokenizer(url) { function loadTokenizerFromHub (line 5093) | async function loadTokenizerFromHub(modelId, options) { method constructor (line 5105) | constructor(config, labels) { method initialize (line 5116) | async initialize() { method setLabels (line 5126) | setLabels(labels) { method run (line 5129) | async run(input, options) { method preprocess (line 5147) | async preprocess(input) { method runInference (line 5158) | async runInference(inputs) { method postprocess (line 5165) | async postprocess(outputs, options) { method constructor (line 5188) | constructor(config) { method analyze (line 5191) | async analyze(text, options) { function createTextClassificationPipeline (line 5195) | function createTextClassificationPipeline(config = {}) { function createSentimentAnalysisPipeline (line 5204) | function createSentimentAnalysisPipeline(config = {}) { method constructor (line 5225) | constructor(config, embeddingDim = DEFAULT_EMBEDDING_DIM) { method initialize (line 5236) | async initialize() { method run (line 5246) | async run(input, options) { method preprocess (line 5264) | async preprocess(input) { method runInference (line 5276) | async runInference(inputs) { method postprocess (line 5284) | async postprocess(outputs, options) { method extractCLSEmbedding (line 5315) | extractCLSEmbedding(hiddenStates) { method meanPooling (line 5320) | meanPooling(hiddenStates) { method maxPooling (line 5332) | maxPooling(hiddenStates) { method normalizeVector (line 5347) | normalizeVector(vec) { function createFeatureExtractionPipeline (line 5358) | function createFeatureExtractionPipeline(config = {}) { method constructor (line 5391) | constructor(options = {}) { method fromConfig (line 5410) | static fromConfig(config) { method fromUrl (line 5467) | static async fromUrl(url) { method fromHuggingFace (line 5478) | static async fromHuggingFace(modelId, options) { method ensureCanvas (line 5486) | ensureCanvas() { method process (line 5499) | async process(input) { method processBatch (line 5522) | async processBatch(inputs) { method loadFromUrl (line 5544) | async loadFromUrl(url) { method loadFromBlob (line 5560) | async loadFromBlob(blob) { method centerCrop (line 5571) | centerCrop(imageData) { method toImageData (line 5598) | toImageData(source) { method resize (line 5609) | resize(imageData) { method toTensor (line 5644) | toTensor(imageData) { method getOptions (line 5697) | getOptions() { method constructor (line 5710) | constructor(options = {}) { method fromConfig (line 5718) | static fromConfig(config) { method fromHuggingFace (line 5741) | static async fromHuggingFace(modelId, options) { method ensureAudioContext (line 5754) | ensureAudioContext() { method process (line 5766) | async process(input) { method processRaw (line 5792) | async processRaw(input) { method loadFromUrl (line 5817) | async loadFromUrl(url) { method loadFromBlob (line 5828) | async loadFromBlob(blob) { method decodeAudioData (line 5835) | async decodeAudioData(data) { method audioBufferToFloat32 (line 5843) | audioBufferToFloat32(buffer) { method normalizeAudio (line 5850) | normalizeAudio(data) { method computeMelSpectrogram (line 5869) | computeMelSpectrogram(audio) { method dispose (line 5894) | dispose() { function preprocessText (line 5901) | function preprocessText(text, options = {}) { function createImagePreprocessor (line 5918) | function createImagePreprocessor(preset = "imagenet", options = {}) { function createAudioPreprocessor (line 5942) | function createAudioPreprocessor(preset = "whisper", options = {}) { method constructor (line 5965) | constructor(config, labels, _numClasses = 1e3) { method initialize (line 5974) | async initialize() { method setLabels (line 5984) | setLabels(labels) { method run (line 5987) | async run(input, options) { method preprocess (line 6005) | async preprocess(input) { method runModelInference (line 6013) | async runModelInference(inputs) { method postprocess (line 6017) | async postprocess(outputs, options) { function createImageClassificationPipeline (line 6036) | function createImageClassificationPipeline(config = {}, labels) { method constructor (line 6054) | constructor(config) { method isModelLoaded (line 6078) | get isModelLoaded() { method setModelUrls (line 6084) | setModelUrls(model, tokenizer) { method loadModel (line 6091) | async loadModel(onProgress) { method fetchModelWithProgress (line 6127) | async fetchModelWithProgress(url, onProgress) { method initialize (line 6161) | async initialize() { method setTokenizer (line 6169) | setTokenizer(tokenizer) { method preprocess (line 6177) | async preprocess(input) { method postprocess (line 6191) | async postprocess(_outputs, _options) { method run (line 6202) | async run(prompt, options) { method stream (line 6211) | async *stream(prompt, options = {}) { method generateSingle (line 6268) | async generateSingle(prompt, options) { method generateNextToken (line 6317) | async generateNextToken(inputIds, temperature, topK, topP, repetitionPen... method greedy (line 6369) | greedy(probs) { method sample (line 6383) | sample(probs, topK, topP) { method setChatTemplate (line 6418) | setChatTemplate(templateType) { method applyChatTemplate (line 6424) | applyChatTemplate(messages, options) { method applyChatMLTemplate (line 6450) | applyChatMLTemplate(messages) { method applyLlama2Template (line 6463) | applyLlama2Template(messages) { method applyLlama3Template (line 6489) | applyLlama3Template(messages) { method applyMistralTemplate (line 6502) | applyMistralTemplate(messages) { method applyPhi3Template (line 6519) | applyPhi3Template(messages) { method applyAlpacaTemplate (line 6532) | applyAlpacaTemplate(messages) { method applyVicunaTemplate (line 6561) | applyVicunaTemplate(messages) { method applyCustomTemplate (line 6582) | applyCustomTemplate(messages, template) { method chat (line 6624) | async chat(userMessage, options) { method chatStream (line 6658) | async *chatStream(userMessage, options) { method getConversationHistory (line 6692) | getConversationHistory() { method setConversationHistory (line 6698) | setConversationHistory(messages) { method clearConversation (line 6704) | clearConversation() { method undoLastExchange (line 6710) | undoLastExchange() { function createTextGenerationPipeline (line 6719) | function createTextGenerationPipeline(config) { method constructor (line 6812) | constructor(config, labels) { method initialize (line 6831) | async initialize() { method setLabels (line 6838) | setLabels(labels) { method run (line 6841) | async run(input, options) { method preprocess (line 6847) | async preprocess(input) { method runModelInference (line 6855) | async runModelInference(inputs) { method postprocess (line 6859) | async postprocess(outputs, options) { method parseDetections (line 6876) | parseDetections(data, shape, threshold) { method nonMaxSuppression (line 6949) | nonMaxSuppression(detections, iouThreshold) { method computeIoU (line 6974) | computeIoU(a, b) { method constructor (line 7002) | constructor(config) { method initialize (line 7025) | async initialize() { method setTokenizer (line 7039) | setTokenizer(tokenizer) { method run (line 7042) | async run(input, options) { method transcribeSingle (line 7054) | async transcribeSingle(audio, options) { method buildInitialTokens (line 7073) | buildInitialTokens(task, language) { method getLanguageToken (line 7080) | getLanguageToken(language) { method autoregressiveDecode (line 7109) | async autoregressiveDecode(encoderHidden, initialTokens) { method extractTimestamps (line 7134) | extractTimestamps(_tokenIds, text) { method processLongAudio (line 7155) | async processLongAudio(audio, options = {}) { method preprocess (line 7186) | async preprocess(input) { method postprocess (line 7195) | async postprocess(outputs, options) { method decodeOutput (line 7210) | decodeOutput(data, shape) { method constructor (line 7249) | constructor(config) { method initialize (line 7262) | async initialize() { method setTokenizer (line 7272) | setTokenizer(tokenizer) { method classify (line 7275) | async classify(text, candidateLabels, options) { method run (line 7278) | async run(input, options) { method classifySingle (line 7288) | async classifySingle(text, candidateLabels, template, multiLabel) { method scoreHypothesis (line 7319) | async scoreHypothesis(premise, hypothesis) { method preprocess (line 7336) | async preprocess(input) { method postprocess (line 7347) | async postprocess(_outputs, _options) { method constructor (line 7365) | constructor(config) { method initialize (line 7377) | async initialize() { method setTokenizer (line 7387) | setTokenizer(tokenizer) { method run (line 7390) | async run(input, options) { method answerQuestion (line 7396) | async answerQuestion(input, options) { method tokenOffsetToCharOffset (line 7447) | tokenOffsetToCharOffset(context, _question, inputIds, tokenIdx) { method preprocess (line 7452) | async preprocess(input) { method postprocess (line 7467) | async postprocess(outputs, _options) { method constructor (line 7506) | constructor(config) { method isModelsLoaded (line 7526) | get isModelsLoaded() { method setModelUrls (line 7532) | setModelUrls(encoder, decoder) { method loadModels (line 7539) | async loadModels(onProgress) { method fetchModelWithProgress (line 7573) | async fetchModelWithProgress(url, onProgress) { method initialize (line 7607) | async initialize() { method loadEncoder (line 7615) | async loadEncoder(modelUrl) { method loadDecoder (line 7623) | async loadDecoder(modelUrl) { method setImage (line 7631) | async setImage(image) { method segment (line 7658) | async segment(options = {}) { method run (line 7684) | async run(input, options) { method loadImage (line 7691) | async loadImage(input) { method loadImageFromUrl (line 7708) | async loadImageFromUrl(url) { method imageElementToImageData (line 7727) | imageElementToImageData(img) { method canvasToImageData (line 7738) | canvasToImageData(canvas) { method imageBitmapToImageData (line 7745) | imageBitmapToImageData(bitmap) { method preprocessImage (line 7756) | preprocessImage(imageData) { method prepareDecoderInputs (line 7800) | prepareDecoderInputs(points, boxes) { method postprocessMasks (line 7829) | postprocessMasks(masks, scores, threshold, returnAllMasks) { method resizeMask (line 7867) | resizeMask(masksData, maskIdx, srcW, srcH, dstW, dstH, threshold) { method clearImage (line 7894) | clearImage() { method preprocess (line 7903) | async preprocess(input) { method postprocess (line 7911) | async postprocess(_outputs, _options) { method dispose (line 7922) | dispose() { function createImageSegmentationPipeline (line 7933) | function createImageSegmentationPipeline(config = {}) { function pipeline (line 7945) | async function pipeline(task, options) { function createPipelines (line 7998) | async function createPipelines(tasks, options) { function compose (line 8009) | function compose(stages) { function parallel (line 8063) | function parallel(stages) { function buildFileUrl (line 8117) | function buildFileUrl(modelId, filename, options = {}) { function fetchWithAuth (line 8123) | async function fetchWithAuth(url, token) { function fileExists (line 8131) | async function fileExists(modelId, filename, options = {}) { function findOnnxModel (line 8140) | async function findOnnxModel(modelId, options = {}) { function downloadFile (line 8148) | async function downloadFile(modelId, filename, options = {}) { function downloadJson (line 8164) | async function downloadJson(modelId, filename, options = {}) { function downloadTokenizer (line 8180) | async function downloadTokenizer(modelId, options = {}) { function downloadConfig (line 8184) | async function downloadConfig(modelId, options = {}) { function downloadModel (line 8187) | async function downloadModel(modelId, options = {}) { function fromHub (line 8254) | async function fromHub(modelId, options = {}) { function modelExists (line 8257) | async function modelExists(modelId, options = {}) { function getModelInfo (line 8265) | async function getModelInfo(modelId, options = {}) { function getDefaultModel (line 8313) | function getDefaultModel(task) { function fromTask (line 8316) | async function fromTask(task, options = {}) { function benchmark (line 8322) | async function benchmark(fn, options = {}) { function compareBenchmarks (line 8391) | async function compareBenchmarks(baseline, comparison, options = {}) { function benchmarkSuite (line 8418) | async function benchmarkSuite(suite, options = {}) { function formatBenchmarkResult (line 8427) | function formatBenchmarkResult(result) { function formatComparisonResult (line 8444) | function formatComparisonResult(result) { function benchmarkMemory (line 8461) | async function benchmarkMemory(fn, options = {}) { function calculateQuantParams (line 8491) | function calculateQuantParams(data, bits, symmetric, perChannel, channel... function quantizeToInt8 (line 8547) | function quantizeToInt8(data, scale, zeroPoint, perChannel, channelSize ... function quantizeToUint8 (line 8570) | function quantizeToUint8(data, scale, zeroPoint, perChannel, channelSize... function quantizeToInt4 (line 8593) | function quantizeToInt4(data, scale, zeroPoint) { function quantizeToFloat16 (line 8605) | function quantizeToFloat16(data) { function float32ToFloat16 (line 8612) | function float32ToFloat16(value) { function dequantizeInt8 (line 8631) | function dequantizeInt8(data, scale, zeroPoint, perChannel = false, chan... function dequantizeUint8 (line 8652) | function dequantizeUint8(data, scale, zeroPoint, perChannel = false, cha... function float16ToFloat32 (line 8673) | function float16ToFloat32(value) { function dequantizeFloat16 (line 8690) | function dequantizeFloat16(data) { function parseModelWeights (line 8697) | function parseModelWeights(modelData) { function serializeQuantizedModel (line 8708) | function serializeQuantizedModel(model) { function quantizeModel (line 8816) | async function quantizeModel(modelData, options) { function quantizeTensor (line 8964) | function quantizeTensor(tensor2, type, options = {}) { function dequantizeTensor (line 8997) | function dequantizeTensor(tensor2, scale, zeroPoint, type) { function pruneTensor (line 9019) | function pruneTensor(tensor2, options = {}) { function pruneModel (line 9058) | async function pruneModel(modelData, options = {}) { function analyzeModel (line 9082) | async function analyzeModel(modelData) { function exportModel (line 9132) | async function exportModel(modelData, options) { function calculateTensorStats (line 9152) | function calculateTensorStats(data) { function createHistogram (line 9197) | function createHistogram(data, bins = 50) { function inspectTensor (line 9230) | function inspectTensor(tensor2, name = "tensor", options = {}) { function formatTensorInspection (line 9254) | function formatTensorInspection(inspection) { function formatBytes (line 9280) | function formatBytes(bytes) { method constructor (line 9290) | constructor(config = {}) { method defaultLogger (line 9320) | defaultLogger(level, message, data) { method log (line 9343) | log(level, message, data) { method addEvent (line 9356) | addEvent(event) { method enable (line 9369) | enable() { method disable (line 9376) | disable() { method on (line 9382) | on(type, callback) { method inspectTensor (line 9395) | inspectTensor(tensor2, name = "tensor") { method startTrace (line 9420) | startTrace(modelId) { method traceInput (line 9439) | traceInput(traceId, tensor2, name) { method traceOutput (line 9448) | traceOutput(traceId, tensor2, name) { method traceOperation (line 9457) | traceOperation(traceId, operation) { method endTrace (line 9466) | endTrace(traceId) { method recordAllocation (line 9493) | recordAllocation(tensor2) { method recordDeallocation (line 9504) | recordDeallocation(tensor2) { method getPerformanceMetrics (line 9514) | getPerformanceMetrics() { method getEvents (line 9520) | getEvents() { method getTraces (line 9526) | getTraces() { method getTrace (line 9532) | getTrace(traceId) { method clear (line 9538) | clear() { method export (line 9556) | export() { method generateReport (line 9567) | generateReport() { function getDebugger (line 9600) | function getDebugger(config) { function enableDebugging (line 9606) | function enableDebugging(config) { function disableDebugging (line 9611) | function disableDebugging() { function createAsciiHistogram (line 9614) | function createAsciiHistogram(histogram, width = 50, height = 10) { function createTensorHeatmap (line 9634) | function createTensorHeatmap(tensor2, width = 40) { function visualizeModelArchitecture (line 9671) | function visualizeModelArchitecture(layers) { method constructor (line 9692) | constructor(config = {}) { method start (line 9725) | start() { method stop (line 9741) | stop() { method monitorFPS (line 9755) | monitorFPS() { method collectSample (line 9771) | collectSample() { method collectMemoryMetrics (line 9816) | collectMemoryMetrics() { method collectSystemMetrics (line 9838) | collectSystemMetrics() { method estimateCPUUsage (line 9861) | estimateCPUUsage() { method checkAlerts (line 9870) | checkAlerts(sample) { method getMetricValue (line 9911) | getMetricValue(sample, metric) { method recordInference (line 9926) | recordInference(duration) { method updateQueueLength (line 9933) | updateQueueLength(length) { method updateActiveCount (line 9939) | updateActiveCount(count) { method updateTensorMemory (line 9945) | updateTensorMemory(bytes) { method updateCacheMemory (line 9951) | updateCacheMemory(bytes) { method addAlert (line 9957) | addAlert(config) { method removeAlert (line 9963) | removeAlert(metric) { method onAlert (line 9969) | onAlert(callback) { method onSample (line 9980) | onSample(callback) { method getCurrentSample (line 9991) | getCurrentSample() { method getSamples (line 9997) | getSamples() { method getSamplesInRange (line 10003) | getSamplesInRange(startTime, endTime) { method getSummary (line 10009) | getSummary() { method clear (line 10040) | clear() { method export (line 10052) | export() { function generateDashboardHTML (line 10061) | function generateDashboardHTML(monitor) { function generateChartPath (line 10441) | function generateChartPath(samples) { function generateAsciiDashboard (line 10467) | function generateAsciiDashboard(monitor) { function getMonitor (line 10526) | function getMonitor(config) { function startMonitoring (line 10532) | function startMonitoring(config) { function stopMonitoring (line 10537) | function stopMonitoring() { function quantize (line 10542) | async function quantize(model, options) { function getModelData (line 10575) | async function getModelData(_model) { function quantizeInt8 (line 10578) | function quantizeInt8(data, _options) { function quantizeUint8 (line 10597) | function quantizeUint8(data, _options) { function quantizeFloat16 (line 10618) | function quantizeFloat16(data, _options) { function quantizeInt4 (line 10630) | function quantizeInt4(data, _options) { function float32ToFloat162 (line 10651) | function float32ToFloat162(value) { function prune (line 10676) | async function prune(model, options) { function analyzeModel2 (line 10698) | async function analyzeModel2(model) { function benchmark2 (line 10715) | async function benchmark2(runFn, options = {}) { function exportModel2 (line 10742) | async function exportModel2(model, format) { function isSupported (line 10756) | async function isSupported() { function getBestRuntimeType (line 10760) | async function getBestRuntimeType() { function preload (line 10770) | async function preload(models) { function getInfo (line 10782) | async function getInfo() { FILE: dist/index.js function isSupported (line 114) | async function isSupported() { function getBestRuntimeType (line 121) | async function getBestRuntimeType() { function preload (line 134) | async function preload(models) { constant VERSION (line 151) | const VERSION = '0.1.0'; function getInfo (line 155) | async function getInfo() { FILE: dist/pipelines/automatic-speech-recognition.d.ts type ASROptions (line 11) | interface ASROptions extends PipelineOptions { type WordTimestamp (line 19) | interface WordTimestamp { type ChunkTimestamp (line 25) | interface ChunkTimestamp { type ASRResult (line 30) | interface ASRResult extends PipelineResult { class AutomaticSpeechRecognitionPipeline (line 36) | class AutomaticSpeechRecognitionPipeline extends BasePipeline BasePipeline; FILE: dist/pipelines/base.js class BasePipeline (line 15) | class BasePipeline { method constructor (line 21) | constructor(config) { method initialize (line 34) | async initialize() { method loadModelWithCache (line 56) | async loadModelWithCache(modelPath) { method run (line 83) | async run(input, options) { method runBatch (line 100) | async runBatch(inputs, options) { method task (line 109) | get task() { method ready (line 115) | get ready() { method dispose (line 121) | dispose() { function registerPipeline (line 136) | function registerPipeline(task, factory) { function getPipelineFactory (line 142) | function getPipelineFactory(task) { constant SENTIMENT_LABELS (line 151) | const SENTIMENT_LABELS = ['negative', 'positive']; constant EMOTION_LABELS (line 155) | const EMOTION_LABELS = [ constant IMAGENET_LABELS (line 161) | const IMAGENET_LABELS = [ FILE: dist/pipelines/feature-extraction.d.ts type FeatureExtractionOptions (line 9) | interface FeatureExtractionOptions extends PipelineOptions { class FeatureExtractionPipeline (line 14) | class FeatureExtractionPipeline extends BasePipeline { class InferenceCache (line 116) | class InferenceCache extends Cache { class ModelDownloadCache (line 129) | class ModelDownloadCache { FILE: dist/utils/cache.js class Cache (line 12) | class Cache { method constructor (line 18) | constructor(options = {}) { method get (line 35) | get(key) { method set (line 56) | set(key, value, size, ttl) { method has (line 88) | has(key) { method delete (line 102) | delete(key) { method clear (line 117) | clear() { method getStats (line 129) | getStats() { method evict (line 142) | evict() { method findLRU (line 165) | findLRU() { method findLFU (line 179) | findLFU() { method findOldest (line 193) | findOldest() { method findExpired (line 207) | findExpired() { method loadFromStorage (line 219) | async loadFromStorage() { method saveToStorage (line 246) | async saveToStorage() { method clearStorage (line 271) | async clearStorage() { method openDB (line 287) | openDB() { class InferenceCache (line 307) | class InferenceCache extends Cache { method generateKey (line 311) | generateKey(modelId, input) { method hashArray (line 320) | hashArray(arr) { class ModelDownloadCache (line 339) | class ModelDownloadCache { method constructor (line 342) | constructor(cacheName = 'edgeflow-models') { method ensureCache (line 348) | async ensureCache() { method get (line 360) | async get(url) { method put (line 372) | async put(url, response) { method delete (line 384) | async delete(url) { method clear (line 396) | async clear() { method keys (line 408) | async keys() { function createCache (line 425) | function createCache(preset = 'medium', options = {}) { FILE: dist/utils/hub.d.ts type HubOptions (line 11) | interface HubOptions { type HubDownloadProgress (line 30) | interface HubDownloadProgress { type ModelConfig (line 45) | interface ModelConfig { type ModelBundle (line 61) | interface ModelBundle { type PopularModelTask (line 148) | type PopularModelTask = keyof typeof POPULAR_MODELS; FILE: dist/utils/hub.js constant DEFAULT_ENDPOINT (line 12) | const DEFAULT_ENDPOINT = 'https://huggingface.co'; constant DEFAULT_REVISION (line 13) | const DEFAULT_REVISION = 'main'; constant ONNX_MODEL_FILES (line 17) | const ONNX_MODEL_FILES = [ function buildFileUrl (line 32) | function buildFileUrl(modelId, filename, options = {}) { function fetchWithAuth (line 41) | async function fetchWithAuth(url, token) { function fileExists (line 52) | async function fileExists(modelId, filename, options = {}) { function findOnnxModel (line 66) | async function findOnnxModel(modelId, options = {}) { function downloadFile (line 78) | async function downloadFile(modelId, filename, options = {}) { function downloadJson (line 98) | async function downloadJson(modelId, filename, options = {}) { function downloadTokenizer (line 119) | async function downloadTokenizer(modelId, options = {}) { function downloadConfig (line 126) | async function downloadConfig(modelId, options = {}) { function downloadModel (line 132) | async function downloadModel(modelId, options = {}) { function fromHub (line 220) | async function fromHub(modelId, options = {}) { function modelExists (line 226) | async function modelExists(modelId, options = {}) { function getModelInfo (line 239) | async function getModelInfo(modelId, options = {}) { constant POPULAR_MODELS (line 259) | const POPULAR_MODELS = { function getDefaultModel (line 296) | function getDefaultModel(task) { function fromTask (line 307) | async function fromTask(task, options = {}) { FILE: dist/utils/model-loader.d.ts type DownloadProgress (line 13) | interface DownloadProgress { type ModelLoaderOptions (line 32) | interface ModelLoaderOptions { type PreloadOptions (line 53) | interface PreloadOptions extends ModelLoaderOptions { FILE: dist/utils/model-loader.js constant DB_NAME (line 13) | const DB_NAME = 'edgeflow-model-cache'; constant DB_VERSION (line 14) | const DB_VERSION = 1; constant STORE_META (line 15) | const STORE_META = 'meta'; constant STORE_CHUNKS (line 16) | const STORE_CHUNKS = 'chunks'; constant STORE_STATE (line 17) | const STORE_STATE = 'download-state'; class ModelCache (line 21) | class ModelCache { method openDB (line 27) | async openDB() { method getMeta (line 61) | async getMeta(url) { method saveMeta (line 74) | async saveMeta(meta) { method saveChunk (line 96) | async saveChunk(url, index, data) { method putInStore (line 118) | async putInStore(storeName, value) { method isQuotaError (line 131) | isQuotaError(err) { method evictOldest (line 141) | async evictOldest(bytesNeeded) { method getChunks (line 162) | async getChunks(url) { method getModel (line 181) | async getModel(url) { method saveDownloadState (line 201) | async saveDownloadState(state) { method getDownloadState (line 217) | async getDownloadState(url) { method deleteDownloadState (line 230) | async deleteDownloadState(url) { method deleteModel (line 243) | async deleteModel(url) { method clear (line 278) | async clear() { method getStats (line 294) | async getStats() { function supportsRangeRequests (line 319) | async function supportsRangeRequests(url) { function downloadChunk (line 338) | async function downloadChunk(url, start, end, timeout) { function downloadWithResume (line 358) | async function downloadWithResume(url, options) { function downloadSimple (line 468) | async function downloadSimple(url, timeout, onProgress) { class PreloadManager (line 520) | class PreloadManager { method preload (line 528) | preload(url, options = {}) { method processQueue (line 569) | async processQueue() { method downloadTask (line 588) | async downloadTask(task) { method isPreloaded (line 602) | isPreloaded(url) { method getStatus (line 609) | getStatus(url) { method get (line 616) | async get(url) { method cancel (line 628) | cancel(url) { method clear (line 639) | clear() { function loadModelData (line 657) | async function loadModelData(url, options = {}) { function preloadModel (line 712) | function preloadModel(url, options = {}) { function preloadModels (line 718) | function preloadModels(urls, options = {}) { function isModelCached (line 724) | async function isModelCached(url) { function getCachedModel (line 731) | async function getCachedModel(url) { function deleteCachedModel (line 737) | async function deleteCachedModel(url) { function clearModelCache (line 743) | async function clearModelCache() { function getModelCacheStats (line 749) | async function getModelCacheStats() { function getPreloadStatus (line 755) | function getPreloadStatus(url) { function cancelPreload (line 761) | function cancelPreload(url) { function getPreloadedModel (line 767) | async function getPreloadedModel(url) { FILE: dist/utils/offline.d.ts type OfflineConfig (line 6) | interface OfflineConfig { type OfflineStatus (line 20) | interface OfflineStatus { type CachedModelInfo (line 32) | interface CachedModelInfo { class OfflineManager (line 42) | class OfflineManager { FILE: dist/utils/offline.js class OfflineManager (line 12) | class OfflineManager { method constructor (line 16) | constructor(config = {}) { method initialize (line 29) | async initialize() { method registerServiceWorker (line 55) | async registerServiceWorker() { method preloadForOffline (line 81) | async preloadForOffline(modelUrls) { method getStatus (line 97) | async getStatus() { method getCachedModels (line 123) | async getCachedModels() { method isModelAvailableOffline (line 146) | async isModelAvailableOffline(url) { method removeFromOffline (line 153) | async removeFromOffline(url) { method clearOfflineData (line 160) | async clearOfflineData() { method getStorageInfo (line 167) | async getStorageInfo() { method requestPersistentStorage (line 181) | async requestPersistentStorage() { method onOnlineStatusChange (line 190) | onOnlineStatusChange(listener) { method isOnline (line 197) | isOnline() { method notifyOnlineStatus (line 203) | notifyOnlineStatus(online) { method openDatabase (line 209) | async openDatabase() { function generateServiceWorker (line 223) | function generateServiceWorker(options = {}) { function generateManifest (line 354) | function generateManifest(options = { name: 'edgeFlow.js App' }) { function getOfflineManager (line 377) | function getOfflineManager(config) { function initOffline (line 386) | async function initOffline(config) { function isOffline (line 394) | function isOffline() { function isPWASupported (line 400) | function isPWASupported() { FILE: dist/utils/preprocessor.d.ts type ImageInput (line 11) | type ImageInput = HTMLImageElement | HTMLCanvasElement | ImageBitmap | I... type AudioInput (line 15) | type AudioInput = AudioBuffer | Float32Array | ArrayBuffer | Blob | File... type ImagePreprocessorOptions (line 19) | interface ImagePreprocessorOptions { class ImagePreprocessor (line 61) | class ImagePreprocessor { type AudioPreprocessorOptions (line 124) | interface AudioPreprocessorOptions { class AudioPreprocessor (line 143) | class AudioPreprocessor { type TextPreprocessorOptions (line 201) | interface TextPreprocessorOptions { FILE: dist/utils/preprocessor.js constant DEFAULT_IMAGE_OPTIONS (line 11) | const DEFAULT_IMAGE_OPTIONS = { class ImagePreprocessor (line 32) | class ImagePreprocessor { method constructor (line 36) | constructor(options = {}) { method fromConfig (line 53) | static fromConfig(config) { method fromUrl (line 114) | static async fromUrl(url) { method fromHuggingFace (line 125) | static async fromHuggingFace(modelId, options) { method ensureCanvas (line 133) | ensureCanvas() { method process (line 147) | async process(input) { method processBatch (line 179) | async processBatch(inputs) { method loadFromUrl (line 202) | async loadFromUrl(url) { method loadFromBlob (line 218) | async loadFromBlob(blob) { method centerCrop (line 230) | centerCrop(imageData) { method toImageData (line 260) | toImageData(source) { method resize (line 271) | resize(imageData) { method toTensor (line 311) | toTensor(imageData) { method getOptions (line 373) | getOptions() { constant DEFAULT_AUDIO_OPTIONS (line 380) | const DEFAULT_AUDIO_OPTIONS = { class AudioPreprocessor (line 393) | class AudioPreprocessor { method constructor (line 396) | constructor(options = {}) { method fromConfig (line 402) | static fromConfig(config) { method fromHuggingFace (line 425) | static async fromHuggingFace(modelId, options) { method ensureAudioContext (line 438) | ensureAudioContext() { method process (line 451) | async process(input) { method processRaw (line 489) | async processRaw(input) { method loadFromUrl (line 520) | async loadFromUrl(url) { method loadFromBlob (line 531) | async loadFromBlob(blob) { method decodeAudioData (line 538) | async decodeAudioData(data) { method audioBufferToFloat32 (line 546) | audioBufferToFloat32(buffer) { method normalizeAudio (line 554) | normalizeAudio(data) { method computeMelSpectrogram (line 573) | computeMelSpectrogram(audio) { method dispose (line 604) | dispose() { function preprocessText (line 614) | function preprocessText(text, options = {}) { function createImagePreprocessor (line 637) | function createImagePreprocessor(preset = 'imagenet', options = {}) { function createAudioPreprocessor (line 664) | function createAudioPreprocessor(preset = 'whisper', options = {}) { FILE: dist/utils/tokenizer.d.ts type TokenizerModel (line 8) | type TokenizerModel = 'BPE' | 'WordPiece' | 'Unigram' | 'basic'; type TokenizerOptions (line 9) | interface TokenizerOptions { type HFTokenizerJSON (line 21) | interface HFTokenizerJSON { class Tokenizer (line 107) | class Tokenizer { FILE: dist/utils/tokenizer.js class Tokenizer (line 14) | class Tokenizer { method constructor (line 40) | constructor() { method initByteEncoder (line 46) | initByteEncoder() { method fromJSON (line 74) | static async fromJSON(json) { method fromUrl (line 156) | static async fromUrl(url) { method fromHuggingFace (line 167) | static async fromHuggingFace(modelId, options) { method normalize (line 175) | normalize(text) { method preTokenize (line 190) | preTokenize(text) { method textToBytes (line 199) | textToBytes(text) { method bytesToText (line 207) | bytesToText(text) { method getPairs (line 215) | getPairs(word) { method bpe (line 225) | bpe(token) { method wordPiece (line 280) | wordPiece(word) { method tokenizeWord (line 314) | tokenizeWord(word) { method unigramTokenize (line 337) | unigramTokenize(word) { method tokenize (line 370) | tokenize(text) { method convertTokensToIds (line 410) | convertTokensToIds(tokens) { method convertIdsToTokens (line 427) | convertIdsToTokens(ids) { method postProcess (line 433) | postProcess(ids, pairIds) { method encode (line 484) | encode(text, options = {}) { method encodeBatch (line 545) | encodeBatch(texts, options = {}) { method decode (line 557) | decode(ids, skipSpecialTokens = true) { method decodeBatch (line 580) | decodeBatch(batchIds, skipSpecialTokens = true) { method vocabSize (line 586) | get vocabSize() { method getSpecialTokenIds (line 592) | getSpecialTokenIds() { method getConfig (line 606) | getConfig() { method isSpecialToken (line 622) | isSpecialToken(token) { method getTokenId (line 628) | getTokenId(token) { method getToken (line 634) | getToken(id) { function createBasicTokenizer (line 644) | function createBasicTokenizer() { function loadTokenizer (line 651) | async function loadTokenizer(url) { function loadTokenizerFromHub (line 657) | async function loadTokenizerFromHub(modelId, options) { FILE: examples/basic-usage.ts function textGenerationExample (line 21) | async function textGenerationExample() { function schedulerExample (line 45) | async function schedulerExample() { function memoryExample (line 76) | async function memoryExample() { function main (line 96) | async function main() { FILE: examples/orchestration.ts function concurrentModelsExample (line 25) | async function concurrentModelsExample() { function cachingExample (line 57) | async function cachingExample() { function memoryScopeExample (line 89) | async function memoryScopeExample() { function cancellationExample (line 117) | async function cancellationExample() { function main (line 158) | async function main() { FILE: scripts/build-browser.js function build (line 12) | async function build() { FILE: src/backends/index.ts function registerAllBackends (line 41) | function registerAllBackends(): void { FILE: src/backends/onnx.ts function getOrt (line 28) | async function getOrt(): Promise { function isOnnxAvailable (line 46) | async function isOnnxAvailable(): Promise { type ONNXSessionData (line 54) | interface ONNXSessionData { class ONNXRuntime (line 69) | class ONNXRuntime implements Runtime { method capabilities (line 75) | get capabilities(): RuntimeCapabilities { method isAvailable (line 89) | async isAvailable(): Promise { method initialize (line 96) | async initialize(): Promise { method loadModel (line 123) | async loadModel( method run (line 208) | async run(model: LoadedModel, inputs: Tensor[]): Promise { method runNamed (line 274) | async runNamed(model: LoadedModel, namedInputs: Map): ... method unloadModel (line 336) | private async unloadModel(modelId: string): Promise { method dispose (line 347) | dispose(): void { function createONNXRuntime (line 357) | function createONNXRuntime(): Runtime { FILE: src/backends/transformers-adapter.ts type TransformersPipelineInstance (line 51) | interface TransformersPipelineInstance { type TransformersPipelineFactory (line 60) | type TransformersPipelineFactory = ( type TransformersAdapterOptions (line 69) | interface TransformersAdapterOptions { class TransformersAdapterRuntime (line 96) | class TransformersAdapterRuntime implements Runtime { method capabilities (line 99) | get capabilities(): RuntimeCapabilities { method isAvailable (line 110) | async isAvailable(): Promise { method initialize (line 114) | async initialize(): Promise { method loadModel (line 124) | async loadModel( method loadPipeline (line 161) | async loadPipeline( method run (line 189) | async run(model: LoadedModel, inputs: Tensor[]): Promise { method runDirect (line 214) | async runDirect( method dispose (line 229) | dispose(): void { function useTransformersBackend (line 260) | function useTransformersBackend(options: TransformersAdapterOptions): vo... function getTransformersAdapter (line 269) | function getTransformersAdapter(): TransformersAdapterRuntime | null { FILE: src/backends/wasm.ts type WASMModule (line 33) | interface WASMModule { type WASMExports (line 41) | interface WASMExports { type WASMModelData (line 70) | interface WASMModelData { type WASMModelConfig (line 82) | interface WASMModelConfig { type WASMLayerConfig (line 93) | interface WASMLayerConfig { class WASMRuntime (line 109) | class WASMRuntime implements Runtime { method capabilities (line 117) | get capabilities(): RuntimeCapabilities { method isAvailable (line 131) | async isAvailable(): Promise { method initialize (line 150) | async initialize(): Promise { method checkSIMDSupport (line 176) | private async checkSIMDSupport(): Promise { method createJSFallback (line 195) | private createJSFallback(memory: WebAssembly.Memory): WASMExports { method loadModel (line 303) | async loadModel( method run (line 360) | async run(model: LoadedModel, inputs: Tensor[]): Promise { method executeModel (line 370) | private async executeModel(inputs: Tensor[], metadata: ModelMetadata):... method parseModelConfig (line 413) | private parseModelConfig(data: ArrayBuffer): WASMModelConfig { method loadWeights (line 448) | private async loadWeights( method unloadModel (line 459) | private unloadModel(modelId: string): void { method ensureInitialized (line 473) | private ensureInitialized(): void { method hasSIMDSupport (line 485) | hasSIMDSupport(): boolean { method dispose (line 492) | dispose(): void { function createWASMRuntime (line 506) | function createWASMRuntime(): Runtime { FILE: src/backends/webgpu.ts type Navigator (line 34) | interface Navigator { type GPU (line 38) | interface GPU { type GPURequestAdapterOptions (line 42) | interface GPURequestAdapterOptions { type GPUAdapter (line 46) | interface GPUAdapter { type GPUDeviceDescriptor (line 50) | interface GPUDeviceDescriptor { type GPUDevice (line 55) | interface GPUDevice { type GPULimits (line 66) | interface GPULimits { type GPUDeviceLostInfo (line 70) | interface GPUDeviceLostInfo { type GPUBuffer (line 75) | interface GPUBuffer { type GPUShaderModule (line 79) | interface GPUShaderModule {} type GPUBindGroupLayout (line 80) | interface GPUBindGroupLayout {} type GPUPipelineLayout (line 81) | interface GPUPipelineLayout {} type GPUComputePipeline (line 82) | interface GPUComputePipeline {} type GPUBufferDescriptor (line 84) | interface GPUBufferDescriptor { type GPUShaderModuleDescriptor (line 89) | interface GPUShaderModuleDescriptor { type GPUBindGroupLayoutDescriptor (line 93) | interface GPUBindGroupLayoutDescriptor { type GPUBindGroupLayoutEntry (line 97) | interface GPUBindGroupLayoutEntry { type GPUPipelineLayoutDescriptor (line 103) | interface GPUPipelineLayoutDescriptor { type GPUComputePipelineDescriptor (line 107) | interface GPUComputePipelineDescriptor { type WebGPUModelData (line 135) | interface WebGPUModelData { type ModelConfig (line 151) | interface ModelConfig { type LayerConfig (line 162) | interface LayerConfig { class WebGPURuntime (line 177) | class WebGPURuntime implements Runtime { method capabilities (line 185) | get capabilities(): RuntimeCapabilities { method isAvailable (line 199) | async isAvailable(): Promise { method initialize (line 214) | async initialize(): Promise { method loadModel (line 255) | async loadModel( method run (line 318) | async run(model: LoadedModel, inputs: Tensor[]): Promise { method executeModel (line 329) | private async executeModel(inputs: Tensor[], metadata: ModelMetadata):... method parseModelData (line 378) | private parseModelData(data: ArrayBuffer): ModelConfig { method uploadWeights (line 410) | private async uploadWeights( method createPipelines (line 431) | private async createPipelines(modelData: WebGPUModelData): Promise { method initialize (line 170) | async initialize(): Promise { method loadModel (line 205) | async loadModel( method run (line 254) | async run(model: LoadedModel, inputs: Tensor[]): Promise { method executeModel (line 264) | private async executeModel(inputs: Tensor[], metadata: ModelMetadata):... method parseModelConfig (line 289) | private parseModelConfig(data: ArrayBuffer): WebNNModelConfig { method unloadModel (line 316) | private unloadModel(modelId: string): void { method ensureInitialized (line 323) | private ensureInitialized(): void { method getDeviceType (line 335) | getDeviceType(): MLContextType { method dispose (line 342) | dispose(): void { function createWebNNRuntime (line 352) | function createWebNNRuntime(): Runtime { FILE: src/core/composer.ts type CompositionStage (line 32) | interface CompositionStage { type CompositionResult (line 53) | interface CompositionResult { type ComposedPipeline (line 67) | interface ComposedPipeline { function compose (line 105) | function compose(stages: CompositionStage[]): ComposedPipeline { function parallel (line 194) | function parallel( FILE: src/core/device-profiler.ts type DeviceTier (line 28) | type DeviceTier = 'high' | 'medium' | 'low'; type DeviceProfile (line 33) | interface DeviceProfile { type ModelRecommendation (line 57) | interface ModelRecommendation { function getDeviceProfile (line 77) | async function getDeviceProfile(): Promise { function recommendQuantization (line 163) | function recommendQuantization(profile: DeviceProfile): QuantizationType { function recommendModelVariant (line 172) | async function recommendModelVariant(): Promise { function resetDeviceProfile (line 186) | function resetDeviceProfile(): void { FILE: src/core/memory.ts type TrackedResource (line 29) | interface TrackedResource { constant DEFAULT_POOL_CONFIG (line 40) | const DEFAULT_POOL_CONFIG: Required = { class MemoryManager (line 61) | class MemoryManager { method constructor (line 74) | private constructor(config: MemoryPoolConfig = {}) { method getInstance (line 81) | static getInstance(): MemoryManager { method configure (line 91) | static configure(config: MemoryPoolConfig): void { method track (line 101) | track(tensor: Tensor, disposer?: () => void): void { method trackModel (line 127) | trackModel(model: LoadedModel, disposer?: () => void): void { method untrack (line 153) | untrack(id: string): void { method release (line 165) | release(resourceOrId: Tensor | LoadedModel | string): void { method estimateTensorSize (line 183) | private estimateTensorSize(tensor: Tensor): number { method getBytesPerElement (line 191) | private getBytesPerElement(dtype: string): number { method captureStackTrace (line 213) | private captureStackTrace(): string | undefined { method checkMemoryThreshold (line 225) | private checkMemoryThreshold(): void { method gc (line 253) | gc(evict = false, maxAge = 5 * 60 * 1000): void { method measureBrowserMemory (line 282) | async measureBrowserMemory(): Promise<{ method getDeviceMemory (line 305) | getDeviceMemory(): number | null { method getStats (line 319) | getStats(): MemoryStats { method getResourceDetails (line 343) | getResourceDetails(): TrackedResource[] { method detectLeaks (line 350) | detectLeaks(maxAge: number = 10 * 60 * 1000): TrackedResource[] { method on (line 366) | on(event: EventType, listener: EventListener): void { method off (line 378) | off(event: EventType, listener: EventListener): void { method emit (line 388) | private emit(type: EventType, data: T): void { method resetStats (line 410) | resetStats(): void { method disposeAll (line 417) | disposeAll(): void { method dispose (line 426) | dispose(): void { class MemoryScope (line 452) | class MemoryScope { method constructor (line 457) | constructor(parent?: MemoryScope) { method track (line 467) | track void }>(resource: T): T { method createChild (line 475) | createChild(): MemoryScope { method keep (line 482) | keep void }>(resource: T): T { method dispose (line 493) | dispose(): void { function withMemoryScope (line 524) | async function withMemoryScope( function withMemoryScopeSync (line 538) | function withMemoryScopeSync( class ModelCache (line 556) | class ModelCache { method constructor (line 562) | constructor(options: { maxSize?: number; maxModels?: number } = {}) { method get (line 570) | get(key: string): LoadedModel | undefined { method set (line 582) | set(key: string, model: LoadedModel): void { method delete (line 605) | delete(key: string): boolean { method has (line 619) | has(key: string): boolean { method evictLRU (line 626) | private evictLRU(): void { method clear (line 645) | clear(): void { method getStats (line 656) | getStats(): { size: number; count: number; maxSize: number; maxModels:... function getMemoryManager (line 673) | function getMemoryManager(): MemoryManager { function getMemoryStats (line 680) | function getMemoryStats(): MemoryStats { function release (line 687) | function release(resource: Tensor | LoadedModel): void { function gc (line 694) | function gc(): void { FILE: src/core/plugin.ts type PluginPipelineEntry (line 34) | interface PluginPipelineEntry { type PluginBackendEntry (line 45) | interface PluginBackendEntry { type PluginMiddleware (line 55) | interface PluginMiddleware { type EdgeFlowPlugin (line 69) | interface EdgeFlowPlugin { function registerPlugin (line 95) | async function registerPlugin(plugin: EdgeFlowPlugin): Promise { function getPluginPipeline (line 132) | function getPluginPipeline(task: string): PluginPipelineEntry | undefined { function getPluginMiddleware (line 139) | function getPluginMiddleware(): ReadonlyArray { function listPlugins (line 146) | function listPlugins(): Array<{ name: string; version: string }> { function unregisterPlugin (line 156) | function unregisterPlugin(name: string): boolean { FILE: src/core/runtime.ts constant RUNTIME_PRIORITY (line 42) | const RUNTIME_PRIORITY: RuntimeType[] = ['webgpu', 'webnn', 'wasm']; class RuntimeManager (line 57) | class RuntimeManager { method constructor (line 63) | private constructor() {} method getInstance (line 68) | static getInstance(): RuntimeManager { method register (line 78) | register(type: RuntimeType, factory: () => Runtime): void { method getRuntime (line 85) | async getRuntime(type: RuntimeType = 'auto'): Promise { method getBestRuntime (line 138) | async getBestRuntime(): Promise { method detectAvailableRuntimes (line 176) | async detectAvailableRuntimes(): Promise> { method getCapabilities (line 200) | async getCapabilities(type: RuntimeType): Promise { method setDefaultRuntime (line 208) | setDefaultRuntime(type: RuntimeType): void { method getDefaultRuntimeType (line 215) | getDefaultRuntimeType(): RuntimeType { method disposeRuntime (line 222) | disposeRuntime(type: RuntimeType): void { method disposeAll (line 233) | disposeAll(): void { method on (line 243) | on(event: EventType, listener: EventListener): void { method off (line 255) | off(event: EventType, listener: EventListener): void { method emit (line 265) | private emit(type: EventType, data: T): void { function generateModelId (line 297) | function generateModelId(): string { class LoadedModelImpl (line 304) | class LoadedModelImpl implements LoadedModel { method constructor (line 312) | constructor( method isLoaded (line 323) | get isLoaded(): boolean { method dispose (line 327) | dispose(): void { function loadModel (line 344) | async function loadModel( function loadModelFromBuffer (line 380) | async function loadModelFromBuffer( function runInference (line 396) | async function runInference( function runInferenceNamed (line 421) | async function runInferenceNamed( function runBatchInference (line 457) | async function runBatchInference( function getRuntimeManager (line 481) | function getRuntimeManager(): RuntimeManager { function registerRuntime (line 488) | function registerRuntime(type: RuntimeType, factory: () => Runtime): void { function getBestRuntime (line 495) | async function getBestRuntime(): Promise { function getAvailableRuntimes (line 502) | async function getAvailableRuntimes(): Promise> { FILE: src/core/scheduler.ts class Task (line 27) | class Task implements InferenceTask { method constructor (line 45) | constructor( method status (line 58) | get status(): TaskStatus { method startedAt (line 62) | get startedAt(): number | undefined { method completedAt (line 66) | get completedAt(): number | undefined { method result (line 70) | get result(): T | undefined { method error (line 74) | get error(): Error | undefined { method cancel (line 81) | cancel(): void { method wait (line 103) | wait(): Promise { method execute (line 128) | async execute(): Promise { constant PRIORITY_ORDER (line 165) | const PRIORITY_ORDER: Record = { class PriorityQueue (line 175) | class PriorityQueue { method length (line 178) | get length(): number { method isEmpty (line 182) | isEmpty(): boolean { method enqueue (line 189) | enqueue(item: T): void { method dequeue (line 209) | dequeue(): T | undefined { method peek (line 216) | peek(): T | undefined { method remove (line 223) | remove(id: string): T | undefined { method getAll (line 235) | getAll(): T[] { method clear (line 242) | clear(): void { class BatchCollector (line 254) | class BatchCollector { method constructor (line 261) | constructor( method add (line 271) | add(task: Task): void { method flush (line 281) | flush(): void { method clear (line 294) | clear(): void { function generateTaskId (line 313) | function generateTaskId(): string { type CircuitState (line 320) | interface CircuitState { constant DEFAULT_OPTIONS (line 329) | const DEFAULT_OPTIONS: Required = { class InferenceScheduler (line 353) | class InferenceScheduler { method constructor (line 365) | constructor(options: SchedulerOptions = {}) { method getCircuit (line 372) | private getCircuit(modelId: string): CircuitState { method isCircuitOpen (line 384) | private isCircuitOpen(modelId: string): boolean { method circuitSuccess (line 401) | private circuitSuccess(modelId: string): void { method circuitFailure (line 411) | private circuitFailure(modelId: string): void { method getQueue (line 428) | private getQueue(modelId: string): PriorityQueue { method getRunningSet (line 440) | private getRunningSet(modelId: string): Set { method canStartTask (line 452) | private canStartTask(modelId: string): boolean { method processQueue (line 468) | private async processQueue(): Promise { method schedule (line 542) | schedule( method scheduleWithTimeout (line 616) | scheduleWithTimeout( method scheduleAll (line 650) | async scheduleAll( method getTask (line 667) | getTask(taskId: string): InferenceTask | undefined { method cancelTask (line 674) | cancelTask(taskId: string): boolean { method cancelAllForModel (line 692) | cancelAllForModel(modelId: string): number { method getStats (line 711) | getStats(): { method on (line 760) | on(event: EventType, listener: EventListener): void { method off (line 772) | off(event: EventType, listener: EventListener): void { method emit (line 782) | private emit(type: EventType, data: T): void { method clearHistory (line 804) | clearHistory(): void { method dispose (line 819) | dispose(): void { function getScheduler (line 852) | function getScheduler(): InferenceScheduler { function setScheduler (line 862) | function setScheduler(scheduler: InferenceScheduler): void { function configureScheduler (line 872) | function configureScheduler(options: SchedulerOptions): void { FILE: src/core/tensor.ts function generateTensorId (line 22) | function generateTensorId(): string { function getTypedArrayConstructor (line 29) | function getTypedArrayConstructor(dtype: DataType): new (length: number)... function calculateSize (line 57) | function calculateSize(shape: Shape): number { function validateShape (line 65) | function validateShape(shape: Shape): void { class EdgeFlowTensor (line 81) | class EdgeFlowTensor implements Tensor { method constructor (line 89) | constructor( method data (line 132) | get data(): TypedArray { method isDisposed (line 137) | get isDisposed(): boolean { method checkDisposed (line 144) | private checkDisposed(): void { method toFloat32Array (line 157) | toFloat32Array(): Float32Array { method toArray (line 174) | toArray(): number[] { method clone (line 191) | clone(): EdgeFlowTensor { method dispose (line 202) | dispose(): void { method get (line 213) | get(...indices: number[]): number { method set (line 249) | set(value: number, ...indices: number[]): void { method reshape (line 285) | reshape(newShape: Shape): EdgeFlowTensor { method transpose (line 305) | transpose(): EdgeFlowTensor { method toString (line 331) | toString(): string { function tensor (line 343) | function tensor( function zeros (line 375) | function zeros(shape: Shape, dtype: DataType = 'float32'): EdgeFlowTensor { function ones (line 385) | function ones(shape: Shape, dtype: DataType = 'float32'): EdgeFlowTensor { function full (line 396) | function full( function random (line 411) | function random(shape: Shape, dtype: DataType = 'float32'): EdgeFlowTens... function randn (line 423) | function randn(shape: Shape, dtype: DataType = 'float32'): EdgeFlowTensor { function arange (line 446) | function arange( function linspace (line 470) | function linspace( function eye (line 489) | function eye(n: number, dtype: DataType = 'float32'): EdgeFlowTensor { function add (line 506) | function add(a: EdgeFlowTensor, b: EdgeFlowTensor | number): EdgeFlowTen... function sub (line 538) | function sub(a: EdgeFlowTensor, b: EdgeFlowTensor | number): EdgeFlowTen... function mul (line 570) | function mul(a: EdgeFlowTensor, b: EdgeFlowTensor | number): EdgeFlowTen... function div (line 602) | function div(a: EdgeFlowTensor, b: EdgeFlowTensor | number): EdgeFlowTen... function matmul (line 634) | function matmul(a: EdgeFlowTensor, b: EdgeFlowTensor): EdgeFlowTensor { function softmax (line 674) | function softmax(t: EdgeFlowTensor, axis: number = -1): EdgeFlowTensor { function relu (line 743) | function relu(t: EdgeFlowTensor): EdgeFlowTensor { function sigmoid (line 757) | function sigmoid(t: EdgeFlowTensor): EdgeFlowTensor { function tanh (line 771) | function tanh(t: EdgeFlowTensor): EdgeFlowTensor { function sum (line 785) | function sum(t: EdgeFlowTensor, axis?: number): EdgeFlowTensor | number { function mean (line 852) | function mean(t: EdgeFlowTensor, axis?: number): EdgeFlowTensor | number { function argmax (line 869) | function argmax(t: EdgeFlowTensor, axis?: number): number | EdgeFlowTens... function concat (line 919) | function concat(tensors: EdgeFlowTensor[], axis: number = 0): EdgeFlowTe... FILE: src/core/types.ts type DataType (line 14) | type DataType = type TypedArray (line 26) | type TypedArray = type Shape (line 37) | type Shape = readonly number[]; type Tensor (line 42) | interface Tensor { type RuntimeType (line 72) | type RuntimeType = 'webgpu' | 'webnn' | 'wasm' | 'auto'; type RuntimeCapabilities (line 77) | interface RuntimeCapabilities { type Runtime (line 95) | interface Runtime { type ModelFormat (line 121) | type ModelFormat = 'onnx' | 'edgeflow' | 'safetensors'; type QuantizationType (line 126) | type QuantizationType = 'float32' | 'float16' | 'int8' | 'uint8' | 'int4'; type ModelMetadata (line 131) | interface ModelMetadata { type ModelIOSpec (line 159) | interface ModelIOSpec { type ModelLoadOptions (line 173) | interface ModelLoadOptions { type LoadedModel (line 187) | interface LoadedModel { type TaskPriority (line 207) | type TaskPriority = 'low' | 'normal' | 'high' | 'critical'; type TaskStatus (line 212) | type TaskStatus = 'pending' | 'running' | 'completed' | 'failed' | 'canc... type InferenceTask (line 217) | interface InferenceTask { type SchedulerOptions (line 245) | interface SchedulerOptions { type MemoryStats (line 277) | interface MemoryStats { type MemoryPoolConfig (line 293) | interface MemoryPoolConfig { type PipelineTask (line 313) | type PipelineTask = type PipelineConfig (line 337) | interface PipelineConfig { type PipelineOptions (line 357) | interface PipelineOptions { type TokenizerConfig (line 377) | interface TokenizerConfig { type TokenizedOutput (line 401) | interface TokenizedOutput { class EdgeFlowError (line 421) | class EdgeFlowError extends Error { method constructor (line 422) | constructor( type ErrorCode (line 471) | type ErrorCode = typeof ErrorCodes[keyof typeof ErrorCodes]; type EventType (line 480) | type EventType = type EdgeFlowEvent (line 495) | interface EdgeFlowEvent { type EventListener (line 504) | type EventListener = (event: EdgeFlowEvent) => void; FILE: src/core/worker.ts type WorkerMessageType (line 16) | type WorkerMessageType = type WorkerMessage (line 29) | interface WorkerMessage { type LoadModelRequest (line 38) | interface LoadModelRequest { type InferenceRequest (line 49) | interface InferenceRequest { type SerializedTensor (line 57) | interface SerializedTensor { type WorkerPoolOptions (line 66) | interface WorkerPoolOptions { function serializeTensor (line 80) | function serializeTensor(tensor: Tensor): SerializedTensor { function deserializeTensor (line 96) | async function deserializeTensor(serialized: SerializedTensor): Promise<... function deserializeTensorSync (line 106) | function deserializeTensorSync( type WorkerHealthState (line 118) | type WorkerHealthState = 'alive' | 'dead' | 'restarting'; constant MAX_RESTART_ATTEMPTS (line 120) | const MAX_RESTART_ATTEMPTS = 3; constant RESTART_BASE_DELAY_MS (line 121) | const RESTART_BASE_DELAY_MS = 1000; class InferenceWorker (line 126) | class InferenceWorker { method constructor (line 139) | constructor(workerUrl?: string) { method health (line 148) | get health(): WorkerHealthState { method initWorker (line 155) | private initWorker(workerUrl?: string): void { method handleCrash (line 177) | private handleCrash(): void { method attemptRestart (line 193) | private attemptRestart(): void { method restart (line 211) | restart(): void { method createWorkerBlob (line 235) | private createWorkerBlob(): string { method handleMessage (line 344) | private handleMessage(message: WorkerMessage): void { method sendRequest (line 367) | private async sendRequest(type: WorkerMessageType, payload?: unknow... method init (line 397) | async init(): Promise { method loadModel (line 406) | async loadModel(url: string, options?: { runtime?: RuntimeType; cache?... method runInference (line 415) | async runInference(modelId: string, inputs: Tensor[]): Promise { method terminate (line 434) | terminate(): void { class WorkerPool (line 451) | class WorkerPool { method constructor (line 457) | constructor(options: WorkerPoolOptions = {}) { method getNextHealthyWorker (line 470) | private getNextHealthyWorker(): InferenceWorker { method getWorkerForModel (line 486) | private getWorkerForModel(modelId: string): InferenceWorker { method replaceWorker (line 502) | replaceWorker(index: number): void { method init (line 512) | async init(): Promise { method loadModel (line 519) | async loadModel( method runInference (line 532) | async runInference(modelId: string, inputs: Tensor[]): Promise { method terminate (line 565) | terminate(): void { method size (line 576) | get size(): number { function getWorkerPool (line 590) | function getWorkerPool(options?: WorkerPoolOptions): WorkerPool { function runInWorker (line 600) | async function runInWorker( function isWorkerSupported (line 617) | function isWorkerSupported(): boolean { FILE: src/index.ts function isSupported (line 446) | async function isSupported(): Promise { function getBestRuntimeType (line 454) | async function getBestRuntimeType(): Promise { function preload (line 467) | async function preload( constant VERSION (line 489) | const VERSION = '0.1.0'; function getInfo (line 494) | async function getInfo(): Promise<{ FILE: src/pipelines/automatic-speech-recognition.ts constant DEFAULT_MODELS (line 19) | const DEFAULT_MODELS = { constant SOT_TOKEN (line 26) | const SOT_TOKEN = 50258; constant TRANSLATE_TOKEN (line 27) | const TRANSLATE_TOKEN = 50358; constant TRANSCRIBE_TOKEN (line 28) | const TRANSCRIBE_TOKEN = 50359; constant EOT_TOKEN (line 29) | const EOT_TOKEN = 50257; constant NO_TIMESTAMPS_TOKEN (line 30) | const NO_TIMESTAMPS_TOKEN = 50363; constant EN_TOKEN (line 31) | const EN_TOKEN = 50259; constant MAX_DECODER_TOKENS (line 33) | const MAX_DECODER_TOKENS = 448; type ASROptions (line 39) | interface ASROptions extends PipelineOptions { type WordTimestamp (line 48) | interface WordTimestamp { type ChunkTimestamp (line 55) | interface ChunkTimestamp { type ASRResult (line 61) | interface ASRResult extends PipelineResult { class AutomaticSpeechRecognitionPipeline (line 72) | class AutomaticSpeechRecognitionPipeline extends BasePipeline { method setTokenizer (line 118) | setTokenizer(tokenizer: Tokenizer): void { method run (line 122) | override async run( method transcribeSingle (line 141) | private async transcribeSingle(audio: AudioInput, options: ASROptions)... method buildInitialTokens (line 180) | private buildInitialTokens(task: 'transcribe' | 'translate', language?... method getLanguageToken (line 188) | private getLanguageToken(language: string): number { method autoregressiveDecode (line 203) | private async autoregressiveDecode( method extractTimestamps (line 246) | private extractTimestamps( method processLongAudio (line 276) | async processLongAudio( method preprocess (line 323) | protected async preprocess(input: AudioInput | AudioInput[]): Promise<... method postprocess (line 342) | protected async postprocess( method decodeOutput (line 367) | private decodeOutput(data: Float32Array, shape: readonly number[]): st... function createASRPipeline (line 405) | function createASRPipeline(config?: PipelineConfig): AutomaticSpeechReco... FILE: src/pipelines/base.ts type PipelineResult (line 25) | interface PipelineResult { type TextClassificationResult (line 33) | interface TextClassificationResult extends PipelineResult { type FeatureExtractionResult (line 41) | interface FeatureExtractionResult extends PipelineResult { type ImageClassificationResult (line 48) | interface ImageClassificationResult extends PipelineResult { type ObjectDetectionResult (line 56) | interface ObjectDetectionResult extends PipelineResult { method constructor (line 76) | constructor(config: PipelineConfig) { method initialize (line 90) | async initialize(): Promise { method loadModelWithCache (line 115) | protected async loadModelWithCache(modelPath: string): Promise { method runBatch (line 168) | async runBatch(inputs: TInput[], options?: PipelineOptions): Promise BasePipeline; function registerPipeline (line 236) | function registerPipeline(task: PipelineTask, factory: PipelineFactory):... function getPipelineFactory (line 243) | function getPipelineFactory(task: PipelineTask): PipelineFactory | undef... constant SENTIMENT_LABELS (line 254) | const SENTIMENT_LABELS = ['negative', 'positive']; constant EMOTION_LABELS (line 259) | const EMOTION_LABELS = [ constant IMAGENET_LABELS (line 266) | const IMAGENET_LABELS = [ FILE: src/pipelines/feature-extraction.ts constant DEFAULT_MODELS (line 26) | const DEFAULT_MODELS = { constant DEFAULT_EMBEDDING_DIM (line 31) | const DEFAULT_EMBEDDING_DIM = 384; type FeatureExtractionOptions (line 37) | interface FeatureExtractionOptions extends PipelineOptions { class FeatureExtractionPipeline (line 43) | class FeatureExtractionPipeline extends BasePipeline< method constructor (line 53) | constructor(config: PipelineConfig, embeddingDim: number = DEFAULT_EMB... method initialize (line 60) | override async initialize(): Promise { method run (line 73) | override async run( method preprocess (line 100) | protected override async preprocess(input: string | string[]): Promise... method runInference (line 130) | private async runInference(inputs: EdgeFlowTensor[]): Promise { method setLabels (line 81) | setLabels(labels: string[]): void { method run (line 85) | override async run( method preprocess (line 112) | protected override async preprocess(input: ImageInput | ImageInput[]):... method runModelInference (line 122) | private async runModelInference(inputs: EdgeFlowTensor[]): Promise { method loadEncoder (line 282) | async loadEncoder(modelUrl: string): Promise { method loadDecoder (line 291) | async loadDecoder(modelUrl: string): Promise { method setImage (line 300) | async setImage(image: ImageInput): Promise { method segment (line 335) | async segment(options: ImageSegmentationOptions = {}): Promise { method loadImageFromUrl (line 408) | private async loadImageFromUrl(url: string): Promise { method imageElementToImageData (line 428) | private imageElementToImageData(img: HTMLImageElement): ImageData { method canvasToImageData (line 440) | private canvasToImageData(canvas: HTMLCanvasElement): ImageData { method imageBitmapToImageData (line 448) | private imageBitmapToImageData(bitmap: ImageBitmap): ImageData { method preprocessImage (line 460) | private preprocessImage(imageData: ImageData): { method prepareDecoderInputs (line 524) | private prepareDecoderInputs( method postprocessMasks (line 589) | private postprocessMasks( method resizeMask (line 659) | private resizeMask( method clearImage (line 710) | clearImage(): void { method preprocess (line 720) | protected override async preprocess(input: ImageInput): Promise( function createPipelines (line 251) | async function createPipelines( FILE: src/pipelines/object-detection.ts type ObjectDetectionOptions (line 18) | interface ObjectDetectionOptions extends PipelineOptions { type BoundingBox (line 25) | interface BoundingBox { type Detection (line 32) | interface Detection extends ObjectDetectionResult { constant DEFAULT_MODELS (line 41) | const DEFAULT_MODELS = { constant COCO_LABELS (line 49) | const COCO_LABELS = [ class ObjectDetectionPipeline (line 68) | class ObjectDetectionPipeline extends BasePipeline { method setLabels (line 100) | setLabels(labels: string[]): void { method run (line 104) | override async run( method preprocess (line 114) | protected async preprocess(input: ImageInput | ImageInput[]): Promise<... method runModelInference (line 129) | private async runModelInference(inputs: EdgeFlowTensor[]): Promise { method setTokenizer (line 82) | setTokenizer(tokenizer: Tokenizer): void { method run (line 86) | override async run( method answerQuestion (line 100) | private async answerQuestion( method tokenOffsetToCharOffset (line 184) | private tokenOffsetToCharOffset( method preprocess (line 197) | protected async preprocess(input: QAInput | QAInput[]): Promise { method setLabels (line 75) | setLabels(labels: string[]): void { method run (line 79) | override async run( method preprocess (line 106) | protected override async preprocess(input: string | string[]): Promise... method runInference (line 130) | private async runInference(inputs: EdgeFlowTensor[]): Promise { method setTokenizer (line 300) | setTokenizer(tokenizer: Tokenizer): void { method preprocess (line 309) | protected async preprocess(input: string | string[]): Promise { method setTokenizer (line 83) | setTokenizer(tokenizer: Tokenizer): void { method classify (line 87) | async classify( method run (line 95) | override async run( method classifySingle (line 114) | private async classifySingle( method scoreHypothesis (line 160) | private async scoreHypothesis(premise: string, hypothesis: string): Pr... method preprocess (line 191) | protected async preprocess( method postprocess (line 211) | protected async postprocess( function createZeroShotClassificationPipeline (line 227) | function createZeroShotClassificationPipeline( FILE: src/tools/benchmark.ts type BenchmarkOptions (line 11) | interface BenchmarkOptions { type BenchmarkResult (line 28) | interface BenchmarkResult { type CompareBenchmarkResult (line 65) | interface CompareBenchmarkResult { function benchmark (line 80) | async function benchmark( function compareBenchmarks (line 174) | async function compareBenchmarks( function benchmarkSuite (line 213) | async function benchmarkSuite( function formatBenchmarkResult (line 230) | function formatBenchmarkResult(result: BenchmarkResult): string { function formatComparisonResult (line 251) | function formatComparisonResult(result: CompareBenchmarkResult): string { type MemoryBenchmarkResult (line 278) | interface MemoryBenchmarkResult { function benchmarkMemory (line 288) | async function benchmarkMemory( FILE: src/tools/debugger.ts type DebuggerConfig (line 16) | interface DebuggerConfig { type TensorInspection (line 39) | interface TensorInspection { type TensorStats (line 53) | interface TensorStats { type HistogramData (line 67) | interface HistogramData { type InferenceTrace (line 76) | interface InferenceTrace { type OperationTrace (line 90) | interface OperationTrace { type DebugEvent (line 102) | interface DebugEvent { type PerformanceMetrics (line 112) | interface PerformanceMetrics { function calculateTensorStats (line 131) | function calculateTensorStats(data: Float32Array | number[]): TensorStats { function createHistogram (line 189) | function createHistogram(data: Float32Array | number[], bins: number = 5... function inspectTensor (line 234) | function inspectTensor( function formatTensorInspection (line 275) | function formatTensorInspection(inspection: TensorInspection): string { function formatBytes (line 309) | function formatBytes(bytes: number): string { class EdgeFlowDebugger (line 323) | class EdgeFlowDebugger { method constructor (line 331) | constructor(config: DebuggerConfig = {}) { method defaultLogger (line 357) | private defaultLogger(level: string, message: string, data?: unknown):... method log (line 382) | log(level: string, message: string, data?: unknown): void { method addEvent (line 397) | private addEvent(event: DebugEvent): void { method enable (line 415) | enable(): void { method disable (line 423) | disable(): void { method on (line 430) | on(type: string, callback: (event: DebugEvent) => void): () => void { method inspectTensor (line 444) | inspectTensor(tensor: EdgeFlowTensor, name: string = 'tensor'): Tensor... method startTrace (line 475) | startTrace(modelId: string): string { method traceInput (line 499) | traceInput(traceId: string, tensor: EdgeFlowTensor, name: string): void { method traceOutput (line 509) | traceOutput(traceId: string, tensor: EdgeFlowTensor, name: string): vo... method traceOperation (line 519) | traceOperation(traceId: string, operation: OperationTrace): void { method endTrace (line 529) | endTrace(traceId: string): InferenceTrace | undefined { method recordAllocation (line 565) | recordAllocation(tensor: EdgeFlowTensor): void { method recordDeallocation (line 580) | recordDeallocation(tensor: EdgeFlowTensor): void { method getPerformanceMetrics (line 591) | getPerformanceMetrics(): PerformanceMetrics { method getEvents (line 598) | getEvents(): DebugEvent[] { method getTraces (line 605) | getTraces(): InferenceTrace[] { method getTrace (line 612) | getTrace(traceId: string): InferenceTrace | undefined { method clear (line 619) | clear(): void { method export (line 638) | export(): { method generateReport (line 655) | generateReport(): string { function getDebugger (line 702) | function getDebugger(config?: DebuggerConfig): EdgeFlowDebugger { function enableDebugging (line 712) | function enableDebugging(config?: DebuggerConfig): EdgeFlowDebugger { function disableDebugging (line 721) | function disableDebugging(): void { function createAsciiHistogram (line 732) | function createAsciiHistogram(histogram: HistogramData, width: number = ... function createTensorHeatmap (line 768) | function createTensorHeatmap(tensor: EdgeFlowTensor, width: number = 40)... function visualizeModelArchitecture (line 818) | function visualizeModelArchitecture( FILE: src/tools/index.ts type QuantizationOptions (line 19) | interface QuantizationOptions { type QuantizationResult (line 33) | interface QuantizationResult { function quantize (line 60) | async function quantize( function getModelData (line 112) | async function getModelData(_model: LoadedModel): Promise { function quantizeInt8 (line 120) | function quantizeInt8( function quantizeUint8 (line 151) | function quantizeUint8( function quantizeFloat16 (line 182) | function quantizeFloat16( function quantizeInt4 (line 204) | function quantizeInt4( function float32ToFloat16 (line 237) | function float32ToFloat16(value: number): number { type PruningOptions (line 279) | interface PruningOptions { type PruningResult (line 291) | interface PruningResult { function prune (line 305) | async function prune( type ModelAnalysis (line 346) | interface ModelAnalysis { function analyzeModel (line 372) | async function analyzeModel( type BenchmarkOptions (line 402) | interface BenchmarkOptions { type BenchmarkResult (line 414) | interface BenchmarkResult { function benchmark (line 432) | async function benchmark( function exportModel (line 584) | async function exportModel( FILE: src/tools/monitor.ts type MonitorConfig (line 14) | interface MonitorConfig { type PerformanceSample (line 37) | interface PerformanceSample { type InferenceMetrics (line 48) | interface InferenceMetrics { type MemoryMetrics (line 74) | interface MemoryMetrics { type SystemMetrics (line 97) | interface SystemMetrics { type AlertConfig (line 120) | interface AlertConfig { type AlertEvent (line 140) | interface AlertEvent { type WidgetData (line 149) | interface WidgetData { class PerformanceMonitor (line 162) | class PerformanceMonitor { method constructor (line 187) | constructor(config: MonitorConfig = {}) { method start (line 201) | start(): void { method stop (line 222) | stop(): void { method monitorFPS (line 239) | private monitorFPS(): void { method collectSample (line 258) | private collectSample(): void { method collectMemoryMetrics (line 329) | private collectMemoryMetrics(): MemoryMetrics { method collectSystemMetrics (line 354) | private collectSystemMetrics(): SystemMetrics { method estimateCPUUsage (line 385) | private estimateCPUUsage(): number { method checkAlerts (line 395) | private checkAlerts(sample: PerformanceSample): void { method getMetricValue (line 427) | private getMetricValue(sample: PerformanceSample, metric: string): num... method recordInference (line 445) | recordInference(duration: number): void { method updateQueueLength (line 453) | updateQueueLength(length: number): void { method updateActiveCount (line 460) | updateActiveCount(count: number): void { method updateTensorMemory (line 467) | updateTensorMemory(bytes: number): void { method updateCacheMemory (line 474) | updateCacheMemory(bytes: number): void { method addAlert (line 481) | addAlert(config: AlertConfig): void { method removeAlert (line 488) | removeAlert(metric: string): void { method onAlert (line 495) | onAlert(callback: (alert: AlertEvent) => void): () => void { method onSample (line 506) | onSample(callback: (sample: PerformanceSample) => void): () => void { method getCurrentSample (line 517) | getCurrentSample(): PerformanceSample | undefined { method getSamples (line 524) | getSamples(): PerformanceSample[] { method getSamplesInRange (line 531) | getSamplesInRange(startTime: number, endTime: number): PerformanceSamp... method getSummary (line 538) | getSummary(): { method clear (line 580) | clear(): void { method export (line 593) | export(): { function generateDashboardHTML (line 622) | function generateDashboardHTML(monitor: PerformanceMonitor): string { function generateChartPath (line 1004) | function generateChartPath(samples: PerformanceSample[]): string { function generateAsciiDashboard (line 1040) | function generateAsciiDashboard(monitor: PerformanceMonitor): string { function getMonitor (line 1113) | function getMonitor(config?: MonitorConfig): PerformanceMonitor { function startMonitoring (line 1123) | function startMonitoring(config?: MonitorConfig): PerformanceMonitor { function stopMonitoring (line 1132) | function stopMonitoring(): void { FILE: src/tools/quantization.ts type QuantizationType (line 17) | type QuantizationType = 'int8' | 'uint8' | 'int4' | 'float16' | 'dynamic'; type QuantizationOptions (line 22) | interface QuantizationOptions { type QuantizationProgress (line 48) | interface QuantizationProgress { type QuantizationResult (line 59) | interface QuantizationResult { type LayerQuantizationStats (line 88) | interface LayerQuantizationStats { type QuantizationStats (line 105) | interface QuantizationStats { type QuantizationParams (line 117) | interface QuantizationParams { function calculateQuantParams (line 131) | function calculateQuantParams( function quantizeToInt8 (line 213) | function quantizeToInt8( function quantizeToUint8 (line 248) | function quantizeToUint8( function quantizeToInt4 (line 283) | function quantizeToInt4( function quantizeToFloat16 (line 309) | function quantizeToFloat16(data: Float32Array): Uint16Array { function float32ToFloat16 (line 322) | function float32ToFloat16(value: number): number { function dequantizeInt8 (line 351) | function dequantizeInt8( function dequantizeUint8 (line 384) | function dequantizeUint8( function float16ToFloat32 (line 417) | function float16ToFloat32(value: number): number { function dequantizeFloat16 (line 441) | function dequantizeFloat16(data: Uint16Array): Float32Array { type ModelWeights (line 456) | interface ModelWeights { type QuantizedModel (line 466) | interface QuantizedModel { function parseModelWeights (line 485) | function parseModelWeights(modelData: ArrayBuffer): ModelWeights[] { function serializeQuantizedModel (line 509) | function serializeQuantizedModel(model: QuantizedModel): ArrayBuffer { function quantizeModel (line 623) | async function quantizeModel( function quantizeTensor (line 865) | function quantizeTensor( function dequantizeTensor (line 944) | function dequantizeTensor( type PruningOptions (line 1001) | interface PruningOptions { type PruningResult (line 1021) | interface PruningResult { function pruneTensor (line 1044) | function pruneTensor( function pruneModel (line 1099) | async function pruneModel( type ModelAnalysis (line 1145) | interface ModelAnalysis { function analyzeModel (line 1174) | async function analyzeModel(modelData: ArrayBuffer): Promise { type CacheOptions (line 31) | interface CacheOptions { type CacheStats (line 49) | interface CacheStats { class Cache (line 69) | class Cache { method constructor (line 76) | constructor(options: CacheOptions = {}) { method get (line 95) | get(key: string): T | undefined { method set (line 121) | set(key: string, value: T, size: number, ttl?: number): void { method has (line 161) | has(key: string): boolean { method delete (line 177) | delete(key: string): boolean { method clear (line 195) | clear(): void { method getStats (line 209) | getStats(): CacheStats { method evict (line 223) | private evict(): void { method findLRU (line 249) | private findLRU(): string | null { method findLFU (line 266) | private findLFU(): string | null { method findOldest (line 283) | private findOldest(): string | null { method findExpired (line 300) | private findExpired(): string | null { method loadFromStorage (line 315) | private async loadFromStorage(): Promise { method saveToStorage (line 343) | private async saveToStorage(): Promise { method clearStorage (line 371) | private async clearStorage(): Promise { method openDB (line 387) | private openDB(): Promise { class InferenceCache (line 411) | class InferenceCache extends Cache { method generateKey (line 415) | generateKey(modelId: string, input: Float32Array | number[]): string { method hashArray (line 425) | private hashArray(arr: number[]): string { class ModelDownloadCache (line 448) | class ModelDownloadCache { method constructor (line 452) | constructor(cacheName: string = 'edgeflow-models') { method ensureCache (line 459) | private async ensureCache(): Promise { method get (line 472) | async get(url: string): Promise { method put (line 484) | async put(url: string, response: Response): Promise { method delete (line 496) | async delete(url: string): Promise { method clear (line 508) | async clear(): Promise { method keys (line 520) | async keys(): Promise { function createCache (line 538) | function createCache( FILE: src/utils/hub.ts type HubOptions (line 18) | interface HubOptions { type HubDownloadProgress (line 38) | interface HubDownloadProgress { type ModelConfig (line 54) | interface ModelConfig { type ModelBundle (line 71) | interface ModelBundle { constant DEFAULT_ENDPOINT (line 92) | const DEFAULT_ENDPOINT = 'https://huggingface.co'; constant DEFAULT_REVISION (line 93) | const DEFAULT_REVISION = 'main'; constant ONNX_MODEL_FILES (line 98) | const ONNX_MODEL_FILES = [ function buildFileUrl (line 115) | function buildFileUrl( function fetchWithAuth (line 130) | async function fetchWithAuth(url: string, token?: string): Promise( function downloadTokenizer (line 238) | async function downloadTokenizer( function downloadConfig (line 249) | async function downloadConfig( function downloadModel (line 259) | async function downloadModel( function fromHub (line 371) | async function fromHub( function modelExists (line 381) | async function modelExists( function getModelInfo (line 397) | async function getModelInfo( constant POPULAR_MODELS (line 429) | const POPULAR_MODELS = { type PopularModelTask (line 477) | type PopularModelTask = keyof typeof POPULAR_MODELS; function getDefaultModel (line 482) | function getDefaultModel(task: PopularModelTask): string { function fromTask (line 494) | async function fromTask( FILE: src/utils/model-loader.ts type DownloadProgress (line 18) | interface DownloadProgress { type ModelLoaderOptions (line 38) | interface ModelLoaderOptions { type PreloadOptions (line 60) | interface PreloadOptions extends ModelLoaderOptions { type CachedModelMeta (line 68) | interface CachedModelMeta { type DownloadState (line 81) | interface DownloadState { type ChunkState (line 92) | interface ChunkState { constant DB_NAME (line 103) | const DB_NAME = 'edgeflow-model-cache'; constant DB_VERSION (line 104) | const DB_VERSION = 1; constant STORE_META (line 105) | const STORE_META = 'meta'; constant STORE_CHUNKS (line 106) | const STORE_CHUNKS = 'chunks'; constant STORE_STATE (line 107) | const STORE_STATE = 'download-state'; class ModelCache (line 112) | class ModelCache { method openDB (line 119) | private async openDB(): Promise { method getMeta (line 160) | async getMeta(url: string): Promise { method saveMeta (line 174) | async saveMeta(meta: CachedModelMeta): Promise { method saveChunk (line 194) | async saveChunk(url: string, index: number, data: ArrayBuffer): Promis... method putInStore (line 214) | private async putInStore(storeName: string, value: unknown): Promise { method getChunks (line 262) | async getChunks(url: string): Promise { method getModel (line 283) | async getModel(url: string): Promise { method saveDownloadState (line 306) | async saveDownloadState(state: DownloadState): Promise { method getDownloadState (line 321) | async getDownloadState(url: string): Promise { method deleteDownloadState (line 335) | async deleteDownloadState(url: string): Promise { method deleteModel (line 349) | async deleteModel(url: string): Promise { method clear (line 390) | async clear(): Promise { method getStats (line 408) | async getStats(): Promise<{ models: number; totalSize: number }> { function supportsRangeRequests (line 437) | async function supportsRangeRequests(url: string): Promise<{ supports: b... function downloadChunk (line 457) | async function downloadChunk( function downloadWithResume (line 485) | async function downloadWithResume( function downloadSimple (line 625) | async function downloadSimple( type PreloadTask (line 692) | interface PreloadTask { class PreloadManager (line 705) | class PreloadManager { method preload (line 714) | preload(url: string, options: PreloadOptions = {}): Promise { method downloadTask (line 784) | private async downloadTask(task: PreloadTask): Promise { method isPreloaded (line 798) | isPreloaded(url: string): boolean { method getStatus (line 806) | getStatus(url: string): 'pending' | 'loading' | 'complete' | 'error' |... method get (line 814) | async get(url: string): Promise { method cancel (line 828) | cancel(url: string): void { method clear (line 840) | clear(): void { function loadModelData (line 861) | async function loadModelData( function preloadModel (line 927) | function preloadModel(url: string, options: PreloadOptions = {}): Promis... function preloadModels (line 934) | function preloadModels( function isModelCached (line 946) | async function isModelCached(url: string): Promise { function getCachedModel (line 954) | async function getCachedModel(url: string): Promise { function deleteCachedModel (line 961) | async function deleteCachedModel(url: string): Promise { function clearModelCache (line 968) | async function clearModelCache(): Promise { function getModelCacheStats (line 975) | async function getModelCacheStats(): Promise<{ models: number; totalSize... function getPreloadStatus (line 982) | function getPreloadStatus(url: string): 'pending' | 'loading' | 'complet... function cancelPreload (line 989) | function cancelPreload(url: string): void { function getPreloadedModel (line 996) | async function getPreloadedModel(url: string): Promise { method registerServiceWorker (line 111) | private async registerServiceWorker(): Promise { method preloadForOffline (line 143) | async preloadForOffline(modelUrls: string[]): Promise { method getStatus (line 160) | async getStatus(): Promise { method getCachedModels (line 186) | async getCachedModels(): Promise { method isModelAvailableOffline (line 211) | async isModelAvailableOffline(url: string): Promise { method removeFromOffline (line 219) | async removeFromOffline(url: string): Promise { method clearOfflineData (line 227) | async clearOfflineData(): Promise { method getStorageInfo (line 235) | async getStorageInfo(): Promise<{ quota: number; usage: number; availa... method requestPersistentStorage (line 250) | async requestPersistentStorage(): Promise { method onOnlineStatusChange (line 260) | onOnlineStatusChange(listener: (online: boolean) => void): () => void { method isOnline (line 268) | isOnline(): boolean { method notifyOnlineStatus (line 275) | private notifyOnlineStatus(online: boolean): void { method openDatabase (line 282) | private async openDatabase(): Promise { function generateServiceWorker (line 298) | function generateServiceWorker(options: { function generateManifest (line 439) | function generateManifest(options: { function getOfflineManager (line 472) | function getOfflineManager(config?: OfflineConfig): OfflineManager { function initOffline (line 482) | async function initOffline(config?: OfflineConfig): Promise): ImagePreprocessor { method fromUrl (line 199) | static async fromUrl(url: string): Promise { method fromHuggingFace (line 211) | static async fromHuggingFace( method ensureCanvas (line 223) | private ensureCanvas(): void { method process (line 237) | async process(input: ImageInput): Promise { method processBatch (line 272) | async processBatch(inputs: ImageInput[]): Promise { method loadFromUrl (line 305) | private async loadFromUrl(url: string): Promise { method loadFromBlob (line 325) | private async loadFromBlob(blob: Blob): Promise { method centerCrop (line 337) | private centerCrop(imageData: ImageData): ImageData { method toImageData (line 373) | private toImageData( method resize (line 389) | private resize(imageData: ImageData): ImageData { method toTensor (line 437) | private toTensor(imageData: ImageData): EdgeFlowTensor { method getOptions (line 518) | getOptions(): ImagePreprocessorOptions { type AudioPreprocessorOptions (line 530) | interface AudioPreprocessorOptions { constant DEFAULT_AUDIO_OPTIONS (line 548) | const DEFAULT_AUDIO_OPTIONS: Required = { class AudioPreprocessor (line 562) | class AudioPreprocessor { method constructor (line 566) | constructor(options: AudioPreprocessorOptions = {}) { method fromConfig (line 573) | static fromConfig(config: Record): AudioPreprocessor { method fromHuggingFace (line 602) | static async fromHuggingFace( method ensureAudioContext (line 620) | private ensureAudioContext(): void { method process (line 633) | async process(input: AudioInput): Promise { method processRaw (line 674) | async processRaw(input: AudioInput): Promise { method loadFromUrl (line 706) | private async loadFromUrl(url: string): Promise { method loadFromBlob (line 719) | private async loadFromBlob(blob: Blob): Promise { method decodeAudioData (line 727) | private async decodeAudioData(data: ArrayBuffer): Promise { method audioBufferToFloat32 (line 736) | private audioBufferToFloat32(buffer: AudioBuffer): Float32Array { method normalizeAudio (line 745) | private normalizeAudio(data: Float32Array): Float32Array { method computeMelSpectrogram (line 766) | private computeMelSpectrogram(audio: Float32Array): EdgeFlowTensor { method dispose (line 806) | dispose(): void { type TextPreprocessorOptions (line 821) | interface TextPreprocessorOptions { function preprocessText (line 835) | function preprocessText( function createImagePreprocessor (line 874) | function createImagePreprocessor( function createAudioPreprocessor (line 906) | function createAudioPreprocessor( FILE: src/utils/tokenizer.ts type TokenizerModel (line 19) | type TokenizerModel = 'BPE' | 'WordPiece' | 'Unigram' | 'basic'; type TokenizerOptions (line 21) | interface TokenizerOptions { type HFTokenizerJSON (line 34) | interface HFTokenizerJSON { class Tokenizer (line 96) | class Tokenizer { method constructor (line 128) | constructor() { method initByteEncoder (line 135) | private initByteEncoder(): void { method fromJSON (line 165) | static async fromJSON(json: HFTokenizerJSON | string): Promise { method fromHuggingFace (line 265) | static async fromHuggingFace(modelId: string, options?: { revision?: s... method normalize (line 274) | private normalize(text: string): string { method preTokenize (line 294) | private preTokenize(text: string): string[] { method textToBytes (line 304) | private textToBytes(text: string): string { method bytesToText (line 313) | private bytesToText(text: string): string { method getPairs (line 324) | private getPairs(word: string[]): Set { method bpe (line 335) | private bpe(token: string): string[] { method wordPiece (line 401) | private wordPiece(word: string): string[] { method tokenizeWord (line 441) | private tokenizeWord(word: string): string[] { method unigramTokenize (line 466) | private unigramTokenize(word: string): string[] { method tokenize (line 502) | private tokenize(text: string): string[] { method convertTokensToIds (line 550) | private convertTokensToIds(tokens: string[]): number[] { method convertIdsToTokens (line 568) | private convertIdsToTokens(ids: number[]): string[] { method postProcess (line 575) | private postProcess( method encode (line 639) | encode(text: string, options: TokenizerOptions = {}): TokenizedOutput { method encodeBatch (line 717) | encodeBatch(texts: string[], options: TokenizerOptions = {}): Tokenize... method decode (line 731) | decode(ids: number[], skipSpecialTokens = true): string { method decodeBatch (line 779) | decodeBatch(batchIds: number[][], skipSpecialTokens = true): string[] { method vocabSize (line 786) | get vocabSize(): number { method getSpecialTokenIds (line 793) | getSpecialTokenIds(): { method getConfig (line 816) | getConfig(): TokenizerConfig { method isSpecialToken (line 833) | isSpecialToken(token: string): boolean { method getTokenId (line 840) | getTokenId(token: string): number | undefined { method getToken (line 847) | getToken(id: number): string | undefined { function createBasicTokenizer (line 859) | function createBasicTokenizer(): Tokenizer { function loadTokenizer (line 867) | async function loadTokenizer(url: string): Promise { function loadTokenizerFromHub (line 874) | async function loadTokenizerFromHub( FILE: tests/e2e/localai-10s-check.spec.ts constant BASE_URL (line 7) | const BASE_URL = 'http://localhost:5173/'; constant SCREENSHOT_DIR (line 8) | const SCREENSHOT_DIR = 'test-results/localai-10s-screenshots'; FILE: tests/e2e/localai-clear-cache-load.spec.ts constant BASE_URL (line 7) | const BASE_URL = 'http://localhost:5173/'; constant SCREENSHOT_DIR (line 8) | const SCREENSHOT_DIR = 'test-results/localai-clear-cache-screenshots'; FILE: tests/e2e/localai-knowledge-base.spec.ts constant BASE_URL (line 8) | const BASE_URL = 'http://localhost:5174/'; constant SCREENSHOT_DIR (line 9) | const SCREENSHOT_DIR = 'test-results/localai-screenshots'; FILE: tests/e2e/localai-load-models.spec.ts constant BASE_URL (line 8) | const BASE_URL = 'http://localhost:5173/'; constant SCREENSHOT_DIR (line 9) | const SCREENSHOT_DIR = 'test-results/localai-load-screenshots'; constant WAIT_SECONDS (line 10) | const WAIT_SECONDS = 150; constant SCREENSHOT_INTERVAL_SEC (line 11) | const SCREENSHOT_INTERVAL_SEC = 30; FILE: tests/e2e/localai-loading-check.spec.ts constant BASE_URL (line 7) | const BASE_URL = 'http://localhost:5173/'; constant SCREENSHOT_DIR (line 8) | const SCREENSHOT_DIR = 'test-results/localai-loading-screenshots'; FILE: tests/e2e/localai-network-audit.spec.ts constant BASE_URL (line 8) | const BASE_URL = 'http://localhost:5173/'; constant SCREENSHOT_DIR (line 9) | const SCREENSHOT_DIR = 'test-results/localai-network-audit'; type RequestRecord (line 11) | interface RequestRecord { FILE: tests/e2e/localai-network-failures.spec.ts constant BASE_URL (line 7) | const BASE_URL = 'http://localhost:5173/'; constant SCREENSHOT_DIR (line 8) | const SCREENSHOT_DIR = 'test-results/localai-network-screenshots'; FILE: tests/e2e/localai-network-full.spec.ts constant BASE_URL (line 7) | const BASE_URL = 'http://localhost:5173/'; constant SCREENSHOT_DIR (line 8) | const SCREENSHOT_DIR = 'test-results/localai-network-full-screenshots'; type NetworkEntry (line 10) | interface NetworkEntry { FILE: tests/integration/pipeline.test.ts function createMockOutputs (line 17) | function createMockOutputs(shapes: { shape: number[]; fill?: number }[])... FILE: tests/unit/memory.test.ts function createMockModel (line 333) | function createMockModel(id: string, sizeBytes: number) { FILE: tests/unit/tokenizer.test.ts constant SAMPLE_TOKENIZER_JSON (line 8) | const SAMPLE_TOKENIZER_JSON = {