SYMBOL INDEX (787 symbols across 233 files) FILE: evi/evi-dotnet-quickstart/EviTests.cs class EviTestFixture (line 18) | public class EviTestFixture : IAsyncLifetime method InitializeAsync (line 23) | public Task InitializeAsync() method DisposeAsync (line 43) | public Task DisposeAsync() class EviConnectionTests (line 49) | [Collection("EviTests")] method EviConnectionTests (line 55) | public EviConnectionTests(EviTestFixture fixture, ITestOutputHelper ou... method TestFixture_HasApiKey (line 61) | [Fact(DisplayName = "test fixture has API key")] method Connects_StartsChat_ReceivesChatId_StaysAlive (line 68) | [Fact(DisplayName = "connects w/ API key, starts a chat, receives a ch... method Connects_VerifiesSessionSettingsOnConnect (line 103) | [Fact(DisplayName = "connects w/ API key, verifies sessionSettings are... method Connects_VerifiesSessionSettingsUpdatedAfterConnect (line 189) | [Fact(DisplayName = "connects w/ API key, verifies sessionSettings can... class EviTestCollection (line 260) | [CollectionDefinition("EviTests")] FILE: evi/evi-flutter/audio/lib/audio.dart class Audio (line 8) | class Audio { method startRecording (line 43) | Future startRecording() method stopRecording (line 62) | Future stopRecording() method mute (line 70) | Future mute() method unmute (line 78) | Future unmute() method enqueueAudio (line 86) | Future enqueueAudio(String base64String) method stopPlayback (line 95) | Future stopPlayback() method dispose (line 103) | Future dispose() FILE: evi/evi-flutter/audio/lib/audio_method_channel.dart class MethodChannelAudio (line 7) | class MethodChannelAudio extends AudioPlatform { method getPlatformVersion (line 13) | Future getPlatformVersion() FILE: evi/evi-flutter/audio/lib/audio_platform_interface.dart class AudioPlatform (line 5) | abstract class AudioPlatform extends PlatformInterface { method getPlatformVersion (line 26) | Future getPlatformVersion() FILE: evi/evi-flutter/audio/lib/dart_audio.dart class DartAudio (line 7) | class DartAudio { method enqueueAudioSegment (line 50) | void enqueueAudioSegment(String base64Bytes) method stopPlayback (line 59) | void stopPlayback() method _playNextAudioSegment (line 64) | void _playNextAudioSegment() method startRecording (line 77) | Future>> startRecording() method stopRecording (line 129) | Future stopRecording() method mute (line 139) | Future mute() method unmute (line 143) | Future unmute() method dispose (line 154) | Future dispose() FILE: evi/evi-flutter/audio/test/audio_method_channel_test.dart function main (line 5) | void main() FILE: evi/evi-flutter/audio/test/audio_test.dart class MockAudioPlatform (line 7) | class MockAudioPlatform method getPlatformVersion (line 12) | Future getPlatformVersion() function main (line 15) | void main() FILE: evi/evi-flutter/lib/chat_card.dart type Role (line 6) | enum Role { user, assistant } class Score (line 8) | class Score { method toJson (line 14) | Map toJson() class ChatEntry (line 22) | class ChatEntry { class ChatCard (line 35) | class ChatCard extends StatelessWidget { method build (line 40) | Widget build(BuildContext context) class ChatDisplay (line 86) | class ChatDisplay extends StatelessWidget { method build (line 91) | Widget build(BuildContext context) FILE: evi/evi-flutter/lib/evi_message.dart class EviMessage (line 11) | sealed class EviMessage { class ErrorMessage (line 39) | class ErrorMessage extends EviMessage { class ChatMetadataMessage (line 46) | class ChatMetadataMessage extends EviMessage { class AudioOutputMessage (line 50) | class AudioOutputMessage extends EviMessage { class UserInterruptionMessage (line 57) | class UserInterruptionMessage extends EviMessage { class ChatMessage (line 61) | class ChatMessage { class ProsodyInference (line 69) | class ProsodyInference { class Inference (line 74) | class Inference { class AssistantMessage (line 79) | class AssistantMessage extends EviMessage { class UserMessage (line 88) | class UserMessage extends EviMessage { class UnknownMessage (line 97) | class UnknownMessage extends EviMessage { FILE: evi/evi-flutter/lib/main.dart class ConfigManager (line 13) | class ConfigManager { method fetchHumeApiKey (line 25) | String fetchHumeApiKey() method fetchAccessToken (line 29) | Future fetchAccessToken() method loadConfig (line 44) | Future loadConfig() function main (line 58) | void main() class MyApp (line 68) | class MyApp extends StatelessWidget { method build (line 72) | Widget build(BuildContext context) method extractTopThreeEmotions (line 90) | List extractTopThreeEmotions(evi.Inference models) class ErrorMessage (line 109) | class ErrorMessage extends StatelessWidget { method build (line 115) | Widget build(BuildContext context) class MyHomePage (line 125) | class MyHomePage extends StatefulWidget { method createState (line 131) | State createState() class _MyHomePageState (line 134) | class _MyHomePageState extends State { method appendNewChatMessage (line 146) | void appendNewChatMessage(evi.ChatMessage chatMessage, evi.Inference m... method build (line 159) | Widget build(BuildContext context) method dispose (line 205) | void dispose() method _connect (line 212) | void _connect() method _disconnect (line 285) | void _disconnect() method _handleConnectionClosed (line 293) | void _handleConnectionClosed() method _handleInterruption (line 300) | void _handleInterruption() method _muteInput (line 304) | void _muteInput() method _prepareAudioSettings (line 311) | void _prepareAudioSettings() method _sendAudio (line 324) | void _sendAudio(String base64) method _startRecording (line 331) | void _startRecording() method _stopRecording (line 342) | void _stopRecording() method _unmuteInput (line 346) | void _unmuteInput() FILE: evi/evi-flutter/test/widget_test.dart function main (line 13) | void main() FILE: evi/evi-next-js-app-router-quickstart/app/actions/set-llm-key.ts function setLlmKeyForChat (line 9) | async function setLlmKeyForChat(chatId: string) { FILE: evi/evi-next-js-app-router-quickstart/app/api-key/page.tsx function ApiKeyPage (line 6) | async function ApiKeyPage() { FILE: evi/evi-next-js-app-router-quickstart/app/error.tsx function Error (line 3) | function Error() { FILE: evi/evi-next-js-app-router-quickstart/app/layout.tsx function RootLayout (line 13) | function RootLayout({ FILE: evi/evi-next-js-app-router-quickstart/app/page.tsx function Page (line 7) | async function Page() { FILE: evi/evi-next-js-app-router-quickstart/app/session-settings/page.tsx function SessionSettingsPage (line 8) | async function SessionSettingsPage() { FILE: evi/evi-next-js-app-router-quickstart/components/Chat.tsx type ChatProps (line 12) | type ChatProps = ( function ClientComponent (line 19) | function ClientComponent({ FILE: evi/evi-next-js-app-router-quickstart/components/ChatLoader.tsx type ChatLoaderProps (line 10) | type ChatLoaderProps = ( function ChatLoader (line 17) | function ChatLoader({ FILE: evi/evi-next-js-app-router-quickstart/components/Controls.tsx constant E2E_ENABLED (line 12) | const E2E_ENABLED = function Controls (line 16) | function Controls() { FILE: evi/evi-next-js-app-router-quickstart/components/Expressions.tsx function Expressions (line 8) | function Expressions({ FILE: evi/evi-next-js-app-router-quickstart/components/MicFFT.tsx function MicFFT (line 7) | function MicFFT({ FILE: evi/evi-next-js-app-router-quickstart/components/StartCall.tsx type StartCallProps (line 7) | type StartCallProps = ( function StartCall (line 14) | function StartCall({ FILE: evi/evi-next-js-app-router-quickstart/components/logos/Hume.tsx type HumeLogoProps (line 4) | type HumeLogoProps = SVGAttributes; function HumeLogo (line 6) | function HumeLogo(props: HumeLogoProps) { FILE: evi/evi-next-js-app-router-quickstart/components/ui/button.tsx type ButtonProps (line 36) | interface ButtonProps FILE: evi/evi-next-js-app-router-quickstart/tests/voice-react.spec.ts function waitForChatMetadataFromPage (line 217) | async function waitForChatMetadataFromPage(page: Page, timeoutMs = 30_00... function fetchChatEvents (line 235) | async function fetchChatEvents( FILE: evi/evi-next-js-app-router-quickstart/utils/e2e-hooks.ts constant ENABLED (line 1) | const ENABLED = type Window (line 6) | interface Window { function getWindow (line 13) | function getWindow(): Window | null { function recordVoiceEvent (line 20) | function recordVoiceEvent(event: unknown) { function trackVoiceStatus (line 28) | function trackVoiceStatus(status: string) { FILE: evi/evi-next-js-app-router-quickstart/utils/index.ts function cn (line 4) | function cn(...inputs: ClassValue[]) { FILE: evi/evi-next-js-app-router-quickstart/utils/session-settings.ts constant E2E_SESSION_SETTINGS (line 3) | const E2E_SESSION_SETTINGS = { FILE: evi/evi-next-js-function-calling/app/api/fetchWeather/route.ts function POST (line 4) | async function POST(request: Request) { FILE: evi/evi-next-js-function-calling/app/error.tsx function Error (line 3) | function Error() { FILE: evi/evi-next-js-function-calling/app/layout.tsx function RootLayout (line 13) | function RootLayout({ FILE: evi/evi-next-js-function-calling/app/page.tsx function Page (line 7) | async function Page() { FILE: evi/evi-next-js-function-calling/components/Chat.tsx type ToolMeta (line 9) | type ToolMeta = { function ClientComponent (line 59) | function ClientComponent({ FILE: evi/evi-next-js-function-calling/components/ChatLoader.tsx function ChatLoader (line 9) | function ChatLoader({ accessToken }: { accessToken: string }) { FILE: evi/evi-next-js-function-calling/components/Controls.tsx function Controls (line 10) | function Controls() { FILE: evi/evi-next-js-function-calling/components/Expressions.tsx function Expressions (line 8) | function Expressions({ FILE: evi/evi-next-js-function-calling/components/MicFFT.tsx function MicFFT (line 7) | function MicFFT({ FILE: evi/evi-next-js-function-calling/components/StartCall.tsx function StartCall (line 6) | function StartCall({ accessToken }: { accessToken: string }) { FILE: evi/evi-next-js-function-calling/components/logos/Hume.tsx type HumeLogoProps (line 4) | type HumeLogoProps = SVGAttributes; function HumeLogo (line 6) | function HumeLogo(props: HumeLogoProps) { FILE: evi/evi-next-js-function-calling/components/ui/button.tsx type ButtonProps (line 36) | interface ButtonProps FILE: evi/evi-next-js-function-calling/utils/index.ts function cn (line 4) | function cn(...inputs: ClassValue[]) { FILE: evi/evi-next-js-pages-router-quickstart/components/Chat.tsx function ClientComponent (line 9) | function ClientComponent({ FILE: evi/evi-next-js-pages-router-quickstart/components/Controls.tsx function Controls (line 10) | function Controls() { FILE: evi/evi-next-js-pages-router-quickstart/components/Expressions.tsx function Expressions (line 8) | function Expressions({ FILE: evi/evi-next-js-pages-router-quickstart/components/MicFFT.tsx function MicFFT (line 7) | function MicFFT({ FILE: evi/evi-next-js-pages-router-quickstart/components/StartCall.tsx function StartCall (line 6) | function StartCall({ accessToken }: { accessToken: string }) { FILE: evi/evi-next-js-pages-router-quickstart/components/logos/Hume.tsx type HumeLogoProps (line 4) | type HumeLogoProps = SVGAttributes; function HumeLogo (line 6) | function HumeLogo(props: HumeLogoProps) { FILE: evi/evi-next-js-pages-router-quickstart/components/ui/button.tsx type ButtonProps (line 36) | interface ButtonProps FILE: evi/evi-next-js-pages-router-quickstart/pages/500.tsx function ErrorPage (line 1) | function ErrorPage() { FILE: evi/evi-next-js-pages-router-quickstart/pages/_app.tsx function App (line 8) | function App({ Component, pageProps }: AppProps) { FILE: evi/evi-next-js-pages-router-quickstart/pages/_document.tsx function Document (line 4) | function Document() { FILE: evi/evi-next-js-pages-router-quickstart/pages/api/control-plane/set-llm-key.ts function handler (line 6) | async function handler( FILE: evi/evi-next-js-pages-router-quickstart/pages/index.tsx type PageProps (line 33) | type PageProps = InferGetServerSidePropsType; function Page (line 35) | function Page({ accessToken }: PageProps) { FILE: evi/evi-next-js-pages-router-quickstart/utils/index.ts function cn (line 4) | function cn(...inputs: ClassValue[]) { FILE: evi/evi-python-chat-history/main.py class EmotionScore (line 12) | class EmotionScore(TypedDict): function main (line 16) | async def main() -> None: function fetch_all_chat_events (line 44) | async def fetch_all_chat_events(chat_id: str) -> list[ReturnChatEvent]: function generate_transcript (line 66) | def generate_transcript(chat_events: list[ReturnChatEvent]) -> str: function get_top_emotions (line 88) | def get_top_emotions(chat_events: list[ReturnChatEvent]) -> dict[str, fl... FILE: evi/evi-python-clm-sse/openai_sse.py function stream_messages_from_openai (line 14) | async def stream_messages_from_openai( function verify_token (line 39) | async def verify_token(credentials: HTTPAuthorizationCredentials = Secur... function root (line 46) | async def root( FILE: evi/evi-python-clm-wss/main.py class ProsodyModel (line 12) | class ProsodyModel(TypedDict): class Models (line 16) | class Models(TypedDict): class MessageContent (line 20) | class MessageContent(TypedDict): class HumeMessage (line 25) | class HumeMessage(TypedDict): class MessagesPayload (line 30) | class MessagesPayload(TypedDict): class ChatHistoryItem (line 34) | class ChatHistoryItem(TypedDict): class Agent (line 38) | class Agent: method __init__ (line 49) | def __init__(self): method _extract_prosody_scores (line 56) | def _extract_prosody_scores(self, message: HumeMessage) -> ProsodyScores: method _get_top_prosody_scores (line 65) | def _get_top_prosody_scores(self, prosody_scores: ProsodyScores, count... method _prosody_report (line 69) | def _prosody_report(self, prosody_scores: ProsodyScores) -> str: method _count_messages_by_role (line 76) | def _count_messages_by_role(self, chat_history: List[ChatHistoryItem])... method parse_hume_payload (line 81) | def parse_hume_payload(self, messages_payload: MessagesPayload) -> Tup... method add_prosody_to_utterance (line 120) | def add_prosody_to_utterance(self, content: str, prosody_scores: Proso... method _generate_eliza_response (line 126) | def _generate_eliza_response(self) -> str: method _should_send_congratulations (line 129) | def _should_send_congratulations(self, user_count: int, assistant_coun... method respond (line 132) | def respond(self, message: str, chat_history: List[ChatHistoryItem], l... function websocket_endpoint (line 152) | async def websocket_endpoint(websocket: WebSocket): FILE: evi/evi-python-control-plane/main.py function load_config (line 32) | def load_config() -> tuple[str, str]: function send_control_message (line 47) | async def send_control_message(client: AsyncHumeClient, chat_id: str, me... function observe_chat (line 75) | async def observe_chat(api_key: str, chat_id: str, on_message_callback) ... function observer_message_handler (line 123) | async def observer_message_handler(message: dict) -> None: function control_plane_demo (line 166) | async def control_plane_demo( function main_new_chat (line 241) | async def main_new_chat() -> None: function find_active_chat (line 309) | def find_active_chat(client: HumeClient, config_id: str): function main_existing_chat (line 326) | async def main_existing_chat() -> None: function main (line 350) | async def main() -> None: FILE: evi/evi-python-function-calling/main.py class WebSocketHandler (line 15) | class WebSocketHandler: method __init__ (line 18) | def __init__(self): method set_socket (line 23) | def set_socket(self, socket: AsyncChatSocketClient): method handle_tool_call (line 35) | async def handle_tool_call(self, message: ToolCallMessage) -> Union[To... method on_open (line 94) | async def on_open(self): method on_message (line 98) | async def on_message(self, message: SubscribeEvent): method on_close (line 156) | async def on_close(self): method on_error (line 160) | async def on_error(self, error): function fetch_weather (line 170) | async def fetch_weather(location: str, format: str) -> str: function sending_handler (line 277) | async def sending_handler(socket: AsyncChatSocketClient): function main (line 297) | async def main() -> None: FILE: evi/evi-python-function-calling/utils.py function print_prompt (line 3) | def print_prompt(text: str) -> None: function extract_top_n_emotions (line 9) | def extract_top_n_emotions(emotion_scores: dict, n: int) -> dict: function print_emotion_scores (line 14) | def print_emotion_scores(emotion_scores: dict) -> None: FILE: evi/evi-python-phone-calling-proxy-server/app.py function serve_homepage (line 25) | def serve_homepage(): function twiml_response (line 30) | def twiml_response(): function media_stream (line 54) | def media_stream(ws): function handle_media_stream (line 63) | async def handle_media_stream(ws): FILE: evi/evi-python-phone-calling-proxy-server/audio_processors/evi_audio_processor.py class AudioProcessingConfig (line 15) | class AudioProcessingConfig: class EviAudioProcessor (line 27) | class EviAudioProcessor: method __init__ (line 32) | def __init__( method postprocess_audio (line 39) | def postprocess_audio(self, evi_audio: bytes) -> bytes: method _read_audio (line 57) | def _read_audio(self, evi_audio: bytes) -> tuple[np.ndarray, int]: method _ensure_float (line 67) | def _ensure_float(self, audio: np.ndarray) -> np.ndarray: method _resample_audio (line 73) | def _resample_audio(self, audio: np.ndarray, original_fs: int, target_... method _apply_filters (line 86) | def _apply_filters(self, audio: np.ndarray, fs: int) -> np.ndarray: method _high_pass_filter (line 92) | def _high_pass_filter(self, audio: np.ndarray, fs: int) -> np.ndarray: method _peak_filter (line 102) | def _peak_filter(self, audio: np.ndarray, fs: int) -> np.ndarray: method _notch_filter (line 117) | def _notch_filter(self, audio: np.ndarray, fs: int) -> np.ndarray: method _normalize_audio (line 125) | def _normalize_audio(self, audio: np.ndarray) -> np.ndarray: FILE: evi/evi-python-phone-calling-proxy-server/audio_processors/twilio_audio_processor.py class TwilioAudioProcessor (line 12) | class TwilioAudioProcessor: method __init__ (line 24) | def __init__(self) -> None: method fill_silence (line 29) | def fill_silence(self, current_timestamp: int) -> None: method buffer_inbound_audio (line 42) | def buffer_inbound_audio(self, twilio_media_payload: Dict[str, Any]) -... method queue_twilio_audio (line 49) | async def queue_twilio_audio(self, twilio_media_payload: Dict[str, Any... FILE: evi/evi-python-phone-calling-proxy-server/tools.py function supportAssistant (line 5) | async def supportAssistant(ticket_id: str) -> str: FILE: evi/evi-python-quickstart/conftest.py function pytest_collection_modifyitems (line 4) | def pytest_collection_modifyitems(config, items): FILE: evi/evi-python-quickstart/quickstart.py function extract_top_n_emotions (line 11) | def extract_top_n_emotions(emotion_scores: dict, n: int) -> dict: function print_emotions (line 18) | def print_emotions(emotion_scores: dict) -> None: function log (line 22) | def log(text: str) -> None: function on_message (line 27) | async def on_message(message: SubscribeEvent, stream: Stream) -> None: function main (line 41) | async def main() -> None: FILE: evi/evi-python-quickstart/test_quickstart.py function api_key (line 21) | def api_key(): function hume_client (line 29) | def hume_client(api_key): function hume_client_sync (line 34) | def hume_client_sync(api_key): function test_connect_to_evi (line 39) | async def test_connect_to_evi(hume_client): function test_session_settings_on_connect (line 84) | async def test_session_settings_on_connect(hume_client, hume_client_sync): function test_session_settings_upd_after_connect (line 152) | async def test_session_settings_upd_after_connect(hume_client, hume_clie... FILE: evi/evi-python-raw-api/src/authenticator.py class Authenticator (line 7) | class Authenticator: method __init__ (line 17) | def __init__(self, api_key: str, secret_key: str, host: str = "test-ap... method fetch_access_token (line 30) | def fetch_access_token(self) -> str: FILE: evi/evi-python-raw-api/src/connection.py class Connection (line 25) | class Connection: method connect (line 31) | async def connect( method _receive_audio_data (line 84) | async def _receive_audio_data(cls, socket): method _read_audio_stream_non_blocking (line 122) | async def _read_audio_stream_non_blocking(cls, audio_stream, chunk_size): method _send_audio_data (line 140) | async def _send_audio_data( FILE: evi/evi-python-raw-api/src/devices.py class AudioDevices (line 6) | class AudioDevices: method list_audio_devices (line 12) | def list_audio_devices( method choose_device (line 46) | def choose_device(cls, devices, device_type="input"): FILE: evi/evi-python-raw-api/src/main.py function main (line 19) | async def main(): function get_access_token (line 74) | def get_access_token() -> str: FILE: evi/evi-python-webhooks/app.py function hume_webhook_handler (line 25) | async def hume_webhook_handler(request: Request, event: WebhookEvent): FILE: evi/evi-python-webhooks/utils.py function fetch_all_chat_events (line 14) | async def fetch_all_chat_events(client: AsyncHumeClient, chat_id: str) -... function construct_transcript (line 23) | def construct_transcript(chat_events: list[ReturnChatEvent]) -> str: function save_transcript_to_file (line 38) | def save_transcript_to_file(transcript: str, chat_id: str) -> None: function get_chat_transcript (line 46) | async def get_chat_transcript(client: AsyncHumeClient, chat_id: str) -> ... function validate_webhook_headers (line 53) | def validate_webhook_headers(payload: str, headers: Headers) -> None: function fetch_weather (line 101) | async def fetch_weather(parameters: str) -> str: function fetch_weather_tool (line 183) | async def fetch_weather_tool( FILE: evi/evi-python-wss-clm-endpoint/agent.py function reflect (line 323) | def reflect(fragment): function eliza_response (line 334) | def eliza_response(user_input): FILE: evi/evi-python-wss-clm-endpoint/app.py function root (line 12) | async def root(): function websocket_handler (line 17) | async def websocket_handler(websocket: WebSocket) -> None: FILE: evi/evi-python-wss-clm-endpoint/cdk/cdk.out/asset.689e46f5ffafa1e0f81f114b5dfd7694d2d1e291d9bd855e4f7b601d2b2403d0/agent.py function reflect (line 323) | def reflect(fragment): function eliza_response (line 334) | def eliza_response(user_input): FILE: evi/evi-python-wss-clm-endpoint/cdk/cdk.out/asset.689e46f5ffafa1e0f81f114b5dfd7694d2d1e291d9bd855e4f7b601d2b2403d0/app.py function root (line 10) | async def root(): function websocket_handler (line 15) | async def websocket_handler(websocket: WebSocket) -> None: FILE: evi/evi-python-wss-clm-endpoint/cdk/cdk.out/asset.689e46f5ffafa1e0f81f114b5dfd7694d2d1e291d9bd855e4f7b601d2b2403d0/modal/modal_app.py function endpoint (line 14) | def endpoint(): FILE: evi/evi-python-wss-clm-endpoint/cdk/cdk.out/asset.ee7de53d64cc9d6248fa6aa550f92358f6c907b5efd6f3298aeab1b5e7ea358a/__entrypoint__.js function handler (line 1) | async function handler(event,context){const sanitizedEvent={...event,Res... function renderResponse (line 1) | function renderResponse(cfnRequest,handlerResponse={}){const physicalRes... function submitResponse (line 1) | async function submitResponse(status,event){const json={Status:status,Re... function defaultSendHttpRequest (line 1) | async function defaultSendHttpRequest(options,requestBody){return new Pr... function defaultLog (line 1) | function defaultLog(fmt,...params){console.log(fmt,...params)} function withRetries (line 1) | function withRetries(options,fn){return async(...xs)=>{let attempts=opti... function sleep (line 1) | async function sleep(ms){return new Promise(ok=>setTimeout(ok,ms))} FILE: evi/evi-python-wss-clm-endpoint/cdk/cdk.out/asset.ee7de53d64cc9d6248fa6aa550f92358f6c907b5efd6f3298aeab1b5e7ea358a/index.js function c (line 1) | function c(r,e){return{GroupId:r,IpPermissions:[{UserIdGroupPairs:[{Grou... function d (line 1) | function d(r){return{GroupId:r,IpPermissions:[{IpRanges:[{CidrIp:"0.0.0.... function f (line 1) | async function f(r){let e=r.ResourceProperties.DefaultSecurityGroupId,o=... function h (line 1) | async function h(r){let e=r.OldResourceProperties.DefaultSecurityGroupId... function p (line 1) | async function p(r,e){try{await u.revokeSecurityGroupEgress(d(r))}catch(... function m (line 1) | async function m(r,e){await u.authorizeSecurityGroupIngress(c(r,e)),awai... FILE: evi/evi-python-wss-clm-endpoint/cdk/cdk.out/asset.f372550efb96be7f549f1d0346e8646080c1fe2b15c7c2e3b1dceb07b6656c54/agent.py function reflect (line 323) | def reflect(fragment): function eliza_response (line 334) | def eliza_response(user_input): FILE: evi/evi-python-wss-clm-endpoint/cdk/cdk.out/asset.f372550efb96be7f549f1d0346e8646080c1fe2b15c7c2e3b1dceb07b6656c54/app.py function root (line 9) | async def root(): function websocket_handler (line 13) | async def websocket_handler(websocket: WebSocket) -> None: FILE: evi/evi-python-wss-clm-endpoint/cdk/cdk.out/asset.f372550efb96be7f549f1d0346e8646080c1fe2b15c7c2e3b1dceb07b6656c54/modal_app.py function endpoint (line 14) | def endpoint(): FILE: evi/evi-python-wss-clm-endpoint/cdk/cdk/eliza_stack.py class ElizaStack (line 9) | class ElizaStack(cdk.Stack): method __init__ (line 11) | def __init__(self, scope: cdk.App, id: str, **kwargs) -> None: FILE: evi/evi-python-wss-clm-endpoint/modal/modal_app.py function endpoint (line 14) | def endpoint(): FILE: evi/evi-python-wss-clm-endpoint/modal_app.py function endpoint (line 14) | def endpoint(): FILE: evi/evi-react-native/App.tsx type ChatEntry (line 27) | interface ChatEntry { FILE: evi/evi-react-native/VoiceIsolationModePrompt.tsx type VoiceIsolationModePromptProps (line 12) | interface VoiceIsolationModePromptProps { FILE: evi/evi-react-native/modules/audio/src/AudioModule.ts class AudioModule (line 5) | class AudioModule extends NativeModule { FILE: evi/evi-react-native/modules/audio/src/AudioModule.types.ts type MicrophoneMode (line 1) | type MicrophoneMode = "N/A" | "Standard" | "Voice Isolation" | "Wide Spe... type AudioModuleEvents (line 3) | type AudioModuleEvents = { type AudioEventPayload (line 8) | type AudioEventPayload = { FILE: evi/evi-react-native/modules/audio/src/AudioModule.web.ts method getPermissions (line 27) | async getPermissions(): Promise { method startRecording (line 34) | async startRecording(): Promise { method stopRecording (line 55) | async stopRecording(): Promise { method enqueueAudio (line 64) | async enqueueAudio(base64EncodedAudio: string): Promise { method mute (line 68) | async mute(): Promise { method unmute (line 72) | async unmute(): Promise { method stopPlayback (line 76) | async stopPlayback(): Promise { method addListener (line 84) | async addListener(eventName: keyof AudioModuleEvents, f: AudioModuleEven... method showMicrophoneModes (line 89) | async showMicrophoneModes(): Promise { method getMicrophoneMode (line 94) | async getMicrophoneMode(): Promise { FILE: evi/evi-swift-chat/access_token_service/run_token_service.py function get_access_token (line 10) | def get_access_token(): FILE: evi/evi-touchdesigner/Scripts/HumeTD.py class HumeTDExt (line 3) | class HumeTDExt: method __init__ (line 4) | def __init__(self, ownerComp): method Send_user_input (line 7) | def Send_user_input(self, user_input: str): FILE: evi/evi-touchdesigner/Scripts/MessagePlaback.py class MessagePlaybackExt (line 7) | class MessagePlaybackExt: method __init__ (line 8) | def __init__(self, owner_comp): method Handle_ws_msg (line 18) | def Handle_ws_msg(self, msg): method check_messages (line 24) | def check_messages(self): method Add_item (line 28) | def Add_item(self, audio_data): method Remove_item (line 35) | def Remove_item(self): method play_next_item (line 44) | def play_next_item(self): method get_audio_duration (line 55) | def get_audio_duration(self, filepath): FILE: evi/evi-typescript-chat-history/src/index.ts function main (line 17) | async function main(): Promise { function fetchAllChatEvents (line 57) | async function fetchAllChatEvents(chatId: string): Promise { function disconnect (line 101) | function disconnect(): void { function captureAudio (line 134) | async function captureAudio(): Promise { function handleWebSocketOpenEvent (line 168) | async function handleWebSocketOpenEvent(): Promise { function handleWebSocketMessageEvent (line 192) | async function handleWebSocketMessageEvent( function handleWebSocketErrorEvent (line 236) | function handleWebSocketErrorEvent(error: Error): void { function handleWebSocketCloseEvent (line 244) | async function handleWebSocketCloseEvent(): Promise { function appendMessage (line 260) | function appendMessage( function toggleBtnStates (line 283) | function toggleBtnStates(): void { function extractTopThreeEmotions (line 292) | function extractTopThreeEmotions( type Score (line 322) | interface Score { type ChatMessage (line 327) | interface ChatMessage { class ChatCard (line 334) | class ChatCard { method constructor (line 337) | constructor(message: ChatMessage) { method createScoreItem (line 341) | private createScoreItem(score: Score): HTMLElement { method render (line 348) | public render(): HTMLElement { FILE: evi/evi-typescript-proxy/app/api.ts class Api (line 4) | class Api { method broadcastState (line 9) | broadcastState(state: State): void { method getNextAPIEvent (line 17) | getNextAPIEvent(): AppEvent | undefined { method hasAPIEvents (line 22) | hasAPIEvents(): boolean { method handleRequest (line 27) | handleRequest( method handlePostAppEvent (line 46) | private handlePostAppEvent( method handleSubscribeAppEvent (line 70) | private handleSubscribeAppEvent( FILE: evi/evi-typescript-proxy/app/cli.ts class CLI (line 15) | class CLI { method constructor (line 22) | constructor(PORT: number, WS_PATH: string) { method setState (line 27) | setState(state: State): void { method getNextCLIEvent (line 32) | getNextCLIEvent(): AppEvent | undefined { method menu (line 36) | private async menu( method playbackMenu (line 63) | private async playbackMenu( method errorSimulationMenu (line 87) | private async errorSimulationMenu( method getNextEvent (line 128) | async getNextEvent(): Promise { method getNextEvent_ (line 142) | private async getNextEvent_(): Promise { method maybePromptUserIfNeeded (line 301) | async maybePromptUserIfNeeded() { method runPromptLoop (line 308) | async runPromptLoop() { FILE: evi/evi-typescript-proxy/app/downstream.ts type InferRaw (line 5) | type InferRaw = T extends { parse: (raw: infer R) => any } ? R : never; type WebSocketErrorRaw (line 6) | type WebSocketErrorRaw = InferRaw string }) { method handleConnection (line 109) | handleConnection(ws: WebSocket): void { FILE: evi/evi-typescript-proxy/app/main.ts constant DIST_DIR (line 21) | const DIST_DIR = path.join(__dirname, "../web/dist"); constant PORT (line 23) | const PORT = 3000; constant DOWNSTREAM_WS_PATH (line 24) | const DOWNSTREAM_WS_PATH = "/v0/evi/chat"; constant UPSTREAM_WS_BASE_URL (line 25) | const UPSTREAM_WS_BASE_URL = "wss://api.hume.ai"; function shallowEqual (line 272) | function shallowEqual(a: any, b: any): boolean { function main (line 284) | async function main() { FILE: evi/evi-typescript-proxy/app/upstream.ts method onMessage (line 13) | public onMessage(handler: (message: Message) => void): void { method onConnect (line 17) | public onConnect(handler: () => void): void { method onDisconnect (line 21) | public onDisconnect(handler: () => void): void { type ConnectArgs (line 26) | type ConnectArgs = { class LiveUpstream (line 31) | class LiveUpstream extends BaseUpstream { method connect (line 36) | public connect(args: ConnectArgs): void { method close (line 82) | public close(): void { method send (line 88) | public send(message: WSMessage): void { class PlaybackUpstream (line 99) | class PlaybackUpstream extends BaseUpstream { method setPlaybackMessages (line 104) | public setPlaybackMessages(messages: Message[]): void { method connect (line 109) | public connect(_args?: any): void { method close (line 115) | public close(): void { method send (line 121) | public send(_message: WSMessage): void { class UninitializedUpstream (line 134) | class UninitializedUpstream extends BaseUpstream { method connect (line 135) | public connect(_args?: any): void { method close (line 141) | public close(): void { method send (line 147) | public send(_message: WSMessage): void { FILE: evi/evi-typescript-proxy/web/ChatControls.tsx function ChatControls (line 4) | function ChatControls() { FILE: evi/evi-typescript-proxy/web/EVIChat.tsx function EVIChat (line 25) | function EVIChat({ accessToken }: { accessToken?: string }) { FILE: evi/evi-typescript-proxy/web/StartCall.tsx function StartCall (line 5) | function StartCall() { FILE: evi/evi-typescript-proxy/web/app.tsx function App (line 6) | function App() { FILE: evi/evi-typescript-proxy/web/useProxyState.ts function useProxyState (line 4) | function useProxyState() { FILE: evi/evi-typescript-quickstart/src/lib/audio.ts function startAudioCapture (line 30) | async function startAudioCapture( FILE: evi/evi-typescript-quickstart/src/lib/evi.test.ts function waitForSocketOpen (line 284) | function waitForSocketOpen(socket: any): Promise { function waitForChatMetadata (line 306) | function waitForChatMetadata(getSocket: () => any): Promise { function fetchChatEvents (line 348) | async function fetchChatEvents( FILE: evi/evi-typescript-quickstart/src/lib/evi.ts function getClient (line 6) | function getClient(apiKey: string): HumeClient { function connectEVI (line 33) | function connectEVI( FILE: evi/evi-typescript-quickstart/src/lib/ui.ts function extractTopThreeEmotions (line 16) | function extractTopThreeEmotions( function appendChatMessage (line 38) | function appendChatMessage( FILE: evi/evi-typescript-quickstart/src/main.ts function setConnected (line 72) | function setConnected(on: boolean): void { function handleOpen (line 78) | async function handleOpen() { function handleMessage (line 84) | async function handleMessage(msg: Hume.empathicVoice.chat.SubscribeEvent) { function handleError (line 111) | function handleError(err: Event | Error) { function handleClose (line 115) | function handleClose(e: unknown) { function connect (line 120) | function connect( function disconnect (line 142) | function disconnect() { FILE: evi/evi-typescript-webhooks/src/main.ts constant PORT (line 15) | const PORT = 5000; FILE: evi/evi-typescript-webhooks/src/util.ts function getWebhookSigningKey (line 14) | function getWebhookSigningKey(): string { function fetchAllChatEvents (line 29) | async function fetchAllChatEvents(client: HumeClient, chatId: string): P... function generateTranscript (line 46) | function generateTranscript(chatEvents: Hume.empathicVoice.ReturnChatEve... function saveTranscriptToFile (line 63) | async function saveTranscriptToFile(transcript: string, chatId: string):... function getChatTranscript (line 82) | async function getChatTranscript(client: HumeClient, chatId: string): Pr... function validateWebhookHeaders (line 96) | function validateWebhookHeaders( FILE: evi/evi-unity-quickstart/Assets/Editor/AutoLoadDefaultScene.cs class AutoLoadDefaultScene (line 5) | [InitializeOnLoad] method AutoLoadDefaultScene (line 10) | static AutoLoadDefaultScene() method LoadDefaultSceneIfNeeded (line 15) | private static void LoadDefaultSceneIfNeeded() FILE: evi/evi-unity-quickstart/Assets/Scripts/HumeEVI.cs class HumeEVI (line 8) | [RequireComponent(typeof(AudioSource))] type ConversationState (line 41) | public enum ConversationState method SetApiKey (line 65) | public void SetApiKey(string key) method Update (line 70) | void Update() method OnDestroy (line 86) | void OnDestroy() method StartConversation (line 91) | public async void StartConversation() method StopConversation (line 131) | public async void StopConversation() method ConnectToEVI (line 167) | private async Task ConnectToEVI() method SubscribeToEvents (line 204) | private void SubscribeToEvents() method ProcessAudioOutput (line 258) | private void ProcessAudioOutput(string base64Audio) method PlayNextAudioChunk (line 284) | private void PlayNextAudioChunk() method WaitForPlaybackComplete (line 330) | private System.Collections.IEnumerator WaitForPlaybackComplete(float d... method StartMicrophoneCapture (line 344) | private void StartMicrophoneCapture() method StopMicrophoneCapture (line 368) | private void StopMicrophoneCapture() method SendMicrophoneAudio (line 385) | private async void SendMicrophoneAudio() method ConvertWavToFloats (line 461) | private float[] ConvertWavToFloats(byte[] wavBytes) method ConvertS16LEToFloats (line 554) | private float[] ConvertS16LEToFloats(byte[] bytes) method ConvertU8ToFloats (line 570) | private float[] ConvertU8ToFloats(byte[] bytes) method ConvertFloatsToPCM (line 585) | private byte[] ConvertFloatsToPCM(float[] samples) class UnityMainThreadDispatcher (line 612) | public class UnityMainThreadDispatcher : MonoBehaviour method Awake (line 617) | void Awake() method Update (line 630) | void Update() method Enqueue (line 641) | public static void Enqueue(Action action) method Initialize (line 651) | [RuntimeInitializeOnLoadMethod(RuntimeInitializeLoadType.BeforeSceneLo... FILE: evi/evi-unity-quickstart/Assets/Scripts/SceneBuilder.cs class SceneBuilder (line 3) | public class SceneBuilder : MonoBehaviour method Awake (line 8) | void Awake() method Start (line 13) | void Start() method BuildScene (line 18) | void BuildScene() class ConversationVisualFeedback (line 78) | public class ConversationVisualFeedback : MonoBehaviour method Initialize (line 106) | public void Initialize(HumeEVI eviComponent, Renderer renderer) method OnDestroy (line 118) | void OnDestroy() method Update (line 128) | void Update() method AnimateColor (line 139) | private void AnimateColor() method GetBaseColorForState (line 157) | private Color GetBaseColorForState(HumeEVI.ConversationState state) method OnStateChanged (line 172) | private void OnStateChanged(HumeEVI.ConversationState newState) method UpdateInstructionText (line 180) | private void UpdateInstructionText() method OnUserTranscript (line 205) | private void OnUserTranscript(string transcript) method OnAssistantMessage (line 211) | private void OnAssistantMessage(string message) method UpdateTranscriptDisplay (line 217) | private void UpdateTranscriptDisplay() class ClickToConverse (line 247) | public class ClickToConverse : MonoBehaviour method Start (line 251) | void Start() method OnMouseDown (line 256) | void OnMouseDown() FILE: expression-measurement/batch/next-js-emotional-language/src/lib/mutations/processTextFile.ts function processTextFile (line 10) | async function processTextFile(fileUrl: string) { function sendFile (line 36) | async function sendFile(fileUrl: string) { function pollForResultsUrl (line 52) | async function pollForResultsUrl(jobId: string, maxAttempts: number) { function fetchResultsFile (line 85) | async function fetchResultsFile(url: string) { FILE: expression-measurement/batch/next-js-emotional-language/src/lib/utils.ts function cn (line 4) | function cn(...inputs: ClassValue[]) { FILE: expression-measurement/batch/next-js-emotional-language/src/pages/_app.tsx function App (line 18) | function App({ Component, pageProps }: AppProps) { FILE: expression-measurement/batch/next-js-emotional-language/src/pages/_document.tsx function Document (line 3) | function Document() { FILE: expression-measurement/batch/next-js-emotional-language/src/pages/api/results.ts function handler (line 8) | async function handler( FILE: expression-measurement/batch/next-js-emotional-language/src/pages/api/send.ts type Data (line 7) | type Data = { function handler (line 11) | async function handler( FILE: expression-measurement/batch/next-js-emotional-language/src/pages/index.tsx function Home (line 19) | function Home() { FILE: expression-measurement/batch/python-top-emotions/top_emotions.py function main (line 10) | async def main(): function poll_for_completion (line 55) | async def poll_for_completion(client: AsyncHumeClient, job_id, timeout=1... function poll_until_complete (line 68) | async def poll_until_complete(client: AsyncHumeClient, job_id): function process_predictions (line 121) | def process_predictions(job_predictions: List[UnionPredictResult], start... FILE: expression-measurement/streaming/next-js-streaming-example/components/inputs/Button.tsx type ButtonProps (line 4) | type ButtonProps = React.HTMLAttributes & { function Button (line 11) | function Button({ className, variant, text, onClick, tooltip }: ButtonPr... FILE: expression-measurement/streaming/next-js-streaming-example/components/inputs/TextArea.tsx type TextAreaProps (line 5) | type TextAreaProps = { function TextArea (line 14) | function TextArea({ className, inputClassName, text, placeholder, onChan... FILE: expression-measurement/streaming/next-js-streaming-example/components/inputs/TextBox.tsx type TextBoxProps (line 4) | type TextBoxProps = { function TextBox (line 15) | function TextBox({ FILE: expression-measurement/streaming/next-js-streaming-example/components/menu/Auth.tsx type ChildElement (line 7) | type ChildElement = JSX.Element | string; type AuthState (line 16) | type AuthState = { type AuthProps (line 23) | type AuthProps = { function Auth (line 27) | function Auth({ children }: AuthProps) { FILE: expression-measurement/streaming/next-js-streaming-example/components/menu/Login.tsx type LoginProps (line 6) | type LoginProps = { function Login (line 10) | function Login({ authenticate }: LoginProps) { FILE: expression-measurement/streaming/next-js-streaming-example/components/menu/Nav.tsx function Nav (line 3) | function Nav() { type NavItemProps (line 25) | type NavItemProps = { function NavItem (line 30) | function NavItem({ route, name }: NavItemProps) { FILE: expression-measurement/streaming/next-js-streaming-example/components/menu/Toolbar.tsx function Toolbar (line 5) | function Toolbar() { FILE: expression-measurement/streaming/next-js-streaming-example/components/widgets/AudioWidgets.tsx type AudioWidgetsProps (line 12) | interface AudioWidgetsProps { function AudioWidgets (line 19) | function AudioWidgets({ modelName, recordingLengthMs, streamWindowLength... FILE: expression-measurement/streaming/next-js-streaming-example/components/widgets/BurstWidgets.tsx type BurstWidgetsProps (line 6) | type BurstWidgetsProps = { function BurstWidgets (line 10) | function BurstWidgets({ onTimeline }: BurstWidgetsProps) { FILE: expression-measurement/streaming/next-js-streaming-example/components/widgets/Descriptor.tsx type DescriptorProps (line 6) | type DescriptorProps = { function Descriptor (line 11) | function Descriptor({ className, emotions }: DescriptorProps) { FILE: expression-measurement/streaming/next-js-streaming-example/components/widgets/DiscreteTimeline.tsx type DiscreteTimelineProps (line 6) | type DiscreteTimelineProps = { function DiscreteTimeline (line 11) | function DiscreteTimeline({ className, predictions }: DiscreteTimelinePr... type DetectionProps (line 99) | type DetectionProps = { function Detection (line 104) | function Detection({ className, detection }: DetectionProps) { FILE: expression-measurement/streaming/next-js-streaming-example/components/widgets/FaceTrackedVideo.tsx type FaceTrackedVideoProps (line 5) | type FaceTrackedVideoProps = { function FaceTrackedVideo (line 13) | function FaceTrackedVideo({ className, trackedFaces, onVideoReady, width... FILE: expression-measurement/streaming/next-js-streaming-example/components/widgets/FaceWidgets.tsx type FaceWidgetsProps (line 16) | type FaceWidgetsProps = { function FaceWidgets (line 20) | function FaceWidgets({ onCalibrate }: FaceWidgetsProps) { FILE: expression-measurement/streaming/next-js-streaming-example/components/widgets/LanguageWidgets.tsx function LanguageWidgets (line 10) | function LanguageWidgets() { FILE: expression-measurement/streaming/next-js-streaming-example/components/widgets/Loader.tsx type LoaderProps (line 5) | type LoaderProps = { function Loader (line 12) | function Loader({ className, emotions, emotionName, numLevels }: LoaderP... FILE: expression-measurement/streaming/next-js-streaming-example/components/widgets/LoaderSet.tsx type LoaderProps (line 5) | type LoaderProps = { function LoaderSet (line 12) | function LoaderSet({ className, emotions, emotionNames, numLevels }: Loa... FILE: expression-measurement/streaming/next-js-streaming-example/components/widgets/ProsodyWidgets.tsx function ProsodyWidgets (line 3) | function ProsodyWidgets() { FILE: expression-measurement/streaming/next-js-streaming-example/components/widgets/TopEmotions.tsx type TopEmotionsProps (line 3) | type TopEmotionsProps = { function TopEmotions (line 9) | function TopEmotions({ className, emotions, numEmotions }: TopEmotionsPr... FILE: expression-measurement/streaming/next-js-streaming-example/lib/data/audioPrediction.ts type AudioPrediction (line 4) | type AudioPrediction = { FILE: expression-measurement/streaming/next-js-streaming-example/lib/data/boundingBox.ts type BoundingBox (line 1) | type BoundingBox = { FILE: expression-measurement/streaming/next-js-streaming-example/lib/data/characterRange.ts type CharacterRange (line 1) | type CharacterRange = { FILE: expression-measurement/streaming/next-js-streaming-example/lib/data/embedding.ts type Embedding (line 1) | type Embedding = number[]; FILE: expression-measurement/streaming/next-js-streaming-example/lib/data/emotion.ts type Emotion (line 1) | type Emotion = { type EmotionName (line 6) | type EmotionName = FILE: expression-measurement/streaming/next-js-streaming-example/lib/data/facePrediction.ts type FacePrediction (line 4) | type FacePrediction = { FILE: expression-measurement/streaming/next-js-streaming-example/lib/data/languagePrediction.ts type LanguagePrediction (line 4) | type LanguagePrediction = { FILE: expression-measurement/streaming/next-js-streaming-example/lib/data/range.ts type Range (line 1) | type Range = { FILE: expression-measurement/streaming/next-js-streaming-example/lib/data/timeRange.ts type TimeRange (line 1) | type TimeRange = { FILE: expression-measurement/streaming/next-js-streaming-example/lib/data/trackedFace.ts type TrackedFace (line 3) | type TrackedFace = { FILE: expression-measurement/streaming/next-js-streaming-example/lib/hooks/keyPress.ts function useKeypress (line 3) | function useKeypress(key: string, callback: () => void, deps: Dependency... FILE: expression-measurement/streaming/next-js-streaming-example/lib/hooks/stability.ts function useStableEmotions (line 6) | function useStableEmotions(emotions: Emotion[], embeddingDistThreshold: ... FILE: expression-measurement/streaming/next-js-streaming-example/lib/hooks/storage.ts function useStorage (line 5) | function useStorage(key: string) { FILE: expression-measurement/streaming/next-js-streaming-example/lib/media/audioRecorder.ts class AudioRecorder (line 3) | class AudioRecorder { method constructor (line 7) | private constructor(recorder: MediaRecorder, mediaStream: MediaStream) { method create (line 12) | static async create(): Promise { method stopRecording (line 19) | async stopRecording() { method record (line 25) | record(length: number): Promise { FILE: expression-measurement/streaming/next-js-streaming-example/lib/media/videoRecorder.ts type Size (line 3) | type Size = { class VideoRecorder (line 8) | class VideoRecorder { method constructor (line 14) | private constructor( method create (line 27) | static async create(videoElement: HTMLVideoElement, photoElement: HTML... method stopRecording (line 38) | async stopRecording() { method setVideoSize (line 44) | private static setVideoSize(videoElement: HTMLVideoElement, photoEleme... method takePhoto (line 64) | async takePhoto(format: string = "image/png"): Promise { FILE: expression-measurement/streaming/next-js-streaming-example/lib/utilities/asyncUtilities.ts function sleep (line 1) | function sleep(delay: number) { FILE: expression-measurement/streaming/next-js-streaming-example/lib/utilities/blobUtilities.ts function blobToBase64 (line 1) | function blobToBase64(blob: Blob) { function canvasToImageBlob (line 14) | function canvasToImageBlob(canvas: HTMLCanvasElement, format: string = "... FILE: expression-measurement/streaming/next-js-streaming-example/lib/utilities/embeddingUtilities.ts function emotionsToEmbedding (line 6) | function emotionsToEmbedding(emotions: Emotion[]): Embedding { function emotionDist (line 21) | function emotionDist(emotionsA: Emotion[], emotionsB: Emotion[]): number { function emotionsToScoreMap (line 25) | function emotionsToScoreMap(emotions: Emotion[]): Map> = new Map([ function getEmotionDescriptor (line 112) | function getEmotionDescriptor(name: EmotionName): Optional { FILE: expression-measurement/streaming/next-js-streaming-example/lib/utilities/environmentUtilities.ts type Environment (line 9) | enum Environment { function parseEnvironment (line 13) | function parseEnvironment(env: string): Environment { function getApiUrlHttp (line 17) | function getApiUrlHttp(environment: Environment): string { function getApiUrlWs (line 21) | function getApiUrlWs(environment: Environment): string { FILE: expression-measurement/streaming/next-js-streaming-example/lib/utilities/scalingUtilities.ts function scaleEmotionsToRanges (line 5) | function scaleEmotionsToRanges(emotions: Emotion[]): Emotion[] { function scale (line 24) | function scale(value: number, range: Range) { function clip (line 29) | function clip(value: number, range: Range) { constant RANGE_MAP (line 33) | const RANGE_MAP = new Map([ FILE: expression-measurement/streaming/next-js-streaming-example/lib/utilities/styleUtilities.ts function cn (line 5) | function cn(...inputs: ClassValue[]) { FILE: expression-measurement/streaming/next-js-streaming-example/lib/utilities/typeUtilities.ts type Optional (line 1) | type Optional = T | undefined; FILE: expression-measurement/streaming/next-js-streaming-example/pages/_app.tsx function App (line 14) | function App({ Component, pageProps }: AppProps) { FILE: expression-measurement/streaming/next-js-streaming-example/pages/burst/index.tsx function BurstPage (line 3) | function BurstPage() { FILE: expression-measurement/streaming/next-js-streaming-example/pages/burst/timeline/index.tsx function BurstTimelinePage (line 4) | function BurstTimelinePage() { FILE: expression-measurement/streaming/next-js-streaming-example/pages/face/calibrate/index.tsx function FaceCalibratePage (line 7) | function FaceCalibratePage() { FILE: expression-measurement/streaming/next-js-streaming-example/pages/face/index.tsx function FacePage (line 3) | function FacePage() { FILE: expression-measurement/streaming/next-js-streaming-example/pages/index.tsx function HomePage (line 10) | function HomePage() { type ModelSectionProps (line 28) | type ModelSectionProps = { function ModelSection (line 34) | function ModelSection(props: ModelSectionProps) { FILE: expression-measurement/streaming/next-js-streaming-example/pages/language/index.tsx function LanguagePage (line 3) | function LanguagePage() { FILE: expression-measurement/streaming/next-js-streaming-example/pages/prosody/index.tsx function ProsodyPage (line 3) | function ProsodyPage() { FILE: expression-measurement/streaming/python-streaming-example/main.py class Result (line 27) | class Result(TypedDict): function process_emotion_scores (line 31) | def process_emotion_scores(event: StreamModelPredictions) -> List[Result]: function print_emotion_summary (line 46) | def print_emotion_summary(result: Result) -> None: function streaming_example (line 57) | async def streaming_example() -> None: function main (line 93) | def main(): FILE: expression-measurement/streaming/python-streaming-example/test_main.py function api_key (line 18) | def api_key(): function test_send_hello_world_returns_non_empty_emotion_analysis (line 26) | async def test_send_hello_world_returns_non_empty_emotion_analysis(api_k... FILE: tts/tts-dotnet-quickstart/Program.cs class Program (line 16) | class Program method RunExamplesAsync (line 47) | static async Task RunExamplesAsync() method Main (line 69) | static async Task Main(string[] args) method Example1Async (line 81) | static async Task Example1Async() method Example2Async (line 101) | static async Task Example2Async() method Example3Async (line 194) | static async Task Example3Async() method StreamAudioToPlayerAsync (line 248) | private static async Task StreamAudioToPlayerAsync( class SilenceFiller (line 267) | public class SilenceFiller : IDisposable method SilenceFiller (line 279) | public SilenceFiller(Stream outputStream) method RunFillerLoop (line 285) | private async Task RunFillerLoop() method WriteAudio (line 321) | public void WriteAudio(byte[] audioBytes) method EndStreamAsync (line 329) | public async Task EndStreamAsync() method Dispose (line 335) | public void Dispose() class StreamingAudioPlayer (line 346) | public class StreamingAudioPlayer : IDisposable method StreamingAudioPlayer (line 358) | public StreamingAudioPlayer(bool usePcmFormat = false) method StartAsync (line 368) | public Task StartAsync() method WriteAudio (line 375) | public void WriteAudio(byte[] audioBytes) method StopAsync (line 390) | public async Task StopAsync() method StartAudioProcess (line 407) | private void StartAudioProcess() method Dispose (line 438) | public void Dispose() method StartAudioPlayer (line 452) | private static StreamingAudioPlayer StartAudioPlayer(bool usePcmFormat... FILE: tts/tts-dotnet-quickstart/StreamingTtsService.cs class Queue (line 27) | internal class Queue method Push (line 34) | public void Push(T x) method End (line 51) | public void End() method GetAsyncEnumerable (line 68) | public async IAsyncEnumerable GetAsyncEnumerable() class StreamingTtsClient (line 115) | public class StreamingTtsClient : IDisposable method StreamingTtsClient (line 132) | public StreamingTtsClient(string apiKey, bool enableDebugLogging = false) method ConnectAsync (line 143) | public async Task ConnectAsync() method SendAsync (line 208) | public async Task SendAsync(object message) method SendFlushAsync (line 216) | public async Task SendFlushAsync() method SendCloseAsync (line 222) | public async Task SendCloseAsync() method ReceiveAudioChunksAsync (line 228) | public async IAsyncEnumerable ReceiveAudioChunksAsy... method LogDebug (line 258) | private void LogDebug(string message) method Dispose (line 266) | public void Dispose() FILE: tts/tts-dotnet-quickstart/TtsTests.cs class TtsTestFixture (line 14) | public class TtsTestFixture : IAsyncLifetime method InitializeAsync (line 19) | public Task InitializeAsync() method DisposeAsync (line 39) | public Task DisposeAsync() class TtsJsonStreamTests (line 45) | [Collection("TtsTests")] method TtsJsonStreamTests (line 50) | public TtsJsonStreamTests(TtsTestFixture fixture) method TestFixture_HasApiKey (line 55) | [Fact(DisplayName = "test fixture has API key")] method GeneratesJsonStream_WithOctave1 (line 62) | [Fact(DisplayName = "connects w/ API key, generates JSON stream w/ Oct... method GeneratesJsonStream_WithOctave2AndTimestamps (line 88) | [Fact(DisplayName = "connects w/ API key, generates JSON stream w/ Oct... class TtsStreamInputTests (line 144) | [Collection("TtsTests")] method TtsStreamInputTests (line 149) | public TtsStreamInputTests(TtsTestFixture fixture) method CreatesStreamAndConnectsSuccessfully (line 154) | [Fact(DisplayName = "StreamingTtsClient: creates a bidirectional strea... method SendsMessagesAndReceivesAudioChunks (line 170) | [Fact(DisplayName = "StreamingTtsClient: sends messages and receives a... class TtsTestCollection (line 208) | [CollectionDefinition("TtsTests")] FILE: tts/tts-next-js-agora/app/api/generate-agora-token/route.ts constant APP_ID (line 4) | const APP_ID = constant APP_CERTIFICATE (line 9) | const APP_CERTIFICATE = constant EXPIRATION_SECONDS (line 13) | const EXPIRATION_SECONDS = Number( function generateChannelName (line 19) | function generateChannelName (): string function GET (line 26) | async function GET ( request: NextRequest ) FILE: tts/tts-next-js-agora/app/api/invite-agent/route.ts constant DEFAULT_BASE_URL (line 10) | const DEFAULT_BASE_URL = constant AGORA_APP_ID (line 13) | const AGORA_APP_ID = constant AGORA_BASE_URL (line 18) | const AGORA_BASE_URL = constant AGORA_CUSTOMER_ID (line 23) | const AGORA_CUSTOMER_ID = constant AGORA_CUSTOMER_SECRET (line 25) | const AGORA_CUSTOMER_SECRET = constant AGORA_AGENT_UID (line 29) | const AGORA_AGENT_UID = constant HUME_API_KEY (line 32) | const HUME_API_KEY = constant HUME_VOICE_ID (line 34) | const HUME_VOICE_ID = constant OPENAI_API_KEY (line 37) | const OPENAI_API_KEY = constant OPENAI_MODEL (line 39) | const OPENAI_MODEL = function validateEnvironment (line 42) | function validateEnvironment () function createAuthorizationHeader (line 61) | function createAuthorizationHeader (): string function getConfig (line 67) | function getConfig () function POST (line 95) | async function POST ( request: Request ) FILE: tts/tts-next-js-agora/app/api/stop-conversation/route.ts constant AGORA_APP_ID (line 3) | const AGORA_APP_ID = constant AGORA_BASE_URL (line 8) | const AGORA_BASE_URL = constant AGORA_CUSTOMER_ID (line 13) | const AGORA_CUSTOMER_ID = constant AGORA_CUSTOMER_SECRET (line 15) | const AGORA_CUSTOMER_SECRET = type StopRequest (line 20) | type StopRequest = { function POST (line 24) | async function POST ( request: Request ) FILE: tts/tts-next-js-agora/app/layout.tsx function RootLayout (line 12) | function RootLayout({ FILE: tts/tts-next-js-agora/app/page.tsx function Page (line 35) | function Page() { FILE: tts/tts-next-js-agora/components/AudioVisualizer.tsx type AudioVisualizerProps (line 6) | interface AudioVisualizerProps { function AudioVisualizer (line 10) | function AudioVisualizer({ track }: AudioVisualizerProps) { FILE: tts/tts-next-js-agora/components/ConversationComponent.tsx function ConversationComponent (line 25) | function ConversationComponent({ FILE: tts/tts-next-js-agora/components/ConvoTextStream.tsx type Props (line 7) | interface Props { function ConvoTextStream (line 13) | function ConvoTextStream({ FILE: tts/tts-next-js-agora/components/MicrophoneButton.tsx type MicrophoneButtonProps (line 7) | interface MicrophoneButtonProps { function MicrophoneButton (line 13) | function MicrophoneButton({ FILE: tts/tts-next-js-agora/lib/message.ts constant CONSOLE_LOG_PREFIX (line 3) | const CONSOLE_LOG_PREFIX = '[MessageService]'; constant DEFAULT_MESSAGE_CACHE_TIMEOUT (line 4) | const DEFAULT_MESSAGE_CACHE_TIMEOUT = 1000 * 60 * 5; type TDataChunk (line 12) | type TDataChunk = { type EMessageStatus (line 19) | enum EMessageStatus type ETranscriptionObjectType (line 26) | enum ETranscriptionObjectType type IMessageListItem (line 33) | interface IMessageListItem type IMessageArrayItem (line 41) | interface IMessageArrayItem type ITranscriptionBase (line 51) | interface ITranscriptionBase type IUserTranscription (line 64) | interface IUserTranscription extends ITranscriptionBase type IAgentTranscription (line 70) | interface IAgentTranscription extends ITranscriptionBase type IMessageInterrupt (line 78) | interface IMessageInterrupt class MessageEngine (line 92) | class MessageEngine method constructor (line 106) | constructor ( method _listenRtcEvents (line 117) | private _listenRtcEvents () method handleStreamMessage (line 127) | public handleStreamMessage ( stream: Uint8Array ) method handleMessage (line 140) | public handleMessage ( method handleTextMessage (line 170) | private handleTextMessage ( message: IUserTranscription ) method handleMessageInterrupt (line 200) | private handleMessageInterrupt ( message: IMessageInterrupt ) method _appendChatHistory (line 212) | private _appendChatHistory ( method _mutateChatHistory (line 220) | private _mutateChatHistory () method handleChunk (line 239) | public handleChunk ( method streamMessage2Chunk (line 311) | public streamMessage2Chunk ( stream: Uint8Array ): string method cleanup (line 316) | public cleanup () FILE: tts/tts-next-js-agora/types/agora-rtc-react.d.ts type UID (line 4) | type UID = number | string; type IRemoteAudioTrack (line 6) | interface IRemoteAudioTrack type ILocalAudioTrack (line 11) | interface ILocalAudioTrack extends IRemoteAudioTrack type IMicrophoneAudioTrack (line 17) | interface IMicrophoneAudioTrack extends ILocalAudioTrack { } type IAgoraRTCClient (line 19) | interface IAgoraRTCClient type AgoraRTCProviderProps (line 59) | interface AgoraRTCProviderProps FILE: tts/tts-next-js-agora/types/conversation.ts type AgoraTokenData (line 3) | interface AgoraTokenData type ClientStartRequest (line 12) | interface ClientStartRequest type StopConversationRequest (line 22) | interface StopConversationRequest type ConversationComponentProps (line 27) | interface ConversationComponentProps type AgentResponse (line 34) | interface AgentResponse type HumeTTSParams (line 41) | interface HumeTTSParams type AgoraStartRequest (line 51) | interface AgoraStartRequest type TTSConfig (line 95) | interface TTSConfig type TokenRenewalHandler (line 101) | interface TokenRenewalHandler FILE: tts/tts-next-js-chat/src/app/api/chat/route.ts function POST (line 4) | async function POST(req: Request) { constant SYSTEM_PROMPT (line 25) | const SYSTEM_PROMPT = ` FILE: tts/tts-next-js-chat/src/app/api/transcribe/route.ts function POST (line 3) | async function POST(req: Request) { FILE: tts/tts-next-js-chat/src/app/api/tts/route.ts function POST (line 10) | async function POST(req: NextRequest) { FILE: tts/tts-next-js-chat/src/app/api/voices/route.ts function GET (line 5) | async function GET(req: NextRequest) { FILE: tts/tts-next-js-chat/src/app/components/AudioPlayer.tsx function AudioPlayer (line 4) | function AudioPlayer({ chunks }: { chunks: Uint8Array[] }) { FILE: tts/tts-next-js-chat/src/app/components/Chat.tsx function Chat (line 14) | function Chat() { FILE: tts/tts-next-js-chat/src/app/components/ControlsPanel.tsx function ControlsPanel (line 8) | function ControlsPanel() { FILE: tts/tts-next-js-chat/src/app/components/VoiceSelector.tsx type Props (line 4) | interface Props { function VoiceSelector (line 10) | function VoiceSelector({ FILE: tts/tts-next-js-chat/src/app/components/logos/Hume.tsx type HumeLogoProps (line 4) | type HumeLogoProps = SVGAttributes; function HumeLogo (line 6) | function HumeLogo(props: HumeLogoProps) { FILE: tts/tts-next-js-chat/src/app/context/VoiceSettingsContext.tsx type VoiceSettings (line 5) | interface VoiceSettings { function VoiceSettingsProvider (line 15) | function VoiceSettingsProvider({ children }: { children: ReactNode }) { function useVoiceSettings (line 35) | function useVoiceSettings() { FILE: tts/tts-next-js-chat/src/app/hooks/useRecording.ts function useRecording (line 3) | function useRecording(onTranscribed: (text: string) => void) { FILE: tts/tts-next-js-chat/src/app/hooks/useTts.ts type AudioChunks (line 5) | type AudioChunks = Record; function useTts (line 7) | function useTts(options: { FILE: tts/tts-next-js-chat/src/app/hooks/useVoices.ts function useVoices (line 4) | function useVoices(provider: VoiceProvider) { FILE: tts/tts-next-js-chat/src/app/layout.tsx function RootLayout (line 21) | function RootLayout({ FILE: tts/tts-next-js-chat/src/app/page.tsx function Home (line 7) | function Home() { FILE: tts/tts-next-js-vercel-ai-sdk/src/actions/generate-speech.ts function tts (line 10) | async function tts(formData: FormData): Promise<{ FILE: tts/tts-next-js-vercel-ai-sdk/src/actions/list-voices.ts function listVoices (line 9) | async function listVoices(): Promise { FILE: tts/tts-next-js-vercel-ai-sdk/src/app/layout.tsx function RootLayout (line 21) | function RootLayout({ FILE: tts/tts-next-js-vercel-ai-sdk/src/app/page.tsx constant DEFAULT_VOICE_ID (line 11) | const DEFAULT_VOICE_ID = "9e068547-5ba4-4c8e-8e03-69282a008f04"; function Page (line 13) | function Page() { FILE: tts/tts-next-js-vercel-ai-sdk/src/components/AudioClipCard.tsx type ClipCardProps (line 5) | interface ClipCardProps { function AudioClipCard (line 12) | function AudioClipCard({ voiceName, text, instructions, url }: ClipCardP... FILE: tts/tts-next-js-vercel-ai-sdk/src/components/AudioGallery.tsx type AudioGalleryProps (line 8) | interface AudioGalleryProps { function AudioGallery (line 13) | function AudioGallery({ clips, voices }: AudioGalleryProps) { FILE: tts/tts-next-js-vercel-ai-sdk/src/components/TextAreaField.tsx type TextAreaFieldProps (line 5) | interface TextAreaFieldProps { function TextAreaField (line 15) | function TextAreaField({ FILE: tts/tts-next-js-vercel-ai-sdk/src/components/TtsForm.tsx type TtsFormProps (line 8) | interface TtsFormProps { function TtsForm (line 16) | function TtsForm({ FILE: tts/tts-next-js-vercel-ai-sdk/src/components/VoiceSelect.tsx type VoiceSelectProps (line 6) | interface VoiceSelectProps { function VoiceSelect (line 12) | function VoiceSelect({ voices, selectedVoiceId, onChange }: VoiceSelectP... FILE: tts/tts-next-js-vercel-ai-sdk/src/components/logos/Hume.tsx type HumeLogoProps (line 4) | type HumeLogoProps = SVGAttributes; function HumeLogo (line 6) | function HumeLogo(props: HumeLogoProps) { FILE: tts/tts-next-js-vercel-ai-sdk/src/hooks/useVoices.ts function useVoices (line 5) | function useVoices(defaultId: string) { FILE: tts/tts-next-js-vercel-ai-sdk/src/types/clip.ts type Clip (line 1) | interface Clip { FILE: tts/tts-python-livekit/src/agent_session/main.py class VoiceAssistant (line 18) | class VoiceAssistant(Agent): method __init__ (line 23) | def __init__(self): function entrypoint (line 27) | async def entrypoint(ctx: JobContext) -> None: FILE: tts/tts-python-livekit/src/standalone_tts/main.py function synthesize_text (line 14) | async def synthesize_text(text: str, session: ClientSession) -> bytes: function interactive_repl (line 37) | async def interactive_repl() -> None: FILE: tts/tts-python-livekit/src/utils.py function validate_env_vars (line 9) | def validate_env_vars(env_vars: list[str]) -> None: FILE: tts/tts-python-quickstart/app.py function example1 (line 45) | async def example1(): function example2 (line 60) | async def example2(): function example3 (line 135) | async def example3(): function main (line 164) | async def main(): FILE: tts/tts-python-quickstart/conftest.py function pytest_collection_modifyitems (line 11) | def pytest_collection_modifyitems(config, items): FILE: tts/tts-python-quickstart/test_app.py function create_audio_collector (line 22) | def create_audio_collector(): function assert_valid_audio_bytes (line 33) | def assert_valid_audio_bytes(chunks: list, *, min_chunks: int = 1): function assert_valid_audio_chunk (line 40) | def assert_valid_audio_chunk(chunk): function test_example1_runs_successfully (line 52) | async def test_example1_runs_successfully(): function test_example2_runs_successfully (line 67) | async def test_example2_runs_successfully(hume_client): function test_example3_runs_successfully (line 94) | async def test_example3_runs_successfully(): function api_key (line 119) | def api_key(): function hume_client (line 127) | def hume_client(api_key): function test_generates_json_with_octave_1 (line 132) | async def test_generates_json_with_octave_1(hume_client): function test_generates_json_with_octave_2_with_timestamps (line 151) | async def test_generates_json_with_octave_2_with_timestamps(hume_client): function test_creates_stream_and_connects_successfully (line 190) | async def test_creates_stream_and_connects_successfully(hume_client): function test_sends_messages_and_receives_audio_chunks (line 201) | async def test_sends_messages_and_receives_audio_chunks(hume_client): FILE: tts/tts-swift-quickstart/access_token_service/run_token_service.py function get_access_token (line 10) | def get_access_token(): FILE: tts/tts-typescript-lipsync/index.tsx function synthesize (line 28) | async function synthesize() { FILE: tts/tts-typescript-lipsync/mouth.ts type Point2D (line 1) | type Point2D = [number, number]; type MouthShape (line 2) | type MouthShape = Point2D[]; type PhonemeEvent (line 4) | type PhonemeEvent = { function mirror (line 11) | function mirror(leftSide: [number, number][]): MouthShape { type Viseme (line 31) | type Viseme = constant VISEME_SHAPES (line 48) | const VISEME_SHAPES: Record = { constant PHONEME_MAP (line 96) | const PHONEME_MAP: Array<[Viseme, string[]]> = [ function phonemeToViseme (line 114) | function phonemeToViseme(phoneme: string): Viseme { class Mouth (line 128) | class Mouth { method addPhoneme (line 134) | addPhoneme(phoneme: string, timestamp: number): void { method getShapeAt (line 149) | getShapeAt(time: number): MouthShape { method clearAfter (line 180) | clearAfter(timestamp: number): void { method reset (line 188) | reset(): void { function drawMouth (line 202) | function drawMouth(ctx: CanvasRenderingContext2D, shape: MouthShape, wid... class MouthAnimation (line 234) | class MouthAnimation { method constructor (line 243) | constructor(mouth: Mouth, width: number = 400, height: number = 300) { method start (line 259) | start(timestamp: number): void { method stop (line 265) | stop(): void { FILE: tts/tts-typescript-quickstart/audio_player.ts constant SAMPLE_RATE (line 3) | const SAMPLE_RATE = 48000; function startAudioPlayer (line 12) | function startAudioPlayer(mode: 'raw' | 'container' = 'container') { FILE: tts/tts-typescript-quickstart/index.test.ts function waitForStreamOpen (line 309) | function waitForStreamOpen(getStream: () => any): Promise { FILE: tts/tts-typescript-quickstart/vite-env.d.ts type ImportMetaEnv (line 3) | interface ImportMetaEnv { type ImportMeta (line 7) | interface ImportMeta { FILE: tts/tts-unity-quickstart/Assets/Scripts/HumeSpeaker.cs class HumeSpeaker (line 9) | [RequireComponent(typeof(AudioSource))] method SetApiKey (line 16) | public void SetApiKey(string key) method Speak (line 21) | public async void Speak() method ConvertBase64ToAudioClip (line 61) | private AudioClip ConvertBase64ToAudioClip(string base64Audio) method ConvertS16LEToFloats (line 82) | private float[] ConvertS16LEToFloats(byte[] bytes) FILE: tts/tts-unity-quickstart/Assets/Scripts/SceneBuilder.cs class SceneBuilder (line 3) | public class SceneBuilder : MonoBehaviour method Awake (line 8) | void Awake() method Start (line 13) | void Start() method BuildScene (line 18) | void BuildScene() class CubeSpinner (line 59) | public class CubeSpinner : MonoBehaviour method Start (line 65) | void Start() method Update (line 70) | void Update() class ClickToSpeak (line 83) | public class ClickToSpeak : MonoBehaviour method Start (line 87) | void Start() method OnMouseDown (line 92) | void OnMouseDown()