SYMBOL INDEX (91 symbols across 16 files) FILE: assets/gen_webm.py function run (line 11) | def run(cmd): function edit (line 16) | def edit(original, start, end, dest): function main (line 37) | def main(argv): FILE: compress_gpt/__init__.py function aclear_cache (line 42) | async def aclear_cache(): function clear_cache (line 46) | def clear_cache(): FILE: compress_gpt/compress.py class Compressor (line 34) | class Compressor: method __init__ (line 35) | def __init__( method _chunks (line 51) | async def _chunks(self, prompt: str, statics: str) -> list[Chunk]: method _static (line 61) | async def _static(self, prompt: str) -> list[StaticChunk]: method _decompress (line 71) | async def _decompress(self, prompt: str, statics: str) -> str: method _format (line 77) | async def _format(self, prompt: str) -> str: method _compare (line 83) | async def _compare( method _fix (line 98) | async def _fix( method _reconstruct (line 113) | def _reconstruct( method _extract_statics (line 149) | def _extract_statics(self, prompt: str, chunks: list[StaticChunk]) -> ... method _compress_segment (line 165) | async def _compress_segment(self, prompt: str, format: str, attempts: ... method _split_and_compress (line 202) | async def _split_and_compress( method _compress (line 218) | async def _compress(self, prompt: str, attempts: int) -> str: method acompress (line 242) | async def acompress(self, prompt: str, attempts: int = 3) -> str: method compress (line 250) | def compress(self, prompt: str, attempts: int = 3) -> str: FILE: compress_gpt/langchain/prompt.py class CompressMixin (line 9) | class CompressMixin(BaseModel): method _compress (line 12) | def _compress(self, prompt: str): class Config (line 15) | class Config: class CompressPrompt (line 20) | class CompressPrompt(CompressMixin, PromptTemplate): method format (line 21) | def format(self, **kwargs) -> str: class CompressTemplate (line 26) | class CompressTemplate(CompressMixin, PromptTemplate): method template (line 28) | def template(self): class CompressSimplePrompt (line 32) | class CompressSimplePrompt(CompressPrompt): class CompressSimpleTemplate (line 36) | class CompressSimpleTemplate(CompressTemplate): FILE: compress_gpt/prompts/__init__.py class Prompt (line 14) | class Prompt(ABC, Generic[M]): method get_prompt (line 17) | def get_prompt() -> ChatPromptTemplate: method get_format (line 21) | def get_format(cls) -> Type[M]: method get_chain (line 25) | def get_chain(cls, model: Optional[BaseLanguageModel]): method run (line 34) | async def run(cls, model: Optional[BaseLanguageModel] = None, **kwargs): class StrPrompt (line 39) | class StrPrompt(Prompt[str]): method get_chain (line 41) | def get_chain(cls, *args, **kwargs): FILE: compress_gpt/prompts/compare_prompts.py class PromptComparison (line 15) | class PromptComparison(BaseModel): class ComparePrompts (line 20) | class ComparePrompts(Prompt[PromptComparison]): method get_prompt (line 22) | def get_prompt() -> ChatPromptTemplate: FILE: compress_gpt/prompts/compress_chunks.py class Chunk (line 19) | class Chunk(BaseModel): class CompressChunks (line 25) | class CompressChunks(Prompt[list[Chunk]]): method get_prompt (line 27) | def get_prompt() -> ChatPromptTemplate: FILE: compress_gpt/prompts/decompress.py class Decompress (line 14) | class Decompress(StrPrompt): method get_prompt (line 16) | def get_prompt() -> ChatPromptTemplate: FILE: compress_gpt/prompts/diff_prompts.py class DiffPrompts (line 14) | class DiffPrompts(StrPrompt): method get_prompt (line 16) | def get_prompt() -> ChatPromptTemplate: FILE: compress_gpt/prompts/fix.py class FixPrompt (line 14) | class FixPrompt(Prompt[list[Chunk]]): method get_prompt (line 16) | def get_prompt() -> ChatPromptTemplate: FILE: compress_gpt/prompts/fix_json.py class FixJSON (line 14) | class FixJSON(StrPrompt): method get_prompt (line 16) | def get_prompt() -> ChatPromptTemplate: FILE: compress_gpt/prompts/identify_format.py class IdentifyFormat (line 16) | class IdentifyFormat(StrPrompt): method get_prompt (line 18) | def get_prompt() -> ChatPromptTemplate: FILE: compress_gpt/prompts/identify_static.py class StaticChunk (line 17) | class StaticChunk(BaseModel): class IdentifyStatic (line 22) | class IdentifyStatic(Prompt[list[StaticChunk]]): method get_prompt (line 24) | def get_prompt() -> ChatPromptTemplate: FILE: compress_gpt/prompts/output_parser.py class OutputParser (line 19) | class OutputParser(PydanticOutputParser, Generic[M]): method set_format (line 24) | def set_format(cls, _, values: dict) -> Type[BaseModel]: method set_pydantic_object (line 28) | def set_pydantic_object(cls, obj: M) -> Type[BaseModel]: method _preprocess (line 31) | def _preprocess(self, text: str) -> str: method _fix (line 41) | async def _fix(self, text: str, error: str) -> str: method aparse (line 46) | async def aparse( method parse (line 60) | def parse(self, text: str) -> Union[BaseModel, list[BaseModel]]: FILE: compress_gpt/tests/test_compress.py function compressor (line 24) | def compressor(): function simple_prompt (line 29) | def simple_prompt(): function complex_prompt (line 44) | def complex_prompt(): function test_prompt (line 109) | async def test_prompt(prompt: ChatPromptTemplate, **kwargs): function test_compress (line 116) | async def test_compress(compressor: Compressor): function test_compress_chunks (line 123) | async def test_compress_chunks(simple_prompt: str, compressor: Compressor): function test_langchain_integration (line 133) | async def test_langchain_integration(simple_prompt: str): function test_complex (line 151) | async def test_complex(complex_prompt: str, compressor: Compressor): function test_output (line 157) | async def test_output(complex_prompt: str, compressor: Compressor): FILE: compress_gpt/utils.py function has_redis (line 9) | def has_redis(): function identity (line 17) | def identity(x=None, *args): function wrap_prompt (line 21) | def wrap_prompt(name): function make_fast (line 26) | def make_fast(model: ChatOpenAI) -> ChatOpenAI: class CompressCallbackHandler (line 40) | class CompressCallbackHandler(BaseCallbackHandler): method __init__ (line 41) | def __init__(self): method on_llm_start (line 44) | def on_llm_start(self, serialized, prompts, **kwargs): method on_llm_end (line 50) | def on_llm_end(self, response, **kwargs): method on_llm_new_token (line 53) | def on_llm_new_token(self, token, **kwargs): method on_llm_error (line 57) | def on_llm_error(self, error, **kwargs): method on_chain_start (line 60) | def on_chain_start(self, serialized, inputs, **kwargs): method on_chain_end (line 63) | def on_chain_end(self, outputs, **kwargs): method on_chain_error (line 66) | def on_chain_error(self, error, **kwargs): method on_tool_start (line 69) | def on_tool_start(self, serialized, input_str, **kwargs): method on_agent_action (line 72) | def on_agent_action(self, action, **kwargs): method on_tool_end (line 75) | def on_tool_end(self, output, **kwargs): method on_tool_error (line 78) | def on_tool_error(self, error, **kwargs): method on_text (line 81) | def on_text(self, text, end="", **kwargs): method on_agent_finish (line 84) | def on_agent_finish(self, finish, **kwargs): method flush_tracker (line 87) | def flush_tracker(self, **kwargs):