SYMBOL INDEX (107 symbols across 15 files) FILE: chunkers/base_chunker.py class TextSplitter (line 12) | class TextSplitter: method __init__ (line 17) | def __init__( method split_text (line 51) | def split_text(self, text: str) -> List[str]: method create_documents (line 54) | def create_documents( method split_documents (line 74) | def split_documents(self, documents: Iterable[Document]) -> List[Docum... method _join_docs (line 82) | def _join_docs(self, docs: List[str], separator: str) -> Optional[str]: method _merge_splits (line 91) | def _merge_splits(self, splits: Iterable[str], separator: str) -> List... method transform_documents (line 133) | def transform_documents( FILE: chunkers/character_chunker.py class RecursiveCharacterTextSplitter (line 10) | class RecursiveCharacterTextSplitter(TextSplitter): method __init__ (line 19) | def __init__(self, chunk_size: int = 4000, chunk_overlap: int = 200, l... method _split_text (line 34) | def _split_text(self, text: str, separators: List[str]) -> List[str]: method split_text (line 74) | def split_text(self, text: str) -> List[str]: function _split_text_with_regex (line 78) | def _split_text_with_regex( FILE: chunkers/ner_chunker.py function batchify (line 15) | def batchify(lst, batch_size): class Token (line 31) | class Token: class TokenClassificationChunker (line 39) | class TokenClassificationChunker(TextSplitter): method __init__ (line 40) | def __init__(self, model_id="mirth/chonky_distilbert_base_uncased_1", ... method split_into_semantic_chunks (line 74) | def split_into_semantic_chunks(self, text, separator_indices: List[int]): method split_text (line 88) | def split_text(self, text: str) -> List[str]: FILE: chunkers/semantic_chunker.py function calculate_cosine_distances (line 17) | def calculate_cosine_distances(sentence_embeddings) -> np.array: class BoundedSemanticChunker (line 45) | class BoundedSemanticChunker(TextSplitter): method __init__ (line 52) | def __init__(self, embedding_model: SentenceTransformer, buffer_size: ... method _calculate_sentence_distances (line 76) | def _calculate_sentence_distances( method _calculate_breakpoint_threshold (line 84) | def _calculate_breakpoint_threshold(self, distances: np.array, alt_bre... method _threshold_from_clusters (line 111) | def _threshold_from_clusters(self, distances: List[float]) -> float: method split_text (line 131) | def split_text( FILE: force_search_box_theme.js function toggleForceSearchDarkMode (line 1) | function toggleForceSearchDarkMode() { FILE: llm_web_search.py function perform_web_search (line 19) | def perform_web_search(query, max_results=3, timeout=10): function retrieve_from_duckduckgo (line 52) | def retrieve_from_duckduckgo(query: str, document_retriever: DocumentRet... function retrieve_from_searxng (line 82) | def retrieve_from_searxng(query: str, url: str, document_retriever: Docu... function get_webpage_content (line 134) | def get_webpage_content(url: str) -> str: FILE: retrieval.py class DocumentRetriever (line 35) | class DocumentRetriever: method __init__ (line 37) | def __init__(self, device="cuda", num_results: int = 5, similarity_thr... method preprocess_text (line 81) | def preprocess_text(self, text: str) -> str: method retrieve_from_snippets (line 87) | def retrieve_from_snippets(self, query: str, documents: list[Document]... method retrieve_from_webpages (line 94) | def retrieve_from_webpages(self, query: str, url_list: list[str]) -> l... function async_download_html (line 174) | async def async_download_html(url: str, headers: Dict, timeout: int): function async_fetch_chunk_websites (line 190) | async def async_fetch_chunk_websites(urls: List[str], function docs_to_pretty_str (line 208) | def docs_to_pretty_str(docs) -> str: function download_html (line 217) | def download_html(url: str) -> bytes: function html_to_plaintext_doc (line 231) | def html_to_plaintext_doc(html_text: str or bytes, url: str) -> Document: function weighted_reciprocal_rank (line 242) | def weighted_reciprocal_rank(doc_lists: List[List[Document]], weights: L... function unique_by_key (line 282) | def unique_by_key(iterable: Iterable, key: Callable) -> Iterator: FILE: retrievers/bm25_retriever.py function default_preprocessing_func (line 11) | def default_preprocessing_func(text: str) -> List[str]: class BM25Retriever (line 15) | class BM25Retriever: method __init__ (line 27) | def __init__(self, vectorizer: Any, docs: List[Document], k: int = 4, method from_texts (line 35) | def from_texts( method from_documents (line 65) | def from_documents( method get_relevant_documents (line 93) | def get_relevant_documents(self, query: str) -> List[Document]: FILE: retrievers/faiss_retriever.py class FaissRetriever (line 12) | class FaissRetriever: method __init__ (line 14) | def __init__(self, embedding_model: MySentenceTransformer, num_results... method add_documents (line 23) | def add_documents(self, documents: List[Document]): method get_relevant_documents (line 32) | def get_relevant_documents(self, query: str) -> List[Document]: FILE: retrievers/splade_retriever.py function neg_dot_dist (line 19) | def neg_dot_dist(x, y): class SpladeRetriever (line 26) | class SpladeRetriever: method __init__ (line 27) | def __init__(self, splade_doc_tokenizer, splade_doc_model, splade_quer... method compute_document_vectors (line 41) | def compute_document_vectors(self, texts: List[str], batch_size: int) ... method compute_query_vector (line 73) | def compute_query_vector(self, text: str): method add_documents (line 91) | def add_documents(self, documents: List[Document])-> List[str]: method add_texts (line 106) | def add_texts( method get_relevant_documents (line 125) | def get_relevant_documents(self, query: str) -> List[Document]: FILE: script.js function set_margins (line 7) | function set_margins(generate_button, stop_button, chat_input, reset=fal... FILE: script.py function patched_is_path_allowed (line 28) | def patched_is_path_allowed(abs_path_str): function setup (line 83) | def setup(): function save_settings (line 104) | def save_settings(): function toggle_extension (line 112) | def toggle_extension(_enable: bool): function get_available_system_prompts (line 144) | def get_available_system_prompts(): function load_system_prompt (line 151) | def load_system_prompt(filename: str or None): function save_system_prompt (line 169) | def save_system_prompt(filename, prompt): function check_file_exists (line 180) | def check_file_exists(filename): function timeout_save_message (line 188) | def timeout_save_message(): function deactivate_system_prompt (line 193) | def deactivate_system_prompt(): function toggle_forced_search (line 198) | def toggle_forced_search(value): function update_chat_id (line 203) | def update_chat_id(_id): function clear_update_history_dict (line 208) | def clear_update_history_dict(): function get_force_search_box_theme_js (line 212) | def get_force_search_box_theme_js(): function create_tool_symlink (line 217) | def create_tool_symlink(): function remove_tool_symlink (line 227) | def remove_tool_symlink(): function handle_add_as_tool (line 233) | def handle_add_as_tool(checked): function ui (line 241) | def ui(): function get_generation_prompt (line 492) | def get_generation_prompt(state, impersonate=False, enable_thinking=True): function custom_generate_reply (line 556) | def custom_generate_reply(question, original_question, state, stopping_s... function output_modifier (line 730) | def output_modifier(string, state, is_chat=False): function custom_css (line 743) | def custom_css(): function custom_js (line 751) | def custom_js(): function chat_input_modifier (line 763) | def chat_input_modifier(text, visible_text, state): function state_modifier (line 775) | def state_modifier(state): function history_modifier (line 785) | def history_modifier(history): FILE: test_basics.py class MyTestCase (line 7) | class MyTestCase(unittest.TestCase): method setUp (line 9) | def setUp(self): method test_basic_search (line 14) | def test_basic_search(self): FILE: tool.py function docs_to_dicts (line 38) | def docs_to_dicts(docs): function execute (line 47) | def execute(arguments): FILE: utils.py class Document (line 17) | class Document: class Generator (line 22) | class Generator: method __init__ (line 25) | def __init__(self, gen): method __iter__ (line 28) | def __iter__(self): function cosine_similarity (line 33) | def cosine_similarity(X, Y) -> np.ndarray: class SimilarLengthsBatchifyer (line 54) | class SimilarLengthsBatchifyer: method __init__ (line 59) | def __init__(self, batch_size, inputs, max_padding_len=10): method __len__ (line 102) | def __len__(self): method __iter__ (line 105) | def __iter__(self): class MySentenceTransformer (line 125) | class MySentenceTransformer(SentenceTransformer): method batch_encode (line 126) | def batch_encode( function sync_device (line 299) | def sync_device(device: torch.device): function filter_similar_embeddings (line 312) | def filter_similar_embeddings( function bow_filter_similar_texts (line 336) | def bow_filter_similar_texts(texts: List[str], similarity_fn: Callable, ...