SYMBOL INDEX (647 symbols across 83 files) FILE: .circleci/grab_notary_id.py function main (line 6) | def main() -> None: FILE: gpt4all-backend/include/gpt4all-backend/llmodel.h function class (line 25) | class LLModel { function class (line 46) | class MissingImplementationError: public std::runtime_error { function class (line 51) | class UnsupportedModelError: public std::runtime_error { function backendIdToName (line 56) | struct GPUDevice { function std (line 78) | static std::string updateSelectionName(const std::string &name) { function class (line 95) | class Implementation { type PromptContext (line 132) | struct PromptContext { function explicit (line 144) | explicit LLModel() {} function virtual (line 145) | virtual ~LLModel() {} function virtual (line 147) | virtual bool supportsEmbedding() const = 0; function virtual (line 151) | virtual bool isEmbeddingModel(const std::string &modelPath) const { (voi... function virtual (line 152) | virtual bool isModelLoaded() const = 0; function virtual (line 185) | virtual std::vector availableGPUDevices(size_t memoryRequired... function virtual (line 190) | virtual bool initializeGPUDevice(size_t memoryRequired, const std::strin... function virtual (line 205) | virtual const char *backendName() const { return "cpu"; } function virtual (line 206) | virtual const char *gpuDeviceName() const { return nullptr; } function setProgressCallback (line 208) | void setProgressCallback(ProgressCallback callback) { m_progressCallback... function virtual (line 210) | virtual int32_t contextLength() const = 0; function virtual (line 237) | virtual int32_t layerCount(std::string const &modelPath) const function virtual (line 243) | virtual auto chatTemplate(const char *modelPath) const -> std::expected<... function staticProgressCallback (line 252) | static bool staticProgressCallback(float progress, void* ctx) FILE: gpt4all-backend/include/gpt4all-backend/llmodel_c.h type token_t (line 29) | typedef int32_t token_t; type llmodel_prompt_context (line 37) | struct llmodel_prompt_context { type llmodel_gpu_device (line 49) | struct llmodel_gpu_device { type llmodel_prompt_context (line 59) | typedef struct llmodel_prompt_context llmodel_prompt_context; type llmodel_gpu_device (line 60) | typedef struct llmodel_gpu_device llmodel_gpu_device; type llmodel_gpu_device (line 266) | struct llmodel_gpu_device FILE: gpt4all-backend/include/gpt4all-backend/sysinfo.h function getSystemTotalRAMInBytes (line 22) | static long long getSystemTotalRAMInBytes() function getSystemTotalRAMInGB (line 53) | static double getSystemTotalRAMInGB() function std (line 58) | static std::string getSystemTotalRAMInGBString() FILE: gpt4all-backend/src/dlhandle.h function class (line 11) | class Dlhandle { FILE: gpt4all-backend/src/llamamodel.cpp function is_embedding_arch (line 99) | static bool is_embedding_arch(const std::string &arch) function llama_verbose (line 104) | static bool llama_verbose() function llama_log_callback (line 110) | static void llama_log_callback(ggml_log_level level, const char *text, v... type gpt_params (line 137) | struct gpt_params { type ggml_type (line 146) | enum ggml_type type gguf_type (line 158) | enum gguf_type function gguf_context (line 165) | static gguf_context *load_gguf(const char *fname) function get_arch_key_u32 (line 187) | static int32_t get_arch_key_u32(std::string const &modelPath, std::strin... type LLamaPrivate (line 217) | struct LLamaPrivate { type llama_file_hparams (line 241) | struct llama_file_hparams { type llama_ftype (line 248) | enum llama_ftype type llama_sampler (line 585) | struct llama_sampler type gguf_type (line 738) | enum gguf_type function llama_batch_add (line 921) | void llama_batch_add( function batch_add_seq (line 938) | static void batch_add_seq(llama_batch &batch, const std::vector std::function function getL2NormScale (line 1072) | double getL2NormScale(T *start, T *end) type split_batch (line 1157) | struct split_batch { unsigned idx; TokenString batch; } type llama_batch (line 1194) | struct llama_batch function DLL_EXPORT (line 1284) | DLL_EXPORT bool is_g4a_backend_model_implementation() function DLL_EXPORT (line 1289) | DLL_EXPORT const char *get_model_type() function DLL_EXPORT (line 1294) | DLL_EXPORT const char *get_build_variant() function DLL_EXPORT (line 1299) | DLL_EXPORT char *get_file_arch(const char *fname) function DLL_EXPORT (line 1325) | DLL_EXPORT bool is_arch_supported(const char *arch) function DLL_EXPORT (line 1330) | DLL_EXPORT LLModel *construct() FILE: gpt4all-backend/src/llamamodel_impl.h type LLamaPrivate (line 16) | struct LLamaPrivate type EmbModelSpec (line 17) | struct EmbModelSpec function class (line 19) | class LLamaModel : public LLModel { FILE: gpt4all-backend/src/llmodel.cpp function get_cpu_info (line 53) | static int get_cpu_info(int func_id, int reg_id) { function isImplementation (line 100) | static bool isImplementation(const Dlhandle &dl) function addCudaSearchPath (line 107) | static void addCudaSearchPath() function applyCPUVariant (line 181) | static std::string applyCPUVariant(const std::string &buildVariant) function LLModel (line 214) | LLModel *LLModel::Implementation::construct(const std::string &modelPath... function LLModel (line 255) | LLModel *LLModel::Implementation::constructGlobalLlama(const std::option... FILE: gpt4all-backend/src/llmodel_c.cpp type LLModelWrapper (line 22) | struct LLModelWrapper { function llmodel_model (line 27) | llmodel_model llmodel_model_create(const char *model_path) function llmodel_set_error (line 37) | static void llmodel_set_error(const char **errptr, const char *message) function llmodel_model (line 46) | llmodel_model llmodel_model_create2(const char *model_path, const char *... function llmodel_model_destroy (line 61) | void llmodel_model_destroy(llmodel_model model) function llmodel_required_mem (line 66) | size_t llmodel_required_mem(llmodel_model model, const char *model_path,... function llmodel_loadModel (line 72) | bool llmodel_loadModel(llmodel_model model, const char *model_path, int ... function llmodel_isModelLoaded (line 85) | bool llmodel_isModelLoaded(llmodel_model model) function llmodel_state_get_size (line 91) | uint64_t llmodel_state_get_size(llmodel_model model) function llmodel_state_get_data (line 97) | uint64_t llmodel_state_get_data(llmodel_model model, uint8_t *state_out,... function llmodel_state_free_input_tokens (line 115) | void llmodel_state_free_input_tokens(LLModel::Token *input_tokens) function llmodel_state_set_data (line 120) | uint64_t llmodel_state_set_data(llmodel_model model, const uint8_t *stat... function llmodel_prompt (line 127) | bool llmodel_prompt(llmodel_model model, function llmodel_free_embedding (line 205) | void llmodel_free_embedding(float *ptr) function llmodel_setThreadCount (line 210) | void llmodel_setThreadCount(llmodel_model model, int32_t n_threads) function llmodel_threadCount (line 216) | int32_t llmodel_threadCount(llmodel_model model) function llmodel_set_implementation_search_path (line 222) | void llmodel_set_implementation_search_path(const char *path) type llmodel_gpu_device_cpp (line 233) | struct llmodel_gpu_device_cpp: llmodel_gpu_device { method llmodel_gpu_device_cpp (line 234) | llmodel_gpu_device_cpp() = default; method llmodel_gpu_device_cpp (line 236) | llmodel_gpu_device_cpp(const llmodel_gpu_device_cpp &) = delete; method llmodel_gpu_device_cpp (line 237) | llmodel_gpu_device_cpp( llmodel_gpu_device_cpp &&) = delete; method llmodel_gpu_device_cpp (line 239) | const llmodel_gpu_device_cpp &operator=(const llmodel_gpu_device_cpp ... method llmodel_gpu_device_cpp (line 240) | llmodel_gpu_device_cpp &operator=( llmodel_gpu_device_cpp &&) = d... type llmodel_gpu_device (line 250) | struct llmodel_gpu_device function llmodel_gpu_init_gpu_device_by_string (line 274) | bool llmodel_gpu_init_gpu_device_by_string(llmodel_model model, size_t m... function llmodel_gpu_init_gpu_device_by_struct (line 280) | bool llmodel_gpu_init_gpu_device_by_struct(llmodel_model model, const ll... function llmodel_gpu_init_gpu_device_by_int (line 286) | bool llmodel_gpu_init_gpu_device_by_int(llmodel_model model, int device) function llmodel_count_prompt_tokens (line 304) | int32_t llmodel_count_prompt_tokens(llmodel_model model, const char *pro... function llmodel_model_foreach_special_token (line 315) | void llmodel_model_foreach_special_token(llmodel_model model, llmodel_sp... FILE: gpt4all-backend/src/llmodel_shared.cpp function stringsOverlap (line 130) | static std::string::size_type stringsOverlap(const std::string &s, const... FILE: gpt4all-bindings/cli/app.py function repl (line 53) | def repl( function _old_loop (line 99) | def _old_loop(gpt4all_instance): function _new_loop (line 135) | def _new_loop(gpt4all_instance): function version (line 178) | def version(): FILE: gpt4all-bindings/python/gpt4all/_pyllmodel.py function _operator_call (line 36) | def _operator_call(obj: Callable[P, T], /, *args: P.args, **kwargs: P.kw... function check_rosetta (line 42) | def check_rosetta() -> None: function find_cuda (line 67) | def find_cuda() -> None: function load_llmodel_library (line 101) | def load_llmodel_library(): class LLModelPromptContext (line 119) | class LLModelPromptContext(ctypes.Structure): class LLModelGPUDevice (line 133) | class LLModelGPUDevice(ctypes.Structure): function empty_response_callback (line 235) | def empty_response_callback(token_id: int, response: str) -> bool: class Sentinel (line 240) | class Sentinel(Enum): class EmbedResult (line 244) | class EmbedResult(Generic[EmbeddingsType], TypedDict): class CancellationError (line 249) | class CancellationError(Exception): class LLModel (line 253) | class LLModel: method __init__ (line 270) | def __init__(self, model_path: str, n_ctx: int, ngl: int, backend: str): method __del__ (line 298) | def __del__(self, llmodel=llmodel): method close (line 302) | def close(self) -> None: method _raise_closed (line 307) | def _raise_closed(self) -> NoReturn: method backend (line 311) | def backend(self) -> Literal["cpu", "kompute", "cuda", "metal"]: method device (line 317) | def device(self) -> str | None: method count_prompt_tokens (line 323) | def count_prompt_tokens(self, prompt: str) -> int: method list_gpus (line 337) | def list_gpus(mem_required: int = 0) -> list[str]: method init_gpu (line 353) | def init_gpu(self, device: str): method load_model (line 372) | def load_model(self) -> bool: method set_thread_count (line 385) | def set_thread_count(self, n_threads): method thread_count (line 392) | def thread_count(self): method generate_embeddings (line 400) | def generate_embeddings( method generate_embeddings (line 405) | def generate_embeddings( method generate_embeddings (line 410) | def generate_embeddings( method generate_embeddings (line 415) | def generate_embeddings( method prompt_model (line 466) | def prompt_model( method prompt_model_streaming (line 531) | def prompt_model_streaming( method _callback_decoder (line 572) | def _callback_decoder(self, callback: ResponseCallbackType) -> RawResp... method _prompt_callback (line 615) | def _prompt_callback(token_ids: ctypes._Pointer[ctypes.c_int32], n_tok... FILE: gpt4all-bindings/python/gpt4all/gpt4all.py function _jinja_env (line 42) | def _jinja_env() -> ImmutableSandboxedEnvironment: class MessageType (line 59) | class MessageType(TypedDict): class ChatSession (line 64) | class ChatSession(NamedTuple): class Embed4All (line 69) | class Embed4All: method __init__ (line 76) | def __init__(self, model_name: str | None = None, *, n_threads: int | ... method __enter__ (line 89) | def __enter__(self) -> Self: method __exit__ (line 92) | def __exit__( method close (line 97) | def close(self) -> None: method embed (line 103) | def embed( method embed (line 108) | def embed( method embed (line 113) | def embed( method embed (line 121) | def embed( method embed (line 126) | def embed( method embed (line 131) | def embed( method embed (line 139) | def embed( method embed (line 145) | def embed( class GPT4All (line 191) | class GPT4All: method __init__ (line 196) | def __init__( method __enter__ (line 271) | def __enter__(self) -> Self: method __exit__ (line 274) | def __exit__( method close (line 279) | def close(self) -> None: method backend (line 284) | def backend(self) -> Literal["cpu", "kompute", "cuda", "metal"]: method device (line 289) | def device(self) -> str | None: method current_chat_session (line 294) | def current_chat_session(self) -> list[MessageType] | None: method current_chat_session (line 298) | def current_chat_session(self, history: list[MessageType]) -> None: method list_models (line 304) | def list_models() -> list[ConfigType]: method retrieve_model (line 317) | def retrieve_model( method download_model (line 378) | def download_model( method generate (line 494) | def generate( method generate (line 500) | def generate( method generate (line 506) | def generate( method generate (line 512) | def generate( method chat_session (line 602) | def chat_session( method list_gpus (line 641) | def list_gpus() -> list[str]: function append_extension_if_missing (line 651) | def append_extension_if_missing(model_name): class _HasFileno (line 657) | class _HasFileno(Protocol): method fileno (line 658) | def fileno(self) -> int: ... function _fsync (line 661) | def _fsync(fd: int | _HasFileno) -> None: function _remove_prefix (line 673) | def _remove_prefix(s: str, prefix: str) -> str: FILE: gpt4all-bindings/python/gpt4all/tests/test_embed_timings.py function time_embedding (line 9) | def time_embedding(i, embedder): FILE: gpt4all-bindings/python/gpt4all/tests/test_gpt4all.py function test_inference (line 10) | def test_inference(): function do_long_input (line 36) | def do_long_input(model): function test_inference_long_orca_3b (line 45) | def test_inference_long_orca_3b(): function test_inference_long_falcon (line 50) | def test_inference_long_falcon(): function test_inference_long_llama_7b (line 55) | def test_inference_long_llama_7b(): function test_inference_long_llama_13b (line 60) | def test_inference_long_llama_13b(): function test_inference_long_mpt (line 65) | def test_inference_long_mpt(): function test_inference_long_replit (line 70) | def test_inference_long_replit(): function test_inference_hparams (line 75) | def test_inference_hparams(): function test_inference_falcon (line 82) | def test_inference_falcon(): function test_inference_mpt (line 90) | def test_inference_mpt(): function test_embedding (line 98) | def test_embedding(): function test_empty_embedding (line 107) | def test_empty_embedding(): function test_download_model (line 113) | def test_download_model(tmp_path: Path): FILE: gpt4all-bindings/python/setup.py function get_c_shared_lib_extension (line 20) | def get_c_shared_lib_extension(): function copy_prebuilt_C_lib (line 33) | def copy_prebuilt_C_lib(src_dir, dest_dir, dest_build_dir): function get_long_description (line 64) | def get_long_description(): FILE: gpt4all-bindings/typescript/index.cc function ChunkedFloatPtr (line 194) | Napi::Array ChunkedFloatPtr(float *embedding_ptr, int embedding_size, in... function llmodel_model (line 440) | llmodel_model NodeModelWrapper::GetInference() function Init (line 446) | Napi::Object Init(Napi::Env env, Napi::Object exports) FILE: gpt4all-bindings/typescript/index.h function class (line 14) | class NodeModelWrapper : public Napi::ObjectWrap FILE: gpt4all-bindings/typescript/prompt.h type ResponseCallbackData (line 13) | struct ResponseCallbackData type PromptCallbackData (line 19) | struct PromptCallbackData type LLModelWrapper (line 24) | struct LLModelWrapper type PromptWorkerConfig (line 34) | struct PromptWorkerConfig function class (line 50) | class PromptWorker : public Napi::AsyncWorker FILE: gpt4all-bindings/typescript/scripts/prebuild.js function createPrebuilds (line 3) | async function createPrebuilds(combinations) { function createPrebuild (line 25) | function createPrebuild(opts) { FILE: gpt4all-bindings/typescript/spec/llmodel.mjs class Extended (line 16) | class Extended extends LLModel {} FILE: gpt4all-bindings/typescript/src/chat-session.js class ChatSession (line 4) | class ChatSession { method constructor (line 24) | constructor(model, chatSessionOpts = {}) { method initialize (line 39) | async initialize(completionOpts = {}) { method ingestMessages (line 73) | async ingestMessages(messages, completionOpts = {}) { method generate (line 91) | async generate(input, completionOpts = {}) { FILE: gpt4all-bindings/typescript/src/config.js constant DEFAULT_DIRECTORY (line 4) | const DEFAULT_DIRECTORY = path.resolve(os.homedir(), ".cache/gpt4all"); constant DEFAULT_LIBRARIES_DIRECTORY (line 23) | const DEFAULT_LIBRARIES_DIRECTORY = librarySearchPaths.join(";"); constant DEFAULT_MODEL_CONFIG (line 25) | const DEFAULT_MODEL_CONFIG = { constant DEFAULT_MODEL_LIST_URL (line 30) | const DEFAULT_MODEL_LIST_URL = "https://gpt4all.io/models/models3.json"; constant DEFAULT_PROMPT_CONTEXT (line 32) | const DEFAULT_PROMPT_CONTEXT = { FILE: gpt4all-bindings/typescript/src/gpt4all.d.ts type LLModelOptions (line 4) | interface LLModelOptions { type ModelConfig (line 14) | interface ModelConfig { type ChatSessionOptions (line 24) | interface ChatSessionOptions extends Partial { class ChatSession (line 39) | class ChatSession implements CompletionProvider { type InferenceResult (line 93) | interface InferenceResult extends LLModelInferenceResult { class InferenceModel (line 101) | class InferenceModel implements CompletionProvider { type EmbedddingOptions (line 142) | interface EmbedddingOptions { type EmbeddingResult (line 198) | interface EmbeddingResult { class EmbeddingModel (line 209) | class EmbeddingModel { type LLModelInferenceResult (line 258) | interface LLModelInferenceResult { type LLModelInferenceOptions (line 263) | interface LLModelInferenceOptions extends Partial { class LLModel (line 281) | class LLModel { type GpuDevice (line 408) | interface GpuDevice { type LoadModelOptions (line 422) | interface LoadModelOptions { type InferenceModelOptions (line 467) | interface InferenceModelOptions extends LoadModelOptions { type EmbeddingModelOptions (line 471) | interface EmbeddingModelOptions extends LoadModelOptions { type CompletionProvider (line 502) | interface CompletionProvider { type CompletionOptions (line 513) | interface CompletionOptions extends LLModelInferenceOptions { type CompletionInput (line 524) | type CompletionInput = string | ChatMessage[]; type CompletionStreamReturn (line 555) | interface CompletionStreamReturn { type ChatMessage (line 576) | interface ChatMessage { type CompletionResult (line 587) | interface CompletionResult { type LLModelPromptContext (line 615) | interface LLModelPromptContext { type DownloadModelOptions (line 770) | interface DownloadModelOptions { type ListModelsOptions (line 795) | interface ListModelsOptions { type RetrieveModelOptions (line 804) | interface RetrieveModelOptions { type DownloadController (line 819) | interface DownloadController { FILE: gpt4all-bindings/typescript/src/gpt4all.js function loadModel (line 33) | async function loadModel(modelName, options = {}) { function createEmbedding (line 87) | function createEmbedding(model, text, options={}) { function createCompletion (line 131) | async function createCompletion( function createCompletionStream (line 166) | function createCompletionStream( FILE: gpt4all-bindings/typescript/src/models.js class InferenceModel (line 5) | class InferenceModel { method constructor (line 11) | constructor(llmodel, config) { method createChatSession (line 17) | async createChatSession(options) { method generate (line 24) | async generate(input, options = DEFAULT_PROMPT_CONTEXT) { method dispose (line 139) | dispose() { class EmbeddingModel (line 144) | class EmbeddingModel { method constructor (line 148) | constructor(llmodel, config) { method embed (line 153) | embed(text, prefix, dimensionality, do_mean, atlas) { method dispose (line 157) | dispose() { FILE: gpt4all-bindings/typescript/src/util.js function listModels (line 12) | async function listModels( function appendBinSuffixIfMissing (line 44) | function appendBinSuffixIfMissing(name) { function prepareMessagesForIngest (line 52) | function prepareMessagesForIngest(messages) { function readChunks (line 111) | function readChunks(reader) { function downloadModel (line 123) | function downloadModel(modelName, options = {}) { function retrieveModel (line 243) | async function retrieveModel(modelName, options = {}) { FILE: gpt4all-bindings/typescript/test/gpt4all.test.js method start (line 109) | start(controller) { FILE: gpt4all-chat/cmake/sign_dmg.py function sign_dmg (line 25) | def sign_dmg(input_dmg: str, output_dmg: str, signing_identity: Optional... FILE: gpt4all-chat/src/chat.cpp function QVariant (line 176) | QVariant Chat::popPrompt(int index) function ModelInfo (line 320) | ModelInfo Chat::modelInfo() const function QString (line 415) | QString Chat::deviceBackend() const function QString (line 420) | QString Chat::device() const function QString (line 425) | QString Chat::fallbackReason() const FILE: gpt4all-chat/src/chat.h function class (line 26) | class Chat : public QObject FILE: gpt4all-chat/src/chatapi.cpp function parsePrompt (line 81) | static auto parsePrompt(QXmlStreamReader &xml) -> std::expected stateOut, std::vector &inputT... function restoreState (line 70) | size_t restoreState(std::span state, std::span loadJinjaTemplate(const std:... class ChatViewResponseHandler (line 958) | class ChatViewResponseHandler : public BaseResponseHandler { method ChatViewResponseHandler (line 960) | ChatViewResponseHandler(ChatLLM *cllm, QElapsedTimer *totalTime, ChatL... method onSplitIntoTwo (line 963) | void onSplitIntoTwo(const QString &startTag, const QString &firstBuffe... method onSplitIntoThree (line 971) | void onSplitIntoThree(const QString &secondBuffer, const QString &thir... method onOldResponseChunk (line 976) | void onOldResponseChunk(const QByteArray &chunk) override method onBufferResponse (line 983) | bool onBufferResponse(const QString &response, int bufferIdx) override method onRegularResponse (line 1000) | bool onRegularResponse() override method getStopGenerating (line 1006) | bool getStopGenerating() const override class SimpleResponseHandler (line 1176) | class SimpleResponseHandler : public BaseResponseHandler { method SimpleResponseHandler (line 1178) | SimpleResponseHandler(ChatLLM *cllm) method onSplitIntoTwo (line 1181) | void onSplitIntoTwo(const QString &startTag, const QString &firstBuffe... method onSplitIntoThree (line 1184) | void onSplitIntoThree(const QString &secondBuffer, const QString &thir... method onOldResponseChunk (line 1187) | void onOldResponseChunk(const QByteArray &chunk) override method onBufferResponse (line 1190) | bool onBufferResponse(const QString &response, int bufferIdx) override method onRegularResponse (line 1197) | bool onRegularResponse() override method getStopGenerating (line 1200) | bool getStopGenerating() const override class NameResponseHandler (line 1211) | class NameResponseHandler : public SimpleResponseHandler { method onSimpleResponse (line 1220) | bool onSimpleResponse(const QString &response) override class QuestionResponseHandler (line 1271) | class QuestionResponseHandler : public SimpleResponseHandler { method onSimpleResponse (line 1276) | bool onSimpleResponse(const QString &response) override FILE: gpt4all-chat/src/chatllm.h type class (line 38) | enum class type class (line 48) | enum class function LLModelTypeV1 (line 61) | inline LLModelTypeV1 parseLLModelTypeV1(int type) function LLModelTypeV1 (line 78) | inline LLModelTypeV1 parseLLModelTypeV0(int v0) type LLModelInfo (line 93) | struct LLModelInfo { function class (line 104) | class TokenTimer : public QObject { function class (line 145) | class ChatLLM : public QObject FILE: gpt4all-chat/src/chatmodel.h type PromptAttachment (line 50) | struct PromptAttachment { function MessageInput (line 87) | Q_DECLARE_METATYPE(PromptAttachment) function std (line 569) | static std::optional getPeerInternal(const MessageItem *arr, ... function const_iterator (line 593) | auto getPeerUnlocked(QList::const_iterator item) const -> st... function appendResponse (line 691) | void appendResponse() function qsizetype (line 712) | qsizetype appendResponseWithHistory(std::span history) function hasError (line 1156) | bool hasError() const { QMutexLocker locker(&m_mutex); return hasErrorUn... function deserialize (line 1189) | bool deserialize(QDataStream &stream, int version) FILE: gpt4all-chat/src/chatviewtextprocessor.cpp type Language (line 31) | enum Language { function Language (line 47) | static Language stringToLanguage(const QString &language) type Code (line 84) | enum Code { type HighlightingRule (line 107) | struct HighlightingRule { function QColor (line 112) | static QColor formatToColor(Code c, const CodeColors &colors) function pythonHighlightingRules (line 140) | static QVector pythonHighlightingRules() function csharpHighlightingRules (line 192) | static QVector csharpHighlightingRules() function cppHighlightingRules (line 252) | static QVector cppHighlightingRules() function typescriptHighlightingRules (line 320) | static QVector typescriptHighlightingRules() function javaHighlightingRules (line 392) | static QVector javaHighlightingRules() function goHighlightingRules (line 453) | static QVector goHighlightingRules() function bashHighlightingRules (line 510) | static QVector bashHighlightingRules() function latexHighlightingRules (line 569) | static QVector latexHighlightingRules() function htmlHighlightingRules (line 591) | static QVector htmlHighlightingRules() function phpHighlightingRules (line 625) | static QVector phpHighlightingRules() function jsonHighlightingRules (line 684) | static QVector jsonHighlightingRules() function QQuickTextDocument (line 784) | QQuickTextDocument* ChatViewTextProcessor::textDocument() const function qreal (line 828) | qreal ChatViewTextProcessor::fontPixelSize() const function CodeColors (line 842) | CodeColors ChatViewTextProcessor::codeColors() const function traverseDocument (line 853) | void traverseDocument(QTextDocument *doc, QTextFrame *frame) function replaceAndInsertMarkdown (line 1066) | void replaceAndInsertMarkdown(int startIndex, int endIndex, QTextDocumen... FILE: gpt4all-chat/src/chatviewtextprocessor.h type CodeColors (line 16) | struct CodeColors { function setCodeColors (line 63) | void setCodeColors(const CodeColors &colors) { m_codeColors = colors; } type ContextLink (line 69) | struct ContextLink { type CodeCopy (line 76) | struct CodeCopy { function class (line 82) | class ChatViewTextProcessor : public QObject FILE: gpt4all-chat/src/codeinterpreter.cpp function QString (line 62) | QString CodeInterpreter::symbolicFormat() const function QString (line 67) | QString CodeInterpreter::examplePrompt() const function QString (line 72) | QString CodeInterpreter::exampleCall() const function QString (line 93) | QString CodeInterpreter::exampleReply() const FILE: gpt4all-chat/src/codeinterpreter.h function class (line 15) | class JavaScriptConsoleCapture : public QObject function class (line 39) | class CodeInterpreterWorker : public QObject { function class (line 66) | class CodeInterpreter : public Tool FILE: gpt4all-chat/src/database.cpp class BinaryDetectingFile (line 57) | class BinaryDetectingFile: public QFile { method binarySeen (line 61) | bool binarySeen() const { return m_binarySeen; } method qint64 (line 64) | qint64 readData(char *data, qint64 maxSize) override { method qint64 (line 69) | qint64 readLineData(char *data, qint64 maxSize) override { method qint64 (line 75) | qint64 checkData(const char *data, qint64 size) { function NAMED_PAIR (line 265) | NAMED_PAIR(EmbeddingKey, QString, embedding_model, int, chunk_id) function selectAllUncompletedChunks (line 271) | static bool selectAllUncompletedChunks(QSqlQuery &q, QHash &chunk_ids) function addCollection (line 373) | static bool addCollection(QSqlQuery &q, const QString &collection_name, ... function selectFoldersFromCollection (line 390) | static bool selectFoldersFromCollection(QSqlQuery &q, const QString &col... function sqlExtractCollections (line 402) | static QList sqlExtractCollections(QSqlQuery &q, bool wi... function selectAllFromCollections (line 437) | static bool selectAllFromCollections(QSqlQuery &q, QList... function selectCollectionByName (line 461) | static bool selectCollectionByName(QSqlQuery &q, const QString &name, st... function setCollectionEmbeddingModel (line 476) | static bool setCollectionEmbeddingModel(QSqlQuery &q, const QString &col... function updateStartUpdateTime (line 485) | static bool updateStartUpdateTime(QSqlQuery &q, int id, qint64 update_time) function updateLastUpdateTime (line 494) | static bool updateLastUpdateTime(QSqlQuery &q, int id, qint64 update_time) function addFolderToDB (line 551) | static bool addFolderToDB(QSqlQuery &q, const QString &folder_path, int ... function removeFolderFromDB (line 562) | static bool removeFolderFromDB(QSqlQuery &q, int folder_id) function selectFolder (line 570) | static bool selectFolder(QSqlQuery &q, const QString &folder_path, int *id) function sqlGetFolderEmbeddingModel (line 583) | static bool sqlGetFolderEmbeddingModel(QSqlQuery &q, int id, QString &em... function addCollectionItem (line 614) | static int addCollectionItem(QSqlQuery &q, int collection_id, int folder... function removeCollectionFolder (line 626) | static int removeCollectionFolder(QSqlQuery &q, const QString &collectio... function sqlPruneCollections (line 637) | static bool sqlPruneCollections(QSqlQuery &q) function addDocument (line 673) | static bool addDocument(QSqlQuery &q, int folder_id, qint64 document_tim... function removeDocument (line 686) | static bool removeDocument(QSqlQuery &q, int document_id) function updateDocument (line 694) | static bool updateDocument(QSqlQuery &q, int id, qint64 document_time) function selectDocument (line 703) | static bool selectDocument(QSqlQuery &q, const QString &document_path, i... function selectDocuments (line 718) | static bool selectDocuments(QSqlQuery &q, int folder_id, QList *doc... function selectCountStatistics (line 730) | static bool selectCountStatistics(QSqlQuery &q, int folder_id, int *tota... type Embedding (line 777) | struct Embedding { QString model; int folder_id; int chunk_id; QByteArra... type EmbeddingStat (line 778) | struct EmbeddingStat { QString lastFile; int nAdded; int nSkipped; } function sqlAddEmbeddings (line 783) | static bool sqlAddEmbeddings(QSqlQuery &q, const QList &embed... function CollectionItem (line 1061) | CollectionItem Database::guiCollectionItem(int folder_id) const function handleDocumentError (line 1108) | static void handleDocumentError(const QString &errorMessage, int documen... class DocumentReader (line 1114) | class DocumentReader { type Metadata (line 1116) | struct Metadata { QString title, author, subject, keywords; } method DocumentInfo (line 1120) | const DocumentInfo &doc () const { return m_info; } method Metadata (line 1121) | const Metadata &metadata() const { return m_metadata; } method getError (line 1124) | virtual std::optional getError() const { return... method page (line 1125) | virtual int page() const { return -1; } method DocumentReader (line 1130) | explicit DocumentReader(DocumentInfo info) method postInit (line 1133) | void postInit(Metadata &&metadata = {}) class PdfDocumentReader (line 1149) | class PdfDocumentReader final : public DocumentReader { method PdfDocumentReader (line 1151) | explicit PdfDocumentReader(DocumentInfo info) method page (line 1166) | int page() const override { return m_currentPage; } method advance (line 1169) | std::optional advance() override method PdfDocumentReader (line 1192) | explicit PdfDocumentReader(DocumentInfo info) method page (line 1218) | int page() const override { return m_currentPage; } method advance (line 1221) | std::optional advance() override method QString (line 1243) | QString getMetadata(FPDF_BYTESTRING key) method QString (line 1256) | QString extractTextFromPage(FPDF_PAGE page) class PdfDocumentReader (line 1190) | class PdfDocumentReader final : public DocumentReader { method PdfDocumentReader (line 1151) | explicit PdfDocumentReader(DocumentInfo info) method page (line 1166) | int page() const override { return m_currentPage; } method advance (line 1169) | std::optional advance() override method PdfDocumentReader (line 1192) | explicit PdfDocumentReader(DocumentInfo info) method page (line 1218) | int page() const override { return m_currentPage; } method advance (line 1221) | std::optional advance() override method QString (line 1243) | QString getMetadata(FPDF_BYTESTRING key) method QString (line 1256) | QString extractTextFromPage(FPDF_PAGE page) class WordDocumentReader (line 1282) | class WordDocumentReader final : public DocumentReader { method WordDocumentReader (line 1284) | explicit WordDocumentReader(DocumentInfo info) method advance (line 1299) | std::optional advance() override method fillBuffer (line 1334) | bool fillBuffer() class TxtDocumentReader (line 1375) | class TxtDocumentReader final : public DocumentReader { method TxtDocumentReader (line 1377) | explicit TxtDocumentReader(DocumentInfo info) method advance (line 1389) | std::optional advance() override method getError (line 1404) | std::optional getError() const override function DocumentInfo (line 1702) | DocumentInfo Database::dequeueDocument() type Result (line 2325) | struct Result { int chunkId; us::distance_punned_t dist; } type SearchResult (line 2454) | struct SearchResult { int chunkId; float score; } FILE: gpt4all-chat/src/database.h type DocumentInfo (line 62) | struct DocumentInfo function CollectionItem (line 132) | Q_DECLARE_METATYPE(ResultInfo) function ChunkStreamer (line 166) | Q_DECLARE_METATYPE(CollectionItem) function class (line 198) | class Database : public QObject FILE: gpt4all-chat/src/download.cpp class MyDownload (line 44) | class MyDownload: public Download { } function Download (line 46) | Download *Download::globalInstance() function ReleaseInfo (line 120) | ReleaseInfo Download::releaseInfo() const FILE: gpt4all-chat/src/download.h type ReleaseInfo (line 24) | struct ReleaseInfo { function class (line 36) | class HashAndSaveFile : public QObject function class (line 54) | class Download : public QObject FILE: gpt4all-chat/src/embllm.cpp function jsonArrayToVector (line 306) | std::vector jsonArrayToVector(const QJsonArray &jsonArray) function jsonArrayToEmbeddingResults (line 323) | QVector jsonArrayToEmbeddingResults(const QVector { function class (line 34) | class JinjaPromptAttachment : public JinjaHelper { function class (line 45) | class JinjaMessage : public JinjaHelper { FILE: gpt4all-chat/src/llm.cpp class MyLLM (line 32) | class MyLLM: public LLM { } function LLM (line 34) | LLM *LLM::globalInstance() function qint64 (line 101) | qint64 LLM::systemTotalRAMInGB() const function QString (line 106) | QString LLM::systemTotalRAMInGBString() const FILE: gpt4all-chat/src/llm.h function class (line 9) | class LLM : public QObject FILE: gpt4all-chat/src/localdocs.cpp class MyLocalDocs (line 17) | class MyLocalDocs: public LocalDocs { } function LocalDocs (line 19) | LocalDocs *LocalDocs::globalInstance() FILE: gpt4all-chat/src/localdocs.h function class (line 14) | class LocalDocs : public QObject FILE: gpt4all-chat/src/localdocsmodel.cpp function QVariant (line 77) | QVariant LocalDocsModel::data(const QModelIndex &index, int role) const FILE: gpt4all-chat/src/localdocsmodel.h function Q_OBJECT (line 51) | Q_OBJECT FILE: gpt4all-chat/src/logger.cpp class MyLogger (line 18) | class MyLogger: public Logger { } function Logger (line 20) | Logger *Logger::globalInstance() FILE: gpt4all-chat/src/logger.h function class (line 10) | class Logger { FILE: gpt4all-chat/src/macosdock.h type MacOSDock (line 4) | struct MacOSDock { FILE: gpt4all-chat/src/main.cpp function raiseWindow (line 55) | static void raiseWindow(QWindow *window) function main (line 73) | int main(int argc, char *argv[]) FILE: gpt4all-chat/src/modellist.cpp function QString (line 83) | QString ModelInfo::id() const function QString (line 93) | QString ModelInfo::name() const function QString (line 104) | QString ModelInfo::filename() const function QString (line 115) | QString ModelInfo::description() const function QString (line 126) | QString ModelInfo::url() const function QString (line 137) | QString ModelInfo::quant() const function QString (line 148) | QString ModelInfo::type() const function QDateTime (line 203) | QDateTime ModelInfo::recency() const function QVariant (line 350) | QVariant ModelInfo::defaultChatTemplate() const function QString (line 391) | QString ModelInfo::defaultSystemMessage() const function QString (line 401) | QString ModelInfo::chatNamePrompt() const function QString (line 412) | QString ModelInfo::suggestedFollowUpPrompt() const function QVariant (line 429) | QVariant ModelInfo::getField(QLatin1StringView name) const class MyModelList (line 548) | class MyModelList: public ModelList { } function ModelList (line 550) | ModelList *ModelList::globalInstance() function QString (line 618) | QString ModelList::compatibleModelNameHash(QUrl baseUrl, QString modelNa... function QString (line 624) | QString ModelList::compatibleModelFilename(QUrl baseUrl, QString modelNa... function QString (line 636) | QString ModelList::incompleteDownloadPath(const QString &modelFile) function ModelInfo (line 652) | ModelInfo ModelList::defaultModelInfo() const function QVariant (line 793) | QVariant ModelList::dataInternal(const ModelInfo *info, int role) const function QVariant (line 898) | QVariant ModelList::data(const QString &id, int role) const function QVariant (line 905) | QVariant ModelList::dataByFilename(const QString &filename, int role) const function QVariant (line 914) | QVariant ModelList::data(const QModelIndex &index, int role) const function ModelInfo (line 1150) | ModelInfo ModelList::modelInfo(const QString &id) const function ModelInfo (line 1158) | ModelInfo ModelList::modelInfoByFilename(const QString &filename, bool a... function QString (line 1177) | QString ModelList::clone(const ModelInfo &model) function QString (line 1282) | QString ModelList::uniqueModelName(const ModelInfo &model) const function QString (line 1464) | static QString modelsJsonFilename() function modelsJsonCacheFile (line 1469) | static std::optional modelsJsonCacheFile() type QuantType (line 2177) | enum QuantType { function QuantType (line 2185) | QuantType toQuantType(const QString& filename) function QString (line 2195) | QString toQuantString(const QString& filename) function QStringList (line 2361) | QStringList ModelList::remoteModelList(const QString &apiKey, const QUrl... FILE: gpt4all-chat/src/modellist.h function class (line 34) | class UpgradeableSetting { type ModelInfo (line 75) | struct ModelInfo { function QString (line 137) | QString id() const; function temperature (line 210) | double temperature() const; type DiscoverSort (line 364) | enum DiscoverSort { type Roles (line 371) | enum Roles { function override (line 422) | const override function ModelInfo (line 493) | ModelInfo defaultModelInfo() const; function GPT4AllDownloadableModels (line 502) | GPT4AllDownloadableModels *gpt4AllDownloadableModels() const { return m_... function HuggingFaceDownloadableModels (line 503) | HuggingFaceDownloadableModels *huggingFaceDownloadableModels() const { r... function QString (line 505) | static inline QString toFileSize(quint64 sz) { function discoverLimit (line 523) | int discoverLimit() const; FILE: gpt4all-chat/src/mysettings.cpp type ModelSettingsKey (line 41) | namespace ModelSettingsKey { namespace { type defaults (line 48) | namespace defaults { function QString (line 78) | static QString defaultLocalModelsPath() function QStringList (line 105) | static QStringList getDevices(bool skipKompute = false) function QString (line 121) | static QString getUiLanguage(const QString directory, const QString file... function QStringList (line 135) | static QStringList getUiLanguages(const QString &modelPath) function QString (line 162) | static QString modelSettingName(const ModelInfo &info, auto &&name) class MyPrivateSettings (line 167) | class MyPrivateSettings: public MySettings { } function MySettings (line 169) | MySettings *MySettings::globalInstance() function QVariant (line 182) | QVariant MySettings::checkJinjaTemplateError(const QString &tmpl) function QVariant (line 210) | QVariant MySettings::getBasicSetting(const QString &name) const function QString (line 280) | QString MySettings::modelName(const ModelInfo &info) const function QVariant (line 299) | QVariant MySettings::getModelSetting(QLatin1StringView name, const Model... function QVariant (line 305) | QVariant MySettings::getModelSetting(const char *name, const ModelInfo &... function QString (line 332) | QString MySettings::modelFilename (const ModelInfo &info... function QString (line 333) | QString MySettings::modelDescription (const ModelInfo &info... function QString (line 334) | QString MySettings::modelUrl (const ModelInfo &info... function QString (line 335) | QString MySettings::modelQuant (const ModelInfo &info... function QString (line 336) | QString MySettings::modelType (const ModelInfo &info... function QDateTime (line 341) | QDateTime MySettings::modelRecency (const ModelInfo &info... function QString (line 352) | QString MySettings::modelChatNamePrompt (const ModelInfo &info... function QString (line 353) | QString MySettings::modelSuggestedFollowUpPrompt(const ModelInfo &info... function QString (line 623) | QString MySettings::userDefaultModel() const { return getBasi... function QString (line 624) | QString MySettings::lastVersionStarted() const { return getBasi... function QStringList (line 628) | QStringList MySettings::localDocsFileExtensions() const { return getBasi... function QString (line 630) | QString MySettings::localDocsNomicAPIKey() const { return getBasi... function QString (line 631) | QString MySettings::localDocsEmbedDevice() const { return getBasi... function QString (line 632) | QString MySettings::networkAttribution() const { return getBasi... function ChatTheme (line 634) | ChatTheme MySettings::chatTheme() const { return ChatTheme ... function FontSize (line 635) | FontSize MySettings::fontSize() const { return FontSize ... function SuggestionMode (line 636) | SuggestionMode MySettings::suggestionMode() const { return SuggestionMod... function QString (line 656) | QString MySettings::modelPath() function QString (line 681) | QString MySettings::device() function QString (line 760) | QString MySettings::languageAndLocale() const function QString (line 768) | QString MySettings::filePathForLocale(const QLocale &locale) FILE: gpt4all-chat/src/mysettings.h function class (line 45) | class FontSize { FILE: gpt4all-chat/src/network.cpp function getSysctl (line 72) | static std::optional getSysctl(const char *name) function QString (line 84) | static QString getCPUModel() { return getSysctl("machdep.cpu.brand_strin... function get_cpuid (line 89) | static void get_cpuid(int level, int *regs) function QString (line 97) | static QString getCPUModel() function QString (line 120) | static QString getCPUModel() { return "(non-x86)"; } class MyNetwork (line 125) | class MyNetwork: public Network { } function Network (line 127) | Network *Network::globalInstance() function QString (line 178) | QString Network::generateUniqueId() const FILE: gpt4all-chat/src/network.h type KeyValue (line 23) | struct KeyValue { function class (line 28) | class Network : public QObject FILE: gpt4all-chat/src/server.cpp class InvalidRequestError (line 60) | class InvalidRequestError: public std::invalid_argument { method QHttpServerResponse (line 64) | QHttpServerResponse asResponse() const function QJsonObject (line 82) | static inline QJsonObject modelToJson(const ModelInfo &info) function QJsonObject (line 111) | static inline QJsonObject resultToJson(const ResultInfo &info) function BaseCompletionRequest (line 135) | BaseCompletionRequest() = default; function BaseCompletionRequest (line 138) | virtual BaseCompletionRequest &parse(QCborMap request) function parseImpl (line 149) | virtual void parseImpl(QCborMap &request) type Type (line 205) | enum class Type : uint8_t { function typeMatches (line 216) | static bool typeMatches(const QCborValue &value, Type type) noexcept { function QCborValue (line 229) | static QCborValue takeValue( class CompletionRequest (line 255) | class CompletionRequest : public BaseCompletionRequest { method CompletionRequest (line 261) | CompletionRequest &parse(QCborMap request) override method parseImpl (line 268) | void parseImpl(QCborMap &request) override class ChatRequest (line 321) | class ChatRequest : public BaseCompletionRequest { type Message (line 323) | struct Message { type Role (line 324) | enum class Role { System, User, Assistant } method ChatRequest (line 331) | ChatRequest &parse(QCborMap request) override method parseImpl (line 338) | void parseImpl(QCborMap &request) override function T (line 431) | T &parseRequest(T &request, QJsonObject &&obj) function QJsonObject (line 446) | static QJsonObject requestFromJson(const QByteArray &request) function makeError (line 621) | static auto makeError(auto &&...args) -> std::pair None: method get (line 25) | def get(self, path: str, *, raise_for_status: bool = True, wait: bool ... method post (line 28) | def post(self, path: str, data: dict[str, Any] | None, *, raise_for_st... method _request (line 31) | def _request( function create_chat_server_config (line 55) | def create_chat_server_config(tmpdir: Path, model_copied: bool = False) ... function prepare_chat_server (line 88) | def prepare_chat_server(model_copied: bool = False) -> Iterator[dict[str... function start_chat_server (line 98) | def start_chat_server(config: dict[str, str]) -> Iterator[None]: function chat_server (line 112) | def chat_server() -> Iterator[None]: function chat_server_with_model (line 118) | def chat_server_with_model() -> Iterator[None]: function test_with_models_empty (line 123) | def test_with_models_empty(chat_server: None) -> None: function test_with_models (line 223) | def test_with_models(chat_server_with_model: None) -> None: function test_with_models_temperature (line 255) | def test_with_models_temperature(chat_server_with_model: None) -> None: FILE: gpt4all-training/data.py function tokenize_inputs (line 10) | def tokenize_inputs(config, tokenizer, examples): function load_data (line 62) | def load_data(config, tokenizer): function load_data_for_inference (line 127) | def load_data_for_inference(config, tokenizer): FILE: gpt4all-training/eval_self_instruct.py function read_jsonl_file (line 17) | def read_jsonl_file(file_path): function setup_model (line 25) | def setup_model(config): function eval_example (line 44) | def eval_example(model, tokenizer, example, config): function do_eval (line 87) | def do_eval(config): FILE: gpt4all-training/generate.py function generate (line 10) | def generate(tokenizer, prompt, model, config): function setup_model (line 20) | def setup_model(config): FILE: gpt4all-training/inference.py function calc_cross_entropy_no_reduction (line 20) | def calc_cross_entropy_no_reduction(lm_logits, labels): function rank0_print (line 31) | def rank0_print(msg): function inference (line 36) | def inference(config): function main (line 191) | def main(): FILE: gpt4all-training/read.py function read_config (line 4) | def read_config(path): FILE: gpt4all-training/train.py function format_metrics (line 18) | def format_metrics(metrics, split, prefix=""): function evaluate (line 25) | def evaluate(model, val_dataloader): function train (line 40) | def train(accelerator, config):