SYMBOL INDEX (41 symbols across 6 files) FILE: SmilesPE/_nbdev.py function custom_doc_links (line 33) | def custom_doc_links(name): return None FILE: SmilesPE/learner.py function randomize_smiles (line 20) | def randomize_smiles(smiles): function corpus_augment (line 38) | def corpus_augment(infile, outdir, cycles): function get_vocabulary (line 63) | def get_vocabulary(smiles, augmentation=0, exclusive_tokens = False): function update_pair_statistics (line 85) | def update_pair_statistics(pair, changed, stats, indices): function get_pair_statistics (line 142) | def get_pair_statistics(vocab): function replace_pair (line 160) | def replace_pair(pair, vocab, indices): function prune_stats (line 184) | def prune_stats(stats, big_stats, threshold): function learn_SPE (line 198) | def learn_SPE(infile, outfile, num_symbols, min_frequency=2, augmentatio... FILE: SmilesPE/pretokenizer.py function atomwise_tokenizer (line 6) | def atomwise_tokenizer(smi, exclusive_tokens = None): function kmer_tokenizer (line 29) | def kmer_tokenizer(smiles, ngram=4, stride=1, remove_last = False, exclu... function tokens_to_mer (line 41) | def tokens_to_mer(toks): FILE: SmilesPE/spe2vec.py class Corpus (line 10) | class Corpus(object): method __init__ (line 18) | def __init__(self, infile, tokenizer, isdir=False, dropout=0): method __iter__ (line 24) | def __iter__(self): function learn_spe2vec (line 33) | def learn_spe2vec(corpus, outfile=None, function load_spe2vec (line 71) | def load_spe2vec(model_path): class SPE2Vec (line 78) | class SPE2Vec(object): method __init__ (line 79) | def __init__(self, model_path, tokenizer): method tokenize (line 91) | def tokenize(self, smi, dropout=0): method smiles2vec (line 97) | def smiles2vec(self, smi, dropout=0, mode = 'average'): method spe2vec (line 138) | def spe2vec(self, smi, dropout=0, skip_unknown=False): FILE: SmilesPE/tokenizer.py class SPE_Tokenizer (line 21) | class SPE_Tokenizer(object): method __init__ (line 37) | def __init__(self, codes, merges=-1, glossaries=None, exclusive_tokens... method tokenize (line 62) | def tokenize(self, smi, dropout=0): method _isolate_glossaries (line 74) | def _isolate_glossaries(self, word): function encode (line 82) | def encode(orig, bpe_codes, bpe_codes_reverse, cache, function isolate_glossary (line 131) | def isolate_glossary(word, glossary): FILE: docs/js/jekyll-search.js function s (line 1) | function s(o,u){if(!n[o]){if(!t[o]){var a="function"==typeof require&&re... function receivedResponse (line 1) | function receivedResponse(xhr){return 200==xhr.status&&4==xhr.readyState} function handleResponse (line 1) | function handleResponse(xhr,callback){xhr.onreadystatechange=function(){... function FuzzySearchStrategy (line 1) | function FuzzySearchStrategy(){function createFuzzyRegExpFromString(stri... function LiteralSearchStrategy (line 1) | function LiteralSearchStrategy(){function doMatch(string,crit){return st... function findMatches (line 1) | function findMatches(store,crit,strategy){for(var data=store.get(),i=0;i... function findMatchesInObject (line 1) | function findMatchesInObject(obj,crit,strategy){for(var key in obj)if(st... function getSearchStrategy (line 1) | function getSearchStrategy(){return fuzzy?fuzzySearchStrategy:literalSea... function isObject (line 1) | function isObject(obj){return!!obj&&"[object Object]"==Object.prototype.... function isArray (line 1) | function isArray(obj){return!!obj&&"[object Array]"==Object.prototype.to... function addObject (line 1) | function addObject(data){return store.push(data),data} function addArray (line 1) | function addArray(data){for(var added=[],i=0;i