SYMBOL INDEX (335 symbols across 37 files) FILE: chatbot/_models.py function gpu_found (line 22) | def gpu_found(): class Model (line 28) | class Model(object): method __init__ (line 32) | def __init__(self, logger, dataset, params): method compile (line 77) | def compile(self): method save (line 115) | def save(self, summaries=None): method close (line 133) | def close(self, save_current=True): method graph (line 150) | def graph(self): method fill_params (line 154) | def fill_params(dataset, params): method freeze (line 167) | def freeze(self): method __getattr__ (line 193) | def __getattr__(self, name): method _build_hparam_path (line 209) | def _build_hparam_path(ckpt_dir, **kwargs): class BucketModel (line 233) | class BucketModel(Model): method __init__ (line 240) | def __init__(self, logger, buckets, dataset, params): method compile (line 247) | def compile(self): method check_input_lengths (line 271) | def check_input_lengths(self, inputs, expected_lengths): method get_batch (line 282) | def get_batch(self, data, bucket_id): method train (line 333) | def train(self, dataset): method decode (line 338) | def decode(self): method step (line 343) | def step(self, encoder_inputs, decoder_inputs, target_weights, bucket_... FILE: chatbot/components/base/_rnn.py class Cell (line 25) | class Cell(RNNCell): method __init__ (line 29) | def __init__(self, state_size, num_layers, dropout_prob, base_cell): method state_size (line 55) | def state_size(self): method shape (line 59) | def shape(self): method single_layer_shape (line 67) | def single_layer_shape(self): method output_size (line 75) | def output_size(self): method __call__ (line 78) | def __call__(self, inputs, state, scope=None): class RNN (line 102) | class RNN(object): method __init__ (line 107) | def __init__(self, method get_cell (line 131) | def get_cell(self, name): method wrapper (line 139) | def wrapper(self, state): method __call__ (line 154) | def __call__(self, *args): class SimpleAttentionWrapper (line 158) | class SimpleAttentionWrapper(RNNCell): method __init__ (line 165) | def __init__(self, method zero_state (line 221) | def zero_state(self, batch_size, dtype): method call (line 250) | def call(self, inputs, state): method output_size (line 304) | def output_size(self): method state_size (line 308) | def state_size(self): method shape (line 317) | def shape(self): class BasicRNNCell (line 326) | class BasicRNNCell(RNNCell): method __init__ (line 335) | def __init__(self, num_units, reuse=None): method state_size (line 340) | def state_size(self): method output_size (line 344) | def output_size(self): method __call__ (line 347) | def __call__(self, inputs, state, scope=None): FILE: chatbot/components/bot_ops.py function dynamic_sampled_softmax_loss (line 9) | def dynamic_sampled_softmax_loss(labels, logits, output_projection, voca... function _dynamic_sampled_map (line 37) | def _dynamic_sampled_map(labels, logits, output_projection, vocab_size, function _dynamic_sampled_from_scratch (line 82) | def _dynamic_sampled_from_scratch(labels, logits, output_projection, voc... function cross_entropy_sequence_loss (line 148) | def cross_entropy_sequence_loss(logits, labels, weights): function dot_prod (line 165) | def dot_prod(x, y): function bahdanau_score (line 169) | def bahdanau_score(attention_dim, h_j, s_i): function luong_score (line 184) | def luong_score(attention_dim, h_j, s_i): function linear_map (line 195) | def linear_map(args, output_size, biases=None): FILE: chatbot/components/decoders.py class Decoder (line 20) | class Decoder(RNN): method __init__ (line 26) | def __init__(self, method __call__ (line 84) | def __call__(self, method apply_projection (line 177) | def apply_projection(self, outputs, scope=None): method sample (line 207) | def sample(self, projected_output): method get_projection_tensors (line 234) | def get_projection_tensors(self): class BasicDecoder (line 241) | class BasicDecoder(Decoder): method __call__ (line 244) | def __call__(self, class AttentionDecoder (line 259) | class AttentionDecoder(Decoder): method __init__ (line 269) | def __init__(self, method __call__ (line 302) | def __call__(self, method get_cell (line 325) | def get_cell(self, name, initial_state): FILE: chatbot/components/embedder.py class Embedder (line 9) | class Embedder: method __init__ (line 14) | def __init__(self, vocab_size, embed_size, l1_reg=0.0): method __call__ (line 20) | def __call__(self, inputs, reuse=None): method assign_visualizers (line 57) | def assign_visualizers(self, writer, scope_names, metadata_path): method get_scope_basename (line 78) | def get_scope_basename(self, scope): class AutoEncoder (line 86) | class AutoEncoder(Model): method __init__ (line 91) | def __init__(self, dataset, params): method build_computation_graph (line 98) | def build_computation_graph(self, dataset): method compile (line 143) | def compile(self): method step (line 178) | def step(self, forward_only=False): method train (line 185) | def train(self, close_when_done=True): method __call__ (line 219) | def __call__(self, sentence): FILE: chatbot/components/encoders.py class BasicEncoder (line 10) | class BasicEncoder(RNN): method __call__ (line 15) | def __call__(self, inputs, initial_state=None): class BidirectionalEncoder (line 37) | class BidirectionalEncoder(RNN): method __call__ (line 46) | def __call__(self, inputs, initial_state=None): FILE: chatbot/components/input_pipeline.py class InputPipeline (line 12) | class InputPipeline: method __init__ (line 22) | def __init__(self, file_paths, batch_size, capacity=None, is_chatting=... method build_pipeline (line 48) | def build_pipeline(self, name): method encoder_inputs (line 74) | def encoder_inputs(self): method decoder_inputs (line 83) | def decoder_inputs(self): method user_input (line 93) | def user_input(self): method feed_dict (line 97) | def feed_dict(self): method feed_user_input (line 100) | def feed_user_input(self, user_input): method toggle_active (line 104) | def toggle_active(self): method _cond_input (line 111) | def _cond_input(self, prefix): method _read_line (line 118) | def _read_line(self, file): method _assign_queue (line 130) | def _assign_queue(self, proto_text): method _padded_bucket_batches (line 154) | def _padded_bucket_batches(self, input_length, data): FILE: chatbot/dynamic_models.py class DynamicBot (line 21) | class DynamicBot(Model): method __init__ (line 32) | def __init__(self, dataset, params): method build_computation_graph (line 54) | def build_computation_graph(self, dataset): method compile (line 142) | def compile(self): method step (line 210) | def step(self, forward_only=False): method train (line 254) | def train(self, dataset=None): method decode (line 333) | def decode(self): method __call__ (line 347) | def __call__(self, sentence): method chat (line 380) | def chat(self): method respond (line 384) | def respond(self, sentence): method close (line 388) | def close(self, save_current=True, rebuild_for_chat=True): method _set_chat_params (line 408) | def _set_chat_params(self): FILE: chatbot/legacy/_decode.py function decode (line 13) | def decode(bot, dataset, teacher_mode=True): function decode_inputs (line 43) | def decode_inputs(inputs, idx_to_word, chatbot): function train_on_feedback (line 55) | def train_on_feedback(chatbot, input_ids, feedback_ids, idx_to_outputs): function _logits_to_outputs (line 67) | def _logits_to_outputs(output_logits, temperature, idx_word): function _sample (line 84) | def _sample(logits, temperature): function _assign_to_bucket (line 97) | def _assign_to_bucket(token_ids, buckets): FILE: chatbot/legacy/_train.py function train (line 9) | def train(bot, dataset): function run_train_step (line 51) | def run_train_step(model, train_set, bucket_id, forward_only=False): function run_checkpoint (line 60) | def run_checkpoint(model, step_time, loss, previous_losses, dev_set): function _get_data_distribution (line 80) | def _get_data_distribution(train_set, buckets): FILE: chatbot/legacy/legacy_models.py class ChatBot (line 18) | class ChatBot(BucketModel): method __init__ (line 37) | def __init__(self, buckets, dataset, params): method step (line 112) | def step(self, encoder_inputs, decoder_inputs, target_weights, bucket_... method _sampled_loss (line 153) | def _sampled_loss(num_samples, hidden_size, vocab_size): method _get_projections (line 187) | def _get_projections(num_buckets, unprojected_vals, projection_operator): method _get_placeholder_list (line 202) | def _get_placeholder_list(name, length, dtype=tf.int32): class SimpleBot (line 213) | class SimpleBot(BucketModel): method __init__ (line 221) | def __init__(self, dataset, params): method _simple_loss (line 324) | def _simple_loss(batch_size, logits, targets, weights): method step (line 339) | def step(self, encoder_inputs, decoder_inputs, target_weights, bucket_... FILE: data/_dataset.py class DatasetABC (line 18) | class DatasetABC(metaclass=ABCMeta): method convert_to_tf_records (line 21) | def convert_to_tf_records(self, *args): method train_generator (line 28) | def train_generator(self, batch_size): method valid_generator (line 35) | def valid_generator(self, batch_size): method word_to_idx (line 42) | def word_to_idx(self): method idx_to_word (line 47) | def idx_to_word(self): method name (line 52) | def name(self): method max_seq_len (line 57) | def max_seq_len(self): class Dataset (line 62) | class Dataset(DatasetABC): method __init__ (line 64) | def __init__(self, dataset_params): method convert_to_tf_records (line 102) | def convert_to_tf_records(self, prefix='train'): method sentence_generator (line 166) | def sentence_generator(self, prefix='from'): method pairs_generator (line 179) | def pairs_generator(self, num_generate=None): method train_generator (line 196) | def train_generator(self, batch_size): method valid_generator (line 203) | def valid_generator(self, batch_size): method _generator (line 210) | def _generator(self, from_path, to_path, batch_size): method word_to_idx (line 272) | def word_to_idx(self): method idx_to_word (line 277) | def idx_to_word(self): method as_words (line 281) | def as_words(self, sentence): method name (line 304) | def name(self): method train_size (line 309) | def train_size(self): method valid_size (line 313) | def valid_size(self): method max_seq_len (line 317) | def max_seq_len(self): method fill_params (line 321) | def fill_params(dataset_params): method __getattr__ (line 328) | def __getattr__(self, name): FILE: data/data_helper.py function prompt (line 38) | def prompt(text, default="", required=False): class DataHelper (line 49) | class DataHelper: method __init__ (line 56) | def __init__(self, log_level=logging.INFO): method safe_load (line 113) | def safe_load(self): method load_random (line 153) | def load_random(self, year=None): method load_next (line 168) | def load_next(self): method set_word_freq (line 181) | def set_word_freq(self, wf): method word_freq (line 188) | def word_freq(self): method next_file_path (line 192) | def next_file_path(self): method get_year_from_path (line 195) | def get_year_from_path(self, path): method generate_files (line 206) | def generate_files(self, method df_generator (line 240) | def df_generator(self): method random_rows_generator (line 250) | def random_rows_generator(num_rows_per_print, num_rows_total): method word_tokenizer (line 259) | def word_tokenizer(sentences): method df_to_json (line 271) | def df_to_json(df, target_file=None, orient='records', lines=False, **... FILE: data/dataset_wrappers.py function check_data (line 11) | def check_data(abs_path, name): class Cornell (line 30) | class Cornell(Dataset): method __init__ (line 33) | def __init__(self, dataset_params): class Ubuntu (line 42) | class Ubuntu(Dataset): method __init__ (line 45) | def __init__(self, dataset_params): class Reddit (line 54) | class Reddit(Dataset): method __init__ (line 57) | def __init__(self, dataset_params): class TestData (line 66) | class TestData(Dataset): method __init__ (line 69) | def __init__(self, dataset_params): FILE: data/reddit_preprocessor.py function timed_function (line 35) | def timed_function(*expected_args): function parallel_map_df (line 52) | def parallel_map_df(fn, df): function parallel_map_list (line 64) | def parallel_map_list(fn, iterable): function sentence_score (line 75) | def sentence_score(sentences): function root_comments (line 86) | def root_comments(df): function remove_extra_columns (line 101) | def remove_extra_columns(df): function regex_replacements (line 112) | def regex_replacements(df): function remove_large_comments (line 131) | def remove_large_comments(max_len, df): function expand_contractions (line 138) | def expand_contractions(df): function children_dict (line 149) | def children_dict(df): function main (line 163) | def main(): FILE: main.py function start_training (line 86) | def start_training(dataset, bot): function start_chatting (line 95) | def start_chatting(bot): function main (line 110) | def main(argv): FILE: tests/test_config.py class TestConfig (line 16) | class TestConfig(unittest.TestCase): method setUp (line 19) | def setUp(self): method test_merge_params (line 24) | def test_merge_params(self): method test_optimize (line 64) | def test_optimize(self): method test_update_config (line 91) | def test_update_config(self): FILE: tests/test_data.py class TestData (line 17) | class TestData(unittest.TestCase): method setUp (line 20) | def setUp(self): method test_basic (line 30) | def test_basic(self): method test_cornell (line 87) | def test_cornell(self): FILE: tests/test_dynamic_models.py class TestDynamicModels (line 18) | class TestDynamicModels(unittest.TestCase): method setUp (line 20) | def setUp(self): method test_create_bot (line 23) | def test_create_bot(self): method test_save_bot (line 29) | def test_save_bot(self): method test_save_bot (line 34) | def test_save_bot(self): method test_train (line 41) | def test_train(self): method test_base_methods (line 51) | def test_base_methods(self): method test_manual_freeze (line 61) | def test_manual_freeze(self): method test_memorize (line 142) | def test_memorize(self): method _quick_train (line 179) | def _quick_train(self, bot, num_iter=10): FILE: tests/test_legacy_models.py class TestLegacyModels (line 13) | class TestLegacyModels(unittest.TestCase): method setUp (line 16) | def setUp(self): method test_create (line 24) | def test_create(self): method test_compile (line 37) | def test_compile(self): FILE: tests/utils.py function create_bot (line 37) | def create_bot(flags=TEST_FLAGS, return_dataset=False): function update_config (line 63) | def update_config(config, **kwargs): FILE: utils/bot_freezer.py function load_graph (line 15) | def load_graph(frozen_model_dir): function unfreeze_bot (line 43) | def unfreeze_bot(frozen_model_path): function unfreeze_and_chat (line 60) | def unfreeze_and_chat(frozen_model_path): function get_frozen_vocab (line 94) | def get_frozen_vocab(config): class FrozenBot (line 103) | class FrozenBot: method __init__ (line 105) | def __init__(self, frozen_model_dir, vocab_size): method as_words (line 115) | def as_words(self, sentence): method __call__ (line 118) | def __call__(self, sentence): FILE: utils/io_utils.py function save_hyper_params (line 60) | def save_hyper_params(hyper_params, fname): function get_sentence (line 67) | def get_sentence(lower=True): function update_config (line 79) | def update_config(config=None, function get_yaml_config (line 129) | def get_yaml_config(path, save_path=True): function load_pretrained_config (line 138) | def load_pretrained_config(pretrained_dir): function print_non_defaults (line 162) | def print_non_defaults(config): function flags_to_dict (line 185) | def flags_to_dict(flags): function merge_dicts (line 228) | def merge_dicts(default_dict, preference_dict): function parse_config (line 252) | def parse_config(flags=None, pretrained_dir=None, config_path=None): function basic_tokenizer (line 309) | def basic_tokenizer(sentence): function num_lines (line 317) | def num_lines(file_path): function get_word_freqs (line 323) | def get_word_freqs(path, counter, norm_digits=True): function create_vocabulary (line 349) | def create_vocabulary(vocab_path, from_path, to_path, max_vocab_size, no... function get_vocab_dicts (line 386) | def get_vocab_dicts(vocabulary_path): function sentence_to_token_ids (line 410) | def sentence_to_token_ids(sentence, vocabulary, normalize_digits=True): function data_to_token_ids (line 434) | def data_to_token_ids(data_path, target_path, vocabulary_path, normalize... function prepare_data (line 461) | def prepare_data(data_dir, FILE: webpage/config.py class Config (line 5) | class Config: method init_app (line 25) | def init_app(app): class DevelopmentConfig (line 29) | class DevelopmentConfig(Config): class TestingConfig (line 35) | class TestingConfig(Config): class ProductionConfig (line 41) | class ProductionConfig(Config): FILE: webpage/deepchat/__init__.py class ReverseProxied (line 29) | class ReverseProxied(object): method __init__ (line 46) | def __init__(self, app): method __call__ (line 49) | def __call__(self, environ, start_response): function create_app (line 68) | def create_app(config_name): FILE: webpage/deepchat/main/errors.py function page_not_found (line 8) | def page_not_found(e): function internal_server_error (line 13) | def internal_server_error(e): FILE: webpage/deepchat/main/forms.py function bad_chars (line 10) | def bad_chars(form, string_field): class ChatForm (line 16) | class ChatForm(FlaskForm): class UserForm (line 22) | class UserForm(FlaskForm): class SentencePairForm (line 30) | class SentencePairForm(FlaskForm): FILE: webpage/deepchat/main/views.py function inject_enumerate (line 26) | def inject_enumerate(): function load_gloabal_data (line 31) | def load_gloabal_data(): function index (line 39) | def index(): function about (line 51) | def about(): function plots (line 56) | def plots(): function update_database (line 61) | def update_database(user_message, bot_response): function get_database_model (line 87) | def get_database_model(class_name, filter=None, **kwargs): class UserAPI (line 111) | class UserAPI(Resource): method post (line 113) | def post(self): class ChatAPI (line 120) | class ChatAPI(Resource): method __init__ (line 126) | def __init__(self, data_name): method post (line 140) | def post(self): class RedditAPI (line 152) | class RedditAPI(ChatAPI): method __init__ (line 153) | def __init__(self): class CornellAPI (line 157) | class CornellAPI(ChatAPI): method __init__ (line 158) | def __init__(self): class UbuntuAPI (line 162) | class UbuntuAPI(ChatAPI): method __init__ (line 163) | def __init__(self): class AuthException (line 176) | class AuthException(HTTPException): method __init__ (line 177) | def __init__(self, message): class ModelView (line 183) | class ModelView(sqla.ModelView): method is_accessible (line 184) | def is_accessible(self): method inaccessible_callback (line 190) | def inaccessible_callback(self, name, **kwargs): FILE: webpage/deepchat/models.py class User (line 15) | class User(db.Model): method __repr__ (line 45) | def __repr__(self): class Chatbot (line 49) | class Chatbot(db.Model): method __init__ (line 62) | def __init__(self, name, **bot_kwargs): method __repr__ (line 73) | def __repr__(self): class Conversation (line 77) | class Conversation(db.Model): method __repr__ (line 84) | def __repr__(self): class Turn (line 88) | class Turn(db.Model): method __repr__ (line 94) | def __repr__(self): FILE: webpage/deepchat/static/js/jqBootstrapValidation.js function regexFromString (line 875) | function regexFromString(inputstring) { function executeFunctionByName (line 885) | function executeFunctionByName(functionName, context /*, args*/) { FILE: webpage/deepchat/static/vendor/bootstrap-3.3.7-dist/js/bootstrap.js function transitionEnd (line 34) | function transitionEnd() { function removeElement (line 126) | function removeElement() { function Plugin (line 142) | function Plugin(option) { function Plugin (line 251) | function Plugin(option) { function Plugin (line 475) | function Plugin(option) { function getTargetFromTrigger (line 695) | function getTargetFromTrigger($trigger) { function Plugin (line 707) | function Plugin(option) { function getParent (line 774) | function getParent($this) { function clearMenus (line 787) | function clearMenus(e) { function Plugin (line 880) | function Plugin(option) { function Plugin (line 1208) | function Plugin(option, _relatedTarget) { function complete (line 1574) | function complete() { function Plugin (line 1750) | function Plugin(option) { function Plugin (line 1859) | function Plugin(option) { function ScrollSpy (line 1902) | function ScrollSpy(element, options) { function Plugin (line 2022) | function Plugin(option) { function next (line 2131) | function next() { function Plugin (line 2177) | function Plugin(option) { function Plugin (line 2334) | function Plugin(option) { FILE: webpage/deepchat/web_bot.py function basic_tokenizer (line 19) | def basic_tokenizer(sentence): function sentence_to_token_ids (line 26) | def sentence_to_token_ids(sentence, vocabulary, normalize_digits=True): function get_vocab_dicts (line 34) | def get_vocab_dicts(vocabulary_path): function load_graph (line 47) | def load_graph(frozen_model_dir): function unfreeze_bot (line 74) | def unfreeze_bot(frozen_model_path): class FrozenBot (line 91) | class FrozenBot: method __init__ (line 94) | def __init__(self, frozen_model_dir, is_testing=False): method load_config (line 115) | def load_config(self, config_path): method __getattr__ (line 121) | def __getattr__(self, name): method get_frozen_vocab (line 134) | def get_frozen_vocab(self, config): method as_words (line 142) | def as_words(self, sentence): method __call__ (line 160) | def __call__(self, sentence): method unfreeze (line 183) | def unfreeze(self): method freeze (line 190) | def freeze(self): FILE: webpage/manage.py function make_shell_context (line 27) | def make_shell_context(): function test (line 44) | def test(): function deploy (line 58) | def deploy(): FILE: webpage/migrations/env.py function run_migrations_offline (line 31) | def run_migrations_offline(): function run_migrations_online (line 50) | def run_migrations_online(): FILE: webpage/migrations/versions/236b966ecd2f_.py function upgrade (line 19) | def upgrade(): function downgrade (line 61) | def downgrade(): FILE: webpage/tests/test_database.py class TestDatabase (line 14) | class TestDatabase(unittest.TestCase): method setUp (line 16) | def setUp(self): method tearDown (line 23) | def tearDown(self): method test_app_exists (line 29) | def test_app_exists(self): FILE: webpage/tests/test_simple.py class TestSimple (line 8) | class TestSimple(unittest.TestCase): method setUp (line 11) | def setUp(self): method tearDown (line 18) | def tearDown(self): method test_app_exists (line 24) | def test_app_exists(self): method test_app_is_testing (line 27) | def test_app_is_testing(self):