SYMBOL INDEX (187 symbols across 9 files) FILE: create_pretraining_data.py class TrainingInstance (line 64) | class TrainingInstance(object): method __init__ (line 67) | def __init__(self, tokens, segment_ids, masked_lm_positions, masked_lm... method __str__ (line 75) | def __str__(self): method __repr__ (line 88) | def __repr__(self): function write_instance_to_example_files (line 92) | def write_instance_to_example_files(instances, tokenizer, max_seq_length, function create_int_feature (line 165) | def create_int_feature(values): function create_float_feature (line 170) | def create_float_feature(values): function create_training_instances (line 175) | def create_training_instances(input_files, tokenizer, max_seq_length, function create_instances_from_document (line 219) | def create_instances_from_document( function create_masked_lm_predictions (line 338) | def create_masked_lm_predictions(tokens, masked_lm_prob, function truncate_seq_pair (line 391) | def truncate_seq_pair(tokens_a, tokens_b, max_num_tokens, rng): function main (line 409) | def main(_): FILE: extract_features.py class InputExample (line 81) | class InputExample(object): method __init__ (line 83) | def __init__(self, unique_id, text_a, text_b): class InputFeatures (line 89) | class InputFeatures(object): method __init__ (line 92) | def __init__(self, unique_id, tokens, input_ids, input_mask, input_typ... function input_fn_builder (line 100) | def input_fn_builder(features, seq_length): function model_fn_builder (line 148) | def model_fn_builder(bert_config, init_checkpoint, layer_indexes, use_tpu, function convert_examples_to_features (line 210) | def convert_examples_to_features(examples, seq_length, tokenizer): function _truncate_seq_pair (line 302) | def _truncate_seq_pair(tokens_a, tokens_b, max_length): function read_examples (line 319) | def read_examples(input_file): function main (line 343) | def main(_): FILE: modeling.py class BertConfig (line 30) | class BertConfig(object): method __init__ (line 33) | def __init__(self, method from_dict (line 82) | def from_dict(cls, json_object): method from_json_file (line 90) | def from_json_file(cls, json_file): method to_dict (line 96) | def to_dict(self): method to_json_string (line 101) | def to_json_string(self): class BertModel (line 106) | class BertModel(object): method __init__ (line 130) | def __init__(self, method get_pooled_output (line 235) | def get_pooled_output(self): method get_sequence_output (line 238) | def get_sequence_output(self): method get_all_encoder_layers (line 247) | def get_all_encoder_layers(self): method get_embedding_output (line 250) | def get_embedding_output(self): method get_embedding_table (line 261) | def get_embedding_table(self): function gelu (line 265) | def gelu(input_tensor): function get_activation (line 281) | def get_activation(activation_string): function get_assignment_map_from_checkpoint (line 318) | def get_assignment_map_from_checkpoint(tvars, init_checkpoint): function dropout (line 345) | def dropout(input_tensor, dropout_prob): function layer_norm (line 363) | def layer_norm(input_tensor, name=None): function layer_norm_and_dropout (line 369) | def layer_norm_and_dropout(input_tensor, dropout_prob, name=None): function create_initializer (line 376) | def create_initializer(initializer_range=0.02): function embedding_lookup (line 381) | def embedding_lookup(input_ids, function embedding_postprocessor (line 430) | def embedding_postprocessor(input_tensor, function create_attention_mask_from_input_mask (line 526) | def create_attention_mask_from_input_mask(from_tensor, to_mask): function attention_layer (line 560) | def attention_layer(from_tensor, function transformer_model (line 756) | def transformer_model(input_tensor, function get_shape_list (line 897) | def get_shape_list(tensor, expected_rank=None, name=None): function reshape_to_matrix (line 934) | def reshape_to_matrix(input_tensor): function reshape_from_matrix (line 948) | def reshape_from_matrix(output_tensor, orig_shape_list): function assert_rank (line 961) | def assert_rank(tensor, expected_rank, name=None): FILE: optimization.py function create_optimizer (line 25) | def create_optimizer(loss, init_lr, num_train_steps, num_warmup_steps, u... class AdamWeightDecayOptimizer (line 87) | class AdamWeightDecayOptimizer(tf.train.Optimizer): method __init__ (line 90) | def __init__(self, method apply_gradients (line 108) | def apply_gradients(self, grads_and_vars, global_step=None, name=None): method _do_use_weight_decay (line 159) | def _do_use_weight_decay(self, param_name): method _get_variable_name (line 169) | def _get_variable_name(self, param_name): FILE: predict.py class InputExample (line 128) | class InputExample(object): method __init__ (line 131) | def __init__(self, guid, text_a, text_b=None, label=None): class PaddingInputExample (line 149) | class PaddingInputExample(object): class InputFeatures (line 162) | class InputFeatures(object): method __init__ (line 165) | def __init__(self, class DataProcessor (line 178) | class DataProcessor(object): method get_train_examples (line 181) | def get_train_examples(self, data_dir): method get_dev_examples (line 185) | def get_dev_examples(self, data_dir): method get_test_examples (line 189) | def get_test_examples(self, data_dir): method get_labels (line 193) | def get_labels(self): method _read_tsv (line 198) | def _read_tsv(cls, input_file, quotechar=None): class SimProcessor (line 207) | class SimProcessor(DataProcessor): method get_train_examples (line 210) | def get_train_examples(self, data_dir): method get_dev_examples (line 213) | def get_dev_examples(self, data_dir): method get_test_examples (line 216) | def get_test_examples(self, data_dir): method get_labels (line 228) | def get_labels(self): function convert_single_example (line 231) | def convert_single_example(ex_index, example, label_list, max_seq_length, function file_based_convert_examples_to_features (line 333) | def file_based_convert_examples_to_features( function file_based_input_fn_builder (line 363) | def file_based_input_fn_builder(input_file, seq_length, is_training, function _truncate_seq_pair (line 411) | def _truncate_seq_pair(tokens_a, tokens_b, max_length): function create_model (line 428) | def create_model(bert_config, is_training, input_ids, input_mask, segmen... function model_fn_builder (line 472) | def model_fn_builder(bert_config, num_labels, init_checkpoint, learning_... function input_fn_builder (line 530) | def input_fn_builder(features, seq_length, is_training, drop_remainder): function convert_examples_to_features (line 584) | def convert_examples_to_features(examples, label_list, max_seq_length, function predicts (line 600) | def predicts(text_data): FILE: run_classifier.py class InputExample (line 129) | class InputExample(object): method __init__ (line 132) | def __init__(self, guid, text_a, text_b=None, label=None): class PaddingInputExample (line 150) | class PaddingInputExample(object): class InputFeatures (line 163) | class InputFeatures(object): method __init__ (line 166) | def __init__(self, class DataProcessor (line 179) | class DataProcessor(object): method get_train_examples (line 182) | def get_train_examples(self, data_dir): method get_dev_examples (line 186) | def get_dev_examples(self, data_dir): method get_test_examples (line 190) | def get_test_examples(self, data_dir): method get_labels (line 194) | def get_labels(self): method _read_tsv (line 199) | def _read_tsv(cls, input_file, quotechar=None): class XnliProcessor (line 209) | class XnliProcessor(DataProcessor): method __init__ (line 212) | def __init__(self): method get_train_examples (line 215) | def get_train_examples(self, data_dir): method get_dev_examples (line 234) | def get_dev_examples(self, data_dir): method get_labels (line 252) | def get_labels(self): class SimProcessor (line 257) | class SimProcessor(DataProcessor): method get_train_examples (line 277) | def get_train_examples(self, data_dir): method get_dev_examples (line 304) | def get_dev_examples(self, data_dir): method get_test_examples (line 318) | def get_test_examples(self, data_dir): method get_labels (line 330) | def get_labels(self): class MnliProcessor (line 334) | class MnliProcessor(DataProcessor): method get_train_examples (line 337) | def get_train_examples(self, data_dir): method get_dev_examples (line 342) | def get_dev_examples(self, data_dir): method get_test_examples (line 348) | def get_test_examples(self, data_dir): method get_labels (line 353) | def get_labels(self): method _create_examples (line 357) | def _create_examples(self, lines, set_type): class MrpcProcessor (line 375) | class MrpcProcessor(DataProcessor): method get_train_examples (line 378) | def get_train_examples(self, data_dir): method get_dev_examples (line 383) | def get_dev_examples(self, data_dir): method get_test_examples (line 388) | def get_test_examples(self, data_dir): method get_labels (line 393) | def get_labels(self): method _create_examples (line 397) | def _create_examples(self, lines, set_type): class ColaProcessor (line 415) | class ColaProcessor(DataProcessor): method get_train_examples (line 418) | def get_train_examples(self, data_dir): method get_dev_examples (line 423) | def get_dev_examples(self, data_dir): method get_test_examples (line 428) | def get_test_examples(self, data_dir): method get_labels (line 433) | def get_labels(self): method _create_examples (line 437) | def _create_examples(self, lines, set_type): function convert_single_example (line 456) | def convert_single_example(ex_index, example, label_list, max_seq_length, function file_based_convert_examples_to_features (line 558) | def file_based_convert_examples_to_features( function file_based_input_fn_builder (line 588) | def file_based_input_fn_builder(input_file, seq_length, is_training, function _truncate_seq_pair (line 636) | def _truncate_seq_pair(tokens_a, tokens_b, max_length): function create_model (line 653) | def create_model(bert_config, is_training, input_ids, input_mask, segmen... function model_fn_builder (line 698) | def model_fn_builder(bert_config, num_labels, init_checkpoint, learning_... function input_fn_builder (line 792) | def input_fn_builder(features, seq_length, is_training, drop_remainder): function convert_examples_to_features (line 846) | def convert_examples_to_features(examples, label_list, max_seq_length, function main (line 862) | def main(_): FILE: run_pretraining.py function model_fn_builder (line 109) | def model_fn_builder(bert_config, init_checkpoint, learning_rate, function get_masked_lm_output (line 240) | def get_masked_lm_output(bert_config, input_tensor, output_weights, posi... function get_next_sentence_output (line 285) | def get_next_sentence_output(bert_config, input_tensor, labels): function gather_indexes (line 308) | def gather_indexes(sequence_tensor, positions): function input_fn_builder (line 324) | def input_fn_builder(input_files, function _decode_record (line 391) | def _decode_record(record, name_to_features): function main (line 406) | def main(_): FILE: run_squad.py class SquadExample (line 157) | class SquadExample(object): method __init__ (line 163) | def __init__(self, method __str__ (line 179) | def __str__(self): method __repr__ (line 182) | def __repr__(self): class InputFeatures (line 197) | class InputFeatures(object): method __init__ (line 200) | def __init__(self, function read_squad_examples (line 227) | def read_squad_examples(input_file, is_training): function convert_examples_to_features (line 309) | def convert_examples_to_features(examples, tokenizer, max_seq_length, function _improve_answer_span (line 476) | def _improve_answer_span(doc_tokens, input_start, input_end, tokenizer, function _check_is_max_context (line 513) | def _check_is_max_context(doc_spans, cur_span_index, position): function create_model (line 550) | def create_model(bert_config, is_training, input_ids, input_mask, segmen... function model_fn_builder (line 590) | def model_fn_builder(bert_config, init_checkpoint, learning_rate, function input_fn_builder (line 687) | def input_fn_builder(input_file, seq_length, is_training, drop_remainder): function write_predictions (line 741) | def write_predictions(all_examples, all_features, all_results, n_best_size, function get_final_text (line 927) | def get_final_text(pred_text, orig_text, do_lower_case): function _get_best_indexes (line 1023) | def _get_best_indexes(logits, n_best_size): function _compute_softmax (line 1035) | def _compute_softmax(scores): class FeatureWriter (line 1058) | class FeatureWriter(object): method __init__ (line 1061) | def __init__(self, filename, is_training): method process_feature (line 1067) | def process_feature(self, feature): method close (line 1093) | def close(self): function validate_flags_or_throw (line 1097) | def validate_flags_or_throw(bert_config): function main (line 1126) | def main(_): FILE: tokenization.py function validate_case_matches_checkpoint (line 28) | def validate_case_matches_checkpoint(do_lower_case, init_checkpoint): function convert_to_unicode (line 78) | def convert_to_unicode(text): function printable_text (line 98) | def printable_text(text): function load_vocab (line 121) | def load_vocab(vocab_file): function convert_by_vocab (line 136) | def convert_by_vocab(vocab, items): function convert_tokens_to_ids (line 144) | def convert_tokens_to_ids(vocab, tokens): function convert_ids_to_tokens (line 148) | def convert_ids_to_tokens(inv_vocab, ids): function whitespace_tokenize (line 152) | def whitespace_tokenize(text): class FullTokenizer (line 161) | class FullTokenizer(object): method __init__ (line 164) | def __init__(self, vocab_file, do_lower_case=True): method tokenize (line 170) | def tokenize(self, text): method convert_tokens_to_ids (line 178) | def convert_tokens_to_ids(self, tokens): method convert_ids_to_tokens (line 181) | def convert_ids_to_tokens(self, ids): class BasicTokenizer (line 185) | class BasicTokenizer(object): method __init__ (line 188) | def __init__(self, do_lower_case=True): method tokenize (line 196) | def tokenize(self, text): method _run_strip_accents (line 220) | def _run_strip_accents(self, text): method _run_split_on_punc (line 231) | def _run_split_on_punc(self, text): method _tokenize_chinese_chars (line 251) | def _tokenize_chinese_chars(self, text): method _is_chinese_char (line 264) | def _is_chinese_char(self, cp): method _clean_text (line 286) | def _clean_text(self, text): class WordpieceTokenizer (line 300) | class WordpieceTokenizer(object): method __init__ (line 303) | def __init__(self, vocab, unk_token="[UNK]", max_input_chars_per_word=... method tokenize (line 308) | def tokenize(self, text): function _is_whitespace (line 362) | def _is_whitespace(char): function _is_control (line 374) | def _is_control(char): function _is_punctuation (line 386) | def _is_punctuation(char):