SYMBOL INDEX (195 symbols across 12 files) FILE: create_pretraining_data.py class TrainingInstance (line 65) | class TrainingInstance(object): method __init__ (line 68) | def __init__(self, tokens, segment_ids, masked_lm_positions, masked_lm... method __str__ (line 76) | def __str__(self): method __repr__ (line 89) | def __repr__(self): function write_instance_to_example_files (line 93) | def write_instance_to_example_files(instances, tokenizer, max_seq_length, function create_int_feature (line 166) | def create_int_feature(values): function create_float_feature (line 171) | def create_float_feature(values): function create_training_instances (line 176) | def create_training_instances(input_files, tokenizer, max_seq_length, function create_instances_from_document (line 220) | def create_instances_from_document( function create_masked_lm_predictions (line 335) | def create_masked_lm_predictions(tokens, masked_lm_prob, function truncate_seq_pair (line 390) | def truncate_seq_pair(tokens_a, tokens_b, max_num_tokens, rng): function main (line 408) | def main(_): FILE: extract_features.py class InputExample (line 81) | class InputExample(object): method __init__ (line 83) | def __init__(self, unique_id, text_a, text_b): class InputFeatures (line 89) | class InputFeatures(object): method __init__ (line 92) | def __init__(self, unique_id, tokens, input_ids, input_mask, input_typ... function input_fn_builder (line 100) | def input_fn_builder(features, seq_length): function model_fn_builder (line 148) | def model_fn_builder(bert_config, init_checkpoint, layer_indexes, use_tpu, function convert_examples_to_features (line 210) | def convert_examples_to_features(examples, seq_length, tokenizer): function _truncate_seq_pair (line 302) | def _truncate_seq_pair(tokens_a, tokens_b, max_length): function read_examples (line 319) | def read_examples(input_file): function main (line 343) | def main(_): FILE: modeling.py class BertConfig (line 30) | class BertConfig(object): method __init__ (line 33) | def __init__(self, method from_dict (line 82) | def from_dict(cls, json_object): method from_json_file (line 90) | def from_json_file(cls, json_file): method to_dict (line 96) | def to_dict(self): method to_json_string (line 101) | def to_json_string(self): class BertModel (line 106) | class BertModel(object): method __init__ (line 130) | def __init__(self, method get_pooled_output (line 235) | def get_pooled_output(self): method get_sequence_output (line 238) | def get_sequence_output(self): method get_all_encoder_layers (line 247) | def get_all_encoder_layers(self): method get_embedding_output (line 250) | def get_embedding_output(self): method get_embedding_table (line 261) | def get_embedding_table(self): function gelu (line 265) | def gelu(input_tensor): function get_activation (line 281) | def get_activation(activation_string): function get_assignment_map_from_checkpoint (line 318) | def get_assignment_map_from_checkpoint(tvars, init_checkpoint): function dropout (line 345) | def dropout(input_tensor, dropout_prob): function layer_norm (line 363) | def layer_norm(input_tensor, name=None): function layer_norm_and_dropout (line 369) | def layer_norm_and_dropout(input_tensor, dropout_prob, name=None): function create_initializer (line 376) | def create_initializer(initializer_range=0.02): function embedding_lookup (line 381) | def embedding_lookup(input_ids, function embedding_postprocessor (line 430) | def embedding_postprocessor(input_tensor, function create_attention_mask_from_input_mask (line 526) | def create_attention_mask_from_input_mask(from_tensor, to_mask): function attention_layer (line 560) | def attention_layer(from_tensor, function transformer_model (line 756) | def transformer_model(input_tensor, function get_shape_list (line 897) | def get_shape_list(tensor, expected_rank=None, name=None): function reshape_to_matrix (line 934) | def reshape_to_matrix(input_tensor): function reshape_from_matrix (line 948) | def reshape_from_matrix(output_tensor, orig_shape_list): function assert_rank (line 961) | def assert_rank(tensor, expected_rank, name=None): FILE: modeling_test.py class BertModelTest (line 29) | class BertModelTest(tf.test.TestCase): class BertModelTester (line 31) | class BertModelTester(object): method __init__ (line 33) | def __init__(self, method create_model (line 71) | def create_model(self): method check_output (line 114) | def check_output(self, result): method test_default (line 126) | def test_default(self): method test_config_to_json_string (line 129) | def test_config_to_json_string(self): method run_tester (line 135) | def run_tester(self, tester): method ids_tensor (line 147) | def ids_tensor(cls, shape, vocab_size, rng=None, name=None): method assert_all_tensors_reachable (line 162) | def assert_all_tensors_reachable(self, sess, outputs): method get_unreachable_ops (line 194) | def get_unreachable_ops(cls, graph, outputs): method flatten_recursive (line 257) | def flatten_recursive(cls, item): FILE: optimization.py function create_optimizer (line 25) | def create_optimizer(loss, init_lr, num_train_steps, num_warmup_steps, u... class AdamWeightDecayOptimizer (line 84) | class AdamWeightDecayOptimizer(tf.train.Optimizer): method __init__ (line 87) | def __init__(self, method apply_gradients (line 105) | def apply_gradients(self, grads_and_vars, global_step=None, name=None): method _do_use_weight_decay (line 156) | def _do_use_weight_decay(self, param_name): method _get_variable_name (line 166) | def _get_variable_name(self, param_name): FILE: optimization_test.py class OptimizationTest (line 23) | class OptimizationTest(tf.test.TestCase): method test_adam (line 25) | def test_adam(self): FILE: run_classifier.py class InputExample (line 127) | class InputExample(object): method __init__ (line 130) | def __init__(self, guid, text_a, text_b=None, label=None): class InputFeatures (line 148) | class InputFeatures(object): method __init__ (line 151) | def __init__(self, input_ids, input_mask, segment_ids, label_id): class DataProcessor (line 158) | class DataProcessor(object): method get_train_examples (line 161) | def get_train_examples(self, data_dir): method get_dev_examples (line 165) | def get_dev_examples(self, data_dir): method get_test_examples (line 169) | def get_test_examples(self, data_dir): method get_labels (line 173) | def get_labels(self): method _read_tsv (line 178) | def _read_tsv(cls, input_file, quotechar=None): class XnliProcessor (line 188) | class XnliProcessor(DataProcessor): method __init__ (line 191) | def __init__(self): method get_train_examples (line 194) | def get_train_examples(self, data_dir): method get_dev_examples (line 213) | def get_dev_examples(self, data_dir): method get_labels (line 231) | def get_labels(self): class MnliProcessor (line 236) | class MnliProcessor(DataProcessor): method get_train_examples (line 239) | def get_train_examples(self, data_dir): method get_dev_examples (line 244) | def get_dev_examples(self, data_dir): method get_test_examples (line 250) | def get_test_examples(self, data_dir): method get_labels (line 255) | def get_labels(self): method _create_examples (line 259) | def _create_examples(self, lines, set_type): class MrpcProcessor (line 277) | class MrpcProcessor(DataProcessor): method get_train_examples (line 280) | def get_train_examples(self, data_dir): method get_dev_examples (line 285) | def get_dev_examples(self, data_dir): method get_test_examples (line 290) | def get_test_examples(self, data_dir): method get_labels (line 295) | def get_labels(self): method _create_examples (line 299) | def _create_examples(self, lines, set_type): class ColaProcessor (line 317) | class ColaProcessor(DataProcessor): method get_train_examples (line 320) | def get_train_examples(self, data_dir): method get_dev_examples (line 325) | def get_dev_examples(self, data_dir): method get_test_examples (line 330) | def get_test_examples(self, data_dir): method get_labels (line 335) | def get_labels(self): method _create_examples (line 339) | def _create_examples(self, lines, set_type): function convert_single_example (line 358) | def convert_single_example(ex_index, example, label_list, max_seq_length, function file_based_convert_examples_to_features (line 450) | def file_based_convert_examples_to_features( function file_based_input_fn_builder (line 477) | def file_based_input_fn_builder(input_file, seq_length, is_training, function _truncate_seq_pair (line 524) | def _truncate_seq_pair(tokens_a, tokens_b, max_length): function create_model (line 541) | def create_model(bert_config, is_training, input_ids, input_mask, segmen... function model_fn_builder (line 586) | def model_fn_builder(bert_config, num_labels, init_checkpoint, learning_... function input_fn_builder (line 671) | def input_fn_builder(features, seq_length, is_training, drop_remainder): function convert_examples_to_features (line 725) | def convert_examples_to_features(examples, label_list, max_seq_length, function main (line 741) | def main(_): FILE: run_lm_predict.py class InputExample (line 97) | class InputExample(object): method __init__ (line 98) | def __init__(self, unique_id, text): function read_examples (line 103) | def read_examples(input_file): function model_fn_builder (line 120) | def model_fn_builder(bert_config, init_checkpoint, use_tpu, function get_masked_lm_output (line 185) | def get_masked_lm_output(bert_config, input_tensor, output_weights, posi... function gather_indexes (line 223) | def gather_indexes(sequence_tensor, positions): function input_fn_builder (line 239) | def input_fn_builder(features, seq_length, max_predictions_per_seq): function convert_examples_to_features (line 299) | def convert_examples_to_features(examples, max_seq_length, tokenizer): function create_masked_lm_prediction (line 322) | def create_masked_lm_prediction(input_ids, mask_position, mask_count=1): class InputFeatures (line 332) | class InputFeatures(object): method __init__ (line 335) | def __init__(self, input_ids, segment_ids, input_mask, masked_lm_posit... function convert_single_example (line 344) | def convert_single_example(ex_index, example, max_seq_length, function is_subtoken (line 393) | def is_subtoken(x): function create_sequential_mask (line 396) | def create_sequential_mask(input_tokens, input_ids, input_mask, segment_... function parse_result (line 422) | def parse_result(result, all_tokens, output_file=None): function main (line 462) | def main(_): FILE: run_pretraining.py function model_fn_builder (line 109) | def model_fn_builder(bert_config, init_checkpoint, learning_rate, function get_masked_lm_output (line 240) | def get_masked_lm_output(bert_config, input_tensor, output_weights, posi... function get_next_sentence_output (line 285) | def get_next_sentence_output(bert_config, input_tensor, labels): function gather_indexes (line 308) | def gather_indexes(sequence_tensor, positions): function input_fn_builder (line 324) | def input_fn_builder(input_files, function _decode_record (line 391) | def _decode_record(record, name_to_features): function main (line 406) | def main(_): FILE: run_squad.py class SquadExample (line 157) | class SquadExample(object): method __init__ (line 163) | def __init__(self, method __str__ (line 179) | def __str__(self): method __repr__ (line 182) | def __repr__(self): class InputFeatures (line 197) | class InputFeatures(object): method __init__ (line 200) | def __init__(self, function read_squad_examples (line 227) | def read_squad_examples(input_file, is_training): function convert_examples_to_features (line 309) | def convert_examples_to_features(examples, tokenizer, max_seq_length, function _improve_answer_span (line 476) | def _improve_answer_span(doc_tokens, input_start, input_end, tokenizer, function _check_is_max_context (line 513) | def _check_is_max_context(doc_spans, cur_span_index, position): function create_model (line 550) | def create_model(bert_config, is_training, input_ids, input_mask, segmen... function model_fn_builder (line 590) | def model_fn_builder(bert_config, init_checkpoint, learning_rate, function input_fn_builder (line 687) | def input_fn_builder(input_file, seq_length, is_training, drop_remainder): function write_predictions (line 741) | def write_predictions(all_examples, all_features, all_results, n_best_size, function get_final_text (line 927) | def get_final_text(pred_text, orig_text, do_lower_case): function _get_best_indexes (line 1023) | def _get_best_indexes(logits, n_best_size): function _compute_softmax (line 1035) | def _compute_softmax(scores): class FeatureWriter (line 1058) | class FeatureWriter(object): method __init__ (line 1061) | def __init__(self, filename, is_training): method process_feature (line 1067) | def process_feature(self, feature): method close (line 1093) | def close(self): function validate_flags_or_throw (line 1097) | def validate_flags_or_throw(bert_config): function main (line 1123) | def main(_): FILE: tokenization.py function convert_to_unicode (line 27) | def convert_to_unicode(text): function printable_text (line 47) | def printable_text(text): function load_vocab (line 70) | def load_vocab(vocab_file): function convert_by_vocab (line 85) | def convert_by_vocab(vocab, items): function convert_tokens_to_ids (line 93) | def convert_tokens_to_ids(vocab, tokens): function convert_ids_to_tokens (line 97) | def convert_ids_to_tokens(inv_vocab, ids): function whitespace_tokenize (line 101) | def whitespace_tokenize(text): class FullTokenizer (line 110) | class FullTokenizer(object): method __init__ (line 113) | def __init__(self, vocab_file, do_lower_case=True): method tokenize (line 119) | def tokenize(self, text): method convert_tokens_to_ids (line 127) | def convert_tokens_to_ids(self, tokens): method convert_ids_to_tokens (line 130) | def convert_ids_to_tokens(self, ids): class BasicTokenizer (line 134) | class BasicTokenizer(object): method __init__ (line 137) | def __init__(self, do_lower_case=True): method tokenize (line 145) | def tokenize(self, text): method _run_strip_accents (line 169) | def _run_strip_accents(self, text): method _run_split_on_punc (line 180) | def _run_split_on_punc(self, text): method _tokenize_chinese_chars (line 200) | def _tokenize_chinese_chars(self, text): method _is_chinese_char (line 213) | def _is_chinese_char(self, cp): method _clean_text (line 235) | def _clean_text(self, text): class WordpieceTokenizer (line 249) | class WordpieceTokenizer(object): method __init__ (line 252) | def __init__(self, vocab, unk_token="[UNK]", max_input_chars_per_word=... method tokenize (line 257) | def tokenize(self, text): function _is_whitespace (line 311) | def _is_whitespace(char): function _is_control (line 323) | def _is_control(char): function _is_punctuation (line 335) | def _is_punctuation(char): FILE: tokenization_test.py class TokenizationTest (line 26) | class TokenizationTest(tf.test.TestCase): method test_full_tokenizer (line 28) | def test_full_tokenizer(self): method test_chinese (line 47) | def test_chinese(self): method test_basic_tokenizer_lower (line 54) | def test_basic_tokenizer_lower(self): method test_basic_tokenizer_no_lower (line 62) | def test_basic_tokenizer_no_lower(self): method test_wordpiece_tokenizer (line 69) | def test_wordpiece_tokenizer(self): method test_convert_tokens_to_ids (line 89) | def test_convert_tokens_to_ids(self): method test_is_whitespace (line 103) | def test_is_whitespace(self): method test_is_control (line 113) | def test_is_control(self): method test_is_punctuation (line 121) | def test_is_punctuation(self):