SYMBOL INDEX (96 symbols across 14 files) FILE: Trainer.py function set_seed (line 82) | def set_seed(seed): function get_dataloader (line 89) | def get_dataloader(dataset, args, mode='train'): function get_optimizer (line 126) | def get_optimizer(model, args, num_training_steps): function print_log (line 153) | def print_log(logs, epoch, global_step, eval_type, tb_writer, iterator=N... function train (line 182) | def train(model, args, train_dataset, dev_dataset, test_dataset, label_v... function evaluate (line 403) | def evaluate(model, args, dataset, label_vocab, global_step, description... function main (line 480) | def main(): FILE: feature/task_dataset.py class TaskDataset (line 17) | class TaskDataset(Dataset): method __init__ (line 18) | def __init__(self, file, params, do_shuffle=False): method init_np_dataset (line 51) | def init_np_dataset(self): method __len__ (line 188) | def __len__(self): method __getitem__ (line 191) | def __getitem__(self, index): FILE: feature/vocab.py class ItemVocabFile (line 20) | class ItemVocabFile(): method __init__ (line 25) | def __init__(self, files, is_word=False, has_default=False, unk_num=0): method init_vocab (line 46) | def init_vocab(self): method get_item_size (line 60) | def get_item_size(self): method convert_item_to_id (line 63) | def convert_item_to_id(self, item): method convert_items_to_ids (line 77) | def convert_items_to_ids(self, items): method convert_id_to_item (line 80) | def convert_id_to_item(self, id): method convert_ids_to_items (line 83) | def convert_ids_to_items(self, ids): class ItemVocabArray (line 88) | class ItemVocabArray(): method __init__ (line 93) | def __init__(self, items_array, is_word=False, has_default=False, unk_... method init_vocab (line 114) | def init_vocab(self): method get_item_size (line 120) | def get_item_size(self): method convert_item_to_id (line 123) | def convert_item_to_id(self, item): method convert_items_to_ids (line 137) | def convert_items_to_ids(self, items): method convert_id_to_item (line 140) | def convert_id_to_item(self, id): method convert_ids_to_items (line 143) | def convert_ids_to_items(self, ids): FILE: function/format_convert.py function BMES_to_json (line 14) | def BMES_to_json(bmes_file, json_file): FILE: function/metrics.py function seq_f1_with_mask (line 8) | def seq_f1_with_mask(all_true_labels, all_pred_labels, all_label_mask, l... FILE: function/preprocess.py function sent_to_matched_words_boundaries (line 12) | def sent_to_matched_words_boundaries(sent, lexicon_tree, max_word_num=No... function sent_to_distinct_matched_words (line 89) | def sent_to_distinct_matched_words(sent, lexicon_tree): function sent_to_matched_words (line 126) | def sent_to_matched_words(sent, lexicon_tree, max_word_num=None): function sent_to_matched_words_set (line 155) | def sent_to_matched_words_set(sent, lexicon_tree, max_word_num=None): function get_corpus_matched_word_from_vocab_files (line 170) | def get_corpus_matched_word_from_vocab_files(files, vocab_files, scan_nu... function get_corpus_matched_word_from_lexicon_tree (line 212) | def get_corpus_matched_word_from_lexicon_tree(files, lexicon_tree): function insert_seg_vocab_to_lexicon_tree (line 251) | def insert_seg_vocab_to_lexicon_tree(seg_vocab, word_vocab, lexicon_tree): function build_lexicon_tree_from_vocabs (line 294) | def build_lexicon_tree_from_vocabs(vocab_files, scan_nums=None): function get_all_labels_from_corpus (line 326) | def get_all_labels_from_corpus(files, label_file, defalut_label='O'): FILE: function/utils.py function load_pretrain_embed (line 13) | def load_pretrain_embed(embedding_path, max_scan_num=1000000, add_seg_vo... function build_pretrained_embedding_for_corpus (line 57) | def build_pretrained_embedding_for_corpus( function reverse_padded_sequence (line 107) | def reverse_padded_sequence(inputs, lengths, batch_first=True): function random_embedding (line 140) | def random_embedding(self, vocab_size, embedding_dim): function gather_indexes (line 148) | def gather_indexes(sequence_tensor, positions): function save_preds_for_seq_labelling (line 175) | def save_preds_for_seq_labelling(token_ids, tokenizer, true_labels, pred... FILE: module/bilstm.py class BiLSTM (line 13) | class BiLSTM(nn.Module): method __init__ (line 14) | def __init__(self, input_size, hidden_size, dropout): method forward (line 21) | def forward(self, inputs, attention_mask): FILE: module/crf.py function log_sum_exp (line 16) | def log_sum_exp(vec, m_size): class CRF (line 29) | class CRF(nn.Module): method __init__ (line 30) | def __init__(self, tagset_size, gpu): method _calculate_PZ (line 43) | def _calculate_PZ(self, feats, mask): method _viterbi_decode (line 80) | def _viterbi_decode(self, feats, mask): method forward (line 148) | def forward(self, feats): method _score_sentence (line 152) | def _score_sentence(self, scores, mask, tags): method neg_log_likelihood_loss (line 188) | def neg_log_likelihood_loss(self, feats, mask, tags): FILE: module/lexicon_tree.py class TrieNode (line 8) | class TrieNode: method __init__ (line 9) | def __init__(self): class Trie (line 13) | class Trie: method __init__ (line 22) | def __init__(self, use_single=True): method insert (line 30) | def insert(self, word): method search (line 40) | def search(self, word): method enumerateMatch (line 49) | def enumerateMatch(self, str, space=""): FILE: module/sampler.py class SequentialDistributedSampler (line 13) | class SequentialDistributedSampler(Sampler): method __init__ (line 23) | def __init__(self, dataset, num_replicas=None, rank=None, do_shuffle=F... method __iter__ (line 39) | def __iter__(self): method __len__ (line 51) | def __len__(self): FILE: to_json.py function convert (line 10) | def convert(task_name, mode="train"): FILE: wcbert_modeling.py class BertEmbeddings (line 38) | class BertEmbeddings(nn.Module): method __init__ (line 42) | def __init__(self, config): method forward (line 57) | def forward(self, input_ids=None, token_type_ids=None, position_ids=No... class BertLayer (line 88) | class BertLayer(nn.Module): method __init__ (line 93) | def __init__(self, config, has_word_attn=False): method forward (line 121) | def forward( method feed_forward_chunk (line 197) | def feed_forward_chunk(self, attention_output): class BertEncoder (line 203) | class BertEncoder(nn.Module): method __init__ (line 204) | def __init__(self, config): method forward (line 218) | def forward( class BertPooler (line 282) | class BertPooler(nn.Module): method __init__ (line 283) | def __init__(self, config): method forward (line 288) | def forward(self, hidden_states): class BertPreTrainedModel (line 296) | class BertPreTrainedModel(PreTrainedModel): method _init_weights (line 306) | def _init_weights(self, module): class WCBertModel (line 319) | class WCBertModel(BertPreTrainedModel): method __init__ (line 320) | def __init__(self, config, add_pooling_layer=True): method get_input_embeddings (line 329) | def get_input_embeddings(self): method set_input_embeddings (line 332) | def set_input_embeddings(self, value): method _prune_heads (line 335) | def _prune_heads(self, heads_to_prune): method forward (line 343) | def forward( class WCBertCRFForTokenClassification (line 459) | class WCBertCRFForTokenClassification(BertPreTrainedModel): method __init__ (line 460) | def __init__(self, config, pretrained_embeddings, num_labels): method forward (line 478) | def forward( class BertWordLSTMCRFForTokenClassification (line 513) | class BertWordLSTMCRFForTokenClassification(BertPreTrainedModel): method __init__ (line 519) | def __init__(self, config, pretrained_embeddings, num_labels): method forward (line 548) | def forward( FILE: wcbert_parser.py function get_argparse (line 7) | def get_argparse():