SYMBOL INDEX (118 symbols across 11 files) FILE: finetune/mc/preprocess_medqa.py function dump_jsonl (line 13) | def dump_jsonl(data, fpath): function process_medqa (line 18) | def process_medqa(fname): FILE: finetune/mc/run_multiple_choice.py class ModelArguments (line 61) | class ModelArguments: class DataTrainingArguments (line 105) | class DataTrainingArguments: method __post_init__ (line 160) | def __post_init__(self): class DataCollatorForMultipleChoice (line 172) | class DataCollatorForMultipleChoice: method __call__ (line 200) | def __call__(self, features): function main (line 225) | def main(): function _mp_fn (line 520) | def _mp_fn(index): FILE: finetune/seqcls/preprocess_blurb_seqcls.py function dump_jsonl (line 11) | def dump_jsonl(data, fpath): function process_pubmedqa (line 22) | def process_pubmedqa(fname): function process_bioasq (line 50) | def process_bioasq(fname): FILE: finetune/seqcls/run_seqcls_gpt.py class DataTrainingArguments (line 79) | class DataTrainingArguments: method __post_init__ (line 156) | def __post_init__(self): class ModelArguments (line 175) | class ModelArguments: function main (line 213) | def main(): function _mp_fn (line 632) | def _mp_fn(index): FILE: finetune/textgen/gpt2/finetune_for_summarization.py class ModelArguments (line 32) | class ModelArguments: class DataArguments (line 56) | class DataArguments: function get_dataset (line 96) | def get_dataset( function finetune (line 121) | def finetune(): FILE: finetune/textgen/gpt2/run_generation_batch.py function prepare_ctrl_input (line 115) | def prepare_ctrl_input(args, _, tokenizer, prompt_text): function prepare_xlm_input (line 125) | def prepare_xlm_input(args, model, tokenizer, prompt_text): function prepare_xlnet_input (line 151) | def prepare_xlnet_input(args, _, tokenizer, prompt_text): function prepare_transfoxl_input (line 157) | def prepare_transfoxl_input(args, _, tokenizer, prompt_text): function read_e2e_files (line 170) | def read_e2e_files(path, tokenizer, lowdata_token=None): function read_wp_files (line 187) | def read_wp_files(path, tokenizer): function read_classifySentiment_files (line 199) | def read_classifySentiment_files(path, tokenizer): function read_classifyTopic_files (line 209) | def read_classifyTopic_files(path, tokenizer): function lmap (line 232) | def lmap(f, x): function ids_to_clean_text (line 236) | def ids_to_clean_text(tokenizer, generated_ids): function flatten_list (line 244) | def flatten_list(summary_ids): function calculate_rouge (line 247) | def calculate_rouge(output_lns, reference_lns, use_stemmer=True): function test_epoch_end (line 258) | def test_epoch_end(outputs, prefix="test"): function test_step (line 282) | def test_step(model, gpt2, batch, batch_idx, args, tokenizer, beam_handl... function read_webnlg_files (line 393) | def read_webnlg_files(path, tokenizer): function read_triples_files2 (line 435) | def read_triples_files2(path, tokenizer): function read_triples_files (line 473) | def read_triples_files(path, tokenizer): function write_e2e_corr (line 515) | def write_e2e_corr(prompt_lst, file_dict, corr_path): function write_e2e_src (line 547) | def write_e2e_src(prompt_lst, corr_path): function get_emb (line 555) | def get_emb(sent_lst, word_lst, num_layer=1): function adjust_length_to_model (line 664) | def adjust_length_to_model(length, max_sequence_length): function read_doc_for_embmatch (line 674) | def read_doc_for_embmatch(file_name, num_layer): function main (line 688) | def main(): FILE: finetune/textgen/gpt2/sum_data_collator.py class DataCollatorForSumLanguageModeling (line 10) | class DataCollatorForSumLanguageModeling: method __call__ (line 21) | def __call__( method _tensorize_batch (line 65) | def _tensorize_batch( class DataCollatorForSumBatchGenLanguageModeling (line 85) | class DataCollatorForSumBatchGenLanguageModeling: method __call__ (line 99) | def __call__( method _tensorize_batch (line 141) | def _tensorize_batch( FILE: finetune/textgen/gpt2/sum_dataset.py class LineByLineSumTextDataset (line 25) | class LineByLineSumTextDataset(Dataset): method __init__ (line 31) | def __init__(self, tokenizer: PreTrainedTokenizer, file_path: str, blo... method __len__ (line 130) | def __len__(self): method __getitem__ (line 134) | def __getitem__(self, i): class LineByLineSumBatchGenTextDataset (line 177) | class LineByLineSumBatchGenTextDataset(Dataset): method __init__ (line 183) | def __init__(self, tokenizer: PreTrainedTokenizer, file_path: str, blo... method __len__ (line 218) | def __len__(self): method __getitem__ (line 222) | def __getitem__(self, i): FILE: finetune/utils/custom_modeling_gpt2.py class GPT2ForTokenClassification (line 57) | class GPT2ForTokenClassification(GPT2PreTrainedModel): method __init__ (line 58) | def __init__(self, config): method forward (line 79) | def forward( class GPT2ForMultipleChoice (line 137) | class GPT2ForMultipleChoice(GPT2PreTrainedModel): method __init__ (line 140) | def __init__(self, config): method forward (line 164) | def forward( class GPT2ForSequenceClassification (line 253) | class GPT2ForSequenceClassification(GPT2PreTrainedModel): method __init__ (line 256) | def __init__(self, config): method forward (line 274) | def forward( FILE: finetune/utils/custom_modeling_gpt_neo.py function load_tf_weights_in_gpt_neo (line 53) | def load_tf_weights_in_gpt_neo(model, config, gpt_neo_checkpoint_path): class GPTNeoAttentionMixin (line 137) | class GPTNeoAttentionMixin: method _get_block_length_and_num_blocks (line 143) | def _get_block_length_and_num_blocks(seq_length, window_size): method _look_back (line 155) | def _look_back(tensor, block_length, window_size, pad_value=0, is_key_... method _split_seq_length_dim_to (line 196) | def _split_seq_length_dim_to(tensors, dim_factor_1, dim_factor_2): method create_local_attention_mask (line 211) | def create_local_attention_mask(batch_size, seq_length, window_size, d... method _split_heads (line 246) | def _split_heads(self, tensor, num_heads, attn_head_size): method _merge_heads (line 259) | def _merge_heads(self, tensor, num_heads, attn_head_size): method _attn (line 272) | def _attn(self, query, key, value, causal_mask, masked_bias, attn_drop... class GPTNeoSelfAttention (line 298) | class GPTNeoSelfAttention(nn.Module, GPTNeoAttentionMixin): method __init__ (line 299) | def __init__(self, config): method forward (line 327) | def forward( class GPTNeoLocalSelfAttention (line 374) | class GPTNeoLocalSelfAttention(nn.Module, GPTNeoAttentionMixin): method __init__ (line 375) | def __init__(self, config): method forward (line 398) | def forward( class GPTNeoAttention (line 472) | class GPTNeoAttention(nn.Module): method __init__ (line 473) | def __init__(self, config, layer_id=0): method forward (line 489) | def forward( class GPTNeoMLP (line 518) | class GPTNeoMLP(nn.Module): method __init__ (line 519) | def __init__(self, intermediate_size, config): # in MLP: intermediate... method forward (line 527) | def forward(self, hidden_states): class GPTNeoBlock (line 535) | class GPTNeoBlock(nn.Module): method __init__ (line 536) | def __init__(self, config, layer_id): method forward (line 545) | def forward( class GPTNeoPreTrainedModel (line 583) | class GPTNeoPreTrainedModel(PreTrainedModel): method __init__ (line 593) | def __init__(self, *inputs, **kwargs): method _init_weights (line 596) | def _init_weights(self, module): class GPTNeoModel (line 701) | class GPTNeoModel(GPTNeoPreTrainedModel): method __init__ (line 702) | def __init__(self, config): method get_input_embeddings (line 714) | def get_input_embeddings(self): method set_input_embeddings (line 717) | def set_input_embeddings(self, new_embeddings): method forward (line 727) | def forward( class GPTNeoForCausalLM (line 900) | class GPTNeoForCausalLM(GPTNeoPreTrainedModel): method __init__ (line 904) | def __init__(self, config): method get_output_embeddings (line 911) | def get_output_embeddings(self): method set_output_embeddings (line 914) | def set_output_embeddings(self, new_embeddings): method prepare_inputs_for_generation (line 917) | def prepare_inputs_for_generation(self, input_ids, past=None, **kwargs): method forward (line 952) | def forward( method _reorder_cache (line 1021) | def _reorder_cache(past: Tuple[Tuple[torch.Tensor]], beam_idx: torch.T... class GPTNeoForSequenceClassification (line 1048) | class GPTNeoForSequenceClassification(GPTNeoPreTrainedModel): method __init__ (line 1051) | def __init__(self, config): method forward (line 1066) | def forward( FILE: finetune/utils/hf_flash_gpt_2.py class GPT2FlashAttention (line 37) | class GPT2FlashAttention(GPT2Attention): method __init__ (line 38) | def __init__(self, config, is_cross_attention=False, layer_idx=None): method _attn (line 42) | def _attn(self, query, key, value, attention_mask=None, head_mask=None): class GPT2FlashBlock (line 72) | class GPT2FlashBlock(GPT2Block): method __init__ (line 73) | def __init__(self, config, layer_idx=None): class GPT2FlashModel (line 89) | class GPT2FlashModel(GPT2Model): method __init__ (line 90) | def __init__(self, config): class GPT2FlashLMHeadModel (line 111) | class GPT2FlashLMHeadModel(GPT2LMHeadModel): method __init__ (line 112) | def __init__(self, config):