SYMBOL INDEX (67 symbols across 8 files) FILE: dino.py function validate_args (line 28) | def validate_args(args) -> None: function validate_task_spec (line 41) | def validate_task_spec(task_spec: Dict[str, Any], with_inputs: bool) -> ... FILE: generation.py class SelfDebiasingLogitsProcessor (line 25) | class SelfDebiasingLogitsProcessor(LogitsProcessor): method __init__ (line 28) | def __init__(self, num_debiasing_prefixes: int, decay_constant: float ... method __call__ (line 44) | def __call__(self, input_ids: torch.LongTensor, scores: torch.FloatTen... method _get_bias_indices (line 53) | def _get_bias_indices(self, regular_sentence_idx: int, batch_size: int... method _debias_scores (line 57) | def _debias_scores(self, scores: torch.FloatTensor, regular_sent_idx: ... method _apply_decay_mask (line 67) | def _apply_decay_mask(self, logits: torch.Tensor, decay_mask: torch.Te... method _generate_decay_mask (line 76) | def _generate_decay_mask(self, logits_regular: torch.FloatTensor, logi... method _get_most_likely_tokens (line 101) | def _get_most_likely_tokens(self, probabilities_tensor: torch.Tensor, ... class SelfDebiasingGPT2LMHeadModel (line 109) | class SelfDebiasingGPT2LMHeadModel(GPT2LMHeadModel, GenerationMixin): method __init__ (line 115) | def __init__(self, *args, **kwargs): method init_logits_processor (line 119) | def init_logits_processor(self, *args, **kwargs): method _get_logits_processor (line 123) | def _get_logits_processor(self, *args, **kwargs) -> LogitsProcessorList: method beam_sample (line 129) | def beam_sample(self, *args, **kwargs): method sample (line 132) | def sample(self, input_ids: torch.LongTensor, logits_processor: Option... FILE: modeling.py class DinoGenerator (line 32) | class DinoGenerator: method __init__ (line 37) | def __init__(self, task_spec: Dict[str, Any], model: Union['str', 'Mod... method generate_dataset (line 76) | def generate_dataset(self, input_texts: Optional[List[str]], num_entri... method _generate_dataset_entries (line 104) | def _generate_dataset_entries(self, input_text_or_id: Union[str, int],... method _build_instruction (line 138) | def _build_instruction(self, label: str, text: str, generate_with_inpu... method _process_output (line 150) | def _process_output(self, input_text: Union[str, int], output_text: st... method _postprocess_dataset (line 159) | def _postprocess_dataset(self, dataset: List[DatasetEntry], generate_w... class ModelWrapper (line 182) | class ModelWrapper(ABC): method __init__ (line 188) | def __init__(self, use_cuda: bool = True): method query_model (line 196) | def query_model(self, input_text: str) -> torch.FloatTensor: method query_model_batch (line 201) | def query_model_batch(self, input_texts: List[str]) -> torch.FloatTensor: method generate (line 206) | def generate(self, input_text: str, **kwargs) -> str: method generate_self_debiasing (line 211) | def generate_self_debiasing(self, input_text: str, debiasing_texts: Li... class GPT2Wrapper (line 226) | class GPT2Wrapper(ModelWrapper): method __init__ (line 228) | def __init__(self, model_name: str = "gpt2-xl", use_cuda: bool = True): method query_model_batch (line 241) | def query_model_batch(self, input_texts: List[str]): method generate (line 248) | def generate(self, input_text: str, **kwargs): method generate_self_debiasing (line 253) | def generate_self_debiasing(self, input_text: str, debiasing_texts: Li... FILE: scripts/imdb/run_supervised.py class IMDbDataset (line 31) | class IMDbDataset(torch.utils.data.Dataset): method __init__ (line 32) | def __init__(self, encodings, labels): method __getitem__ (line 36) | def __getitem__(self, idx): method __len__ (line 41) | def __len__(self): function compute_metrics (line 45) | def compute_metrics(eval_pred): function load_datasets (line 51) | def load_datasets(args) -> Dict[str, IMDbDataset]: FILE: scripts/imdb/run_unsupervised.py class CausalLMWrapper (line 29) | class CausalLMWrapper: method __init__ (line 32) | def __init__(self, model_name: str, use_cuda: bool = True): method get_token_probabilities (line 39) | def get_token_probabilities(self, input_text: str, prompt: str) -> tor... class MaskedLMWrapper (line 47) | class MaskedLMWrapper: method __init__ (line 50) | def __init__(self, model_name: str, use_cuda: bool = True): method get_token_probabilities (line 55) | def get_token_probabilities(self, input_text: str, prompt: str) -> tor... class GPT3Wrapper (line 72) | class GPT3Wrapper: method __init__ (line 75) | def __init__(self, engine: str): method get_scores (line 78) | def get_scores(self, prompt: str): FILE: scripts/sts/postprocess_dataset.py function postprocess_dataset (line 9) | def postprocess_dataset( FILE: scripts/sts/run_training.py function download_sts_dataset (line 22) | def download_sts_dataset(sts_dataset_path: str) -> None: function set_seed (line 28) | def set_seed(seed: int) -> None: function build_sentence_transformer (line 37) | def build_sentence_transformer(model_name: str) -> SentenceTransformer: function split_dataset (line 48) | def split_dataset(ds: List[DatasetEntry], dev_size: float = 0.1, seed: i... FILE: utils.py function set_seed (line 26) | def set_seed(seed: int) -> None: function read_inputs (line 35) | def read_inputs(input_file: str, input_file_type: str) -> List[str]: function read_plaintext_inputs (line 58) | def read_plaintext_inputs(path: str) -> List[str]: function read_jsonl_inputs (line 66) | def read_jsonl_inputs(path: str) -> List[str]: function read_sts_inputs (line 73) | def read_sts_inputs(path: str) -> List[str]: class DatasetEntry (line 89) | class DatasetEntry: method __init__ (line 92) | def __init__(self, text_a: str, text_b: Optional[str], label: Any): method __repr__ (line 97) | def __repr__(self): method __key (line 103) | def __key(self): method __hash__ (line 106) | def __hash__(self): method __eq__ (line 109) | def __eq__(self, other): method save_list (line 115) | def save_list(entries: List['DatasetEntry'], path: str): method read_list (line 121) | def read_list(path: str) -> List['DatasetEntry']: