SYMBOL INDEX (19 symbols across 7 files) FILE: examples/question_answering.py function question_answering_example (line 3) | def question_answering_example(): FILE: examples/text_generation.py function text_generation_example (line 3) | def text_generation_example(): FILE: run_inference.py function run_inference (line 6) | def run_inference(model_name, prompt): FILE: src/amd_setup.py function setup_amd_environment (line 4) | def setup_amd_environment(): function optimize_for_amd (line 21) | def optimize_for_amd(): FILE: src/engine.py class InferenceEngine (line 5) | class InferenceEngine: method __init__ (line 6) | def __init__(self, model_path, seed=42): method run_inference (line 11) | def run_inference(self, prompt, max_length=100, temperature=0.7): method batch_inference (line 38) | def batch_inference(self, prompts, **kwargs): FILE: src/model.py class LlamaModel (line 4) | class LlamaModel: method __init__ (line 5) | def __init__(self, model_name): method generate (line 19) | def generate(self, prompt, max_length=100, temperature=0.7): method __call__ (line 35) | def __call__(self, prompt, **kwargs): FILE: src/utils.py function set_seed (line 5) | def set_seed(seed): function format_time (line 12) | def format_time(time_ms): function get_gpu_memory_usage (line 23) | def get_gpu_memory_usage(): function print_gpu_utilization (line 30) | def print_gpu_utilization(): function is_amd_gpu (line 40) | def is_amd_gpu(): function get_gpu_info (line 44) | def get_gpu_info():