id: 1 unit: def prefill() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 106 LOC McCabe index: 33 number of parameters: 2 id: 2 unit: def decode() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 64 LOC McCabe index: 21 number of parameters: 2 id: 3 unit: def __init__() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 59 LOC McCabe index: 3 number of parameters: 5 id: 4 unit: def prefill() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 51 LOC McCabe index: 12 number of parameters: 2 id: 5 unit: def _mp_fn() file: optimum/tpu/distributed_model.py start line: 0 end line: 0 size: 42 LOC McCabe index: 11 number of parameters: 4 id: 6 unit: def warmup() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 39 LOC McCabe index: 14 number of parameters: 2 id: 7 unit: def warmup() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 36 LOC McCabe index: 9 number of parameters: 2 id: 8 unit: def _upad_input() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 34 LOC McCabe index: 3 number of parameters: 6 id: 9 unit: def _upad_input() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 34 LOC McCabe index: 3 number of parameters: 6 id: 10 unit: def _upad_input() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 33 LOC McCabe index: 4 number of parameters: 6 id: 11 unit: def _post_generate() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 31 LOC McCabe index: 6 number of parameters: 4 id: 12 unit: def load_hook() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 31 LOC McCabe index: 8 number of parameters: 4 id: 13 unit: def load_hook() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 31 LOC McCabe index: 8 number of parameters: 4 id: 14 unit: def decode() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 30 LOC McCabe index: 19 number of parameters: 2 id: 15 unit: def load_hook() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 29 LOC McCabe index: 7 number of parameters: 4 id: 16 unit: def __init__() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 29 LOC McCabe index: 1 number of parameters: 4 id: 17 unit: def __init__() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 26 LOC McCabe index: 2 number of parameters: 4 id: 18 unit: def _init_rope() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 26 LOC McCabe index: 4 number of parameters: 1 id: 19 unit: def get_fsdp_training_args() file: optimum/tpu/fsdp_v2.py start line: 0 end line: 0 size: 23 LOC McCabe index: 8 number of parameters: 1 id: 20 unit: def _create_dummy_request() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 22 LOC McCabe index: 1 number of parameters: 2 id: 21 unit: def _create_dummy_request() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 22 LOC McCabe index: 1 number of parameters: 2 id: 22 unit: def load_model_info() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/engine_loader.py start line: 0 end line: 0 size: 22 LOC McCabe index: 4 number of parameters: 1 id: 23 unit: def __call__() file: optimum/tpu/generation/logits_process.py start line: 0 end line: 0 size: 22 LOC McCabe index: 8 number of parameters: 2 id: 24 unit: def __init__() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 22 LOC McCabe index: 5 number of parameters: 4 id: 25 unit: def apply_scaling() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/llama_model_exportable_hf.py start line: 0 end line: 0 size: 21 LOC McCabe index: 4 number of parameters: 2 id: 26 unit: def __init__() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 21 LOC McCabe index: 5 number of parameters: 4 id: 27 unit: def __init__() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 21 LOC McCabe index: 5 number of parameters: 4 id: 28 unit: def _token_encode() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 20 LOC McCabe index: 2 number of parameters: 3 id: 29 unit: def __call__() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/logits_process.py start line: 0 end line: 0 size: 20 LOC McCabe index: 8 number of parameters: 2 id: 30 unit: def _find_qparams() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 20 LOC McCabe index: 3 number of parameters: 2 id: 31 unit: def __init__() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 19 LOC McCabe index: 5 number of parameters: 4 id: 32 unit: def forward() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 19 LOC McCabe index: 6 number of parameters: 2 id: 33 unit: def _sample() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/token_selector.py start line: 0 end line: 0 size: 18 LOC McCabe index: 5 number of parameters: 3 id: 34 unit: def forward() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 18 LOC McCabe index: 5 number of parameters: 2 id: 35 unit: def forward() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 18 LOC McCabe index: 4 number of parameters: 2 id: 36 unit: def clear() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 17 LOC McCabe index: 1 number of parameters: 1 id: 37 unit: def assign() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 16 LOC McCabe index: 1 number of parameters: 4 id: 38 unit: def assign() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 16 LOC McCabe index: 1 number of parameters: 4 id: 39 unit: def my_gather() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 16 LOC McCabe index: 4 number of parameters: 4 id: 40 unit: def __init__() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 16 LOC McCabe index: 5 number of parameters: 4 id: 41 unit: def __init__() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 15 LOC McCabe index: 5 number of parameters: 4 id: 42 unit: def __init__() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 15 LOC McCabe index: 1 number of parameters: 0 id: 43 unit: def forward() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 15 LOC McCabe index: 5 number of parameters: 2 id: 44 unit: def clear() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 14 LOC McCabe index: 1 number of parameters: 1 id: 45 unit: def quantize() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 14 LOC McCabe index: 1 number of parameters: 1 id: 46 unit: def quantize() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 14 LOC McCabe index: 1 number of parameters: 1 id: 47 unit: def __init__() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 14 LOC McCabe index: 1 number of parameters: 4 id: 48 unit: def shard_weights() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/engine_loader.py start line: 0 end line: 0 size: 13 LOC McCabe index: 3 number of parameters: 3 id: 49 unit: def forward() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 13 LOC McCabe index: 1 number of parameters: 0 id: 50 unit: def forward() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 13 LOC McCabe index: 1 number of parameters: 0 id: 51 unit: def __init__() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 13 LOC McCabe index: 1 number of parameters: 0 id: 52 unit: def create() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 13 LOC McCabe index: 1 number of parameters: 0 id: 53 unit: def __init__() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 13 LOC McCabe index: 1 number of parameters: 0 id: 54 unit: def create() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 13 LOC McCabe index: 1 number of parameters: 0 id: 55 unit: def forward() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 13 LOC McCabe index: 1 number of parameters: 0 id: 56 unit: def __init__() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 13 LOC McCabe index: 4 number of parameters: 2 id: 57 unit: def forward() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 12 LOC McCabe index: 3 number of parameters: 4 id: 58 unit: def forward() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 12 LOC McCabe index: 1 number of parameters: 0 id: 59 unit: def config_name_to_class() file: optimum/tpu/modeling.py start line: 0 end line: 0 size: 12 LOC McCabe index: 4 number of parameters: 1 id: 60 unit: def forward() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 12 LOC McCabe index: 2 number of parameters: 3 id: 61 unit: def forward() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 12 LOC McCabe index: 1 number of parameters: 0 id: 62 unit: def forward() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 12 LOC McCabe index: 1 number of parameters: 0 id: 63 unit: def forward() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 12 LOC McCabe index: 1 number of parameters: 0 id: 64 unit: def get_export_kwargs_from_env() file: optimum/tpu/model.py start line: 0 end line: 0 size: 12 LOC McCabe index: 3 number of parameters: 0 id: 65 unit: def forward() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 12 LOC McCabe index: 1 number of parameters: 0 id: 66 unit: def forward() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 12 LOC McCabe index: 1 number of parameters: 0 id: 67 unit: def forward() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 12 LOC McCabe index: 1 number of parameters: 0 id: 68 unit: def __init__() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 11 LOC McCabe index: 1 number of parameters: 5 id: 69 unit: def forward() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 11 LOC McCabe index: 3 number of parameters: 3 id: 70 unit: def __init__() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 11 LOC McCabe index: 1 number of parameters: 5 id: 71 unit: def _initialize_affine_weight() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 11 LOC McCabe index: 1 number of parameters: 0 id: 72 unit: def forward() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 11 LOC McCabe index: 3 number of parameters: 3 id: 73 unit: def reset() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 10 LOC McCabe index: 2 number of parameters: 4 id: 74 unit: def from_pretrained() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 10 LOC McCabe index: 1 number of parameters: 5 id: 75 unit: def serve() file: text-generation-inference/server/text_generation_server/cli.py start line: 0 end line: 0 size: 10 LOC McCabe index: 1 number of parameters: 0 id: 76 unit: def pad_to_shard() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/engine_loader.py start line: 0 end line: 0 size: 10 LOC McCabe index: 6 number of parameters: 3 id: 77 unit: def send() file: optimum/tpu/xla_mp_comm.py start line: 0 end line: 0 size: 10 LOC McCabe index: 2 number of parameters: 3 id: 78 unit: def jetstream_pt_available() file: optimum/tpu/jetstream_pt_support.py start line: 0 end line: 0 size: 10 LOC McCabe index: 3 number of parameters: 0 id: 79 unit: def _get_unpad_data() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 10 LOC McCabe index: 1 number of parameters: 1 id: 80 unit: def forward() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 10 LOC McCabe index: 1 number of parameters: 0 id: 81 unit: def forward() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 10 LOC McCabe index: 1 number of parameters: 0 id: 82 unit: def _init_weights() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 10 LOC McCabe index: 5 number of parameters: 2 id: 83 unit: def _get_unpad_data() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 10 LOC McCabe index: 1 number of parameters: 1 id: 84 unit: def forward() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 10 LOC McCabe index: 1 number of parameters: 0 id: 85 unit: def forward() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 10 LOC McCabe index: 1 number of parameters: 0 id: 86 unit: def _init_weights() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 10 LOC McCabe index: 5 number of parameters: 2 id: 87 unit: def _get_unpad_data() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 10 LOC McCabe index: 1 number of parameters: 1 id: 88 unit: def _flash_attention_forward() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 10 LOC McCabe index: 1 number of parameters: 0 id: 89 unit: def _init_weights() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 10 LOC McCabe index: 5 number of parameters: 2 id: 90 unit: def append() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 9 LOC McCabe index: 3 number of parameters: 2 id: 91 unit: def download_weights() file: text-generation-inference/server/text_generation_server/cli.py start line: 0 end line: 0 size: 9 LOC McCabe index: 1 number of parameters: 0 id: 92 unit: def __init__() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/token_selector.py start line: 0 end line: 0 size: 9 LOC McCabe index: 1 number of parameters: 0 id: 93 unit: def select() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/token_selector.py start line: 0 end line: 0 size: 9 LOC McCabe index: 5 number of parameters: 3 id: 94 unit: def __init__() file: optimum/tpu/generation/token_selector.py start line: 0 end line: 0 size: 9 LOC McCabe index: 1 number of parameters: 0 id: 95 unit: def forward() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 9 LOC McCabe index: 1 number of parameters: 0 id: 96 unit: def forward() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 9 LOC McCabe index: 1 number of parameters: 0 id: 97 unit: def prepare_inputs_for_generation() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 9 LOC McCabe index: 1 number of parameters: 0 id: 98 unit: def __init__() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 9 LOC McCabe index: 2 number of parameters: 6 id: 99 unit: def forward() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 9 LOC McCabe index: 1 number of parameters: 0 id: 100 unit: def forward() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 9 LOC McCabe index: 1 number of parameters: 0 id: 101 unit: def prepare_inputs_for_generation() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 9 LOC McCabe index: 1 number of parameters: 0 id: 102 unit: def forward() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 9 LOC McCabe index: 1 number of parameters: 0 id: 103 unit: def forward() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 9 LOC McCabe index: 1 number of parameters: 0 id: 104 unit: def forward() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 9 LOC McCabe index: 1 number of parameters: 0 id: 105 unit: def forward() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 9 LOC McCabe index: 1 number of parameters: 0 id: 106 unit: def prepare_inputs_for_generation() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 9 LOC McCabe index: 1 number of parameters: 0 id: 107 unit: def __init__() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 8 LOC McCabe index: 1 number of parameters: 5 id: 108 unit: def leave() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 8 LOC McCabe index: 2 number of parameters: 1 id: 109 unit: def create() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/token_selector.py start line: 0 end line: 0 size: 8 LOC McCabe index: 1 number of parameters: 0 id: 110 unit: def _select_from_slots() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 8 LOC McCabe index: 2 number of parameters: 3 id: 111 unit: def _cached_batch() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 8 LOC McCabe index: 4 number of parameters: 3 id: 112 unit: def model_can_use_jetstream_pt() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/compatibility.py start line: 0 end line: 0 size: 8 LOC McCabe index: 3 number of parameters: 1 id: 113 unit: def create() file: optimum/tpu/generation/token_selector.py start line: 0 end line: 0 size: 8 LOC McCabe index: 1 number of parameters: 0 id: 114 unit: def from_pretrained() file: optimum/tpu/modeling.py start line: 0 end line: 0 size: 8 LOC McCabe index: 1 number of parameters: 0 id: 115 unit: def _unwrap_model() file: optimum/tpu/fsdp_v2.py start line: 0 end line: 0 size: 8 LOC McCabe index: 4 number of parameters: 1 id: 116 unit: def _find_per_channel_min_max() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 8 LOC McCabe index: 1 number of parameters: 2 id: 117 unit: def my_reduce() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 8 LOC McCabe index: 4 number of parameters: 4 id: 118 unit: def set_quantize() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 8 LOC McCabe index: 1 number of parameters: 1 id: 119 unit: def set_quantize() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 8 LOC McCabe index: 1 number of parameters: 1 id: 120 unit: def leave() file: optimum/tpu/distributed_model.py start line: 0 end line: 0 size: 8 LOC McCabe index: 2 number of parameters: 1 id: 121 unit: def _update_causal_mask() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 8 LOC McCabe index: 1 number of parameters: 0 id: 122 unit: def info() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 7 LOC McCabe index: 1 number of parameters: 1 id: 123 unit: def append() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 7 LOC McCabe index: 1 number of parameters: 2 id: 124 unit: def info() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 7 LOC McCabe index: 2 number of parameters: 1 id: 125 unit: def serve() file: text-generation-inference/server/text_generation_server/server.py start line: 0 end line: 0 size: 7 LOC McCabe index: 1 number of parameters: 0 id: 126 unit: def __init__() file: optimum/tpu/xla_mp_comm.py start line: 0 end line: 0 size: 7 LOC McCabe index: 1 number of parameters: 2 id: 127 unit: def __init__() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 7 LOC McCabe index: 1 number of parameters: 5 id: 128 unit: def _update_causal_mask() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 7 LOC McCabe index: 1 number of parameters: 0 id: 129 unit: def _reorder_cache() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 7 LOC McCabe index: 3 number of parameters: 2 id: 130 unit: def _update_causal_mask() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 7 LOC McCabe index: 1 number of parameters: 0 id: 131 unit: def _reorder_cache() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 7 LOC McCabe index: 3 number of parameters: 2 id: 132 unit: def get_fsdp_config() file: optimum/tpu/fsdp_v2.py start line: 0 end line: 0 size: 7 LOC McCabe index: 1 number of parameters: 1 id: 133 unit: def get_master_weight() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 7 LOC McCabe index: 1 number of parameters: 1 id: 134 unit: def __init__() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 7 LOC McCabe index: 1 number of parameters: 5 id: 135 unit: def __init__() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 7 LOC McCabe index: 1 number of parameters: 5 id: 136 unit: def _reorder_cache() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 7 LOC McCabe index: 3 number of parameters: 2 id: 137 unit: def decode() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 6 LOC McCabe index: 4 number of parameters: 2 id: 138 unit: def patch_precompute_freqs_cis() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/llama_model_exportable_hf.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 1 id: 139 unit: def reset() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 3 id: 140 unit: def create_engine() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/compatibility.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 0 id: 141 unit: def create_engine_env_data() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/engine_loader.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 0 id: 142 unit: def create_engine() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/engine_loader.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 0 id: 143 unit: def __init__() file: optimum/tpu/xla_mp_comm.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 2 id: 144 unit: def select() file: optimum/tpu/generation/token_selector.py start line: 0 end line: 0 size: 6 LOC McCabe index: 2 number of parameters: 3 id: 145 unit: def _sample() file: optimum/tpu/generation/token_selector.py start line: 0 end line: 0 size: 6 LOC McCabe index: 2 number of parameters: 2 id: 146 unit: def apply_rotary_pos_emb() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 6 id: 147 unit: def repeat_kv() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 6 LOC McCabe index: 2 number of parameters: 2 id: 148 unit: def __init__() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 0 id: 149 unit: def update() file: optimum/tpu/static_cache_xla.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 0 id: 150 unit: def get_seq_length() file: optimum/tpu/static_cache_xla.py start line: 0 end line: 0 size: 6 LOC McCabe index: 4 number of parameters: 2 id: 151 unit: def forward() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 2 id: 152 unit: def apply_rotary_pos_emb() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 6 id: 153 unit: def repeat_kv() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 6 LOC McCabe index: 2 number of parameters: 2 id: 154 unit: def __init__() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 0 id: 155 unit: def __init__() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 6 LOC McCabe index: 2 number of parameters: 2 id: 156 unit: def set_g_group() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 0 id: 157 unit: def my_split() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 6 LOC McCabe index: 3 number of parameters: 4 id: 158 unit: def forward() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 2 id: 159 unit: def apply_rotary_pos_emb() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 6 LOC McCabe index: 1 number of parameters: 6 id: 160 unit: def repeat_kv() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 6 LOC McCabe index: 2 number of parameters: 2 id: 161 unit: def __init__() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 6 LOC McCabe index: 2 number of parameters: 2 id: 162 unit: def take_nearest_length() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 5 LOC McCabe index: 2 number of parameters: 1 id: 163 unit: def __init__() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 5 LOC McCabe index: 1 number of parameters: 5 id: 164 unit: def clear() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 5 LOC McCabe index: 4 number of parameters: 2 id: 165 unit: def _clear() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 5 LOC McCabe index: 4 number of parameters: 2 id: 166 unit: def prefill() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 5 LOC McCabe index: 3 number of parameters: 2 id: 167 unit: def __init__() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/gemma_model_hf.py start line: 0 end line: 0 size: 5 LOC McCabe index: 1 number of parameters: 0 id: 168 unit: def precompute_freqs_cis() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/llama_model_exportable_hf.py start line: 0 end line: 0 size: 5 LOC McCabe index: 1 number of parameters: 0 id: 169 unit: def __init__() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/llama_model_exportable_hf.py start line: 0 end line: 0 size: 5 LOC McCabe index: 1 number of parameters: 0 id: 170 unit: def __init__() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/mixtral_model_hf.py start line: 0 end line: 0 size: 5 LOC McCabe index: 1 number of parameters: 0 id: 171 unit: def _get_slot() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 5 LOC McCabe index: 4 number of parameters: 1 id: 172 unit: def clear() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 5 LOC McCabe index: 4 number of parameters: 2 id: 173 unit: def _clear() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 5 LOC McCabe index: 4 number of parameters: 2 id: 174 unit: def __init__() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/engine_loader.py start line: 0 end line: 0 size: 5 LOC McCabe index: 1 number of parameters: 0 id: 175 unit: def from_pretrained() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 5 LOC McCabe index: 1 number of parameters: 4 id: 176 unit: def __init__() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 5 LOC McCabe index: 2 number of parameters: 2 id: 177 unit: def _quantize_to_dtype() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 5 LOC McCabe index: 1 number of parameters: 0 id: 178 unit: def __init__() file: optimum/tpu/distributed_model.py start line: 0 end line: 0 size: 5 LOC McCabe index: 1 number of parameters: 3 id: 179 unit: def from_pretrained() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 5 LOC McCabe index: 2 number of parameters: 4 id: 180 unit: def pause() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 4 LOC McCabe index: 2 number of parameters: 1 id: 181 unit: def __init__() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 0 id: 182 unit: def _cached_batch() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 3 id: 183 unit: def filter() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 4 LOC McCabe index: 3 number of parameters: 3 id: 184 unit: def return_to_caller() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 4 LOC McCabe index: 2 number of parameters: 1 id: 185 unit: def from_config() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/gemma_model_hf.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 3 id: 186 unit: def from_config() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/llama_model_exportable_hf.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 3 id: 187 unit: def from_config() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/mixtral_model_hf.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 3 id: 188 unit: def __init__() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 3 id: 189 unit: def select() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 4 LOC McCabe index: 2 number of parameters: 2 id: 190 unit: def __init__() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 0 id: 191 unit: def filter() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 4 LOC McCabe index: 3 number of parameters: 3 id: 192 unit: def _get_head_dim() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/engine_loader.py start line: 0 end line: 0 size: 4 LOC McCabe index: 2 number of parameters: 1 id: 193 unit: def _pad_array_up_to() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/engine_loader.py start line: 0 end line: 0 size: 4 LOC McCabe index: 2 number of parameters: 3 id: 194 unit: def receive() file: optimum/tpu/xla_mp_comm.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 1 id: 195 unit: def command_data() file: optimum/tpu/xla_mp_comm.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 1 id: 196 unit: def __init__() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 3 id: 197 unit: def forward() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 2 id: 198 unit: def rotate_half() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 1 id: 199 unit: def __init__() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 3 id: 200 unit: def forward() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 3 id: 201 unit: def rotate_half() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 1 id: 202 unit: def quantize_tensor() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 2 id: 203 unit: def get_model_parallel_rank() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 4 LOC McCabe index: 2 number of parameters: 0 id: 204 unit: def get_model_parallel_world_size() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 4 LOC McCabe index: 2 number of parameters: 0 id: 205 unit: def __init__() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 3 id: 206 unit: def rotate_half() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 4 LOC McCabe index: 1 number of parameters: 1 id: 207 unit: def stopped() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 1 id: 208 unit: def info() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 1 id: 209 unit: def filter() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 3 id: 210 unit: def __init__() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/gemma_model_hf.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 3 id: 211 unit: def __init__() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/mixtral_model_hf.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 3 id: 212 unit: def stopped() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 1 id: 213 unit: def instantiate_model_from_repo_id() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/engine_loader.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 0 id: 214 unit: def __init__() file: text-generation-inference/server/text_generation_server/server.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 3 id: 215 unit: def __init__() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 3 LOC McCabe index: 5 number of parameters: 3 id: 216 unit: def _check_module() file: optimum/tpu/cli.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 1 id: 217 unit: def _run() file: optimum/tpu/cli.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 1 id: 218 unit: def install_jetstream_pytorch() file: optimum/tpu/cli.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 0 id: 219 unit: def warning() file: optimum/tpu/xla_logger.py start line: 0 end line: 0 size: 3 LOC McCabe index: 2 number of parameters: 1 id: 220 unit: def info() file: optimum/tpu/xla_logger.py start line: 0 end line: 0 size: 3 LOC McCabe index: 2 number of parameters: 1 id: 221 unit: def debug() file: optimum/tpu/xla_logger.py start line: 0 end line: 0 size: 3 LOC McCabe index: 2 number of parameters: 1 id: 222 unit: def error() file: optimum/tpu/xla_logger.py start line: 0 end line: 0 size: 3 LOC McCabe index: 2 number of parameters: 1 id: 223 unit: def __init__() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 3 LOC McCabe index: 5 number of parameters: 3 id: 224 unit: def from_pretrained() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 4 id: 225 unit: def use_fsdp_v2() file: optimum/tpu/fsdp_v2.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 0 id: 226 unit: def forward() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 5 id: 227 unit: def backward() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 2 id: 228 unit: def forward() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 5 id: 229 unit: def backward() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 2 id: 230 unit: def forward() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 5 id: 231 unit: def backward() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 2 id: 232 unit: def divide_and_check_no_remainder() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 2 id: 233 unit: def prefill() file: optimum/tpu/distributed_model.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 2 id: 234 unit: def decode() file: optimum/tpu/distributed_model.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 2 id: 235 unit: def fetch_model() file: optimum/tpu/model.py start line: 0 end line: 0 size: 3 LOC McCabe index: 1 number of parameters: 0 id: 236 unit: def __init__() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 3 LOC McCabe index: 5 number of parameters: 3 id: 237 unit: def from_pretrained() file: text-generation-inference/server/text_generation_server/auto_generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 0 id: 238 unit: def id() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 239 unit: def state() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 240 unit: def batch_id() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 241 unit: def request_id() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 242 unit: def cached_text() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 243 unit: def generation_config() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 244 unit: def generated_tokens() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 245 unit: def truncate() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 246 unit: def position_id() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 247 unit: def position_id() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 248 unit: def cache_position() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 2 number of parameters: 1 id: 249 unit: def resume() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 250 unit: def _decode_next_tokens() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 0 id: 251 unit: def select() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 3 id: 252 unit: def generated_text() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 253 unit: def next_token() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 2 number of parameters: 1 id: 254 unit: def attention_mask() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 255 unit: def max_token() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 256 unit: def max_new_tokens() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 3 number of parameters: 1 id: 257 unit: def _generate_token() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 0 id: 258 unit: def _post_generate() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 0 id: 259 unit: def _mp_fn() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 0 id: 260 unit: def model_loop_fn() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 261 unit: def warmup() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 262 unit: def clear() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 263 unit: def config() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 264 unit: def __del__() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 265 unit: def from_pretrained() file: text-generation-inference/server/text_generation_server/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 5 id: 266 unit: def block_size() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/mixtral_model_hf.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 267 unit: def n_layer() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/mixtral_model_hf.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 268 unit: def n_head() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/mixtral_model_hf.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 269 unit: def dim() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/mixtral_model_hf.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 270 unit: def n_local_heads() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/mixtral_model_hf.py start line: 0 end line: 0 size: 2 LOC McCabe index: 2 number of parameters: 1 id: 271 unit: def num_activated_experts() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/models/mixtral_model_hf.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 272 unit: def id() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 273 unit: def state() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 274 unit: def batch_id() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 275 unit: def request_id() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 276 unit: def generation_config() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 277 unit: def generated_tokens() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 278 unit: def truncate() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 279 unit: def tokens() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 280 unit: def _decode_next_tokens() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 0 id: 281 unit: def generated_text() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 282 unit: def next_token() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 2 number of parameters: 1 id: 283 unit: def empty() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 284 unit: def seed() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 285 unit: def max_new_tokens() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 3 number of parameters: 1 id: 286 unit: def __init__() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 287 unit: def set() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 288 unit: def select() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 289 unit: def model() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 290 unit: def from_pretrained() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/generator.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 0 id: 291 unit: def from_config() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/logits_process.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 292 unit: def _get_needed_padding() file: text-generation-inference/server/text_generation_server/jetstream_pt_support/engine_loader.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 293 unit: def info() file: text-generation-inference/server/text_generation_server/generator_base.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 294 unit: def warmup() file: text-generation-inference/server/text_generation_server/generator_base.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 295 unit: def prefill() file: text-generation-inference/server/text_generation_server/generator_base.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 296 unit: def decode() file: text-generation-inference/server/text_generation_server/generator_base.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 297 unit: def filter() file: text-generation-inference/server/text_generation_server/generator_base.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 3 id: 298 unit: def clear() file: text-generation-inference/server/text_generation_server/generator_base.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 299 unit: def from_pretrained() file: text-generation-inference/server/text_generation_server/generator_base.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 3 id: 300 unit: def send() file: optimum/tpu/xla_mp_comm.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 301 unit: def from_config() file: optimum/tpu/generation/logits_process.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 302 unit: def _norm() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 303 unit: def forward() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 304 unit: def _flash_attention_forward() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 0 id: 305 unit: def get_input_embeddings() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 306 unit: def set_input_embeddings() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 307 unit: def get_input_embeddings() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 308 unit: def set_input_embeddings() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 309 unit: def get_output_embeddings() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 310 unit: def set_output_embeddings() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 311 unit: def set_decoder() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 312 unit: def get_decoder() file: optimum/tpu/modeling_gemma.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 313 unit: def _install_torch_cpu() file: optimum/tpu/cli.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 0 id: 314 unit: def install_pytorch_xla() file: optimum/tpu/cli.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 0 id: 315 unit: def _flash_attention_forward() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 0 id: 316 unit: def get_input_embeddings() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 317 unit: def set_input_embeddings() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 318 unit: def get_input_embeddings() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 319 unit: def set_input_embeddings() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 320 unit: def get_output_embeddings() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 321 unit: def set_output_embeddings() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 322 unit: def set_decoder() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 323 unit: def get_decoder() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 324 unit: def get_input_embeddings() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 325 unit: def set_input_embeddings() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 326 unit: def get_input_embeddings() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 327 unit: def set_input_embeddings() file: optimum/tpu/modeling_llama.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 328 unit: def get_model_parallel_group() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 0 id: 329 unit: def forward() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 5 id: 330 unit: def backward() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 331 unit: def copy_to_model_parallel_region() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 4 id: 332 unit: def reduce_from_model_parallel_region() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 4 id: 333 unit: def scatter_to_model_parallel_region() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 4 id: 334 unit: def gather_from_model_parallel_region() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 4 id: 335 unit: def ensure_divisibility() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 336 unit: def split_tensor_along_last_dim() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 0 id: 337 unit: def get_master_weight() file: optimum/tpu/xla_model_parallel.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 338 unit: def model_loop_fn() file: optimum/tpu/distributed_model.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 339 unit: def __del__() file: optimum/tpu/distributed_model.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 340 unit: def forward() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 341 unit: def get_input_embeddings() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 342 unit: def set_input_embeddings() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 343 unit: def get_input_embeddings() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 344 unit: def set_input_embeddings() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 345 unit: def get_output_embeddings() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 346 unit: def set_output_embeddings() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 347 unit: def set_decoder() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 348 unit: def get_decoder() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 349 unit: def get_input_embeddings() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 350 unit: def set_input_embeddings() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2 id: 351 unit: def get_input_embeddings() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 1 id: 352 unit: def set_input_embeddings() file: optimum/tpu/modeling_mistral.py start line: 0 end line: 0 size: 2 LOC McCabe index: 1 number of parameters: 2