in llm_perf/benchmark_runners/cpu/update_llm_perf_cpu_onnxruntime.py [0:0]
def get_list_of_benchmarks_to_run(self) -> List[Dict[str, Any]]:
return [
{
"model": model,
"attn_implementation": attn_impl,
"weights_config": weights_cfg,
}
for model, attn_impl, weights_cfg in product(
CANONICAL_PRETRAINED_OPEN_LLM_LIST,
self.attention_configs,
self.weights_configs.keys(),
)
]