in cookbook-efforts/kto-preference/preference_gen.py [0:0]
def load_llama2(task: Task) -> LLM:
return InferenceEndpointsLLM(
"meta-llama/Llama-2-70b-chat-hf",
token=HF_TOKEN,
task=task,
max_new_tokens=512,
prompt_format="llama2",
)