def load_llama2()

in cookbook-efforts/kto-preference/preference_gen.py [0:0]


def load_llama2(task: Task) -> LLM:
    return InferenceEndpointsLLM(
        "meta-llama/Llama-2-70b-chat-hf",
        token=HF_TOKEN,
        task=task,
        max_new_tokens=512,
        prompt_format="llama2",
    )