in src/lighteval/metrics/llm_as_judge.py [0:0]
def __call_api(self, prompt):
for _ in range(self.API_MAX_RETRY):
try:
# Base model
response = self.client.beta.chat.completions.parse(
model=self.model,
messages=as_list(prompt),
response_format=self.response_format,
max_tokens=4096,
temperature=0.0,
n=1,
)
answer = response.choices[0].message.parsed
return answer
except TypeError:
try:
# Finetune
response = self.client.chat.completions.create(
model=self.model,
messages=as_list(prompt),
response_format=self.response_format,
max_tokens=512,
n=1,
)
text = response.choices[0].message.content
return text
except Exception as e:
logger.warning(f"{type(e), e}")
time.sleep(self.API_RETRY_SLEEP)
except Exception as e:
logger.warning(f"{type(e), e}")
time.sleep(self.API_RETRY_SLEEP)
raise Exception("Failed to get response from the API")