lm_eval/tasks/eus_proficiency/eus_proficiency.yaml (16 lines of code) (raw):
dataset_path: HiTZ/EusProficiency
dataset_name: default
task: eus_proficiency
doc_to_text: "Galdera: {{question}}\nA: {{candidates[0]}}\nB: {{candidates[1]}}\nC: {{candidates[2]}}\nD: {{candidates[3]}}\nErantzuna:"
doc_to_choice: ["A", "B", "C", "D"]
validation_split: null
test_split: test
fewshot_split: test
output_type: multiple_choice
doc_to_target: answer
metric_list:
- metric: acc
aggregation: mean
higher_is_better: true
metadata:
version: 0.0