lm_eval/tasks/super_glue/cb/t5-prompt.yaml (25 lines of code) (raw):
tag:
- super-glue-t5-prompt
task: super_glue-cb-t5-prompt
dataset_path: super_glue
dataset_name: cb
training_split: train
validation_split: validation
output_type: generate_until
doc_to_text: "cb hypothesis: {{hypothesis}} premise: {{premise}}"
doc_to_target: label
doc_to_choice: ['entailment', 'contradiction', 'neutral']
generation_kwargs:
until:
- "</s>"
metric_list:
- metric: exact_match
aggregation: mean
higher_is_better: true
ignore_case: true
ignore_punctuation: true
- metric: !function "t5_utils.mean_3class_f1"
aggregation: !function "t5_utils.agg_mean_3class_f1"
higher_is_better: true
metadata:
version: 0.0