lm_eval/tasks/super_glue/multirc/t5-prompt.yaml (23 lines of code) (raw):

tag: - super-glue-t5-prompt task: super_glue-multirc-t5-prompt dataset_path: super_glue dataset_name: multirc training_split: train validation_split: validation output_type: generate_until doc_to_text: "multirc question: {{question}} answer: {{answer}} paragraph: {{paragraph}}" doc_to_target: label doc_to_choice: "{% set group_id = idx.question|string %}{{[group_id+'_False', group_id+'_True']}}" generation_kwargs: until: - "</s>" metric_list: - metric: !function t5_utils.f1 aggregation: !function t5_utils.agg_f1 higher_is_better: true - metric: !function t5_utils.em aggregation: !function t5_utils.agg_em higher_is_better: true metadata: version: 0.0