lmms_eval/tasks/scienceqa/scienceqa_img.yaml (33 lines of code) (raw):

dataset_path: lmms-lab/ScienceQA dataset_name: ScienceQA-IMG task: "scienceqa_img" dataset_kwargs: token: True test_split: test output_type: generate_until doc_to_visual: !function utils.sqa_doc_to_visual doc_to_text: !function utils.sqa_doc_to_text doc_to_target: !function utils.sqa_doc_to_target generation_kwargs: max_new_tokens: 16 temperature: 0 do_sample: False metric_list: - metric: exact_match aggregation: mean higher_is_better: true ignore_case: true ignore_punctuation: true process_results: !function utils.sqa_process_results metadata: - version: 0.0 model_specific_prompt_kwargs: default: format: default pre_prompt: "" post_prompt: "\nAnswer with the option's letter from the given choices directly." qwen_vl: format: qwen_vl model_specific_generation_kwargs: llava: image_aspect_ratio: original