lmms_eval/tasks/mathvista/mathvista_testmini.yaml (28 lines of code) (raw):

dataset_path: AI4Math/MathVista dataset_kwargs: token: True task: "mathvista_testmini" test_split: testmini output_type: generate_until doc_to_visual: !function utils.mathvista_doc_to_visual doc_to_text: !function utils.mathvista_doc_to_text doc_to_target: "answer" generation_kwargs: until: - "ASSISTANT:" max_new_tokens: 1024 temperature: 0 top_p: 0 num_beams: 1 do_sample: false process_results: !function utils.mathvista_process_results metric_list: - metric: gpt_eval_score aggregation: !function utils.mathvista_aggregate_results higher_is_better: true model_specific_prompt_kwargs: default: shot_type: "format-prompt" # can be "reason-first", "solution", "step-by-step" model_specific_generation_kwargs: llava: image_aspect_ratio: original