src/lighteval/tasks/multilingual/tasks.py [2400:2417]:
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        hf_revision="08663706ee7cab30c4b7dc1bb00042a3227ce1ff",
        trust_dataset=True,
        evaluation_splits=["test"],
        few_shots_split="validation",
        metrics=get_metrics_for_formulation(
            formulation,
            [
                loglikelihood_acc_metric(normalization=LogProbTokenNorm()),
                loglikelihood_acc_metric(normalization=LogProbCharNorm()),
            ],
        ),
    )
    for formulation in [
        MCFFormulation(),
        CFFormulation(),
        HybridFormulation(),
    ]
]
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -



src/lighteval/tasks/multilingual/tasks.py [3107:3124]:
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        hf_revision="08663706ee7cab30c4b7dc1bb00042a3227ce1ff",
        trust_dataset=True,
        evaluation_splits=["test"],
        few_shots_split="validation",
        metrics=get_metrics_for_formulation(
            formulation,
            [
                loglikelihood_acc_metric(normalization=LogProbTokenNorm()),
                loglikelihood_acc_metric(normalization=LogProbCharNorm()),
            ],
        ),
    )
    for formulation in [
        MCFFormulation(),
        CFFormulation(),
        HybridFormulation(),
    ]
]
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -



