wisent 0.7.379__py3-none-any.whl → 0.7.901__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- wisent/__init__.py +1 -1
- wisent/core/activations/__init__.py +22 -6
- wisent/core/activations/activation_cache.py +393 -0
- wisent/core/activations/activations.py +22 -40
- wisent/core/activations/activations_collector.py +145 -373
- wisent/core/activations/classifier_inference_strategy.py +195 -0
- wisent/core/activations/core/atoms.py +8 -92
- wisent/core/activations/extraction_strategy.py +480 -0
- wisent/core/agent/diagnose/response_diagnostics.py +3 -3
- wisent/core/agent/diagnose.py +3 -3
- wisent/core/autonomous_agent.py +2 -2
- wisent/core/classifiers/classifiers/core/atoms.py +3 -2
- wisent/core/cli/__init__.py +2 -1
- wisent/core/cli/agent/apply_steering.py +25 -31
- wisent/core/cli/agent/evaluate_response.py +18 -20
- wisent/core/cli/agent/train_classifier.py +36 -26
- wisent/core/cli/check_linearity.py +35 -3
- wisent/core/cli/cluster_benchmarks.py +470 -0
- wisent/core/cli/create_steering_vector.py +19 -9
- wisent/core/cli/diagnose_vectors.py +7 -4
- wisent/core/cli/estimate_unified_goodness_time.py +6 -4
- wisent/core/cli/generate_pairs_from_task.py +9 -56
- wisent/core/cli/generate_vector_from_task.py +4 -0
- wisent/core/cli/geometry_search.py +137 -0
- wisent/core/cli/get_activations.py +13 -37
- wisent/core/cli/method_optimizer.py +860 -0
- wisent/core/cli/modify_weights.py +3 -2
- wisent/core/cli/optimize.py +44 -5
- wisent/core/cli/optimize_classification.py +5 -6
- wisent/core/cli/optimize_sample_size.py +9 -23
- wisent/core/cli/optimize_steering.py +433 -159
- wisent/core/cli/optimize_weights.py +67 -7
- wisent/core/cli/preview_pairs.py +203 -0
- wisent/core/cli/steering_method_trainer.py +8 -7
- wisent/core/cli/steering_search_space.py +20 -15
- wisent/core/cli/tasks.py +31 -117
- wisent/core/cli/train_unified_goodness.py +18 -19
- wisent/core/contrastive_pairs/diagnostics/control_vectors.py +1582 -177
- wisent/core/contrastive_pairs/diagnostics/linearity.py +70 -80
- wisent/core/contrastive_pairs/diagnostics/vector_quality.py +6 -5
- wisent/core/contrastive_pairs/huggingface_pairs/hf_extractor_manifest.py +5 -19
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/__init__.py +11 -5
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/agentic_search.py +37 -347
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/aider_polyglot.py +113 -136
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/apps.py +146 -32
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/codeforces.py +2 -12
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/codexglue.py +2 -2
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/coding_benchmarks.py +124 -504
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/faithbench.py +40 -63
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/flames.py +46 -89
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/flores.py +15 -4
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/frames.py +36 -20
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/hallucinations_leaderboard.py +3 -45
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/humaneval.py +98 -57
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/livemathbench.py +42 -4
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/longform_writing.py +2 -112
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/math500.py +39 -4
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/medium_priority_benchmarks.py +475 -525
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/mercury.py +65 -42
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/olympiadbench.py +2 -12
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/planbench.py +78 -219
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/polymath.py +37 -4
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/recode.py +84 -69
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/refusalbench.py +168 -160
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/simpleqa.py +44 -25
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/tau_bench.py +3 -103
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/toolbench.py +3 -97
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/toolemu.py +48 -182
- wisent/core/contrastive_pairs/lm_eval_pairs/group_task_manifests/code_x_glue.py +8 -8
- wisent/core/contrastive_pairs/lm_eval_pairs/group_task_manifests/freebase.py +1 -1
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_extractor_manifest.py +11 -5
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_extractor_registry.py +19 -1
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/aclue.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/acp_bench.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/acp_bench_hard.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/advanced.py +2 -4
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/aexams.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/afrimmlu.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/afrixnli.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/agieval_aqua_rat.py +129 -0
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/arabculture.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/arabic.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/arabic_exams.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/arabic_leaderboard_complete.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/arabic_leaderboard_light.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/arabicmmlu.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/aradice.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/arc.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/arc_challenge.py +1 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/arc_easy.py +1 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/arithmetic.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/asdiv.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/babi.py +36 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/basque_bench.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/bbq.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/belebele.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/benchmarks.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/bertaqa.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/bhs.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/bhtc.py +3 -5
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/blimp.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/blimp_nl.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/boolq.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/c4.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/cabbq.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/careqa.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/catalan_bench.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/catalanqa.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/catcola.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/cb.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/ceval.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/ceval_valid.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/chain.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/chartqa.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/claim.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/click.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/cmmlu.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/cnn.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/cocoteros.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/code_x_glue.py +11 -6
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/coedit.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/commonsense.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/commonsense_qa.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/copa.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/copal_id.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/coqa.py +3 -4
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/csatqa.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/cycle.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/darija_bench.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/darijahellaswag.py +2 -6
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/darijammlu.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/dbpedia.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/discrim_eval.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/doc.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/drop.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/epec.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/eq.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/eq_bench.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/eq_bench_ca.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/eq_bench_es.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/esbbq.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/ethics.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/eus.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/eus_exams.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/eus_proficiency.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/eus_reading.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/eus_trivia.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/evalita_llm.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/financial.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/flan.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/french_bench.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/galician_bench.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/gaokao.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/glianorex.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/global_mmlu.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/global_piqa.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/gpt3.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/groundcocoa.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/gsm8k.py +1 -1
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/haerae.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/headqa.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/hellaswag.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/hendrycks_ethics.py +5 -9
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/hendrycks_math.py +63 -16
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/histoires_morales.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/hrm8k.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/humaneval_infilling.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/icelandic_winogrande.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/inverse.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/inverse_scaling.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/ja.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/japanese_leaderboard.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/japanese_leaderboard_mc.py +1 -1
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/kmmlu.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/kobest.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/kormedmcqa.py +5 -17
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/lambada_cloze.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/lambada_multilingual.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/law.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/leaderboard.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/lingoly.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/llama3.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/lm_syneval.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/logiqa.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/logiqa2.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/longbench.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/longbenchv2.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/mastermind.py +2 -4
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/mbpp.py +47 -6
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/mc-taco.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/med_concepts_qa.py +2 -4
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/meddialog.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/medical.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/medmcqa.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/medqa.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/mela.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/metabench.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/minerva_math.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/mmlu.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/mmlusr.py +3 -4
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/mrpc.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/multiblimp.py +2 -5
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/multirc.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/mutual.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/non.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/noreval.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/noreval_exact.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/noreval_gen_exact.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/noreval_mc.py +4 -8
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/noreval_mc_log_likelihoods.py +4 -8
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/nq_open.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/okapi_arc_multilingual.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/okapi_hellaswag_multilingual.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/okapi_mmlu_multilingual.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/okapi_truthfulqa_multilingual.py +2 -5
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/olaph.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/openbookqa.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/option.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/parafraseja.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/parafrases.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/paws.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/paws_x.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/pawsx.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/persona.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/phrases.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/pile.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/piqa.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/portuguese_bench.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/prompt.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/prost.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/pubmedqa.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/qa4mre.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/qasper.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/qasper_bool.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/qnli.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/qnlieu.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/qqp.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/race.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/random.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/record.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/reversed.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/rte.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/ruler.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/sciq.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/score.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/scrolls.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/scrolls_mc.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/self.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/sglue.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/sglue_rte.py +2 -1
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/siqa.py +4 -7
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/social_iqa.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/spanish_bench.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/storycloze.py +2 -6
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/summarization.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/super.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/super_glue.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/swag.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/swde.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/sycophancy.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/t0.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/teca.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/tinyarc.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/tinybenchmarks.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/tinygsm8k.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/tinyhellaswag.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/tinymmlu.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/tinytruthfulqa.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/tinywinogrande.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/tmmluplus.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/triviaqa.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/truthfulqa.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/truthfulqa_mc1.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/truthfulqa_mc2.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/turblimp_core.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/turkishmmlu.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/turkishmmlu_mc.py +0 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/unscramble.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/vaxx.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/webqs.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/wic.py +3 -4
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/winogrande.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/wmdp.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/wnli.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/wsc.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/wsc273.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/xcopa.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/xlsum.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/xnli.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/xquad.py +2 -4
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/xstorycloze.py +2 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/xwinograd.py +2 -2
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/zhoblimp.py +1 -3
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_pairs_generation.py +173 -6
- wisent/core/data_loaders/loaders/lm_loader.py +12 -1
- wisent/core/evaluators/benchmark_specific/apps_evaluator.py +133 -0
- wisent/core/evaluators/benchmark_specific/coding/metrics/evaluator.py +6 -1
- wisent/core/evaluators/benchmark_specific/conala_evaluator.py +31 -168
- wisent/core/evaluators/custom/examples/humanization_coherent.py +89 -35
- wisent/core/evaluators/oracles/truthfulqa_gen_evaluator.py +2 -20
- wisent/core/evaluators/personalization/coherence.py +46 -0
- wisent/core/geometry_runner.py +995 -0
- wisent/core/geometry_search_space.py +237 -0
- wisent/core/hyperparameter_optimizer.py +14 -14
- wisent/core/lm_eval_harness_ground_truth.py +7 -11
- wisent/core/main.py +6 -0
- wisent/core/models/core/atoms.py +5 -3
- wisent/core/models/wisent_model.py +9 -8
- wisent/core/opti/methods/opti_weights.py +29 -2
- wisent/core/optuna/classifier/activation_generator.py +14 -12
- wisent/core/optuna/classifier/optuna_classifier_optimizer.py +2 -2
- wisent/core/optuna/steering/steering_optimization.py +14 -9
- wisent/core/parser_arguments/check_linearity_parser.py +12 -2
- wisent/core/parser_arguments/cluster_benchmarks_parser.py +31 -0
- wisent/core/parser_arguments/generate_vector_from_synthetic_parser.py +2 -2
- wisent/core/parser_arguments/generate_vector_from_task_parser.py +22 -2
- wisent/core/parser_arguments/geometry_search_parser.py +61 -0
- wisent/core/parser_arguments/main_parser.py +16 -0
- wisent/core/parser_arguments/optimize_steering_parser.py +117 -10
- wisent/core/parser_arguments/optimize_weights_parser.py +6 -0
- wisent/core/parser_arguments/tasks_parser.py +7 -19
- wisent/core/parser_arguments/train_unified_goodness_parser.py +2 -2
- wisent/core/steering.py +5 -3
- wisent/core/steering_methods/core/atoms.py +1 -2
- wisent/core/steering_methods/methods/caa.py +1 -1
- wisent/core/steering_methods/methods/hyperplane.py +75 -0
- wisent/core/steering_methods/methods/prism.py +1 -2
- wisent/core/steering_methods/methods/pulse.py +39 -8
- wisent/core/steering_methods/methods/titan.py +59 -14
- wisent/core/steering_methods/registry.py +52 -12
- wisent/core/steering_optimizer.py +15 -15
- wisent/core/synthetic/generators/nonsense_generator.py +30 -18
- wisent/core/trainers/steering_trainer.py +11 -20
- wisent/core/utils/device.py +27 -27
- wisent/core/utils/layer_combinations.py +70 -0
- wisent/examples/__init__.py +1 -0
- wisent/examples/scripts/__init__.py +1 -0
- wisent/examples/scripts/count_all_benchmarks.py +121 -0
- wisent/examples/scripts/discover_directions.py +469 -0
- wisent/examples/scripts/extract_benchmark_info.py +71 -0
- wisent/examples/scripts/generate_paper_data.py +384 -0
- wisent/examples/scripts/intervention_validation.py +626 -0
- wisent/examples/scripts/results/test_AraDiCE_ArabicMMLU_lev_evaluation.json +324 -0
- wisent/examples/scripts/results/test_AraDiCE_ArabicMMLU_lev_pairs.json +92 -0
- wisent/examples/scripts/results/test_aexams_IslamicStudies_evaluation.json +324 -0
- wisent/examples/scripts/results/test_aexams_IslamicStudies_pairs.json +92 -0
- wisent/examples/scripts/results/test_afrimgsm_pairs.json +92 -0
- wisent/examples/scripts/results/test_afrimmlu_evaluation.json +324 -0
- wisent/examples/scripts/results/test_afrimmlu_pairs.json +92 -0
- wisent/examples/scripts/search_all_short_names.py +31 -0
- wisent/examples/scripts/test_all_benchmarks.py +138 -0
- wisent/examples/scripts/test_all_benchmarks_new.py +28 -0
- wisent/examples/scripts/test_contrastive_pairs_all_supported.py +230 -0
- wisent/examples/scripts/test_nonsense_baseline.py +261 -0
- wisent/examples/scripts/test_one_benchmark.py +324 -0
- wisent/examples/scripts/test_one_coding_benchmark.py +293 -0
- wisent/examples/scripts/threshold_analysis.py +434 -0
- wisent/examples/scripts/visualization_gallery.py +582 -0
- wisent/parameters/lm_eval/broken_in_lm_eval.json +179 -2
- wisent/parameters/lm_eval/category_directions.json +137 -0
- wisent/parameters/lm_eval/repair_plan.json +282 -0
- wisent/parameters/lm_eval/track_progress_not_lm_eval_tasks.json +19 -70
- wisent/parameters/lm_eval/weak_contrastive_pairs.json +38 -0
- wisent/parameters/lm_eval/working_benchmarks.json +206 -0
- wisent/parameters/lm_eval/working_benchmarks_categorized.json +236 -0
- wisent/scripts/run_quality_metrics_sweep.sh +22 -27
- wisent/tests/test_aggregation_geometry.py +236 -0
- wisent/tests/test_detector_accuracy.py +163 -0
- wisent/tests/test_geometry_exhaustive.py +1202 -0
- wisent/tests/visualize_geometry.py +255 -61
- {wisent-0.7.379.dist-info → wisent-0.7.901.dist-info}/METADATA +1 -1
- {wisent-0.7.379.dist-info → wisent-0.7.901.dist-info}/RECORD +376 -974
- wisent/core/activations/prompt_construction_strategy.py +0 -47
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/browsecomp.py +0 -245
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/codexglue_code_to_text.py +0 -15
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/codexglue_code_to_text_go.py +0 -64
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/codexglue_code_to_text_java.py +0 -65
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/codexglue_code_to_text_javascript.py +0 -65
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/codexglue_code_to_text_php.py +0 -65
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/codexglue_code_to_text_python.py +0 -65
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/codexglue_code_to_text_ruby.py +0 -65
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/freebase.py +0 -99
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/instruct_humaneval.py +0 -180
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/instructhumaneval.py +0 -129
- wisent/core/contrastive_pairs/huggingface_pairs/hf_task_extractors/mbpp.py +0 -142
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/agieval.py +0 -155
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/code2text.py +0 -161
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/codexglue.py +0 -107
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/livemathbench.py +0 -155
- wisent/core/contrastive_pairs/lm_eval_pairs/lm_task_extractors/polymath.py +0 -155
- wisent/examples/scripts/results/benchmark_descriptions.json +0 -1244
- wisent/examples/scripts/results/benchmark_evaluation_methods.json +0 -66
- wisent/examples/scripts/results/benchmark_evaluator_mapping.json +0 -2781
- wisent/examples/scripts/results/benchmark_evaluator_mapping_updated.json +0 -30536
- wisent/examples/scripts/results/benchmark_evaluators_clean.json +0 -469
- wisent/examples/scripts/results/benchmark_methods_summary.json +0 -260
- wisent/examples/scripts/results/benchmark_pair_creation_methods.json +0 -66
- wisent/examples/scripts/results/benchmark_pair_totals.json +0 -269
- wisent/examples/scripts/results/benchmark_tags.json +0 -917
- wisent/examples/scripts/results/benchmark_test_summary_nov4.json +0 -71
- wisent/examples/scripts/results/coding_benchmarks_test_code_status.json +0 -150
- wisent/examples/scripts/results/failing_benchmarks.json +0 -946
- wisent/examples/scripts/results/failing_benchmarks_list.json +0 -41
- wisent/examples/scripts/results/failing_benchmarks_test_results.json +0 -945
- wisent/examples/scripts/results/missing_benchmark_tags.json +0 -341
- wisent/examples/scripts/results/test_20_newsgroups_evaluation.json +0 -30
- wisent/examples/scripts/results/test_20_newsgroups_pairs.json +0 -8
- wisent/examples/scripts/results/test_AraDICE_evaluation.json +0 -51
- wisent/examples/scripts/results/test_AraDICE_pairs.json +0 -14
- wisent/examples/scripts/results/test_AraDiCE_boolq_egy/test_AraDiCE_boolq_egy_evaluation.json +0 -30
- wisent/examples/scripts/results/test_AraDiCE_boolq_egy/test_AraDiCE_boolq_egy_pairs.json +0 -8
- wisent/examples/scripts/results/test_ArabCulture_evaluation.json +0 -51
- wisent/examples/scripts/results/test_ArabCulture_pairs.json +0 -14
- wisent/examples/scripts/results/test_Tag_evaluation.json +0 -30
- wisent/examples/scripts/results/test_Tag_pairs.json +0 -8
- wisent/examples/scripts/results/test_aclue_evaluation.json +0 -51
- wisent/examples/scripts/results/test_aclue_pairs.json +0 -14
- wisent/examples/scripts/results/test_acp_bench_evaluation.json +0 -51
- wisent/examples/scripts/results/test_acp_bench_hard_evaluation.json +0 -51
- wisent/examples/scripts/results/test_acp_bench_hard_pairs.json +0 -14
- wisent/examples/scripts/results/test_acp_bench_pairs.json +0 -14
- wisent/examples/scripts/results/test_advanced_ai_risk_evaluation.json +0 -51
- wisent/examples/scripts/results/test_advanced_ai_risk_pairs.json +0 -14
- wisent/examples/scripts/results/test_aexams_evaluation.json +0 -51
- wisent/examples/scripts/results/test_aexams_pairs.json +0 -14
- wisent/examples/scripts/results/test_afrimgsm_direct_amh_evaluation.json +0 -30
- wisent/examples/scripts/results/test_afrimgsm_direct_amh_pairs.json +0 -8
- wisent/examples/scripts/results/test_afrimmlu_direct_amh_evaluation.json +0 -30
- wisent/examples/scripts/results/test_afrimmlu_direct_amh_pairs.json +0 -8
- wisent/examples/scripts/results/test_afrixnli_en_direct_amh_evaluation.json +0 -30
- wisent/examples/scripts/results/test_afrixnli_en_direct_amh_pairs.json +0 -8
- wisent/examples/scripts/results/test_ag_news_evaluation.json +0 -30
- wisent/examples/scripts/results/test_ag_news_pairs.json +0 -8
- wisent/examples/scripts/results/test_agieval_evaluation.json +0 -51
- wisent/examples/scripts/results/test_agieval_pairs.json +0 -14
- wisent/examples/scripts/results/test_aime2024_evaluation.json +0 -30
- wisent/examples/scripts/results/test_aime2024_pairs.json +0 -8
- wisent/examples/scripts/results/test_aime2025_evaluation.json +0 -30
- wisent/examples/scripts/results/test_aime2025_pairs.json +0 -8
- wisent/examples/scripts/results/test_aime_evaluation.json +0 -30
- wisent/examples/scripts/results/test_aime_pairs.json +0 -8
- wisent/examples/scripts/results/test_anagrams1_evaluation.json +0 -30
- wisent/examples/scripts/results/test_anagrams1_pairs.json +0 -8
- wisent/examples/scripts/results/test_anagrams2_evaluation.json +0 -30
- wisent/examples/scripts/results/test_anagrams2_pairs.json +0 -8
- wisent/examples/scripts/results/test_anli_evaluation.json +0 -30
- wisent/examples/scripts/results/test_anli_pairs.json +0 -8
- wisent/examples/scripts/results/test_apps_evaluation.json +0 -30
- wisent/examples/scripts/results/test_apps_pairs.json +0 -8
- wisent/examples/scripts/results/test_arabic_exams_evaluation.json +0 -30
- wisent/examples/scripts/results/test_arabic_exams_pairs.json +0 -8
- wisent/examples/scripts/results/test_arabic_leaderboard_complete_evaluation.json +0 -51
- wisent/examples/scripts/results/test_arabic_leaderboard_complete_pairs.json +0 -14
- wisent/examples/scripts/results/test_arabic_leaderboard_light_evaluation.json +0 -51
- wisent/examples/scripts/results/test_arabic_leaderboard_light_pairs.json +0 -14
- wisent/examples/scripts/results/test_arabicmmlu_evaluation.json +0 -51
- wisent/examples/scripts/results/test_arabicmmlu_pairs.json +0 -14
- wisent/examples/scripts/results/test_aradice/test_aradice_evaluation.json +0 -51
- wisent/examples/scripts/results/test_aradice/test_aradice_pairs.json +0 -14
- wisent/examples/scripts/results/test_aradice3/test_aradice_evaluation.json +0 -51
- wisent/examples/scripts/results/test_aradice3/test_aradice_pairs.json +0 -14
- wisent/examples/scripts/results/test_arc_ar_evaluation.json +0 -30
- wisent/examples/scripts/results/test_arc_ar_pairs.json +0 -8
- wisent/examples/scripts/results/test_arc_challenge_evaluation.json +0 -30
- wisent/examples/scripts/results/test_arc_challenge_pairs.json +0 -8
- wisent/examples/scripts/results/test_arc_easy_evaluation.json +0 -30
- wisent/examples/scripts/results/test_arc_easy_pairs.json +0 -8
- wisent/examples/scripts/results/test_argument_topic_evaluation.json +0 -30
- wisent/examples/scripts/results/test_argument_topic_pairs.json +0 -8
- wisent/examples/scripts/results/test_arithmetic_evaluation.json +0 -51
- wisent/examples/scripts/results/test_arithmetic_pairs.json +0 -14
- wisent/examples/scripts/results/test_asdiv_evaluation.json +0 -30
- wisent/examples/scripts/results/test_asdiv_pairs.json +0 -8
- wisent/examples/scripts/results/test_assin_entailment_evaluation.json +0 -30
- wisent/examples/scripts/results/test_assin_entailment_pairs.json +0 -8
- wisent/examples/scripts/results/test_atis_evaluation.json +0 -30
- wisent/examples/scripts/results/test_atis_pairs.json +0 -8
- wisent/examples/scripts/results/test_babi_evaluation.json +0 -30
- wisent/examples/scripts/results/test_babi_pairs.json +0 -8
- wisent/examples/scripts/results/test_babilong_evaluation.json +0 -30
- wisent/examples/scripts/results/test_babilong_pairs.json +0 -8
- wisent/examples/scripts/results/test_bangla_mmlu_evaluation.json +0 -30
- wisent/examples/scripts/results/test_bangla_mmlu_pairs.json +0 -8
- wisent/examples/scripts/results/test_banking77_evaluation.json +0 -30
- wisent/examples/scripts/results/test_banking77_pairs.json +0 -8
- wisent/examples/scripts/results/test_basque/test_basque-glue_pairs.json +0 -14
- wisent/examples/scripts/results/test_basque-glue_evaluation.json +0 -51
- wisent/examples/scripts/results/test_basque-glue_pairs.json +0 -14
- wisent/examples/scripts/results/test_basque2/test_basque-glue_evaluation.json +0 -51
- wisent/examples/scripts/results/test_basque2/test_basque-glue_pairs.json +0 -14
- wisent/examples/scripts/results/test_basque_bench_evaluation.json +0 -51
- wisent/examples/scripts/results/test_basque_bench_pairs.json +0 -14
- wisent/examples/scripts/results/test_basque_glue/test_basque-glue_evaluation.json +0 -51
- wisent/examples/scripts/results/test_basque_glue/test_basque-glue_pairs.json +0 -14
- wisent/examples/scripts/results/test_basqueglue_evaluation.json +0 -51
- wisent/examples/scripts/results/test_basqueglue_pairs.json +0 -14
- wisent/examples/scripts/results/test_bbh_evaluation.json +0 -51
- wisent/examples/scripts/results/test_bbh_pairs.json +0 -14
- wisent/examples/scripts/results/test_bbq_evaluation.json +0 -30
- wisent/examples/scripts/results/test_bbq_pairs.json +0 -8
- wisent/examples/scripts/results/test_bec2016eu_evaluation.json +0 -51
- wisent/examples/scripts/results/test_bec2016eu_pairs.json +0 -14
- wisent/examples/scripts/results/test_belebele_evaluation.json +0 -51
- wisent/examples/scripts/results/test_belebele_pairs.json +0 -14
- wisent/examples/scripts/results/test_benchmarks_evaluation.json +0 -51
- wisent/examples/scripts/results/test_benchmarks_pairs.json +0 -14
- wisent/examples/scripts/results/test_bertaqa_evaluation.json +0 -51
- wisent/examples/scripts/results/test_bertaqa_pairs.json +0 -14
- wisent/examples/scripts/results/test_bhtc_v2_evaluation.json +0 -30
- wisent/examples/scripts/results/test_bhtc_v2_pairs.json +0 -8
- wisent/examples/scripts/results/test_bigbench_evaluation.json +0 -51
- wisent/examples/scripts/results/test_bigbench_pairs.json +0 -14
- wisent/examples/scripts/results/test_blimp_evaluation.json +0 -51
- wisent/examples/scripts/results/test_blimp_pairs.json +0 -14
- wisent/examples/scripts/results/test_boolq/test_boolq_evaluation.json +0 -30
- wisent/examples/scripts/results/test_boolq/test_boolq_pairs.json +0 -8
- wisent/examples/scripts/results/test_boolq-seq2seq_evaluation.json +0 -30
- wisent/examples/scripts/results/test_boolq-seq2seq_pairs.json +0 -8
- wisent/examples/scripts/results/test_boolq_evaluation.json +0 -30
- wisent/examples/scripts/results/test_boolq_pairs.json +0 -8
- wisent/examples/scripts/results/test_c4_evaluation.json +0 -30
- wisent/examples/scripts/results/test_c4_pairs.json +0 -8
- wisent/examples/scripts/results/test_cabreu_evaluation.json +0 -30
- wisent/examples/scripts/results/test_cabreu_pairs.json +0 -8
- wisent/examples/scripts/results/test_careqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_careqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_catalan_bench_evaluation.json +0 -51
- wisent/examples/scripts/results/test_catalan_bench_pairs.json +0 -14
- wisent/examples/scripts/results/test_catalanqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_catalanqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_catcola_evaluation.json +0 -30
- wisent/examples/scripts/results/test_catcola_pairs.json +0 -8
- wisent/examples/scripts/results/test_cb_evaluation.json +0 -30
- wisent/examples/scripts/results/test_cb_pairs.json +0 -8
- wisent/examples/scripts/results/test_ceval/test_ceval_evaluation.json +0 -51
- wisent/examples/scripts/results/test_ceval/test_ceval_pairs.json +0 -14
- wisent/examples/scripts/results/test_ceval_accountant/test_ceval-valid_accountant_evaluation.json +0 -30
- wisent/examples/scripts/results/test_ceval_accountant/test_ceval-valid_accountant_pairs.json +0 -8
- wisent/examples/scripts/results/test_ceval_evaluation.json +0 -51
- wisent/examples/scripts/results/test_ceval_pairs.json +0 -14
- wisent/examples/scripts/results/test_ceval_valid/test_ceval_valid_evaluation.json +0 -51
- wisent/examples/scripts/results/test_ceval_valid/test_ceval_valid_pairs.json +0 -14
- wisent/examples/scripts/results/test_chain_of_thought_evaluation.json +0 -51
- wisent/examples/scripts/results/test_chain_of_thought_pairs.json +0 -14
- wisent/examples/scripts/results/test_chartqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_chartqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_claim_stance_topic_evaluation.json +0 -30
- wisent/examples/scripts/results/test_claim_stance_topic_pairs.json +0 -8
- wisent/examples/scripts/results/test_cmmlu_evaluation.json +0 -51
- wisent/examples/scripts/results/test_cmmlu_pairs.json +0 -14
- wisent/examples/scripts/results/test_cnn_dailymail_evaluation.json +0 -30
- wisent/examples/scripts/results/test_cnn_dailymail_pairs.json +0 -8
- wisent/examples/scripts/results/test_cocoteros_es_evaluation.json +0 -30
- wisent/examples/scripts/results/test_cocoteros_es_pairs.json +0 -8
- wisent/examples/scripts/results/test_codexglue_code_to_text_go_evaluation.json +0 -30
- wisent/examples/scripts/results/test_codexglue_code_to_text_go_pairs.json +0 -8
- wisent/examples/scripts/results/test_codexglue_code_to_text_java_evaluation.json +0 -30
- wisent/examples/scripts/results/test_codexglue_code_to_text_java_pairs.json +0 -8
- wisent/examples/scripts/results/test_codexglue_code_to_text_javascript_evaluation.json +0 -30
- wisent/examples/scripts/results/test_codexglue_code_to_text_javascript_pairs.json +0 -8
- wisent/examples/scripts/results/test_codexglue_code_to_text_php_evaluation.json +0 -30
- wisent/examples/scripts/results/test_codexglue_code_to_text_php_pairs.json +0 -8
- wisent/examples/scripts/results/test_codexglue_code_to_text_python_evaluation.json +0 -30
- wisent/examples/scripts/results/test_codexglue_code_to_text_python_pairs.json +0 -8
- wisent/examples/scripts/results/test_codexglue_code_to_text_ruby_evaluation.json +0 -30
- wisent/examples/scripts/results/test_codexglue_code_to_text_ruby_pairs.json +0 -8
- wisent/examples/scripts/results/test_coedit_gec_evaluation.json +0 -30
- wisent/examples/scripts/results/test_coedit_gec_pairs.json +0 -8
- wisent/examples/scripts/results/test_cola_evaluation.json +0 -30
- wisent/examples/scripts/results/test_cola_pairs.json +0 -8
- wisent/examples/scripts/results/test_commonsense_qa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_commonsense_qa_pairs.json +0 -8
- wisent/examples/scripts/results/test_conala_evaluation.json +0 -30
- wisent/examples/scripts/results/test_conala_pairs.json +0 -8
- wisent/examples/scripts/results/test_concode_evaluation.json +0 -30
- wisent/examples/scripts/results/test_concode_pairs.json +0 -8
- wisent/examples/scripts/results/test_copa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_copa_pairs.json +0 -8
- wisent/examples/scripts/results/test_copal_id_evaluation.json +0 -30
- wisent/examples/scripts/results/test_copal_id_pairs.json +0 -8
- wisent/examples/scripts/results/test_coqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_coqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_coqcat_evaluation.json +0 -30
- wisent/examples/scripts/results/test_coqcat_pairs.json +0 -8
- wisent/examples/scripts/results/test_crows_pairs_evaluation.json +0 -51
- wisent/examples/scripts/results/test_crows_pairs_pairs.json +0 -14
- wisent/examples/scripts/results/test_csatqa_evaluation.json +0 -51
- wisent/examples/scripts/results/test_csatqa_pairs.json +0 -14
- wisent/examples/scripts/results/test_cycle_letters_evaluation.json +0 -30
- wisent/examples/scripts/results/test_cycle_letters_pairs.json +0 -8
- wisent/examples/scripts/results/test_darija_bench/test_darija_bench_evaluation.json +0 -51
- wisent/examples/scripts/results/test_darija_bench/test_darija_bench_pairs.json +0 -14
- wisent/examples/scripts/results/test_darija_bench_evaluation.json +0 -51
- wisent/examples/scripts/results/test_darija_bench_pairs.json +0 -14
- wisent/examples/scripts/results/test_darijahellaswag_evaluation.json +0 -30
- wisent/examples/scripts/results/test_darijahellaswag_pairs.json +0 -8
- wisent/examples/scripts/results/test_darijammlu_evaluation.json +0 -51
- wisent/examples/scripts/results/test_darijammlu_pairs.json +0 -14
- wisent/examples/scripts/results/test_dbpedia_14_evaluation.json +0 -30
- wisent/examples/scripts/results/test_dbpedia_14_pairs.json +0 -8
- wisent/examples/scripts/results/test_drop_evaluation.json +0 -30
- wisent/examples/scripts/results/test_drop_pairs.json +0 -8
- wisent/examples/scripts/results/test_ds1000_evaluation.json +0 -30
- wisent/examples/scripts/results/test_ds1000_pairs.json +0 -8
- wisent/examples/scripts/results/test_egyhellaswag_evaluation.json +0 -30
- wisent/examples/scripts/results/test_egyhellaswag_pairs.json +0 -8
- wisent/examples/scripts/results/test_egymmlu_evaluation.json +0 -51
- wisent/examples/scripts/results/test_egymmlu_pairs.json +0 -14
- wisent/examples/scripts/results/test_epec_koref_bin_evaluation.json +0 -30
- wisent/examples/scripts/results/test_epec_koref_bin_pairs.json +0 -8
- wisent/examples/scripts/results/test_eq_bench_evaluation.json +0 -30
- wisent/examples/scripts/results/test_eq_bench_pairs.json +0 -8
- wisent/examples/scripts/results/test_escola_evaluation.json +0 -30
- wisent/examples/scripts/results/test_escola_pairs.json +0 -8
- wisent/examples/scripts/results/test_ethics_cm_evaluation.json +0 -30
- wisent/examples/scripts/results/test_ethics_cm_pairs.json +0 -8
- wisent/examples/scripts/results/test_ethos_binary_evaluation.json +0 -30
- wisent/examples/scripts/results/test_ethos_binary_pairs.json +0 -8
- wisent/examples/scripts/results/test_eus_exams/test_eus_exams_evaluation.json +0 -51
- wisent/examples/scripts/results/test_eus_exams/test_eus_exams_pairs.json +0 -14
- wisent/examples/scripts/results/test_eus_exams_es_evaluation.json +0 -51
- wisent/examples/scripts/results/test_eus_exams_es_pairs.json +0 -14
- wisent/examples/scripts/results/test_eus_exams_evaluation.json +0 -51
- wisent/examples/scripts/results/test_eus_exams_pairs.json +0 -14
- wisent/examples/scripts/results/test_eus_proficiency_evaluation.json +0 -30
- wisent/examples/scripts/results/test_eus_proficiency_pairs.json +0 -8
- wisent/examples/scripts/results/test_eus_reading_evaluation.json +0 -30
- wisent/examples/scripts/results/test_eus_reading_pairs.json +0 -8
- wisent/examples/scripts/results/test_eus_trivia_evaluation.json +0 -30
- wisent/examples/scripts/results/test_eus_trivia_pairs.json +0 -8
- wisent/examples/scripts/results/test_evalita-mp_evaluation.json +0 -51
- wisent/examples/scripts/results/test_evalita-mp_pairs.json +0 -14
- wisent/examples/scripts/results/test_evalita-sp_sum_task_fp-small_p1_evaluation.json +0 -30
- wisent/examples/scripts/results/test_evalita-sp_sum_task_fp-small_p1_pairs.json +0 -8
- wisent/examples/scripts/results/test_evalita_LLM_evaluation.json +0 -51
- wisent/examples/scripts/results/test_evalita_LLM_pairs.json +0 -14
- wisent/examples/scripts/results/test_evalita_llm/test_evalita_llm_evaluation.json +0 -51
- wisent/examples/scripts/results/test_evalita_llm/test_evalita_llm_pairs.json +0 -14
- wisent/examples/scripts/results/test_evalita_mp/test_evalita-mp_te_prompt-1_evaluation.json +0 -30
- wisent/examples/scripts/results/test_evalita_mp/test_evalita-mp_te_prompt-1_pairs.json +0 -8
- wisent/examples/scripts/results/test_evalita_mp2/test_evalita_mp_evaluation.json +0 -51
- wisent/examples/scripts/results/test_evalita_mp2/test_evalita_mp_pairs.json +0 -14
- wisent/examples/scripts/results/test_evalita_sp2/test_evalita-sp_sum_task_fp-small_p1_evaluation.json +0 -30
- wisent/examples/scripts/results/test_evalita_sp2/test_evalita-sp_sum_task_fp-small_p1_pairs.json +0 -8
- wisent/examples/scripts/results/test_fda_evaluation.json +0 -30
- wisent/examples/scripts/results/test_fda_pairs.json +0 -8
- wisent/examples/scripts/results/test_financial_tweets_evaluation.json +0 -30
- wisent/examples/scripts/results/test_financial_tweets_pairs.json +0 -8
- wisent/examples/scripts/results/test_fld/test_fld_evaluation.json +0 -30
- wisent/examples/scripts/results/test_fld/test_fld_pairs.json +0 -8
- wisent/examples/scripts/results/test_fld_evaluation.json +0 -30
- wisent/examples/scripts/results/test_fld_fixed/test_fld_evaluation.json +0 -30
- wisent/examples/scripts/results/test_fld_fixed/test_fld_pairs.json +0 -8
- wisent/examples/scripts/results/test_fld_pairs.json +0 -8
- wisent/examples/scripts/results/test_flores_evaluation.json +0 -51
- wisent/examples/scripts/results/test_flores_pairs.json +0 -14
- wisent/examples/scripts/results/test_freebase_evaluation.json +0 -30
- wisent/examples/scripts/results/test_freebase_pairs.json +0 -8
- wisent/examples/scripts/results/test_french_bench_evaluation.json +0 -51
- wisent/examples/scripts/results/test_french_bench_pairs.json +0 -14
- wisent/examples/scripts/results/test_galcola_evaluation.json +0 -30
- wisent/examples/scripts/results/test_galcola_pairs.json +0 -8
- wisent/examples/scripts/results/test_galician_bench_evaluation.json +0 -51
- wisent/examples/scripts/results/test_galician_bench_pairs.json +0 -14
- wisent/examples/scripts/results/test_glianorex_evaluation.json +0 -30
- wisent/examples/scripts/results/test_glianorex_pairs.json +0 -8
- wisent/examples/scripts/results/test_global_mmlu_evaluation.json +0 -51
- wisent/examples/scripts/results/test_global_mmlu_pairs.json +0 -14
- wisent/examples/scripts/results/test_glue_evaluation.json +0 -51
- wisent/examples/scripts/results/test_glue_pairs.json +0 -14
- wisent/examples/scripts/results/test_gpqa_evaluation.json +0 -51
- wisent/examples/scripts/results/test_gpqa_pairs.json +0 -14
- wisent/examples/scripts/results/test_gpt3_translation_benchmarks_evaluation.json +0 -51
- wisent/examples/scripts/results/test_gpt3_translation_benchmarks_pairs.json +0 -14
- wisent/examples/scripts/results/test_groundcocoa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_groundcocoa_pairs.json +0 -8
- wisent/examples/scripts/results/test_gsm8k_evaluation.json +0 -30
- wisent/examples/scripts/results/test_gsm8k_pairs.json +0 -8
- wisent/examples/scripts/results/test_haerae_evaluation.json +0 -51
- wisent/examples/scripts/results/test_haerae_pairs.json +0 -14
- wisent/examples/scripts/results/test_headqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_headqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_hellaswag_evaluation.json +0 -30
- wisent/examples/scripts/results/test_hellaswag_pairs.json +0 -8
- wisent/examples/scripts/results/test_hendrycks_ethics_evaluation.json +0 -51
- wisent/examples/scripts/results/test_hendrycks_ethics_pairs.json +0 -14
- wisent/examples/scripts/results/test_hendrycks_math_evaluation.json +0 -51
- wisent/examples/scripts/results/test_hendrycks_math_pairs.json +0 -14
- wisent/examples/scripts/results/test_histoires_morales_evaluation.json +0 -30
- wisent/examples/scripts/results/test_histoires_morales_pairs.json +0 -8
- wisent/examples/scripts/results/test_hmmt_evaluation.json +0 -30
- wisent/examples/scripts/results/test_hmmt_feb_2025_evaluation.json +0 -30
- wisent/examples/scripts/results/test_hmmt_feb_2025_pairs.json +0 -8
- wisent/examples/scripts/results/test_hmmt_pairs.json +0 -8
- wisent/examples/scripts/results/test_hrm8k_evaluation.json +0 -51
- wisent/examples/scripts/results/test_hrm8k_pairs.json +0 -14
- wisent/examples/scripts/results/test_humaneval_evaluation.json +0 -30
- wisent/examples/scripts/results/test_humaneval_pairs.json +0 -8
- wisent/examples/scripts/results/test_humaneval_plus_evaluation.json +0 -30
- wisent/examples/scripts/results/test_humaneval_plus_pairs.json +0 -8
- wisent/examples/scripts/results/test_ifeval_evaluation.json +0 -30
- wisent/examples/scripts/results/test_ifeval_pairs.json +0 -8
- wisent/examples/scripts/results/test_instruct_humaneval/test_instruct_humaneval_evaluation.json +0 -30
- wisent/examples/scripts/results/test_instruct_humaneval/test_instruct_humaneval_pairs.json +0 -8
- wisent/examples/scripts/results/test_instruct_humaneval_evaluation.json +0 -30
- wisent/examples/scripts/results/test_instruct_humaneval_pairs.json +0 -8
- wisent/examples/scripts/results/test_inverse_scaling_evaluation.json +0 -51
- wisent/examples/scripts/results/test_inverse_scaling_hindsight_neglect_10shot_evaluation.json +0 -30
- wisent/examples/scripts/results/test_inverse_scaling_hindsight_neglect_10shot_pairs.json +0 -8
- wisent/examples/scripts/results/test_inverse_scaling_mc/test_inverse_scaling_mc_evaluation.json +0 -51
- wisent/examples/scripts/results/test_inverse_scaling_mc/test_inverse_scaling_mc_pairs.json +0 -14
- wisent/examples/scripts/results/test_inverse_scaling_pairs.json +0 -14
- wisent/examples/scripts/results/test_iwslt2017-ar-en_evaluation.json +0 -30
- wisent/examples/scripts/results/test_iwslt2017-ar-en_pairs.json +0 -8
- wisent/examples/scripts/results/test_iwslt2017-en-ar_evaluation.json +0 -30
- wisent/examples/scripts/results/test_iwslt2017-en-ar_pairs.json +0 -8
- wisent/examples/scripts/results/test_iwslt2017_ar_en/test_iwslt2017-ar-en_evaluation.json +0 -30
- wisent/examples/scripts/results/test_iwslt2017_ar_en/test_iwslt2017-ar-en_pairs.json +0 -8
- wisent/examples/scripts/results/test_iwslt2017_en_ar/test_iwslt2017-en-ar_evaluation.json +0 -30
- wisent/examples/scripts/results/test_iwslt2017_en_ar/test_iwslt2017-en-ar_pairs.json +0 -8
- wisent/examples/scripts/results/test_iwslt2017_group/test_iwslt2017_evaluation.json +0 -30
- wisent/examples/scripts/results/test_iwslt2017_group/test_iwslt2017_pairs.json +0 -8
- wisent/examples/scripts/results/test_japanese_leaderboard_evaluation.json +0 -51
- wisent/examples/scripts/results/test_japanese_leaderboard_pairs.json +0 -14
- wisent/examples/scripts/results/test_jsonschema_bench/test_jsonschema_bench_evaluation.json +0 -30
- wisent/examples/scripts/results/test_jsonschema_bench/test_jsonschema_bench_pairs.json +0 -8
- wisent/examples/scripts/results/test_jsonschema_bench_evaluation.json +0 -30
- wisent/examples/scripts/results/test_jsonschema_bench_final/test_jsonschema_bench_evaluation.json +0 -30
- wisent/examples/scripts/results/test_jsonschema_bench_final/test_jsonschema_bench_pairs.json +0 -8
- wisent/examples/scripts/results/test_jsonschema_bench_pairs.json +0 -8
- wisent/examples/scripts/results/test_kbl_evaluation.json +0 -51
- wisent/examples/scripts/results/test_kbl_fixed/test_kbl_evaluation.json +0 -51
- wisent/examples/scripts/results/test_kbl_fixed/test_kbl_pairs.json +0 -14
- wisent/examples/scripts/results/test_kbl_pairs.json +0 -14
- wisent/examples/scripts/results/test_kmmlu_evaluation.json +0 -51
- wisent/examples/scripts/results/test_kmmlu_pairs.json +0 -14
- wisent/examples/scripts/results/test_kobest_evaluation.json +0 -51
- wisent/examples/scripts/results/test_kobest_pairs.json +0 -14
- wisent/examples/scripts/results/test_kormedmcqa/test_kormedmcqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_kormedmcqa/test_kormedmcqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_kormedmcqa_dentist/test_kormedmcqa_dentist_evaluation.json +0 -30
- wisent/examples/scripts/results/test_kormedmcqa_dentist/test_kormedmcqa_dentist_pairs.json +0 -8
- wisent/examples/scripts/results/test_kormedmcqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_kormedmcqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_lambada_cloze_evaluation.json +0 -30
- wisent/examples/scripts/results/test_lambada_cloze_pairs.json +0 -8
- wisent/examples/scripts/results/test_lambada_evaluation.json +0 -30
- wisent/examples/scripts/results/test_lambada_final/test_lambada_openai_mt_stablelm_en_evaluation.json +0 -30
- wisent/examples/scripts/results/test_lambada_final/test_lambada_openai_mt_stablelm_en_pairs.json +0 -8
- wisent/examples/scripts/results/test_lambada_multilingual/test_lambada_multilingual_evaluation.json +0 -51
- wisent/examples/scripts/results/test_lambada_multilingual/test_lambada_multilingual_pairs.json +0 -14
- wisent/examples/scripts/results/test_lambada_multilingual_evaluation.json +0 -51
- wisent/examples/scripts/results/test_lambada_multilingual_pairs.json +0 -14
- wisent/examples/scripts/results/test_lambada_multilingual_stablelm_evaluation.json +0 -51
- wisent/examples/scripts/results/test_lambada_multilingual_stablelm_pairs.json +0 -14
- wisent/examples/scripts/results/test_lambada_openai_evaluation.json +0 -30
- wisent/examples/scripts/results/test_lambada_openai_pairs.json +0 -8
- wisent/examples/scripts/results/test_lambada_pairs.json +0 -8
- wisent/examples/scripts/results/test_lambada_stablelm_en_fixed/test_lambada_openai_mt_stablelm_en_evaluation.json +0 -30
- wisent/examples/scripts/results/test_lambada_stablelm_en_fixed/test_lambada_openai_mt_stablelm_en_pairs.json +0 -8
- wisent/examples/scripts/results/test_lambada_stablelm_fixed/test_lambada_openai_mt_stablelm_en_evaluation.json +0 -30
- wisent/examples/scripts/results/test_lambada_stablelm_fixed/test_lambada_openai_mt_stablelm_en_pairs.json +0 -8
- wisent/examples/scripts/results/test_lambada_standard_evaluation.json +0 -30
- wisent/examples/scripts/results/test_lambada_standard_pairs.json +0 -8
- wisent/examples/scripts/results/test_leaderboard_evaluation.json +0 -51
- wisent/examples/scripts/results/test_leaderboard_pairs.json +0 -14
- wisent/examples/scripts/results/test_libra/test_libra_evaluation.json +0 -51
- wisent/examples/scripts/results/test_libra/test_libra_pairs.json +0 -14
- wisent/examples/scripts/results/test_libra_evaluation.json +0 -51
- wisent/examples/scripts/results/test_libra_pairs.json +0 -14
- wisent/examples/scripts/results/test_lingoly_evaluation.json +0 -30
- wisent/examples/scripts/results/test_lingoly_pairs.json +0 -8
- wisent/examples/scripts/results/test_livecodebench_evaluation.json +0 -30
- wisent/examples/scripts/results/test_livecodebench_pairs.json +0 -8
- wisent/examples/scripts/results/test_livemathbench_cnmo_en_evaluation.json +0 -30
- wisent/examples/scripts/results/test_livemathbench_cnmo_en_pairs.json +0 -8
- wisent/examples/scripts/results/test_livemathbench_cnmo_zh_evaluation.json +0 -30
- wisent/examples/scripts/results/test_livemathbench_cnmo_zh_pairs.json +0 -8
- wisent/examples/scripts/results/test_llama_evaluation.json +0 -30
- wisent/examples/scripts/results/test_llama_pairs.json +0 -8
- wisent/examples/scripts/results/test_logiqa2_evaluation.json +0 -30
- wisent/examples/scripts/results/test_logiqa2_pairs.json +0 -8
- wisent/examples/scripts/results/test_logiqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_logiqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_m_mmlu_evaluation.json +0 -51
- wisent/examples/scripts/results/test_m_mmlu_pairs.json +0 -14
- wisent/examples/scripts/results/test_mastermind/test_mastermind_evaluation.json +0 -51
- wisent/examples/scripts/results/test_mastermind/test_mastermind_pairs.json +0 -14
- wisent/examples/scripts/results/test_mastermind_24_easy/test_mastermind_24_easy_evaluation.json +0 -30
- wisent/examples/scripts/results/test_mastermind_24_easy/test_mastermind_24_easy_pairs.json +0 -8
- wisent/examples/scripts/results/test_mastermind_evaluation.json +0 -51
- wisent/examples/scripts/results/test_mastermind_pairs.json +0 -14
- wisent/examples/scripts/results/test_math500_evaluation.json +0 -30
- wisent/examples/scripts/results/test_math500_pairs.json +0 -8
- wisent/examples/scripts/results/test_math_evaluation.json +0 -30
- wisent/examples/scripts/results/test_math_pairs.json +0 -8
- wisent/examples/scripts/results/test_mathqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_mathqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_mbpp_evaluation.json +0 -30
- wisent/examples/scripts/results/test_mbpp_pairs.json +0 -8
- wisent/examples/scripts/results/test_mbpp_plus_evaluation.json +0 -30
- wisent/examples/scripts/results/test_mbpp_plus_pairs.json +0 -8
- wisent/examples/scripts/results/test_mc_taco_evaluation.json +0 -30
- wisent/examples/scripts/results/test_mc_taco_pairs.json +0 -8
- wisent/examples/scripts/results/test_med_concepts_qa/test_med_concepts_qa_evaluation.json +0 -51
- wisent/examples/scripts/results/test_med_concepts_qa/test_med_concepts_qa_pairs.json +0 -14
- wisent/examples/scripts/results/test_med_concepts_qa_atc_easy/test_med_concepts_qa_atc_easy_evaluation.json +0 -30
- wisent/examples/scripts/results/test_med_concepts_qa_atc_easy/test_med_concepts_qa_atc_easy_pairs.json +0 -8
- wisent/examples/scripts/results/test_med_concepts_qa_evaluation.json +0 -51
- wisent/examples/scripts/results/test_med_concepts_qa_pairs.json +0 -14
- wisent/examples/scripts/results/test_meddialog_evaluation.json +0 -30
- wisent/examples/scripts/results/test_meddialog_pairs.json +0 -8
- wisent/examples/scripts/results/test_meddialog_raw_perplexity/test_meddialog_raw_perplexity_evaluation.json +0 -30
- wisent/examples/scripts/results/test_meddialog_raw_perplexity/test_meddialog_raw_perplexity_pairs.json +0 -8
- wisent/examples/scripts/results/test_mediqa_qa2019_evaluation.json +0 -30
- wisent/examples/scripts/results/test_mediqa_qa2019_pairs.json +0 -8
- wisent/examples/scripts/results/test_medmcqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_medmcqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_medqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_medqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_medtext_evaluation.json +0 -30
- wisent/examples/scripts/results/test_medtext_pairs.json +0 -8
- wisent/examples/scripts/results/test_mela_evaluation.json +0 -51
- wisent/examples/scripts/results/test_mela_pairs.json +0 -14
- wisent/examples/scripts/results/test_meqsum_evaluation.json +0 -30
- wisent/examples/scripts/results/test_meqsum_pairs.json +0 -8
- wisent/examples/scripts/results/test_mercury_evaluation.json +0 -30
- wisent/examples/scripts/results/test_mercury_pairs.json +0 -8
- wisent/examples/scripts/results/test_metabench_evaluation.json +0 -51
- wisent/examples/scripts/results/test_metabench_pairs.json +0 -14
- wisent/examples/scripts/results/test_mgsm_evaluation.json +0 -51
- wisent/examples/scripts/results/test_mgsm_pairs.json +0 -14
- wisent/examples/scripts/results/test_mimic_repsum_evaluation.json +0 -30
- wisent/examples/scripts/results/test_mimic_repsum_pairs.json +0 -8
- wisent/examples/scripts/results/test_minerva_math_evaluation.json +0 -51
- wisent/examples/scripts/results/test_minerva_math_pairs.json +0 -14
- wisent/examples/scripts/results/test_mlqa_evaluation.json +0 -51
- wisent/examples/scripts/results/test_mlqa_pairs.json +0 -14
- wisent/examples/scripts/results/test_mmlu-pro-plus_evaluation.json +0 -51
- wisent/examples/scripts/results/test_mmlu-pro-plus_pairs.json +0 -14
- wisent/examples/scripts/results/test_mmlu_evaluation.json +0 -51
- wisent/examples/scripts/results/test_mmlu_pairs.json +0 -14
- wisent/examples/scripts/results/test_mmlu_pro_evaluation.json +0 -51
- wisent/examples/scripts/results/test_mmlu_pro_pairs.json +0 -14
- wisent/examples/scripts/results/test_mmlu_prox_evaluation.json +0 -51
- wisent/examples/scripts/results/test_mmlu_prox_pairs.json +0 -14
- wisent/examples/scripts/results/test_mmlusr_evaluation.json +0 -30
- wisent/examples/scripts/results/test_mmlusr_pairs.json +0 -8
- wisent/examples/scripts/results/test_mmmu_evaluation.json +0 -51
- wisent/examples/scripts/results/test_mmmu_pairs.json +0 -14
- wisent/examples/scripts/results/test_mnli_evaluation.json +0 -30
- wisent/examples/scripts/results/test_mnli_pairs.json +0 -8
- wisent/examples/scripts/results/test_model_written_evals_evaluation.json +0 -51
- wisent/examples/scripts/results/test_model_written_evals_pairs.json +0 -14
- wisent/examples/scripts/results/test_moral_stories_evaluation.json +0 -30
- wisent/examples/scripts/results/test_moral_stories_pairs.json +0 -8
- wisent/examples/scripts/results/test_mts_dialog_evaluation.json +0 -30
- wisent/examples/scripts/results/test_mts_dialog_pairs.json +0 -8
- wisent/examples/scripts/results/test_multiblimp_evaluation.json +0 -51
- wisent/examples/scripts/results/test_multiblimp_pairs.json +0 -14
- wisent/examples/scripts/results/test_multimedqa_evaluation.json +0 -51
- wisent/examples/scripts/results/test_multimedqa_pairs.json +0 -14
- wisent/examples/scripts/results/test_multipl_e_evaluation.json +0 -30
- wisent/examples/scripts/results/test_multipl_e_pairs.json +0 -8
- wisent/examples/scripts/results/test_mutual_evaluation.json +0 -30
- wisent/examples/scripts/results/test_mutual_pairs.json +0 -8
- wisent/examples/scripts/results/test_non_greedy_robustness_agieval_aqua_rat_evaluation.json +0 -30
- wisent/examples/scripts/results/test_non_greedy_robustness_agieval_aqua_rat_pairs.json +0 -8
- wisent/examples/scripts/results/test_noreval_evaluation.json +0 -51
- wisent/examples/scripts/results/test_noreval_pairs.json +0 -14
- wisent/examples/scripts/results/test_noticia_evaluation.json +0 -30
- wisent/examples/scripts/results/test_noticia_pairs.json +0 -8
- wisent/examples/scripts/results/test_nq_open_evaluation.json +0 -30
- wisent/examples/scripts/results/test_nq_open_pairs.json +0 -8
- wisent/examples/scripts/results/test_olaph_evaluation.json +0 -30
- wisent/examples/scripts/results/test_olaph_pairs.json +0 -8
- wisent/examples/scripts/results/test_openbookqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_openbookqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_openllm_evaluation.json +0 -51
- wisent/examples/scripts/results/test_openllm_pairs.json +0 -14
- wisent/examples/scripts/results/test_option_order_robustness_agieval_aqua_rat_evaluation.json +0 -30
- wisent/examples/scripts/results/test_option_order_robustness_agieval_aqua_rat_pairs.json +0 -8
- wisent/examples/scripts/results/test_paloma_evaluation.json +0 -51
- wisent/examples/scripts/results/test_paloma_pairs.json +0 -14
- wisent/examples/scripts/results/test_passkey/test_passkey_evaluation.json +0 -30
- wisent/examples/scripts/results/test_passkey/test_passkey_pairs.json +0 -8
- wisent/examples/scripts/results/test_paws-x_evaluation.json +0 -51
- wisent/examples/scripts/results/test_paws-x_pairs.json +0 -14
- wisent/examples/scripts/results/test_paws_en/test_paws_en_evaluation.json +0 -30
- wisent/examples/scripts/results/test_paws_en/test_paws_en_pairs.json +0 -8
- wisent/examples/scripts/results/test_penn_treebank_evaluation.json +0 -30
- wisent/examples/scripts/results/test_penn_treebank_pairs.json +0 -8
- wisent/examples/scripts/results/test_pile_10k/test_pile_10k_evaluation.json +0 -30
- wisent/examples/scripts/results/test_pile_10k/test_pile_10k_pairs.json +0 -8
- wisent/examples/scripts/results/test_piqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_piqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_polemo2_evaluation.json +0 -30
- wisent/examples/scripts/results/test_polemo2_pairs.json +0 -8
- wisent/examples/scripts/results/test_polymath_en_high_evaluation.json +0 -30
- wisent/examples/scripts/results/test_polymath_en_high_pairs.json +0 -8
- wisent/examples/scripts/results/test_polymath_en_medium_evaluation.json +0 -30
- wisent/examples/scripts/results/test_polymath_en_medium_pairs.json +0 -8
- wisent/examples/scripts/results/test_polymath_zh_high_evaluation.json +0 -30
- wisent/examples/scripts/results/test_polymath_zh_high_pairs.json +0 -8
- wisent/examples/scripts/results/test_polymath_zh_medium_evaluation.json +0 -30
- wisent/examples/scripts/results/test_polymath_zh_medium_pairs.json +0 -8
- wisent/examples/scripts/results/test_portuguese_bench_evaluation.json +0 -51
- wisent/examples/scripts/results/test_portuguese_bench_pairs.json +0 -14
- wisent/examples/scripts/results/test_prompt_robustness_agieval_aqua_rat/test_prompt_robustness_agieval_aqua_rat_evaluation.json +0 -30
- wisent/examples/scripts/results/test_prompt_robustness_agieval_aqua_rat/test_prompt_robustness_agieval_aqua_rat_pairs.json +0 -8
- wisent/examples/scripts/results/test_prompt_robustness_agieval_aqua_rat_evaluation.json +0 -30
- wisent/examples/scripts/results/test_prompt_robustness_agieval_aqua_rat_pairs.json +0 -8
- wisent/examples/scripts/results/test_prost_evaluation.json +0 -30
- wisent/examples/scripts/results/test_prost_pairs.json +0 -8
- wisent/examples/scripts/results/test_ptb_evaluation.json +0 -30
- wisent/examples/scripts/results/test_ptb_pairs.json +0 -8
- wisent/examples/scripts/results/test_pubmedqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_pubmedqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_pythia_evaluation.json +0 -51
- wisent/examples/scripts/results/test_pythia_pairs.json +0 -14
- wisent/examples/scripts/results/test_qa4mre_evaluation.json +0 -30
- wisent/examples/scripts/results/test_qa4mre_pairs.json +0 -8
- wisent/examples/scripts/results/test_qasper_evaluation.json +0 -30
- wisent/examples/scripts/results/test_qasper_pairs.json +0 -8
- wisent/examples/scripts/results/test_race_evaluation.json +0 -30
- wisent/examples/scripts/results/test_race_pairs.json +0 -8
- wisent/examples/scripts/results/test_realtoxicityprompts_evaluation.json +0 -30
- wisent/examples/scripts/results/test_realtoxicityprompts_pairs.json +0 -8
- wisent/examples/scripts/results/test_recode_evaluation.json +0 -30
- wisent/examples/scripts/results/test_recode_pairs.json +0 -8
- wisent/examples/scripts/results/test_record_evaluation.json +0 -30
- wisent/examples/scripts/results/test_record_pairs.json +0 -8
- wisent/examples/scripts/results/test_ruler_evaluation.json +0 -51
- wisent/examples/scripts/results/test_ruler_pairs.json +0 -14
- wisent/examples/scripts/results/test_sciq_evaluation.json +0 -30
- wisent/examples/scripts/results/test_sciq_pairs.json +0 -8
- wisent/examples/scripts/results/test_score_evaluation.json +0 -51
- wisent/examples/scripts/results/test_score_pairs.json +0 -14
- wisent/examples/scripts/results/test_self_consistency_evaluation.json +0 -30
- wisent/examples/scripts/results/test_self_consistency_pairs.json +0 -8
- wisent/examples/scripts/results/test_siqa/test_siqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_siqa/test_siqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_siqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_siqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_spanish_bench_evaluation.json +0 -51
- wisent/examples/scripts/results/test_spanish_bench_pairs.json +0 -14
- wisent/examples/scripts/results/test_squad2_evaluation.json +0 -30
- wisent/examples/scripts/results/test_squad2_pairs.json +0 -8
- wisent/examples/scripts/results/test_squadv2_evaluation.json +0 -30
- wisent/examples/scripts/results/test_squadv2_pairs.json +0 -8
- wisent/examples/scripts/results/test_super-glue-lm-eval-v1-seq2seq_evaluation.json +0 -30
- wisent/examples/scripts/results/test_super-glue-lm-eval-v1-seq2seq_pairs.json +0 -8
- wisent/examples/scripts/results/test_super-glue-lm-eval-v1_evaluation.json +0 -51
- wisent/examples/scripts/results/test_super-glue-lm-eval-v1_pairs.json +0 -14
- wisent/examples/scripts/results/test_swag_evaluation.json +0 -30
- wisent/examples/scripts/results/test_swag_pairs.json +0 -8
- wisent/examples/scripts/results/test_tinyBenchmarks_evaluation.json +0 -51
- wisent/examples/scripts/results/test_tinyBenchmarks_pairs.json +0 -14
- wisent/examples/scripts/results/test_tmmluplus_evaluation.json +0 -51
- wisent/examples/scripts/results/test_tmmluplus_pairs.json +0 -14
- wisent/examples/scripts/results/test_translation_evaluation.json +0 -51
- wisent/examples/scripts/results/test_translation_pairs.json +0 -14
- wisent/examples/scripts/results/test_triviaqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_triviaqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_truthfulqa-multi_evaluation.json +0 -51
- wisent/examples/scripts/results/test_truthfulqa-multi_pairs.json +0 -14
- wisent/examples/scripts/results/test_truthfulqa_evaluation.json +0 -30
- wisent/examples/scripts/results/test_truthfulqa_mc1_evaluation.json +0 -30
- wisent/examples/scripts/results/test_truthfulqa_mc1_pairs.json +0 -8
- wisent/examples/scripts/results/test_truthfulqa_mc2_evaluation.json +0 -30
- wisent/examples/scripts/results/test_truthfulqa_mc2_pairs.json +0 -8
- wisent/examples/scripts/results/test_truthfulqa_pairs.json +0 -8
- wisent/examples/scripts/results/test_turkishmmlu_evaluation.json +0 -51
- wisent/examples/scripts/results/test_turkishmmlu_pairs.json +0 -14
- wisent/examples/scripts/results/test_unfair_tos_evaluation.json +0 -30
- wisent/examples/scripts/results/test_unfair_tos_pairs.json +0 -8
- wisent/examples/scripts/results/test_unscramble_evaluation.json +0 -51
- wisent/examples/scripts/results/test_unscramble_pairs.json +0 -14
- wisent/examples/scripts/results/test_webqs_evaluation.json +0 -30
- wisent/examples/scripts/results/test_webqs_pairs.json +0 -8
- wisent/examples/scripts/results/test_wikitext103_evaluation.json +0 -30
- wisent/examples/scripts/results/test_wikitext103_pairs.json +0 -8
- wisent/examples/scripts/results/test_wikitext_evaluation.json +0 -30
- wisent/examples/scripts/results/test_wikitext_pairs.json +0 -8
- wisent/examples/scripts/results/test_winogender_evaluation.json +0 -51
- wisent/examples/scripts/results/test_winogender_pairs.json +0 -14
- wisent/examples/scripts/results/test_winogrande_evaluation.json +0 -30
- wisent/examples/scripts/results/test_winogrande_pairs.json +0 -8
- wisent/examples/scripts/results/test_wmdp_evaluation.json +0 -30
- wisent/examples/scripts/results/test_wmdp_pairs.json +0 -8
- wisent/examples/scripts/results/test_wmt-ro-en-t5-prompt_evaluation.json +0 -30
- wisent/examples/scripts/results/test_wmt-ro-en-t5-prompt_pairs.json +0 -8
- wisent/examples/scripts/results/test_wmt14_en_fr_evaluation.json +0 -30
- wisent/examples/scripts/results/test_wmt14_en_fr_pairs.json +0 -8
- wisent/examples/scripts/results/test_wmt16_en_de_evaluation.json +0 -30
- wisent/examples/scripts/results/test_wmt16_en_de_pairs.json +0 -8
- wisent/examples/scripts/results/test_wmt16_ro_en_evaluation.json +0 -30
- wisent/examples/scripts/results/test_wmt16_ro_en_pairs.json +0 -8
- wisent/examples/scripts/results/test_wsc273_evaluation.json +0 -30
- wisent/examples/scripts/results/test_wsc273_pairs.json +0 -8
- wisent/examples/scripts/results/test_xcopa_evaluation.json +0 -51
- wisent/examples/scripts/results/test_xcopa_pairs.json +0 -14
- wisent/examples/scripts/results/test_xnli_eu_evaluation.json +0 -30
- wisent/examples/scripts/results/test_xnli_eu_pairs.json +0 -8
- wisent/examples/scripts/results/test_xnli_evaluation.json +0 -51
- wisent/examples/scripts/results/test_xnli_pairs.json +0 -14
- wisent/examples/scripts/results/test_xquad_evaluation.json +0 -51
- wisent/examples/scripts/results/test_xquad_pairs.json +0 -14
- wisent/examples/scripts/results/test_xstorycloze_evaluation.json +0 -51
- wisent/examples/scripts/results/test_xstorycloze_pairs.json +0 -14
- wisent/examples/scripts/results/test_xsum_evaluation.json +0 -30
- wisent/examples/scripts/results/test_xsum_pairs.json +0 -8
- wisent/examples/scripts/results/test_xwinograd_evaluation.json +0 -51
- wisent/examples/scripts/results/test_xwinograd_pairs.json +0 -14
- wisent/examples/scripts/results/test_yahoo_answers_topics_evaluation.json +0 -30
- wisent/examples/scripts/results/test_yahoo_answers_topics_pairs.json +0 -8
- {wisent-0.7.379.dist-info → wisent-0.7.901.dist-info}/WHEEL +0 -0
- {wisent-0.7.379.dist-info → wisent-0.7.901.dist-info}/entry_points.txt +0 -0
- {wisent-0.7.379.dist-info → wisent-0.7.901.dist-info}/licenses/LICENSE +0 -0
- {wisent-0.7.379.dist-info → wisent-0.7.901.dist-info}/top_level.txt +0 -0
|
@@ -1,469 +0,0 @@
|
|
|
1
|
-
{
|
|
2
|
-
"wmt14_en_fr": {
|
|
3
|
-
"wisent_evaluator": "generation",
|
|
4
|
-
"actual_lm_eval_harness_metrics": ["BLEU", "TER", "CHRF"],
|
|
5
|
-
"notes": "Wisent uses generic 'generation' evaluator with string comparison. Actual WMT14 benchmark uses BLEU (n-gram overlap), TER (Translation Edit Rate), and CHRF (character-level F-score) - translation-specific metrics."
|
|
6
|
-
},
|
|
7
|
-
"mmlu": {
|
|
8
|
-
"wisent_evaluator": "log_likelihoods",
|
|
9
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
10
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
11
|
-
"notes": "Both use multiple-choice format. LM-eval measures accuracy via mean. Wisent uses log_likelihoods to compare probabilities of answer choices A/B/C/D."
|
|
12
|
-
},
|
|
13
|
-
"gsm8k": {
|
|
14
|
-
"wisent_evaluator": "exact_match",
|
|
15
|
-
"actual_lm_eval_harness_metrics": ["exact_match"],
|
|
16
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
17
|
-
"notes": "Both use exact_match for numerical answers. LM-eval uses regex to extract answers from generated text with patterns like '#### 42' or 'answer is 42'. Wisent also extracts numerical answers for comparison."
|
|
18
|
-
},
|
|
19
|
-
"hellaswag": {
|
|
20
|
-
"wisent_evaluator": "log_likelihoods",
|
|
21
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
22
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
23
|
-
"notes": "Both use multiple-choice format for sentence completion. LM-eval measures accuracy (acc) and normalized accuracy (acc_norm). Wisent uses log_likelihoods to compare probabilities of completion choices."
|
|
24
|
-
},
|
|
25
|
-
"arc_easy": {
|
|
26
|
-
"wisent_evaluator": "log_likelihoods",
|
|
27
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
28
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
29
|
-
"notes": "Both use multiple-choice format for science questions. LM-eval measures accuracy (acc) and normalized accuracy (acc_norm). Wisent uses log_likelihoods to compare probabilities of answer choices."
|
|
30
|
-
},
|
|
31
|
-
"humaneval": {
|
|
32
|
-
"wisent_evaluator": "MISSING",
|
|
33
|
-
"actual_lm_eval_harness_metrics": ["pass_at_k"],
|
|
34
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
35
|
-
"notes": "LM-eval uses pass@k metric by executing generated code against test cases (allows unsafe code execution). Wisent has NO evaluator defined for HumanEval - benchmark is broken. Should use docker_code evaluator."
|
|
36
|
-
},
|
|
37
|
-
"truthfulqa_mc1": {
|
|
38
|
-
"wisent_evaluator": "log_likelihoods",
|
|
39
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
40
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
41
|
-
"notes": "Both use multiple-choice format. LM-eval measures accuracy where first choice is correct answer. Wisent uses log_likelihoods to compare probabilities."
|
|
42
|
-
},
|
|
43
|
-
"winogrande": {
|
|
44
|
-
"wisent_evaluator": "log_likelihoods",
|
|
45
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
46
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
47
|
-
"notes": "Both use multiple-choice format for commonsense reasoning (partial evaluation method). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities of fill-in-blank options."
|
|
48
|
-
},
|
|
49
|
-
"drop": {
|
|
50
|
-
"wisent_evaluator": "exact_match",
|
|
51
|
-
"actual_lm_eval_harness_metrics": ["exact_match", "f1"],
|
|
52
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
53
|
-
"notes": "LM-eval uses both exact_match and F1 score for reading comprehension with numerical reasoning. Wisent only uses exact_match. Both generate text until period."
|
|
54
|
-
},
|
|
55
|
-
"boolq": {
|
|
56
|
-
"wisent_evaluator": "log_likelihoods",
|
|
57
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
58
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
59
|
-
"notes": "Both use multiple-choice format for yes/no questions. LM-eval measures accuracy with choices ['no', 'yes']. Wisent uses log_likelihoods to compare probabilities."
|
|
60
|
-
},
|
|
61
|
-
"piqa": {
|
|
62
|
-
"wisent_evaluator": "log_likelihoods",
|
|
63
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
64
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
65
|
-
"notes": "Both use multiple-choice format for physical commonsense QA with 2 solutions. LM-eval measures accuracy and normalized accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
66
|
-
},
|
|
67
|
-
"openbookqa": {
|
|
68
|
-
"wisent_evaluator": "log_likelihoods",
|
|
69
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
70
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
71
|
-
"notes": "Both use multiple-choice format for open-domain reading comprehension. LM-eval measures accuracy and normalized accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
72
|
-
},
|
|
73
|
-
"copa": {
|
|
74
|
-
"wisent_evaluator": "log_likelihoods",
|
|
75
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
76
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
77
|
-
"notes": "Both use multiple-choice format for Choice of Plausible Alternatives (SuperGLUE). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
78
|
-
},
|
|
79
|
-
"mathqa": {
|
|
80
|
-
"wisent_evaluator": "log_likelihoods",
|
|
81
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
82
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
83
|
-
"notes": "Both use multiple-choice format for math word problems with 5 options (a-e). LM-eval measures accuracy and normalized accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
84
|
-
},
|
|
85
|
-
"lambada_openai": {
|
|
86
|
-
"wisent_evaluator": "log_likelihoods",
|
|
87
|
-
"actual_lm_eval_harness_metrics": ["perplexity", "acc"],
|
|
88
|
-
"actual_lm_eval_harness_output_type": "loglikelihood",
|
|
89
|
-
"notes": "Both use loglikelihood for predicting final word in context. LM-eval measures both perplexity (lower is better) and accuracy. Wisent uses log_likelihoods similarly."
|
|
90
|
-
},
|
|
91
|
-
"coqa": {
|
|
92
|
-
"wisent_evaluator": "exact_match",
|
|
93
|
-
"actual_lm_eval_harness_metrics": ["exact_match", "f1"],
|
|
94
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
95
|
-
"notes": "LM-eval uses both exact_match and F1 score for conversational question answering. Wisent only uses exact_match. Both generate text until '\\nQ:'."
|
|
96
|
-
},
|
|
97
|
-
"record": {
|
|
98
|
-
"wisent_evaluator": "log_likelihoods",
|
|
99
|
-
"actual_lm_eval_harness_metrics": ["f1", "em"],
|
|
100
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
101
|
-
"notes": "LM-eval uses F1 and exact_match for SuperGLUE reading comprehension. Wisent uses log_likelihoods for multiple-choice format. Different approaches - LM-eval treats as generation task, Wisent as probability comparison."
|
|
102
|
-
},
|
|
103
|
-
"race": {
|
|
104
|
-
"wisent_evaluator": "log_likelihoods",
|
|
105
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
106
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
107
|
-
"notes": "Both use multiple-choice format for reading comprehension. LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
108
|
-
},
|
|
109
|
-
"headqa": {
|
|
110
|
-
"wisent_evaluator": "log_likelihoods",
|
|
111
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
112
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
113
|
-
"notes": "Both use multiple-choice format for medical knowledge questions. LM-eval measures accuracy and normalized accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
114
|
-
},
|
|
115
|
-
"logiqa": {
|
|
116
|
-
"wisent_evaluator": "log_likelihoods",
|
|
117
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
118
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
119
|
-
"notes": "Both use multiple-choice format for logical reasoning. LM-eval measures accuracy and normalized accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
120
|
-
},
|
|
121
|
-
"storycloze": {
|
|
122
|
-
"wisent_evaluator": "log_likelihoods",
|
|
123
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
124
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
125
|
-
"notes": "Both use multiple-choice format for story ending selection (2 choices). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
126
|
-
},
|
|
127
|
-
"swag": {
|
|
128
|
-
"wisent_evaluator": "log_likelihoods",
|
|
129
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
130
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
131
|
-
"notes": "Both use multiple-choice format for situation completion with 4 endings. LM-eval measures accuracy and normalized accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
132
|
-
},
|
|
133
|
-
"arithmetic": {
|
|
134
|
-
"wisent_evaluator": "log_likelihoods",
|
|
135
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
136
|
-
"actual_lm_eval_harness_output_type": "loglikelihood",
|
|
137
|
-
"notes": "Both use loglikelihood for arithmetic problems. LM-eval measures accuracy. Wisent uses log_likelihoods similarly."
|
|
138
|
-
},
|
|
139
|
-
"asdiv": {
|
|
140
|
-
"wisent_evaluator": "log_likelihoods",
|
|
141
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
142
|
-
"actual_lm_eval_harness_output_type": "loglikelihood",
|
|
143
|
-
"notes": "Both use loglikelihood for arithmetic word problems. LM-eval measures accuracy by extracting answers before parentheses. Wisent uses log_likelihoods similarly."
|
|
144
|
-
},
|
|
145
|
-
"medqa": {
|
|
146
|
-
"wisent_evaluator": "log_likelihoods",
|
|
147
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
148
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
149
|
-
"notes": "Both use multiple-choice format for USMLE medical questions with 4 options (A/B/C/D). LM-eval measures accuracy and normalized accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
150
|
-
},
|
|
151
|
-
"cb": {
|
|
152
|
-
"wisent_evaluator": "log_likelihoods",
|
|
153
|
-
"actual_lm_eval_harness_metrics": ["acc", "f1"],
|
|
154
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
155
|
-
"notes": "Both use multiple-choice format for SuperGLUE CommitmentBank with 3 options (True/False/Neither). LM-eval measures accuracy and F1 score with multi-class aggregation. Wisent uses log_likelihoods to compare probabilities."
|
|
156
|
-
},
|
|
157
|
-
"rte": {
|
|
158
|
-
"wisent_evaluator": "log_likelihoods",
|
|
159
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
160
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
161
|
-
"notes": "Both use multiple-choice format for SuperGLUE Recognizing Textual Entailment with binary choices (True/False). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
162
|
-
},
|
|
163
|
-
"wic": {
|
|
164
|
-
"wisent_evaluator": "log_likelihoods",
|
|
165
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
166
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
167
|
-
"notes": "Both use multiple-choice format for SuperGLUE Word-in-Context with binary choices (no/yes). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
168
|
-
},
|
|
169
|
-
"wsc": {
|
|
170
|
-
"wisent_evaluator": "log_likelihoods",
|
|
171
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
172
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
173
|
-
"notes": "Both use multiple-choice format for SuperGLUE Winograd Schema Challenge with binary choices (no/yes). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
174
|
-
},
|
|
175
|
-
"multirc": {
|
|
176
|
-
"wisent_evaluator": "log_likelihoods",
|
|
177
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
178
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
179
|
-
"notes": "Both use multiple-choice format for SuperGLUE MultiRC reading comprehension. LM-eval measures accuracy by asking 'Is the answer correct? yes/no' for each option. Wisent uses log_likelihoods to compare probabilities."
|
|
180
|
-
},
|
|
181
|
-
"mrpc": {
|
|
182
|
-
"wisent_evaluator": "log_likelihoods",
|
|
183
|
-
"actual_lm_eval_harness_metrics": ["acc", "f1"],
|
|
184
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
185
|
-
"notes": "Both use multiple-choice format for GLUE Microsoft Research Paraphrase Corpus with binary choices (no/yes). LM-eval measures accuracy and F1 score. Wisent uses log_likelihoods to compare probabilities."
|
|
186
|
-
},
|
|
187
|
-
"qqp": {
|
|
188
|
-
"wisent_evaluator": "log_likelihoods",
|
|
189
|
-
"actual_lm_eval_harness_metrics": ["acc", "f1"],
|
|
190
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
191
|
-
"notes": "Both use multiple-choice format for GLUE Quora Question Pairs with binary choices (no/yes). LM-eval measures accuracy and F1 score. Wisent uses log_likelihoods to compare probabilities."
|
|
192
|
-
},
|
|
193
|
-
"sst2": {
|
|
194
|
-
"wisent_evaluator": "log_likelihoods",
|
|
195
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
196
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
197
|
-
"notes": "Both use multiple-choice format for GLUE sentiment classification with binary choices (negative/positive). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
198
|
-
},
|
|
199
|
-
"qnli": {
|
|
200
|
-
"wisent_evaluator": "log_likelihoods",
|
|
201
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
202
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
203
|
-
"notes": "Both use multiple-choice format for GLUE Question NLI with binary choices (yes/no). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
204
|
-
},
|
|
205
|
-
"wnli": {
|
|
206
|
-
"wisent_evaluator": "log_likelihoods",
|
|
207
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
208
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
209
|
-
"notes": "Both use multiple-choice format for GLUE Winograd NLI with binary choices (False/True). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
210
|
-
},
|
|
211
|
-
"sciq": {
|
|
212
|
-
"wisent_evaluator": "log_likelihoods",
|
|
213
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
214
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
215
|
-
"notes": "Both use multiple-choice format for science questions with 4 options (3 distractors + correct). LM-eval measures accuracy and normalized accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
216
|
-
},
|
|
217
|
-
"commonsense_qa": {
|
|
218
|
-
"wisent_evaluator": "log_likelihoods",
|
|
219
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
220
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
221
|
-
"notes": "Both use multiple-choice format for commonsense reasoning with 5 options (A-E). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
222
|
-
},
|
|
223
|
-
"triviaqa": {
|
|
224
|
-
"wisent_evaluator": "exact_match",
|
|
225
|
-
"actual_lm_eval_harness_metrics": ["exact_match"],
|
|
226
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
227
|
-
"notes": "Both use exact_match for trivia QA. LM-eval generates text until newline/period/comma with case-insensitive and punctuation-normalized matching. Wisent uses exact_match similarly."
|
|
228
|
-
},
|
|
229
|
-
"nq_open": {
|
|
230
|
-
"wisent_evaluator": "exact_match",
|
|
231
|
-
"actual_lm_eval_harness_metrics": ["exact_match"],
|
|
232
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
233
|
-
"notes": "Both use exact_match for Natural Questions Open. LM-eval generates text until newline/period/comma with case-insensitive matching and article stripping. Wisent uses exact_match similarly."
|
|
234
|
-
},
|
|
235
|
-
"webqs": {
|
|
236
|
-
"wisent_evaluator": "exact_match",
|
|
237
|
-
"actual_lm_eval_harness_metrics": ["exact_match"],
|
|
238
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
239
|
-
"notes": "Both use exact_match for web questions. LM-eval uses multiple-choice format with Freebase knowledge. Wisent uses exact_match similarly."
|
|
240
|
-
},
|
|
241
|
-
"arc_challenge": {
|
|
242
|
-
"wisent_evaluator": "log_likelihoods",
|
|
243
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
244
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
245
|
-
"notes": "Both use multiple-choice format for AI2 ARC Challenge science questions. LM-eval measures accuracy and normalized accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
246
|
-
},
|
|
247
|
-
"toxigen": {
|
|
248
|
-
"wisent_evaluator": "log_likelihoods",
|
|
249
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
250
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
251
|
-
"notes": "Both use multiple-choice format for hateful content detection with binary choices (No/Yes). LM-eval measures accuracy and normalized accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
252
|
-
},
|
|
253
|
-
"pubmedqa": {
|
|
254
|
-
"wisent_evaluator": "log_likelihoods",
|
|
255
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
256
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
257
|
-
"notes": "Both use multiple-choice format for PubMed abstracts with 3 options (yes/no/maybe). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
258
|
-
},
|
|
259
|
-
"anli_r1": {
|
|
260
|
-
"wisent_evaluator": "log_likelihoods",
|
|
261
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
262
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
263
|
-
"notes": "Both use multiple-choice format for Adversarial NLI Round 1 with 3 options (True/False/Neither). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
264
|
-
},
|
|
265
|
-
"mnli": {
|
|
266
|
-
"wisent_evaluator": "log_likelihoods",
|
|
267
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
268
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
269
|
-
"notes": "Both use multiple-choice format for GLUE Multi-Genre NLI with 3 options (True/Neither/False). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
270
|
-
},
|
|
271
|
-
"cola": {
|
|
272
|
-
"wisent_evaluator": "log_likelihoods",
|
|
273
|
-
"actual_lm_eval_harness_metrics": ["mcc"],
|
|
274
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
275
|
-
"notes": "LM-eval uses Matthews Correlation Coefficient for GLUE Corpus of Linguistic Acceptability with binary choices (no/yes). Wisent uses log_likelihoods. Different metrics - MCC vs log probability comparison."
|
|
276
|
-
},
|
|
277
|
-
"logiqa2": {
|
|
278
|
-
"wisent_evaluator": "log_likelihoods",
|
|
279
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
280
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
281
|
-
"notes": "Both use multiple-choice format for logical reasoning. LM-eval measures accuracy and normalized accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
282
|
-
},
|
|
283
|
-
"mc_taco": {
|
|
284
|
-
"wisent_evaluator": "log_likelihoods",
|
|
285
|
-
"actual_lm_eval_harness_metrics": ["acc", "f1"],
|
|
286
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
287
|
-
"notes": "Both use multiple-choice format for temporal reasoning with binary plausibility judgments (no/yes). LM-eval measures accuracy and F1 score. Wisent uses log_likelihoods to compare probabilities."
|
|
288
|
-
},
|
|
289
|
-
"glue_rte": {
|
|
290
|
-
"wisent_evaluator": "log_likelihoods",
|
|
291
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
292
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
293
|
-
"notes": "Both use multiple-choice format for GLUE RTE with binary choices (True/False). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
294
|
-
},
|
|
295
|
-
"polemo2_in": {
|
|
296
|
-
"wisent_evaluator": "log_likelihoods",
|
|
297
|
-
"actual_lm_eval_harness_metrics": ["f1", "acc"],
|
|
298
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
299
|
-
"notes": "LM-eval uses micro-averaged F1 and accuracy for Polish sentiment with 4 classes (A-D). Generates text until period/comma, extracts letter. Wisent uses log_likelihoods for probability comparison."
|
|
300
|
-
},
|
|
301
|
-
"mutual": {
|
|
302
|
-
"wisent_evaluator": "log_likelihoods",
|
|
303
|
-
"actual_lm_eval_harness_metrics": ["r@1", "r@2", "mrr"],
|
|
304
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
305
|
-
"notes": "LM-eval uses ranking metrics (recall@1, recall@2, MRR) for multiple-choice dialogue with 4 options. Wisent uses log_likelihoods. Different evaluation approach - ranking vs probability comparison."
|
|
306
|
-
},
|
|
307
|
-
"siqa": {
|
|
308
|
-
"wisent_evaluator": "log_likelihoods",
|
|
309
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
310
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
311
|
-
"notes": "Both use multiple-choice format for Social IQA with 3 options (answerA/B/C). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
312
|
-
},
|
|
313
|
-
"ifeval": {
|
|
314
|
-
"wisent_evaluator": "generation",
|
|
315
|
-
"actual_lm_eval_harness_metrics": ["prompt_level_strict_acc", "inst_level_strict_acc", "prompt_level_loose_acc", "inst_level_loose_acc"],
|
|
316
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
317
|
-
"notes": "LM-eval uses 4 instruction-following accuracy metrics (strict/loose at prompt/instruction level) for IFEval. Generates up to 1280 tokens. Wisent uses generic generation evaluator. Different evaluation - instruction compliance vs string comparison."
|
|
318
|
-
},
|
|
319
|
-
"babi": {
|
|
320
|
-
"wisent_evaluator": "exact_match",
|
|
321
|
-
"actual_lm_eval_harness_metrics": ["exact_match"],
|
|
322
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
323
|
-
"notes": "Both use exact_match for bAbI reading comprehension. LM-eval generates text until newline or 'Passage:'. Wisent uses exact_match similarly."
|
|
324
|
-
},
|
|
325
|
-
"crows_pairs_english": {
|
|
326
|
-
"wisent_evaluator": "log_likelihoods",
|
|
327
|
-
"actual_lm_eval_harness_metrics": ["likelihood_difference", "pct_stereotype"],
|
|
328
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
329
|
-
"notes": "LM-eval uses bias metrics (likelihood difference, stereotype percentage) for CROWS-Pairs bias detection. Lower scores are better. Wisent uses log_likelihoods. Different purpose - bias measurement vs task performance."
|
|
330
|
-
},
|
|
331
|
-
"mbpp": {
|
|
332
|
-
"wisent_evaluator": "MISSING",
|
|
333
|
-
"actual_lm_eval_harness_metrics": ["pass@1"],
|
|
334
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
335
|
-
"notes": "LM-eval uses pass@1 metric by executing generated Python code against test cases (allows unsafe code execution). Wisent has NO evaluator defined for MBPP - benchmark is broken. Should use docker_code evaluator."
|
|
336
|
-
},
|
|
337
|
-
"realtoxicityprompts": {
|
|
338
|
-
"wisent_evaluator": "generation",
|
|
339
|
-
"actual_lm_eval_harness_metrics": ["toxicity_perspective_api", "perspective_api_toxicity_score"],
|
|
340
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
341
|
-
"notes": "LM-eval uses Perspective API to measure toxicity in generated completions. Lower scores are better. Wisent uses generic generation evaluator. Different purpose - toxicity measurement vs string comparison."
|
|
342
|
-
},
|
|
343
|
-
"leaderboard_mmlu_pro": {
|
|
344
|
-
"wisent_evaluator": "log_likelihoods",
|
|
345
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
346
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
347
|
-
"notes": "Both use multiple-choice format for MMLU-Pro leaderboard task. LM-eval uses 5-shot with validation examples, measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
348
|
-
},
|
|
349
|
-
"leaderboard_ifeval": {
|
|
350
|
-
"wisent_evaluator": "generation",
|
|
351
|
-
"actual_lm_eval_harness_metrics": ["prompt_level_strict_acc", "inst_level_strict_acc", "prompt_level_loose_acc", "inst_level_loose_acc"],
|
|
352
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
353
|
-
"notes": "LM-eval uses 4 instruction-following accuracy metrics for leaderboard IFEval. Generates up to 1280 tokens. Wisent uses generic generation evaluator. Different evaluation - instruction compliance vs string comparison."
|
|
354
|
-
},
|
|
355
|
-
"medmcqa": {
|
|
356
|
-
"wisent_evaluator": "log_likelihoods",
|
|
357
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
358
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
359
|
-
"notes": "Both use multiple-choice format for medical MCQ with 4 options (A/B/C/D). LM-eval measures accuracy and normalized accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
360
|
-
},
|
|
361
|
-
"tinyHellaswag": {
|
|
362
|
-
"wisent_evaluator": "log_likelihoods",
|
|
363
|
-
"actual_lm_eval_harness_metrics": ["acc_norm"],
|
|
364
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
365
|
-
"notes": "Both use multiple-choice format for tiny HellaSwag subset. LM-eval uses normalized accuracy with custom GPIRT aggregation. Wisent uses log_likelihoods to compare probabilities."
|
|
366
|
-
},
|
|
367
|
-
"tinyMMLU": {
|
|
368
|
-
"wisent_evaluator": "log_likelihoods",
|
|
369
|
-
"actual_lm_eval_harness_metrics": ["acc_norm"],
|
|
370
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
371
|
-
"notes": "Both use multiple-choice format for tiny MMLU subset with 4 options (A/B/C/D). LM-eval uses normalized accuracy with custom GPIRT aggregation. Wisent uses log_likelihoods to compare probabilities."
|
|
372
|
-
},
|
|
373
|
-
"tinyGSM8k": {
|
|
374
|
-
"wisent_evaluator": "exact_match",
|
|
375
|
-
"actual_lm_eval_harness_metrics": ["exact_match"],
|
|
376
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
377
|
-
"notes": "Both use exact_match for tiny GSM8k subset. LM-eval uses custom GPIRT aggregation with regex answer extraction. Wisent uses exact_match similarly."
|
|
378
|
-
},
|
|
379
|
-
"tinyArc": {
|
|
380
|
-
"wisent_evaluator": "log_likelihoods",
|
|
381
|
-
"actual_lm_eval_harness_metrics": ["acc_norm"],
|
|
382
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
383
|
-
"notes": "Both use multiple-choice format for tiny ARC-Challenge subset with 25-shot. LM-eval uses normalized accuracy with custom GPIRT aggregation. Wisent uses log_likelihoods to compare probabilities."
|
|
384
|
-
},
|
|
385
|
-
"tinyWinogrande": {
|
|
386
|
-
"wisent_evaluator": "log_likelihoods",
|
|
387
|
-
"actual_lm_eval_harness_metrics": ["acc_norm"],
|
|
388
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
389
|
-
"notes": "Both use multiple-choice format for tiny Winogrande subset with 5-shot. LM-eval uses normalized accuracy with custom GPIRT aggregation. Wisent uses log_likelihoods to compare probabilities."
|
|
390
|
-
},
|
|
391
|
-
"wikitext": {
|
|
392
|
-
"wisent_evaluator": "log_likelihoods",
|
|
393
|
-
"actual_lm_eval_harness_metrics": ["word_perplexity", "byte_perplexity", "bits_per_byte"],
|
|
394
|
-
"actual_lm_eval_harness_output_type": "loglikelihood_rolling",
|
|
395
|
-
"notes": "LM-eval uses perplexity metrics for language modeling with loglikelihood_rolling evaluation. Wisent uses log_likelihoods. Different purpose - language modeling evaluation vs task performance."
|
|
396
|
-
},
|
|
397
|
-
"qa4mre_2011": {
|
|
398
|
-
"wisent_evaluator": "log_likelihoods",
|
|
399
|
-
"actual_lm_eval_harness_metrics": ["acc", "acc_norm"],
|
|
400
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
401
|
-
"notes": "Both use multiple-choice format for QA4MRE 2011 reading comprehension. LM-eval measures accuracy and normalized accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
402
|
-
},
|
|
403
|
-
"agieval_en": {
|
|
404
|
-
"wisent_evaluator": "log_likelihoods",
|
|
405
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
406
|
-
"actual_lm_eval_harness_output_type": "unknown",
|
|
407
|
-
"notes": "LM-eval uses accuracy metric with mean aggregation and weight_by_size: true. Output type not explicitly specified in config. Wisent uses log_likelihoods to compare probabilities."
|
|
408
|
-
},
|
|
409
|
-
"squadv2": {
|
|
410
|
-
"wisent_evaluator": "exact_match",
|
|
411
|
-
"actual_lm_eval_harness_metrics": ["exact_match", "f1"],
|
|
412
|
-
"actual_lm_eval_harness_output_type": "function-based (SQuAD2)",
|
|
413
|
-
"notes": "LM-eval uses function-based task class for SQuAD v2 with exact_match and F1 metrics. Wisent uses exact_match. Both handle questions where no answer is possible."
|
|
414
|
-
},
|
|
415
|
-
"xwinograd_en": {
|
|
416
|
-
"wisent_evaluator": "log_likelihoods",
|
|
417
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
418
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
419
|
-
"notes": "Both use multiple-choice format for cross-lingual Winograd Schema Challenge. LM-eval measures accuracy with mean aggregation. Wisent uses log_likelihoods to compare probabilities."
|
|
420
|
-
},
|
|
421
|
-
"scrolls_qasper": {
|
|
422
|
-
"wisent_evaluator": "exact_match",
|
|
423
|
-
"actual_lm_eval_harness_metrics": ["exact_match", "f1"],
|
|
424
|
-
"actual_lm_eval_harness_output_type": "function-based (Qasper)",
|
|
425
|
-
"notes": "LM-eval uses function-based task class for SCROLLS Qasper with exact_match and F1 metrics for scientific paper QA. Wisent uses exact_match. Both handle long document reasoning."
|
|
426
|
-
},
|
|
427
|
-
"truthfulqa_mc2": {
|
|
428
|
-
"wisent_evaluator": "log_likelihoods",
|
|
429
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
430
|
-
"actual_lm_eval_harness_output_type": "mc2",
|
|
431
|
-
"notes": "LM-eval uses mc2 (multiple choice with 2 options) format for TruthfulQA. Measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
432
|
-
},
|
|
433
|
-
"truthfulqa_gen": {
|
|
434
|
-
"wisent_evaluator": "generation",
|
|
435
|
-
"actual_lm_eval_harness_metrics": ["bleu_max", "bleu_acc", "bleu_diff", "rouge1_max", "rouge1_acc", "rouge1_diff", "rouge2_max", "rouge2_acc", "rouge2_diff", "rougeL_max", "rougeL_acc", "rougeL_diff"],
|
|
436
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
437
|
-
"notes": "LM-eval uses 12 BLEU/ROUGE metrics (max/acc/diff variants) for generation-based TruthfulQA. Wisent uses generic generation evaluator. Different evaluation - semantic similarity vs string comparison."
|
|
438
|
-
},
|
|
439
|
-
"xnli_en": {
|
|
440
|
-
"wisent_evaluator": "log_likelihoods",
|
|
441
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
442
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
443
|
-
"notes": "Both use multiple-choice format for cross-lingual NLI with 3 options (True/Neither/False). LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
444
|
-
},
|
|
445
|
-
"winogender": {
|
|
446
|
-
"wisent_evaluator": "log_likelihoods",
|
|
447
|
-
"actual_lm_eval_harness_metrics": ["acc"],
|
|
448
|
-
"actual_lm_eval_harness_output_type": "multiple_choice",
|
|
449
|
-
"notes": "Both use multiple-choice format for gender bias detection in Winograd schemas. LM-eval measures accuracy. Wisent uses log_likelihoods to compare probabilities."
|
|
450
|
-
},
|
|
451
|
-
"hendrycks_math": {
|
|
452
|
-
"wisent_evaluator": "exact_match",
|
|
453
|
-
"actual_lm_eval_harness_metrics": ["exact_match"],
|
|
454
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
455
|
-
"notes": "Both use exact_match for Hendrycks MATH dataset with LaTeX math problems. LM-eval generates text with weighted aggregation by dataset size. Wisent uses exact_match similarly."
|
|
456
|
-
},
|
|
457
|
-
"aime": {
|
|
458
|
-
"wisent_evaluator": "exact_match",
|
|
459
|
-
"actual_lm_eval_harness_metrics": ["exact_match"],
|
|
460
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
461
|
-
"notes": "Both use exact_match for AIME (American Invitational Mathematics Examination) competition problems. LM-eval generates text with mean aggregation. Wisent uses exact_match similarly."
|
|
462
|
-
},
|
|
463
|
-
"xquad_en": {
|
|
464
|
-
"wisent_evaluator": "exact_match",
|
|
465
|
-
"actual_lm_eval_harness_metrics": ["exact_match", "f1"],
|
|
466
|
-
"actual_lm_eval_harness_output_type": "generate_until",
|
|
467
|
-
"notes": "LM-eval uses exact_match and F1 for cross-lingual QA (English subset of XQuAD). Generates text for reading comprehension. Wisent uses exact_match. Both evaluate extractive QA."
|
|
468
|
-
}
|
|
469
|
-
}
|