EuroEval 16.2.0__tar.gz → 16.2.2__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of EuroEval might be problematic. Click here for more details.
- {euroeval-16.2.0 → euroeval-16.2.2}/CHANGELOG.md +13 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/PKG-INFO +1 -1
- {euroeval-16.2.0 → euroeval-16.2.2}/pyproject.toml +1 -1
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/benchmark_config_factory.py +41 -129
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/benchmark_modules/vllm.py +10 -7
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/benchmarker.py +183 -337
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/data_models.py +35 -34
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/utils.py +9 -4
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/conftest.py +9 -0
- euroeval-16.2.2/tests/test_cli.py +69 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_data_models.py +47 -9
- {euroeval-16.2.0 → euroeval-16.2.2}/uv.lock +1 -1
- euroeval-16.2.0/tests/test_cli.py +0 -81
- {euroeval-16.2.0 → euroeval-16.2.2}/.github/ISSUE_TEMPLATE/benchmark_dataset_request.yaml +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/.github/ISSUE_TEMPLATE/bug.yaml +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/.github/ISSUE_TEMPLATE/feature_request.yaml +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/.github/ISSUE_TEMPLATE/language_request.yaml +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/.github/ISSUE_TEMPLATE/model_evaluation_request.yaml +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/.github/workflows/ci.yaml +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/.gitignore +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/.pre-commit-config.yaml +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/CITATION.cff +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/CODE_OF_CONDUCT.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/CONTRIBUTING.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/Dockerfile.cuda +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/LICENSE +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/NEW_DATASET_GUIDE.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/README.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/CNAME +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/README.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/README.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/danish.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/dutch.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/english.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/estonian.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/faroese.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/finnish.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/french.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/german.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/icelandic.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/italian.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/latvian.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/norwegian.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/polish.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/portuguese.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/spanish.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/datasets/swedish.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/extras/radial_plotter.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/faq.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/gfx/favicon.png +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Monolingual/danish.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Monolingual/dutch.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Monolingual/english.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Monolingual/estonian.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Monolingual/faroese.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Monolingual/finnish.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Monolingual/french.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Monolingual/german.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Monolingual/icelandic.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Monolingual/italian.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Monolingual/norwegian.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Monolingual/portuguese.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Monolingual/spanish.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Monolingual/swedish.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Multilingual/european.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Multilingual/finnic.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Multilingual/germanic.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Multilingual/mainland-scandinavian.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/Multilingual/romance.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/leaderboards/README.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/methodology.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/python-package.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/tasks/README.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/tasks/common-sense-reasoning.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/tasks/knowledge.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/tasks/linguistic-acceptability.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/tasks/named-entity-recognition.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/tasks/reading-comprehension.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/tasks/sentiment-classification.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/tasks/speed.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/docs/tasks/summarization.md +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/gfx/euroeval.png +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/gfx/euroeval.xcf +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/gfx/scandeval.png +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/makefile +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/mkdocs.yaml +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/__init__.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/benchmark_modules/__init__.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/benchmark_modules/base.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/benchmark_modules/fresh.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/benchmark_modules/hf.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/benchmark_modules/litellm.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/callbacks.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/cli.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/constants.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/data_loading.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/__init__.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/danish.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/dutch.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/english.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/estonian.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/faroese.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/finnish.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/french.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/german.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/icelandic.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/italian.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/latvian.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/norwegian.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/polish.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/portuguese.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/spanish.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/dataset_configs/swedish.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/enums.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/exceptions.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/finetuning.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/generation.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/generation_utils.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/languages.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/metrics/__init__.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/metrics/base.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/metrics/huggingface.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/metrics/llm_as_a_judge.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/metrics/pipeline.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/metrics/speed.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/model_cache.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/model_config.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/model_loading.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/prompt_templates/__init__.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/prompt_templates/linguistic_acceptability.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/prompt_templates/multiple_choice.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/prompt_templates/named_entity_recognition.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/prompt_templates/reading_comprehension.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/prompt_templates/sentiment_classification.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/prompt_templates/summarization.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/scores.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/speed_benchmark.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/task_group_utils/__init__.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/task_group_utils/multiple_choice_classification.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/task_group_utils/question_answering.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/task_group_utils/sequence_classification.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/task_group_utils/text_to_text.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/task_group_utils/token_classification.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/tasks.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/tokenisation_utils.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/euroeval/types.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/constants.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_allocine.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_angry_tweets.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_arc.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_arc_is.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_belebele.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_boolq_pt.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_cnn_dailymail.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_conll_en.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_conll_es.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_conll_nl.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_copa_lv.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_dane.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_danish_citizen_tests.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_dansk.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_danske_talemaader.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_danske_talemaader_old.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_dbrd.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_dutch_cola.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_eltec.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_err_news.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_estner.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_estonian_valence.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_european_values.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_exam_et.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_fone.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_foqa.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_fosent.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_fquad.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_fullstack_ner.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_germanquad.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_germeval.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_goldenswag.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_grammar_et.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_harem.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_hellaswag.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_hellaswag_fi.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_hotter_and_colder_sentiment.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_ice_linguistic.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_icelandic_error_corpus.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_icelandic_knowledge.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_icelandic_qa.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_icesum.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_idioms_no.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_ilpost_sum.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_jentoft.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_kpwr_ner.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_latvian_lsm_summary.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_latvian_twitter_sentiment.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_life_in_the_uk.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_llmzszl.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_mim_gold_ner.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_mlqa_es.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_mlsum_de.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_mlsum_es.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_mmlu.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_mmlu_lv.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_multi_wiki_qa.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_multinerd-it.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_no_cola.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_no_sammendrag.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_nor_common_sense_qa.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_nordjylland_news.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_norec.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_norglm_multiqa.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_norglm_multisum.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_norne.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_norquad.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_nqii.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_nrk_quiz_qa.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_orange_sum.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_personal_sum.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_polemo2.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_poquad.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_psc.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_publico.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_rrn.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_sb10k.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_scala.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_scandiqa.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_scandisent_fi.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_schibsted.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_sentiment_headlines_es.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_sentipolc16.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_squad.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_squad_it.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_squad_nl.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_squad_nl_old.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_sst2_pt.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_sst5.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_suc3.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_swedish_skolprov.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_swedn.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_swerec.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_trivia_et.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_turku_ner_fi.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_tydiqa_fi.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_wiki_lingua_nl.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_wikiann_lv.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_wikineural-it.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_winogrande.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_winogrande_et.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_winogrande_is.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_xlsum_fi.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/create_xquad.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/fix_dot_env_file.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/load_ud_pos.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/src/scripts/versioning.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/__init__.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_benchmark_config_factory.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_benchmark_modules/__init__.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_benchmark_modules/test_hf.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_benchmarker.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_callbacks.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_constants.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_data_loading.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_dataset_configs.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_enums.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_exceptions.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_finetuning.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_languages.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_model_config.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_model_loading.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_scores.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_speed_benchmark.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_tasks.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_tokenisation_utils.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_types.py +0 -0
- {euroeval-16.2.0 → euroeval-16.2.2}/tests/test_utils.py +0 -0
|
@@ -10,6 +10,19 @@ and this project adheres to [Semantic Versioning](http://semver.org/spec/v2.0.0.
|
|
|
10
10
|
|
|
11
11
|
|
|
12
12
|
|
|
13
|
+
## [v16.2.2] - 2025-09-15
|
|
14
|
+
### Fixed
|
|
15
|
+
- Added missing benchmark arguments to the `Benchmarker.benchmark` method.
|
|
16
|
+
- Fixed another issue related to the `download_only` mode, causing model evaluations to
|
|
17
|
+
fail, as it could not find the model locally. This has been fixed now.
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
## [v16.2.1] - 2025-09-15
|
|
21
|
+
### Fixed
|
|
22
|
+
- Some of the `download_only` arguments were missing in the code, and have now been
|
|
23
|
+
added.
|
|
24
|
+
|
|
25
|
+
|
|
13
26
|
## [v16.2.0] - 2025-09-15
|
|
14
27
|
### Added
|
|
15
28
|
- Now supports evaluating models in an offline environment. This is done by first
|
|
@@ -6,9 +6,9 @@ import typing as t
|
|
|
6
6
|
|
|
7
7
|
import torch
|
|
8
8
|
|
|
9
|
-
from .data_models import BenchmarkConfig
|
|
9
|
+
from .data_models import BenchmarkConfig, BenchmarkConfigParams
|
|
10
10
|
from .dataset_configs import get_all_dataset_configs
|
|
11
|
-
from .enums import Device
|
|
11
|
+
from .enums import Device
|
|
12
12
|
from .exceptions import InvalidBenchmark
|
|
13
13
|
from .languages import get_all_languages
|
|
14
14
|
from .tasks import SPEED, get_all_tasks
|
|
@@ -21,154 +21,66 @@ logger = logging.getLogger("euroeval")
|
|
|
21
21
|
|
|
22
22
|
|
|
23
23
|
def build_benchmark_config(
|
|
24
|
-
|
|
25
|
-
save_results: bool,
|
|
26
|
-
task: str | list[str] | None,
|
|
27
|
-
dataset: str | list[str] | None,
|
|
28
|
-
language: str | list[str],
|
|
29
|
-
model_language: str | list[str] | None,
|
|
30
|
-
dataset_language: str | list[str] | None,
|
|
31
|
-
device: Device | None,
|
|
32
|
-
batch_size: int,
|
|
33
|
-
raise_errors: bool,
|
|
34
|
-
cache_dir: str,
|
|
35
|
-
api_key: str | None,
|
|
36
|
-
force: bool,
|
|
37
|
-
verbose: bool,
|
|
38
|
-
trust_remote_code: bool,
|
|
39
|
-
clear_model_cache: bool,
|
|
40
|
-
evaluate_test_split: bool,
|
|
41
|
-
few_shot: bool,
|
|
42
|
-
num_iterations: int,
|
|
43
|
-
api_base: str | None,
|
|
44
|
-
api_version: str | None,
|
|
45
|
-
gpu_memory_utilization: float,
|
|
46
|
-
generative_type: GenerativeType | None,
|
|
47
|
-
debug: bool,
|
|
48
|
-
run_with_cli: bool,
|
|
49
|
-
requires_safetensors: bool,
|
|
50
|
-
download_only: bool,
|
|
24
|
+
benchmark_config_params: BenchmarkConfigParams,
|
|
51
25
|
) -> BenchmarkConfig:
|
|
52
26
|
"""Create a benchmark configuration.
|
|
53
27
|
|
|
54
28
|
Args:
|
|
55
|
-
|
|
56
|
-
|
|
57
|
-
save_results:
|
|
58
|
-
Whether to save the benchmark results to a file.
|
|
59
|
-
task:
|
|
60
|
-
The tasks to include for dataset. If None then datasets will not be
|
|
61
|
-
filtered based on their task.
|
|
62
|
-
dataset:
|
|
63
|
-
The datasets to include for task. If None then all datasets will be
|
|
64
|
-
included, limited by the `task` parameter.
|
|
65
|
-
language:
|
|
66
|
-
The language codes of the languages to include, both for models and
|
|
67
|
-
datasets. Here 'no' means both Bokmål (nb) and Nynorsk (nn). Set this
|
|
68
|
-
to 'all' if all languages should be considered.
|
|
69
|
-
model_language:
|
|
70
|
-
The language codes of the languages to include for models. If None then
|
|
71
|
-
the `language` parameter will be used.
|
|
72
|
-
dataset_language:
|
|
73
|
-
The language codes of the languages to include for datasets. If None then
|
|
74
|
-
the `language` parameter will be used.
|
|
75
|
-
device:
|
|
76
|
-
The device to use for running the models. If None then the device will be
|
|
77
|
-
set automatically.
|
|
78
|
-
batch_size:
|
|
79
|
-
The batch size to use for running the models.
|
|
80
|
-
raise_errors:
|
|
81
|
-
Whether to raise errors when running the benchmark.
|
|
82
|
-
cache_dir:
|
|
83
|
-
The directory to use for caching the models.
|
|
84
|
-
api_key:
|
|
85
|
-
The API key to use for a given inference server.
|
|
86
|
-
force:
|
|
87
|
-
Whether to force the benchmark to run even if the results are already
|
|
88
|
-
cached.
|
|
89
|
-
verbose:
|
|
90
|
-
Whether to print verbose output when running the benchmark. This is
|
|
91
|
-
automatically set if `debug` is True.
|
|
92
|
-
trust_remote_code:
|
|
93
|
-
Whether to trust remote code when running the benchmark.
|
|
94
|
-
clear_model_cache:
|
|
95
|
-
Whether to clear the model cache before running the benchmark.
|
|
96
|
-
evaluate_test_split:
|
|
97
|
-
Whether to use the test split for the datasets.
|
|
98
|
-
few_shot:
|
|
99
|
-
Whether to use few-shot learning for the models.
|
|
100
|
-
num_iterations:
|
|
101
|
-
The number of iterations each model should be evaluated for.
|
|
102
|
-
api_base:
|
|
103
|
-
The base URL for a given inference API. Only relevant if `model` refers to a
|
|
104
|
-
model on an inference API.
|
|
105
|
-
api_version:
|
|
106
|
-
The version of the API to use for a given inference API.
|
|
107
|
-
gpu_memory_utilization:
|
|
108
|
-
The GPU memory utilization to use for vLLM. A larger value will result in
|
|
109
|
-
faster evaluation, but at the risk of running out of GPU memory. Only reduce
|
|
110
|
-
this if you are running out of GPU memory. Only relevant if the model is
|
|
111
|
-
generative.
|
|
112
|
-
generative_type:
|
|
113
|
-
The type of generative model. Only relevant if the model is generative. If
|
|
114
|
-
not specified, the type will be inferred automatically.
|
|
115
|
-
debug:
|
|
116
|
-
Whether to run the benchmark in debug mode.
|
|
117
|
-
run_with_cli:
|
|
118
|
-
Whether the benchmark is being run with the CLI.
|
|
119
|
-
requires_safetensors:
|
|
120
|
-
Whether to only allow evaluations of models stored as safetensors.
|
|
121
|
-
download_only:
|
|
122
|
-
Whether to only download the requested model weights and datasets.
|
|
29
|
+
benchmark_config_params:
|
|
30
|
+
The parameters for creating the benchmark configuration.
|
|
123
31
|
|
|
124
32
|
Returns:
|
|
125
33
|
The benchmark configuration.
|
|
126
34
|
"""
|
|
127
|
-
language_codes = get_correct_language_codes(
|
|
35
|
+
language_codes = get_correct_language_codes(
|
|
36
|
+
language_codes=benchmark_config_params.language
|
|
37
|
+
)
|
|
128
38
|
model_languages = prepare_languages(
|
|
129
|
-
language_codes=model_language,
|
|
39
|
+
language_codes=benchmark_config_params.model_language,
|
|
40
|
+
default_language_codes=language_codes,
|
|
130
41
|
)
|
|
131
42
|
dataset_languages = prepare_languages(
|
|
132
|
-
language_codes=dataset_language,
|
|
43
|
+
language_codes=benchmark_config_params.dataset_language,
|
|
44
|
+
default_language_codes=language_codes,
|
|
133
45
|
)
|
|
134
46
|
|
|
135
47
|
tasks, datasets = prepare_tasks_and_datasets(
|
|
136
|
-
task=task,
|
|
48
|
+
task=benchmark_config_params.task,
|
|
49
|
+
dataset=benchmark_config_params.dataset,
|
|
50
|
+
dataset_languages=dataset_languages,
|
|
137
51
|
)
|
|
138
52
|
|
|
139
|
-
torch_device = prepare_device(device=device)
|
|
140
|
-
|
|
141
|
-
# Set variable with number of iterations
|
|
142
|
-
if hasattr(sys, "_called_from_test"):
|
|
143
|
-
num_iterations = 1
|
|
144
|
-
|
|
145
53
|
return BenchmarkConfig(
|
|
146
54
|
model_languages=model_languages,
|
|
147
55
|
dataset_languages=dataset_languages,
|
|
148
56
|
tasks=tasks,
|
|
149
57
|
datasets=datasets,
|
|
150
|
-
batch_size=batch_size,
|
|
151
|
-
raise_errors=raise_errors,
|
|
152
|
-
cache_dir=cache_dir,
|
|
153
|
-
api_key=api_key,
|
|
154
|
-
force=force,
|
|
155
|
-
progress_bar=progress_bar,
|
|
156
|
-
save_results=save_results,
|
|
157
|
-
verbose=verbose or debug,
|
|
158
|
-
device=
|
|
159
|
-
trust_remote_code=trust_remote_code,
|
|
160
|
-
clear_model_cache=clear_model_cache,
|
|
161
|
-
evaluate_test_split=evaluate_test_split,
|
|
162
|
-
few_shot=few_shot,
|
|
163
|
-
num_iterations=
|
|
164
|
-
|
|
165
|
-
|
|
166
|
-
|
|
167
|
-
|
|
168
|
-
|
|
169
|
-
|
|
170
|
-
|
|
171
|
-
|
|
58
|
+
batch_size=benchmark_config_params.batch_size,
|
|
59
|
+
raise_errors=benchmark_config_params.raise_errors,
|
|
60
|
+
cache_dir=benchmark_config_params.cache_dir,
|
|
61
|
+
api_key=benchmark_config_params.api_key,
|
|
62
|
+
force=benchmark_config_params.force,
|
|
63
|
+
progress_bar=benchmark_config_params.progress_bar,
|
|
64
|
+
save_results=benchmark_config_params.save_results,
|
|
65
|
+
verbose=benchmark_config_params.verbose or benchmark_config_params.debug,
|
|
66
|
+
device=prepare_device(device=benchmark_config_params.device),
|
|
67
|
+
trust_remote_code=benchmark_config_params.trust_remote_code,
|
|
68
|
+
clear_model_cache=benchmark_config_params.clear_model_cache,
|
|
69
|
+
evaluate_test_split=benchmark_config_params.evaluate_test_split,
|
|
70
|
+
few_shot=benchmark_config_params.few_shot,
|
|
71
|
+
num_iterations=(
|
|
72
|
+
1
|
|
73
|
+
if hasattr(sys, "_called_from_test")
|
|
74
|
+
else benchmark_config_params.num_iterations
|
|
75
|
+
),
|
|
76
|
+
api_base=benchmark_config_params.api_base,
|
|
77
|
+
api_version=benchmark_config_params.api_version,
|
|
78
|
+
gpu_memory_utilization=benchmark_config_params.gpu_memory_utilization,
|
|
79
|
+
generative_type=benchmark_config_params.generative_type,
|
|
80
|
+
debug=benchmark_config_params.debug,
|
|
81
|
+
run_with_cli=benchmark_config_params.run_with_cli,
|
|
82
|
+
requires_safetensors=benchmark_config_params.requires_safetensors,
|
|
83
|
+
download_only=benchmark_config_params.download_only,
|
|
172
84
|
)
|
|
173
85
|
|
|
174
86
|
|
|
@@ -836,15 +836,18 @@ def load_model_and_tokeniser(
|
|
|
836
836
|
|
|
837
837
|
clear_vllm()
|
|
838
838
|
|
|
839
|
-
# if we do not have an internet connection we need to give the path to the folder
|
|
840
|
-
# that contains the model weights and config files, otherwise vLLM will try to
|
|
841
|
-
# download them regardless if they are already present in the download_dir
|
|
842
|
-
model_path = resolve_model_path(download_dir)
|
|
843
|
-
|
|
844
839
|
try:
|
|
845
840
|
model = LLM(
|
|
846
|
-
model=
|
|
847
|
-
|
|
841
|
+
model=(
|
|
842
|
+
model_id
|
|
843
|
+
if internet_connection_available()
|
|
844
|
+
else resolve_model_path(download_dir=download_dir)
|
|
845
|
+
),
|
|
846
|
+
tokenizer=(
|
|
847
|
+
model_id
|
|
848
|
+
if internet_connection_available()
|
|
849
|
+
else resolve_model_path(download_dir=download_dir)
|
|
850
|
+
),
|
|
848
851
|
gpu_memory_utilization=benchmark_config.gpu_memory_utilization,
|
|
849
852
|
max_model_len=min(true_max_model_len, MAX_CONTEXT_LENGTH),
|
|
850
853
|
download_dir=download_dir,
|