EuroEval 15.12.0__py3-none-any.whl → 16.7.1__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (87) hide show
  1. euroeval/__init__.py +32 -14
  2. euroeval/benchmark_config_factory.py +92 -180
  3. euroeval/benchmark_modules/base.py +49 -39
  4. euroeval/benchmark_modules/fresh.py +35 -21
  5. euroeval/benchmark_modules/hf.py +280 -244
  6. euroeval/benchmark_modules/litellm.py +752 -312
  7. euroeval/benchmark_modules/vllm.py +570 -268
  8. euroeval/benchmarker.py +651 -528
  9. euroeval/caching_utils.py +79 -0
  10. euroeval/callbacks.py +5 -7
  11. euroeval/cli.py +49 -38
  12. euroeval/constants.py +44 -25
  13. euroeval/data_loading.py +111 -55
  14. euroeval/data_models.py +490 -323
  15. euroeval/dataset_configs/__init__.py +26 -4
  16. euroeval/dataset_configs/bosnian.py +39 -0
  17. euroeval/dataset_configs/bulgarian.py +56 -0
  18. euroeval/dataset_configs/croatian.py +56 -0
  19. euroeval/dataset_configs/czech.py +75 -0
  20. euroeval/dataset_configs/danish.py +78 -50
  21. euroeval/dataset_configs/dutch.py +74 -44
  22. euroeval/dataset_configs/english.py +71 -36
  23. euroeval/dataset_configs/estonian.py +111 -0
  24. euroeval/dataset_configs/faroese.py +25 -18
  25. euroeval/dataset_configs/finnish.py +63 -26
  26. euroeval/dataset_configs/french.py +65 -32
  27. euroeval/dataset_configs/german.py +77 -36
  28. euroeval/dataset_configs/greek.py +64 -0
  29. euroeval/dataset_configs/icelandic.py +68 -57
  30. euroeval/dataset_configs/italian.py +68 -36
  31. euroeval/dataset_configs/latvian.py +87 -0
  32. euroeval/dataset_configs/lithuanian.py +64 -0
  33. euroeval/dataset_configs/norwegian.py +98 -72
  34. euroeval/dataset_configs/polish.py +96 -0
  35. euroeval/dataset_configs/portuguese.py +63 -40
  36. euroeval/dataset_configs/serbian.py +64 -0
  37. euroeval/dataset_configs/slovak.py +55 -0
  38. euroeval/dataset_configs/slovene.py +56 -0
  39. euroeval/dataset_configs/spanish.py +68 -34
  40. euroeval/dataset_configs/swedish.py +82 -41
  41. euroeval/dataset_configs/ukrainian.py +64 -0
  42. euroeval/enums.py +12 -6
  43. euroeval/exceptions.py +21 -1
  44. euroeval/finetuning.py +34 -26
  45. euroeval/generation.py +76 -41
  46. euroeval/generation_utils.py +169 -34
  47. euroeval/languages.py +1020 -188
  48. euroeval/logging_utils.py +268 -0
  49. euroeval/metrics/__init__.py +6 -0
  50. euroeval/metrics/base.py +85 -0
  51. euroeval/metrics/huggingface.py +216 -0
  52. euroeval/metrics/llm_as_a_judge.py +260 -0
  53. euroeval/metrics/pipeline.py +289 -0
  54. euroeval/metrics/speed.py +48 -0
  55. euroeval/model_cache.py +40 -21
  56. euroeval/model_config.py +4 -5
  57. euroeval/model_loading.py +3 -0
  58. euroeval/prompt_templates/__init__.py +2 -0
  59. euroeval/prompt_templates/classification.py +206 -0
  60. euroeval/prompt_templates/linguistic_acceptability.py +157 -22
  61. euroeval/prompt_templates/multiple_choice.py +159 -17
  62. euroeval/prompt_templates/named_entity_recognition.py +318 -21
  63. euroeval/prompt_templates/reading_comprehension.py +207 -16
  64. euroeval/prompt_templates/sentiment_classification.py +205 -22
  65. euroeval/prompt_templates/summarization.py +122 -22
  66. euroeval/prompt_templates/token_classification.py +279 -0
  67. euroeval/scores.py +20 -9
  68. euroeval/speed_benchmark.py +11 -12
  69. euroeval/task_group_utils/multiple_choice_classification.py +21 -12
  70. euroeval/task_group_utils/question_answering.py +101 -73
  71. euroeval/task_group_utils/sequence_classification.py +144 -61
  72. euroeval/task_group_utils/text_to_text.py +33 -12
  73. euroeval/task_group_utils/token_classification.py +86 -89
  74. euroeval/tasks.py +75 -16
  75. euroeval/tokenisation_utils.py +603 -0
  76. euroeval/types.py +17 -11
  77. euroeval/utils.py +332 -137
  78. euroeval-16.7.1.dist-info/METADATA +623 -0
  79. euroeval-16.7.1.dist-info/RECORD +84 -0
  80. {euroeval-15.12.0.dist-info → euroeval-16.7.1.dist-info}/entry_points.txt +0 -1
  81. euroeval/human_evaluation.py +0 -737
  82. euroeval/metrics.py +0 -452
  83. euroeval/tokenization_utils.py +0 -498
  84. euroeval-15.12.0.dist-info/METADATA +0 -285
  85. euroeval-15.12.0.dist-info/RECORD +0 -63
  86. {euroeval-15.12.0.dist-info → euroeval-16.7.1.dist-info}/WHEEL +0 -0
  87. {euroeval-15.12.0.dist-info → euroeval-16.7.1.dist-info}/licenses/LICENSE +0 -0
@@ -1,285 +0,0 @@
1
- Metadata-Version: 2.4
2
- Name: EuroEval
3
- Version: 15.12.0
4
- Summary: The robust European language model benchmark.
5
- Project-URL: Repository, https://github.com/EuroEval/EuroEval
6
- Project-URL: Issues, https://github.com/EuroEval/EuroEval/issues
7
- Author-email: Dan Saattrup Smart <dan.smart@alexandra.dk>
8
- Maintainer-email: Dan Saattrup Smart <dan.smart@alexandra.dk>
9
- License: MIT License
10
-
11
- Copyright (c) 2022-2025 Dan Saattrup Smart
12
-
13
- Permission is hereby granted, free of charge, to any person obtaining a copy
14
- of this software and associated documentation files (the "Software"), to deal
15
- in the Software without restriction, including without limitation the rights
16
- to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
17
- copies of the Software, and to permit persons to whom the Software is
18
- furnished to do so, subject to the following conditions:
19
-
20
- The above copyright notice and this permission notice shall be included in all
21
- copies or substantial portions of the Software.
22
-
23
- THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
24
- IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
25
- FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
26
- AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
27
- LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
28
- OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
29
- SOFTWARE.
30
- License-File: LICENSE
31
- Requires-Python: <4.0,>=3.10
32
- Requires-Dist: accelerate>=0.34.2
33
- Requires-Dist: bert-score>=0.3.13
34
- Requires-Dist: click>=8.1.3
35
- Requires-Dist: datasets>=3.5.0
36
- Requires-Dist: demjson3>=3.0.6
37
- Requires-Dist: evaluate>=0.4.1
38
- Requires-Dist: huggingface-hub>=0.30.1
39
- Requires-Dist: levenshtein>=0.24.0
40
- Requires-Dist: litellm>=1.72.2
41
- Requires-Dist: more-itertools>=10.5.0
42
- Requires-Dist: numpy<2.0.0,>=1.23.0
43
- Requires-Dist: ollama>=0.5.1
44
- Requires-Dist: pandas>=2.2.0
45
- Requires-Dist: peft>=0.15.0
46
- Requires-Dist: protobuf>=2.0.0
47
- Requires-Dist: pydantic>=2.6.0
48
- Requires-Dist: pyinfer>=0.0.3
49
- Requires-Dist: python-dotenv>=1.0.1
50
- Requires-Dist: rouge-score>=0.1.2
51
- Requires-Dist: sacremoses>=0.1.1
52
- Requires-Dist: scikit-learn<1.6.0
53
- Requires-Dist: sentencepiece>=0.1.96
54
- Requires-Dist: seqeval>=1.2.2
55
- Requires-Dist: setuptools>=75.8.2
56
- Requires-Dist: tenacity>=9.0.0
57
- Requires-Dist: termcolor>=2.0.0
58
- Requires-Dist: torch>=2.6.0
59
- Requires-Dist: transformers>=4.51.0
60
- Provides-Extra: all
61
- Requires-Dist: bitsandbytes>=0.43.1; (platform_system == 'Linux') and extra == 'all'
62
- Requires-Dist: fbgemm-gpu>=1.0.0; (platform_system == 'Linux') and extra == 'all'
63
- Requires-Dist: gradio>=4.26.0; extra == 'all'
64
- Requires-Dist: outlines>=0.1.11; extra == 'all'
65
- Requires-Dist: vllm>=0.9.1; (platform_system == 'Linux') and extra == 'all'
66
- Provides-Extra: generative
67
- Requires-Dist: bitsandbytes>=0.43.1; (platform_system == 'Linux') and extra == 'generative'
68
- Requires-Dist: fbgemm-gpu>=1.0.0; (platform_system == 'Linux') and extra == 'generative'
69
- Requires-Dist: outlines>=0.1.11; extra == 'generative'
70
- Requires-Dist: vllm>=0.9.1; (platform_system == 'Linux') and extra == 'generative'
71
- Provides-Extra: human-evaluation
72
- Requires-Dist: gradio>=4.26.0; extra == 'human-evaluation'
73
- Provides-Extra: test
74
- Requires-Dist: gradio>=4.26.0; extra == 'test'
75
- Description-Content-Type: text/markdown
76
-
77
- <div align='center'>
78
- <img src="https://raw.githubusercontent.com/EuroEval/EuroEval/main/gfx/euroeval.png" height="500" width="372">
79
- </div>
80
-
81
- ### The robust European language model benchmark.
82
-
83
- _(formerly known as ScandEval)_
84
-
85
- ______________________________________________________________________
86
- [![Documentation](https://img.shields.io/badge/docs-passing-green)](https://euroeval.com)
87
- [![PyPI Status](https://badge.fury.io/py/euroeval.svg)](https://pypi.org/project/euroeval/)
88
- [![First paper](https://img.shields.io/badge/arXiv-2304.00906-b31b1b.svg)](https://arxiv.org/abs/2304.00906)
89
- [![Second paper](https://img.shields.io/badge/arXiv-2406.13469-b31b1b.svg)](https://arxiv.org/abs/2406.13469)
90
- [![License](https://img.shields.io/github/license/EuroEval/EuroEval)](https://github.com/EuroEval/EuroEval/blob/main/LICENSE)
91
- [![LastCommit](https://img.shields.io/github/last-commit/EuroEval/EuroEval)](https://github.com/EuroEval/EuroEval/commits/main)
92
- [![Code Coverage](https://img.shields.io/badge/Coverage-67%25-yellow.svg)](https://github.com/EuroEval/EuroEval/tree/main/tests)
93
- [![Contributor Covenant](https://img.shields.io/badge/Contributor%20Covenant-2.0-4baaaa.svg)](https://github.com/EuroEval/EuroEval/blob/main/CODE_OF_CONDUCT.md)
94
-
95
-
96
- ## Maintainer
97
-
98
- - Dan Saattrup Smart ([@saattrupdan](https://github.com/saattrupdan), dan.smart@alexandra.dk)
99
-
100
-
101
- ## Installation
102
- To install the package simply write the following command in your favorite terminal:
103
- ```
104
- $ pip install euroeval[all]
105
- ```
106
-
107
- This will install the EuroEval package with all extras. You can also install the
108
- minimal version by leaving out the `[all]`, in which case the package will let you know
109
- when an evaluation requires a certain extra dependency, and how you install it.
110
-
111
- ## Quickstart
112
- ### Benchmarking from the Command Line
113
- The easiest way to benchmark pretrained models is via the command line interface. After
114
- having installed the package, you can benchmark your favorite model like so:
115
- ```
116
- $ euroeval --model <model-id>
117
- ```
118
-
119
- Here `model` is the HuggingFace model ID, which can be found on the [HuggingFace
120
- Hub](https://huggingface.co/models). By default this will benchmark the model on all
121
- the tasks available. If you want to benchmark on a particular task, then use the
122
- `--task` argument:
123
- ```
124
- $ euroeval --model <model-id> --task sentiment-classification
125
- ```
126
-
127
- We can also narrow down which languages we would like to benchmark on. This can be done
128
- by setting the `--language` argument. Here we thus benchmark the model on the Danish
129
- sentiment classification task:
130
- ```
131
- $ euroeval --model <model-id> --task sentiment-classification --language da
132
- ```
133
-
134
- Multiple models, datasets and/or languages can be specified by just attaching multiple
135
- arguments. Here is an example with two models:
136
- ```
137
- $ euroeval --model <model-id1> --model <model-id2>
138
- ```
139
-
140
- The specific model version/revision to use can also be added after the suffix '@':
141
- ```
142
- $ euroeval --model <model-id>@<commit>
143
- ```
144
-
145
- This can be a branch name, a tag name, or a commit id. It defaults to 'main' for latest.
146
-
147
- See all the arguments and options available for the `euroeval` command by typing
148
- ```
149
- $ euroeval --help
150
- ```
151
-
152
- ### Benchmarking from a Script
153
- In a script, the syntax is similar to the command line interface. You simply initialise
154
- an object of the `Benchmarker` class, and call this benchmark object with your favorite
155
- model:
156
- ```
157
- >>> from euroeval import Benchmarker
158
- >>> benchmark = Benchmarker()
159
- >>> benchmark(model="<model>")
160
- ```
161
-
162
- To benchmark on a specific task and/or language, you simply specify the `task` or
163
- `language` arguments, shown here with same example as above:
164
- ```
165
- >>> benchmark(model="<model>", task="sentiment-classification", language="da")
166
- ```
167
-
168
- If you want to benchmark a subset of all the models on the Hugging Face Hub, you can
169
- simply leave out the `model` argument. In this example, we're benchmarking all Danish
170
- models on the Danish sentiment classification task:
171
- ```
172
- >>> benchmark(task="sentiment-classification", language="da")
173
- ```
174
-
175
- ### Benchmarking from Docker
176
- A Dockerfile is provided in the repo, which can be downloaded and run, without needing
177
- to clone the repo and installing from source. This can be fetched programmatically by
178
- running the following:
179
- ```
180
- $ wget https://raw.githubusercontent.com/EuroEval/EuroEval/main/Dockerfile.cuda
181
- ```
182
-
183
- Next, to be able to build the Docker image, first ensure that the NVIDIA Container
184
- Toolkit is
185
- [installed](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html#installation)
186
- and
187
- [configured](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html#configuring-docker).
188
- Ensure that the the CUDA version stated at the top of the Dockerfile matches the CUDA
189
- version installed (which you can check using `nvidia-smi`). After that, we build the
190
- image as follows:
191
- ```
192
- $ docker build --pull -t euroeval -f Dockerfile.cuda .
193
- ```
194
-
195
- With the Docker image built, we can now evaluate any model as follows:
196
- ```
197
- $ docker run -e args="<euroeval-arguments>" --gpus 1 --name euroeval --rm euroeval
198
- ```
199
-
200
- Here `<euroeval-arguments>` consists of the arguments added to the `euroeval` CLI
201
- argument. This could for instance be `--model <model-id> --task
202
- sentiment-classification`.
203
-
204
-
205
- ### Reproducing the datasets
206
- All datasets used in this project are generated using the scripts located in the
207
- [src/scripts](src/scripts) folder. To reproduce a dataset, run the corresponding script
208
- with the following command
209
-
210
- ```shell
211
- $ uv run src/scripts/<name-of-script>.py
212
- ```
213
-
214
- Replace <name-of-script> with the specific script you wish to execute, e.g.,
215
-
216
- ```shell
217
- $ uv run src/scripts/create_allocine.py
218
- ```
219
-
220
- ## Contributors :pray:
221
-
222
- A huge thank you to all the contributors who have helped make this project a success!
223
-
224
- <a href="https://github.com/peter-sk"><img src="https://avatars.githubusercontent.com/u/6168908" width=50 alt="Contributor avatar for peter-sk"/></a>
225
- <a href="https://github.com/AJDERS"><img src="https://avatars.githubusercontent.com/u/38854604" width=50 alt="Contributor avatar for AJDERS"/></a>
226
- <a href="https://github.com/oliverkinch"><img src="https://avatars.githubusercontent.com/u/71556498" width=50 alt="Contributor avatar for oliverkinch"/></a>
227
- <a href="https://github.com/versae"><img src="https://avatars.githubusercontent.com/u/173537" width=50 alt="Contributor avatar for versae"/></a>
228
- <a href="https://github.com/viggo-gascou"><img src="https://avatars.githubusercontent.com/u/94069687" width=50 alt="Contributor avatar for viggo-gascou"/></a>
229
- <a href="https://github.com/mathiasesn"><img src="https://avatars.githubusercontent.com/u/27091759" width=50 alt="Contributor avatar for mathiasesn"/></a>
230
- <a href="https://github.com/Alkarex"><img src="https://avatars.githubusercontent.com/u/1008324" width=50 alt="Contributor avatar for Alkarex"/></a>
231
- <a href="https://github.com/marksverdhei"><img src="https://avatars.githubusercontent.com/u/46672778" width=50 alt="Contributor avatar for marksverdhei"/></a>
232
- <a href="https://github.com/Mikeriess"><img src="https://avatars.githubusercontent.com/u/19728563" width=50 alt="Contributor avatar for Mikeriess"/></a>
233
- <a href="https://github.com/pakagronglb"><img src="https://avatars.githubusercontent.com/u/178713124" width=50 alt="Contributor avatar for pakagronglb"/></a>
234
- <a href="https://github.com/ThomasKluiters"><img src="https://avatars.githubusercontent.com/u/8137941" width=50 alt="Contributor avatar for ThomasKluiters"/></a>
235
- <a href="https://github.com/BramVanroy"><img src="https://avatars.githubusercontent.com/u/2779410" width=50 alt="Contributor avatar for BramVanroy"/></a>
236
- <a href="https://github.com/peregilk"><img src="https://avatars.githubusercontent.com/u/9079808" width=50 alt="Contributor avatar for peregilk"/></a>
237
- <a href="https://github.com/Rijgersberg"><img src="https://avatars.githubusercontent.com/u/8604946" width=50 alt="Contributor avatar for Rijgersberg"/></a>
238
-
239
-
240
- ### Contribute to EuroEval
241
-
242
- We welcome contributions to EuroEval! Whether you're fixing bugs, adding features, or
243
- contributing new datasets, your help makes this project better for everyone.
244
-
245
- - **General contributions**: Check out our [contribution guidelines](CONTRIBUTING.md)
246
- for information on how to get started.
247
- - **Adding datasets**: If you're interested in adding a new dataset to EuroEval, we have
248
- a [dedicated guide](NEW_DATASET_GUIDE.md) with step-by-step instructions.
249
-
250
-
251
- ### Special Thanks
252
- - Thanks to [Google](https://google.com/) for sponsoring Gemini credits as part of their
253
- [Google Cloud for Researchers Program](https://cloud.google.com/edu/researchers).
254
- - Thanks [@Mikeriess](https://github.com/Mikeriess) for evaluating many of the larger
255
- models on the leaderboards.
256
- - Thanks to [OpenAI](https://openai.com/) for sponsoring OpenAI credits as part of their
257
- [Researcher Access Program](https://openai.com/form/researcher-access-program/).
258
- - Thanks to [UWV](https://www.uwv.nl/) and [KU
259
- Leuven](https://www.arts.kuleuven.be/ling/ccl) for sponsoring the Azure OpenAI
260
- credits used to evaluate GPT-4-turbo in Dutch.
261
- - Thanks to [Miðeind](https://mideind.is/english.html) for sponsoring the OpenAI
262
- credits used to evaluate GPT-4-turbo in Icelandic and Faroese.
263
- - Thanks to [CHC](https://chc.au.dk/) for sponsoring the OpenAI credits used to
264
- evaluate GPT-4-turbo in German.
265
-
266
-
267
- ## Citing EuroEval
268
- If you want to cite the framework then feel free to use this:
269
-
270
- ```
271
- @article{smart2024encoder,
272
- title={Encoder vs Decoder: Comparative Analysis of Encoder and Decoder Language Models on Multilingual NLU Tasks},
273
- author={Smart, Dan Saattrup and Enevoldsen, Kenneth and Schneider-Kamp, Peter},
274
- journal={arXiv preprint arXiv:2406.13469},
275
- year={2024}
276
- }
277
- @inproceedings{smart2023scandeval,
278
- author = {Smart, Dan Saattrup},
279
- booktitle = {Proceedings of the 24th Nordic Conference on Computational Linguistics (NoDaLiDa)},
280
- month = may,
281
- pages = {185--201},
282
- title = {{ScandEval: A Benchmark for Scandinavian Natural Language Processing}},
283
- year = {2023}
284
- }
285
- ```
@@ -1,63 +0,0 @@
1
- euroeval/__init__.py,sha256=fZyR9R3C3vwGJS3CrCJ6ySr_FDnMu_Aqnz0FdadWEEs,3399
2
- euroeval/benchmark_config_factory.py,sha256=jKC8bEzJSGGCcG8aWsPxiyHX6fjOQYQWvkp1MIUuHYM,11564
3
- euroeval/benchmarker.py,sha256=SDBzdCa4I8u1XDeN_1mKTFzfaaQbbY_oWcHt3niADxk,48497
4
- euroeval/callbacks.py,sha256=5BTlDvBJ60xRvj01EpXZSZu3MFdKa3LgVuhxoLb3i3E,2565
5
- euroeval/cli.py,sha256=h81Lswm_q9htkYz-GQQQVIsdsUPnfe3LDH8AZdBcpKs,8602
6
- euroeval/constants.py,sha256=0KHrH74zGM8vNF4uZG_a5qFJRZH5YgyQULYZtCKlo68,2452
7
- euroeval/data_loading.py,sha256=DP-cqwN_d0Y-KaN8P8c3fDr6PX80UYROHgRwX82ix4w,4156
8
- euroeval/data_models.py,sha256=gPHyIoN2A5_O-cJgyb6jhn6enH8zsiIBI09W_wdHMQs,22031
9
- euroeval/enums.py,sha256=L9LcNeruuhHvze9vKRogXY9vonRzoBqDzWSP6hxKQ7A,3195
10
- euroeval/exceptions.py,sha256=5kQ-YvHyFO3aaA-zfOTaS07LRFH8xlSqlOiATvnIObY,5116
11
- euroeval/finetuning.py,sha256=BrPZ-6qFY8K-dwfaRwNetVYfYburoQwLQty6pn6iP_s,11340
12
- euroeval/generation.py,sha256=1fqFEWwM2RzI3uPZem95VFWbN8EfrKZQTrHEP34ihHs,11622
13
- euroeval/generation_utils.py,sha256=zRsaOHcbhysbMa983BZXxfd-qMe4NYts-ZbQxfvNTK4,13310
14
- euroeval/human_evaluation.py,sha256=Jtz3K5Lqne48wPZWf4EAd3d-n_wX27nGJHigjhV1D7s,27537
15
- euroeval/languages.py,sha256=cr_Z5jtaHb2XY0zeOhuk3ATHX74PODzt6gMPC2zMD7c,8594
16
- euroeval/metrics.py,sha256=nxosyoRjlk7TcoAOkjU7zx2TB43b9tA8M1m4V1s5eKU,15516
17
- euroeval/model_cache.py,sha256=HgXTgn4RMBqIjKaTmYzxu0f4NIwbXx1XJFbvbITqy4E,8686
18
- euroeval/model_config.py,sha256=64KKHPTrpsFhFAANtBnAKkOs7PWZ50GXkXeDl4jICgs,2748
19
- euroeval/model_loading.py,sha256=B6dyjYO0Dg7NOcUXls8Sjwe6W0c2UqJ1OGw-RkzoSSQ,2239
20
- euroeval/scores.py,sha256=TatSbjia7Zwj71gQFyV_gCHyppMbOgeaZgNCib8G86k,2849
21
- euroeval/speed_benchmark.py,sha256=6bFGeMmtdl_6owkxNQ3ZKiyQQS58k0NApzlsbDgBW5s,4037
22
- euroeval/tasks.py,sha256=btxf29M5rUP7JjBl6u9aQlHQAxrJNP4bRbdEQtDnmDA,3376
23
- euroeval/tokenization_utils.py,sha256=LxgGs7juS5PuMYt5LL2X6eVXdtnpi-A2jFxqcWpF6NA,17931
24
- euroeval/types.py,sha256=EIYMNOqqHqibnbNw-fvdst6HwTvq32gtxhr7jL7i-xM,2511
25
- euroeval/utils.py,sha256=5R7y67xe0ODaje7k8nOu2AFS3Ph2gcsiWpIq5rjSSuA,11613
26
- euroeval/benchmark_modules/__init__.py,sha256=TNO-sNDwlXE-LMFXfwwqjQqUy55gywSmwRBcoPUFuaU,236
27
- euroeval/benchmark_modules/base.py,sha256=D1oKD16KBvxEoBUfqwvzvcDc1hx6letdD3v1PnBmF4A,10669
28
- euroeval/benchmark_modules/fresh.py,sha256=sg_AXNPApFObCzCRWhCgKxfr-eqQsT6Ri0xx0_Yy5JM,10293
29
- euroeval/benchmark_modules/hf.py,sha256=-W_bWEdm0zePkn4nDz4l0T4hhJJnlfwHrtIO3m5BrUs,44725
30
- euroeval/benchmark_modules/litellm.py,sha256=_gKBbJsXzo_cHJVaeuQpHRBENEZUGS_vcC-uGIhhmHA,52111
31
- euroeval/benchmark_modules/vllm.py,sha256=kq3PMUuRT0NOky6XSHl1JeHTDGehwcub0HcGC5S_Wv4,38834
32
- euroeval/dataset_configs/__init__.py,sha256=EbjEyHwBtSztASl8_xblD8hessruDdV4Eg1vXrmGOuY,1935
33
- euroeval/dataset_configs/danish.py,sha256=MTt9EcriSer0QaFQ7_6evYxh-g9OPjroWegYdFpiKag,3395
34
- euroeval/dataset_configs/dutch.py,sha256=r21nxEvMmBkKqPXVW082batPsxJ9d0RB4DzngOTMJSk,3185
35
- euroeval/dataset_configs/english.py,sha256=1q8XJqIVWBBNkldL7t-cVnU2O9EUb9_xoVRSN8arN90,2561
36
- euroeval/dataset_configs/faroese.py,sha256=QQgLe5gv0f3AtXe5rV65xZ98gFgyITQPDr3UwO4Bnv4,1350
37
- euroeval/dataset_configs/finnish.py,sha256=_8YWIlZNpO8Qi233bH7cKwm3tq3WETLfC_6mzg7LLog,2045
38
- euroeval/dataset_configs/french.py,sha256=ATsj8_9_GxFTQgmfrniPQFZ1R9hoQCI1_ieWTnscFHU,2382
39
- euroeval/dataset_configs/german.py,sha256=QO6PrBQY6kyZeQMU1vg6KrC_sKyj9U2ukS9nbKO19is,2560
40
- euroeval/dataset_configs/icelandic.py,sha256=mncl7X4yO9gBmYqXMBfm7FKU1jcKryerSgd0dqlIA_4,4198
41
- euroeval/dataset_configs/italian.py,sha256=KNjCvTzsEqH_EEk3At8slKqNwWWiIdbv_t5ke7n9nZI,2660
42
- euroeval/dataset_configs/norwegian.py,sha256=30YGdDPtDszG10BNDVHb-XXTGgGIIgDUNGoeM9q0K_E,5385
43
- euroeval/dataset_configs/portuguese.py,sha256=-HSDsujWfK__nV2SCu-z0ne0AXLDszOT05oYphQUDTw,2063
44
- euroeval/dataset_configs/spanish.py,sha256=Yzm1kiilEKoHyd3xD2wrw596Ac9UcaWhlE93GlOFjlc,2558
45
- euroeval/dataset_configs/swedish.py,sha256=SOD2nKQTVwTpTvr362mDPHon42kr9vWs5C0mK02Fh-o,2811
46
- euroeval/prompt_templates/__init__.py,sha256=HWMZpybxs2xHPnVeJ43893conARahIVLWNXeRhXEGZw,357
47
- euroeval/prompt_templates/linguistic_acceptability.py,sha256=ZN71BEt4HAhSYY-GWjh-S-iVvq5AODQJThkrjDhy4oM,7138
48
- euroeval/prompt_templates/multiple_choice.py,sha256=F9ItGQtnaaez15A8MQ1UCpKRDsLM-AZyRdYetGAofa0,5494
49
- euroeval/prompt_templates/named_entity_recognition.py,sha256=ga21s9T4_Hhbf88boWm7gnL7OgD7txuS_EeDgXaxEoE,13602
50
- euroeval/prompt_templates/reading_comprehension.py,sha256=yLqryWQAW04GULz_EyNDLOS7ZrDUeasuLFt-dtqCnYk,6585
51
- euroeval/prompt_templates/sentiment_classification.py,sha256=2Xsmj8lbaAXACHhwbbR4dWhoKyKB87TqpMO-ssQ-Djo,7649
52
- euroeval/prompt_templates/summarization.py,sha256=I98LlUOBVa_xo02npq7BWKKZOXGqm-_15i64QzbEsb0,5334
53
- euroeval/task_group_utils/__init__.py,sha256=CorGVkixkoEDOQuDsrOGlTmF1zmM0wnGHs8psWTfD28,72
54
- euroeval/task_group_utils/multiple_choice_classification.py,sha256=yfy8lczpZ_MY-Y4FQx3Et9vEUpuD3YMFjF3wQGCfMNw,6632
55
- euroeval/task_group_utils/question_answering.py,sha256=agwtWOmctgat98yqgFiMSPY6zmoaPgYVyzMmOkNjr58,27284
56
- euroeval/task_group_utils/sequence_classification.py,sha256=igmD24aMNN7QBJ8NDzgEnGwM-jq_zhC37QxazNm7GZ4,12711
57
- euroeval/task_group_utils/text_to_text.py,sha256=xOpja-W4E-1peMjZX8G-3G5iRgmFHHygrQ5WN1hB3FI,4550
58
- euroeval/task_group_utils/token_classification.py,sha256=wCy3aI-Sn9f-87tHzAnYDA6EbY3ah3xao1SnfnoRNz4,17490
59
- euroeval-15.12.0.dist-info/METADATA,sha256=8cY6HWgAZgrCkIA20lVKuf42y-e7U1MZQZSTdF3e7ig,13479
60
- euroeval-15.12.0.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
61
- euroeval-15.12.0.dist-info/entry_points.txt,sha256=tKQRxN0HX2mGtbZbZQdCRFUDZIecA_z4mZduueor3Ug,135
62
- euroeval-15.12.0.dist-info/licenses/LICENSE,sha256=guvz_zBHgkQSY_QiUU0Bkc1k-L_PFZuLjIPfuKne2OY,1080
63
- euroeval-15.12.0.dist-info/RECORD,,