EuroEval 15.6.0__py3-none-any.whl → 15.6.1__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of EuroEval might be problematic. Click here for more details.

euroeval/data_models.py CHANGED
@@ -388,8 +388,10 @@ class DatasetConfig:
388
388
  language.
389
389
  _prompt_label_mapping (optional):
390
390
  A mapping from the labels to another phrase which is used as a substitute
391
- for the label in few-shot evaluation. Defaults to the template for the task
392
- and language.
391
+ for the label in few-shot evaluation. If "auto" then the mapping will be set
392
+ to a 1:1 mapping between the labels and themselves. If None then the mapping
393
+ will be set to the default mapping for the task and language. Defaults to
394
+ None.
393
395
  unofficial (optional):
394
396
  Whether the dataset is unofficial. Defaults to False.
395
397
  """
@@ -405,7 +407,7 @@ class DatasetConfig:
405
407
  _num_few_shot_examples: int | None = None
406
408
  _max_generated_tokens: int | None = None
407
409
  _labels: list[str] | None = None
408
- _prompt_label_mapping: dict[str, str] | None = None
410
+ _prompt_label_mapping: dict[str, str] | t.Literal["auto"] | None = None
409
411
  unofficial: bool = False
410
412
 
411
413
  @property
@@ -475,7 +477,9 @@ class DatasetConfig:
475
477
  @property
476
478
  def prompt_label_mapping(self) -> dict[str, str]:
477
479
  """Mapping from English labels to localised labels."""
478
- if self._prompt_label_mapping is not None:
480
+ if self._prompt_label_mapping == "auto":
481
+ return {label: label for label in self.labels}
482
+ elif self._prompt_label_mapping is not None:
479
483
  return self._prompt_label_mapping
480
484
 
481
485
  main_language = self.languages[0]
@@ -12,6 +12,7 @@ FOSENT_CONFIG = DatasetConfig(
12
12
  huggingface_id="EuroEval/fosent",
13
13
  task=SENT,
14
14
  languages=[FO],
15
+ _num_few_shot_examples=5,
15
16
  )
16
17
 
17
18
  SCALA_FO_CONFIG = DatasetConfig(
@@ -83,6 +83,7 @@ NOR_COMMON_SENSE_QA_CONFIG = DatasetConfig(
83
83
  huggingface_id="EuroEval/nor-common-sense-qa",
84
84
  task=COMMON_SENSE,
85
85
  languages=[NB, NN, NO],
86
+ _labels=["a", "b", "c", "d", "e"],
86
87
  )
87
88
 
88
89
 
@@ -105,7 +106,6 @@ NORGLM_MULTI_QA = DatasetConfig(
105
106
  huggingface_id="EuroEval/norglm-multi-qa",
106
107
  task=RC,
107
108
  languages=[NB, NN, NO],
108
- _num_few_shot_examples=2,
109
109
  unofficial=True,
110
110
  )
111
111
 
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: EuroEval
3
- Version: 15.6.0
3
+ Version: 15.6.1
4
4
  Summary: The robust European language model benchmark.
5
5
  Project-URL: Repository, https://github.com/EuroEval/EuroEval
6
6
  Project-URL: Issues, https://github.com/EuroEval/EuroEval/issues
@@ -237,6 +237,7 @@ A huge thank you to all the contributors who have helped make this project a suc
237
237
  <a href="https://github.com/ThomasKluiters"><img src="https://avatars.githubusercontent.com/u/8137941" width=50 alt="Contributor avatar for ThomasKluiters"/></a>
238
238
  <a href="https://github.com/BramVanroy"><img src="https://avatars.githubusercontent.com/u/2779410" width=50 alt="Contributor avatar for BramVanroy"/></a>
239
239
  <a href="https://github.com/peregilk"><img src="https://avatars.githubusercontent.com/u/9079808" width=50 alt="Contributor avatar for peregilk"/></a>
240
+ <a href="https://github.com/Rijgersberg"><img src="https://avatars.githubusercontent.com/u/8604946" width=50 alt="Contributor avatar for Rijgersberg"/></a>
240
241
 
241
242
  ### Special Thanks
242
243
  - Thanks to [Google](https://google.com/) for sponsoring Gemini credits as part of their
@@ -5,7 +5,7 @@ euroeval/callbacks.py,sha256=F1AJCLB8FJpxqYprwLi_PsH4Bc0x4lyR8UiTG-GlFLY,2452
5
5
  euroeval/cli.py,sha256=EMB6g6kRvxIqlfYLSoMzwLAtEd-fqXipo4A_HTkhjkA,8575
6
6
  euroeval/constants.py,sha256=t2mAT8tE3Dn2lXWHTnaFoaOIaUcdiBjJTASCt7nSdkg,1984
7
7
  euroeval/data_loading.py,sha256=7xXdoFSvEDzpw1FNR8E8YV4c9Vy86hlU5-qLm9RUejE,3318
8
- euroeval/data_models.py,sha256=RjU7REmUMxSMeZfTeUNYb6XRlHMUri7Tk_zwexUOupU,22840
8
+ euroeval/data_models.py,sha256=oZLrGg1dhIIwbgtEzq4U_fu_ZbBsz35mrqsyizuZNPw,23138
9
9
  euroeval/enums.py,sha256=L9LcNeruuhHvze9vKRogXY9vonRzoBqDzWSP6hxKQ7A,3195
10
10
  euroeval/exceptions.py,sha256=LRd7HoudupRp5-AX3L0X4hIAWCa6JVx-LViHPg7u7dg,5821
11
11
  euroeval/finetuning.py,sha256=IieAhgvxjeLHAHBief1Ay-STcCosQmrDHFTRTXFZX0Q,10743
@@ -31,12 +31,12 @@ euroeval/dataset_configs/__init__.py,sha256=fkD1hzW7szJLc1MdK-AY4EBFWBUX5Z8t4f9u
31
31
  euroeval/dataset_configs/danish.py,sha256=MTt9EcriSer0QaFQ7_6evYxh-g9OPjroWegYdFpiKag,3395
32
32
  euroeval/dataset_configs/dutch.py,sha256=N3zL0vGe4OyPgVU_AiYNNfk96jSc_JDtKrVIHbaEYCU,3536
33
33
  euroeval/dataset_configs/english.py,sha256=yHw7D0zSNVbiSBAjR1mWX4V5FSkhqy4y-o-pnyWCLxE,2323
34
- euroeval/dataset_configs/faroese.py,sha256=9QYFtd3GqaFcyQjsmru_yvJuTjParyz8Ra_ekw_3xbA,1320
34
+ euroeval/dataset_configs/faroese.py,sha256=QQgLe5gv0f3AtXe5rV65xZ98gFgyITQPDr3UwO4Bnv4,1350
35
35
  euroeval/dataset_configs/french.py,sha256=ATsj8_9_GxFTQgmfrniPQFZ1R9hoQCI1_ieWTnscFHU,2382
36
36
  euroeval/dataset_configs/german.py,sha256=QO6PrBQY6kyZeQMU1vg6KrC_sKyj9U2ukS9nbKO19is,2560
37
37
  euroeval/dataset_configs/icelandic.py,sha256=mncl7X4yO9gBmYqXMBfm7FKU1jcKryerSgd0dqlIA_4,4198
38
38
  euroeval/dataset_configs/italian.py,sha256=5yYMMBbxkfSDpLgJ9IH_pgkpzEp-74vMMvx-dT8x4WY,2345
39
- euroeval/dataset_configs/norwegian.py,sha256=3kKhri5qWIiFwNSzNFHjNbRpkW1NSK_PUltAGQpxmAY,5172
39
+ euroeval/dataset_configs/norwegian.py,sha256=2SD5681gZFa1Ig-AEpnyStbivan_bq_Pada4qwE7tw0,5181
40
40
  euroeval/dataset_configs/spanish.py,sha256=fc0dHWU7-g_p6kaSGA8nD1vLVQF_yqR2PkixrYyWywc,2212
41
41
  euroeval/dataset_configs/swedish.py,sha256=SOD2nKQTVwTpTvr362mDPHon42kr9vWs5C0mK02Fh-o,2811
42
42
  euroeval/prompt_templates/__init__.py,sha256=HWMZpybxs2xHPnVeJ43893conARahIVLWNXeRhXEGZw,357
@@ -52,8 +52,8 @@ euroeval/task_group_utils/question_answering.py,sha256=kZBABJ_WYNTH4Xgo2jIvfx7iY
52
52
  euroeval/task_group_utils/sequence_classification.py,sha256=gqd0-l5o7vAY5QIpGSkSqwJwez3Y0r5SqOiywfPNW8A,12239
53
53
  euroeval/task_group_utils/text_to_text.py,sha256=QECnGdZ0YLjsbMc6LwXqVi4KMuITdiOjmJUNQtAAOW0,5712
54
54
  euroeval/task_group_utils/token_classification.py,sha256=3idWB81Fcx9UhTuk-gxMfXENrCBmiWBDUWdULXoIhpw,17863
55
- euroeval-15.6.0.dist-info/METADATA,sha256=m1NE2zaj_hbP-3kW-2_oC9Ug-POilMU1fVWQTt-SNIU,13027
56
- euroeval-15.6.0.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
57
- euroeval-15.6.0.dist-info/entry_points.txt,sha256=tKQRxN0HX2mGtbZbZQdCRFUDZIecA_z4mZduueor3Ug,135
58
- euroeval-15.6.0.dist-info/licenses/LICENSE,sha256=oZp5fpOSQ7w-vFui8QNwrBIosrO7cnpArItdbvn52Ao,1082
59
- euroeval-15.6.0.dist-info/RECORD,,
55
+ euroeval-15.6.1.dist-info/METADATA,sha256=4i98IBxn6yWh4ugBW-SnljmDfKEXBSfRGjZyf_dlOUs,13183
56
+ euroeval-15.6.1.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
57
+ euroeval-15.6.1.dist-info/entry_points.txt,sha256=tKQRxN0HX2mGtbZbZQdCRFUDZIecA_z4mZduueor3Ug,135
58
+ euroeval-15.6.1.dist-info/licenses/LICENSE,sha256=oZp5fpOSQ7w-vFui8QNwrBIosrO7cnpArItdbvn52Ao,1082
59
+ euroeval-15.6.1.dist-info/RECORD,,