mteb 2.1.1__py3-none-any.whl → 2.1.3__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -5,7 +5,6 @@ from typing import Any
5
5
  import numpy as np
6
6
  import pandas as pd
7
7
  import pytrec_eval
8
- import torch
9
8
  from packaging.version import Version
10
9
  from sklearn.metrics import auc
11
10
 
@@ -14,14 +13,6 @@ from mteb.types import RelevantDocumentsType, RetrievalEvaluationResult
14
13
  logger = logging.getLogger(__name__)
15
14
 
16
15
 
17
- try:
18
- # speeds up computation if available
19
- torch.set_float32_matmul_precision("high")
20
- logger.info("Setting torch float32 matmul precision to high for a speedup")
21
- except Exception:
22
- pass
23
-
24
-
25
16
  def mrr(
26
17
  qrels: RelevantDocumentsType,
27
18
  results: dict[str, dict[str, float]],
@@ -1634,7 +1634,7 @@ BEIR_NL = Benchmark(
1634
1634
 
1635
1635
  MTEB_NL = Benchmark(
1636
1636
  name="MTEB(nld, v1)",
1637
- display_name="MTEB-NL",
1637
+ display_name="Dutch",
1638
1638
  icon="https://github.com/lipis/flag-icons/raw/260c91531be024944c6514130c5defb2ebb02b7d/flags/4x3/nl.svg",
1639
1639
  tasks=MTEBTasks(
1640
1640
  get_tasks(
@@ -1678,7 +1678,7 @@ MTEB_NL = Benchmark(
1678
1678
  "SciFact-NL",
1679
1679
  "NFCorpus-NL",
1680
1680
  "BelebeleRetrieval",
1681
- # "WebFAQRetrieval",
1681
+ "WebFAQRetrieval",
1682
1682
  "DutchNewsArticlesRetrieval",
1683
1683
  "bBSARDNLRetrieval",
1684
1684
  "LegalQANLRetrieval",
mteb/evaluate.py CHANGED
@@ -333,7 +333,7 @@ def evaluate(
333
333
  task = cast(AbsTaskAggregate, tasks)
334
334
  results = evaluate(
335
335
  model,
336
- task.metadata.task_list,
336
+ task.metadata.tasks,
337
337
  co2_tracker=co2_tracker,
338
338
  raise_error=raise_error,
339
339
  encode_kwargs=encode_kwargs,
mteb/leaderboard/app.py CHANGED
@@ -318,7 +318,7 @@ def get_leaderboard_app(cache: ResultCache = ResultCache()) -> gr.Blocks:
318
318
  """
319
319
  ## Embedding Leaderboard
320
320
 
321
- This leaderboard compares 100+ text and image embedding models across 1000+ languages. We refer to the publication of each selectable benchmark for details on metrics, languages, tasks, and task types. Anyone is welcome [to add a model](https://github.com/embeddings-benchmark/mteb/blob/main/docs/adding_a_model.md), [add benchmarks](https://github.com/embeddings-benchmark/mteb/blob/main/docs/adding_a_benchmark.md), [help us improve zero-shot annotations](https://github.com/embeddings-benchmark/mteb/blob/06489abca007261c7e6b11f36d4844c5ed5efdcb/mteb/models/bge_models.py#L91) or [propose other changes to the leaderboard](https://github.com/embeddings-benchmark/mteb/tree/main/mteb/leaderboard).
321
+ This leaderboard compares 100+ text and image embedding models across 1000+ languages. We refer to the publication of each selectable benchmark for details on metrics, languages, tasks, and task types. Anyone is welcome [to add a model](https://embeddings-benchmark.github.io/mteb/contributing/adding_a_model/), [add benchmarks](https://embeddings-benchmark.github.io/mteb/contributing/adding_a_benchmark/), [help us improve zero-shot annotations](https://github.com/embeddings-benchmark/mteb/blob/06489abca007261c7e6b11f36d4844c5ed5efdcb/mteb/models/bge_models.py#L91) or [propose other changes to the leaderboard](https://github.com/embeddings-benchmark/mteb/issues/new?template=enhancement.yaml).
322
322
  """
323
323
  )
324
324
  gr.Markdown(
@@ -73,6 +73,7 @@ GP_BENCHMARK_ENTRIES = [
73
73
  "MTEB(fra, v1)",
74
74
  "MTEB(jpn, v1)",
75
75
  "MTEB(kor, v1)",
76
+ "MTEB(nld, v1)",
76
77
  "MTEB(pol, v1)",
77
78
  "MTEB(rus, v1)",
78
79
  "MTEB(fas, v2)",
@@ -18,7 +18,7 @@ def instruction_template(
18
18
  ) -> str:
19
19
  return (
20
20
  # https://github.com/facebookresearch/ReasonIR/blob/0aac96269e455965949df16520fab72da68ffc22/evaluation/bright/configs/reasonir/economics.json#L3
21
- f"<|user|>\n{instruction}<|embed|>\n"
21
+ f"<|user|>\n{instruction}\n<|embed|>\n"
22
22
  if (prompt_type is None or prompt_type == PromptType.query) and instruction
23
23
  else "<|embed|>\n"
24
24
  )
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: mteb
3
- Version: 2.1.1
3
+ Version: 2.1.3
4
4
  Summary: Massive Text Embedding Benchmark
5
5
  Author-email: MTEB Contributors <niklas@huggingface.co>, Kenneth Enevoldsen <kenneth.enevoldsen@cas.au.dk>, Nouamane Tazi <nouamane@huggingface.co>, Nils Reimers <info@nils-reimers.de>
6
6
  Maintainer-email: Kenneth Enevoldsen <kenneth.enevoldsen@cas.au.dk>, Roman Solomatin <risolomatin@gmail.com>, Isaac Chung <chungisaac1217@gmail.com>
@@ -7,7 +7,7 @@ mteb/_requires_package.py,sha256=eHg_TD9BVZRzNCcQQrUP17d8M1DF_vOd_tVx54AmAnM,301
7
7
  mteb/_set_seed.py,sha256=HPlPRl__Pe6IG-4UgJqTfplcivJ_wA2kaClbXoHQedM,1178
8
8
  mteb/cache.py,sha256=rjpXTo9FTuYNwuZGQxZ1v9Yj8pIeMvwW7WNWM4CI6zA,20010
9
9
  mteb/deprecated_evaluator.py,sha256=t13Eluvm5ByVIOqgT7fqiVfLb8Ud3A4bbF2djRfs8iA,26901
10
- mteb/evaluate.py,sha256=nk7OzWuLTgcEJsrHY6KcRlcUJyeTJu3xujytmDkHXZ0,17157
10
+ mteb/evaluate.py,sha256=a8dE6GtDC8TC4Q9rvJfzPuPNDbuNJSx8TAhR-aFMnLQ,17153
11
11
  mteb/filter_tasks.py,sha256=5XE1OYmgDDoJYnXwFf4ma_PIT_Lekzs420sQF_kpCiY,7240
12
12
  mteb/get_tasks.py,sha256=6Gc18a2bZoLQV1Ms_qdr2KieAqIXg8TDg4l7ZN8rW2I,14218
13
13
  mteb/load_results.py,sha256=Xw2ZX7BToU92WwUTQUQKPAgPhX7ucyRRdoCrxAoPHdI,6414
@@ -19,7 +19,7 @@ mteb/_evaluators/clustering_evaluator.py,sha256=uw9mNJ_swuZFwjU8b1XRMsI49nip-_fO
19
19
  mteb/_evaluators/evaluator.py,sha256=gwaeftcAKoGcIQs8jIaafynbcYrYErj6AitHBxgjn2w,807
20
20
  mteb/_evaluators/pair_classification_evaluator.py,sha256=m-xxDvSLrsqccbpmHur3h2dLl5YAC8Noy4n3-p0ewBM,6770
21
21
  mteb/_evaluators/retrieval_evaluator.py,sha256=HsowKZkqRCNzTwM7EcsHX18KhVKAjrm0sa_wFrreCb8,3031
22
- mteb/_evaluators/retrieval_metrics.py,sha256=EzeasJMK2aDCXLtcXKUsdZwLrJF3pV_eLh6H3pd1BD0,23803
22
+ mteb/_evaluators/retrieval_metrics.py,sha256=we0damQCJrdaRUD6JlU2MM7Ls9xERP_OBS5gHt53u9Q,23588
23
23
  mteb/_evaluators/sklearn_evaluator.py,sha256=P6YRx4we1AHmIZJV_VTzHiZ-tx1kV7iG_t4vutlXixA,4491
24
24
  mteb/_evaluators/zeroshot_classification_evaluator.py,sha256=i_4XXqZ_naSWHWSJChzxby4x9EGG0FjOSfDzB_7IVNc,2288
25
25
  mteb/_evaluators/image/__init__.py,sha256=CsQd7OMkeV2Phun7paPWjayZ5qRnvj8H0TYBFeqMxag,148
@@ -56,7 +56,7 @@ mteb/benchmarks/_create_table.py,sha256=gYgwhACmY8JSN00y6r3HfHNQphCIuvcw_IwRZNQw
56
56
  mteb/benchmarks/benchmark.py,sha256=9pbrn5vsSB9Xo23YqQXjRdzorXnMulkm1QvJUkqvuGk,3445
57
57
  mteb/benchmarks/get_benchmark.py,sha256=Zh_C_RXl02hBxg6ffQlxEFERrokaern1RAN4yh2pa2E,3833
58
58
  mteb/benchmarks/benchmarks/__init__.py,sha256=Et7Gf7pKDxa939bIvAJB6OCirQrWAbmbbI3LHA6eMa0,2070
59
- mteb/benchmarks/benchmarks/benchmarks.py,sha256=_qo3bfJsEhqGdovIPxZnYCT-4sJynNMpBdV0AjhMTgw,88420
59
+ mteb/benchmarks/benchmarks/benchmarks.py,sha256=iddOlyNIqnxd6Mp_-cQcADKbZ3gc2uZKxyEutJvv78o,88416
60
60
  mteb/benchmarks/benchmarks/rteb_benchmarks.py,sha256=QnCSrTTaBfcRlAQp2Nu81tgv1idMXqiM16Fp2zKJ5Ys,10607
61
61
  mteb/cli/__init__.py,sha256=v-csUr3eUZElIvrGB6QGtaIdndDfNWEe9oZchsGsJpg,64
62
62
  mteb/cli/_display_tasks.py,sha256=7A06dT9sSoTz6shyMvskPxuc5eHY_H7PGPlROzMP0yw,2196
@@ -1408,8 +1408,8 @@ mteb/languages/language_family.json,sha256=OUGcHeOIPcZPb2FWmYLhxTS0JxjK5y3Fo6x0P
1408
1408
  mteb/languages/language_scripts.py,sha256=5wix9HTYolNIpTiS5oXf2pGJyL7ftdGKs_m432w81V8,3998
1409
1409
  mteb/languages/programming_languages.py,sha256=zxAakT3OSUnAuTnQ34VyeFIECnNXMlleZmAake6jsZE,211
1410
1410
  mteb/leaderboard/__init__.py,sha256=991roXmtRwEQysV-37hWEzWpkvPgMCGRqZTHR-hm2io,88
1411
- mteb/leaderboard/app.py,sha256=5XcuTCJUS5WPDTBejeVETWY3vpn78Fq7zVR2vv0iS7U,32575
1412
- mteb/leaderboard/benchmark_selector.py,sha256=TnW2Y-xoDkEH3ZhR_8EW42WZN_XrtfucUbiuiNjChCA,7580
1411
+ mteb/leaderboard/app.py,sha256=xvOcK_ICmAwl1Mo5muI7fmCjYeSt1ztSgsQDrPF6OvM,32575
1412
+ mteb/leaderboard/benchmark_selector.py,sha256=1PRTDcNU4o2ZVHAzAUVp6UxWNElZDcGbnKHsNGv4aco,7621
1413
1413
  mteb/leaderboard/figures.py,sha256=Rq20LFpaUhQD4tuKp7P7ExQtAjonMLibgO3ud0ykMag,7491
1414
1414
  mteb/leaderboard/table.py,sha256=qs0H_Gt9FzRvzb-AL0YlqEe0YAsdYsVX3QlncfCBEqg,7828
1415
1415
  mteb/leaderboard/text_segments.py,sha256=iMIkS04QQjPbT-SkU0x6fOcS8xRbUYevryu9HydipKM,6570
@@ -1505,7 +1505,7 @@ mteb/models/model_implementations/qtack_models.py,sha256=biZLH5E3UWIcMZXIZNGgBZF
1505
1505
  mteb/models/model_implementations/qwen3_models.py,sha256=IpPJC_y-Kt_1Bq6nT2lX3-lMl-hl0BV6VhAL8Czfyfg,5133
1506
1506
  mteb/models/model_implementations/qzhou_models.py,sha256=90tsGftdCX90Gvu63bxtMyzczGi_u9bsUlg4WAVeH9Q,3536
1507
1507
  mteb/models/model_implementations/random_baseline.py,sha256=NH-epZ73BYvV6qYgu09BsoiFxBWypUNLdJIXRp5QPhM,8030
1508
- mteb/models/model_implementations/reasonir_model.py,sha256=MJWHFiu4x2s3GwpuI5_E9bcwvU_KL-FIROzGaw0SMJQ,2216
1508
+ mteb/models/model_implementations/reasonir_model.py,sha256=wSCcJpUgZ0pG2g3vTEzYNmPlPG_CVn_rR0ENVCines0,2218
1509
1509
  mteb/models/model_implementations/repllama_models.py,sha256=89HoqEpzkNysHeuf_-YhU8WETamHTogSRztGIRo6G1s,7321
1510
1510
  mteb/models/model_implementations/rerankers_custom.py,sha256=ro73A9-hHudy3_qIMrhP-ja-3Xqu78r_aORm856zHQc,10651
1511
1511
  mteb/models/model_implementations/rerankers_monot5_based.py,sha256=rxVwzapNnHl4gCw79XVCaTXj3-wbToyj7XVL97tpAF4,34302
@@ -2536,9 +2536,9 @@ mteb/types/_metadata.py,sha256=NN-W0S6a5TDV7UkpRx1pyWtGF4TyyCyoPUfHOwdeci8,2290
2536
2536
  mteb/types/_result.py,sha256=CRAUc5IvqI3_9SyXDwv-PWLCXwXdZem9RePeYESRtuw,996
2537
2537
  mteb/types/_string_validators.py,sha256=PY-dYq4E8O50VS3bLYdldPWp400fl_WzUjfVSkNWe8U,523
2538
2538
  mteb/types/statistics.py,sha256=YwJsxTf1eaCI_RE-J37a-gK5wDeGAsmkeZKoZCFihSo,3755
2539
- mteb-2.1.1.dist-info/licenses/LICENSE,sha256=xx0jnfkXJvxRnG63LTGOxlggYnIysveWIZ6H3PNdCrQ,11357
2540
- mteb-2.1.1.dist-info/METADATA,sha256=T4fFqeirUqZMgxCCsxYn0bCZDj13uyWBNKXR_DWnRKE,13573
2541
- mteb-2.1.1.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
2542
- mteb-2.1.1.dist-info/entry_points.txt,sha256=8IJoEJFKoDHmVnNev-qJ9pp4Ln7_1-ma9QsXnzVCzGU,39
2543
- mteb-2.1.1.dist-info/top_level.txt,sha256=OLVIjcQAlWBz0bdmutKlWHLF42FF0hp4uVAg3ZyiG4U,5
2544
- mteb-2.1.1.dist-info/RECORD,,
2539
+ mteb-2.1.3.dist-info/licenses/LICENSE,sha256=xx0jnfkXJvxRnG63LTGOxlggYnIysveWIZ6H3PNdCrQ,11357
2540
+ mteb-2.1.3.dist-info/METADATA,sha256=fSSp_uFAo3SSrjhhMQGHJMDIyEjeinv-7QjmK11VrPQ,13573
2541
+ mteb-2.1.3.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
2542
+ mteb-2.1.3.dist-info/entry_points.txt,sha256=8IJoEJFKoDHmVnNev-qJ9pp4Ln7_1-ma9QsXnzVCzGU,39
2543
+ mteb-2.1.3.dist-info/top_level.txt,sha256=OLVIjcQAlWBz0bdmutKlWHLF42FF0hp4uVAg3ZyiG4U,5
2544
+ mteb-2.1.3.dist-info/RECORD,,
File without changes