mteb 2.7.2__py3-none-any.whl → 2.7.4__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- mteb/_create_dataloaders.py +16 -9
- mteb/_evaluators/any_sts_evaluator.py +10 -5
- mteb/_evaluators/clustering_evaluator.py +10 -4
- mteb/_evaluators/evaluator.py +9 -4
- mteb/_evaluators/image/imagetext_pairclassification_evaluator.py +6 -4
- mteb/_evaluators/pair_classification_evaluator.py +10 -5
- mteb/_evaluators/retrieval_evaluator.py +19 -13
- mteb/_evaluators/retrieval_metrics.py +9 -3
- mteb/_evaluators/sklearn_evaluator.py +14 -10
- mteb/_evaluators/text/bitext_mining_evaluator.py +8 -3
- mteb/_evaluators/text/summarization_evaluator.py +8 -4
- mteb/_evaluators/zeroshot_classification_evaluator.py +10 -3
- mteb/_helpful_enum.py +5 -1
- mteb/abstasks/_data_filter/filters.py +8 -2
- mteb/abstasks/_data_filter/task_pipelines.py +7 -2
- mteb/abstasks/_statistics_calculation.py +6 -4
- mteb/abstasks/abstask.py +17 -9
- mteb/abstasks/aggregate_task_metadata.py +20 -9
- mteb/abstasks/aggregated_task.py +15 -8
- mteb/abstasks/classification.py +15 -6
- mteb/abstasks/clustering.py +17 -8
- mteb/abstasks/clustering_legacy.py +14 -6
- mteb/abstasks/image/image_text_pair_classification.py +17 -7
- mteb/abstasks/multilabel_classification.py +11 -5
- mteb/abstasks/pair_classification.py +19 -9
- mteb/abstasks/regression.py +14 -6
- mteb/abstasks/retrieval.py +28 -17
- mteb/abstasks/retrieval_dataset_loaders.py +11 -8
- mteb/abstasks/sts.py +19 -10
- mteb/abstasks/task_metadata.py +17 -8
- mteb/abstasks/text/bitext_mining.py +14 -7
- mteb/abstasks/text/summarization.py +17 -7
- mteb/abstasks/zeroshot_classification.py +15 -7
- mteb/benchmarks/_create_table.py +13 -3
- mteb/benchmarks/benchmark.py +11 -1
- mteb/benchmarks/benchmarks/__init__.py +2 -0
- mteb/benchmarks/benchmarks/benchmarks.py +41 -2
- mteb/benchmarks/benchmarks/rteb_benchmarks.py +20 -9
- mteb/cache.py +10 -5
- mteb/cli/_display_tasks.py +9 -3
- mteb/cli/build_cli.py +5 -2
- mteb/cli/generate_model_card.py +9 -2
- mteb/deprecated_evaluator.py +16 -12
- mteb/descriptive_stats/Retrieval/BrightAopsRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightBiologyLongRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightBiologyRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightEarthScienceLongRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightEarthScienceRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightEconomicsLongRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightEconomicsRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightLeetcodeRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightPonyLongRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightPonyRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightPsychologyLongRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightPsychologyRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightRoboticsLongRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightRoboticsRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightStackoverflowLongRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightStackoverflowRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightSustainableLivingLongRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightSustainableLivingRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightTheoremQAQuestionsRetrieval.json +35 -0
- mteb/descriptive_stats/Retrieval/BrightTheoremQATheoremsRetrieval.json +35 -0
- mteb/evaluate.py +20 -18
- mteb/filter_tasks.py +12 -7
- mteb/get_tasks.py +9 -4
- mteb/languages/language_scripts.py +8 -3
- mteb/leaderboard/app.py +7 -3
- mteb/leaderboard/table.py +7 -2
- mteb/load_results.py +9 -3
- mteb/models/abs_encoder.py +22 -12
- mteb/models/cache_wrappers/cache_backend_protocol.py +5 -3
- mteb/models/cache_wrappers/cache_backends/_hash_utils.py +8 -4
- mteb/models/cache_wrappers/cache_backends/faiss_cache.py +8 -3
- mteb/models/cache_wrappers/cache_wrapper.py +14 -9
- mteb/models/get_model_meta.py +11 -4
- mteb/models/instruct_wrapper.py +13 -5
- mteb/models/model_implementations/align_models.py +10 -4
- mteb/models/model_implementations/amazon_models.py +1 -0
- mteb/models/model_implementations/andersborges.py +2 -0
- mteb/models/model_implementations/ara_models.py +1 -0
- mteb/models/model_implementations/arctic_models.py +8 -0
- mteb/models/model_implementations/b1ade_models.py +1 -0
- mteb/models/model_implementations/bedrock_models.py +20 -6
- mteb/models/model_implementations/bge_models.py +40 -1
- mteb/models/model_implementations/bica_model.py +1 -0
- mteb/models/model_implementations/blip2_models.py +11 -4
- mteb/models/model_implementations/blip_models.py +17 -4
- mteb/models/model_implementations/bm25.py +22 -14
- mteb/models/model_implementations/bmretriever_models.py +10 -2
- mteb/models/model_implementations/cadet_models.py +1 -0
- mteb/models/model_implementations/cde_models.py +11 -5
- mteb/models/model_implementations/clip_models.py +12 -4
- mteb/models/model_implementations/clips_models.py +3 -0
- mteb/models/model_implementations/codefuse_models.py +5 -0
- mteb/models/model_implementations/codesage_models.py +3 -0
- mteb/models/model_implementations/cohere_models.py +14 -4
- mteb/models/model_implementations/cohere_v.py +14 -4
- mteb/models/model_implementations/colpali_models.py +7 -3
- mteb/models/model_implementations/colqwen_models.py +17 -31
- mteb/models/model_implementations/colsmol_models.py +3 -1
- mteb/models/model_implementations/conan_models.py +11 -4
- mteb/models/model_implementations/dino_models.py +28 -4
- mteb/models/model_implementations/e5_instruct.py +4 -0
- mteb/models/model_implementations/e5_models.py +9 -0
- mteb/models/model_implementations/e5_v.py +10 -4
- mteb/models/model_implementations/eagerworks_models.py +11 -4
- mteb/models/model_implementations/emillykkejensen_models.py +3 -0
- mteb/models/model_implementations/en_code_retriever.py +1 -0
- mteb/models/model_implementations/euler_models.py +1 -0
- mteb/models/model_implementations/evaclip_models.py +13 -4
- mteb/models/model_implementations/fa_models.py +9 -0
- mteb/models/model_implementations/facebookai.py +2 -0
- mteb/models/model_implementations/geogpt_models.py +1 -0
- mteb/models/model_implementations/gme_v_models.py +7 -3
- mteb/models/model_implementations/google_models.py +15 -4
- mteb/models/model_implementations/granite_vision_embedding_models.py +7 -5
- mteb/models/model_implementations/gritlm_models.py +2 -0
- mteb/models/model_implementations/gte_models.py +9 -0
- mteb/models/model_implementations/hinvec_models.py +6 -1
- mteb/models/model_implementations/human.py +1 -0
- mteb/models/model_implementations/ibm_granite_models.py +6 -0
- mteb/models/model_implementations/inf_models.py +2 -0
- mteb/models/model_implementations/jasper_models.py +14 -5
- mteb/models/model_implementations/jina_clip.py +10 -4
- mteb/models/model_implementations/jina_models.py +17 -5
- mteb/models/model_implementations/kalm_models.py +24 -12
- mteb/models/model_implementations/kblab.py +1 -0
- mteb/models/model_implementations/kennethenevoldsen_models.py +2 -0
- mteb/models/model_implementations/kfst.py +1 -0
- mteb/models/model_implementations/kowshik24_models.py +1 -0
- mteb/models/model_implementations/lens_models.py +2 -0
- mteb/models/model_implementations/lgai_embedding_models.py +1 -0
- mteb/models/model_implementations/linq_models.py +7 -1
- mteb/models/model_implementations/listconranker.py +10 -4
- mteb/models/model_implementations/llm2clip_models.py +12 -4
- mteb/models/model_implementations/llm2vec_models.py +20 -6
- mteb/models/model_implementations/mcinext_models.py +8 -2
- mteb/models/model_implementations/mdbr_models.py +2 -0
- mteb/models/model_implementations/misc_models.py +63 -0
- mteb/models/model_implementations/mixedbread_ai_models.py +3 -0
- mteb/models/model_implementations/mme5_models.py +2 -1
- mteb/models/model_implementations/moco_models.py +11 -4
- mteb/models/model_implementations/mod_models.py +2 -1
- mteb/models/model_implementations/model2vec_models.py +23 -4
- mteb/models/model_implementations/moka_models.py +3 -0
- mteb/models/model_implementations/nbailab.py +3 -0
- mteb/models/model_implementations/no_instruct_sentence_models.py +13 -5
- mteb/models/model_implementations/nomic_models.py +16 -4
- mteb/models/model_implementations/nomic_models_vision.py +5 -3
- mteb/models/model_implementations/nvidia_llama_nemoretriever_colemb.py +9 -3
- mteb/models/model_implementations/nvidia_models.py +15 -4
- mteb/models/model_implementations/octen_models.py +3 -1
- mteb/models/model_implementations/openai_models.py +14 -4
- mteb/models/model_implementations/openclip_models.py +17 -4
- mteb/models/model_implementations/opensearch_neural_sparse_models.py +15 -4
- mteb/models/model_implementations/ops_moa_models.py +9 -2
- mteb/models/model_implementations/ordalietech_solon_embeddings_mini_beta_1_1.py +1 -0
- mteb/models/model_implementations/pawan_models.py +1 -0
- mteb/models/model_implementations/piccolo_models.py +2 -0
- mteb/models/model_implementations/promptriever_models.py +16 -6
- mteb/models/model_implementations/pylate_models.py +22 -13
- mteb/models/model_implementations/qodo_models.py +2 -0
- mteb/models/model_implementations/qtack_models.py +1 -0
- mteb/models/model_implementations/qwen3_models.py +11 -1
- mteb/models/model_implementations/qzhou_models.py +2 -0
- mteb/models/model_implementations/random_baseline.py +4 -3
- mteb/models/model_implementations/rasgaard_models.py +1 -0
- mteb/models/model_implementations/reasonir_model.py +65 -0
- mteb/models/model_implementations/repllama_models.py +15 -6
- mteb/models/model_implementations/rerankers_custom.py +13 -4
- mteb/models/model_implementations/rerankers_monot5_based.py +24 -4
- mteb/models/model_implementations/richinfoai_models.py +1 -0
- mteb/models/model_implementations/ru_sentence_models.py +20 -0
- mteb/models/model_implementations/ruri_models.py +10 -0
- mteb/models/model_implementations/salesforce_models.py +10 -1
- mteb/models/model_implementations/samilpwc_models.py +1 -0
- mteb/models/model_implementations/sarashina_embedding_models.py +2 -0
- mteb/models/model_implementations/searchmap_models.py +1 -0
- mteb/models/model_implementations/seed_1_6_embedding_models.py +5 -2
- mteb/models/model_implementations/seed_1_6_embedding_models_1215.py +6 -2
- mteb/models/model_implementations/seed_models.py +2 -1
- mteb/models/model_implementations/sentence_transformers_models.py +18 -0
- mteb/models/model_implementations/shuu_model.py +1 -0
- mteb/models/model_implementations/siglip_models.py +19 -4
- mteb/models/model_implementations/slm_models.py +7 -4
- mteb/models/model_implementations/sonar_models.py +2 -1
- mteb/models/model_implementations/spartan8806_atles_champion.py +1 -0
- mteb/models/model_implementations/stella_models.py +6 -0
- mteb/models/model_implementations/tarka_models.py +2 -0
- mteb/models/model_implementations/text2vec_models.py +3 -0
- mteb/models/model_implementations/ua_sentence_models.py +1 -0
- mteb/models/model_implementations/uae_models.py +10 -4
- mteb/models/model_implementations/vdr_models.py +8 -1
- mteb/models/model_implementations/vi_vn_models.py +6 -0
- mteb/models/model_implementations/vista_models.py +11 -4
- mteb/models/model_implementations/vlm2vec_models.py +11 -4
- mteb/models/model_implementations/voyage_models.py +25 -4
- mteb/models/model_implementations/voyage_v.py +11 -6
- mteb/models/model_implementations/xyz_models.py +1 -0
- mteb/models/model_implementations/youtu_models.py +1 -0
- mteb/models/model_implementations/yuan_models.py +1 -0
- mteb/models/model_implementations/yuan_models_en.py +2 -1
- mteb/models/model_meta.py +47 -9
- mteb/models/models_protocols.py +19 -18
- mteb/models/search_encoder_index/search_backend_protocol.py +7 -3
- mteb/models/search_encoder_index/search_indexes/faiss_search_index.py +12 -4
- mteb/models/search_wrappers.py +19 -12
- mteb/models/sentence_transformer_wrapper.py +4 -3
- mteb/models/vllm_wrapper.py +8 -6
- mteb/results/benchmark_results.py +22 -17
- mteb/results/model_result.py +21 -15
- mteb/results/task_result.py +15 -9
- mteb/similarity_functions.py +8 -2
- mteb/tasks/aggregated_tasks/eng/cqadupstack_retrieval.py +3 -3
- mteb/tasks/aggregated_tasks/eng/sts17_multilingual_visual_sts_eng.py +3 -3
- mteb/tasks/aggregated_tasks/eng/sts_benchmark_multilingual_visual_sts_eng.py +3 -3
- mteb/tasks/aggregated_tasks/fas/cqadupstack_retrieval_fa.py +3 -3
- mteb/tasks/aggregated_tasks/fas/syn_per_chatbot_conv_sa_classification.py +3 -3
- mteb/tasks/aggregated_tasks/multilingual/sts17_multilingual_vision_sts.py +3 -3
- mteb/tasks/aggregated_tasks/multilingual/sts_benchmark_multilingual_visual_sts.py +3 -3
- mteb/tasks/aggregated_tasks/nld/cqadupstack_nl_retrieval.py +3 -3
- mteb/tasks/aggregated_tasks/pol/cqadupstack_retrieval_pl.py +3 -3
- mteb/tasks/clustering/nob/snl_clustering.py +7 -2
- mteb/tasks/clustering/nob/vg_clustering.py +7 -2
- mteb/tasks/retrieval/eng/__init__.py +42 -0
- mteb/tasks/retrieval/eng/bright_retrieval.py +9 -1
- mteb/tasks/retrieval/eng/bright_v1_1_retrieval.py +968 -0
- mteb/tasks/retrieval/eng/limit_retrieval.py +6 -1
- mteb/tasks/retrieval/multilingual/ru_sci_bench_retrieval.py +3 -3
- mteb/types/_encoder_io.py +1 -1
- mteb/types/statistics.py +9 -2
- {mteb-2.7.2.dist-info → mteb-2.7.4.dist-info}/METADATA +1 -1
- {mteb-2.7.2.dist-info → mteb-2.7.4.dist-info}/RECORD +238 -217
- {mteb-2.7.2.dist-info → mteb-2.7.4.dist-info}/WHEEL +0 -0
- {mteb-2.7.2.dist-info → mteb-2.7.4.dist-info}/entry_points.txt +0 -0
- {mteb-2.7.2.dist-info → mteb-2.7.4.dist-info}/licenses/LICENSE +0 -0
- {mteb-2.7.2.dist-info → mteb-2.7.4.dist-info}/top_level.txt +0 -0
|
@@ -9,6 +9,7 @@ kowshik24_bangla_embedding_model = ModelMeta(
|
|
|
9
9
|
revision="6689c21e69be5950596bad084457cbaa138728d8",
|
|
10
10
|
release_date="2025-11-10",
|
|
11
11
|
n_parameters=278_000_000,
|
|
12
|
+
n_embedding_parameters=192_001_536,
|
|
12
13
|
memory_usage_mb=1061,
|
|
13
14
|
embed_dim=768,
|
|
14
15
|
license="apache-2.0",
|
|
@@ -18,6 +18,7 @@ lens_d4000 = ModelMeta(
|
|
|
18
18
|
revision="e473b33364e6c48a324796fd1411d3b93670c6fe",
|
|
19
19
|
release_date="2025-01-17",
|
|
20
20
|
n_parameters=int(7.11 * 1e9),
|
|
21
|
+
n_embedding_parameters=131_084_288,
|
|
21
22
|
memory_usage_mb=27125,
|
|
22
23
|
embed_dim=4000,
|
|
23
24
|
license="apache-2.0",
|
|
@@ -41,6 +42,7 @@ lens_d8000 = ModelMeta(
|
|
|
41
42
|
revision="a0b87bd91cb27b6f2f0b0fe22c28026da1d464ef",
|
|
42
43
|
release_date="2025-01-17",
|
|
43
44
|
n_parameters=int(7.11 * 1e9),
|
|
45
|
+
n_embedding_parameters=131_084_288,
|
|
44
46
|
memory_usage_mb=27125,
|
|
45
47
|
embed_dim=8000,
|
|
46
48
|
license="apache-2.0",
|
|
@@ -1,11 +1,16 @@
|
|
|
1
|
+
from __future__ import annotations
|
|
2
|
+
|
|
3
|
+
from typing import TYPE_CHECKING
|
|
4
|
+
|
|
1
5
|
import torch
|
|
2
6
|
|
|
3
7
|
from mteb.models.instruct_wrapper import instruct_wrapper
|
|
4
8
|
from mteb.models.model_meta import ModelMeta, ScoringFunction
|
|
5
|
-
from mteb.types import PromptType
|
|
6
9
|
|
|
7
10
|
from .e5_instruct import E5_MISTRAL_TRAINING_DATA
|
|
8
11
|
|
|
12
|
+
if TYPE_CHECKING:
|
|
13
|
+
from mteb.types import PromptType
|
|
9
14
|
LINQ_EMBED_MISTRAL_CITATION = """@misc{LinqAIResearch2024,
|
|
10
15
|
title={Linq-Embed-Mistral:Elevating Text Retrieval with Improved GPT Data Through Task-Specific Control and Quality Refinement},
|
|
11
16
|
author={Junseong Kim and Seolhwa Lee and Jihoon Kwon and Sangmo Gu and Yejin Kim and Minkyung Cho and Jy-yong Sohn and Chanyeol Choi},
|
|
@@ -38,6 +43,7 @@ Linq_Embed_Mistral = ModelMeta(
|
|
|
38
43
|
revision="0c1a0b0589177079acc552433cad51d7c9132379",
|
|
39
44
|
release_date="2024-05-29", # initial commit of hf model.
|
|
40
45
|
n_parameters=7_110_000_000,
|
|
46
|
+
n_embedding_parameters=None,
|
|
41
47
|
memory_usage_mb=13563,
|
|
42
48
|
embed_dim=4096,
|
|
43
49
|
license="cc-by-nc-4.0",
|
|
@@ -1,14 +1,19 @@
|
|
|
1
|
-
from
|
|
1
|
+
from __future__ import annotations
|
|
2
|
+
|
|
3
|
+
from typing import TYPE_CHECKING, Any
|
|
2
4
|
|
|
3
5
|
import torch
|
|
4
|
-
from torch.utils.data import DataLoader
|
|
5
6
|
|
|
6
|
-
from mteb.abstasks.task_metadata import TaskMetadata
|
|
7
7
|
from mteb.models.model_meta import ModelMeta
|
|
8
|
-
from mteb.types import BatchedInput, PromptType
|
|
9
8
|
|
|
10
9
|
from .rerankers_custom import RerankerWrapper
|
|
11
10
|
|
|
11
|
+
if TYPE_CHECKING:
|
|
12
|
+
from torch.utils.data import DataLoader
|
|
13
|
+
|
|
14
|
+
from mteb.abstasks.task_metadata import TaskMetadata
|
|
15
|
+
from mteb.types import BatchedInput, PromptType
|
|
16
|
+
|
|
12
17
|
LISTCONRANKER_CITATION = """@article{liu2025listconranker,
|
|
13
18
|
title={ListConRanker: A Contrastive Text Reranker with Listwise Encoding},
|
|
14
19
|
author={Liu, Junlong and Ma, Yue and Zhao, Ruihui and Zheng, Junhao and Ma, Qianli and Kang, Yangyang},
|
|
@@ -118,6 +123,7 @@ listconranker = ModelMeta(
|
|
|
118
123
|
revision="95ae6a5f422a916bc36520f0f3e198e7d91520a0",
|
|
119
124
|
release_date="2024-12-11",
|
|
120
125
|
n_parameters=401_000_000,
|
|
126
|
+
n_embedding_parameters=None,
|
|
121
127
|
memory_usage_mb=1242,
|
|
122
128
|
similarity_fn_name="cosine",
|
|
123
129
|
training_datasets=listconranker_training_datasets,
|
|
@@ -1,15 +1,20 @@
|
|
|
1
|
+
from __future__ import annotations
|
|
2
|
+
|
|
1
3
|
from pathlib import Path
|
|
2
|
-
from typing import Any
|
|
4
|
+
from typing import TYPE_CHECKING, Any
|
|
3
5
|
|
|
4
6
|
import torch
|
|
5
|
-
from torch.utils.data import DataLoader
|
|
6
7
|
from tqdm.auto import tqdm
|
|
7
8
|
|
|
8
9
|
from mteb._requires_package import requires_image_dependencies, requires_package
|
|
9
|
-
from mteb.abstasks.task_metadata import TaskMetadata
|
|
10
10
|
from mteb.models.abs_encoder import AbsEncoder
|
|
11
11
|
from mteb.models.model_meta import ModelMeta, ScoringFunction
|
|
12
|
-
|
|
12
|
+
|
|
13
|
+
if TYPE_CHECKING:
|
|
14
|
+
from torch.utils.data import DataLoader
|
|
15
|
+
|
|
16
|
+
from mteb.abstasks.task_metadata import TaskMetadata
|
|
17
|
+
from mteb.types import Array, BatchedInput, PromptType
|
|
13
18
|
|
|
14
19
|
LLM2CLIP_CITATION = """@misc{huang2024llm2clippowerfullanguagemodel,
|
|
15
20
|
title={LLM2CLIP: Powerful Language Model Unlock Richer Visual Representation},
|
|
@@ -189,6 +194,7 @@ llm2clip_openai_l_14_336 = ModelMeta(
|
|
|
189
194
|
release_date="2024-11-07",
|
|
190
195
|
modalities=["image", "text"],
|
|
191
196
|
n_parameters=579_000_000,
|
|
197
|
+
n_embedding_parameters=None,
|
|
192
198
|
memory_usage_mb=None,
|
|
193
199
|
max_tokens=None,
|
|
194
200
|
embed_dim=1280,
|
|
@@ -214,6 +220,7 @@ llm2clip_openai_l_14_224 = ModelMeta(
|
|
|
214
220
|
release_date="2024-11-07",
|
|
215
221
|
modalities=["image", "text"],
|
|
216
222
|
n_parameters=578_000_000,
|
|
223
|
+
n_embedding_parameters=None,
|
|
217
224
|
memory_usage_mb=None,
|
|
218
225
|
max_tokens=None,
|
|
219
226
|
embed_dim=1280,
|
|
@@ -238,6 +245,7 @@ llm2clip_openai_b_16 = ModelMeta(
|
|
|
238
245
|
release_date="2024-11-07",
|
|
239
246
|
modalities=["image", "text"],
|
|
240
247
|
n_parameters=361_000_000,
|
|
248
|
+
n_embedding_parameters=None,
|
|
241
249
|
memory_usage_mb=None,
|
|
242
250
|
max_tokens=None,
|
|
243
251
|
embed_dim=1280,
|
|
@@ -1,16 +1,22 @@
|
|
|
1
|
+
from __future__ import annotations
|
|
2
|
+
|
|
1
3
|
import logging
|
|
2
|
-
from
|
|
3
|
-
from typing import Any
|
|
4
|
+
from typing import TYPE_CHECKING, Any
|
|
4
5
|
|
|
5
6
|
import torch
|
|
6
|
-
from torch.utils.data import DataLoader
|
|
7
7
|
|
|
8
8
|
from mteb._requires_package import requires_package, suggest_package
|
|
9
|
-
from mteb.abstasks.task_metadata import TaskMetadata
|
|
10
9
|
from mteb.models.abs_encoder import AbsEncoder
|
|
11
10
|
from mteb.models.model_meta import ModelMeta, ScoringFunction
|
|
12
|
-
|
|
13
|
-
|
|
11
|
+
|
|
12
|
+
if TYPE_CHECKING:
|
|
13
|
+
from collections.abc import Callable
|
|
14
|
+
|
|
15
|
+
from torch.utils.data import DataLoader
|
|
16
|
+
|
|
17
|
+
from mteb.abstasks.task_metadata import TaskMetadata
|
|
18
|
+
from mteb.models.models_protocols import EncoderProtocol
|
|
19
|
+
from mteb.types import Array, BatchedInput, PromptType
|
|
14
20
|
|
|
15
21
|
logger = logging.getLogger(__name__)
|
|
16
22
|
|
|
@@ -139,6 +145,7 @@ llm2vec_llama3_8b_supervised = ModelMeta(
|
|
|
139
145
|
# TODO: Not sure what to put here as a model is made of two peft repos, each with a different revision
|
|
140
146
|
release_date="2024-04-09",
|
|
141
147
|
n_parameters=7_505_000_000,
|
|
148
|
+
n_embedding_parameters=None,
|
|
142
149
|
memory_usage_mb=28629,
|
|
143
150
|
max_tokens=8192,
|
|
144
151
|
embed_dim=4096,
|
|
@@ -168,6 +175,7 @@ llm2vec_llama3_8b_unsupervised = ModelMeta(
|
|
|
168
175
|
revision="1cb7b735326d13a8541db8f57f35da5373f5e9c6",
|
|
169
176
|
release_date="2024-04-09",
|
|
170
177
|
n_parameters=7_505_000_000,
|
|
178
|
+
n_embedding_parameters=None,
|
|
171
179
|
memory_usage_mb=28629,
|
|
172
180
|
max_tokens=8192,
|
|
173
181
|
embed_dim=4096,
|
|
@@ -197,6 +205,7 @@ llm2vec_mistral7b_supervised = ModelMeta(
|
|
|
197
205
|
revision="0ae69bdd5816105778b971c3138e8f8a18eaa3ae",
|
|
198
206
|
release_date="2024-04-09",
|
|
199
207
|
n_parameters=7_111_000_000,
|
|
208
|
+
n_embedding_parameters=131_072_000,
|
|
200
209
|
memory_usage_mb=27126,
|
|
201
210
|
max_tokens=32768,
|
|
202
211
|
embed_dim=4096,
|
|
@@ -226,6 +235,7 @@ llm2vec_mistral7b_unsupervised = ModelMeta(
|
|
|
226
235
|
revision="2c055a5d77126c0d3dc6cd8ffa30e2908f4f45f8",
|
|
227
236
|
release_date="2024-04-09",
|
|
228
237
|
n_parameters=7_111_000_000,
|
|
238
|
+
n_embedding_parameters=131_072_000,
|
|
229
239
|
memory_usage_mb=27126,
|
|
230
240
|
max_tokens=32768,
|
|
231
241
|
embed_dim=4096,
|
|
@@ -255,6 +265,7 @@ llm2vec_llama2_7b_supervised = ModelMeta(
|
|
|
255
265
|
revision="2c055a5d77126c0d3dc6cd8ffa30e2908f4f45f8",
|
|
256
266
|
release_date="2024-04-09",
|
|
257
267
|
n_parameters=7_111_000_000,
|
|
268
|
+
n_embedding_parameters=None,
|
|
258
269
|
memory_usage_mb=27126,
|
|
259
270
|
max_tokens=32768,
|
|
260
271
|
embed_dim=4096,
|
|
@@ -284,6 +295,7 @@ llm2vec_llama2_7b_unsupervised = ModelMeta(
|
|
|
284
295
|
revision="a76944871d169ebe7c97eb921764cd063afed785",
|
|
285
296
|
release_date="2024-04-09",
|
|
286
297
|
n_parameters=7_111_000_000,
|
|
298
|
+
n_embedding_parameters=None,
|
|
287
299
|
memory_usage_mb=27126,
|
|
288
300
|
max_tokens=32768,
|
|
289
301
|
embed_dim=4096,
|
|
@@ -313,6 +325,7 @@ llm2vec_sheared_llama_supervised = ModelMeta(
|
|
|
313
325
|
revision="a5943d406c6b016fef3f07906aac183cf1a0b47d",
|
|
314
326
|
release_date="2024-04-09",
|
|
315
327
|
n_parameters=7_111_000_000,
|
|
328
|
+
n_embedding_parameters=65_536_000,
|
|
316
329
|
memory_usage_mb=27126,
|
|
317
330
|
max_tokens=32768,
|
|
318
331
|
embed_dim=4096,
|
|
@@ -342,6 +355,7 @@ llm2vec_sheared_llama_unsupervised = ModelMeta(
|
|
|
342
355
|
revision="a5943d406c6b016fef3f07906aac183cf1a0b47d",
|
|
343
356
|
release_date="2024-04-09",
|
|
344
357
|
n_parameters=7_111_000_000,
|
|
358
|
+
n_embedding_parameters=65_536_000,
|
|
345
359
|
memory_usage_mb=27126,
|
|
346
360
|
max_tokens=32768,
|
|
347
361
|
embed_dim=4096,
|
|
@@ -1,16 +1,19 @@
|
|
|
1
|
+
from __future__ import annotations
|
|
2
|
+
|
|
1
3
|
import logging
|
|
2
4
|
import os
|
|
3
5
|
import time
|
|
4
6
|
import warnings
|
|
5
|
-
from typing import Any
|
|
7
|
+
from typing import TYPE_CHECKING, Any
|
|
6
8
|
|
|
7
9
|
import numpy as np
|
|
8
10
|
import requests
|
|
9
11
|
|
|
10
12
|
from mteb.models.abs_encoder import AbsEncoder
|
|
11
13
|
from mteb.models.model_meta import ModelMeta
|
|
12
|
-
from mteb.types import PromptType
|
|
13
14
|
|
|
15
|
+
if TYPE_CHECKING:
|
|
16
|
+
from mteb.types import PromptType
|
|
14
17
|
logger = logging.getLogger(__name__)
|
|
15
18
|
|
|
16
19
|
HAKIM_CITATION = """@article{sarmadi2025hakim,
|
|
@@ -355,6 +358,7 @@ hakim = ModelMeta(
|
|
|
355
358
|
revision="1",
|
|
356
359
|
release_date="2025-05-10",
|
|
357
360
|
n_parameters=124_441_344,
|
|
361
|
+
n_embedding_parameters=None,
|
|
358
362
|
memory_usage_mb=475,
|
|
359
363
|
embed_dim=768,
|
|
360
364
|
license="not specified",
|
|
@@ -423,6 +427,7 @@ hakim_small = ModelMeta(
|
|
|
423
427
|
revision="1",
|
|
424
428
|
release_date="2025-05-10",
|
|
425
429
|
n_parameters=38_736_384,
|
|
430
|
+
n_embedding_parameters=None,
|
|
426
431
|
memory_usage_mb=148,
|
|
427
432
|
embed_dim=512,
|
|
428
433
|
license="not specified",
|
|
@@ -490,6 +495,7 @@ hakim_unsup = ModelMeta(
|
|
|
490
495
|
revision="1",
|
|
491
496
|
release_date="2025-05-10",
|
|
492
497
|
n_parameters=124_441_344,
|
|
498
|
+
n_embedding_parameters=None,
|
|
493
499
|
memory_usage_mb=475,
|
|
494
500
|
embed_dim=768,
|
|
495
501
|
license="not specified",
|
|
@@ -45,6 +45,7 @@ mdbr_leaf_ir = ModelMeta(
|
|
|
45
45
|
"Transformers",
|
|
46
46
|
],
|
|
47
47
|
n_parameters=22_861_056,
|
|
48
|
+
n_embedding_parameters=11_720_448,
|
|
48
49
|
memory_usage_mb=86,
|
|
49
50
|
max_tokens=512,
|
|
50
51
|
embed_dim=768,
|
|
@@ -79,6 +80,7 @@ mdbr_leaf_mt = ModelMeta(
|
|
|
79
80
|
"Transformers",
|
|
80
81
|
],
|
|
81
82
|
n_parameters=22_958_592,
|
|
83
|
+
n_embedding_parameters=11_720_448,
|
|
82
84
|
memory_usage_mb=86,
|
|
83
85
|
max_tokens=512,
|
|
84
86
|
embed_dim=1024,
|