mteb 2.6.4__py3-none-any.whl → 2.6.5__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (126) hide show
  1. mteb/abstasks/classification.py +2 -3
  2. mteb/abstasks/multilabel_classification.py +3 -3
  3. mteb/abstasks/regression.py +1 -1
  4. mteb/abstasks/retrieval.py +1 -1
  5. mteb/abstasks/task_metadata.py +9 -14
  6. mteb/models/model_implementations/align_models.py +1 -1
  7. mteb/models/model_implementations/andersborges.py +2 -2
  8. mteb/models/model_implementations/ara_models.py +1 -1
  9. mteb/models/model_implementations/arctic_models.py +8 -8
  10. mteb/models/model_implementations/b1ade_models.py +1 -1
  11. mteb/models/model_implementations/bge_models.py +45 -21
  12. mteb/models/model_implementations/bica_model.py +3 -3
  13. mteb/models/model_implementations/blip2_models.py +2 -2
  14. mteb/models/model_implementations/blip_models.py +8 -8
  15. mteb/models/model_implementations/bmretriever_models.py +4 -4
  16. mteb/models/model_implementations/cadet_models.py +1 -1
  17. mteb/models/model_implementations/cde_models.py +2 -2
  18. mteb/models/model_implementations/clip_models.py +3 -3
  19. mteb/models/model_implementations/clips_models.py +3 -3
  20. mteb/models/model_implementations/codefuse_models.py +5 -5
  21. mteb/models/model_implementations/codesage_models.py +3 -3
  22. mteb/models/model_implementations/cohere_models.py +4 -4
  23. mteb/models/model_implementations/colpali_models.py +3 -3
  24. mteb/models/model_implementations/colqwen_models.py +8 -8
  25. mteb/models/model_implementations/colsmol_models.py +2 -2
  26. mteb/models/model_implementations/conan_models.py +1 -1
  27. mteb/models/model_implementations/dino_models.py +19 -19
  28. mteb/models/model_implementations/e5_instruct.py +23 -4
  29. mteb/models/model_implementations/e5_models.py +9 -9
  30. mteb/models/model_implementations/e5_v.py +1 -1
  31. mteb/models/model_implementations/eagerworks_models.py +1 -1
  32. mteb/models/model_implementations/emillykkejensen_models.py +3 -3
  33. mteb/models/model_implementations/en_code_retriever.py +1 -1
  34. mteb/models/model_implementations/euler_models.py +2 -2
  35. mteb/models/model_implementations/fa_models.py +9 -9
  36. mteb/models/model_implementations/facebookai.py +14 -2
  37. mteb/models/model_implementations/geogpt_models.py +1 -1
  38. mteb/models/model_implementations/gme_v_models.py +2 -2
  39. mteb/models/model_implementations/google_models.py +1 -1
  40. mteb/models/model_implementations/granite_vision_embedding_models.py +1 -1
  41. mteb/models/model_implementations/gritlm_models.py +2 -2
  42. mteb/models/model_implementations/gte_models.py +25 -13
  43. mteb/models/model_implementations/hinvec_models.py +1 -1
  44. mteb/models/model_implementations/ibm_granite_models.py +30 -6
  45. mteb/models/model_implementations/inf_models.py +2 -2
  46. mteb/models/model_implementations/jasper_models.py +2 -2
  47. mteb/models/model_implementations/jina_clip.py +1 -1
  48. mteb/models/model_implementations/jina_models.py +11 -5
  49. mteb/models/model_implementations/kblab.py +12 -6
  50. mteb/models/model_implementations/kennethenevoldsen_models.py +2 -2
  51. mteb/models/model_implementations/kfst.py +1 -1
  52. mteb/models/model_implementations/kowshik24_models.py +1 -1
  53. mteb/models/model_implementations/lgai_embedding_models.py +1 -1
  54. mteb/models/model_implementations/linq_models.py +1 -1
  55. mteb/models/model_implementations/listconranker.py +1 -1
  56. mteb/models/model_implementations/llm2clip_models.py +3 -3
  57. mteb/models/model_implementations/llm2vec_models.py +8 -8
  58. mteb/models/model_implementations/mdbr_models.py +14 -2
  59. mteb/models/model_implementations/misc_models.py +68 -68
  60. mteb/models/model_implementations/mme5_models.py +1 -1
  61. mteb/models/model_implementations/moco_models.py +2 -2
  62. mteb/models/model_implementations/mod_models.py +1 -1
  63. mteb/models/model_implementations/model2vec_models.py +13 -13
  64. mteb/models/model_implementations/moka_models.py +1 -1
  65. mteb/models/model_implementations/mxbai_models.py +16 -3
  66. mteb/models/model_implementations/nbailab.py +3 -3
  67. mteb/models/model_implementations/no_instruct_sentence_models.py +1 -1
  68. mteb/models/model_implementations/nomic_models.py +18 -6
  69. mteb/models/model_implementations/nomic_models_vision.py +1 -1
  70. mteb/models/model_implementations/nvidia_llama_nemoretriever_colemb.py +2 -2
  71. mteb/models/model_implementations/nvidia_models.py +3 -3
  72. mteb/models/model_implementations/octen_models.py +2 -2
  73. mteb/models/model_implementations/openclip_models.py +6 -6
  74. mteb/models/model_implementations/opensearch_neural_sparse_models.py +5 -5
  75. mteb/models/model_implementations/ops_moa_models.py +1 -1
  76. mteb/models/model_implementations/ordalietech_solon_embeddings_mini_beta_1_1.py +1 -1
  77. mteb/models/model_implementations/pawan_models.py +1 -1
  78. mteb/models/model_implementations/piccolo_models.py +1 -1
  79. mteb/models/model_implementations/promptriever_models.py +4 -4
  80. mteb/models/model_implementations/pylate_models.py +5 -5
  81. mteb/models/model_implementations/qodo_models.py +2 -2
  82. mteb/models/model_implementations/qtack_models.py +1 -1
  83. mteb/models/model_implementations/qwen3_models.py +3 -3
  84. mteb/models/model_implementations/qzhou_models.py +2 -2
  85. mteb/models/model_implementations/rasgaard_models.py +1 -1
  86. mteb/models/model_implementations/reasonir_model.py +1 -1
  87. mteb/models/model_implementations/repllama_models.py +1 -1
  88. mteb/models/model_implementations/rerankers_custom.py +9 -3
  89. mteb/models/model_implementations/rerankers_monot5_based.py +14 -14
  90. mteb/models/model_implementations/richinfoai_models.py +1 -1
  91. mteb/models/model_implementations/ru_sentence_models.py +20 -20
  92. mteb/models/model_implementations/ruri_models.py +10 -10
  93. mteb/models/model_implementations/salesforce_models.py +3 -3
  94. mteb/models/model_implementations/samilpwc_models.py +1 -1
  95. mteb/models/model_implementations/sarashina_embedding_models.py +2 -2
  96. mteb/models/model_implementations/searchmap_models.py +1 -1
  97. mteb/models/model_implementations/sentence_transformers_models.py +58 -22
  98. mteb/models/model_implementations/shuu_model.py +1 -1
  99. mteb/models/model_implementations/siglip_models.py +10 -10
  100. mteb/models/model_implementations/slm_models.py +416 -0
  101. mteb/models/model_implementations/spartan8806_atles_champion.py +1 -1
  102. mteb/models/model_implementations/stella_models.py +17 -4
  103. mteb/models/model_implementations/tarka_models.py +2 -2
  104. mteb/models/model_implementations/text2vec_models.py +9 -3
  105. mteb/models/model_implementations/ua_sentence_models.py +1 -1
  106. mteb/models/model_implementations/uae_models.py +7 -1
  107. mteb/models/model_implementations/vdr_models.py +1 -1
  108. mteb/models/model_implementations/vi_vn_models.py +6 -6
  109. mteb/models/model_implementations/vlm2vec_models.py +2 -2
  110. mteb/models/model_implementations/youtu_models.py +1 -1
  111. mteb/models/model_implementations/yuan_models.py +1 -1
  112. mteb/models/model_implementations/yuan_models_en.py +1 -1
  113. mteb/models/model_meta.py +46 -17
  114. mteb/results/benchmark_results.py +2 -2
  115. mteb/tasks/classification/kur/kurdish_sentiment_classification.py +2 -2
  116. mteb/tasks/clustering/eng/hume_wiki_cities_clustering.py +1 -1
  117. mteb/tasks/clustering/eng/wiki_cities_clustering.py +1 -1
  118. mteb/tasks/clustering/zho/cmteb_clustering.py +2 -2
  119. mteb/tasks/reranking/multilingual/wikipedia_reranking_multilingual.py +1 -1
  120. mteb/tasks/retrieval/eng/cub200_i2i_retrieval.py +1 -1
  121. {mteb-2.6.4.dist-info → mteb-2.6.5.dist-info}/METADATA +3 -1
  122. {mteb-2.6.4.dist-info → mteb-2.6.5.dist-info}/RECORD +126 -125
  123. {mteb-2.6.4.dist-info → mteb-2.6.5.dist-info}/WHEEL +0 -0
  124. {mteb-2.6.4.dist-info → mteb-2.6.5.dist-info}/entry_points.txt +0 -0
  125. {mteb-2.6.4.dist-info → mteb-2.6.5.dist-info}/licenses/LICENSE +0 -0
  126. {mteb-2.6.4.dist-info → mteb-2.6.5.dist-info}/top_level.txt +0 -0
@@ -16,7 +16,7 @@ Euler_Legal_Embedding_V1 = ModelMeta(
16
16
  open_weights=True,
17
17
  public_training_code=None,
18
18
  public_training_data=None,
19
- framework=["PyTorch", "Sentence Transformers"],
19
+ framework=["PyTorch", "Sentence Transformers", "safetensors", "Transformers"],
20
20
  reference="https://huggingface.co/Mira190/Euler-Legal-Embedding-V1",
21
21
  similarity_fn_name="cosine",
22
22
  use_instructions=False,
@@ -24,7 +24,7 @@ Euler_Legal_Embedding_V1 = ModelMeta(
24
24
  adapted_from="Qwen/Qwen3-Embedding-8B",
25
25
  superseded_by=None,
26
26
  citation="""@misc{euler2025legal,
27
- title={Euler-Legal-Embedding: Advanced Legal Representation Learning},
27
+ title={Euler-Legal-Embedding: Advanced Legal Representation Learning},
28
28
  author={LawRank Team},
29
29
  year={2025},
30
30
  publisher={Hugging Face}
@@ -18,7 +18,7 @@ parsbert = ModelMeta(
18
18
  max_tokens=512,
19
19
  reference="https://huggingface.co/HooshvareLab/bert-base-parsbert-uncased",
20
20
  similarity_fn_name=ScoringFunction.COSINE,
21
- framework=["Sentence Transformers", "PyTorch"],
21
+ framework=["Sentence Transformers", "PyTorch", "Transformers"],
22
22
  use_instructions=False,
23
23
  public_training_code=None,
24
24
  public_training_data=None,
@@ -54,7 +54,7 @@ bert_zwnj = ModelMeta(
54
54
  max_tokens=512,
55
55
  reference="https://huggingface.co/m3hrdadfi/bert-zwnj-wnli-mean-tokens",
56
56
  similarity_fn_name=ScoringFunction.COSINE,
57
- framework=["Sentence Transformers", "PyTorch"],
57
+ framework=["Sentence Transformers", "PyTorch", "Transformers"],
58
58
  use_instructions=False,
59
59
  public_training_code=None,
60
60
  public_training_data=None,
@@ -80,7 +80,7 @@ roberta_zwnj = ModelMeta(
80
80
  max_tokens=514,
81
81
  reference="https://huggingface.co/m3hrdadfi/roberta-zwnj-wnli-mean-tokens",
82
82
  similarity_fn_name=ScoringFunction.COSINE,
83
- framework=["Sentence Transformers", "PyTorch"],
83
+ framework=["Sentence Transformers", "PyTorch", "Transformers"],
84
84
  use_instructions=False,
85
85
  public_training_code=None,
86
86
  public_training_data=None,
@@ -105,7 +105,7 @@ sentence_transformer_parsbert = ModelMeta(
105
105
  max_tokens=512,
106
106
  reference="https://huggingface.co/myrkur/sentence-transformer-parsbert-fa",
107
107
  similarity_fn_name=ScoringFunction.COSINE,
108
- framework=["Sentence Transformers", "PyTorch"],
108
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
109
109
  use_instructions=False,
110
110
  public_training_code=None,
111
111
  public_training_data=None,
@@ -129,7 +129,7 @@ tooka_bert_base = ModelMeta(
129
129
  max_tokens=512,
130
130
  reference="https://huggingface.co/PartAI/TookaBERT-Base",
131
131
  similarity_fn_name=ScoringFunction.COSINE,
132
- framework=["Sentence Transformers", "PyTorch"],
132
+ framework=["Sentence Transformers", "PyTorch", "Transformers", "safetensors"],
133
133
  use_instructions=False,
134
134
  public_training_code=None,
135
135
  public_training_data=None,
@@ -156,7 +156,7 @@ tooka_sbert = ModelMeta(
156
156
  max_tokens=512,
157
157
  reference="https://huggingface.co/PartAI/Tooka-SBERT",
158
158
  similarity_fn_name=ScoringFunction.COSINE,
159
- framework=["Sentence Transformers", "PyTorch"],
159
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
160
160
  use_instructions=False,
161
161
  public_training_code=None,
162
162
  public_training_data=None,
@@ -187,7 +187,7 @@ fa_bert = ModelMeta(
187
187
  max_tokens=512,
188
188
  reference="https://huggingface.co/sbunlp/fabert",
189
189
  similarity_fn_name=ScoringFunction.COSINE,
190
- framework=["Sentence Transformers", "PyTorch"],
190
+ framework=["Sentence Transformers", "PyTorch", "Transformers", "safetensors"],
191
191
  use_instructions=False,
192
192
  public_training_code=None,
193
193
  public_training_data=None,
@@ -235,7 +235,7 @@ tooka_sbert_v2_small = ModelMeta(
235
235
  max_tokens=512,
236
236
  reference="https://huggingface.co/PartAI/Tooka-SBERT-V2-Small",
237
237
  similarity_fn_name="cosine",
238
- framework=["Sentence Transformers", "PyTorch"],
238
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
239
239
  use_instructions=False,
240
240
  public_training_code=None,
241
241
  public_training_data=None,
@@ -266,7 +266,7 @@ tooka_sbert_v2_large = ModelMeta(
266
266
  max_tokens=512,
267
267
  reference="https://huggingface.co/PartAI/Tooka-SBERT-V2-Large",
268
268
  similarity_fn_name="cosine",
269
- framework=["Sentence Transformers", "PyTorch"],
269
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
270
270
  use_instructions=False,
271
271
  public_training_code=None,
272
272
  public_training_data=None,
@@ -119,7 +119,13 @@ xlmr_base = ModelMeta(
119
119
  max_tokens=512,
120
120
  reference="https://huggingface.co/FacebookAI/xlm-roberta-base",
121
121
  similarity_fn_name=ScoringFunction.COSINE,
122
- framework=["Sentence Transformers", "PyTorch"],
122
+ framework=[
123
+ "Sentence Transformers",
124
+ "PyTorch",
125
+ "Transformers",
126
+ "ONNX",
127
+ "safetensors",
128
+ ],
123
129
  use_instructions=False,
124
130
  public_training_code=None,
125
131
  public_training_data=None,
@@ -163,7 +169,13 @@ xlmr_large = ModelMeta(
163
169
  max_tokens=512,
164
170
  reference="https://huggingface.co/FacebookAI/xlm-roberta-large",
165
171
  similarity_fn_name=ScoringFunction.COSINE,
166
- framework=["Sentence Transformers", "PyTorch"],
172
+ framework=[
173
+ "Sentence Transformers",
174
+ "PyTorch",
175
+ "Transformers",
176
+ "ONNX",
177
+ "safetensors",
178
+ ],
167
179
  use_instructions=False,
168
180
  public_training_code=None,
169
181
  public_training_data=None,
@@ -26,7 +26,7 @@ geoembedding = ModelMeta(
26
26
  max_tokens=32768,
27
27
  reference="https://huggingface.co/GeoGPT-Research-Project/GeoEmbedding",
28
28
  similarity_fn_name="cosine",
29
- framework=["Sentence Transformers", "PyTorch"],
29
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
30
30
  use_instructions=True,
31
31
  public_training_code=None,
32
32
  public_training_data=None,
@@ -360,7 +360,7 @@ gme_qwen2vl_2b = ModelMeta(
360
360
  max_tokens=32768,
361
361
  reference="https://huggingface.co/Alibaba-NLP/gme-Qwen2-VL-2B-Instruct",
362
362
  similarity_fn_name=ScoringFunction.COSINE,
363
- framework=["PyTorch"],
363
+ framework=["PyTorch", "Sentence Transformers", "safetensors", "Transformers"],
364
364
  use_instructions=True,
365
365
  public_training_code=None,
366
366
  public_training_data=None,
@@ -384,7 +384,7 @@ gme_qwen2vl_7b = ModelMeta(
384
384
  max_tokens=32768,
385
385
  reference="https://huggingface.co/Alibaba-NLP/gme-Qwen2-VL-7B-Instruct",
386
386
  similarity_fn_name=ScoringFunction.COSINE,
387
- framework=["PyTorch"],
387
+ framework=["PyTorch", "Sentence Transformers", "safetensors", "Transformers"],
388
388
  use_instructions=True,
389
389
  public_training_code=None,
390
390
  public_training_data=None,
@@ -270,7 +270,7 @@ embedding_gemma_300m = ModelMeta(
270
270
  max_tokens=2048,
271
271
  license="gemma",
272
272
  reference="https://ai.google.dev/gemma/docs/embeddinggemma/model_card",
273
- framework=["Sentence Transformers", "PyTorch"],
273
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
274
274
  use_instructions=True,
275
275
  public_training_code=None,
276
276
  public_training_data=None,
@@ -179,7 +179,7 @@ granite_vision_embedding = ModelMeta(
179
179
  open_weights=True,
180
180
  public_training_code=None,
181
181
  public_training_data=None,
182
- framework=["PyTorch"],
182
+ framework=["PyTorch", "Transformers", "safetensors"],
183
183
  reference="https://huggingface.co/ibm-granite/granite-vision-3.3-2b-embedding",
184
184
  similarity_fn_name="MaxSim",
185
185
  use_instructions=True,
@@ -50,7 +50,7 @@ gritlm7b = ModelMeta(
50
50
  max_tokens=32768,
51
51
  reference="https://huggingface.co/GritLM/GritLM-7B",
52
52
  similarity_fn_name=ScoringFunction.COSINE,
53
- framework=["GritLM", "PyTorch"],
53
+ framework=["GritLM", "PyTorch", "Transformers", "safetensors"],
54
54
  use_instructions=True,
55
55
  training_datasets=GRIT_LM_TRAINING_DATA,
56
56
  # section 3.1 "We finetune our final models from Mistral 7B [68] and Mixtral 8x7B [69] using adaptations of E5 [160] and the Tülu 2 data
@@ -79,7 +79,7 @@ gritlm8x7b = ModelMeta(
79
79
  max_tokens=32768,
80
80
  reference="https://huggingface.co/GritLM/GritLM-8x7B",
81
81
  similarity_fn_name=ScoringFunction.COSINE,
82
- framework=["GritLM", "PyTorch"],
82
+ framework=["GritLM", "PyTorch", "Transformers", "safetensors"],
83
83
  use_instructions=True,
84
84
  training_datasets=GRIT_LM_TRAINING_DATA,
85
85
  citation=GRITLM_CITATION,
@@ -53,7 +53,7 @@ gte_qwen2_7b_instruct = ModelMeta(
53
53
  license="apache-2.0",
54
54
  reference="https://huggingface.co/Alibaba-NLP/gte-Qwen2-7B-instruct",
55
55
  similarity_fn_name=ScoringFunction.COSINE,
56
- framework=["Sentence Transformers", "PyTorch"],
56
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
57
57
  use_instructions=True,
58
58
  citation=GTE_CITATION,
59
59
  public_training_code=None,
@@ -86,7 +86,7 @@ gte_qwen1_5_7b_instruct = ModelMeta(
86
86
  max_tokens=32_768,
87
87
  reference="https://huggingface.co/Alibaba-NLP/gte-Qwen1.5-7B-instruct",
88
88
  similarity_fn_name=ScoringFunction.COSINE,
89
- framework=["Sentence Transformers", "PyTorch"],
89
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
90
90
  use_instructions=True,
91
91
  public_training_code=None,
92
92
  public_training_data=None,
@@ -123,7 +123,7 @@ gte_qwen2_1_5b_instruct = ModelMeta(
123
123
  max_tokens=32_768,
124
124
  reference="https://huggingface.co/Alibaba-NLP/gte-Qwen2-1.5B-instruct",
125
125
  similarity_fn_name=ScoringFunction.COSINE,
126
- framework=["Sentence Transformers", "PyTorch"],
126
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
127
127
  use_instructions=True,
128
128
  public_training_code=None,
129
129
  public_training_data=None,
@@ -151,7 +151,7 @@ gte_small_zh = ModelMeta(
151
151
  max_tokens=512,
152
152
  reference="https://huggingface.co/thenlper/gte-small-zh",
153
153
  similarity_fn_name=ScoringFunction.COSINE,
154
- framework=["Sentence Transformers", "PyTorch"],
154
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
155
155
  use_instructions=False,
156
156
  public_training_code=None,
157
157
  public_training_data=None,
@@ -179,7 +179,7 @@ gte_base_zh = ModelMeta(
179
179
  max_tokens=512,
180
180
  reference="https://huggingface.co/thenlper/gte-base-zh",
181
181
  similarity_fn_name=ScoringFunction.COSINE,
182
- framework=["Sentence Transformers", "PyTorch"],
182
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
183
183
  use_instructions=False,
184
184
  public_training_code=None,
185
185
  public_training_data=None,
@@ -207,7 +207,7 @@ gte_large_zh = ModelMeta(
207
207
  max_tokens=512,
208
208
  reference="https://huggingface.co/thenlper/gte-large-zh",
209
209
  similarity_fn_name=ScoringFunction.COSINE,
210
- framework=["Sentence Transformers", "PyTorch"],
210
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
211
211
  use_instructions=False,
212
212
  public_training_code=None,
213
213
  public_training_data=None,
@@ -336,7 +336,7 @@ gte_multilingual_base = ModelMeta(
336
336
  max_tokens=8192,
337
337
  reference="https://huggingface.co/Alibaba-NLP/gte-multilingual-base",
338
338
  similarity_fn_name=ScoringFunction.COSINE,
339
- framework=["Sentence Transformers", "PyTorch"],
339
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
340
340
  use_instructions=False,
341
341
  public_training_code=None,
342
342
  public_training_data=None, # couldn't find
@@ -365,7 +365,13 @@ gte_modernbert_base = ModelMeta(
365
365
  max_tokens=8192,
366
366
  reference="https://huggingface.co/Alibaba-NLP/gte-modernbert-base",
367
367
  similarity_fn_name=ScoringFunction.COSINE,
368
- framework=["Sentence Transformers", "PyTorch"],
368
+ framework=[
369
+ "Sentence Transformers",
370
+ "PyTorch",
371
+ "Transformers",
372
+ "ONNX",
373
+ "safetensors",
374
+ ],
369
375
  use_instructions=False,
370
376
  public_training_code=None, # couldn't find
371
377
  public_training_data=None,
@@ -402,7 +408,13 @@ gte_base_en_v15 = ModelMeta(
402
408
  max_tokens=8192,
403
409
  reference="https://huggingface.co/Alibaba-NLP/gte-base-en-v1.5",
404
410
  similarity_fn_name=ScoringFunction.COSINE,
405
- framework=["Sentence Transformers", "PyTorch"],
411
+ framework=[
412
+ "Sentence Transformers",
413
+ "PyTorch",
414
+ "Transformers",
415
+ "ONNX",
416
+ "safetensors",
417
+ ],
406
418
  use_instructions=False,
407
419
  superseded_by=None,
408
420
  adapted_from=None,
@@ -410,21 +422,21 @@ gte_base_en_v15 = ModelMeta(
410
422
  public_training_data=None,
411
423
  training_datasets=None,
412
424
  citation="""@misc{zhang2024mgte,
413
- title={mGTE: Generalized Long-Context Text Representation and Reranking Models for Multilingual Text Retrieval},
425
+ title={mGTE: Generalized Long-Context Text Representation and Reranking Models for Multilingual Text Retrieval},
414
426
  author={Xin Zhang and Yanzhao Zhang and Dingkun Long and Wen Xie and Ziqi Dai and Jialong Tang and Huan Lin and Baosong Yang and Pengjun Xie and Fei Huang and Meishan Zhang and Wenjie Li and Min Zhang},
415
427
  year={2024},
416
428
  eprint={2407.19669},
417
429
  archivePrefix={arXiv},
418
430
  primaryClass={cs.CL},
419
- url={https://arxiv.org/abs/2407.19669},
431
+ url={https://arxiv.org/abs/2407.19669},
420
432
  }
421
433
  @misc{li2023gte,
422
- title={Towards General Text Embeddings with Multi-stage Contrastive Learning},
434
+ title={Towards General Text Embeddings with Multi-stage Contrastive Learning},
423
435
  author={Zehan Li and Xin Zhang and Yanzhao Zhang and Dingkun Long and Pengjun Xie and Meishan Zhang},
424
436
  year={2023},
425
437
  eprint={2308.03281},
426
438
  archivePrefix={arXiv},
427
439
  primaryClass={cs.CL},
428
- url={https://arxiv.org/abs/2308.03281},
440
+ url={https://arxiv.org/abs/2308.03281},
429
441
  }""",
430
442
  )
@@ -49,7 +49,7 @@ Hinvec_bidir = ModelMeta(
49
49
  max_tokens=2048,
50
50
  reference="https://huggingface.co/Sailesh97/Hinvec",
51
51
  similarity_fn_name="cosine",
52
- framework=["Sentence Transformers", "PyTorch"],
52
+ framework=["Sentence Transformers", "PyTorch", "Transformers", "safetensors"],
53
53
  use_instructions=True,
54
54
  training_datasets=hinvec_training_datasets,
55
55
  public_training_code=None,
@@ -106,7 +106,13 @@ granite_107m_multilingual = ModelMeta(
106
106
  max_tokens=512,
107
107
  reference="https://huggingface.co/ibm-granite/granite-embedding-107m-multilingual",
108
108
  similarity_fn_name=ScoringFunction.COSINE,
109
- framework=["Sentence Transformers", "PyTorch"],
109
+ framework=[
110
+ "Sentence Transformers",
111
+ "PyTorch",
112
+ "Transformers",
113
+ "ONNX",
114
+ "safetensors",
115
+ ],
110
116
  adapted_from=None,
111
117
  superseded_by=None,
112
118
  public_training_code=None,
@@ -131,7 +137,13 @@ granite_278m_multilingual = ModelMeta(
131
137
  max_tokens=512,
132
138
  reference="https://huggingface.co/ibm-granite/granite-embedding-278m-multilingual",
133
139
  similarity_fn_name=ScoringFunction.COSINE,
134
- framework=["Sentence Transformers", "PyTorch"],
140
+ framework=[
141
+ "Sentence Transformers",
142
+ "PyTorch",
143
+ "Transformers",
144
+ "ONNX",
145
+ "safetensors",
146
+ ],
135
147
  adapted_from=None,
136
148
  superseded_by=None,
137
149
  public_training_code=None,
@@ -156,7 +168,13 @@ granite_30m_english = ModelMeta(
156
168
  max_tokens=512,
157
169
  reference="https://huggingface.co/ibm-granite/granite-embedding-30m-english",
158
170
  similarity_fn_name=ScoringFunction.COSINE,
159
- framework=["Sentence Transformers", "PyTorch"],
171
+ framework=[
172
+ "Sentence Transformers",
173
+ "PyTorch",
174
+ "ONNX",
175
+ "safetensors",
176
+ "Transformers",
177
+ ],
160
178
  adapted_from=None,
161
179
  superseded_by=None,
162
180
  public_training_code=None,
@@ -181,7 +199,13 @@ granite_125m_english = ModelMeta(
181
199
  max_tokens=512,
182
200
  reference="https://huggingface.co/ibm-granite/granite-embedding-125m-english",
183
201
  similarity_fn_name=ScoringFunction.COSINE,
184
- framework=["Sentence Transformers", "PyTorch"],
202
+ framework=[
203
+ "Sentence Transformers",
204
+ "PyTorch",
205
+ "ONNX",
206
+ "safetensors",
207
+ "Transformers",
208
+ ],
185
209
  adapted_from=None,
186
210
  superseded_by=None,
187
211
  public_training_code=None,
@@ -207,7 +231,7 @@ granite_english_r2 = ModelMeta(
207
231
  max_tokens=8192,
208
232
  reference="https://huggingface.co/ibm-granite/granite-embedding-english-r2",
209
233
  similarity_fn_name="cosine",
210
- framework=["Sentence Transformers", "PyTorch"],
234
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
211
235
  adapted_from=None,
212
236
  superseded_by=None,
213
237
  public_training_code=None,
@@ -232,7 +256,7 @@ granite_small_english_r2 = ModelMeta(
232
256
  max_tokens=8192,
233
257
  reference="https://huggingface.co/ibm-granite/granite-embedding-small-english-r2",
234
258
  similarity_fn_name="cosine",
235
- framework=["Sentence Transformers", "PyTorch"],
259
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
236
260
  adapted_from=None,
237
261
  superseded_by=None,
238
262
  public_training_code=None,
@@ -62,7 +62,7 @@ inf_retriever_v1 = ModelMeta(
62
62
  max_tokens=32768,
63
63
  reference="https://huggingface.co/infly/inf-retriever-v1",
64
64
  similarity_fn_name=ScoringFunction.COSINE,
65
- framework=["Sentence Transformers", "PyTorch"],
65
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
66
66
  use_instructions=True,
67
67
  adapted_from="Alibaba-NLP/gte-Qwen2-7B-instruct",
68
68
  public_training_code=None,
@@ -89,7 +89,7 @@ inf_retriever_v1_1_5b = ModelMeta(
89
89
  max_tokens=32768,
90
90
  reference="https://huggingface.co/infly/inf-retriever-v1-1.5b",
91
91
  similarity_fn_name=ScoringFunction.COSINE,
92
- framework=["Sentence Transformers", "PyTorch"],
92
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
93
93
  use_instructions=True,
94
94
  adapted_from="Alibaba-NLP/gte-Qwen2-1.5B-instruct",
95
95
  public_training_code=None,
@@ -298,7 +298,7 @@ jasper_en_v1 = ModelMeta(
298
298
  license="apache-2.0",
299
299
  reference="https://huggingface.co/infgrad/jasper_en_vision_language_v1",
300
300
  similarity_fn_name=ScoringFunction.COSINE,
301
- framework=["Sentence Transformers", "PyTorch"],
301
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
302
302
  use_instructions=True,
303
303
  adapted_from=None,
304
304
  superseded_by=None,
@@ -345,7 +345,7 @@ Jasper_Token_Compression_600M = ModelMeta(
345
345
  max_tokens=32768,
346
346
  reference="https://huggingface.co/infgrad/Jasper-Token-Compression-600M",
347
347
  similarity_fn_name="cosine",
348
- framework=["Sentence Transformers", "PyTorch"],
348
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
349
349
  use_instructions=True,
350
350
  public_training_code="https://github.com/DunZhang/Jasper-Token-Compression-Training",
351
351
  # public_training_data: unsupervised data for distillation
@@ -136,7 +136,7 @@ jina_clip_v1 = ModelMeta(
136
136
  open_weights=True,
137
137
  public_training_code=None,
138
138
  public_training_data=None,
139
- framework=["PyTorch"],
139
+ framework=["PyTorch", "Transformers", "ONNX", "safetensors"],
140
140
  reference="https://huggingface.co/jinaai/jina-clip-v1",
141
141
  similarity_fn_name=ScoringFunction.COSINE,
142
142
  use_instructions=True,
@@ -733,7 +733,7 @@ jina_reranker_v3 = ModelMeta(
733
733
  embed_dim=None,
734
734
  license="cc-by-nc-4.0",
735
735
  similarity_fn_name=None,
736
- framework=["PyTorch"],
736
+ framework=["PyTorch", "Transformers", "safetensors"],
737
737
  use_instructions=None,
738
738
  reference="https://huggingface.co/jinaai/jina-reranker-v3",
739
739
  public_training_code=None,
@@ -776,7 +776,7 @@ jina_embeddings_v4 = ModelMeta(
776
776
  embed_dim=2048,
777
777
  license="cc-by-nc-4.0",
778
778
  similarity_fn_name="cosine",
779
- framework=["Sentence Transformers", "PyTorch"],
779
+ framework=["Sentence Transformers", "PyTorch", "Transformers", "safetensors"],
780
780
  use_instructions=True,
781
781
  reference="https://huggingface.co/jinaai/jina-embeddings-v4",
782
782
  public_training_code=None,
@@ -824,7 +824,13 @@ jina_embeddings_v3 = ModelMeta(
824
824
  embed_dim=1024,
825
825
  license="cc-by-nc-4.0",
826
826
  similarity_fn_name=ScoringFunction.COSINE,
827
- framework=["Sentence Transformers", "PyTorch"],
827
+ framework=[
828
+ "Sentence Transformers",
829
+ "PyTorch",
830
+ "Transformers",
831
+ "ONNX",
832
+ "safetensors",
833
+ ],
828
834
  use_instructions=True,
829
835
  reference="https://huggingface.co/jinaai/jina-embeddings-v3",
830
836
  public_training_code=None,
@@ -879,7 +885,7 @@ jina_embeddings_v2_base_en = ModelMeta(
879
885
  max_tokens=8192,
880
886
  reference="https://huggingface.co/jinaai/jina-embeddings-v2-base-en",
881
887
  similarity_fn_name=ScoringFunction.COSINE,
882
- framework=["Sentence Transformers", "PyTorch"],
888
+ framework=["Sentence Transformers", "PyTorch", "ONNX", "safetensors"],
883
889
  use_instructions=False,
884
890
  superseded_by=None,
885
891
  adapted_from="jina-bert-base-en-v1", # pretrained on C4 with Alibi to support longer context.
@@ -943,7 +949,7 @@ jina_embeddings_v2_small_en = ModelMeta(
943
949
  max_tokens=8192,
944
950
  reference="https://huggingface.co/jinaai/jina-embeddings-v2-small-en",
945
951
  similarity_fn_name=ScoringFunction.COSINE,
946
- framework=["Sentence Transformers", "PyTorch"],
952
+ framework=["Sentence Transformers", "PyTorch", "ONNX", "safetensors"],
947
953
  use_instructions=False,
948
954
  superseded_by=None,
949
955
  adapted_from="jina-bert-smalll-en-v1", # pretrained on C4 with Alibi to support longer context
@@ -16,16 +16,22 @@ sbert_swedish = ModelMeta(
16
16
  max_tokens=384,
17
17
  reference="https://huggingface.co/KBLab/sentence-bert-swedish-cased",
18
18
  similarity_fn_name=ScoringFunction.COSINE,
19
- framework=["Sentence Transformers", "PyTorch"],
19
+ framework=[
20
+ "Sentence Transformers",
21
+ "PyTorch",
22
+ "safetensors",
23
+ "GGUF",
24
+ "Transformers",
25
+ ],
20
26
  use_instructions=False,
21
27
  public_training_code=None,
22
28
  public_training_data=None,
23
29
  training_datasets=None,
24
30
  adapted_from="sentence-transformers/all-mpnet-base-v2",
25
- citation="""@misc{rekathati2021introducing,
26
- author = {Rekathati, Faton},
27
- title = {The KBLab Blog: Introducing a Swedish Sentence Transformer},
28
- url = {https://kb-labb.github.io/posts/2021-08-23-a-swedish-sentence-transformer/},
29
- year = {2021}
31
+ citation="""@misc{rekathati2021introducing,
32
+ author = {Rekathati, Faton},
33
+ title = {The KBLab Blog: Introducing a Swedish Sentence Transformer},
34
+ url = {https://kb-labb.github.io/posts/2021-08-23-a-swedish-sentence-transformer/},
35
+ year = {2021}
30
36
  }""",
31
37
  )
@@ -18,7 +18,7 @@ dfm_enc_large = ModelMeta(
18
18
  max_tokens=512,
19
19
  reference="https://huggingface.co/KennethEnevoldsen/dfm-sentence-encoder-large",
20
20
  similarity_fn_name=ScoringFunction.COSINE,
21
- framework=["Sentence Transformers", "PyTorch"],
21
+ framework=["Sentence Transformers", "PyTorch", "Transformers", "safetensors"],
22
22
  use_instructions=False,
23
23
  superseded_by=None,
24
24
  adapted_from="chcaa/dfm-encoder-large-v1",
@@ -53,7 +53,7 @@ dfm_enc_med = ModelMeta(
53
53
  max_tokens=512,
54
54
  reference="https://huggingface.co/KennethEnevoldsen/dfm-sentence-encoder-medium",
55
55
  similarity_fn_name=ScoringFunction.COSINE,
56
- framework=["Sentence Transformers", "PyTorch"],
56
+ framework=["Sentence Transformers", "PyTorch", "Transformers"],
57
57
  use_instructions=False,
58
58
  superseded_by=None,
59
59
  adapted_from=None,
@@ -16,7 +16,7 @@ xlmr_scandi = ModelMeta(
16
16
  max_tokens=512,
17
17
  reference="https://huggingface.co/KFST/XLMRoberta-en-da-sv-nb",
18
18
  similarity_fn_name=ScoringFunction.COSINE,
19
- framework=["Sentence Transformers", "PyTorch"],
19
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
20
20
  use_instructions=False,
21
21
  public_training_code=None,
22
22
  public_training_data=None,
@@ -15,7 +15,7 @@ kowshik24_bangla_embedding_model = ModelMeta(
15
15
  max_tokens=128,
16
16
  reference="https://huggingface.co/Kowshik24/bangla-sentence-transformer-ft-matryoshka-paraphrase-multilingual-mpnet-base-v2",
17
17
  similarity_fn_name="cosine",
18
- framework=["Sentence Transformers", "PyTorch"],
18
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
19
19
  use_instructions=False,
20
20
  public_training_code="https://github.com/kowshik24/Bangla-Embedding",
21
21
  public_training_data="https://huggingface.co/datasets/sartajekram/BanglaRQA",
@@ -58,7 +58,7 @@ lgai_embedding_en = ModelMeta(
58
58
  max_tokens=32768,
59
59
  reference="https://huggingface.co/annamodels/LGAI-Embedding-Preview",
60
60
  similarity_fn_name="cosine",
61
- framework=["Sentence Transformers", "PyTorch"],
61
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
62
62
  use_instructions=True,
63
63
  public_training_code=None,
64
64
  public_training_data=None,
@@ -44,7 +44,7 @@ Linq_Embed_Mistral = ModelMeta(
44
44
  max_tokens=32768,
45
45
  reference="https://huggingface.co/Linq-AI-Research/Linq-Embed-Mistral",
46
46
  similarity_fn_name=ScoringFunction.COSINE,
47
- framework=["Sentence Transformers", "PyTorch"],
47
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
48
48
  use_instructions=True,
49
49
  public_training_code=None,
50
50
  public_training_data=None,
@@ -125,7 +125,7 @@ listconranker = ModelMeta(
125
125
  license="mit",
126
126
  max_tokens=512,
127
127
  reference="https://huggingface.co/ByteDance/ListConRanker",
128
- framework=["PyTorch"],
128
+ framework=["PyTorch", "Sentence Transformers", "safetensors", "Transformers"],
129
129
  use_instructions=False,
130
130
  public_training_code=None,
131
131
  public_training_data=None,
@@ -196,7 +196,7 @@ llm2clip_openai_l_14_336 = ModelMeta(
196
196
  open_weights=True,
197
197
  public_training_code="https://github.com/microsoft/LLM2CLIP",
198
198
  public_training_data=None,
199
- framework=["PyTorch"],
199
+ framework=["PyTorch", "safetensors"],
200
200
  reference="https://huggingface.co/microsoft/LLM2CLIP-Openai-L-14-336",
201
201
  similarity_fn_name=ScoringFunction.COSINE,
202
202
  use_instructions=True,
@@ -221,7 +221,7 @@ llm2clip_openai_l_14_224 = ModelMeta(
221
221
  open_weights=True,
222
222
  public_training_code="https://github.com/microsoft/LLM2CLIP",
223
223
  public_training_data=None,
224
- framework=["PyTorch"],
224
+ framework=["PyTorch", "safetensors"],
225
225
  reference="https://huggingface.co/microsoft/LLM2CLIP-Openai-L-14-224",
226
226
  similarity_fn_name=ScoringFunction.COSINE,
227
227
  use_instructions=True,
@@ -245,7 +245,7 @@ llm2clip_openai_b_16 = ModelMeta(
245
245
  open_weights=True,
246
246
  public_training_code="https://github.com/microsoft/LLM2CLIP",
247
247
  public_training_data=None,
248
- framework=["PyTorch"],
248
+ framework=["PyTorch", "safetensors"],
249
249
  reference="https://huggingface.co/microsoft/LLM2CLIP-Openai-B-16",
250
250
  similarity_fn_name=ScoringFunction.COSINE,
251
251
  use_instructions=True,