mteb 2.6.4__py3-none-any.whl → 2.6.6__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (142) hide show
  1. mteb/abstasks/classification.py +2 -3
  2. mteb/abstasks/multilabel_classification.py +3 -3
  3. mteb/abstasks/regression.py +1 -1
  4. mteb/abstasks/retrieval.py +1 -1
  5. mteb/abstasks/task_metadata.py +9 -14
  6. mteb/descriptive_stats/Retrieval/NanoClimateFEVER-VN.json +30 -0
  7. mteb/descriptive_stats/Retrieval/NanoDBPedia-VN.json +30 -0
  8. mteb/descriptive_stats/Retrieval/NanoFEVER-VN.json +30 -0
  9. mteb/descriptive_stats/Retrieval/NanoHotpotQA-VN.json +30 -0
  10. mteb/descriptive_stats/Retrieval/NanoMSMARCO-VN.json +30 -0
  11. mteb/descriptive_stats/Retrieval/NanoNQ-VN.json +30 -0
  12. mteb/descriptive_stats/Retrieval/TVPLRetrieval.json +30 -0
  13. mteb/models/model_implementations/align_models.py +1 -1
  14. mteb/models/model_implementations/andersborges.py +2 -2
  15. mteb/models/model_implementations/ara_models.py +1 -1
  16. mteb/models/model_implementations/arctic_models.py +8 -8
  17. mteb/models/model_implementations/b1ade_models.py +1 -1
  18. mteb/models/model_implementations/bge_models.py +45 -21
  19. mteb/models/model_implementations/bica_model.py +3 -3
  20. mteb/models/model_implementations/blip2_models.py +2 -2
  21. mteb/models/model_implementations/blip_models.py +8 -8
  22. mteb/models/model_implementations/bmretriever_models.py +4 -4
  23. mteb/models/model_implementations/cadet_models.py +1 -1
  24. mteb/models/model_implementations/cde_models.py +2 -2
  25. mteb/models/model_implementations/clip_models.py +3 -3
  26. mteb/models/model_implementations/clips_models.py +3 -3
  27. mteb/models/model_implementations/codefuse_models.py +5 -5
  28. mteb/models/model_implementations/codesage_models.py +3 -3
  29. mteb/models/model_implementations/cohere_models.py +4 -4
  30. mteb/models/model_implementations/colpali_models.py +3 -3
  31. mteb/models/model_implementations/colqwen_models.py +8 -8
  32. mteb/models/model_implementations/colsmol_models.py +2 -2
  33. mteb/models/model_implementations/conan_models.py +1 -1
  34. mteb/models/model_implementations/dino_models.py +19 -19
  35. mteb/models/model_implementations/e5_instruct.py +23 -4
  36. mteb/models/model_implementations/e5_models.py +9 -9
  37. mteb/models/model_implementations/e5_v.py +1 -1
  38. mteb/models/model_implementations/eagerworks_models.py +1 -1
  39. mteb/models/model_implementations/emillykkejensen_models.py +3 -3
  40. mteb/models/model_implementations/en_code_retriever.py +1 -1
  41. mteb/models/model_implementations/euler_models.py +2 -2
  42. mteb/models/model_implementations/fa_models.py +9 -9
  43. mteb/models/model_implementations/facebookai.py +14 -2
  44. mteb/models/model_implementations/geogpt_models.py +1 -1
  45. mteb/models/model_implementations/gme_v_models.py +2 -2
  46. mteb/models/model_implementations/google_models.py +1 -1
  47. mteb/models/model_implementations/granite_vision_embedding_models.py +1 -1
  48. mteb/models/model_implementations/gritlm_models.py +2 -2
  49. mteb/models/model_implementations/gte_models.py +25 -13
  50. mteb/models/model_implementations/hinvec_models.py +1 -1
  51. mteb/models/model_implementations/ibm_granite_models.py +30 -6
  52. mteb/models/model_implementations/inf_models.py +2 -2
  53. mteb/models/model_implementations/jasper_models.py +2 -2
  54. mteb/models/model_implementations/jina_clip.py +1 -1
  55. mteb/models/model_implementations/jina_models.py +11 -5
  56. mteb/models/model_implementations/kblab.py +12 -6
  57. mteb/models/model_implementations/kennethenevoldsen_models.py +2 -2
  58. mteb/models/model_implementations/kfst.py +1 -1
  59. mteb/models/model_implementations/kowshik24_models.py +1 -1
  60. mteb/models/model_implementations/lgai_embedding_models.py +1 -1
  61. mteb/models/model_implementations/linq_models.py +1 -1
  62. mteb/models/model_implementations/listconranker.py +1 -1
  63. mteb/models/model_implementations/llm2clip_models.py +3 -3
  64. mteb/models/model_implementations/llm2vec_models.py +8 -8
  65. mteb/models/model_implementations/mdbr_models.py +14 -2
  66. mteb/models/model_implementations/misc_models.py +68 -68
  67. mteb/models/model_implementations/mme5_models.py +1 -1
  68. mteb/models/model_implementations/moco_models.py +2 -2
  69. mteb/models/model_implementations/mod_models.py +1 -1
  70. mteb/models/model_implementations/model2vec_models.py +13 -13
  71. mteb/models/model_implementations/moka_models.py +1 -1
  72. mteb/models/model_implementations/mxbai_models.py +16 -3
  73. mteb/models/model_implementations/nbailab.py +3 -3
  74. mteb/models/model_implementations/no_instruct_sentence_models.py +1 -1
  75. mteb/models/model_implementations/nomic_models.py +18 -6
  76. mteb/models/model_implementations/nomic_models_vision.py +1 -1
  77. mteb/models/model_implementations/nvidia_llama_nemoretriever_colemb.py +2 -2
  78. mteb/models/model_implementations/nvidia_models.py +3 -3
  79. mteb/models/model_implementations/octen_models.py +3 -3
  80. mteb/models/model_implementations/openclip_models.py +6 -6
  81. mteb/models/model_implementations/opensearch_neural_sparse_models.py +5 -5
  82. mteb/models/model_implementations/ops_moa_models.py +1 -1
  83. mteb/models/model_implementations/ordalietech_solon_embeddings_mini_beta_1_1.py +1 -1
  84. mteb/models/model_implementations/pawan_models.py +1 -1
  85. mteb/models/model_implementations/piccolo_models.py +1 -1
  86. mteb/models/model_implementations/promptriever_models.py +4 -4
  87. mteb/models/model_implementations/pylate_models.py +5 -5
  88. mteb/models/model_implementations/qodo_models.py +2 -2
  89. mteb/models/model_implementations/qtack_models.py +1 -1
  90. mteb/models/model_implementations/qwen3_models.py +3 -3
  91. mteb/models/model_implementations/qzhou_models.py +2 -2
  92. mteb/models/model_implementations/rasgaard_models.py +1 -1
  93. mteb/models/model_implementations/reasonir_model.py +1 -1
  94. mteb/models/model_implementations/repllama_models.py +1 -1
  95. mteb/models/model_implementations/rerankers_custom.py +9 -3
  96. mteb/models/model_implementations/rerankers_monot5_based.py +14 -14
  97. mteb/models/model_implementations/richinfoai_models.py +1 -1
  98. mteb/models/model_implementations/ru_sentence_models.py +20 -20
  99. mteb/models/model_implementations/ruri_models.py +10 -10
  100. mteb/models/model_implementations/salesforce_models.py +3 -3
  101. mteb/models/model_implementations/samilpwc_models.py +1 -1
  102. mteb/models/model_implementations/sarashina_embedding_models.py +2 -2
  103. mteb/models/model_implementations/searchmap_models.py +1 -1
  104. mteb/models/model_implementations/sentence_transformers_models.py +58 -22
  105. mteb/models/model_implementations/shuu_model.py +1 -1
  106. mteb/models/model_implementations/siglip_models.py +10 -10
  107. mteb/models/model_implementations/slm_models.py +416 -0
  108. mteb/models/model_implementations/spartan8806_atles_champion.py +1 -1
  109. mteb/models/model_implementations/stella_models.py +17 -4
  110. mteb/models/model_implementations/tarka_models.py +2 -2
  111. mteb/models/model_implementations/text2vec_models.py +9 -3
  112. mteb/models/model_implementations/ua_sentence_models.py +1 -1
  113. mteb/models/model_implementations/uae_models.py +7 -1
  114. mteb/models/model_implementations/vdr_models.py +1 -1
  115. mteb/models/model_implementations/vi_vn_models.py +6 -6
  116. mteb/models/model_implementations/vlm2vec_models.py +2 -2
  117. mteb/models/model_implementations/youtu_models.py +1 -1
  118. mteb/models/model_implementations/yuan_models.py +1 -1
  119. mteb/models/model_implementations/yuan_models_en.py +1 -1
  120. mteb/models/model_meta.py +46 -17
  121. mteb/results/benchmark_results.py +2 -2
  122. mteb/tasks/classification/kur/kurdish_sentiment_classification.py +2 -2
  123. mteb/tasks/clustering/eng/hume_wiki_cities_clustering.py +1 -1
  124. mteb/tasks/clustering/eng/wiki_cities_clustering.py +1 -1
  125. mteb/tasks/clustering/zho/cmteb_clustering.py +2 -2
  126. mteb/tasks/reranking/multilingual/wikipedia_reranking_multilingual.py +1 -1
  127. mteb/tasks/retrieval/eng/cub200_i2i_retrieval.py +1 -1
  128. mteb/tasks/retrieval/vie/__init__.py +14 -6
  129. mteb/tasks/retrieval/vie/climate_fevervn_retrieval.py +39 -0
  130. mteb/tasks/retrieval/vie/db_pedia_vn_retrieval.py +39 -0
  131. mteb/tasks/retrieval/vie/fevervn_retrieval.py +39 -0
  132. mteb/tasks/retrieval/vie/hotpot_qavn_retrieval.py +39 -0
  133. mteb/tasks/retrieval/vie/msmarcovn_retrieval.py +48 -0
  134. mteb/tasks/retrieval/vie/nqvn_retrieval.py +39 -0
  135. mteb/tasks/retrieval/vie/tvpl_retrieval.py +42 -0
  136. mteb/tasks/retrieval/vie/zac_legal_text_retrieval.py +15 -1
  137. {mteb-2.6.4.dist-info → mteb-2.6.6.dist-info}/METADATA +3 -3
  138. {mteb-2.6.4.dist-info → mteb-2.6.6.dist-info}/RECORD +142 -133
  139. {mteb-2.6.4.dist-info → mteb-2.6.6.dist-info}/WHEEL +0 -0
  140. {mteb-2.6.4.dist-info → mteb-2.6.6.dist-info}/entry_points.txt +0 -0
  141. {mteb-2.6.4.dist-info → mteb-2.6.6.dist-info}/licenses/LICENSE +0 -0
  142. {mteb-2.6.4.dist-info → mteb-2.6.6.dist-info}/top_level.txt +0 -0
@@ -53,7 +53,7 @@ gte_qwen2_7b_instruct = ModelMeta(
53
53
  license="apache-2.0",
54
54
  reference="https://huggingface.co/Alibaba-NLP/gte-Qwen2-7B-instruct",
55
55
  similarity_fn_name=ScoringFunction.COSINE,
56
- framework=["Sentence Transformers", "PyTorch"],
56
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
57
57
  use_instructions=True,
58
58
  citation=GTE_CITATION,
59
59
  public_training_code=None,
@@ -86,7 +86,7 @@ gte_qwen1_5_7b_instruct = ModelMeta(
86
86
  max_tokens=32_768,
87
87
  reference="https://huggingface.co/Alibaba-NLP/gte-Qwen1.5-7B-instruct",
88
88
  similarity_fn_name=ScoringFunction.COSINE,
89
- framework=["Sentence Transformers", "PyTorch"],
89
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
90
90
  use_instructions=True,
91
91
  public_training_code=None,
92
92
  public_training_data=None,
@@ -123,7 +123,7 @@ gte_qwen2_1_5b_instruct = ModelMeta(
123
123
  max_tokens=32_768,
124
124
  reference="https://huggingface.co/Alibaba-NLP/gte-Qwen2-1.5B-instruct",
125
125
  similarity_fn_name=ScoringFunction.COSINE,
126
- framework=["Sentence Transformers", "PyTorch"],
126
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
127
127
  use_instructions=True,
128
128
  public_training_code=None,
129
129
  public_training_data=None,
@@ -151,7 +151,7 @@ gte_small_zh = ModelMeta(
151
151
  max_tokens=512,
152
152
  reference="https://huggingface.co/thenlper/gte-small-zh",
153
153
  similarity_fn_name=ScoringFunction.COSINE,
154
- framework=["Sentence Transformers", "PyTorch"],
154
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
155
155
  use_instructions=False,
156
156
  public_training_code=None,
157
157
  public_training_data=None,
@@ -179,7 +179,7 @@ gte_base_zh = ModelMeta(
179
179
  max_tokens=512,
180
180
  reference="https://huggingface.co/thenlper/gte-base-zh",
181
181
  similarity_fn_name=ScoringFunction.COSINE,
182
- framework=["Sentence Transformers", "PyTorch"],
182
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
183
183
  use_instructions=False,
184
184
  public_training_code=None,
185
185
  public_training_data=None,
@@ -207,7 +207,7 @@ gte_large_zh = ModelMeta(
207
207
  max_tokens=512,
208
208
  reference="https://huggingface.co/thenlper/gte-large-zh",
209
209
  similarity_fn_name=ScoringFunction.COSINE,
210
- framework=["Sentence Transformers", "PyTorch"],
210
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
211
211
  use_instructions=False,
212
212
  public_training_code=None,
213
213
  public_training_data=None,
@@ -336,7 +336,7 @@ gte_multilingual_base = ModelMeta(
336
336
  max_tokens=8192,
337
337
  reference="https://huggingface.co/Alibaba-NLP/gte-multilingual-base",
338
338
  similarity_fn_name=ScoringFunction.COSINE,
339
- framework=["Sentence Transformers", "PyTorch"],
339
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
340
340
  use_instructions=False,
341
341
  public_training_code=None,
342
342
  public_training_data=None, # couldn't find
@@ -365,7 +365,13 @@ gte_modernbert_base = ModelMeta(
365
365
  max_tokens=8192,
366
366
  reference="https://huggingface.co/Alibaba-NLP/gte-modernbert-base",
367
367
  similarity_fn_name=ScoringFunction.COSINE,
368
- framework=["Sentence Transformers", "PyTorch"],
368
+ framework=[
369
+ "Sentence Transformers",
370
+ "PyTorch",
371
+ "Transformers",
372
+ "ONNX",
373
+ "safetensors",
374
+ ],
369
375
  use_instructions=False,
370
376
  public_training_code=None, # couldn't find
371
377
  public_training_data=None,
@@ -402,7 +408,13 @@ gte_base_en_v15 = ModelMeta(
402
408
  max_tokens=8192,
403
409
  reference="https://huggingface.co/Alibaba-NLP/gte-base-en-v1.5",
404
410
  similarity_fn_name=ScoringFunction.COSINE,
405
- framework=["Sentence Transformers", "PyTorch"],
411
+ framework=[
412
+ "Sentence Transformers",
413
+ "PyTorch",
414
+ "Transformers",
415
+ "ONNX",
416
+ "safetensors",
417
+ ],
406
418
  use_instructions=False,
407
419
  superseded_by=None,
408
420
  adapted_from=None,
@@ -410,21 +422,21 @@ gte_base_en_v15 = ModelMeta(
410
422
  public_training_data=None,
411
423
  training_datasets=None,
412
424
  citation="""@misc{zhang2024mgte,
413
- title={mGTE: Generalized Long-Context Text Representation and Reranking Models for Multilingual Text Retrieval},
425
+ title={mGTE: Generalized Long-Context Text Representation and Reranking Models for Multilingual Text Retrieval},
414
426
  author={Xin Zhang and Yanzhao Zhang and Dingkun Long and Wen Xie and Ziqi Dai and Jialong Tang and Huan Lin and Baosong Yang and Pengjun Xie and Fei Huang and Meishan Zhang and Wenjie Li and Min Zhang},
415
427
  year={2024},
416
428
  eprint={2407.19669},
417
429
  archivePrefix={arXiv},
418
430
  primaryClass={cs.CL},
419
- url={https://arxiv.org/abs/2407.19669},
431
+ url={https://arxiv.org/abs/2407.19669},
420
432
  }
421
433
  @misc{li2023gte,
422
- title={Towards General Text Embeddings with Multi-stage Contrastive Learning},
434
+ title={Towards General Text Embeddings with Multi-stage Contrastive Learning},
423
435
  author={Zehan Li and Xin Zhang and Yanzhao Zhang and Dingkun Long and Pengjun Xie and Meishan Zhang},
424
436
  year={2023},
425
437
  eprint={2308.03281},
426
438
  archivePrefix={arXiv},
427
439
  primaryClass={cs.CL},
428
- url={https://arxiv.org/abs/2308.03281},
440
+ url={https://arxiv.org/abs/2308.03281},
429
441
  }""",
430
442
  )
@@ -49,7 +49,7 @@ Hinvec_bidir = ModelMeta(
49
49
  max_tokens=2048,
50
50
  reference="https://huggingface.co/Sailesh97/Hinvec",
51
51
  similarity_fn_name="cosine",
52
- framework=["Sentence Transformers", "PyTorch"],
52
+ framework=["Sentence Transformers", "PyTorch", "Transformers", "safetensors"],
53
53
  use_instructions=True,
54
54
  training_datasets=hinvec_training_datasets,
55
55
  public_training_code=None,
@@ -106,7 +106,13 @@ granite_107m_multilingual = ModelMeta(
106
106
  max_tokens=512,
107
107
  reference="https://huggingface.co/ibm-granite/granite-embedding-107m-multilingual",
108
108
  similarity_fn_name=ScoringFunction.COSINE,
109
- framework=["Sentence Transformers", "PyTorch"],
109
+ framework=[
110
+ "Sentence Transformers",
111
+ "PyTorch",
112
+ "Transformers",
113
+ "ONNX",
114
+ "safetensors",
115
+ ],
110
116
  adapted_from=None,
111
117
  superseded_by=None,
112
118
  public_training_code=None,
@@ -131,7 +137,13 @@ granite_278m_multilingual = ModelMeta(
131
137
  max_tokens=512,
132
138
  reference="https://huggingface.co/ibm-granite/granite-embedding-278m-multilingual",
133
139
  similarity_fn_name=ScoringFunction.COSINE,
134
- framework=["Sentence Transformers", "PyTorch"],
140
+ framework=[
141
+ "Sentence Transformers",
142
+ "PyTorch",
143
+ "Transformers",
144
+ "ONNX",
145
+ "safetensors",
146
+ ],
135
147
  adapted_from=None,
136
148
  superseded_by=None,
137
149
  public_training_code=None,
@@ -156,7 +168,13 @@ granite_30m_english = ModelMeta(
156
168
  max_tokens=512,
157
169
  reference="https://huggingface.co/ibm-granite/granite-embedding-30m-english",
158
170
  similarity_fn_name=ScoringFunction.COSINE,
159
- framework=["Sentence Transformers", "PyTorch"],
171
+ framework=[
172
+ "Sentence Transformers",
173
+ "PyTorch",
174
+ "ONNX",
175
+ "safetensors",
176
+ "Transformers",
177
+ ],
160
178
  adapted_from=None,
161
179
  superseded_by=None,
162
180
  public_training_code=None,
@@ -181,7 +199,13 @@ granite_125m_english = ModelMeta(
181
199
  max_tokens=512,
182
200
  reference="https://huggingface.co/ibm-granite/granite-embedding-125m-english",
183
201
  similarity_fn_name=ScoringFunction.COSINE,
184
- framework=["Sentence Transformers", "PyTorch"],
202
+ framework=[
203
+ "Sentence Transformers",
204
+ "PyTorch",
205
+ "ONNX",
206
+ "safetensors",
207
+ "Transformers",
208
+ ],
185
209
  adapted_from=None,
186
210
  superseded_by=None,
187
211
  public_training_code=None,
@@ -207,7 +231,7 @@ granite_english_r2 = ModelMeta(
207
231
  max_tokens=8192,
208
232
  reference="https://huggingface.co/ibm-granite/granite-embedding-english-r2",
209
233
  similarity_fn_name="cosine",
210
- framework=["Sentence Transformers", "PyTorch"],
234
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
211
235
  adapted_from=None,
212
236
  superseded_by=None,
213
237
  public_training_code=None,
@@ -232,7 +256,7 @@ granite_small_english_r2 = ModelMeta(
232
256
  max_tokens=8192,
233
257
  reference="https://huggingface.co/ibm-granite/granite-embedding-small-english-r2",
234
258
  similarity_fn_name="cosine",
235
- framework=["Sentence Transformers", "PyTorch"],
259
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
236
260
  adapted_from=None,
237
261
  superseded_by=None,
238
262
  public_training_code=None,
@@ -62,7 +62,7 @@ inf_retriever_v1 = ModelMeta(
62
62
  max_tokens=32768,
63
63
  reference="https://huggingface.co/infly/inf-retriever-v1",
64
64
  similarity_fn_name=ScoringFunction.COSINE,
65
- framework=["Sentence Transformers", "PyTorch"],
65
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
66
66
  use_instructions=True,
67
67
  adapted_from="Alibaba-NLP/gte-Qwen2-7B-instruct",
68
68
  public_training_code=None,
@@ -89,7 +89,7 @@ inf_retriever_v1_1_5b = ModelMeta(
89
89
  max_tokens=32768,
90
90
  reference="https://huggingface.co/infly/inf-retriever-v1-1.5b",
91
91
  similarity_fn_name=ScoringFunction.COSINE,
92
- framework=["Sentence Transformers", "PyTorch"],
92
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
93
93
  use_instructions=True,
94
94
  adapted_from="Alibaba-NLP/gte-Qwen2-1.5B-instruct",
95
95
  public_training_code=None,
@@ -298,7 +298,7 @@ jasper_en_v1 = ModelMeta(
298
298
  license="apache-2.0",
299
299
  reference="https://huggingface.co/infgrad/jasper_en_vision_language_v1",
300
300
  similarity_fn_name=ScoringFunction.COSINE,
301
- framework=["Sentence Transformers", "PyTorch"],
301
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
302
302
  use_instructions=True,
303
303
  adapted_from=None,
304
304
  superseded_by=None,
@@ -345,7 +345,7 @@ Jasper_Token_Compression_600M = ModelMeta(
345
345
  max_tokens=32768,
346
346
  reference="https://huggingface.co/infgrad/Jasper-Token-Compression-600M",
347
347
  similarity_fn_name="cosine",
348
- framework=["Sentence Transformers", "PyTorch"],
348
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
349
349
  use_instructions=True,
350
350
  public_training_code="https://github.com/DunZhang/Jasper-Token-Compression-Training",
351
351
  # public_training_data: unsupervised data for distillation
@@ -136,7 +136,7 @@ jina_clip_v1 = ModelMeta(
136
136
  open_weights=True,
137
137
  public_training_code=None,
138
138
  public_training_data=None,
139
- framework=["PyTorch"],
139
+ framework=["PyTorch", "Transformers", "ONNX", "safetensors"],
140
140
  reference="https://huggingface.co/jinaai/jina-clip-v1",
141
141
  similarity_fn_name=ScoringFunction.COSINE,
142
142
  use_instructions=True,
@@ -733,7 +733,7 @@ jina_reranker_v3 = ModelMeta(
733
733
  embed_dim=None,
734
734
  license="cc-by-nc-4.0",
735
735
  similarity_fn_name=None,
736
- framework=["PyTorch"],
736
+ framework=["PyTorch", "Transformers", "safetensors"],
737
737
  use_instructions=None,
738
738
  reference="https://huggingface.co/jinaai/jina-reranker-v3",
739
739
  public_training_code=None,
@@ -776,7 +776,7 @@ jina_embeddings_v4 = ModelMeta(
776
776
  embed_dim=2048,
777
777
  license="cc-by-nc-4.0",
778
778
  similarity_fn_name="cosine",
779
- framework=["Sentence Transformers", "PyTorch"],
779
+ framework=["Sentence Transformers", "PyTorch", "Transformers", "safetensors"],
780
780
  use_instructions=True,
781
781
  reference="https://huggingface.co/jinaai/jina-embeddings-v4",
782
782
  public_training_code=None,
@@ -824,7 +824,13 @@ jina_embeddings_v3 = ModelMeta(
824
824
  embed_dim=1024,
825
825
  license="cc-by-nc-4.0",
826
826
  similarity_fn_name=ScoringFunction.COSINE,
827
- framework=["Sentence Transformers", "PyTorch"],
827
+ framework=[
828
+ "Sentence Transformers",
829
+ "PyTorch",
830
+ "Transformers",
831
+ "ONNX",
832
+ "safetensors",
833
+ ],
828
834
  use_instructions=True,
829
835
  reference="https://huggingface.co/jinaai/jina-embeddings-v3",
830
836
  public_training_code=None,
@@ -879,7 +885,7 @@ jina_embeddings_v2_base_en = ModelMeta(
879
885
  max_tokens=8192,
880
886
  reference="https://huggingface.co/jinaai/jina-embeddings-v2-base-en",
881
887
  similarity_fn_name=ScoringFunction.COSINE,
882
- framework=["Sentence Transformers", "PyTorch"],
888
+ framework=["Sentence Transformers", "PyTorch", "ONNX", "safetensors"],
883
889
  use_instructions=False,
884
890
  superseded_by=None,
885
891
  adapted_from="jina-bert-base-en-v1", # pretrained on C4 with Alibi to support longer context.
@@ -943,7 +949,7 @@ jina_embeddings_v2_small_en = ModelMeta(
943
949
  max_tokens=8192,
944
950
  reference="https://huggingface.co/jinaai/jina-embeddings-v2-small-en",
945
951
  similarity_fn_name=ScoringFunction.COSINE,
946
- framework=["Sentence Transformers", "PyTorch"],
952
+ framework=["Sentence Transformers", "PyTorch", "ONNX", "safetensors"],
947
953
  use_instructions=False,
948
954
  superseded_by=None,
949
955
  adapted_from="jina-bert-smalll-en-v1", # pretrained on C4 with Alibi to support longer context
@@ -16,16 +16,22 @@ sbert_swedish = ModelMeta(
16
16
  max_tokens=384,
17
17
  reference="https://huggingface.co/KBLab/sentence-bert-swedish-cased",
18
18
  similarity_fn_name=ScoringFunction.COSINE,
19
- framework=["Sentence Transformers", "PyTorch"],
19
+ framework=[
20
+ "Sentence Transformers",
21
+ "PyTorch",
22
+ "safetensors",
23
+ "GGUF",
24
+ "Transformers",
25
+ ],
20
26
  use_instructions=False,
21
27
  public_training_code=None,
22
28
  public_training_data=None,
23
29
  training_datasets=None,
24
30
  adapted_from="sentence-transformers/all-mpnet-base-v2",
25
- citation="""@misc{rekathati2021introducing,
26
- author = {Rekathati, Faton},
27
- title = {The KBLab Blog: Introducing a Swedish Sentence Transformer},
28
- url = {https://kb-labb.github.io/posts/2021-08-23-a-swedish-sentence-transformer/},
29
- year = {2021}
31
+ citation="""@misc{rekathati2021introducing,
32
+ author = {Rekathati, Faton},
33
+ title = {The KBLab Blog: Introducing a Swedish Sentence Transformer},
34
+ url = {https://kb-labb.github.io/posts/2021-08-23-a-swedish-sentence-transformer/},
35
+ year = {2021}
30
36
  }""",
31
37
  )
@@ -18,7 +18,7 @@ dfm_enc_large = ModelMeta(
18
18
  max_tokens=512,
19
19
  reference="https://huggingface.co/KennethEnevoldsen/dfm-sentence-encoder-large",
20
20
  similarity_fn_name=ScoringFunction.COSINE,
21
- framework=["Sentence Transformers", "PyTorch"],
21
+ framework=["Sentence Transformers", "PyTorch", "Transformers", "safetensors"],
22
22
  use_instructions=False,
23
23
  superseded_by=None,
24
24
  adapted_from="chcaa/dfm-encoder-large-v1",
@@ -53,7 +53,7 @@ dfm_enc_med = ModelMeta(
53
53
  max_tokens=512,
54
54
  reference="https://huggingface.co/KennethEnevoldsen/dfm-sentence-encoder-medium",
55
55
  similarity_fn_name=ScoringFunction.COSINE,
56
- framework=["Sentence Transformers", "PyTorch"],
56
+ framework=["Sentence Transformers", "PyTorch", "Transformers"],
57
57
  use_instructions=False,
58
58
  superseded_by=None,
59
59
  adapted_from=None,
@@ -16,7 +16,7 @@ xlmr_scandi = ModelMeta(
16
16
  max_tokens=512,
17
17
  reference="https://huggingface.co/KFST/XLMRoberta-en-da-sv-nb",
18
18
  similarity_fn_name=ScoringFunction.COSINE,
19
- framework=["Sentence Transformers", "PyTorch"],
19
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
20
20
  use_instructions=False,
21
21
  public_training_code=None,
22
22
  public_training_data=None,
@@ -15,7 +15,7 @@ kowshik24_bangla_embedding_model = ModelMeta(
15
15
  max_tokens=128,
16
16
  reference="https://huggingface.co/Kowshik24/bangla-sentence-transformer-ft-matryoshka-paraphrase-multilingual-mpnet-base-v2",
17
17
  similarity_fn_name="cosine",
18
- framework=["Sentence Transformers", "PyTorch"],
18
+ framework=["Sentence Transformers", "PyTorch", "safetensors"],
19
19
  use_instructions=False,
20
20
  public_training_code="https://github.com/kowshik24/Bangla-Embedding",
21
21
  public_training_data="https://huggingface.co/datasets/sartajekram/BanglaRQA",
@@ -58,7 +58,7 @@ lgai_embedding_en = ModelMeta(
58
58
  max_tokens=32768,
59
59
  reference="https://huggingface.co/annamodels/LGAI-Embedding-Preview",
60
60
  similarity_fn_name="cosine",
61
- framework=["Sentence Transformers", "PyTorch"],
61
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
62
62
  use_instructions=True,
63
63
  public_training_code=None,
64
64
  public_training_data=None,
@@ -44,7 +44,7 @@ Linq_Embed_Mistral = ModelMeta(
44
44
  max_tokens=32768,
45
45
  reference="https://huggingface.co/Linq-AI-Research/Linq-Embed-Mistral",
46
46
  similarity_fn_name=ScoringFunction.COSINE,
47
- framework=["Sentence Transformers", "PyTorch"],
47
+ framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
48
48
  use_instructions=True,
49
49
  public_training_code=None,
50
50
  public_training_data=None,
@@ -125,7 +125,7 @@ listconranker = ModelMeta(
125
125
  license="mit",
126
126
  max_tokens=512,
127
127
  reference="https://huggingface.co/ByteDance/ListConRanker",
128
- framework=["PyTorch"],
128
+ framework=["PyTorch", "Sentence Transformers", "safetensors", "Transformers"],
129
129
  use_instructions=False,
130
130
  public_training_code=None,
131
131
  public_training_data=None,
@@ -196,7 +196,7 @@ llm2clip_openai_l_14_336 = ModelMeta(
196
196
  open_weights=True,
197
197
  public_training_code="https://github.com/microsoft/LLM2CLIP",
198
198
  public_training_data=None,
199
- framework=["PyTorch"],
199
+ framework=["PyTorch", "safetensors"],
200
200
  reference="https://huggingface.co/microsoft/LLM2CLIP-Openai-L-14-336",
201
201
  similarity_fn_name=ScoringFunction.COSINE,
202
202
  use_instructions=True,
@@ -221,7 +221,7 @@ llm2clip_openai_l_14_224 = ModelMeta(
221
221
  open_weights=True,
222
222
  public_training_code="https://github.com/microsoft/LLM2CLIP",
223
223
  public_training_data=None,
224
- framework=["PyTorch"],
224
+ framework=["PyTorch", "safetensors"],
225
225
  reference="https://huggingface.co/microsoft/LLM2CLIP-Openai-L-14-224",
226
226
  similarity_fn_name=ScoringFunction.COSINE,
227
227
  use_instructions=True,
@@ -245,7 +245,7 @@ llm2clip_openai_b_16 = ModelMeta(
245
245
  open_weights=True,
246
246
  public_training_code="https://github.com/microsoft/LLM2CLIP",
247
247
  public_training_data=None,
248
- framework=["PyTorch"],
248
+ framework=["PyTorch", "safetensors"],
249
249
  reference="https://huggingface.co/microsoft/LLM2CLIP-Openai-B-16",
250
250
  similarity_fn_name=ScoringFunction.COSINE,
251
251
  use_instructions=True,
@@ -145,7 +145,7 @@ llm2vec_llama3_8b_supervised = ModelMeta(
145
145
  license="mit",
146
146
  reference="https://huggingface.co/McGill-NLP/LLM2Vec-Meta-Llama-3-8B-Instruct-mntp-supervised",
147
147
  similarity_fn_name=ScoringFunction.COSINE,
148
- framework=["LLM2Vec", "PyTorch"],
148
+ framework=["LLM2Vec", "PyTorch", "safetensors"],
149
149
  use_instructions=True,
150
150
  public_training_code="https://github.com/McGill-NLP/llm2vec/tree/250292a307428240d801fadd85825464e71c3277/train_configs",
151
151
  training_datasets=llm2vec_supervised_training_data,
@@ -174,7 +174,7 @@ llm2vec_llama3_8b_unsupervised = ModelMeta(
174
174
  license="mit",
175
175
  reference="https://huggingface.co/McGill-NLP/LLM2Vec-Meta-Llama-3-8B-Instruct-mntp-unsup-simcse",
176
176
  similarity_fn_name=ScoringFunction.COSINE,
177
- framework=["LLM2Vec", "PyTorch"],
177
+ framework=["LLM2Vec", "PyTorch", "safetensors"],
178
178
  use_instructions=True,
179
179
  citation=LLM2VEC_CITATION,
180
180
  public_training_code="https://github.com/McGill-NLP/llm2vec/tree/250292a307428240d801fadd85825464e71c3277/train_configs",
@@ -203,7 +203,7 @@ llm2vec_mistral7b_supervised = ModelMeta(
203
203
  license="mit",
204
204
  reference="https://huggingface.co/McGill-NLP/LLM2Vec-Mistral-7B-Instruct-v2-mntp-supervised",
205
205
  similarity_fn_name=ScoringFunction.COSINE,
206
- framework=["LLM2Vec", "PyTorch"],
206
+ framework=["LLM2Vec", "PyTorch", "safetensors"],
207
207
  use_instructions=True,
208
208
  citation=LLM2VEC_CITATION,
209
209
  public_training_code="https://github.com/McGill-NLP/llm2vec/tree/250292a307428240d801fadd85825464e71c3277/train_configs",
@@ -232,7 +232,7 @@ llm2vec_mistral7b_unsupervised = ModelMeta(
232
232
  license="mit",
233
233
  reference="https://huggingface.co/McGill-NLP/LLM2Vec-Mistral-7B-Instruct-v2-mntp-unsup-simcse",
234
234
  similarity_fn_name=ScoringFunction.COSINE,
235
- framework=["LLM2Vec", "PyTorch"],
235
+ framework=["LLM2Vec", "PyTorch", "safetensors"],
236
236
  use_instructions=True,
237
237
  citation=LLM2VEC_CITATION,
238
238
  public_training_code="https://github.com/McGill-NLP/llm2vec/tree/250292a307428240d801fadd85825464e71c3277/train_configs",
@@ -261,7 +261,7 @@ llm2vec_llama2_7b_supervised = ModelMeta(
261
261
  license="mit",
262
262
  reference="https://huggingface.co/McGill-NLP/LLM2Vec-Llama-2-7b-chat-hf-mntp-supervised",
263
263
  similarity_fn_name=ScoringFunction.COSINE,
264
- framework=["LLM2Vec", "PyTorch"],
264
+ framework=["LLM2Vec", "PyTorch", "safetensors"],
265
265
  use_instructions=True,
266
266
  citation=LLM2VEC_CITATION,
267
267
  public_training_code="https://github.com/McGill-NLP/llm2vec/tree/250292a307428240d801fadd85825464e71c3277/train_configs",
@@ -290,7 +290,7 @@ llm2vec_llama2_7b_unsupervised = ModelMeta(
290
290
  license="mit",
291
291
  reference="https://huggingface.co/McGill-NLP/LLM2Vec-Llama-2-7b-chat-hf-mntp-unsup-simcse",
292
292
  similarity_fn_name=ScoringFunction.COSINE,
293
- framework=["LLM2Vec", "PyTorch"],
293
+ framework=["LLM2Vec", "PyTorch", "safetensors"],
294
294
  use_instructions=True,
295
295
  public_training_code="https://github.com/McGill-NLP/llm2vec/tree/250292a307428240d801fadd85825464e71c3277/train_configs",
296
296
  training_datasets=set(),
@@ -319,7 +319,7 @@ llm2vec_sheared_llama_supervised = ModelMeta(
319
319
  license="mit",
320
320
  reference="https://huggingface.co/McGill-NLP/LLM2Vec-Sheared-LLaMA-mntp-supervised",
321
321
  similarity_fn_name=ScoringFunction.COSINE,
322
- framework=["LLM2Vec", "PyTorch"],
322
+ framework=["LLM2Vec", "PyTorch", "safetensors"],
323
323
  use_instructions=True,
324
324
  citation=LLM2VEC_CITATION,
325
325
  public_training_code="https://github.com/McGill-NLP/llm2vec/tree/250292a307428240d801fadd85825464e71c3277/train_configs",
@@ -348,7 +348,7 @@ llm2vec_sheared_llama_unsupervised = ModelMeta(
348
348
  license="mit",
349
349
  reference="https://huggingface.co/McGill-NLP/LLM2Vec-Sheared-LLaMA-mntp-unsup-simcse",
350
350
  similarity_fn_name=ScoringFunction.COSINE,
351
- framework=["LLM2Vec", "PyTorch"],
351
+ framework=["LLM2Vec", "PyTorch", "safetensors"],
352
352
  use_instructions=True,
353
353
  citation=LLM2VEC_CITATION,
354
354
  public_training_code="https://github.com/McGill-NLP/llm2vec/tree/250292a307428240d801fadd85825464e71c3277/train_configs",
@@ -35,7 +35,13 @@ mdbr_leaf_ir = ModelMeta(
35
35
  release_date="2025-08-27",
36
36
  languages=["eng-Latn"],
37
37
  open_weights=True,
38
- framework=["Sentence Transformers", "PyTorch"],
38
+ framework=[
39
+ "Sentence Transformers",
40
+ "PyTorch",
41
+ "ONNX",
42
+ "safetensors",
43
+ "Transformers",
44
+ ],
39
45
  n_parameters=22_861_056,
40
46
  memory_usage_mb=86,
41
47
  max_tokens=512,
@@ -63,7 +69,13 @@ mdbr_leaf_mt = ModelMeta(
63
69
  release_date="2025-08-27",
64
70
  languages=["eng-Latn"],
65
71
  open_weights=True,
66
- framework=["Sentence Transformers", "PyTorch"],
72
+ framework=[
73
+ "Sentence Transformers",
74
+ "PyTorch",
75
+ "ONNX",
76
+ "safetensors",
77
+ "Transformers",
78
+ ],
67
79
  n_parameters=22_958_592,
68
80
  memory_usage_mb=86,
69
81
  max_tokens=512,