mteb 2.6.4__py3-none-any.whl → 2.6.5__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- mteb/abstasks/classification.py +2 -3
- mteb/abstasks/multilabel_classification.py +3 -3
- mteb/abstasks/regression.py +1 -1
- mteb/abstasks/retrieval.py +1 -1
- mteb/abstasks/task_metadata.py +9 -14
- mteb/models/model_implementations/align_models.py +1 -1
- mteb/models/model_implementations/andersborges.py +2 -2
- mteb/models/model_implementations/ara_models.py +1 -1
- mteb/models/model_implementations/arctic_models.py +8 -8
- mteb/models/model_implementations/b1ade_models.py +1 -1
- mteb/models/model_implementations/bge_models.py +45 -21
- mteb/models/model_implementations/bica_model.py +3 -3
- mteb/models/model_implementations/blip2_models.py +2 -2
- mteb/models/model_implementations/blip_models.py +8 -8
- mteb/models/model_implementations/bmretriever_models.py +4 -4
- mteb/models/model_implementations/cadet_models.py +1 -1
- mteb/models/model_implementations/cde_models.py +2 -2
- mteb/models/model_implementations/clip_models.py +3 -3
- mteb/models/model_implementations/clips_models.py +3 -3
- mteb/models/model_implementations/codefuse_models.py +5 -5
- mteb/models/model_implementations/codesage_models.py +3 -3
- mteb/models/model_implementations/cohere_models.py +4 -4
- mteb/models/model_implementations/colpali_models.py +3 -3
- mteb/models/model_implementations/colqwen_models.py +8 -8
- mteb/models/model_implementations/colsmol_models.py +2 -2
- mteb/models/model_implementations/conan_models.py +1 -1
- mteb/models/model_implementations/dino_models.py +19 -19
- mteb/models/model_implementations/e5_instruct.py +23 -4
- mteb/models/model_implementations/e5_models.py +9 -9
- mteb/models/model_implementations/e5_v.py +1 -1
- mteb/models/model_implementations/eagerworks_models.py +1 -1
- mteb/models/model_implementations/emillykkejensen_models.py +3 -3
- mteb/models/model_implementations/en_code_retriever.py +1 -1
- mteb/models/model_implementations/euler_models.py +2 -2
- mteb/models/model_implementations/fa_models.py +9 -9
- mteb/models/model_implementations/facebookai.py +14 -2
- mteb/models/model_implementations/geogpt_models.py +1 -1
- mteb/models/model_implementations/gme_v_models.py +2 -2
- mteb/models/model_implementations/google_models.py +1 -1
- mteb/models/model_implementations/granite_vision_embedding_models.py +1 -1
- mteb/models/model_implementations/gritlm_models.py +2 -2
- mteb/models/model_implementations/gte_models.py +25 -13
- mteb/models/model_implementations/hinvec_models.py +1 -1
- mteb/models/model_implementations/ibm_granite_models.py +30 -6
- mteb/models/model_implementations/inf_models.py +2 -2
- mteb/models/model_implementations/jasper_models.py +2 -2
- mteb/models/model_implementations/jina_clip.py +1 -1
- mteb/models/model_implementations/jina_models.py +11 -5
- mteb/models/model_implementations/kblab.py +12 -6
- mteb/models/model_implementations/kennethenevoldsen_models.py +2 -2
- mteb/models/model_implementations/kfst.py +1 -1
- mteb/models/model_implementations/kowshik24_models.py +1 -1
- mteb/models/model_implementations/lgai_embedding_models.py +1 -1
- mteb/models/model_implementations/linq_models.py +1 -1
- mteb/models/model_implementations/listconranker.py +1 -1
- mteb/models/model_implementations/llm2clip_models.py +3 -3
- mteb/models/model_implementations/llm2vec_models.py +8 -8
- mteb/models/model_implementations/mdbr_models.py +14 -2
- mteb/models/model_implementations/misc_models.py +68 -68
- mteb/models/model_implementations/mme5_models.py +1 -1
- mteb/models/model_implementations/moco_models.py +2 -2
- mteb/models/model_implementations/mod_models.py +1 -1
- mteb/models/model_implementations/model2vec_models.py +13 -13
- mteb/models/model_implementations/moka_models.py +1 -1
- mteb/models/model_implementations/mxbai_models.py +16 -3
- mteb/models/model_implementations/nbailab.py +3 -3
- mteb/models/model_implementations/no_instruct_sentence_models.py +1 -1
- mteb/models/model_implementations/nomic_models.py +18 -6
- mteb/models/model_implementations/nomic_models_vision.py +1 -1
- mteb/models/model_implementations/nvidia_llama_nemoretriever_colemb.py +2 -2
- mteb/models/model_implementations/nvidia_models.py +3 -3
- mteb/models/model_implementations/octen_models.py +2 -2
- mteb/models/model_implementations/openclip_models.py +6 -6
- mteb/models/model_implementations/opensearch_neural_sparse_models.py +5 -5
- mteb/models/model_implementations/ops_moa_models.py +1 -1
- mteb/models/model_implementations/ordalietech_solon_embeddings_mini_beta_1_1.py +1 -1
- mteb/models/model_implementations/pawan_models.py +1 -1
- mteb/models/model_implementations/piccolo_models.py +1 -1
- mteb/models/model_implementations/promptriever_models.py +4 -4
- mteb/models/model_implementations/pylate_models.py +5 -5
- mteb/models/model_implementations/qodo_models.py +2 -2
- mteb/models/model_implementations/qtack_models.py +1 -1
- mteb/models/model_implementations/qwen3_models.py +3 -3
- mteb/models/model_implementations/qzhou_models.py +2 -2
- mteb/models/model_implementations/rasgaard_models.py +1 -1
- mteb/models/model_implementations/reasonir_model.py +1 -1
- mteb/models/model_implementations/repllama_models.py +1 -1
- mteb/models/model_implementations/rerankers_custom.py +9 -3
- mteb/models/model_implementations/rerankers_monot5_based.py +14 -14
- mteb/models/model_implementations/richinfoai_models.py +1 -1
- mteb/models/model_implementations/ru_sentence_models.py +20 -20
- mteb/models/model_implementations/ruri_models.py +10 -10
- mteb/models/model_implementations/salesforce_models.py +3 -3
- mteb/models/model_implementations/samilpwc_models.py +1 -1
- mteb/models/model_implementations/sarashina_embedding_models.py +2 -2
- mteb/models/model_implementations/searchmap_models.py +1 -1
- mteb/models/model_implementations/sentence_transformers_models.py +58 -22
- mteb/models/model_implementations/shuu_model.py +1 -1
- mteb/models/model_implementations/siglip_models.py +10 -10
- mteb/models/model_implementations/slm_models.py +416 -0
- mteb/models/model_implementations/spartan8806_atles_champion.py +1 -1
- mteb/models/model_implementations/stella_models.py +17 -4
- mteb/models/model_implementations/tarka_models.py +2 -2
- mteb/models/model_implementations/text2vec_models.py +9 -3
- mteb/models/model_implementations/ua_sentence_models.py +1 -1
- mteb/models/model_implementations/uae_models.py +7 -1
- mteb/models/model_implementations/vdr_models.py +1 -1
- mteb/models/model_implementations/vi_vn_models.py +6 -6
- mteb/models/model_implementations/vlm2vec_models.py +2 -2
- mteb/models/model_implementations/youtu_models.py +1 -1
- mteb/models/model_implementations/yuan_models.py +1 -1
- mteb/models/model_implementations/yuan_models_en.py +1 -1
- mteb/models/model_meta.py +46 -17
- mteb/results/benchmark_results.py +2 -2
- mteb/tasks/classification/kur/kurdish_sentiment_classification.py +2 -2
- mteb/tasks/clustering/eng/hume_wiki_cities_clustering.py +1 -1
- mteb/tasks/clustering/eng/wiki_cities_clustering.py +1 -1
- mteb/tasks/clustering/zho/cmteb_clustering.py +2 -2
- mteb/tasks/reranking/multilingual/wikipedia_reranking_multilingual.py +1 -1
- mteb/tasks/retrieval/eng/cub200_i2i_retrieval.py +1 -1
- {mteb-2.6.4.dist-info → mteb-2.6.5.dist-info}/METADATA +3 -1
- {mteb-2.6.4.dist-info → mteb-2.6.5.dist-info}/RECORD +126 -125
- {mteb-2.6.4.dist-info → mteb-2.6.5.dist-info}/WHEEL +0 -0
- {mteb-2.6.4.dist-info → mteb-2.6.5.dist-info}/entry_points.txt +0 -0
- {mteb-2.6.4.dist-info → mteb-2.6.5.dist-info}/licenses/LICENSE +0 -0
- {mteb-2.6.4.dist-info → mteb-2.6.5.dist-info}/top_level.txt +0 -0
|
@@ -16,7 +16,7 @@ Euler_Legal_Embedding_V1 = ModelMeta(
|
|
|
16
16
|
open_weights=True,
|
|
17
17
|
public_training_code=None,
|
|
18
18
|
public_training_data=None,
|
|
19
|
-
framework=["PyTorch", "Sentence Transformers"],
|
|
19
|
+
framework=["PyTorch", "Sentence Transformers", "safetensors", "Transformers"],
|
|
20
20
|
reference="https://huggingface.co/Mira190/Euler-Legal-Embedding-V1",
|
|
21
21
|
similarity_fn_name="cosine",
|
|
22
22
|
use_instructions=False,
|
|
@@ -24,7 +24,7 @@ Euler_Legal_Embedding_V1 = ModelMeta(
|
|
|
24
24
|
adapted_from="Qwen/Qwen3-Embedding-8B",
|
|
25
25
|
superseded_by=None,
|
|
26
26
|
citation="""@misc{euler2025legal,
|
|
27
|
-
title={Euler-Legal-Embedding: Advanced Legal Representation Learning},
|
|
27
|
+
title={Euler-Legal-Embedding: Advanced Legal Representation Learning},
|
|
28
28
|
author={LawRank Team},
|
|
29
29
|
year={2025},
|
|
30
30
|
publisher={Hugging Face}
|
|
@@ -18,7 +18,7 @@ parsbert = ModelMeta(
|
|
|
18
18
|
max_tokens=512,
|
|
19
19
|
reference="https://huggingface.co/HooshvareLab/bert-base-parsbert-uncased",
|
|
20
20
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
21
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
21
|
+
framework=["Sentence Transformers", "PyTorch", "Transformers"],
|
|
22
22
|
use_instructions=False,
|
|
23
23
|
public_training_code=None,
|
|
24
24
|
public_training_data=None,
|
|
@@ -54,7 +54,7 @@ bert_zwnj = ModelMeta(
|
|
|
54
54
|
max_tokens=512,
|
|
55
55
|
reference="https://huggingface.co/m3hrdadfi/bert-zwnj-wnli-mean-tokens",
|
|
56
56
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
57
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
57
|
+
framework=["Sentence Transformers", "PyTorch", "Transformers"],
|
|
58
58
|
use_instructions=False,
|
|
59
59
|
public_training_code=None,
|
|
60
60
|
public_training_data=None,
|
|
@@ -80,7 +80,7 @@ roberta_zwnj = ModelMeta(
|
|
|
80
80
|
max_tokens=514,
|
|
81
81
|
reference="https://huggingface.co/m3hrdadfi/roberta-zwnj-wnli-mean-tokens",
|
|
82
82
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
83
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
83
|
+
framework=["Sentence Transformers", "PyTorch", "Transformers"],
|
|
84
84
|
use_instructions=False,
|
|
85
85
|
public_training_code=None,
|
|
86
86
|
public_training_data=None,
|
|
@@ -105,7 +105,7 @@ sentence_transformer_parsbert = ModelMeta(
|
|
|
105
105
|
max_tokens=512,
|
|
106
106
|
reference="https://huggingface.co/myrkur/sentence-transformer-parsbert-fa",
|
|
107
107
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
108
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
108
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors"],
|
|
109
109
|
use_instructions=False,
|
|
110
110
|
public_training_code=None,
|
|
111
111
|
public_training_data=None,
|
|
@@ -129,7 +129,7 @@ tooka_bert_base = ModelMeta(
|
|
|
129
129
|
max_tokens=512,
|
|
130
130
|
reference="https://huggingface.co/PartAI/TookaBERT-Base",
|
|
131
131
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
132
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
132
|
+
framework=["Sentence Transformers", "PyTorch", "Transformers", "safetensors"],
|
|
133
133
|
use_instructions=False,
|
|
134
134
|
public_training_code=None,
|
|
135
135
|
public_training_data=None,
|
|
@@ -156,7 +156,7 @@ tooka_sbert = ModelMeta(
|
|
|
156
156
|
max_tokens=512,
|
|
157
157
|
reference="https://huggingface.co/PartAI/Tooka-SBERT",
|
|
158
158
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
159
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
159
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors"],
|
|
160
160
|
use_instructions=False,
|
|
161
161
|
public_training_code=None,
|
|
162
162
|
public_training_data=None,
|
|
@@ -187,7 +187,7 @@ fa_bert = ModelMeta(
|
|
|
187
187
|
max_tokens=512,
|
|
188
188
|
reference="https://huggingface.co/sbunlp/fabert",
|
|
189
189
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
190
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
190
|
+
framework=["Sentence Transformers", "PyTorch", "Transformers", "safetensors"],
|
|
191
191
|
use_instructions=False,
|
|
192
192
|
public_training_code=None,
|
|
193
193
|
public_training_data=None,
|
|
@@ -235,7 +235,7 @@ tooka_sbert_v2_small = ModelMeta(
|
|
|
235
235
|
max_tokens=512,
|
|
236
236
|
reference="https://huggingface.co/PartAI/Tooka-SBERT-V2-Small",
|
|
237
237
|
similarity_fn_name="cosine",
|
|
238
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
238
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors"],
|
|
239
239
|
use_instructions=False,
|
|
240
240
|
public_training_code=None,
|
|
241
241
|
public_training_data=None,
|
|
@@ -266,7 +266,7 @@ tooka_sbert_v2_large = ModelMeta(
|
|
|
266
266
|
max_tokens=512,
|
|
267
267
|
reference="https://huggingface.co/PartAI/Tooka-SBERT-V2-Large",
|
|
268
268
|
similarity_fn_name="cosine",
|
|
269
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
269
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors"],
|
|
270
270
|
use_instructions=False,
|
|
271
271
|
public_training_code=None,
|
|
272
272
|
public_training_data=None,
|
|
@@ -119,7 +119,13 @@ xlmr_base = ModelMeta(
|
|
|
119
119
|
max_tokens=512,
|
|
120
120
|
reference="https://huggingface.co/FacebookAI/xlm-roberta-base",
|
|
121
121
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
122
|
-
framework=[
|
|
122
|
+
framework=[
|
|
123
|
+
"Sentence Transformers",
|
|
124
|
+
"PyTorch",
|
|
125
|
+
"Transformers",
|
|
126
|
+
"ONNX",
|
|
127
|
+
"safetensors",
|
|
128
|
+
],
|
|
123
129
|
use_instructions=False,
|
|
124
130
|
public_training_code=None,
|
|
125
131
|
public_training_data=None,
|
|
@@ -163,7 +169,13 @@ xlmr_large = ModelMeta(
|
|
|
163
169
|
max_tokens=512,
|
|
164
170
|
reference="https://huggingface.co/FacebookAI/xlm-roberta-large",
|
|
165
171
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
166
|
-
framework=[
|
|
172
|
+
framework=[
|
|
173
|
+
"Sentence Transformers",
|
|
174
|
+
"PyTorch",
|
|
175
|
+
"Transformers",
|
|
176
|
+
"ONNX",
|
|
177
|
+
"safetensors",
|
|
178
|
+
],
|
|
167
179
|
use_instructions=False,
|
|
168
180
|
public_training_code=None,
|
|
169
181
|
public_training_data=None,
|
|
@@ -26,7 +26,7 @@ geoembedding = ModelMeta(
|
|
|
26
26
|
max_tokens=32768,
|
|
27
27
|
reference="https://huggingface.co/GeoGPT-Research-Project/GeoEmbedding",
|
|
28
28
|
similarity_fn_name="cosine",
|
|
29
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
29
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
|
|
30
30
|
use_instructions=True,
|
|
31
31
|
public_training_code=None,
|
|
32
32
|
public_training_data=None,
|
|
@@ -360,7 +360,7 @@ gme_qwen2vl_2b = ModelMeta(
|
|
|
360
360
|
max_tokens=32768,
|
|
361
361
|
reference="https://huggingface.co/Alibaba-NLP/gme-Qwen2-VL-2B-Instruct",
|
|
362
362
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
363
|
-
framework=["PyTorch"],
|
|
363
|
+
framework=["PyTorch", "Sentence Transformers", "safetensors", "Transformers"],
|
|
364
364
|
use_instructions=True,
|
|
365
365
|
public_training_code=None,
|
|
366
366
|
public_training_data=None,
|
|
@@ -384,7 +384,7 @@ gme_qwen2vl_7b = ModelMeta(
|
|
|
384
384
|
max_tokens=32768,
|
|
385
385
|
reference="https://huggingface.co/Alibaba-NLP/gme-Qwen2-VL-7B-Instruct",
|
|
386
386
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
387
|
-
framework=["PyTorch"],
|
|
387
|
+
framework=["PyTorch", "Sentence Transformers", "safetensors", "Transformers"],
|
|
388
388
|
use_instructions=True,
|
|
389
389
|
public_training_code=None,
|
|
390
390
|
public_training_data=None,
|
|
@@ -270,7 +270,7 @@ embedding_gemma_300m = ModelMeta(
|
|
|
270
270
|
max_tokens=2048,
|
|
271
271
|
license="gemma",
|
|
272
272
|
reference="https://ai.google.dev/gemma/docs/embeddinggemma/model_card",
|
|
273
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
273
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors"],
|
|
274
274
|
use_instructions=True,
|
|
275
275
|
public_training_code=None,
|
|
276
276
|
public_training_data=None,
|
|
@@ -179,7 +179,7 @@ granite_vision_embedding = ModelMeta(
|
|
|
179
179
|
open_weights=True,
|
|
180
180
|
public_training_code=None,
|
|
181
181
|
public_training_data=None,
|
|
182
|
-
framework=["PyTorch"],
|
|
182
|
+
framework=["PyTorch", "Transformers", "safetensors"],
|
|
183
183
|
reference="https://huggingface.co/ibm-granite/granite-vision-3.3-2b-embedding",
|
|
184
184
|
similarity_fn_name="MaxSim",
|
|
185
185
|
use_instructions=True,
|
|
@@ -50,7 +50,7 @@ gritlm7b = ModelMeta(
|
|
|
50
50
|
max_tokens=32768,
|
|
51
51
|
reference="https://huggingface.co/GritLM/GritLM-7B",
|
|
52
52
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
53
|
-
framework=["GritLM", "PyTorch"],
|
|
53
|
+
framework=["GritLM", "PyTorch", "Transformers", "safetensors"],
|
|
54
54
|
use_instructions=True,
|
|
55
55
|
training_datasets=GRIT_LM_TRAINING_DATA,
|
|
56
56
|
# section 3.1 "We finetune our final models from Mistral 7B [68] and Mixtral 8x7B [69] using adaptations of E5 [160] and the Tülu 2 data
|
|
@@ -79,7 +79,7 @@ gritlm8x7b = ModelMeta(
|
|
|
79
79
|
max_tokens=32768,
|
|
80
80
|
reference="https://huggingface.co/GritLM/GritLM-8x7B",
|
|
81
81
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
82
|
-
framework=["GritLM", "PyTorch"],
|
|
82
|
+
framework=["GritLM", "PyTorch", "Transformers", "safetensors"],
|
|
83
83
|
use_instructions=True,
|
|
84
84
|
training_datasets=GRIT_LM_TRAINING_DATA,
|
|
85
85
|
citation=GRITLM_CITATION,
|
|
@@ -53,7 +53,7 @@ gte_qwen2_7b_instruct = ModelMeta(
|
|
|
53
53
|
license="apache-2.0",
|
|
54
54
|
reference="https://huggingface.co/Alibaba-NLP/gte-Qwen2-7B-instruct",
|
|
55
55
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
56
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
56
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
|
|
57
57
|
use_instructions=True,
|
|
58
58
|
citation=GTE_CITATION,
|
|
59
59
|
public_training_code=None,
|
|
@@ -86,7 +86,7 @@ gte_qwen1_5_7b_instruct = ModelMeta(
|
|
|
86
86
|
max_tokens=32_768,
|
|
87
87
|
reference="https://huggingface.co/Alibaba-NLP/gte-Qwen1.5-7B-instruct",
|
|
88
88
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
89
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
89
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
|
|
90
90
|
use_instructions=True,
|
|
91
91
|
public_training_code=None,
|
|
92
92
|
public_training_data=None,
|
|
@@ -123,7 +123,7 @@ gte_qwen2_1_5b_instruct = ModelMeta(
|
|
|
123
123
|
max_tokens=32_768,
|
|
124
124
|
reference="https://huggingface.co/Alibaba-NLP/gte-Qwen2-1.5B-instruct",
|
|
125
125
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
126
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
126
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
|
|
127
127
|
use_instructions=True,
|
|
128
128
|
public_training_code=None,
|
|
129
129
|
public_training_data=None,
|
|
@@ -151,7 +151,7 @@ gte_small_zh = ModelMeta(
|
|
|
151
151
|
max_tokens=512,
|
|
152
152
|
reference="https://huggingface.co/thenlper/gte-small-zh",
|
|
153
153
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
154
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
154
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors"],
|
|
155
155
|
use_instructions=False,
|
|
156
156
|
public_training_code=None,
|
|
157
157
|
public_training_data=None,
|
|
@@ -179,7 +179,7 @@ gte_base_zh = ModelMeta(
|
|
|
179
179
|
max_tokens=512,
|
|
180
180
|
reference="https://huggingface.co/thenlper/gte-base-zh",
|
|
181
181
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
182
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
182
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors"],
|
|
183
183
|
use_instructions=False,
|
|
184
184
|
public_training_code=None,
|
|
185
185
|
public_training_data=None,
|
|
@@ -207,7 +207,7 @@ gte_large_zh = ModelMeta(
|
|
|
207
207
|
max_tokens=512,
|
|
208
208
|
reference="https://huggingface.co/thenlper/gte-large-zh",
|
|
209
209
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
210
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
210
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors"],
|
|
211
211
|
use_instructions=False,
|
|
212
212
|
public_training_code=None,
|
|
213
213
|
public_training_data=None,
|
|
@@ -336,7 +336,7 @@ gte_multilingual_base = ModelMeta(
|
|
|
336
336
|
max_tokens=8192,
|
|
337
337
|
reference="https://huggingface.co/Alibaba-NLP/gte-multilingual-base",
|
|
338
338
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
339
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
339
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
|
|
340
340
|
use_instructions=False,
|
|
341
341
|
public_training_code=None,
|
|
342
342
|
public_training_data=None, # couldn't find
|
|
@@ -365,7 +365,13 @@ gte_modernbert_base = ModelMeta(
|
|
|
365
365
|
max_tokens=8192,
|
|
366
366
|
reference="https://huggingface.co/Alibaba-NLP/gte-modernbert-base",
|
|
367
367
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
368
|
-
framework=[
|
|
368
|
+
framework=[
|
|
369
|
+
"Sentence Transformers",
|
|
370
|
+
"PyTorch",
|
|
371
|
+
"Transformers",
|
|
372
|
+
"ONNX",
|
|
373
|
+
"safetensors",
|
|
374
|
+
],
|
|
369
375
|
use_instructions=False,
|
|
370
376
|
public_training_code=None, # couldn't find
|
|
371
377
|
public_training_data=None,
|
|
@@ -402,7 +408,13 @@ gte_base_en_v15 = ModelMeta(
|
|
|
402
408
|
max_tokens=8192,
|
|
403
409
|
reference="https://huggingface.co/Alibaba-NLP/gte-base-en-v1.5",
|
|
404
410
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
405
|
-
framework=[
|
|
411
|
+
framework=[
|
|
412
|
+
"Sentence Transformers",
|
|
413
|
+
"PyTorch",
|
|
414
|
+
"Transformers",
|
|
415
|
+
"ONNX",
|
|
416
|
+
"safetensors",
|
|
417
|
+
],
|
|
406
418
|
use_instructions=False,
|
|
407
419
|
superseded_by=None,
|
|
408
420
|
adapted_from=None,
|
|
@@ -410,21 +422,21 @@ gte_base_en_v15 = ModelMeta(
|
|
|
410
422
|
public_training_data=None,
|
|
411
423
|
training_datasets=None,
|
|
412
424
|
citation="""@misc{zhang2024mgte,
|
|
413
|
-
title={mGTE: Generalized Long-Context Text Representation and Reranking Models for Multilingual Text Retrieval},
|
|
425
|
+
title={mGTE: Generalized Long-Context Text Representation and Reranking Models for Multilingual Text Retrieval},
|
|
414
426
|
author={Xin Zhang and Yanzhao Zhang and Dingkun Long and Wen Xie and Ziqi Dai and Jialong Tang and Huan Lin and Baosong Yang and Pengjun Xie and Fei Huang and Meishan Zhang and Wenjie Li and Min Zhang},
|
|
415
427
|
year={2024},
|
|
416
428
|
eprint={2407.19669},
|
|
417
429
|
archivePrefix={arXiv},
|
|
418
430
|
primaryClass={cs.CL},
|
|
419
|
-
url={https://arxiv.org/abs/2407.19669},
|
|
431
|
+
url={https://arxiv.org/abs/2407.19669},
|
|
420
432
|
}
|
|
421
433
|
@misc{li2023gte,
|
|
422
|
-
title={Towards General Text Embeddings with Multi-stage Contrastive Learning},
|
|
434
|
+
title={Towards General Text Embeddings with Multi-stage Contrastive Learning},
|
|
423
435
|
author={Zehan Li and Xin Zhang and Yanzhao Zhang and Dingkun Long and Pengjun Xie and Meishan Zhang},
|
|
424
436
|
year={2023},
|
|
425
437
|
eprint={2308.03281},
|
|
426
438
|
archivePrefix={arXiv},
|
|
427
439
|
primaryClass={cs.CL},
|
|
428
|
-
url={https://arxiv.org/abs/2308.03281},
|
|
440
|
+
url={https://arxiv.org/abs/2308.03281},
|
|
429
441
|
}""",
|
|
430
442
|
)
|
|
@@ -49,7 +49,7 @@ Hinvec_bidir = ModelMeta(
|
|
|
49
49
|
max_tokens=2048,
|
|
50
50
|
reference="https://huggingface.co/Sailesh97/Hinvec",
|
|
51
51
|
similarity_fn_name="cosine",
|
|
52
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
52
|
+
framework=["Sentence Transformers", "PyTorch", "Transformers", "safetensors"],
|
|
53
53
|
use_instructions=True,
|
|
54
54
|
training_datasets=hinvec_training_datasets,
|
|
55
55
|
public_training_code=None,
|
|
@@ -106,7 +106,13 @@ granite_107m_multilingual = ModelMeta(
|
|
|
106
106
|
max_tokens=512,
|
|
107
107
|
reference="https://huggingface.co/ibm-granite/granite-embedding-107m-multilingual",
|
|
108
108
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
109
|
-
framework=[
|
|
109
|
+
framework=[
|
|
110
|
+
"Sentence Transformers",
|
|
111
|
+
"PyTorch",
|
|
112
|
+
"Transformers",
|
|
113
|
+
"ONNX",
|
|
114
|
+
"safetensors",
|
|
115
|
+
],
|
|
110
116
|
adapted_from=None,
|
|
111
117
|
superseded_by=None,
|
|
112
118
|
public_training_code=None,
|
|
@@ -131,7 +137,13 @@ granite_278m_multilingual = ModelMeta(
|
|
|
131
137
|
max_tokens=512,
|
|
132
138
|
reference="https://huggingface.co/ibm-granite/granite-embedding-278m-multilingual",
|
|
133
139
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
134
|
-
framework=[
|
|
140
|
+
framework=[
|
|
141
|
+
"Sentence Transformers",
|
|
142
|
+
"PyTorch",
|
|
143
|
+
"Transformers",
|
|
144
|
+
"ONNX",
|
|
145
|
+
"safetensors",
|
|
146
|
+
],
|
|
135
147
|
adapted_from=None,
|
|
136
148
|
superseded_by=None,
|
|
137
149
|
public_training_code=None,
|
|
@@ -156,7 +168,13 @@ granite_30m_english = ModelMeta(
|
|
|
156
168
|
max_tokens=512,
|
|
157
169
|
reference="https://huggingface.co/ibm-granite/granite-embedding-30m-english",
|
|
158
170
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
159
|
-
framework=[
|
|
171
|
+
framework=[
|
|
172
|
+
"Sentence Transformers",
|
|
173
|
+
"PyTorch",
|
|
174
|
+
"ONNX",
|
|
175
|
+
"safetensors",
|
|
176
|
+
"Transformers",
|
|
177
|
+
],
|
|
160
178
|
adapted_from=None,
|
|
161
179
|
superseded_by=None,
|
|
162
180
|
public_training_code=None,
|
|
@@ -181,7 +199,13 @@ granite_125m_english = ModelMeta(
|
|
|
181
199
|
max_tokens=512,
|
|
182
200
|
reference="https://huggingface.co/ibm-granite/granite-embedding-125m-english",
|
|
183
201
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
184
|
-
framework=[
|
|
202
|
+
framework=[
|
|
203
|
+
"Sentence Transformers",
|
|
204
|
+
"PyTorch",
|
|
205
|
+
"ONNX",
|
|
206
|
+
"safetensors",
|
|
207
|
+
"Transformers",
|
|
208
|
+
],
|
|
185
209
|
adapted_from=None,
|
|
186
210
|
superseded_by=None,
|
|
187
211
|
public_training_code=None,
|
|
@@ -207,7 +231,7 @@ granite_english_r2 = ModelMeta(
|
|
|
207
231
|
max_tokens=8192,
|
|
208
232
|
reference="https://huggingface.co/ibm-granite/granite-embedding-english-r2",
|
|
209
233
|
similarity_fn_name="cosine",
|
|
210
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
234
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
|
|
211
235
|
adapted_from=None,
|
|
212
236
|
superseded_by=None,
|
|
213
237
|
public_training_code=None,
|
|
@@ -232,7 +256,7 @@ granite_small_english_r2 = ModelMeta(
|
|
|
232
256
|
max_tokens=8192,
|
|
233
257
|
reference="https://huggingface.co/ibm-granite/granite-embedding-small-english-r2",
|
|
234
258
|
similarity_fn_name="cosine",
|
|
235
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
259
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
|
|
236
260
|
adapted_from=None,
|
|
237
261
|
superseded_by=None,
|
|
238
262
|
public_training_code=None,
|
|
@@ -62,7 +62,7 @@ inf_retriever_v1 = ModelMeta(
|
|
|
62
62
|
max_tokens=32768,
|
|
63
63
|
reference="https://huggingface.co/infly/inf-retriever-v1",
|
|
64
64
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
65
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
65
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
|
|
66
66
|
use_instructions=True,
|
|
67
67
|
adapted_from="Alibaba-NLP/gte-Qwen2-7B-instruct",
|
|
68
68
|
public_training_code=None,
|
|
@@ -89,7 +89,7 @@ inf_retriever_v1_1_5b = ModelMeta(
|
|
|
89
89
|
max_tokens=32768,
|
|
90
90
|
reference="https://huggingface.co/infly/inf-retriever-v1-1.5b",
|
|
91
91
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
92
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
92
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
|
|
93
93
|
use_instructions=True,
|
|
94
94
|
adapted_from="Alibaba-NLP/gte-Qwen2-1.5B-instruct",
|
|
95
95
|
public_training_code=None,
|
|
@@ -298,7 +298,7 @@ jasper_en_v1 = ModelMeta(
|
|
|
298
298
|
license="apache-2.0",
|
|
299
299
|
reference="https://huggingface.co/infgrad/jasper_en_vision_language_v1",
|
|
300
300
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
301
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
301
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors"],
|
|
302
302
|
use_instructions=True,
|
|
303
303
|
adapted_from=None,
|
|
304
304
|
superseded_by=None,
|
|
@@ -345,7 +345,7 @@ Jasper_Token_Compression_600M = ModelMeta(
|
|
|
345
345
|
max_tokens=32768,
|
|
346
346
|
reference="https://huggingface.co/infgrad/Jasper-Token-Compression-600M",
|
|
347
347
|
similarity_fn_name="cosine",
|
|
348
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
348
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors"],
|
|
349
349
|
use_instructions=True,
|
|
350
350
|
public_training_code="https://github.com/DunZhang/Jasper-Token-Compression-Training",
|
|
351
351
|
# public_training_data: unsupervised data for distillation
|
|
@@ -136,7 +136,7 @@ jina_clip_v1 = ModelMeta(
|
|
|
136
136
|
open_weights=True,
|
|
137
137
|
public_training_code=None,
|
|
138
138
|
public_training_data=None,
|
|
139
|
-
framework=["PyTorch"],
|
|
139
|
+
framework=["PyTorch", "Transformers", "ONNX", "safetensors"],
|
|
140
140
|
reference="https://huggingface.co/jinaai/jina-clip-v1",
|
|
141
141
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
142
142
|
use_instructions=True,
|
|
@@ -733,7 +733,7 @@ jina_reranker_v3 = ModelMeta(
|
|
|
733
733
|
embed_dim=None,
|
|
734
734
|
license="cc-by-nc-4.0",
|
|
735
735
|
similarity_fn_name=None,
|
|
736
|
-
framework=["PyTorch"],
|
|
736
|
+
framework=["PyTorch", "Transformers", "safetensors"],
|
|
737
737
|
use_instructions=None,
|
|
738
738
|
reference="https://huggingface.co/jinaai/jina-reranker-v3",
|
|
739
739
|
public_training_code=None,
|
|
@@ -776,7 +776,7 @@ jina_embeddings_v4 = ModelMeta(
|
|
|
776
776
|
embed_dim=2048,
|
|
777
777
|
license="cc-by-nc-4.0",
|
|
778
778
|
similarity_fn_name="cosine",
|
|
779
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
779
|
+
framework=["Sentence Transformers", "PyTorch", "Transformers", "safetensors"],
|
|
780
780
|
use_instructions=True,
|
|
781
781
|
reference="https://huggingface.co/jinaai/jina-embeddings-v4",
|
|
782
782
|
public_training_code=None,
|
|
@@ -824,7 +824,13 @@ jina_embeddings_v3 = ModelMeta(
|
|
|
824
824
|
embed_dim=1024,
|
|
825
825
|
license="cc-by-nc-4.0",
|
|
826
826
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
827
|
-
framework=[
|
|
827
|
+
framework=[
|
|
828
|
+
"Sentence Transformers",
|
|
829
|
+
"PyTorch",
|
|
830
|
+
"Transformers",
|
|
831
|
+
"ONNX",
|
|
832
|
+
"safetensors",
|
|
833
|
+
],
|
|
828
834
|
use_instructions=True,
|
|
829
835
|
reference="https://huggingface.co/jinaai/jina-embeddings-v3",
|
|
830
836
|
public_training_code=None,
|
|
@@ -879,7 +885,7 @@ jina_embeddings_v2_base_en = ModelMeta(
|
|
|
879
885
|
max_tokens=8192,
|
|
880
886
|
reference="https://huggingface.co/jinaai/jina-embeddings-v2-base-en",
|
|
881
887
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
882
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
888
|
+
framework=["Sentence Transformers", "PyTorch", "ONNX", "safetensors"],
|
|
883
889
|
use_instructions=False,
|
|
884
890
|
superseded_by=None,
|
|
885
891
|
adapted_from="jina-bert-base-en-v1", # pretrained on C4 with Alibi to support longer context.
|
|
@@ -943,7 +949,7 @@ jina_embeddings_v2_small_en = ModelMeta(
|
|
|
943
949
|
max_tokens=8192,
|
|
944
950
|
reference="https://huggingface.co/jinaai/jina-embeddings-v2-small-en",
|
|
945
951
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
946
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
952
|
+
framework=["Sentence Transformers", "PyTorch", "ONNX", "safetensors"],
|
|
947
953
|
use_instructions=False,
|
|
948
954
|
superseded_by=None,
|
|
949
955
|
adapted_from="jina-bert-smalll-en-v1", # pretrained on C4 with Alibi to support longer context
|
|
@@ -16,16 +16,22 @@ sbert_swedish = ModelMeta(
|
|
|
16
16
|
max_tokens=384,
|
|
17
17
|
reference="https://huggingface.co/KBLab/sentence-bert-swedish-cased",
|
|
18
18
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
19
|
-
framework=[
|
|
19
|
+
framework=[
|
|
20
|
+
"Sentence Transformers",
|
|
21
|
+
"PyTorch",
|
|
22
|
+
"safetensors",
|
|
23
|
+
"GGUF",
|
|
24
|
+
"Transformers",
|
|
25
|
+
],
|
|
20
26
|
use_instructions=False,
|
|
21
27
|
public_training_code=None,
|
|
22
28
|
public_training_data=None,
|
|
23
29
|
training_datasets=None,
|
|
24
30
|
adapted_from="sentence-transformers/all-mpnet-base-v2",
|
|
25
|
-
citation="""@misc{rekathati2021introducing,
|
|
26
|
-
author = {Rekathati, Faton},
|
|
27
|
-
title = {The KBLab Blog: Introducing a Swedish Sentence Transformer},
|
|
28
|
-
url = {https://kb-labb.github.io/posts/2021-08-23-a-swedish-sentence-transformer/},
|
|
29
|
-
year = {2021}
|
|
31
|
+
citation="""@misc{rekathati2021introducing,
|
|
32
|
+
author = {Rekathati, Faton},
|
|
33
|
+
title = {The KBLab Blog: Introducing a Swedish Sentence Transformer},
|
|
34
|
+
url = {https://kb-labb.github.io/posts/2021-08-23-a-swedish-sentence-transformer/},
|
|
35
|
+
year = {2021}
|
|
30
36
|
}""",
|
|
31
37
|
)
|
|
@@ -18,7 +18,7 @@ dfm_enc_large = ModelMeta(
|
|
|
18
18
|
max_tokens=512,
|
|
19
19
|
reference="https://huggingface.co/KennethEnevoldsen/dfm-sentence-encoder-large",
|
|
20
20
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
21
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
21
|
+
framework=["Sentence Transformers", "PyTorch", "Transformers", "safetensors"],
|
|
22
22
|
use_instructions=False,
|
|
23
23
|
superseded_by=None,
|
|
24
24
|
adapted_from="chcaa/dfm-encoder-large-v1",
|
|
@@ -53,7 +53,7 @@ dfm_enc_med = ModelMeta(
|
|
|
53
53
|
max_tokens=512,
|
|
54
54
|
reference="https://huggingface.co/KennethEnevoldsen/dfm-sentence-encoder-medium",
|
|
55
55
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
56
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
56
|
+
framework=["Sentence Transformers", "PyTorch", "Transformers"],
|
|
57
57
|
use_instructions=False,
|
|
58
58
|
superseded_by=None,
|
|
59
59
|
adapted_from=None,
|
|
@@ -16,7 +16,7 @@ xlmr_scandi = ModelMeta(
|
|
|
16
16
|
max_tokens=512,
|
|
17
17
|
reference="https://huggingface.co/KFST/XLMRoberta-en-da-sv-nb",
|
|
18
18
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
19
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
19
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
|
|
20
20
|
use_instructions=False,
|
|
21
21
|
public_training_code=None,
|
|
22
22
|
public_training_data=None,
|
|
@@ -15,7 +15,7 @@ kowshik24_bangla_embedding_model = ModelMeta(
|
|
|
15
15
|
max_tokens=128,
|
|
16
16
|
reference="https://huggingface.co/Kowshik24/bangla-sentence-transformer-ft-matryoshka-paraphrase-multilingual-mpnet-base-v2",
|
|
17
17
|
similarity_fn_name="cosine",
|
|
18
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
18
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors"],
|
|
19
19
|
use_instructions=False,
|
|
20
20
|
public_training_code="https://github.com/kowshik24/Bangla-Embedding",
|
|
21
21
|
public_training_data="https://huggingface.co/datasets/sartajekram/BanglaRQA",
|
|
@@ -58,7 +58,7 @@ lgai_embedding_en = ModelMeta(
|
|
|
58
58
|
max_tokens=32768,
|
|
59
59
|
reference="https://huggingface.co/annamodels/LGAI-Embedding-Preview",
|
|
60
60
|
similarity_fn_name="cosine",
|
|
61
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
61
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
|
|
62
62
|
use_instructions=True,
|
|
63
63
|
public_training_code=None,
|
|
64
64
|
public_training_data=None,
|
|
@@ -44,7 +44,7 @@ Linq_Embed_Mistral = ModelMeta(
|
|
|
44
44
|
max_tokens=32768,
|
|
45
45
|
reference="https://huggingface.co/Linq-AI-Research/Linq-Embed-Mistral",
|
|
46
46
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
47
|
-
framework=["Sentence Transformers", "PyTorch"],
|
|
47
|
+
framework=["Sentence Transformers", "PyTorch", "safetensors", "Transformers"],
|
|
48
48
|
use_instructions=True,
|
|
49
49
|
public_training_code=None,
|
|
50
50
|
public_training_data=None,
|
|
@@ -125,7 +125,7 @@ listconranker = ModelMeta(
|
|
|
125
125
|
license="mit",
|
|
126
126
|
max_tokens=512,
|
|
127
127
|
reference="https://huggingface.co/ByteDance/ListConRanker",
|
|
128
|
-
framework=["PyTorch"],
|
|
128
|
+
framework=["PyTorch", "Sentence Transformers", "safetensors", "Transformers"],
|
|
129
129
|
use_instructions=False,
|
|
130
130
|
public_training_code=None,
|
|
131
131
|
public_training_data=None,
|
|
@@ -196,7 +196,7 @@ llm2clip_openai_l_14_336 = ModelMeta(
|
|
|
196
196
|
open_weights=True,
|
|
197
197
|
public_training_code="https://github.com/microsoft/LLM2CLIP",
|
|
198
198
|
public_training_data=None,
|
|
199
|
-
framework=["PyTorch"],
|
|
199
|
+
framework=["PyTorch", "safetensors"],
|
|
200
200
|
reference="https://huggingface.co/microsoft/LLM2CLIP-Openai-L-14-336",
|
|
201
201
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
202
202
|
use_instructions=True,
|
|
@@ -221,7 +221,7 @@ llm2clip_openai_l_14_224 = ModelMeta(
|
|
|
221
221
|
open_weights=True,
|
|
222
222
|
public_training_code="https://github.com/microsoft/LLM2CLIP",
|
|
223
223
|
public_training_data=None,
|
|
224
|
-
framework=["PyTorch"],
|
|
224
|
+
framework=["PyTorch", "safetensors"],
|
|
225
225
|
reference="https://huggingface.co/microsoft/LLM2CLIP-Openai-L-14-224",
|
|
226
226
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
227
227
|
use_instructions=True,
|
|
@@ -245,7 +245,7 @@ llm2clip_openai_b_16 = ModelMeta(
|
|
|
245
245
|
open_weights=True,
|
|
246
246
|
public_training_code="https://github.com/microsoft/LLM2CLIP",
|
|
247
247
|
public_training_data=None,
|
|
248
|
-
framework=["PyTorch"],
|
|
248
|
+
framework=["PyTorch", "safetensors"],
|
|
249
249
|
reference="https://huggingface.co/microsoft/LLM2CLIP-Openai-B-16",
|
|
250
250
|
similarity_fn_name=ScoringFunction.COSINE,
|
|
251
251
|
use_instructions=True,
|