xinference 1.7.1__py3-none-any.whl → 1.8.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of xinference might be problematic. Click here for more details.
- xinference/_version.py +3 -3
- xinference/client/restful/async_restful_client.py +8 -13
- xinference/client/restful/restful_client.py +6 -2
- xinference/core/chat_interface.py +6 -4
- xinference/core/media_interface.py +5 -0
- xinference/core/model.py +1 -5
- xinference/core/supervisor.py +117 -68
- xinference/core/worker.py +49 -37
- xinference/deploy/test/test_cmdline.py +2 -6
- xinference/model/audio/__init__.py +26 -23
- xinference/model/audio/chattts.py +3 -2
- xinference/model/audio/core.py +49 -98
- xinference/model/audio/cosyvoice.py +3 -2
- xinference/model/audio/custom.py +28 -73
- xinference/model/audio/f5tts.py +3 -2
- xinference/model/audio/f5tts_mlx.py +3 -2
- xinference/model/audio/fish_speech.py +3 -2
- xinference/model/audio/funasr.py +17 -4
- xinference/model/audio/kokoro.py +3 -2
- xinference/model/audio/megatts.py +3 -2
- xinference/model/audio/melotts.py +3 -2
- xinference/model/audio/model_spec.json +572 -171
- xinference/model/audio/utils.py +0 -6
- xinference/model/audio/whisper.py +3 -2
- xinference/model/audio/whisper_mlx.py +3 -2
- xinference/model/cache_manager.py +141 -0
- xinference/model/core.py +6 -49
- xinference/model/custom.py +174 -0
- xinference/model/embedding/__init__.py +67 -56
- xinference/model/embedding/cache_manager.py +35 -0
- xinference/model/embedding/core.py +104 -84
- xinference/model/embedding/custom.py +55 -78
- xinference/model/embedding/embed_family.py +80 -31
- xinference/model/embedding/flag/core.py +21 -5
- xinference/model/embedding/llama_cpp/__init__.py +0 -0
- xinference/model/embedding/llama_cpp/core.py +234 -0
- xinference/model/embedding/model_spec.json +968 -103
- xinference/model/embedding/sentence_transformers/core.py +30 -20
- xinference/model/embedding/vllm/core.py +11 -5
- xinference/model/flexible/__init__.py +8 -2
- xinference/model/flexible/core.py +26 -119
- xinference/model/flexible/custom.py +69 -0
- xinference/model/flexible/launchers/image_process_launcher.py +1 -0
- xinference/model/flexible/launchers/modelscope_launcher.py +5 -1
- xinference/model/flexible/launchers/transformers_launcher.py +15 -3
- xinference/model/flexible/launchers/yolo_launcher.py +5 -1
- xinference/model/image/__init__.py +20 -20
- xinference/model/image/cache_manager.py +62 -0
- xinference/model/image/core.py +70 -182
- xinference/model/image/custom.py +28 -72
- xinference/model/image/model_spec.json +402 -119
- xinference/model/image/ocr/got_ocr2.py +3 -2
- xinference/model/image/stable_diffusion/core.py +22 -7
- xinference/model/image/stable_diffusion/mlx.py +6 -6
- xinference/model/image/utils.py +2 -2
- xinference/model/llm/__init__.py +71 -94
- xinference/model/llm/cache_manager.py +292 -0
- xinference/model/llm/core.py +37 -111
- xinference/model/llm/custom.py +88 -0
- xinference/model/llm/llama_cpp/core.py +5 -7
- xinference/model/llm/llm_family.json +16260 -8151
- xinference/model/llm/llm_family.py +138 -839
- xinference/model/llm/lmdeploy/core.py +5 -7
- xinference/model/llm/memory.py +3 -4
- xinference/model/llm/mlx/core.py +6 -8
- xinference/model/llm/reasoning_parser.py +3 -1
- xinference/model/llm/sglang/core.py +32 -14
- xinference/model/llm/transformers/chatglm.py +3 -7
- xinference/model/llm/transformers/core.py +49 -27
- xinference/model/llm/transformers/deepseek_v2.py +2 -2
- xinference/model/llm/transformers/gemma3.py +2 -2
- xinference/model/llm/transformers/multimodal/cogagent.py +2 -2
- xinference/model/llm/transformers/multimodal/deepseek_vl2.py +2 -2
- xinference/model/llm/transformers/multimodal/gemma3.py +2 -2
- xinference/model/llm/transformers/multimodal/glm4_1v.py +167 -0
- xinference/model/llm/transformers/multimodal/glm4v.py +2 -2
- xinference/model/llm/transformers/multimodal/intern_vl.py +2 -2
- xinference/model/llm/transformers/multimodal/minicpmv26.py +3 -3
- xinference/model/llm/transformers/multimodal/ovis2.py +2 -2
- xinference/model/llm/transformers/multimodal/qwen-omni.py +2 -2
- xinference/model/llm/transformers/multimodal/qwen2_audio.py +2 -2
- xinference/model/llm/transformers/multimodal/qwen2_vl.py +2 -2
- xinference/model/llm/transformers/opt.py +3 -7
- xinference/model/llm/utils.py +34 -49
- xinference/model/llm/vllm/core.py +77 -27
- xinference/model/llm/vllm/xavier/engine.py +5 -3
- xinference/model/llm/vllm/xavier/scheduler.py +10 -6
- xinference/model/llm/vllm/xavier/transfer.py +1 -1
- xinference/model/rerank/__init__.py +26 -25
- xinference/model/rerank/core.py +47 -87
- xinference/model/rerank/custom.py +25 -71
- xinference/model/rerank/model_spec.json +158 -33
- xinference/model/rerank/utils.py +2 -2
- xinference/model/utils.py +115 -54
- xinference/model/video/__init__.py +13 -17
- xinference/model/video/core.py +44 -102
- xinference/model/video/diffusers.py +4 -3
- xinference/model/video/model_spec.json +90 -21
- xinference/types.py +5 -3
- xinference/web/ui/build/asset-manifest.json +3 -3
- xinference/web/ui/build/index.html +1 -1
- xinference/web/ui/build/static/js/main.7d24df53.js +3 -0
- xinference/web/ui/build/static/js/main.7d24df53.js.map +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/2704ff66a5f73ca78b341eb3edec60154369df9d87fbc8c6dd60121abc5e1b0a.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/607dfef23d33e6b594518c0c6434567639f24f356b877c80c60575184ec50ed0.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/9be3d56173aacc3efd0b497bcb13c4f6365de30069176ee9403b40e717542326.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/9f9dd6c32c78a222d07da5987ae902effe16bcf20aac00774acdccc4de3c9ff2.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/b2ab5ee972c60d15eb9abf5845705f8ab7e1d125d324d9a9b1bcae5d6fd7ffb2.json +1 -0
- xinference/web/ui/src/locales/en.json +0 -1
- xinference/web/ui/src/locales/ja.json +0 -1
- xinference/web/ui/src/locales/ko.json +0 -1
- xinference/web/ui/src/locales/zh.json +0 -1
- {xinference-1.7.1.dist-info → xinference-1.8.0.dist-info}/METADATA +9 -11
- {xinference-1.7.1.dist-info → xinference-1.8.0.dist-info}/RECORD +119 -119
- xinference/model/audio/model_spec_modelscope.json +0 -231
- xinference/model/embedding/model_spec_modelscope.json +0 -293
- xinference/model/embedding/utils.py +0 -18
- xinference/model/image/model_spec_modelscope.json +0 -375
- xinference/model/llm/llama_cpp/memory.py +0 -457
- xinference/model/llm/llm_family_csghub.json +0 -56
- xinference/model/llm/llm_family_modelscope.json +0 -8700
- xinference/model/llm/llm_family_openmind_hub.json +0 -1019
- xinference/model/rerank/model_spec_modelscope.json +0 -85
- xinference/model/video/model_spec_modelscope.json +0 -184
- xinference/web/ui/build/static/js/main.9b12b7f9.js +0 -3
- xinference/web/ui/build/static/js/main.9b12b7f9.js.map +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/1460361af6975e63576708039f1cb732faf9c672d97c494d4055fc6331460be0.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/4efd8dda58fda83ed9546bf2f587df67f8d98e639117bee2d9326a9a1d9bebb2.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/55b9fb40b57fa926e8f05f31c2f96467e76e5ad62f033dca97c03f9e8c4eb4fe.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/5b2dafe5aa9e1105e0244a2b6751807342fa86aa0144b4e84d947a1686102715.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/611fa2c6c53b66039991d06dfb0473b5ab37fc63b4564e0f6e1718523768a045.json +0 -1
- /xinference/web/ui/build/static/js/{main.9b12b7f9.js.LICENSE.txt → main.7d24df53.js.LICENSE.txt} +0 -0
- {xinference-1.7.1.dist-info → xinference-1.8.0.dist-info}/WHEEL +0 -0
- {xinference-1.7.1.dist-info → xinference-1.8.0.dist-info}/entry_points.txt +0 -0
- {xinference-1.7.1.dist-info → xinference-1.8.0.dist-info}/licenses/LICENSE +0 -0
- {xinference-1.7.1.dist-info → xinference-1.8.0.dist-info}/top_level.txt +0 -0
|
@@ -1,231 +0,0 @@
|
|
|
1
|
-
[
|
|
2
|
-
{
|
|
3
|
-
"model_name": "whisper-large-v3",
|
|
4
|
-
"model_family": "whisper",
|
|
5
|
-
"model_hub": "modelscope",
|
|
6
|
-
"model_id": "AI-ModelScope/whisper-large-v3",
|
|
7
|
-
"model_revision": "master",
|
|
8
|
-
"model_ability": ["audio2text"],
|
|
9
|
-
"multilingual": true
|
|
10
|
-
},
|
|
11
|
-
{
|
|
12
|
-
"model_name": "whisper-large-v3-turbo",
|
|
13
|
-
"model_family": "whisper",
|
|
14
|
-
"model_hub": "modelscope",
|
|
15
|
-
"model_id": "AI-ModelScope/whisper-large-v3-turbo",
|
|
16
|
-
"model_revision": "master",
|
|
17
|
-
"model_ability": ["audio2text"],
|
|
18
|
-
"multilingual": true
|
|
19
|
-
},
|
|
20
|
-
{
|
|
21
|
-
"model_name": "Belle-whisper-large-v3-zh",
|
|
22
|
-
"model_family": "whisper",
|
|
23
|
-
"model_hub": "modelscope",
|
|
24
|
-
"model_id": "Xorbits/Belle-whisper-large-v3-zh",
|
|
25
|
-
"model_revision": "master",
|
|
26
|
-
"model_ability": ["audio2text"],
|
|
27
|
-
"multilingual": false
|
|
28
|
-
},
|
|
29
|
-
{
|
|
30
|
-
"model_name": "SenseVoiceSmall",
|
|
31
|
-
"model_family": "funasr",
|
|
32
|
-
"model_hub": "modelscope",
|
|
33
|
-
"model_id": "iic/SenseVoiceSmall",
|
|
34
|
-
"model_revision": "master",
|
|
35
|
-
"model_ability": ["audio2text"],
|
|
36
|
-
"multilingual": true,
|
|
37
|
-
"default_model_config": {
|
|
38
|
-
"vad_model": "fsmn-vad",
|
|
39
|
-
"vad_kwargs": {
|
|
40
|
-
"max_single_segment_time": 30000
|
|
41
|
-
}
|
|
42
|
-
},
|
|
43
|
-
"default_transcription_config": {
|
|
44
|
-
"use_itn": true,
|
|
45
|
-
"batch_size_s": 60,
|
|
46
|
-
"merge_vad": true,
|
|
47
|
-
"merge_length_s": 15
|
|
48
|
-
}
|
|
49
|
-
},
|
|
50
|
-
{
|
|
51
|
-
"model_name": "paraformer-zh",
|
|
52
|
-
"model_family": "funasr",
|
|
53
|
-
"model_hub": "modelscope",
|
|
54
|
-
"model_id": "iic/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch",
|
|
55
|
-
"model_revision": "master",
|
|
56
|
-
"model_ability": ["audio2text"],
|
|
57
|
-
"multilingual": false,
|
|
58
|
-
"default_model_config": {
|
|
59
|
-
"vad_model": "fsmn-vad",
|
|
60
|
-
"punc_model": "ct-punc"
|
|
61
|
-
},
|
|
62
|
-
"default_transcription_config": {
|
|
63
|
-
"batch_size_s": 300
|
|
64
|
-
}
|
|
65
|
-
},
|
|
66
|
-
{
|
|
67
|
-
"model_name": "paraformer-zh-hotword",
|
|
68
|
-
"model_family": "funasr",
|
|
69
|
-
"model_hub": "modelscope",
|
|
70
|
-
"model_id": "iic/speech_paraformer-large-contextual_asr_nat-zh-cn-16k-common-vocab8404",
|
|
71
|
-
"model_revision": "master",
|
|
72
|
-
"model_ability": ["audio2text"],
|
|
73
|
-
"multilingual": false,
|
|
74
|
-
"default_model_config": {
|
|
75
|
-
"vad_model": "fsmn-vad",
|
|
76
|
-
"punc_model": "ct-punc"
|
|
77
|
-
},
|
|
78
|
-
"default_transcription_config": {
|
|
79
|
-
"hotword": "",
|
|
80
|
-
"batch_size_s": 300
|
|
81
|
-
}
|
|
82
|
-
},
|
|
83
|
-
{
|
|
84
|
-
"model_name": "paraformer-zh-long",
|
|
85
|
-
"model_family": "funasr",
|
|
86
|
-
"model_hub": "modelscope",
|
|
87
|
-
"model_id": "iic/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch",
|
|
88
|
-
"model_revision": "master",
|
|
89
|
-
"model_ability": ["audio2text"],
|
|
90
|
-
"multilingual": false,
|
|
91
|
-
"default_model_config": {
|
|
92
|
-
"vad_model": "fsmn-vad",
|
|
93
|
-
"punc_model": "ct-punc"
|
|
94
|
-
},
|
|
95
|
-
"default_transcription_config": {
|
|
96
|
-
"batch_size_s": 300
|
|
97
|
-
}
|
|
98
|
-
},
|
|
99
|
-
{
|
|
100
|
-
"model_name": "paraformer-zh-spk",
|
|
101
|
-
"model_family": "funasr",
|
|
102
|
-
"model_hub": "modelscope",
|
|
103
|
-
"model_id": "iic/speech_paraformer-large-vad-punc-spk_asr_nat-zh-cn",
|
|
104
|
-
"model_revision": "master",
|
|
105
|
-
"model_ability": ["audio2text"],
|
|
106
|
-
"multilingual": false,
|
|
107
|
-
"default_model_config": {
|
|
108
|
-
"vad_model": "fsmn-vad",
|
|
109
|
-
"punc_model": "ct-punc",
|
|
110
|
-
"spk_model":"cam++"
|
|
111
|
-
},
|
|
112
|
-
"default_transcription_config": {
|
|
113
|
-
"batch_size_s": 300
|
|
114
|
-
}
|
|
115
|
-
},
|
|
116
|
-
{
|
|
117
|
-
"model_name": "seaco-paraformer-zh",
|
|
118
|
-
"model_family": "funasr",
|
|
119
|
-
"model_hub": "modelscope",
|
|
120
|
-
"model_id": "iic/speech_seaco_paraformer_large_asr_nat-zh-cn-16k-common-vocab8404-pytorch",
|
|
121
|
-
"model_revision": "master",
|
|
122
|
-
"model_ability": ["audio2text"],
|
|
123
|
-
"multilingual": false,
|
|
124
|
-
"default_model_config": {
|
|
125
|
-
"vad_model": "fsmn-vad",
|
|
126
|
-
"punc_model": "ct-punc"
|
|
127
|
-
},
|
|
128
|
-
"default_transcription_config": {
|
|
129
|
-
"hotword": "",
|
|
130
|
-
"batch_size_s": 300
|
|
131
|
-
}
|
|
132
|
-
},
|
|
133
|
-
{
|
|
134
|
-
"model_name": "ChatTTS",
|
|
135
|
-
"model_family": "ChatTTS",
|
|
136
|
-
"model_hub": "modelscope",
|
|
137
|
-
"model_id": "AI-ModelScope/ChatTTS",
|
|
138
|
-
"model_revision": "master",
|
|
139
|
-
"model_ability": ["text2audio"],
|
|
140
|
-
"multilingual": true,
|
|
141
|
-
"virtualenv": {
|
|
142
|
-
"packages": [
|
|
143
|
-
"ChatTTS>=0.2.1",
|
|
144
|
-
"#system_torch#",
|
|
145
|
-
"#system_numpy#"
|
|
146
|
-
]
|
|
147
|
-
}
|
|
148
|
-
},
|
|
149
|
-
{
|
|
150
|
-
"model_name": "CosyVoice-300M",
|
|
151
|
-
"model_family": "CosyVoice",
|
|
152
|
-
"model_hub": "modelscope",
|
|
153
|
-
"model_id": "iic/CosyVoice-300M",
|
|
154
|
-
"model_revision": "master",
|
|
155
|
-
"model_ability": ["text2audio"],
|
|
156
|
-
"multilingual": true
|
|
157
|
-
},
|
|
158
|
-
{
|
|
159
|
-
"model_name": "CosyVoice-300M-SFT",
|
|
160
|
-
"model_family": "CosyVoice",
|
|
161
|
-
"model_hub": "modelscope",
|
|
162
|
-
"model_id": "iic/CosyVoice-300M-SFT",
|
|
163
|
-
"model_revision": "master",
|
|
164
|
-
"model_ability": ["text2audio"],
|
|
165
|
-
"multilingual": true
|
|
166
|
-
},
|
|
167
|
-
{
|
|
168
|
-
"model_name": "CosyVoice-300M-Instruct",
|
|
169
|
-
"model_family": "CosyVoice",
|
|
170
|
-
"model_hub": "modelscope",
|
|
171
|
-
"model_id": "iic/CosyVoice-300M-Instruct",
|
|
172
|
-
"model_revision": "master",
|
|
173
|
-
"model_ability": ["text2audio"],
|
|
174
|
-
"multilingual": true
|
|
175
|
-
},
|
|
176
|
-
{
|
|
177
|
-
"model_name": "CosyVoice2-0.5B",
|
|
178
|
-
"model_family": "CosyVoice",
|
|
179
|
-
"model_hub": "modelscope",
|
|
180
|
-
"model_id": "iic/CosyVoice2-0.5B",
|
|
181
|
-
"model_revision": "master",
|
|
182
|
-
"model_ability": ["text2audio"],
|
|
183
|
-
"multilingual": true,
|
|
184
|
-
"virtualenv": {
|
|
185
|
-
"packages": [
|
|
186
|
-
"librosa",
|
|
187
|
-
"tiktoken",
|
|
188
|
-
"lightning>=2.0.0",
|
|
189
|
-
"hydra-core>=1.3.2",
|
|
190
|
-
"inflect",
|
|
191
|
-
"conformer",
|
|
192
|
-
"diffusers==0.29.0",
|
|
193
|
-
"gdown",
|
|
194
|
-
"pyarrow",
|
|
195
|
-
"HyperPyYAML",
|
|
196
|
-
"onnxruntime>=1.16.0",
|
|
197
|
-
"pyworld>=0.3.4",
|
|
198
|
-
"WeTextProcessing<1.0.4",
|
|
199
|
-
"#system_numpy#",
|
|
200
|
-
"#system_torch#"
|
|
201
|
-
]
|
|
202
|
-
}
|
|
203
|
-
},
|
|
204
|
-
{
|
|
205
|
-
"model_name": "F5-TTS",
|
|
206
|
-
"model_family": "F5-TTS",
|
|
207
|
-
"model_hub": "modelscope",
|
|
208
|
-
"model_id": "SWivid/F5-TTS_Emilia-ZH-EN",
|
|
209
|
-
"model_revision": "master",
|
|
210
|
-
"model_ability": ["text2audio"],
|
|
211
|
-
"multilingual": true
|
|
212
|
-
},
|
|
213
|
-
{
|
|
214
|
-
"model_name": "Kokoro-82M",
|
|
215
|
-
"model_family": "Kokoro",
|
|
216
|
-
"model_hub": "modelscope",
|
|
217
|
-
"model_id": "AI-ModelScope/Kokoro-82M",
|
|
218
|
-
"model_revision": "master",
|
|
219
|
-
"model_ability": ["text2audio"],
|
|
220
|
-
"multilingual": true
|
|
221
|
-
},
|
|
222
|
-
{
|
|
223
|
-
"model_name": "MegaTTS3",
|
|
224
|
-
"model_family": "MegaTTS",
|
|
225
|
-
"model_hub": "modelscope",
|
|
226
|
-
"model_id": "ByteDance/MegaTTS3",
|
|
227
|
-
"model_revision": "master",
|
|
228
|
-
"model_ability": ["text2audio"],
|
|
229
|
-
"multilingual": true
|
|
230
|
-
}
|
|
231
|
-
]
|
|
@@ -1,293 +0,0 @@
|
|
|
1
|
-
[
|
|
2
|
-
{
|
|
3
|
-
"model_name": "bge-large-en",
|
|
4
|
-
"dimensions": 1024,
|
|
5
|
-
"max_tokens": 512,
|
|
6
|
-
"language": ["en"],
|
|
7
|
-
"model_id": "Xorbits/bge-large-en",
|
|
8
|
-
"model_revision": "v0.0.1",
|
|
9
|
-
"model_hub": "modelscope"
|
|
10
|
-
},
|
|
11
|
-
{
|
|
12
|
-
"model_name": "bge-base-en",
|
|
13
|
-
"dimensions": 768,
|
|
14
|
-
"max_tokens": 512,
|
|
15
|
-
"language": ["en"],
|
|
16
|
-
"model_id": "Xorbits/bge-base-en",
|
|
17
|
-
"model_revision": "v0.0.1",
|
|
18
|
-
"model_hub": "modelscope"
|
|
19
|
-
},
|
|
20
|
-
{
|
|
21
|
-
"model_name": "gte-large",
|
|
22
|
-
"dimensions": 1024,
|
|
23
|
-
"max_tokens": 512,
|
|
24
|
-
"language": ["en"],
|
|
25
|
-
"model_id": "Xorbits/gte-large",
|
|
26
|
-
"model_revision": "v0.0.1",
|
|
27
|
-
"model_hub": "modelscope"
|
|
28
|
-
},
|
|
29
|
-
{
|
|
30
|
-
"model_name": "gte-base",
|
|
31
|
-
"dimensions": 768,
|
|
32
|
-
"max_tokens": 512,
|
|
33
|
-
"language": ["en"],
|
|
34
|
-
"model_id": "Xorbits/gte-base",
|
|
35
|
-
"model_revision": "v0.0.1",
|
|
36
|
-
"model_hub": "modelscope"
|
|
37
|
-
},
|
|
38
|
-
{
|
|
39
|
-
"model_name": "e5-large-v2",
|
|
40
|
-
"dimensions": 1024,
|
|
41
|
-
"max_tokens": 512,
|
|
42
|
-
"language": ["en"],
|
|
43
|
-
"model_id": "Xorbits/e5-large-v2",
|
|
44
|
-
"model_revision": "v0.0.1",
|
|
45
|
-
"model_hub": "modelscope"
|
|
46
|
-
},
|
|
47
|
-
{
|
|
48
|
-
"model_name": "bge-large-zh",
|
|
49
|
-
"dimensions": 1024,
|
|
50
|
-
"max_tokens": 512,
|
|
51
|
-
"language": ["zh"],
|
|
52
|
-
"model_id": "Xorbits/bge-large-zh",
|
|
53
|
-
"model_revision": "v0.0.1",
|
|
54
|
-
"model_hub": "modelscope"
|
|
55
|
-
},
|
|
56
|
-
{
|
|
57
|
-
"model_name": "bge-large-zh-noinstruct",
|
|
58
|
-
"dimensions": 1024,
|
|
59
|
-
"max_tokens": 512,
|
|
60
|
-
"language": ["zh"],
|
|
61
|
-
"model_id": "Xorbits/bge-large-zh-noinstruct",
|
|
62
|
-
"model_revision": "v0.0.1",
|
|
63
|
-
"model_hub": "modelscope"
|
|
64
|
-
},
|
|
65
|
-
{
|
|
66
|
-
"model_name": "bge-base-zh",
|
|
67
|
-
"dimensions": 768,
|
|
68
|
-
"max_tokens": 512,
|
|
69
|
-
"language": ["zh"],
|
|
70
|
-
"model_id": "Xorbits/bge-base-zh",
|
|
71
|
-
"model_revision": "v0.0.2",
|
|
72
|
-
"model_hub": "modelscope"
|
|
73
|
-
},
|
|
74
|
-
{
|
|
75
|
-
"model_name": "multilingual-e5-large",
|
|
76
|
-
"dimensions": 1024,
|
|
77
|
-
"max_tokens": 514,
|
|
78
|
-
"language": ["zh"],
|
|
79
|
-
"model_id": "Xorbits/multilingual-e5-large",
|
|
80
|
-
"model_revision": "v0.0.1",
|
|
81
|
-
"model_hub": "modelscope"
|
|
82
|
-
},
|
|
83
|
-
{
|
|
84
|
-
"model_name": "bge-small-zh",
|
|
85
|
-
"dimensions": 512,
|
|
86
|
-
"max_tokens": 512,
|
|
87
|
-
"language": ["zh"],
|
|
88
|
-
"model_id": "Xorbits/bge-small-zh",
|
|
89
|
-
"model_revision": "v0.0.1",
|
|
90
|
-
"model_hub": "modelscope"
|
|
91
|
-
},
|
|
92
|
-
{
|
|
93
|
-
"model_name": "bge-small-zh-v1.5",
|
|
94
|
-
"dimensions": 512,
|
|
95
|
-
"max_tokens": 512,
|
|
96
|
-
"language": ["zh"],
|
|
97
|
-
"model_id": "Xorbits/bge-small-zh-v1.5",
|
|
98
|
-
"model_revision": "v0.0.2",
|
|
99
|
-
"model_hub": "modelscope"
|
|
100
|
-
},
|
|
101
|
-
{
|
|
102
|
-
"model_name": "bge-base-zh-v1.5",
|
|
103
|
-
"dimensions": 768,
|
|
104
|
-
"max_tokens": 512,
|
|
105
|
-
"language": ["zh"],
|
|
106
|
-
"model_id": "Xorbits/bge-base-zh-v1.5",
|
|
107
|
-
"model_revision": "v0.0.1",
|
|
108
|
-
"model_hub": "modelscope"
|
|
109
|
-
},
|
|
110
|
-
{
|
|
111
|
-
"model_name": "bge-large-zh-v1.5",
|
|
112
|
-
"dimensions": 1024,
|
|
113
|
-
"max_tokens": 512,
|
|
114
|
-
"language": ["zh"],
|
|
115
|
-
"model_id": "Xorbits/bge-large-zh-v1.5",
|
|
116
|
-
"model_revision": "v0.0.1",
|
|
117
|
-
"model_hub": "modelscope"
|
|
118
|
-
},
|
|
119
|
-
{
|
|
120
|
-
"model_name": "bge-small-en-v1.5",
|
|
121
|
-
"dimensions": 384,
|
|
122
|
-
"max_tokens": 512,
|
|
123
|
-
"language": ["en"],
|
|
124
|
-
"model_id": "Xorbits/bge-small-en-v1.5",
|
|
125
|
-
"model_revision": "v0.0.2",
|
|
126
|
-
"model_hub": "modelscope"
|
|
127
|
-
},
|
|
128
|
-
{
|
|
129
|
-
"model_name": "bge-base-en-v1.5",
|
|
130
|
-
"dimensions": 768,
|
|
131
|
-
"max_tokens": 512,
|
|
132
|
-
"language": ["en"],
|
|
133
|
-
"model_id": "Xorbits/bge-base-en-v1.5",
|
|
134
|
-
"model_revision": "v0.0.1",
|
|
135
|
-
"model_hub": "modelscope"
|
|
136
|
-
},
|
|
137
|
-
{
|
|
138
|
-
"model_name": "bge-large-en-v1.5",
|
|
139
|
-
"dimensions": 1024,
|
|
140
|
-
"max_tokens": 512,
|
|
141
|
-
"language": ["en"],
|
|
142
|
-
"model_id": "Xorbits/bge-large-en-v1.5",
|
|
143
|
-
"model_revision": "v0.0.1",
|
|
144
|
-
"model_hub": "modelscope"
|
|
145
|
-
},
|
|
146
|
-
{
|
|
147
|
-
"model_name": "jina-embeddings-v2-small-en",
|
|
148
|
-
"dimensions": 512,
|
|
149
|
-
"max_tokens": 8192,
|
|
150
|
-
"language": ["en"],
|
|
151
|
-
"model_id": "Xorbits/jina-embeddings-v2-small-en",
|
|
152
|
-
"model_revision": "v0.0.1",
|
|
153
|
-
"model_hub": "modelscope"
|
|
154
|
-
},
|
|
155
|
-
{
|
|
156
|
-
"model_name": "jina-embeddings-v2-base-en",
|
|
157
|
-
"dimensions": 512,
|
|
158
|
-
"max_tokens": 8192,
|
|
159
|
-
"language": ["en"],
|
|
160
|
-
"model_id": "Xorbits/jina-embeddings-v2-base-en",
|
|
161
|
-
"model_revision": "v0.0.1",
|
|
162
|
-
"model_hub": "modelscope"
|
|
163
|
-
},
|
|
164
|
-
{
|
|
165
|
-
"model_name": "jina-embeddings-v2-base-zh",
|
|
166
|
-
"dimensions": 768,
|
|
167
|
-
"max_tokens": 8192,
|
|
168
|
-
"language": ["zh", "en"],
|
|
169
|
-
"model_id": "jinaai/jina-embeddings-v2-base-zh",
|
|
170
|
-
"model_hub": "modelscope"
|
|
171
|
-
},
|
|
172
|
-
{
|
|
173
|
-
"model_name": "text2vec-large-chinese",
|
|
174
|
-
"dimensions": 1024,
|
|
175
|
-
"max_tokens": 256,
|
|
176
|
-
"language": ["zh"],
|
|
177
|
-
"model_id": "Jerry0/text2vec-large-chinese",
|
|
178
|
-
"model_hub": "modelscope"
|
|
179
|
-
},
|
|
180
|
-
{
|
|
181
|
-
"model_name": "text2vec-base-chinese",
|
|
182
|
-
"dimensions": 768,
|
|
183
|
-
"max_tokens": 128,
|
|
184
|
-
"language": ["zh"],
|
|
185
|
-
"model_id": "Jerry0/text2vec-base-chinese",
|
|
186
|
-
"model_hub": "modelscope"
|
|
187
|
-
},
|
|
188
|
-
{
|
|
189
|
-
"model_name": "text2vec-base-chinese-paraphrase",
|
|
190
|
-
"dimensions": 768,
|
|
191
|
-
"max_tokens": 256,
|
|
192
|
-
"language": ["zh"],
|
|
193
|
-
"model_id": "mwei23/text2vec-base-chinese-paraphrase",
|
|
194
|
-
"model_hub": "modelscope"
|
|
195
|
-
},
|
|
196
|
-
{
|
|
197
|
-
"model_name": "bge-m3",
|
|
198
|
-
"dimensions": 1024,
|
|
199
|
-
"max_tokens": 8192,
|
|
200
|
-
"language": ["zh", "en"],
|
|
201
|
-
"model_id": "Xorbits/bge-m3",
|
|
202
|
-
"model_hub": "modelscope"
|
|
203
|
-
},
|
|
204
|
-
{
|
|
205
|
-
"model_name": "bce-embedding-base_v1",
|
|
206
|
-
"dimensions": 768,
|
|
207
|
-
"max_tokens": 512,
|
|
208
|
-
"language": ["zh", "en"],
|
|
209
|
-
"model_id": "maidalun/bce-embedding-base_v1",
|
|
210
|
-
"model_hub": "modelscope"
|
|
211
|
-
},
|
|
212
|
-
{
|
|
213
|
-
"model_name": "m3e-small",
|
|
214
|
-
"dimensions": 512,
|
|
215
|
-
"max_tokens": 512,
|
|
216
|
-
"language": ["zh", "en"],
|
|
217
|
-
"model_id": "AI-ModelScope/m3e-small",
|
|
218
|
-
"model_hub": "modelscope"
|
|
219
|
-
},
|
|
220
|
-
{
|
|
221
|
-
"model_name": "m3e-base",
|
|
222
|
-
"dimensions": 768,
|
|
223
|
-
"max_tokens": 512,
|
|
224
|
-
"language": ["zh", "en"],
|
|
225
|
-
"model_id": "AI-ModelScope/m3e-base",
|
|
226
|
-
"model_hub": "modelscope"
|
|
227
|
-
},
|
|
228
|
-
{
|
|
229
|
-
"model_name": "m3e-large",
|
|
230
|
-
"dimensions": 1024,
|
|
231
|
-
"max_tokens": 512,
|
|
232
|
-
"language": ["zh", "en"],
|
|
233
|
-
"model_id": "AI-ModelScope/m3e-large",
|
|
234
|
-
"model_hub": "modelscope"
|
|
235
|
-
},
|
|
236
|
-
{
|
|
237
|
-
"model_name": "gte-Qwen2",
|
|
238
|
-
"dimensions": 3584,
|
|
239
|
-
"max_tokens": 32000,
|
|
240
|
-
"language": ["zh", "en"],
|
|
241
|
-
"model_id": "iic/gte_Qwen2-7B-instruct",
|
|
242
|
-
"model_hub": "modelscope"
|
|
243
|
-
},
|
|
244
|
-
{
|
|
245
|
-
"model_name": "Qwen3-Embedding-0.6B",
|
|
246
|
-
"dimensions": 1024,
|
|
247
|
-
"max_tokens": 32768,
|
|
248
|
-
"language": ["zh", "en"],
|
|
249
|
-
"model_id": "Qwen/Qwen3-Embedding-0.6B",
|
|
250
|
-
"model_hub": "modelscope"
|
|
251
|
-
},
|
|
252
|
-
{
|
|
253
|
-
"model_name": "Qwen3-Embedding-4B",
|
|
254
|
-
"dimensions": 2560,
|
|
255
|
-
"max_tokens": 32768,
|
|
256
|
-
"language": ["zh", "en"],
|
|
257
|
-
"model_id": "Qwen/Qwen3-Embedding-4B",
|
|
258
|
-
"model_hub": "modelscope"
|
|
259
|
-
},
|
|
260
|
-
{
|
|
261
|
-
"model_name": "Qwen3-Embedding-8B",
|
|
262
|
-
"dimensions": 4096,
|
|
263
|
-
"max_tokens": 32768,
|
|
264
|
-
"language": ["zh", "en"],
|
|
265
|
-
"model_id": "Qwen/Qwen3-Embedding-8B",
|
|
266
|
-
"model_hub": "modelscope"
|
|
267
|
-
},
|
|
268
|
-
{
|
|
269
|
-
"model_name": "jina-embeddings-v3",
|
|
270
|
-
"dimensions": 1024,
|
|
271
|
-
"max_tokens": 8192,
|
|
272
|
-
"language": ["zh", "en"],
|
|
273
|
-
"model_id": "jinaai/jina-embeddings-v3",
|
|
274
|
-
"model_hub": "modelscope"
|
|
275
|
-
},
|
|
276
|
-
{
|
|
277
|
-
"model_name": "jina-clip-v2",
|
|
278
|
-
"dimensions": 1024,
|
|
279
|
-
"max_tokens": 8192,
|
|
280
|
-
"language": ["89 languages supported"],
|
|
281
|
-
"model_id": "jinaai/jina-clip-v2",
|
|
282
|
-
"model_hub": "modelscope",
|
|
283
|
-
"virtualenv": {
|
|
284
|
-
"packages": [
|
|
285
|
-
"sentence_transformers",
|
|
286
|
-
"transformers==4.51.3",
|
|
287
|
-
"xformers",
|
|
288
|
-
"flash_attn==2.7.3 ; sys_platform=='linux'"
|
|
289
|
-
],
|
|
290
|
-
"no_build_isolation": true
|
|
291
|
-
}
|
|
292
|
-
}
|
|
293
|
-
]
|
|
@@ -1,18 +0,0 @@
|
|
|
1
|
-
# Copyright 2022-2024 XProbe Inc.
|
|
2
|
-
#
|
|
3
|
-
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
4
|
-
# you may not use this file except in compliance with the License.
|
|
5
|
-
# You may obtain a copy of the License at
|
|
6
|
-
#
|
|
7
|
-
# http://www.apache.org/licenses/LICENSE-2.0
|
|
8
|
-
#
|
|
9
|
-
# Unless required by applicable law or agreed to in writing, software
|
|
10
|
-
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
11
|
-
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
12
|
-
# See the License for the specific language governing permissions and
|
|
13
|
-
# limitations under the License.
|
|
14
|
-
from .core import EmbeddingModelSpec
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
def get_model_version(embedding_model: EmbeddingModelSpec) -> str:
|
|
18
|
-
return f"{embedding_model.model_name}--{embedding_model.max_tokens}--{embedding_model.dimensions}"
|