xinference 1.7.1.post1__py3-none-any.whl → 1.8.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of xinference might be problematic. Click here for more details.

Files changed (136) hide show
  1. xinference/_version.py +3 -3
  2. xinference/client/restful/async_restful_client.py +8 -13
  3. xinference/client/restful/restful_client.py +6 -2
  4. xinference/core/chat_interface.py +6 -4
  5. xinference/core/media_interface.py +5 -0
  6. xinference/core/model.py +1 -5
  7. xinference/core/supervisor.py +117 -68
  8. xinference/core/worker.py +49 -37
  9. xinference/deploy/test/test_cmdline.py +2 -6
  10. xinference/model/audio/__init__.py +26 -23
  11. xinference/model/audio/chattts.py +3 -2
  12. xinference/model/audio/core.py +49 -98
  13. xinference/model/audio/cosyvoice.py +3 -2
  14. xinference/model/audio/custom.py +28 -73
  15. xinference/model/audio/f5tts.py +3 -2
  16. xinference/model/audio/f5tts_mlx.py +3 -2
  17. xinference/model/audio/fish_speech.py +3 -2
  18. xinference/model/audio/funasr.py +17 -4
  19. xinference/model/audio/kokoro.py +3 -2
  20. xinference/model/audio/megatts.py +3 -2
  21. xinference/model/audio/melotts.py +3 -2
  22. xinference/model/audio/model_spec.json +572 -171
  23. xinference/model/audio/utils.py +0 -6
  24. xinference/model/audio/whisper.py +3 -2
  25. xinference/model/audio/whisper_mlx.py +3 -2
  26. xinference/model/cache_manager.py +141 -0
  27. xinference/model/core.py +6 -49
  28. xinference/model/custom.py +174 -0
  29. xinference/model/embedding/__init__.py +67 -56
  30. xinference/model/embedding/cache_manager.py +35 -0
  31. xinference/model/embedding/core.py +104 -84
  32. xinference/model/embedding/custom.py +55 -78
  33. xinference/model/embedding/embed_family.py +80 -31
  34. xinference/model/embedding/flag/core.py +21 -5
  35. xinference/model/embedding/llama_cpp/__init__.py +0 -0
  36. xinference/model/embedding/llama_cpp/core.py +234 -0
  37. xinference/model/embedding/model_spec.json +968 -103
  38. xinference/model/embedding/sentence_transformers/core.py +30 -20
  39. xinference/model/embedding/vllm/core.py +11 -5
  40. xinference/model/flexible/__init__.py +8 -2
  41. xinference/model/flexible/core.py +26 -119
  42. xinference/model/flexible/custom.py +69 -0
  43. xinference/model/flexible/launchers/image_process_launcher.py +1 -0
  44. xinference/model/flexible/launchers/modelscope_launcher.py +5 -1
  45. xinference/model/flexible/launchers/transformers_launcher.py +15 -3
  46. xinference/model/flexible/launchers/yolo_launcher.py +5 -1
  47. xinference/model/image/__init__.py +20 -20
  48. xinference/model/image/cache_manager.py +62 -0
  49. xinference/model/image/core.py +70 -182
  50. xinference/model/image/custom.py +28 -72
  51. xinference/model/image/model_spec.json +402 -119
  52. xinference/model/image/ocr/got_ocr2.py +3 -2
  53. xinference/model/image/stable_diffusion/core.py +22 -7
  54. xinference/model/image/stable_diffusion/mlx.py +6 -6
  55. xinference/model/image/utils.py +2 -2
  56. xinference/model/llm/__init__.py +71 -94
  57. xinference/model/llm/cache_manager.py +292 -0
  58. xinference/model/llm/core.py +37 -111
  59. xinference/model/llm/custom.py +88 -0
  60. xinference/model/llm/llama_cpp/core.py +5 -7
  61. xinference/model/llm/llm_family.json +16260 -8151
  62. xinference/model/llm/llm_family.py +138 -839
  63. xinference/model/llm/lmdeploy/core.py +5 -7
  64. xinference/model/llm/memory.py +3 -4
  65. xinference/model/llm/mlx/core.py +6 -8
  66. xinference/model/llm/reasoning_parser.py +3 -1
  67. xinference/model/llm/sglang/core.py +32 -14
  68. xinference/model/llm/transformers/chatglm.py +3 -7
  69. xinference/model/llm/transformers/core.py +49 -27
  70. xinference/model/llm/transformers/deepseek_v2.py +2 -2
  71. xinference/model/llm/transformers/gemma3.py +2 -2
  72. xinference/model/llm/transformers/multimodal/cogagent.py +2 -2
  73. xinference/model/llm/transformers/multimodal/deepseek_vl2.py +2 -2
  74. xinference/model/llm/transformers/multimodal/gemma3.py +2 -2
  75. xinference/model/llm/transformers/multimodal/glm4_1v.py +167 -0
  76. xinference/model/llm/transformers/multimodal/glm4v.py +2 -2
  77. xinference/model/llm/transformers/multimodal/intern_vl.py +2 -2
  78. xinference/model/llm/transformers/multimodal/minicpmv26.py +3 -3
  79. xinference/model/llm/transformers/multimodal/ovis2.py +2 -2
  80. xinference/model/llm/transformers/multimodal/qwen-omni.py +2 -2
  81. xinference/model/llm/transformers/multimodal/qwen2_audio.py +2 -2
  82. xinference/model/llm/transformers/multimodal/qwen2_vl.py +2 -2
  83. xinference/model/llm/transformers/opt.py +3 -7
  84. xinference/model/llm/utils.py +34 -49
  85. xinference/model/llm/vllm/core.py +77 -27
  86. xinference/model/llm/vllm/xavier/engine.py +5 -3
  87. xinference/model/llm/vllm/xavier/scheduler.py +10 -6
  88. xinference/model/llm/vllm/xavier/transfer.py +1 -1
  89. xinference/model/rerank/__init__.py +26 -25
  90. xinference/model/rerank/core.py +47 -87
  91. xinference/model/rerank/custom.py +25 -71
  92. xinference/model/rerank/model_spec.json +158 -33
  93. xinference/model/rerank/utils.py +2 -2
  94. xinference/model/utils.py +115 -54
  95. xinference/model/video/__init__.py +13 -17
  96. xinference/model/video/core.py +44 -102
  97. xinference/model/video/diffusers.py +4 -3
  98. xinference/model/video/model_spec.json +90 -21
  99. xinference/types.py +5 -3
  100. xinference/web/ui/build/asset-manifest.json +3 -3
  101. xinference/web/ui/build/index.html +1 -1
  102. xinference/web/ui/build/static/js/main.7d24df53.js +3 -0
  103. xinference/web/ui/build/static/js/main.7d24df53.js.map +1 -0
  104. xinference/web/ui/node_modules/.cache/babel-loader/2704ff66a5f73ca78b341eb3edec60154369df9d87fbc8c6dd60121abc5e1b0a.json +1 -0
  105. xinference/web/ui/node_modules/.cache/babel-loader/607dfef23d33e6b594518c0c6434567639f24f356b877c80c60575184ec50ed0.json +1 -0
  106. xinference/web/ui/node_modules/.cache/babel-loader/9be3d56173aacc3efd0b497bcb13c4f6365de30069176ee9403b40e717542326.json +1 -0
  107. xinference/web/ui/node_modules/.cache/babel-loader/9f9dd6c32c78a222d07da5987ae902effe16bcf20aac00774acdccc4de3c9ff2.json +1 -0
  108. xinference/web/ui/node_modules/.cache/babel-loader/b2ab5ee972c60d15eb9abf5845705f8ab7e1d125d324d9a9b1bcae5d6fd7ffb2.json +1 -0
  109. xinference/web/ui/src/locales/en.json +0 -1
  110. xinference/web/ui/src/locales/ja.json +0 -1
  111. xinference/web/ui/src/locales/ko.json +0 -1
  112. xinference/web/ui/src/locales/zh.json +0 -1
  113. {xinference-1.7.1.post1.dist-info → xinference-1.8.0.dist-info}/METADATA +9 -11
  114. {xinference-1.7.1.post1.dist-info → xinference-1.8.0.dist-info}/RECORD +119 -119
  115. xinference/model/audio/model_spec_modelscope.json +0 -231
  116. xinference/model/embedding/model_spec_modelscope.json +0 -293
  117. xinference/model/embedding/utils.py +0 -18
  118. xinference/model/image/model_spec_modelscope.json +0 -375
  119. xinference/model/llm/llama_cpp/memory.py +0 -457
  120. xinference/model/llm/llm_family_csghub.json +0 -56
  121. xinference/model/llm/llm_family_modelscope.json +0 -8700
  122. xinference/model/llm/llm_family_openmind_hub.json +0 -1019
  123. xinference/model/rerank/model_spec_modelscope.json +0 -85
  124. xinference/model/video/model_spec_modelscope.json +0 -184
  125. xinference/web/ui/build/static/js/main.9b12b7f9.js +0 -3
  126. xinference/web/ui/build/static/js/main.9b12b7f9.js.map +0 -1
  127. xinference/web/ui/node_modules/.cache/babel-loader/1460361af6975e63576708039f1cb732faf9c672d97c494d4055fc6331460be0.json +0 -1
  128. xinference/web/ui/node_modules/.cache/babel-loader/4efd8dda58fda83ed9546bf2f587df67f8d98e639117bee2d9326a9a1d9bebb2.json +0 -1
  129. xinference/web/ui/node_modules/.cache/babel-loader/55b9fb40b57fa926e8f05f31c2f96467e76e5ad62f033dca97c03f9e8c4eb4fe.json +0 -1
  130. xinference/web/ui/node_modules/.cache/babel-loader/5b2dafe5aa9e1105e0244a2b6751807342fa86aa0144b4e84d947a1686102715.json +0 -1
  131. xinference/web/ui/node_modules/.cache/babel-loader/611fa2c6c53b66039991d06dfb0473b5ab37fc63b4564e0f6e1718523768a045.json +0 -1
  132. /xinference/web/ui/build/static/js/{main.9b12b7f9.js.LICENSE.txt → main.7d24df53.js.LICENSE.txt} +0 -0
  133. {xinference-1.7.1.post1.dist-info → xinference-1.8.0.dist-info}/WHEEL +0 -0
  134. {xinference-1.7.1.post1.dist-info → xinference-1.8.0.dist-info}/entry_points.txt +0 -0
  135. {xinference-1.7.1.post1.dist-info → xinference-1.8.0.dist-info}/licenses/LICENSE +0 -0
  136. {xinference-1.7.1.post1.dist-info → xinference-1.8.0.dist-info}/top_level.txt +0 -0
@@ -1,231 +0,0 @@
1
- [
2
- {
3
- "model_name": "whisper-large-v3",
4
- "model_family": "whisper",
5
- "model_hub": "modelscope",
6
- "model_id": "AI-ModelScope/whisper-large-v3",
7
- "model_revision": "master",
8
- "model_ability": ["audio2text"],
9
- "multilingual": true
10
- },
11
- {
12
- "model_name": "whisper-large-v3-turbo",
13
- "model_family": "whisper",
14
- "model_hub": "modelscope",
15
- "model_id": "AI-ModelScope/whisper-large-v3-turbo",
16
- "model_revision": "master",
17
- "model_ability": ["audio2text"],
18
- "multilingual": true
19
- },
20
- {
21
- "model_name": "Belle-whisper-large-v3-zh",
22
- "model_family": "whisper",
23
- "model_hub": "modelscope",
24
- "model_id": "Xorbits/Belle-whisper-large-v3-zh",
25
- "model_revision": "master",
26
- "model_ability": ["audio2text"],
27
- "multilingual": false
28
- },
29
- {
30
- "model_name": "SenseVoiceSmall",
31
- "model_family": "funasr",
32
- "model_hub": "modelscope",
33
- "model_id": "iic/SenseVoiceSmall",
34
- "model_revision": "master",
35
- "model_ability": ["audio2text"],
36
- "multilingual": true,
37
- "default_model_config": {
38
- "vad_model": "fsmn-vad",
39
- "vad_kwargs": {
40
- "max_single_segment_time": 30000
41
- }
42
- },
43
- "default_transcription_config": {
44
- "use_itn": true,
45
- "batch_size_s": 60,
46
- "merge_vad": true,
47
- "merge_length_s": 15
48
- }
49
- },
50
- {
51
- "model_name": "paraformer-zh",
52
- "model_family": "funasr",
53
- "model_hub": "modelscope",
54
- "model_id": "iic/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch",
55
- "model_revision": "master",
56
- "model_ability": ["audio2text"],
57
- "multilingual": false,
58
- "default_model_config": {
59
- "vad_model": "fsmn-vad",
60
- "punc_model": "ct-punc"
61
- },
62
- "default_transcription_config": {
63
- "batch_size_s": 300
64
- }
65
- },
66
- {
67
- "model_name": "paraformer-zh-hotword",
68
- "model_family": "funasr",
69
- "model_hub": "modelscope",
70
- "model_id": "iic/speech_paraformer-large-contextual_asr_nat-zh-cn-16k-common-vocab8404",
71
- "model_revision": "master",
72
- "model_ability": ["audio2text"],
73
- "multilingual": false,
74
- "default_model_config": {
75
- "vad_model": "fsmn-vad",
76
- "punc_model": "ct-punc"
77
- },
78
- "default_transcription_config": {
79
- "hotword": "",
80
- "batch_size_s": 300
81
- }
82
- },
83
- {
84
- "model_name": "paraformer-zh-long",
85
- "model_family": "funasr",
86
- "model_hub": "modelscope",
87
- "model_id": "iic/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch",
88
- "model_revision": "master",
89
- "model_ability": ["audio2text"],
90
- "multilingual": false,
91
- "default_model_config": {
92
- "vad_model": "fsmn-vad",
93
- "punc_model": "ct-punc"
94
- },
95
- "default_transcription_config": {
96
- "batch_size_s": 300
97
- }
98
- },
99
- {
100
- "model_name": "paraformer-zh-spk",
101
- "model_family": "funasr",
102
- "model_hub": "modelscope",
103
- "model_id": "iic/speech_paraformer-large-vad-punc-spk_asr_nat-zh-cn",
104
- "model_revision": "master",
105
- "model_ability": ["audio2text"],
106
- "multilingual": false,
107
- "default_model_config": {
108
- "vad_model": "fsmn-vad",
109
- "punc_model": "ct-punc",
110
- "spk_model":"cam++"
111
- },
112
- "default_transcription_config": {
113
- "batch_size_s": 300
114
- }
115
- },
116
- {
117
- "model_name": "seaco-paraformer-zh",
118
- "model_family": "funasr",
119
- "model_hub": "modelscope",
120
- "model_id": "iic/speech_seaco_paraformer_large_asr_nat-zh-cn-16k-common-vocab8404-pytorch",
121
- "model_revision": "master",
122
- "model_ability": ["audio2text"],
123
- "multilingual": false,
124
- "default_model_config": {
125
- "vad_model": "fsmn-vad",
126
- "punc_model": "ct-punc"
127
- },
128
- "default_transcription_config": {
129
- "hotword": "",
130
- "batch_size_s": 300
131
- }
132
- },
133
- {
134
- "model_name": "ChatTTS",
135
- "model_family": "ChatTTS",
136
- "model_hub": "modelscope",
137
- "model_id": "AI-ModelScope/ChatTTS",
138
- "model_revision": "master",
139
- "model_ability": ["text2audio"],
140
- "multilingual": true,
141
- "virtualenv": {
142
- "packages": [
143
- "ChatTTS>=0.2.1",
144
- "#system_torch#",
145
- "#system_numpy#"
146
- ]
147
- }
148
- },
149
- {
150
- "model_name": "CosyVoice-300M",
151
- "model_family": "CosyVoice",
152
- "model_hub": "modelscope",
153
- "model_id": "iic/CosyVoice-300M",
154
- "model_revision": "master",
155
- "model_ability": ["text2audio"],
156
- "multilingual": true
157
- },
158
- {
159
- "model_name": "CosyVoice-300M-SFT",
160
- "model_family": "CosyVoice",
161
- "model_hub": "modelscope",
162
- "model_id": "iic/CosyVoice-300M-SFT",
163
- "model_revision": "master",
164
- "model_ability": ["text2audio"],
165
- "multilingual": true
166
- },
167
- {
168
- "model_name": "CosyVoice-300M-Instruct",
169
- "model_family": "CosyVoice",
170
- "model_hub": "modelscope",
171
- "model_id": "iic/CosyVoice-300M-Instruct",
172
- "model_revision": "master",
173
- "model_ability": ["text2audio"],
174
- "multilingual": true
175
- },
176
- {
177
- "model_name": "CosyVoice2-0.5B",
178
- "model_family": "CosyVoice",
179
- "model_hub": "modelscope",
180
- "model_id": "iic/CosyVoice2-0.5B",
181
- "model_revision": "master",
182
- "model_ability": ["text2audio"],
183
- "multilingual": true,
184
- "virtualenv": {
185
- "packages": [
186
- "librosa",
187
- "tiktoken",
188
- "lightning>=2.0.0",
189
- "hydra-core>=1.3.2",
190
- "inflect",
191
- "conformer",
192
- "diffusers==0.29.0",
193
- "gdown",
194
- "pyarrow",
195
- "HyperPyYAML",
196
- "onnxruntime>=1.16.0",
197
- "pyworld>=0.3.4",
198
- "WeTextProcessing<1.0.4",
199
- "#system_numpy#",
200
- "#system_torch#"
201
- ]
202
- }
203
- },
204
- {
205
- "model_name": "F5-TTS",
206
- "model_family": "F5-TTS",
207
- "model_hub": "modelscope",
208
- "model_id": "SWivid/F5-TTS_Emilia-ZH-EN",
209
- "model_revision": "master",
210
- "model_ability": ["text2audio"],
211
- "multilingual": true
212
- },
213
- {
214
- "model_name": "Kokoro-82M",
215
- "model_family": "Kokoro",
216
- "model_hub": "modelscope",
217
- "model_id": "AI-ModelScope/Kokoro-82M",
218
- "model_revision": "master",
219
- "model_ability": ["text2audio"],
220
- "multilingual": true
221
- },
222
- {
223
- "model_name": "MegaTTS3",
224
- "model_family": "MegaTTS",
225
- "model_hub": "modelscope",
226
- "model_id": "ByteDance/MegaTTS3",
227
- "model_revision": "master",
228
- "model_ability": ["text2audio"],
229
- "multilingual": true
230
- }
231
- ]
@@ -1,293 +0,0 @@
1
- [
2
- {
3
- "model_name": "bge-large-en",
4
- "dimensions": 1024,
5
- "max_tokens": 512,
6
- "language": ["en"],
7
- "model_id": "Xorbits/bge-large-en",
8
- "model_revision": "v0.0.1",
9
- "model_hub": "modelscope"
10
- },
11
- {
12
- "model_name": "bge-base-en",
13
- "dimensions": 768,
14
- "max_tokens": 512,
15
- "language": ["en"],
16
- "model_id": "Xorbits/bge-base-en",
17
- "model_revision": "v0.0.1",
18
- "model_hub": "modelscope"
19
- },
20
- {
21
- "model_name": "gte-large",
22
- "dimensions": 1024,
23
- "max_tokens": 512,
24
- "language": ["en"],
25
- "model_id": "Xorbits/gte-large",
26
- "model_revision": "v0.0.1",
27
- "model_hub": "modelscope"
28
- },
29
- {
30
- "model_name": "gte-base",
31
- "dimensions": 768,
32
- "max_tokens": 512,
33
- "language": ["en"],
34
- "model_id": "Xorbits/gte-base",
35
- "model_revision": "v0.0.1",
36
- "model_hub": "modelscope"
37
- },
38
- {
39
- "model_name": "e5-large-v2",
40
- "dimensions": 1024,
41
- "max_tokens": 512,
42
- "language": ["en"],
43
- "model_id": "Xorbits/e5-large-v2",
44
- "model_revision": "v0.0.1",
45
- "model_hub": "modelscope"
46
- },
47
- {
48
- "model_name": "bge-large-zh",
49
- "dimensions": 1024,
50
- "max_tokens": 512,
51
- "language": ["zh"],
52
- "model_id": "Xorbits/bge-large-zh",
53
- "model_revision": "v0.0.1",
54
- "model_hub": "modelscope"
55
- },
56
- {
57
- "model_name": "bge-large-zh-noinstruct",
58
- "dimensions": 1024,
59
- "max_tokens": 512,
60
- "language": ["zh"],
61
- "model_id": "Xorbits/bge-large-zh-noinstruct",
62
- "model_revision": "v0.0.1",
63
- "model_hub": "modelscope"
64
- },
65
- {
66
- "model_name": "bge-base-zh",
67
- "dimensions": 768,
68
- "max_tokens": 512,
69
- "language": ["zh"],
70
- "model_id": "Xorbits/bge-base-zh",
71
- "model_revision": "v0.0.2",
72
- "model_hub": "modelscope"
73
- },
74
- {
75
- "model_name": "multilingual-e5-large",
76
- "dimensions": 1024,
77
- "max_tokens": 514,
78
- "language": ["zh"],
79
- "model_id": "Xorbits/multilingual-e5-large",
80
- "model_revision": "v0.0.1",
81
- "model_hub": "modelscope"
82
- },
83
- {
84
- "model_name": "bge-small-zh",
85
- "dimensions": 512,
86
- "max_tokens": 512,
87
- "language": ["zh"],
88
- "model_id": "Xorbits/bge-small-zh",
89
- "model_revision": "v0.0.1",
90
- "model_hub": "modelscope"
91
- },
92
- {
93
- "model_name": "bge-small-zh-v1.5",
94
- "dimensions": 512,
95
- "max_tokens": 512,
96
- "language": ["zh"],
97
- "model_id": "Xorbits/bge-small-zh-v1.5",
98
- "model_revision": "v0.0.2",
99
- "model_hub": "modelscope"
100
- },
101
- {
102
- "model_name": "bge-base-zh-v1.5",
103
- "dimensions": 768,
104
- "max_tokens": 512,
105
- "language": ["zh"],
106
- "model_id": "Xorbits/bge-base-zh-v1.5",
107
- "model_revision": "v0.0.1",
108
- "model_hub": "modelscope"
109
- },
110
- {
111
- "model_name": "bge-large-zh-v1.5",
112
- "dimensions": 1024,
113
- "max_tokens": 512,
114
- "language": ["zh"],
115
- "model_id": "Xorbits/bge-large-zh-v1.5",
116
- "model_revision": "v0.0.1",
117
- "model_hub": "modelscope"
118
- },
119
- {
120
- "model_name": "bge-small-en-v1.5",
121
- "dimensions": 384,
122
- "max_tokens": 512,
123
- "language": ["en"],
124
- "model_id": "Xorbits/bge-small-en-v1.5",
125
- "model_revision": "v0.0.2",
126
- "model_hub": "modelscope"
127
- },
128
- {
129
- "model_name": "bge-base-en-v1.5",
130
- "dimensions": 768,
131
- "max_tokens": 512,
132
- "language": ["en"],
133
- "model_id": "Xorbits/bge-base-en-v1.5",
134
- "model_revision": "v0.0.1",
135
- "model_hub": "modelscope"
136
- },
137
- {
138
- "model_name": "bge-large-en-v1.5",
139
- "dimensions": 1024,
140
- "max_tokens": 512,
141
- "language": ["en"],
142
- "model_id": "Xorbits/bge-large-en-v1.5",
143
- "model_revision": "v0.0.1",
144
- "model_hub": "modelscope"
145
- },
146
- {
147
- "model_name": "jina-embeddings-v2-small-en",
148
- "dimensions": 512,
149
- "max_tokens": 8192,
150
- "language": ["en"],
151
- "model_id": "Xorbits/jina-embeddings-v2-small-en",
152
- "model_revision": "v0.0.1",
153
- "model_hub": "modelscope"
154
- },
155
- {
156
- "model_name": "jina-embeddings-v2-base-en",
157
- "dimensions": 512,
158
- "max_tokens": 8192,
159
- "language": ["en"],
160
- "model_id": "Xorbits/jina-embeddings-v2-base-en",
161
- "model_revision": "v0.0.1",
162
- "model_hub": "modelscope"
163
- },
164
- {
165
- "model_name": "jina-embeddings-v2-base-zh",
166
- "dimensions": 768,
167
- "max_tokens": 8192,
168
- "language": ["zh", "en"],
169
- "model_id": "jinaai/jina-embeddings-v2-base-zh",
170
- "model_hub": "modelscope"
171
- },
172
- {
173
- "model_name": "text2vec-large-chinese",
174
- "dimensions": 1024,
175
- "max_tokens": 256,
176
- "language": ["zh"],
177
- "model_id": "Jerry0/text2vec-large-chinese",
178
- "model_hub": "modelscope"
179
- },
180
- {
181
- "model_name": "text2vec-base-chinese",
182
- "dimensions": 768,
183
- "max_tokens": 128,
184
- "language": ["zh"],
185
- "model_id": "Jerry0/text2vec-base-chinese",
186
- "model_hub": "modelscope"
187
- },
188
- {
189
- "model_name": "text2vec-base-chinese-paraphrase",
190
- "dimensions": 768,
191
- "max_tokens": 256,
192
- "language": ["zh"],
193
- "model_id": "mwei23/text2vec-base-chinese-paraphrase",
194
- "model_hub": "modelscope"
195
- },
196
- {
197
- "model_name": "bge-m3",
198
- "dimensions": 1024,
199
- "max_tokens": 8192,
200
- "language": ["zh", "en"],
201
- "model_id": "Xorbits/bge-m3",
202
- "model_hub": "modelscope"
203
- },
204
- {
205
- "model_name": "bce-embedding-base_v1",
206
- "dimensions": 768,
207
- "max_tokens": 512,
208
- "language": ["zh", "en"],
209
- "model_id": "maidalun/bce-embedding-base_v1",
210
- "model_hub": "modelscope"
211
- },
212
- {
213
- "model_name": "m3e-small",
214
- "dimensions": 512,
215
- "max_tokens": 512,
216
- "language": ["zh", "en"],
217
- "model_id": "AI-ModelScope/m3e-small",
218
- "model_hub": "modelscope"
219
- },
220
- {
221
- "model_name": "m3e-base",
222
- "dimensions": 768,
223
- "max_tokens": 512,
224
- "language": ["zh", "en"],
225
- "model_id": "AI-ModelScope/m3e-base",
226
- "model_hub": "modelscope"
227
- },
228
- {
229
- "model_name": "m3e-large",
230
- "dimensions": 1024,
231
- "max_tokens": 512,
232
- "language": ["zh", "en"],
233
- "model_id": "AI-ModelScope/m3e-large",
234
- "model_hub": "modelscope"
235
- },
236
- {
237
- "model_name": "gte-Qwen2",
238
- "dimensions": 3584,
239
- "max_tokens": 32000,
240
- "language": ["zh", "en"],
241
- "model_id": "iic/gte_Qwen2-7B-instruct",
242
- "model_hub": "modelscope"
243
- },
244
- {
245
- "model_name": "Qwen3-Embedding-0.6B",
246
- "dimensions": 1024,
247
- "max_tokens": 32768,
248
- "language": ["zh", "en"],
249
- "model_id": "Qwen/Qwen3-Embedding-0.6B",
250
- "model_hub": "modelscope"
251
- },
252
- {
253
- "model_name": "Qwen3-Embedding-4B",
254
- "dimensions": 2560,
255
- "max_tokens": 32768,
256
- "language": ["zh", "en"],
257
- "model_id": "Qwen/Qwen3-Embedding-4B",
258
- "model_hub": "modelscope"
259
- },
260
- {
261
- "model_name": "Qwen3-Embedding-8B",
262
- "dimensions": 4096,
263
- "max_tokens": 32768,
264
- "language": ["zh", "en"],
265
- "model_id": "Qwen/Qwen3-Embedding-8B",
266
- "model_hub": "modelscope"
267
- },
268
- {
269
- "model_name": "jina-embeddings-v3",
270
- "dimensions": 1024,
271
- "max_tokens": 8192,
272
- "language": ["zh", "en"],
273
- "model_id": "jinaai/jina-embeddings-v3",
274
- "model_hub": "modelscope"
275
- },
276
- {
277
- "model_name": "jina-clip-v2",
278
- "dimensions": 1024,
279
- "max_tokens": 8192,
280
- "language": ["89 languages supported"],
281
- "model_id": "jinaai/jina-clip-v2",
282
- "model_hub": "modelscope",
283
- "virtualenv": {
284
- "packages": [
285
- "sentence_transformers",
286
- "transformers==4.51.3",
287
- "xformers",
288
- "flash_attn==2.7.3 ; sys_platform=='linux'"
289
- ],
290
- "no_build_isolation": true
291
- }
292
- }
293
- ]
@@ -1,18 +0,0 @@
1
- # Copyright 2022-2024 XProbe Inc.
2
- #
3
- # Licensed under the Apache License, Version 2.0 (the "License");
4
- # you may not use this file except in compliance with the License.
5
- # You may obtain a copy of the License at
6
- #
7
- # http://www.apache.org/licenses/LICENSE-2.0
8
- #
9
- # Unless required by applicable law or agreed to in writing, software
10
- # distributed under the License is distributed on an "AS IS" BASIS,
11
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
- # See the License for the specific language governing permissions and
13
- # limitations under the License.
14
- from .core import EmbeddingModelSpec
15
-
16
-
17
- def get_model_version(embedding_model: EmbeddingModelSpec) -> str:
18
- return f"{embedding_model.model_name}--{embedding_model.max_tokens}--{embedding_model.dimensions}"