xinference 1.2.1__py3-none-any.whl → 1.2.2__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of xinference might be problematic. Click here for more details.

@@ -11,16 +11,28 @@
11
11
  # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
12
  # See the License for the specific language governing permissions and
13
13
  # limitations under the License.
14
+
14
15
  import base64
15
16
  import functools
16
17
  import json
17
18
  import logging
18
19
  import os
20
+ import re
19
21
  import time
20
22
  import typing
21
23
  import uuid
22
24
  from io import BytesIO
23
- from typing import AsyncGenerator, Dict, Iterator, List, Optional, Tuple, cast
25
+ from typing import (
26
+ Any,
27
+ AsyncGenerator,
28
+ Dict,
29
+ Iterable,
30
+ Iterator,
31
+ List,
32
+ Optional,
33
+ Tuple,
34
+ cast,
35
+ )
24
36
 
25
37
  import requests
26
38
  from PIL import Image
@@ -64,6 +76,18 @@ LLAMA3_TOOL_CALL_FAMILY = [
64
76
  "llama-3.1-instruct",
65
77
  ]
66
78
 
79
+ DEEPSEEK_TOOL_CALL_FAMILY = [
80
+ "deepseek-r1-distill-qwen",
81
+ "deepseek-r1-distill-llama",
82
+ ]
83
+
84
+ TOOL_CALL_FAMILY = (
85
+ QWEN_TOOL_CALL_FAMILY
86
+ + GLM4_TOOL_CALL_FAMILY
87
+ + LLAMA3_TOOL_CALL_FAMILY
88
+ + DEEPSEEK_TOOL_CALL_FAMILY
89
+ )
90
+
67
91
  QWEN_TOOL_CALL_SYMBOLS = ["<tool_call>", "</tool_call>"]
68
92
 
69
93
 
@@ -308,6 +332,35 @@ class ChatModelMixin:
308
332
  else:
309
333
  yield cls._to_chat_completion_chunk(chunk)
310
334
 
335
+ @classmethod
336
+ def _tools_to_messages_for_deepseek(
337
+ cls, messages: List[dict], tools: Iterable[dict]
338
+ ):
339
+ # deepseek integrates tool calls into messages
340
+ # we follow the chat template rule to integrate tools into messages
341
+ tool_call_message: Dict[str, Any] = {
342
+ "role": "assistant",
343
+ "content": None,
344
+ "tool_calls": [],
345
+ }
346
+
347
+ for tool in tools:
348
+ function_name = tool["function"]["name"]
349
+ parameters = tool["function"].get("parameters", {}).get("properties", {})
350
+ function_args_json = json.dumps(parameters)
351
+
352
+ tool_call_message["tool_calls"].append(
353
+ {
354
+ "type": "function",
355
+ "function": {
356
+ "name": function_name,
357
+ "arguments": function_args_json,
358
+ },
359
+ }
360
+ )
361
+
362
+ messages.append(tool_call_message)
363
+
311
364
  @classmethod
312
365
  async def _async_to_chat_completion_chunks(
313
366
  cls,
@@ -401,6 +454,61 @@ class ChatModelMixin:
401
454
  except Exception:
402
455
  return [(text, None, None)]
403
456
 
457
+ @classmethod
458
+ def _eval_deepseek_chat_arguments(cls, c) -> List[Tuple]:
459
+ """
460
+ Parses tool calls from deepseek-r1 format and removes duplicates.
461
+
462
+ Returns:
463
+ List[Tuple[Optional[str], Optional[str], Optional[dict]]]
464
+ - (None, function_name, arguments) if successfully parsed.
465
+ - (content, None, None) if parsing failed (content is raw JSON text).
466
+
467
+ Example input:
468
+ <|tool▁call|>get_current_weather
469
+ ```json
470
+ {"location": "tokyo", "unit": "fahrenheit"}
471
+ ```
472
+
473
+ Output:
474
+ [
475
+ (None, "get_current_weather", {"location": "tokyo", "unit": "fahrenheit"})
476
+ ]
477
+ """
478
+
479
+ text = c["choices"][0]["text"]
480
+
481
+ pattern = r"<|tool▁call|>(\w+)\s*```json\s*(.*?)\s*```"
482
+ matches = re.findall(pattern, text, re.DOTALL)
483
+
484
+ if not matches:
485
+ return [(text, None, None)]
486
+
487
+ tool_calls = set() # Used for deduplication
488
+ results = []
489
+
490
+ for function_name, args_json in matches:
491
+ try:
492
+ arguments = json.loads(args_json)
493
+ # Convert dictionary to frozenset for deduplication
494
+ arguments_hashable = frozenset(arguments.items())
495
+ tool_call_tuple = (None, function_name, arguments)
496
+ except json.JSONDecodeError:
497
+ tool_call_tuple = (
498
+ args_json,
499
+ None,
500
+ None,
501
+ ) # If parsing fails, treat as raw content
502
+ arguments_hashable = None # No need for hashing
503
+
504
+ # Avoid duplicate entries
505
+ dedup_key = (function_name, arguments_hashable)
506
+ if dedup_key not in tool_calls:
507
+ tool_calls.add(dedup_key)
508
+ results.append(tool_call_tuple)
509
+
510
+ return results
511
+
404
512
  @classmethod
405
513
  def _eval_tool_arguments(cls, model_family, c):
406
514
  family = model_family.model_family or model_family.model_name
@@ -410,6 +518,8 @@ class ChatModelMixin:
410
518
  result = cls._eval_qwen_chat_arguments(c)
411
519
  elif family in LLAMA3_TOOL_CALL_FAMILY:
412
520
  result = cls._eval_llama3_chat_arguments(c)
521
+ elif family in DEEPSEEK_TOOL_CALL_FAMILY:
522
+ result = cls._eval_deepseek_chat_arguments(c)
413
523
  else:
414
524
  raise Exception(
415
525
  f"Model {model_family.model_name} is not support tool calls."
@@ -44,6 +44,7 @@ from ....types import (
44
44
  from .. import LLM, LLMFamilyV1, LLMSpecV1
45
45
  from ..llm_family import CustomLLMFamilyV1
46
46
  from ..utils import (
47
+ DEEPSEEK_TOOL_CALL_FAMILY,
47
48
  QWEN_TOOL_CALL_FAMILY,
48
49
  QWEN_TOOL_CALL_SYMBOLS,
49
50
  ChatModelMixin,
@@ -185,6 +186,7 @@ if VLLM_INSTALLED and vllm.__version__ > "0.5.3":
185
186
  VLLM_SUPPORTED_MODELS.append("llama-3.1")
186
187
  VLLM_SUPPORTED_CHAT_MODELS.append("llama-3.1-instruct")
187
188
  VLLM_SUPPORTED_CHAT_MODELS.append("llama-3.3-instruct")
189
+ VLLM_SUPPORTED_CHAT_MODELS.append("deepseek-r1-distill-llama")
188
190
 
189
191
  if VLLM_INSTALLED and vllm.__version__ >= "0.6.1":
190
192
  VLLM_SUPPORTED_VISION_MODEL_LIST.append("internvl2")
@@ -198,6 +200,12 @@ if VLLM_INSTALLED and vllm.__version__ >= "0.6.3":
198
200
  VLLM_SUPPORTED_VISION_MODEL_LIST.append("qwen2-vl-instruct")
199
201
  VLLM_SUPPORTED_VISION_MODEL_LIST.append("QvQ-72B-Preview")
200
202
 
203
+ if VLLM_INSTALLED and vllm.__version__ >= "0.7.0":
204
+ VLLM_SUPPORTED_CHAT_MODELS.append("internlm3-instruct")
205
+
206
+ if VLLM_INSTALLED and vllm.__version__ >= "0.7.2":
207
+ VLLM_SUPPORTED_VISION_MODEL_LIST.append("qwen2.5-vl-instruct")
208
+
201
209
 
202
210
  class VLLMModel(LLM):
203
211
  def __init__(
@@ -807,8 +815,11 @@ class VLLMChatModel(VLLMModel, ChatModelMixin):
807
815
  tools = generate_config.pop("tools", []) if generate_config else None
808
816
  model_family = self.model_family.model_family or self.model_family.model_name
809
817
  full_context_kwargs = {}
810
- if tools and model_family in QWEN_TOOL_CALL_FAMILY:
811
- full_context_kwargs["tools"] = tools
818
+ if tools:
819
+ if model_family in QWEN_TOOL_CALL_FAMILY:
820
+ full_context_kwargs["tools"] = tools
821
+ elif model_family in DEEPSEEK_TOOL_CALL_FAMILY:
822
+ self._tools_to_messages_for_deepseek(messages, tools)
812
823
  assert self.model_family.chat_template is not None
813
824
  full_prompt = self.get_full_context(
814
825
  messages, self.model_family.chat_template, **full_context_kwargs
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: xinference
3
- Version: 1.2.1
3
+ Version: 1.2.2
4
4
  Summary: Model Serving Made Easy
5
5
  Home-page: https://github.com/xorbitsai/inference
6
6
  Author: Qin Xuye
@@ -98,7 +98,7 @@ Requires-Dist: tomli; extra == "all"
98
98
  Requires-Dist: vocos; extra == "all"
99
99
  Requires-Dist: jieba; extra == "all"
100
100
  Requires-Dist: soundfile; extra == "all"
101
- Requires-Dist: qwen-vl-utils; extra == "all"
101
+ Requires-Dist: qwen-vl-utils!=0.0.9; extra == "all"
102
102
  Requires-Dist: datamodel-code-generator; extra == "all"
103
103
  Requires-Dist: jsonschema; extra == "all"
104
104
  Requires-Dist: verovio>=4.3.1; extra == "all"
@@ -154,6 +154,7 @@ Requires-Dist: fugashi; extra == "audio"
154
154
  Requires-Dist: g2p-en; extra == "audio"
155
155
  Requires-Dist: anyascii; extra == "audio"
156
156
  Requires-Dist: gruut[de,es,fr]; extra == "audio"
157
+ Requires-Dist: kokoro>=0.7.9; extra == "audio"
157
158
  Requires-Dist: nemo-text-processing<1.1.0; sys_platform == "linux" and extra == "audio"
158
159
  Requires-Dist: WeTextProcessing<1.0.4; sys_platform == "linux" and extra == "audio"
159
160
  Provides-Extra: benchmark
@@ -211,7 +212,7 @@ Requires-Dist: mlx-lm>=0.21.1; extra == "mlx"
211
212
  Requires-Dist: mlx-vlm>=0.1.11; extra == "mlx"
212
213
  Requires-Dist: mlx-whisper; extra == "mlx"
213
214
  Requires-Dist: f5-tts-mlx; extra == "mlx"
214
- Requires-Dist: qwen-vl-utils; extra == "mlx"
215
+ Requires-Dist: qwen-vl-utils!=0.0.9; extra == "mlx"
215
216
  Requires-Dist: tomli; extra == "mlx"
216
217
  Provides-Extra: rerank
217
218
  Requires-Dist: FlagEmbedding; extra == "rerank"
@@ -225,7 +226,6 @@ Requires-Dist: torch; extra == "transformers"
225
226
  Requires-Dist: accelerate>=0.28.0; extra == "transformers"
226
227
  Requires-Dist: sentencepiece; extra == "transformers"
227
228
  Requires-Dist: transformers-stream-generator; extra == "transformers"
228
- Requires-Dist: bitsandbytes; extra == "transformers"
229
229
  Requires-Dist: protobuf; extra == "transformers"
230
230
  Requires-Dist: einops; extra == "transformers"
231
231
  Requires-Dist: tiktoken; extra == "transformers"
@@ -236,11 +236,12 @@ Requires-Dist: torchvision; extra == "transformers"
236
236
  Requires-Dist: peft; extra == "transformers"
237
237
  Requires-Dist: eva-decord; extra == "transformers"
238
238
  Requires-Dist: jj-pytorchvideo; extra == "transformers"
239
- Requires-Dist: qwen-vl-utils; extra == "transformers"
239
+ Requires-Dist: qwen-vl-utils!=0.0.9; extra == "transformers"
240
240
  Requires-Dist: datamodel-code-generator; extra == "transformers"
241
241
  Requires-Dist: jsonschema; extra == "transformers"
242
242
  Requires-Dist: auto-gptq; sys_platform != "darwin" and extra == "transformers"
243
243
  Requires-Dist: autoawq<0.2.6; sys_platform != "darwin" and extra == "transformers"
244
+ Requires-Dist: bitsandbytes; sys_platform == "linux" and extra == "transformers"
244
245
  Provides-Extra: video
245
246
  Requires-Dist: diffusers>=0.32.0; extra == "video"
246
247
  Requires-Dist: imageio-ffmpeg; extra == "video"
@@ -296,14 +297,14 @@ potential of cutting-edge AI models.
296
297
  - Support speech recognition model: [#929](https://github.com/xorbitsai/inference/pull/929)
297
298
  - Metrics support: [#906](https://github.com/xorbitsai/inference/pull/906)
298
299
  ### New Models
300
+ - Built-in support for [DeepSeek-R1-Distill-Qwen](https://github.com/deepseek-ai/DeepSeek-R1?tab=readme-ov-file#deepseek-r1-distill-models): [#2781](https://github.com/xorbitsai/inference/pull/2781)
301
+ - Built-in support for [qwen2.5-vl](https://github.com/QwenLM/Qwen2.5-VL): [#2788](https://github.com/xorbitsai/inference/pull/2788)
302
+ - Built-in support for [internlm3-instruct](https://github.com/InternLM/InternLM): [#2789](https://github.com/xorbitsai/inference/pull/2789)
299
303
  - Built-in support for [MeloTTS](https://github.com/myshell-ai/MeloTTS): [#2760](https://github.com/xorbitsai/inference/pull/2760)
300
304
  - Built-in support for [CogAgent](https://github.com/THUDM/CogAgent): [#2740](https://github.com/xorbitsai/inference/pull/2740)
301
305
  - Built-in support for [HunyuanVideo](https://github.com/Tencent/HunyuanVideo): [#2721](https://github.com/xorbitsai/inference/pull/2721)
302
306
  - Built-in support for [HunyuanDiT](https://github.com/Tencent/HunyuanDiT): [#2727](https://github.com/xorbitsai/inference/pull/2727)
303
307
  - Built-in support for [Macro-o1](https://github.com/AIDC-AI/Marco-o1): [#2749](https://github.com/xorbitsai/inference/pull/2749)
304
- - Built-in support for [Stable Diffusion 3.5](https://huggingface.co/collections/stabilityai/stable-diffusion-35-671785cca799084f71fa2838): [#2706](https://github.com/xorbitsai/inference/pull/2706)
305
- - Built-in support for [CosyVoice 2](https://huggingface.co/FunAudioLLM/CosyVoice2-0.5B): [#2684](https://github.com/xorbitsai/inference/pull/2684)
306
- - Built-in support for [Fish Speech V1.5](https://huggingface.co/fishaudio/fish-speech-1.5): [#2672](https://github.com/xorbitsai/inference/pull/2672)
307
308
  ### Integrations
308
309
  - [Dify](https://docs.dify.ai/advanced/model-configuration/xinference): an LLMOps platform that enables developers (and even non-developers) to quickly build useful applications based on large language models, ensuring they are visual, operable, and improvable.
309
310
  - [FastGPT](https://github.com/labring/FastGPT): a knowledge-based platform built on the LLM, offers out-of-the-box data processing and model invocation capabilities, allows for workflow orchestration through Flow visualization.
@@ -1,6 +1,6 @@
1
1
  xinference/__init__.py,sha256=nmTTrYbIpj964ZF6ojtgOM7E85JBOj1EyQbmYjbj1jw,915
2
2
  xinference/_compat.py,sha256=URSJQLXrcsTO9B_4x0wVDPijYQDhuVJmZ95npID560w,4197
3
- xinference/_version.py,sha256=cRjgwmUs9T7LZabp8_lbu3KxEJZrH641zBpN4hZkBEY,497
3
+ xinference/_version.py,sha256=tyoMIoWH7JrsR3odtmfwgrhewV2IdEEE1IRf42ATx28,497
4
4
  xinference/conftest.py,sha256=vETDpRBVIlWbWi7OTwf7og89U25KyYGyI7yPIB3O8N8,9564
5
5
  xinference/constants.py,sha256=mEW4HDzjXtDXN61Mt6TtJrJ4ljbB6VUkh97e3oDbNx4,3905
6
6
  xinference/device_utils.py,sha256=zswJiws3VyTIaNO8z-MOcsJH_UiPoePPiKK5zoNrjTA,3285
@@ -9,14 +9,14 @@ xinference/isolation.py,sha256=gTU1em5fxg1m-7hxieWBMZvVkXZX4GZYmeT7XxXsYrU,2699
9
9
  xinference/types.py,sha256=p9IV2vmbtFfpo9P66IynkEgU2N6310uVMKBuq0rbtNc,13102
10
10
  xinference/utils.py,sha256=zYgf9bCvfbybRt3gEog6r5WJCpj0czZCf0qgRdYjkN8,720
11
11
  xinference/api/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
12
- xinference/api/restful_api.py,sha256=NlWmA1pDu8oPybI9daU4pplQXkRD-NExTO2L_E4Y2zM,92876
12
+ xinference/api/restful_api.py,sha256=XENr9892-WHZ1XKaCwijdTAAd1uX2D0oMKUF6Pw99GQ,92730
13
13
  xinference/api/oauth2/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
14
14
  xinference/api/oauth2/auth_service.py,sha256=74JzB42fbbmBu4Q1dW3A9Fp_N7167KgRGB42Z0NHjAM,6119
15
15
  xinference/api/oauth2/types.py,sha256=K923sv_XySIUtM2Eozl9IG082IJcDOS5SFLrPZ5ELBg,996
16
16
  xinference/api/oauth2/utils.py,sha256=SIiiUj6VuTEsj3bZ2TYUyhx3cGlLSX3ZNWDOgUwRtXc,1410
17
17
  xinference/client/__init__.py,sha256=Gc4HOzAy_1cic5kXlso7hahYgw89CKvZSJDicEU461k,669
18
18
  xinference/client/common.py,sha256=iciZRs5YjM2gYsXnwACPMaiBZp4_XpawWwfym0Iyu40,1617
19
- xinference/client/handlers.py,sha256=OKl_i5FA341wsQf_0onSOPbbW6V861WJrSP7ghtDc8c,527
19
+ xinference/client/handlers.py,sha256=HYmQjG0BPK5PKcWVEDJqQz5iLjS8yjOq5j-HP_CAi_k,630
20
20
  xinference/client/restful/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
21
21
  xinference/client/restful/restful_client.py,sha256=JwzP7etUZBR0mmU7y3dUOEWN_D7ol_2hXN9KMAzKZaw,53601
22
22
  xinference/core/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
@@ -28,14 +28,14 @@ xinference/core/metrics.py,sha256=ScmTG15Uq3h_ob72ybZSMWdnk8P4sUZFcm60f4ikSXc,26
28
28
  xinference/core/model.py,sha256=8FaEJK4tTL7wvYivLEqWKkZOrAMODBSoxg9A4fYzUH0,43365
29
29
  xinference/core/progress_tracker.py,sha256=LIF6CLIlnEoSBkuDCraJktDOzZ31mQ4HOo6EVr3KpQM,6453
30
30
  xinference/core/resource.py,sha256=FQ0aRt3T4ZQo0P6CZZf5QUKHiCsr5llBvKb1f7wfnxg,1611
31
- xinference/core/scheduler.py,sha256=gdj3SyP_jelJ86vTRrgnFynhxz5JSwLRsQgx8PTtBi8,15671
31
+ xinference/core/scheduler.py,sha256=WHWOlbzAPoURXNzjFNdfvhkgBYO8ufZtIu3JXfq7fmY,15576
32
32
  xinference/core/status_guard.py,sha256=4an1KjUOhCStgRQUw1VSzXcycXUtvhxwiMREKKcl1UI,2828
33
33
  xinference/core/supervisor.py,sha256=4beE9EZx5xWpSLZtZwT3CPYWT25zPFl9gRj2oK8Za3w,60208
34
34
  xinference/core/utils.py,sha256=RR3XDioh52Wy8vmAqhCZ6EQskr-HPDfUp0vCCAPIAvs,11302
35
35
  xinference/core/worker.py,sha256=FVy_zXOHovYLPQQr6DSpjSGWaLyH425gTUUH_vwKVfk,50666
36
36
  xinference/deploy/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
37
37
  xinference/deploy/cmdline.py,sha256=gPwp9IngaXCypUEnPDS_22U8GntsKr7qHDST7duyAoI,48478
38
- xinference/deploy/local.py,sha256=gcH6WfTxfhjvNkxxKZH3tcGtXV48BEPoaLWYztZHaeo,3954
38
+ xinference/deploy/local.py,sha256=sO3BcpEH9oCF87CxWVA4AXAYcfHGnrcop40ew5NOA2g,3979
39
39
  xinference/deploy/supervisor.py,sha256=68rB2Ey5KFeF6zto9YGbw3P8QLZmF_KSh1NwH_pNP4w,2986
40
40
  xinference/deploy/utils.py,sha256=jdL7i2WV6u_BZ8IiE1d3YktvCARcB3ntzMQ5rHGD5DM,6756
41
41
  xinference/deploy/worker.py,sha256=VQ71ClWpeGsyFgDmcOes2ub1cil10cBjhFLHYeuVwC4,2974
@@ -46,16 +46,17 @@ xinference/model/core.py,sha256=_NEH4wkjjJgRDdLHNVY_hL3V0kT67CvTay89uIzx1Ns,4736
46
46
  xinference/model/utils.py,sha256=_yJ5h4RUzt7Kjs2WdjSzbVM3FTWEkX0ycOnXANZ9KVg,11394
47
47
  xinference/model/audio/__init__.py,sha256=KasWsaNPeij6sGpHKqXaUc_bxUw1yYbD7-fwxkcoAVE,3731
48
48
  xinference/model/audio/chattts.py,sha256=ny3DZTCTt2MzdkLw994_QHZ_4qIEUZcNexNJkCejCyo,4998
49
- xinference/model/audio/core.py,sha256=jboohBIr8GmXH4ASoKJvU0lWJzsCqUszTsJdn-pByaI,7517
49
+ xinference/model/audio/core.py,sha256=AAuss7fGwGktcB_aGRwuR4TJIM6-Rrm0YCU752QoxfQ,7710
50
50
  xinference/model/audio/cosyvoice.py,sha256=vw7OR2t7zNimQn3Q74iiL1-2en5o6gvdcZsDgkmYpy4,7796
51
51
  xinference/model/audio/custom.py,sha256=8GXBRmTtR-GY03-E91nlRGTIuabCRzlt20ecU6Un6Y8,4985
52
52
  xinference/model/audio/f5tts.py,sha256=RyfEYVvKhV7JygIv4F45C8wBr-u0mx9WpXj3gIOtL7o,6809
53
53
  xinference/model/audio/f5tts_mlx.py,sha256=SbYIo1C3EfZ8L30P7OTx0Dx7KuXIIpQKf8uZqR1e760,8527
54
54
  xinference/model/audio/fish_speech.py,sha256=U1NtEhQFtzVYZ0vpx10EqBnqUz-hmx1ZTAzD9OSPskQ,6767
55
55
  xinference/model/audio/funasr.py,sha256=65z7U7_F14CCP-jg6BpeY3_49FK7Y5OCRSzrhhsklCg,4075
56
+ xinference/model/audio/kokoro.py,sha256=7n2l0MWRvTz6uI4rtf-22MlrB6-HivmG8rp-JKdJCgY,4063
56
57
  xinference/model/audio/melotts.py,sha256=lJdOsce2mMOTQIslv6xq1JTEO7vC6Q9ga2xjY-_E7uw,3493
57
- xinference/model/audio/model_spec.json,sha256=rGe6_DBmit4CUCFv1y8r9qHFgiFXJePzeaIWhEyyYaE,10239
58
- xinference/model/audio/model_spec_modelscope.json,sha256=q4LKNjZ1ZG0hI-eWhLj9FWNcAVqiaLtsc6_SCrLkIVk,2788
58
+ xinference/model/audio/model_spec.json,sha256=VbeNoQIWW1qe9TOEgU_m9wlfaaDjE_iKAMMhLs-kCqM,10477
59
+ xinference/model/audio/model_spec_modelscope.json,sha256=aFCqDC74wlisqhWzt42k0vE42b8eshfz2YXhJf95nNQ,3029
59
60
  xinference/model/audio/utils.py,sha256=fnnQfZlhH6DRw6beXPUIO28u6qu2GgCONrWDIsTCNkw,1591
60
61
  xinference/model/audio/whisper.py,sha256=NePQOYy2CnVqi7g6uY9hq5tlUIxZ995FOIOPsPZCfJ8,9058
61
62
  xinference/model/audio/whisper_mlx.py,sha256=zBuCd7GUlsN9FC_-J11gqIkOCicihfbqxoabiXTvH2Q,7237
@@ -87,19 +88,19 @@ xinference/model/image/stable_diffusion/core.py,sha256=SaeBYNdRdqarEwoyjHDz_SfRI
87
88
  xinference/model/image/stable_diffusion/mlx.py,sha256=GZsozzGB04NfHAdU9MI6gwWE1t_A-s_Ddn_ic8DlkKQ,7476
88
89
  xinference/model/llm/__init__.py,sha256=zKmkoNsUoQk1_jnLHNpno4UPKGl5O_vJK5R5R-TYF9w,14140
89
90
  xinference/model/llm/core.py,sha256=g-luuAjZizrPunhyFE9IRjn57l0g6FY_1xUwtlRegbs,8151
90
- xinference/model/llm/llm_family.json,sha256=Q67g29KCIgnRQVS92ortD_RY1RhYU_47WnwFMM8GDbY,328868
91
- xinference/model/llm/llm_family.py,sha256=HiWobT1SInGXWT78g673K5-FoDkDi1dSUU4pmfFAPPI,39050
91
+ xinference/model/llm/llm_family.json,sha256=6GXYmnqKZAh2dwnaDyg1tzk583hRLy33KAI4B5rDQc8,339783
92
+ xinference/model/llm/llm_family.py,sha256=tee7rQYayo46-lK3QhrZqoGKYQM5ijA5tb6xQTPGZa0,39125
92
93
  xinference/model/llm/llm_family_csghub.json,sha256=zMKWbihsxQNVB1u5iKJbZUkbOfQ4IPNq1KQ-8IDPQQA,8759
93
- xinference/model/llm/llm_family_modelscope.json,sha256=jEp2-wzMsJYTR4xXqt5eETghsxgYgEGLiDSzc12pUzY,258710
94
+ xinference/model/llm/llm_family_modelscope.json,sha256=-JAca-vryl6svsL1Nk4gd3eq-03BFJ8KfdS3q0g7A-U,269784
94
95
  xinference/model/llm/llm_family_openmind_hub.json,sha256=jl9pfbe5DztoxgEwKBxDk1Wd7TziTiJ48_Ie_lJdYjA,67872
95
96
  xinference/model/llm/memory.py,sha256=GLNmXBI-AtMbuaJfEf50fnhN4rdbOZjLyT6L_Vjqa5g,10206
96
- xinference/model/llm/utils.py,sha256=nSI5vtOp3qfAB-RgYXNrMZH9f-X1JbNOyCBVFz21eMw,25407
97
+ xinference/model/llm/utils.py,sha256=AUm_q51rmV8HgVMcRttd8B8TA-EXxU1oJVdkLYYGhes,28641
97
98
  xinference/model/llm/llama_cpp/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
98
- xinference/model/llm/llama_cpp/core.py,sha256=vjuTapwbn-ZjUX-8WA0nFyicE4UGUSehU_csSetvcZw,10928
99
+ xinference/model/llm/llama_cpp/core.py,sha256=3GSGk42c8Oy_jTqRv4nLC482V2tUis3V0LlohQy_I1U,11312
99
100
  xinference/model/llm/lmdeploy/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
100
101
  xinference/model/llm/lmdeploy/core.py,sha256=WvSP3x6t-HBv6hKh1qWZatFAzlcZCyyKqvc3ua8yPTI,19835
101
102
  xinference/model/llm/mlx/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
102
- xinference/model/llm/mlx/core.py,sha256=rpgwZ7R2WDd-ayhi28UDglGxIutBuIXLyX7n7fHIEYw,22936
103
+ xinference/model/llm/mlx/core.py,sha256=WQN2iURiWSL_MY5hR0GkCYa08qr5wtOFx522_c2vW30,23130
103
104
  xinference/model/llm/sglang/__init__.py,sha256=-sjSIQ4K6w-TEzx49kVaWeWC443fnZqODU91GCQ_JNo,581
104
105
  xinference/model/llm/sglang/core.py,sha256=Ab0i6Q3M-DqQi5bHMyfa9klPElGSk1ThEke4mdsBHXU,16747
105
106
  xinference/model/llm/transformers/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
@@ -108,7 +109,7 @@ xinference/model/llm/transformers/cogagent.py,sha256=JbIiqVW-S9MA3d4CN2DlI7-WDB_
108
109
  xinference/model/llm/transformers/cogvlm2.py,sha256=I5Ftm0VYjbTAv5ZARZCo32Ggpw58PJfHs5B_nX_BIlU,15972
109
110
  xinference/model/llm/transformers/cogvlm2_video.py,sha256=ZGkpC4x2uEtjwoMrLSODmAUYTjOeSNYxZi9VpQrpnhU,11857
110
111
  xinference/model/llm/transformers/compression.py,sha256=U0vMJ-JaBt4oC2LffgWg6HbPj1CeUi_YdwVbjDd0mRA,8112
111
- xinference/model/llm/transformers/core.py,sha256=t1tujy8-IfD83CjKkvxZ-jqZISDQVFJpSiDkeIL4LGM,28286
112
+ xinference/model/llm/transformers/core.py,sha256=7VkOA04mpiuyA8-P62NSC74_WkKLFb_O51hIEtUHPBw,28493
112
113
  xinference/model/llm/transformers/deepseek_v2.py,sha256=-RKlI3mhja730md4evQ2vfIxBnZD5vWyrgmg_3eovms,4096
113
114
  xinference/model/llm/transformers/deepseek_vl.py,sha256=pB6i6DW5oyfHdqTgKpi2DkIKVGlPLGIDR_Op0sB1uKA,10445
114
115
  xinference/model/llm/transformers/glm4v.py,sha256=goph2HhpV8gUm2t8-T1P-jTF2r_kPeH6QNe64lmlm0g,13871
@@ -119,14 +120,14 @@ xinference/model/llm/transformers/minicpmv25.py,sha256=mr80-OlSlK_opSuAO3cz_Qlkq
119
120
  xinference/model/llm/transformers/minicpmv26.py,sha256=_e2C4vmyKIzKt7S7AvKgiqhDOhGiBXa6Xoiix4UaYtI,13440
120
121
  xinference/model/llm/transformers/omnilmm.py,sha256=2ZLW979ETqDDKo9CaTNwi9uLBZ2d6itHAYqjUA4jdro,5172
121
122
  xinference/model/llm/transformers/opt.py,sha256=dkZFNwtw_sUuVaz9He6LWfEojRGfOQFQ5atvC5OYPuY,2429
122
- xinference/model/llm/transformers/qwen2_audio.py,sha256=1XmlawVF-Xh2pgGoLDX7kOYIiF_bDUR3doSOnM59QbQ,6107
123
- xinference/model/llm/transformers/qwen2_vl.py,sha256=TKn7p0-bNzzv7ZVZ18mOD9NYgJ9q_y_CJMA3OWWvv2c,7768
123
+ xinference/model/llm/transformers/qwen2_audio.py,sha256=RkKSUFTgyolVSuFoLj2GdzsFfU8goOIYtDFLoRLiZ2s,6259
124
+ xinference/model/llm/transformers/qwen2_vl.py,sha256=HEtxPJyhfsvcULu8guLwasB0DaXRgOGxP1WrhgDUWec,8437
124
125
  xinference/model/llm/transformers/qwen_vl.py,sha256=LG19qJW30bFiZOS-t9OM3JP6K1KCLj_Sv3nKSCLvyts,14058
125
126
  xinference/model/llm/transformers/tensorizer_utils.py,sha256=VXSYbPZtCbd8lVvsnjDLPZjfCMil67Pkywd_Ze4dTx4,11362
126
- xinference/model/llm/transformers/utils.py,sha256=a4-X5P9_L--rgSx5jI8haYA6GSpKhMdOYE97VNh54yM,19389
127
+ xinference/model/llm/transformers/utils.py,sha256=GHJsjBjEXpzZAFbcfKiMxMjFnOBYeZt-eXF5S4HQ8I8,19582
127
128
  xinference/model/llm/transformers/yi_vl.py,sha256=iCdRLw-wizbU-qXXc8CT4DhC0Pt-uYg0vFwXEhAZjQg,8961
128
129
  xinference/model/llm/vllm/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
129
- xinference/model/llm/vllm/core.py,sha256=hrHFC3Q3J-MLJnyqX9HcdiIGBfxfrn_27UPsjldLSHc,37233
130
+ xinference/model/llm/vllm/core.py,sha256=p6s5qsz9564yOfkD6nG2dISenwrRAtYUaGOe5plHs04,37708
130
131
  xinference/model/llm/vllm/utils.py,sha256=LKOmwfFRrlSecawxT-uE39tC2RQbf1UIiSH9Uz90X6w,1313
131
132
  xinference/model/llm/vllm/xavier/__init__.py,sha256=CyLLkbImZouAk4lePIgKXT4WQoqyauIEwdqea5IOUVU,581
132
133
  xinference/model/llm/vllm/xavier/allocator.py,sha256=SJ2eCOxF6CWTBZIP39FRxeK6fxIE8pRshOPnSRc72d4,2691
@@ -15718,9 +15719,9 @@ xinference/web/ui/node_modules/yup/package.json,sha256=xRFSROB9NKxqSWHEVFvSTsPs9
15718
15719
  xinference/web/ui/node_modules/yup/node_modules/type-fest/package.json,sha256=JTv2zTTVgxQ2H82m1-6qEpdMv08lHjFx4Puf_MsbB_Q,1134
15719
15720
  xinference/web/ui/src/locales/en.json,sha256=MahpAAKmZPqtK5-M_kwdI9IUbBP-GcNqI0jSTVXHEE8,8169
15720
15721
  xinference/web/ui/src/locales/zh.json,sha256=9-Hu72a9FSB1ZCUMkKDzopBTh7Aer6b-3PB62cYxsOg,7933
15721
- xinference-1.2.1.dist-info/LICENSE,sha256=QwcOLU5TJoTeUhuIXzhdCEEDDvorGiC6-3YTOl4TecE,11356
15722
- xinference-1.2.1.dist-info/METADATA,sha256=NUwKoNtMeFa3TkRJMqGYa7E-HLCZqOxBysi6gNHY-c4,23919
15723
- xinference-1.2.1.dist-info/WHEEL,sha256=tZoeGjtWxWRfdplE7E3d45VPlLNQnvbKiYnx7gwAy8A,92
15724
- xinference-1.2.1.dist-info/entry_points.txt,sha256=-lDyyzqWMFQF0Rgm7VxBNz0V-bMBMQLRR3pvQ-Y8XTY,226
15725
- xinference-1.2.1.dist-info/top_level.txt,sha256=L1rQt7pl6m8tmKXpWVHzP-GtmzAxp663rXxGE7qnK00,11
15726
- xinference-1.2.1.dist-info/RECORD,,
15722
+ xinference-1.2.2.dist-info/LICENSE,sha256=QwcOLU5TJoTeUhuIXzhdCEEDDvorGiC6-3YTOl4TecE,11356
15723
+ xinference-1.2.2.dist-info/METADATA,sha256=IKf5udmeYA8AAAPMHGwry4YSehJPGvxp30AKB57P80Y,23991
15724
+ xinference-1.2.2.dist-info/WHEEL,sha256=tZoeGjtWxWRfdplE7E3d45VPlLNQnvbKiYnx7gwAy8A,92
15725
+ xinference-1.2.2.dist-info/entry_points.txt,sha256=-lDyyzqWMFQF0Rgm7VxBNz0V-bMBMQLRR3pvQ-Y8XTY,226
15726
+ xinference-1.2.2.dist-info/top_level.txt,sha256=L1rQt7pl6m8tmKXpWVHzP-GtmzAxp663rXxGE7qnK00,11
15727
+ xinference-1.2.2.dist-info/RECORD,,