xinference 1.2.1__py3-none-any.whl → 1.2.2__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of xinference might be problematic. Click here for more details.
- xinference/_version.py +3 -3
- xinference/api/restful_api.py +4 -7
- xinference/client/handlers.py +3 -0
- xinference/core/scheduler.py +4 -7
- xinference/deploy/local.py +2 -1
- xinference/model/audio/core.py +5 -0
- xinference/model/audio/kokoro.py +139 -0
- xinference/model/audio/model_spec.json +8 -0
- xinference/model/audio/model_spec_modelscope.json +9 -0
- xinference/model/llm/llama_cpp/core.py +21 -14
- xinference/model/llm/llm_family.json +306 -1
- xinference/model/llm/llm_family.py +4 -1
- xinference/model/llm/llm_family_modelscope.json +307 -3
- xinference/model/llm/mlx/core.py +11 -3
- xinference/model/llm/transformers/core.py +9 -1
- xinference/model/llm/transformers/qwen2_audio.py +3 -1
- xinference/model/llm/transformers/qwen2_vl.py +20 -3
- xinference/model/llm/transformers/utils.py +22 -11
- xinference/model/llm/utils.py +111 -1
- xinference/model/llm/vllm/core.py +13 -2
- {xinference-1.2.1.dist-info → xinference-1.2.2.dist-info}/METADATA +9 -8
- {xinference-1.2.1.dist-info → xinference-1.2.2.dist-info}/RECORD +26 -25
- {xinference-1.2.1.dist-info → xinference-1.2.2.dist-info}/LICENSE +0 -0
- {xinference-1.2.1.dist-info → xinference-1.2.2.dist-info}/WHEEL +0 -0
- {xinference-1.2.1.dist-info → xinference-1.2.2.dist-info}/entry_points.txt +0 -0
- {xinference-1.2.1.dist-info → xinference-1.2.2.dist-info}/top_level.txt +0 -0
xinference/model/llm/utils.py
CHANGED
|
@@ -11,16 +11,28 @@
|
|
|
11
11
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
12
12
|
# See the License for the specific language governing permissions and
|
|
13
13
|
# limitations under the License.
|
|
14
|
+
|
|
14
15
|
import base64
|
|
15
16
|
import functools
|
|
16
17
|
import json
|
|
17
18
|
import logging
|
|
18
19
|
import os
|
|
20
|
+
import re
|
|
19
21
|
import time
|
|
20
22
|
import typing
|
|
21
23
|
import uuid
|
|
22
24
|
from io import BytesIO
|
|
23
|
-
from typing import
|
|
25
|
+
from typing import (
|
|
26
|
+
Any,
|
|
27
|
+
AsyncGenerator,
|
|
28
|
+
Dict,
|
|
29
|
+
Iterable,
|
|
30
|
+
Iterator,
|
|
31
|
+
List,
|
|
32
|
+
Optional,
|
|
33
|
+
Tuple,
|
|
34
|
+
cast,
|
|
35
|
+
)
|
|
24
36
|
|
|
25
37
|
import requests
|
|
26
38
|
from PIL import Image
|
|
@@ -64,6 +76,18 @@ LLAMA3_TOOL_CALL_FAMILY = [
|
|
|
64
76
|
"llama-3.1-instruct",
|
|
65
77
|
]
|
|
66
78
|
|
|
79
|
+
DEEPSEEK_TOOL_CALL_FAMILY = [
|
|
80
|
+
"deepseek-r1-distill-qwen",
|
|
81
|
+
"deepseek-r1-distill-llama",
|
|
82
|
+
]
|
|
83
|
+
|
|
84
|
+
TOOL_CALL_FAMILY = (
|
|
85
|
+
QWEN_TOOL_CALL_FAMILY
|
|
86
|
+
+ GLM4_TOOL_CALL_FAMILY
|
|
87
|
+
+ LLAMA3_TOOL_CALL_FAMILY
|
|
88
|
+
+ DEEPSEEK_TOOL_CALL_FAMILY
|
|
89
|
+
)
|
|
90
|
+
|
|
67
91
|
QWEN_TOOL_CALL_SYMBOLS = ["<tool_call>", "</tool_call>"]
|
|
68
92
|
|
|
69
93
|
|
|
@@ -308,6 +332,35 @@ class ChatModelMixin:
|
|
|
308
332
|
else:
|
|
309
333
|
yield cls._to_chat_completion_chunk(chunk)
|
|
310
334
|
|
|
335
|
+
@classmethod
|
|
336
|
+
def _tools_to_messages_for_deepseek(
|
|
337
|
+
cls, messages: List[dict], tools: Iterable[dict]
|
|
338
|
+
):
|
|
339
|
+
# deepseek integrates tool calls into messages
|
|
340
|
+
# we follow the chat template rule to integrate tools into messages
|
|
341
|
+
tool_call_message: Dict[str, Any] = {
|
|
342
|
+
"role": "assistant",
|
|
343
|
+
"content": None,
|
|
344
|
+
"tool_calls": [],
|
|
345
|
+
}
|
|
346
|
+
|
|
347
|
+
for tool in tools:
|
|
348
|
+
function_name = tool["function"]["name"]
|
|
349
|
+
parameters = tool["function"].get("parameters", {}).get("properties", {})
|
|
350
|
+
function_args_json = json.dumps(parameters)
|
|
351
|
+
|
|
352
|
+
tool_call_message["tool_calls"].append(
|
|
353
|
+
{
|
|
354
|
+
"type": "function",
|
|
355
|
+
"function": {
|
|
356
|
+
"name": function_name,
|
|
357
|
+
"arguments": function_args_json,
|
|
358
|
+
},
|
|
359
|
+
}
|
|
360
|
+
)
|
|
361
|
+
|
|
362
|
+
messages.append(tool_call_message)
|
|
363
|
+
|
|
311
364
|
@classmethod
|
|
312
365
|
async def _async_to_chat_completion_chunks(
|
|
313
366
|
cls,
|
|
@@ -401,6 +454,61 @@ class ChatModelMixin:
|
|
|
401
454
|
except Exception:
|
|
402
455
|
return [(text, None, None)]
|
|
403
456
|
|
|
457
|
+
@classmethod
|
|
458
|
+
def _eval_deepseek_chat_arguments(cls, c) -> List[Tuple]:
|
|
459
|
+
"""
|
|
460
|
+
Parses tool calls from deepseek-r1 format and removes duplicates.
|
|
461
|
+
|
|
462
|
+
Returns:
|
|
463
|
+
List[Tuple[Optional[str], Optional[str], Optional[dict]]]
|
|
464
|
+
- (None, function_name, arguments) if successfully parsed.
|
|
465
|
+
- (content, None, None) if parsing failed (content is raw JSON text).
|
|
466
|
+
|
|
467
|
+
Example input:
|
|
468
|
+
<|tool▁call|>get_current_weather
|
|
469
|
+
```json
|
|
470
|
+
{"location": "tokyo", "unit": "fahrenheit"}
|
|
471
|
+
```
|
|
472
|
+
|
|
473
|
+
Output:
|
|
474
|
+
[
|
|
475
|
+
(None, "get_current_weather", {"location": "tokyo", "unit": "fahrenheit"})
|
|
476
|
+
]
|
|
477
|
+
"""
|
|
478
|
+
|
|
479
|
+
text = c["choices"][0]["text"]
|
|
480
|
+
|
|
481
|
+
pattern = r"<|tool▁call|>(\w+)\s*```json\s*(.*?)\s*```"
|
|
482
|
+
matches = re.findall(pattern, text, re.DOTALL)
|
|
483
|
+
|
|
484
|
+
if not matches:
|
|
485
|
+
return [(text, None, None)]
|
|
486
|
+
|
|
487
|
+
tool_calls = set() # Used for deduplication
|
|
488
|
+
results = []
|
|
489
|
+
|
|
490
|
+
for function_name, args_json in matches:
|
|
491
|
+
try:
|
|
492
|
+
arguments = json.loads(args_json)
|
|
493
|
+
# Convert dictionary to frozenset for deduplication
|
|
494
|
+
arguments_hashable = frozenset(arguments.items())
|
|
495
|
+
tool_call_tuple = (None, function_name, arguments)
|
|
496
|
+
except json.JSONDecodeError:
|
|
497
|
+
tool_call_tuple = (
|
|
498
|
+
args_json,
|
|
499
|
+
None,
|
|
500
|
+
None,
|
|
501
|
+
) # If parsing fails, treat as raw content
|
|
502
|
+
arguments_hashable = None # No need for hashing
|
|
503
|
+
|
|
504
|
+
# Avoid duplicate entries
|
|
505
|
+
dedup_key = (function_name, arguments_hashable)
|
|
506
|
+
if dedup_key not in tool_calls:
|
|
507
|
+
tool_calls.add(dedup_key)
|
|
508
|
+
results.append(tool_call_tuple)
|
|
509
|
+
|
|
510
|
+
return results
|
|
511
|
+
|
|
404
512
|
@classmethod
|
|
405
513
|
def _eval_tool_arguments(cls, model_family, c):
|
|
406
514
|
family = model_family.model_family or model_family.model_name
|
|
@@ -410,6 +518,8 @@ class ChatModelMixin:
|
|
|
410
518
|
result = cls._eval_qwen_chat_arguments(c)
|
|
411
519
|
elif family in LLAMA3_TOOL_CALL_FAMILY:
|
|
412
520
|
result = cls._eval_llama3_chat_arguments(c)
|
|
521
|
+
elif family in DEEPSEEK_TOOL_CALL_FAMILY:
|
|
522
|
+
result = cls._eval_deepseek_chat_arguments(c)
|
|
413
523
|
else:
|
|
414
524
|
raise Exception(
|
|
415
525
|
f"Model {model_family.model_name} is not support tool calls."
|
|
@@ -44,6 +44,7 @@ from ....types import (
|
|
|
44
44
|
from .. import LLM, LLMFamilyV1, LLMSpecV1
|
|
45
45
|
from ..llm_family import CustomLLMFamilyV1
|
|
46
46
|
from ..utils import (
|
|
47
|
+
DEEPSEEK_TOOL_CALL_FAMILY,
|
|
47
48
|
QWEN_TOOL_CALL_FAMILY,
|
|
48
49
|
QWEN_TOOL_CALL_SYMBOLS,
|
|
49
50
|
ChatModelMixin,
|
|
@@ -185,6 +186,7 @@ if VLLM_INSTALLED and vllm.__version__ > "0.5.3":
|
|
|
185
186
|
VLLM_SUPPORTED_MODELS.append("llama-3.1")
|
|
186
187
|
VLLM_SUPPORTED_CHAT_MODELS.append("llama-3.1-instruct")
|
|
187
188
|
VLLM_SUPPORTED_CHAT_MODELS.append("llama-3.3-instruct")
|
|
189
|
+
VLLM_SUPPORTED_CHAT_MODELS.append("deepseek-r1-distill-llama")
|
|
188
190
|
|
|
189
191
|
if VLLM_INSTALLED and vllm.__version__ >= "0.6.1":
|
|
190
192
|
VLLM_SUPPORTED_VISION_MODEL_LIST.append("internvl2")
|
|
@@ -198,6 +200,12 @@ if VLLM_INSTALLED and vllm.__version__ >= "0.6.3":
|
|
|
198
200
|
VLLM_SUPPORTED_VISION_MODEL_LIST.append("qwen2-vl-instruct")
|
|
199
201
|
VLLM_SUPPORTED_VISION_MODEL_LIST.append("QvQ-72B-Preview")
|
|
200
202
|
|
|
203
|
+
if VLLM_INSTALLED and vllm.__version__ >= "0.7.0":
|
|
204
|
+
VLLM_SUPPORTED_CHAT_MODELS.append("internlm3-instruct")
|
|
205
|
+
|
|
206
|
+
if VLLM_INSTALLED and vllm.__version__ >= "0.7.2":
|
|
207
|
+
VLLM_SUPPORTED_VISION_MODEL_LIST.append("qwen2.5-vl-instruct")
|
|
208
|
+
|
|
201
209
|
|
|
202
210
|
class VLLMModel(LLM):
|
|
203
211
|
def __init__(
|
|
@@ -807,8 +815,11 @@ class VLLMChatModel(VLLMModel, ChatModelMixin):
|
|
|
807
815
|
tools = generate_config.pop("tools", []) if generate_config else None
|
|
808
816
|
model_family = self.model_family.model_family or self.model_family.model_name
|
|
809
817
|
full_context_kwargs = {}
|
|
810
|
-
if tools
|
|
811
|
-
|
|
818
|
+
if tools:
|
|
819
|
+
if model_family in QWEN_TOOL_CALL_FAMILY:
|
|
820
|
+
full_context_kwargs["tools"] = tools
|
|
821
|
+
elif model_family in DEEPSEEK_TOOL_CALL_FAMILY:
|
|
822
|
+
self._tools_to_messages_for_deepseek(messages, tools)
|
|
812
823
|
assert self.model_family.chat_template is not None
|
|
813
824
|
full_prompt = self.get_full_context(
|
|
814
825
|
messages, self.model_family.chat_template, **full_context_kwargs
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.1
|
|
2
2
|
Name: xinference
|
|
3
|
-
Version: 1.2.
|
|
3
|
+
Version: 1.2.2
|
|
4
4
|
Summary: Model Serving Made Easy
|
|
5
5
|
Home-page: https://github.com/xorbitsai/inference
|
|
6
6
|
Author: Qin Xuye
|
|
@@ -98,7 +98,7 @@ Requires-Dist: tomli; extra == "all"
|
|
|
98
98
|
Requires-Dist: vocos; extra == "all"
|
|
99
99
|
Requires-Dist: jieba; extra == "all"
|
|
100
100
|
Requires-Dist: soundfile; extra == "all"
|
|
101
|
-
Requires-Dist: qwen-vl-utils; extra == "all"
|
|
101
|
+
Requires-Dist: qwen-vl-utils!=0.0.9; extra == "all"
|
|
102
102
|
Requires-Dist: datamodel-code-generator; extra == "all"
|
|
103
103
|
Requires-Dist: jsonschema; extra == "all"
|
|
104
104
|
Requires-Dist: verovio>=4.3.1; extra == "all"
|
|
@@ -154,6 +154,7 @@ Requires-Dist: fugashi; extra == "audio"
|
|
|
154
154
|
Requires-Dist: g2p-en; extra == "audio"
|
|
155
155
|
Requires-Dist: anyascii; extra == "audio"
|
|
156
156
|
Requires-Dist: gruut[de,es,fr]; extra == "audio"
|
|
157
|
+
Requires-Dist: kokoro>=0.7.9; extra == "audio"
|
|
157
158
|
Requires-Dist: nemo-text-processing<1.1.0; sys_platform == "linux" and extra == "audio"
|
|
158
159
|
Requires-Dist: WeTextProcessing<1.0.4; sys_platform == "linux" and extra == "audio"
|
|
159
160
|
Provides-Extra: benchmark
|
|
@@ -211,7 +212,7 @@ Requires-Dist: mlx-lm>=0.21.1; extra == "mlx"
|
|
|
211
212
|
Requires-Dist: mlx-vlm>=0.1.11; extra == "mlx"
|
|
212
213
|
Requires-Dist: mlx-whisper; extra == "mlx"
|
|
213
214
|
Requires-Dist: f5-tts-mlx; extra == "mlx"
|
|
214
|
-
Requires-Dist: qwen-vl-utils; extra == "mlx"
|
|
215
|
+
Requires-Dist: qwen-vl-utils!=0.0.9; extra == "mlx"
|
|
215
216
|
Requires-Dist: tomli; extra == "mlx"
|
|
216
217
|
Provides-Extra: rerank
|
|
217
218
|
Requires-Dist: FlagEmbedding; extra == "rerank"
|
|
@@ -225,7 +226,6 @@ Requires-Dist: torch; extra == "transformers"
|
|
|
225
226
|
Requires-Dist: accelerate>=0.28.0; extra == "transformers"
|
|
226
227
|
Requires-Dist: sentencepiece; extra == "transformers"
|
|
227
228
|
Requires-Dist: transformers-stream-generator; extra == "transformers"
|
|
228
|
-
Requires-Dist: bitsandbytes; extra == "transformers"
|
|
229
229
|
Requires-Dist: protobuf; extra == "transformers"
|
|
230
230
|
Requires-Dist: einops; extra == "transformers"
|
|
231
231
|
Requires-Dist: tiktoken; extra == "transformers"
|
|
@@ -236,11 +236,12 @@ Requires-Dist: torchvision; extra == "transformers"
|
|
|
236
236
|
Requires-Dist: peft; extra == "transformers"
|
|
237
237
|
Requires-Dist: eva-decord; extra == "transformers"
|
|
238
238
|
Requires-Dist: jj-pytorchvideo; extra == "transformers"
|
|
239
|
-
Requires-Dist: qwen-vl-utils; extra == "transformers"
|
|
239
|
+
Requires-Dist: qwen-vl-utils!=0.0.9; extra == "transformers"
|
|
240
240
|
Requires-Dist: datamodel-code-generator; extra == "transformers"
|
|
241
241
|
Requires-Dist: jsonschema; extra == "transformers"
|
|
242
242
|
Requires-Dist: auto-gptq; sys_platform != "darwin" and extra == "transformers"
|
|
243
243
|
Requires-Dist: autoawq<0.2.6; sys_platform != "darwin" and extra == "transformers"
|
|
244
|
+
Requires-Dist: bitsandbytes; sys_platform == "linux" and extra == "transformers"
|
|
244
245
|
Provides-Extra: video
|
|
245
246
|
Requires-Dist: diffusers>=0.32.0; extra == "video"
|
|
246
247
|
Requires-Dist: imageio-ffmpeg; extra == "video"
|
|
@@ -296,14 +297,14 @@ potential of cutting-edge AI models.
|
|
|
296
297
|
- Support speech recognition model: [#929](https://github.com/xorbitsai/inference/pull/929)
|
|
297
298
|
- Metrics support: [#906](https://github.com/xorbitsai/inference/pull/906)
|
|
298
299
|
### New Models
|
|
300
|
+
- Built-in support for [DeepSeek-R1-Distill-Qwen](https://github.com/deepseek-ai/DeepSeek-R1?tab=readme-ov-file#deepseek-r1-distill-models): [#2781](https://github.com/xorbitsai/inference/pull/2781)
|
|
301
|
+
- Built-in support for [qwen2.5-vl](https://github.com/QwenLM/Qwen2.5-VL): [#2788](https://github.com/xorbitsai/inference/pull/2788)
|
|
302
|
+
- Built-in support for [internlm3-instruct](https://github.com/InternLM/InternLM): [#2789](https://github.com/xorbitsai/inference/pull/2789)
|
|
299
303
|
- Built-in support for [MeloTTS](https://github.com/myshell-ai/MeloTTS): [#2760](https://github.com/xorbitsai/inference/pull/2760)
|
|
300
304
|
- Built-in support for [CogAgent](https://github.com/THUDM/CogAgent): [#2740](https://github.com/xorbitsai/inference/pull/2740)
|
|
301
305
|
- Built-in support for [HunyuanVideo](https://github.com/Tencent/HunyuanVideo): [#2721](https://github.com/xorbitsai/inference/pull/2721)
|
|
302
306
|
- Built-in support for [HunyuanDiT](https://github.com/Tencent/HunyuanDiT): [#2727](https://github.com/xorbitsai/inference/pull/2727)
|
|
303
307
|
- Built-in support for [Macro-o1](https://github.com/AIDC-AI/Marco-o1): [#2749](https://github.com/xorbitsai/inference/pull/2749)
|
|
304
|
-
- Built-in support for [Stable Diffusion 3.5](https://huggingface.co/collections/stabilityai/stable-diffusion-35-671785cca799084f71fa2838): [#2706](https://github.com/xorbitsai/inference/pull/2706)
|
|
305
|
-
- Built-in support for [CosyVoice 2](https://huggingface.co/FunAudioLLM/CosyVoice2-0.5B): [#2684](https://github.com/xorbitsai/inference/pull/2684)
|
|
306
|
-
- Built-in support for [Fish Speech V1.5](https://huggingface.co/fishaudio/fish-speech-1.5): [#2672](https://github.com/xorbitsai/inference/pull/2672)
|
|
307
308
|
### Integrations
|
|
308
309
|
- [Dify](https://docs.dify.ai/advanced/model-configuration/xinference): an LLMOps platform that enables developers (and even non-developers) to quickly build useful applications based on large language models, ensuring they are visual, operable, and improvable.
|
|
309
310
|
- [FastGPT](https://github.com/labring/FastGPT): a knowledge-based platform built on the LLM, offers out-of-the-box data processing and model invocation capabilities, allows for workflow orchestration through Flow visualization.
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
xinference/__init__.py,sha256=nmTTrYbIpj964ZF6ojtgOM7E85JBOj1EyQbmYjbj1jw,915
|
|
2
2
|
xinference/_compat.py,sha256=URSJQLXrcsTO9B_4x0wVDPijYQDhuVJmZ95npID560w,4197
|
|
3
|
-
xinference/_version.py,sha256=
|
|
3
|
+
xinference/_version.py,sha256=tyoMIoWH7JrsR3odtmfwgrhewV2IdEEE1IRf42ATx28,497
|
|
4
4
|
xinference/conftest.py,sha256=vETDpRBVIlWbWi7OTwf7og89U25KyYGyI7yPIB3O8N8,9564
|
|
5
5
|
xinference/constants.py,sha256=mEW4HDzjXtDXN61Mt6TtJrJ4ljbB6VUkh97e3oDbNx4,3905
|
|
6
6
|
xinference/device_utils.py,sha256=zswJiws3VyTIaNO8z-MOcsJH_UiPoePPiKK5zoNrjTA,3285
|
|
@@ -9,14 +9,14 @@ xinference/isolation.py,sha256=gTU1em5fxg1m-7hxieWBMZvVkXZX4GZYmeT7XxXsYrU,2699
|
|
|
9
9
|
xinference/types.py,sha256=p9IV2vmbtFfpo9P66IynkEgU2N6310uVMKBuq0rbtNc,13102
|
|
10
10
|
xinference/utils.py,sha256=zYgf9bCvfbybRt3gEog6r5WJCpj0czZCf0qgRdYjkN8,720
|
|
11
11
|
xinference/api/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
|
|
12
|
-
xinference/api/restful_api.py,sha256=
|
|
12
|
+
xinference/api/restful_api.py,sha256=XENr9892-WHZ1XKaCwijdTAAd1uX2D0oMKUF6Pw99GQ,92730
|
|
13
13
|
xinference/api/oauth2/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
|
|
14
14
|
xinference/api/oauth2/auth_service.py,sha256=74JzB42fbbmBu4Q1dW3A9Fp_N7167KgRGB42Z0NHjAM,6119
|
|
15
15
|
xinference/api/oauth2/types.py,sha256=K923sv_XySIUtM2Eozl9IG082IJcDOS5SFLrPZ5ELBg,996
|
|
16
16
|
xinference/api/oauth2/utils.py,sha256=SIiiUj6VuTEsj3bZ2TYUyhx3cGlLSX3ZNWDOgUwRtXc,1410
|
|
17
17
|
xinference/client/__init__.py,sha256=Gc4HOzAy_1cic5kXlso7hahYgw89CKvZSJDicEU461k,669
|
|
18
18
|
xinference/client/common.py,sha256=iciZRs5YjM2gYsXnwACPMaiBZp4_XpawWwfym0Iyu40,1617
|
|
19
|
-
xinference/client/handlers.py,sha256=
|
|
19
|
+
xinference/client/handlers.py,sha256=HYmQjG0BPK5PKcWVEDJqQz5iLjS8yjOq5j-HP_CAi_k,630
|
|
20
20
|
xinference/client/restful/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
|
|
21
21
|
xinference/client/restful/restful_client.py,sha256=JwzP7etUZBR0mmU7y3dUOEWN_D7ol_2hXN9KMAzKZaw,53601
|
|
22
22
|
xinference/core/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
|
|
@@ -28,14 +28,14 @@ xinference/core/metrics.py,sha256=ScmTG15Uq3h_ob72ybZSMWdnk8P4sUZFcm60f4ikSXc,26
|
|
|
28
28
|
xinference/core/model.py,sha256=8FaEJK4tTL7wvYivLEqWKkZOrAMODBSoxg9A4fYzUH0,43365
|
|
29
29
|
xinference/core/progress_tracker.py,sha256=LIF6CLIlnEoSBkuDCraJktDOzZ31mQ4HOo6EVr3KpQM,6453
|
|
30
30
|
xinference/core/resource.py,sha256=FQ0aRt3T4ZQo0P6CZZf5QUKHiCsr5llBvKb1f7wfnxg,1611
|
|
31
|
-
xinference/core/scheduler.py,sha256=
|
|
31
|
+
xinference/core/scheduler.py,sha256=WHWOlbzAPoURXNzjFNdfvhkgBYO8ufZtIu3JXfq7fmY,15576
|
|
32
32
|
xinference/core/status_guard.py,sha256=4an1KjUOhCStgRQUw1VSzXcycXUtvhxwiMREKKcl1UI,2828
|
|
33
33
|
xinference/core/supervisor.py,sha256=4beE9EZx5xWpSLZtZwT3CPYWT25zPFl9gRj2oK8Za3w,60208
|
|
34
34
|
xinference/core/utils.py,sha256=RR3XDioh52Wy8vmAqhCZ6EQskr-HPDfUp0vCCAPIAvs,11302
|
|
35
35
|
xinference/core/worker.py,sha256=FVy_zXOHovYLPQQr6DSpjSGWaLyH425gTUUH_vwKVfk,50666
|
|
36
36
|
xinference/deploy/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
|
|
37
37
|
xinference/deploy/cmdline.py,sha256=gPwp9IngaXCypUEnPDS_22U8GntsKr7qHDST7duyAoI,48478
|
|
38
|
-
xinference/deploy/local.py,sha256=
|
|
38
|
+
xinference/deploy/local.py,sha256=sO3BcpEH9oCF87CxWVA4AXAYcfHGnrcop40ew5NOA2g,3979
|
|
39
39
|
xinference/deploy/supervisor.py,sha256=68rB2Ey5KFeF6zto9YGbw3P8QLZmF_KSh1NwH_pNP4w,2986
|
|
40
40
|
xinference/deploy/utils.py,sha256=jdL7i2WV6u_BZ8IiE1d3YktvCARcB3ntzMQ5rHGD5DM,6756
|
|
41
41
|
xinference/deploy/worker.py,sha256=VQ71ClWpeGsyFgDmcOes2ub1cil10cBjhFLHYeuVwC4,2974
|
|
@@ -46,16 +46,17 @@ xinference/model/core.py,sha256=_NEH4wkjjJgRDdLHNVY_hL3V0kT67CvTay89uIzx1Ns,4736
|
|
|
46
46
|
xinference/model/utils.py,sha256=_yJ5h4RUzt7Kjs2WdjSzbVM3FTWEkX0ycOnXANZ9KVg,11394
|
|
47
47
|
xinference/model/audio/__init__.py,sha256=KasWsaNPeij6sGpHKqXaUc_bxUw1yYbD7-fwxkcoAVE,3731
|
|
48
48
|
xinference/model/audio/chattts.py,sha256=ny3DZTCTt2MzdkLw994_QHZ_4qIEUZcNexNJkCejCyo,4998
|
|
49
|
-
xinference/model/audio/core.py,sha256=
|
|
49
|
+
xinference/model/audio/core.py,sha256=AAuss7fGwGktcB_aGRwuR4TJIM6-Rrm0YCU752QoxfQ,7710
|
|
50
50
|
xinference/model/audio/cosyvoice.py,sha256=vw7OR2t7zNimQn3Q74iiL1-2en5o6gvdcZsDgkmYpy4,7796
|
|
51
51
|
xinference/model/audio/custom.py,sha256=8GXBRmTtR-GY03-E91nlRGTIuabCRzlt20ecU6Un6Y8,4985
|
|
52
52
|
xinference/model/audio/f5tts.py,sha256=RyfEYVvKhV7JygIv4F45C8wBr-u0mx9WpXj3gIOtL7o,6809
|
|
53
53
|
xinference/model/audio/f5tts_mlx.py,sha256=SbYIo1C3EfZ8L30P7OTx0Dx7KuXIIpQKf8uZqR1e760,8527
|
|
54
54
|
xinference/model/audio/fish_speech.py,sha256=U1NtEhQFtzVYZ0vpx10EqBnqUz-hmx1ZTAzD9OSPskQ,6767
|
|
55
55
|
xinference/model/audio/funasr.py,sha256=65z7U7_F14CCP-jg6BpeY3_49FK7Y5OCRSzrhhsklCg,4075
|
|
56
|
+
xinference/model/audio/kokoro.py,sha256=7n2l0MWRvTz6uI4rtf-22MlrB6-HivmG8rp-JKdJCgY,4063
|
|
56
57
|
xinference/model/audio/melotts.py,sha256=lJdOsce2mMOTQIslv6xq1JTEO7vC6Q9ga2xjY-_E7uw,3493
|
|
57
|
-
xinference/model/audio/model_spec.json,sha256=
|
|
58
|
-
xinference/model/audio/model_spec_modelscope.json,sha256=
|
|
58
|
+
xinference/model/audio/model_spec.json,sha256=VbeNoQIWW1qe9TOEgU_m9wlfaaDjE_iKAMMhLs-kCqM,10477
|
|
59
|
+
xinference/model/audio/model_spec_modelscope.json,sha256=aFCqDC74wlisqhWzt42k0vE42b8eshfz2YXhJf95nNQ,3029
|
|
59
60
|
xinference/model/audio/utils.py,sha256=fnnQfZlhH6DRw6beXPUIO28u6qu2GgCONrWDIsTCNkw,1591
|
|
60
61
|
xinference/model/audio/whisper.py,sha256=NePQOYy2CnVqi7g6uY9hq5tlUIxZ995FOIOPsPZCfJ8,9058
|
|
61
62
|
xinference/model/audio/whisper_mlx.py,sha256=zBuCd7GUlsN9FC_-J11gqIkOCicihfbqxoabiXTvH2Q,7237
|
|
@@ -87,19 +88,19 @@ xinference/model/image/stable_diffusion/core.py,sha256=SaeBYNdRdqarEwoyjHDz_SfRI
|
|
|
87
88
|
xinference/model/image/stable_diffusion/mlx.py,sha256=GZsozzGB04NfHAdU9MI6gwWE1t_A-s_Ddn_ic8DlkKQ,7476
|
|
88
89
|
xinference/model/llm/__init__.py,sha256=zKmkoNsUoQk1_jnLHNpno4UPKGl5O_vJK5R5R-TYF9w,14140
|
|
89
90
|
xinference/model/llm/core.py,sha256=g-luuAjZizrPunhyFE9IRjn57l0g6FY_1xUwtlRegbs,8151
|
|
90
|
-
xinference/model/llm/llm_family.json,sha256=
|
|
91
|
-
xinference/model/llm/llm_family.py,sha256=
|
|
91
|
+
xinference/model/llm/llm_family.json,sha256=6GXYmnqKZAh2dwnaDyg1tzk583hRLy33KAI4B5rDQc8,339783
|
|
92
|
+
xinference/model/llm/llm_family.py,sha256=tee7rQYayo46-lK3QhrZqoGKYQM5ijA5tb6xQTPGZa0,39125
|
|
92
93
|
xinference/model/llm/llm_family_csghub.json,sha256=zMKWbihsxQNVB1u5iKJbZUkbOfQ4IPNq1KQ-8IDPQQA,8759
|
|
93
|
-
xinference/model/llm/llm_family_modelscope.json,sha256
|
|
94
|
+
xinference/model/llm/llm_family_modelscope.json,sha256=-JAca-vryl6svsL1Nk4gd3eq-03BFJ8KfdS3q0g7A-U,269784
|
|
94
95
|
xinference/model/llm/llm_family_openmind_hub.json,sha256=jl9pfbe5DztoxgEwKBxDk1Wd7TziTiJ48_Ie_lJdYjA,67872
|
|
95
96
|
xinference/model/llm/memory.py,sha256=GLNmXBI-AtMbuaJfEf50fnhN4rdbOZjLyT6L_Vjqa5g,10206
|
|
96
|
-
xinference/model/llm/utils.py,sha256=
|
|
97
|
+
xinference/model/llm/utils.py,sha256=AUm_q51rmV8HgVMcRttd8B8TA-EXxU1oJVdkLYYGhes,28641
|
|
97
98
|
xinference/model/llm/llama_cpp/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
|
|
98
|
-
xinference/model/llm/llama_cpp/core.py,sha256=
|
|
99
|
+
xinference/model/llm/llama_cpp/core.py,sha256=3GSGk42c8Oy_jTqRv4nLC482V2tUis3V0LlohQy_I1U,11312
|
|
99
100
|
xinference/model/llm/lmdeploy/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
100
101
|
xinference/model/llm/lmdeploy/core.py,sha256=WvSP3x6t-HBv6hKh1qWZatFAzlcZCyyKqvc3ua8yPTI,19835
|
|
101
102
|
xinference/model/llm/mlx/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
|
|
102
|
-
xinference/model/llm/mlx/core.py,sha256=
|
|
103
|
+
xinference/model/llm/mlx/core.py,sha256=WQN2iURiWSL_MY5hR0GkCYa08qr5wtOFx522_c2vW30,23130
|
|
103
104
|
xinference/model/llm/sglang/__init__.py,sha256=-sjSIQ4K6w-TEzx49kVaWeWC443fnZqODU91GCQ_JNo,581
|
|
104
105
|
xinference/model/llm/sglang/core.py,sha256=Ab0i6Q3M-DqQi5bHMyfa9klPElGSk1ThEke4mdsBHXU,16747
|
|
105
106
|
xinference/model/llm/transformers/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
|
|
@@ -108,7 +109,7 @@ xinference/model/llm/transformers/cogagent.py,sha256=JbIiqVW-S9MA3d4CN2DlI7-WDB_
|
|
|
108
109
|
xinference/model/llm/transformers/cogvlm2.py,sha256=I5Ftm0VYjbTAv5ZARZCo32Ggpw58PJfHs5B_nX_BIlU,15972
|
|
109
110
|
xinference/model/llm/transformers/cogvlm2_video.py,sha256=ZGkpC4x2uEtjwoMrLSODmAUYTjOeSNYxZi9VpQrpnhU,11857
|
|
110
111
|
xinference/model/llm/transformers/compression.py,sha256=U0vMJ-JaBt4oC2LffgWg6HbPj1CeUi_YdwVbjDd0mRA,8112
|
|
111
|
-
xinference/model/llm/transformers/core.py,sha256=
|
|
112
|
+
xinference/model/llm/transformers/core.py,sha256=7VkOA04mpiuyA8-P62NSC74_WkKLFb_O51hIEtUHPBw,28493
|
|
112
113
|
xinference/model/llm/transformers/deepseek_v2.py,sha256=-RKlI3mhja730md4evQ2vfIxBnZD5vWyrgmg_3eovms,4096
|
|
113
114
|
xinference/model/llm/transformers/deepseek_vl.py,sha256=pB6i6DW5oyfHdqTgKpi2DkIKVGlPLGIDR_Op0sB1uKA,10445
|
|
114
115
|
xinference/model/llm/transformers/glm4v.py,sha256=goph2HhpV8gUm2t8-T1P-jTF2r_kPeH6QNe64lmlm0g,13871
|
|
@@ -119,14 +120,14 @@ xinference/model/llm/transformers/minicpmv25.py,sha256=mr80-OlSlK_opSuAO3cz_Qlkq
|
|
|
119
120
|
xinference/model/llm/transformers/minicpmv26.py,sha256=_e2C4vmyKIzKt7S7AvKgiqhDOhGiBXa6Xoiix4UaYtI,13440
|
|
120
121
|
xinference/model/llm/transformers/omnilmm.py,sha256=2ZLW979ETqDDKo9CaTNwi9uLBZ2d6itHAYqjUA4jdro,5172
|
|
121
122
|
xinference/model/llm/transformers/opt.py,sha256=dkZFNwtw_sUuVaz9He6LWfEojRGfOQFQ5atvC5OYPuY,2429
|
|
122
|
-
xinference/model/llm/transformers/qwen2_audio.py,sha256=
|
|
123
|
-
xinference/model/llm/transformers/qwen2_vl.py,sha256=
|
|
123
|
+
xinference/model/llm/transformers/qwen2_audio.py,sha256=RkKSUFTgyolVSuFoLj2GdzsFfU8goOIYtDFLoRLiZ2s,6259
|
|
124
|
+
xinference/model/llm/transformers/qwen2_vl.py,sha256=HEtxPJyhfsvcULu8guLwasB0DaXRgOGxP1WrhgDUWec,8437
|
|
124
125
|
xinference/model/llm/transformers/qwen_vl.py,sha256=LG19qJW30bFiZOS-t9OM3JP6K1KCLj_Sv3nKSCLvyts,14058
|
|
125
126
|
xinference/model/llm/transformers/tensorizer_utils.py,sha256=VXSYbPZtCbd8lVvsnjDLPZjfCMil67Pkywd_Ze4dTx4,11362
|
|
126
|
-
xinference/model/llm/transformers/utils.py,sha256=
|
|
127
|
+
xinference/model/llm/transformers/utils.py,sha256=GHJsjBjEXpzZAFbcfKiMxMjFnOBYeZt-eXF5S4HQ8I8,19582
|
|
127
128
|
xinference/model/llm/transformers/yi_vl.py,sha256=iCdRLw-wizbU-qXXc8CT4DhC0Pt-uYg0vFwXEhAZjQg,8961
|
|
128
129
|
xinference/model/llm/vllm/__init__.py,sha256=h_JgzSqV5lP6vQ6XX_17kE4IY4BRnvKta_7VLQAL1ms,581
|
|
129
|
-
xinference/model/llm/vllm/core.py,sha256=
|
|
130
|
+
xinference/model/llm/vllm/core.py,sha256=p6s5qsz9564yOfkD6nG2dISenwrRAtYUaGOe5plHs04,37708
|
|
130
131
|
xinference/model/llm/vllm/utils.py,sha256=LKOmwfFRrlSecawxT-uE39tC2RQbf1UIiSH9Uz90X6w,1313
|
|
131
132
|
xinference/model/llm/vllm/xavier/__init__.py,sha256=CyLLkbImZouAk4lePIgKXT4WQoqyauIEwdqea5IOUVU,581
|
|
132
133
|
xinference/model/llm/vllm/xavier/allocator.py,sha256=SJ2eCOxF6CWTBZIP39FRxeK6fxIE8pRshOPnSRc72d4,2691
|
|
@@ -15718,9 +15719,9 @@ xinference/web/ui/node_modules/yup/package.json,sha256=xRFSROB9NKxqSWHEVFvSTsPs9
|
|
|
15718
15719
|
xinference/web/ui/node_modules/yup/node_modules/type-fest/package.json,sha256=JTv2zTTVgxQ2H82m1-6qEpdMv08lHjFx4Puf_MsbB_Q,1134
|
|
15719
15720
|
xinference/web/ui/src/locales/en.json,sha256=MahpAAKmZPqtK5-M_kwdI9IUbBP-GcNqI0jSTVXHEE8,8169
|
|
15720
15721
|
xinference/web/ui/src/locales/zh.json,sha256=9-Hu72a9FSB1ZCUMkKDzopBTh7Aer6b-3PB62cYxsOg,7933
|
|
15721
|
-
xinference-1.2.
|
|
15722
|
-
xinference-1.2.
|
|
15723
|
-
xinference-1.2.
|
|
15724
|
-
xinference-1.2.
|
|
15725
|
-
xinference-1.2.
|
|
15726
|
-
xinference-1.2.
|
|
15722
|
+
xinference-1.2.2.dist-info/LICENSE,sha256=QwcOLU5TJoTeUhuIXzhdCEEDDvorGiC6-3YTOl4TecE,11356
|
|
15723
|
+
xinference-1.2.2.dist-info/METADATA,sha256=IKf5udmeYA8AAAPMHGwry4YSehJPGvxp30AKB57P80Y,23991
|
|
15724
|
+
xinference-1.2.2.dist-info/WHEEL,sha256=tZoeGjtWxWRfdplE7E3d45VPlLNQnvbKiYnx7gwAy8A,92
|
|
15725
|
+
xinference-1.2.2.dist-info/entry_points.txt,sha256=-lDyyzqWMFQF0Rgm7VxBNz0V-bMBMQLRR3pvQ-Y8XTY,226
|
|
15726
|
+
xinference-1.2.2.dist-info/top_level.txt,sha256=L1rQt7pl6m8tmKXpWVHzP-GtmzAxp663rXxGE7qnK00,11
|
|
15727
|
+
xinference-1.2.2.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|