xinference 0.11.2.post1__py3-none-any.whl → 0.12.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of xinference might be problematic. Click here for more details.
- xinference/_version.py +3 -3
- xinference/api/restful_api.py +83 -8
- xinference/client/restful/restful_client.py +70 -0
- xinference/constants.py +8 -0
- xinference/core/__init__.py +0 -2
- xinference/core/cache_tracker.py +22 -1
- xinference/core/chat_interface.py +71 -10
- xinference/core/model.py +141 -12
- xinference/core/scheduler.py +428 -0
- xinference/core/supervisor.py +31 -3
- xinference/core/worker.py +8 -3
- xinference/isolation.py +9 -2
- xinference/model/audio/chattts.py +84 -0
- xinference/model/audio/core.py +10 -3
- xinference/model/audio/model_spec.json +20 -0
- xinference/model/llm/__init__.py +6 -0
- xinference/model/llm/llm_family.json +1063 -260
- xinference/model/llm/llm_family_modelscope.json +686 -13
- xinference/model/llm/pytorch/baichuan.py +2 -1
- xinference/model/llm/pytorch/chatglm.py +2 -1
- xinference/model/llm/pytorch/cogvlm2.py +316 -0
- xinference/model/llm/pytorch/core.py +92 -6
- xinference/model/llm/pytorch/glm4v.py +258 -0
- xinference/model/llm/pytorch/intern_vl.py +5 -10
- xinference/model/llm/pytorch/minicpmv25.py +232 -0
- xinference/model/llm/pytorch/utils.py +386 -2
- xinference/model/llm/vllm/core.py +7 -1
- xinference/thirdparty/ChatTTS/__init__.py +1 -0
- xinference/thirdparty/ChatTTS/core.py +200 -0
- xinference/types.py +3 -0
- {xinference-0.11.2.post1.dist-info → xinference-0.12.0.dist-info}/METADATA +28 -11
- {xinference-0.11.2.post1.dist-info → xinference-0.12.0.dist-info}/RECORD +36 -29
- {xinference-0.11.2.post1.dist-info → xinference-0.12.0.dist-info}/LICENSE +0 -0
- {xinference-0.11.2.post1.dist-info → xinference-0.12.0.dist-info}/WHEEL +0 -0
- {xinference-0.11.2.post1.dist-info → xinference-0.12.0.dist-info}/entry_points.txt +0 -0
- {xinference-0.11.2.post1.dist-info → xinference-0.12.0.dist-info}/top_level.txt +0 -0
|
@@ -4,6 +4,7 @@
|
|
|
4
4
|
"model_family": "whisper",
|
|
5
5
|
"model_id": "openai/whisper-tiny",
|
|
6
6
|
"model_revision": "167c219b21f11ef214220b8fdb7536b8a88c2475",
|
|
7
|
+
"ability": "audio-to-text",
|
|
7
8
|
"multilingual": true
|
|
8
9
|
},
|
|
9
10
|
{
|
|
@@ -11,6 +12,7 @@
|
|
|
11
12
|
"model_family": "whisper",
|
|
12
13
|
"model_id": "openai/whisper-tiny.en",
|
|
13
14
|
"model_revision": "87c7102498dcde7456f24cfd30239ca606ed9063",
|
|
15
|
+
"ability": "audio-to-text",
|
|
14
16
|
"multilingual": false
|
|
15
17
|
},
|
|
16
18
|
{
|
|
@@ -18,6 +20,7 @@
|
|
|
18
20
|
"model_family": "whisper",
|
|
19
21
|
"model_id": "openai/whisper-base",
|
|
20
22
|
"model_revision": "8c1db9b51951100007a96a525d83a8ec81b3c237",
|
|
23
|
+
"ability": "audio-to-text",
|
|
21
24
|
"multilingual": true
|
|
22
25
|
},
|
|
23
26
|
{
|
|
@@ -25,6 +28,7 @@
|
|
|
25
28
|
"model_family": "whisper",
|
|
26
29
|
"model_id": "openai/whisper-base.en",
|
|
27
30
|
"model_revision": "911407f4214e0e1d82085af863093ec0b66f9cd6",
|
|
31
|
+
"ability": "audio-to-text",
|
|
28
32
|
"multilingual": false
|
|
29
33
|
},
|
|
30
34
|
{
|
|
@@ -32,6 +36,7 @@
|
|
|
32
36
|
"model_family": "whisper",
|
|
33
37
|
"model_id": "openai/whisper-small",
|
|
34
38
|
"model_revision": "998cb1a777c20db53d6033a61b977ed4c3792cac",
|
|
39
|
+
"ability": "audio-to-text",
|
|
35
40
|
"multilingual": true
|
|
36
41
|
},
|
|
37
42
|
{
|
|
@@ -39,6 +44,7 @@
|
|
|
39
44
|
"model_family": "whisper",
|
|
40
45
|
"model_id": "openai/whisper-small.en",
|
|
41
46
|
"model_revision": "e8727524f962ee844a7319d92be39ac1bd25655a",
|
|
47
|
+
"ability": "audio-to-text",
|
|
42
48
|
"multilingual": false
|
|
43
49
|
},
|
|
44
50
|
{
|
|
@@ -46,6 +52,7 @@
|
|
|
46
52
|
"model_family": "whisper",
|
|
47
53
|
"model_id": "openai/whisper-medium",
|
|
48
54
|
"model_revision": "16688beb1294bedd0a6f5cd86fe7eec57bce41ed",
|
|
55
|
+
"ability": "audio-to-text",
|
|
49
56
|
"multilingual": true
|
|
50
57
|
},
|
|
51
58
|
{
|
|
@@ -53,6 +60,7 @@
|
|
|
53
60
|
"model_family": "whisper",
|
|
54
61
|
"model_id": "openai/whisper-medium.en",
|
|
55
62
|
"model_revision": "2e98eb6279edf5095af0c8dedb36bdec0acd172b",
|
|
63
|
+
"ability": "audio-to-text",
|
|
56
64
|
"multilingual": false
|
|
57
65
|
},
|
|
58
66
|
{
|
|
@@ -60,6 +68,7 @@
|
|
|
60
68
|
"model_family": "whisper",
|
|
61
69
|
"model_id": "openai/whisper-large-v3",
|
|
62
70
|
"model_revision": "6cdf07a7e3ec3806e5d55f787915b85d4cd020b1",
|
|
71
|
+
"ability": "audio-to-text",
|
|
63
72
|
"multilingual": true
|
|
64
73
|
},
|
|
65
74
|
{
|
|
@@ -67,6 +76,7 @@
|
|
|
67
76
|
"model_family": "whisper",
|
|
68
77
|
"model_id": "BELLE-2/Belle-distilwhisper-large-v2-zh",
|
|
69
78
|
"model_revision": "ed25d13498fa5bac758b2fc479435b698532dfe8",
|
|
79
|
+
"ability": "audio-to-text",
|
|
70
80
|
"multilingual": false
|
|
71
81
|
},
|
|
72
82
|
{
|
|
@@ -74,6 +84,7 @@
|
|
|
74
84
|
"model_family": "whisper",
|
|
75
85
|
"model_id": "BELLE-2/Belle-whisper-large-v2-zh",
|
|
76
86
|
"model_revision": "ec5bd5d78598545b7585814edde86dac2002b5b9",
|
|
87
|
+
"ability": "audio-to-text",
|
|
77
88
|
"multilingual": false
|
|
78
89
|
},
|
|
79
90
|
{
|
|
@@ -81,6 +92,15 @@
|
|
|
81
92
|
"model_family": "whisper",
|
|
82
93
|
"model_id": "BELLE-2/Belle-whisper-large-v3-zh",
|
|
83
94
|
"model_revision": "3bebc7247696b39f5ab9ed22db426943ac33f600",
|
|
95
|
+
"ability": "audio-to-text",
|
|
84
96
|
"multilingual": false
|
|
97
|
+
},
|
|
98
|
+
{
|
|
99
|
+
"model_name": "ChatTTS",
|
|
100
|
+
"model_family": "ChatTTS",
|
|
101
|
+
"model_id": "2Noise/ChatTTS",
|
|
102
|
+
"model_revision": "ce5913842aebd78e4a01a02d47244b8d62ac4ee3",
|
|
103
|
+
"ability": "text-to-audio",
|
|
104
|
+
"multilingual": true
|
|
85
105
|
}
|
|
86
106
|
]
|
xinference/model/llm/__init__.py
CHANGED
|
@@ -113,12 +113,15 @@ def _install():
|
|
|
113
113
|
from .ggml.llamacpp import LlamaCppChatModel, LlamaCppModel
|
|
114
114
|
from .pytorch.baichuan import BaichuanPytorchChatModel
|
|
115
115
|
from .pytorch.chatglm import ChatglmPytorchChatModel
|
|
116
|
+
from .pytorch.cogvlm2 import CogVLM2Model
|
|
116
117
|
from .pytorch.core import PytorchChatModel, PytorchModel
|
|
117
118
|
from .pytorch.deepseek_vl import DeepSeekVLChatModel
|
|
118
119
|
from .pytorch.falcon import FalconPytorchChatModel, FalconPytorchModel
|
|
120
|
+
from .pytorch.glm4v import Glm4VModel
|
|
119
121
|
from .pytorch.intern_vl import InternVLChatModel
|
|
120
122
|
from .pytorch.internlm2 import Internlm2PytorchChatModel
|
|
121
123
|
from .pytorch.llama_2 import LlamaPytorchChatModel, LlamaPytorchModel
|
|
124
|
+
from .pytorch.minicpmv25 import MiniCPMV25Model
|
|
122
125
|
from .pytorch.qwen_vl import QwenVLChatModel
|
|
123
126
|
from .pytorch.vicuna import VicunaPytorchChatModel
|
|
124
127
|
from .pytorch.yi_vl import YiVLChatModel
|
|
@@ -159,6 +162,9 @@ def _install():
|
|
|
159
162
|
DeepSeekVLChatModel,
|
|
160
163
|
InternVLChatModel,
|
|
161
164
|
PytorchModel,
|
|
165
|
+
CogVLM2Model,
|
|
166
|
+
MiniCPMV25Model,
|
|
167
|
+
Glm4VModel,
|
|
162
168
|
]
|
|
163
169
|
)
|
|
164
170
|
if OmniLMMModel: # type: ignore
|