xinference 1.1.1__py3-none-any.whl → 1.2.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of xinference might be problematic. Click here for more details.
- xinference/_version.py +3 -3
- xinference/api/restful_api.py +49 -65
- xinference/core/model.py +77 -19
- xinference/core/supervisor.py +81 -10
- xinference/core/utils.py +2 -2
- xinference/core/worker.py +32 -0
- xinference/model/image/model_spec.json +18 -0
- xinference/model/image/model_spec_modelscope.json +20 -0
- xinference/model/llm/__init__.py +2 -0
- xinference/model/llm/llm_family.json +96 -0
- xinference/model/llm/llm_family_modelscope.json +99 -0
- xinference/model/llm/mlx/core.py +23 -73
- xinference/model/llm/transformers/cogagent.py +272 -0
- xinference/model/llm/transformers/core.py +1 -0
- xinference/model/llm/transformers/qwen2_vl.py +10 -1
- xinference/model/llm/utils.py +27 -3
- xinference/model/llm/vllm/core.py +37 -7
- xinference/model/llm/vllm/xavier/__init__.py +13 -0
- xinference/model/llm/vllm/xavier/allocator.py +74 -0
- xinference/model/llm/vllm/xavier/block.py +112 -0
- xinference/model/llm/vllm/xavier/block_manager.py +71 -0
- xinference/model/llm/vllm/xavier/block_tracker.py +116 -0
- xinference/model/llm/vllm/xavier/engine.py +247 -0
- xinference/model/llm/vllm/xavier/executor.py +132 -0
- xinference/model/llm/vllm/xavier/scheduler.py +422 -0
- xinference/model/llm/vllm/xavier/test/__init__.py +13 -0
- xinference/model/llm/vllm/xavier/test/test_xavier.py +122 -0
- xinference/model/llm/vllm/xavier/transfer.py +298 -0
- xinference/model/video/diffusers.py +14 -0
- xinference/model/video/model_spec.json +15 -0
- xinference/model/video/model_spec_modelscope.json +16 -0
- xinference/types.py +13 -0
- xinference/web/ui/build/asset-manifest.json +6 -6
- xinference/web/ui/build/index.html +1 -1
- xinference/web/ui/build/static/css/main.51a587ff.css +2 -0
- xinference/web/ui/build/static/css/main.51a587ff.css.map +1 -0
- xinference/web/ui/build/static/js/main.1eb206d1.js +3 -0
- xinference/web/ui/build/static/js/main.1eb206d1.js.map +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/03c4052f1b91f6ba0c5389bdcf49c43319b4076c08e4b8585dab312538ae290a.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/1786b83003b8e9605a0f5f855a185d4d16e38fc893dfb326a2a9cca206b4240a.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/17cbc181dd674b9150b80c73ed6a82656de0082d857f6e5f66d9716129ac0b38.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/185ceb8872d562e032b47e79df6a45670e06345b8ed70aad1a131e0476783c5c.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/2213d49de260e1f67c888081b18f120f5225462b829ae57c9e05a05cec83689d.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/26b8c9f34b0bed789b3a833767672e39302d1e0c09b4276f4d58d1df7b6bd93b.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/2b484da66c724d0d56a40849c109327408796a668b1381511b6e9e03baa48658.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/2cbbbce9b84df73330d4c42b82436ed881b3847628f2fbc346aa62e2859fd88c.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/2ec9b14431ed33ce6901bf9f27007be4e6e472709c99d6e22b50ce528e4b78ee.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/3b966db018f96be4a055d6ca205f0990d4d0b370e2980c17d8bca2c9a021819c.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/3eefb411b24c2b3ce053570ef50daccf154022f0e168be5ed0fec21394baf9f4.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/522b229e3cac219123f0d69673f5570e191c2d2a505dc65b312d336eae2279c0.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/52e45f17ba300580ea3fcc9f9228ccba194bb092b76f25e9255af311f8b05aab.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/5a0bc4631f936459afc1a3b1d3ec2420118b1f00e11f60ccac3e08088f3f27a8.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/611fa2c6c53b66039991d06dfb0473b5ab37fc63b4564e0f6e1718523768a045.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/6329bc76c406fe5eb305412383fbde5950f847bb5e43261f73f37622c365acb4.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/63c8e07687ea53a4f8a910ee5e42e0eb26cd1acbfbe820f3e3248a786ee51401.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/69b2d5001684174ec9da57e07914eed3eac4960018bceb6cbfa801d861301d7c.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/710c1acda69e561e30a933b98c6a56d50197868b15c21e2aad55ab6d46649eb6.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/720deca1fce5a1dc5056048fa8258fd138a82ea855f350b6613f104a73fb761f.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/76a23b92d26a499c57e61eea2b895fbc9771bd0849a72e66f8e633192017978b.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/858063f23b34dfe600254eb5afd85518b0002ec4b30b7386616c45600826e3b2.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/920b82c1c89124cf217109eeedbfcd3aae3b917be50c9dfb6bbb4ce26bdfd2e7.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/94d8b7aeb0076f2ce07db598cea0e87b13bc8d5614eb530b8d6e696c2daf6f88.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/9e917fe7022d01b2ccbe5cc0ce73d70bb72bee584ff293bad71bdff6695dee28.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/9f28fdb8399f1d0474f0aca86f1658dc94f5bf0c90f6146352de150692de8862.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/a0dfafa06b2bb7cba8cad41c482503f61944f759f4318139362602ef5cc47ccb.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/afb8084f539534cd594755ea2205ecd5bd1f62dddcfdf75a2eace59a28131278.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/b57b1438b77294c1f3f6cfce12ac487d8106c6f016975ba0aec94d98997e2e1e.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/b9917b0bf8e4d55ccbac1c334aa04d6ff3c5b6ed9e5d38b9ea2c687fa7d3f5a9.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/bbcc94b0149963d1d6f267ee1f4f03d3925b758392ce2f516c3fe8af0e0169fc.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/bdee44abeadc4abc17d41c52eb49c6e19a4b1a267b6e16876ce91bdeeebfc52d.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/beb112b70f4a56db95920a9e20efb6c97c37b68450716730217a9ee1a9ae92be.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/c88db97be0cdf440193b3995996e83510a04cb00048135485fc0e26d197e80b5.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/d49e5314d34310a62d01a03067ce1bec5da00abce84c5196aa9c6842fa79a430.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/d7664d18c4ddbad9c3a6a31b91f7c00fb0dde804608674a9860ee50f33e54708.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/d9072c318b819b7c90a0f7e9cc0b6413b4dbeb8e9859898e53d75ea882fcde99.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/db16a983bc08a05f0439cc61ca0840e49e1d8400eef678909f16c032a418a3d6.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/dc249829767b8abcbc3677e0b07b6d3ecbfdfe6d08cfe23a665eb33373a9aa9d.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/e242c583c2dbc2784f0fcf513523975f7d5df447e106c1c17e49e8578a6fc3ed.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/eac5f1296513e69e4b96f750ddccd4d0264e2bae4e4c449144e83274a48698d9.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/ed57202cb79649bb716400436590245547df241988fc7c8e1d85d132299542d2.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/f125bf72e773a14cdaebd0c343e80adb909d12e317ee5c00cd4a57442fbe2c62.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/f91af913d7f91c410719ab13136aaed3aaf0f8dda06652f25c42cb5231587398.json +1 -0
- xinference/web/ui/node_modules/.package-lock.json +67 -3
- xinference/web/ui/node_modules/@babel/runtime/package.json +592 -538
- xinference/web/ui/node_modules/html-parse-stringify/package.json +50 -0
- xinference/web/ui/node_modules/i18next/dist/esm/package.json +1 -0
- xinference/web/ui/node_modules/i18next/package.json +129 -0
- xinference/web/ui/node_modules/react-i18next/.eslintrc.json +74 -0
- xinference/web/ui/node_modules/react-i18next/dist/es/package.json +1 -0
- xinference/web/ui/node_modules/react-i18next/package.json +162 -0
- xinference/web/ui/node_modules/void-elements/package.json +34 -0
- xinference/web/ui/package-lock.json +69 -3
- xinference/web/ui/package.json +2 -0
- xinference/web/ui/src/locales/en.json +186 -0
- xinference/web/ui/src/locales/zh.json +186 -0
- {xinference-1.1.1.dist-info → xinference-1.2.0.dist-info}/METADATA +9 -6
- {xinference-1.1.1.dist-info → xinference-1.2.0.dist-info}/RECORD +102 -56
- xinference/web/ui/build/static/css/main.5061c4c3.css +0 -2
- xinference/web/ui/build/static/css/main.5061c4c3.css.map +0 -1
- xinference/web/ui/build/static/js/main.4eb4ee80.js +0 -3
- xinference/web/ui/build/static/js/main.4eb4ee80.js.map +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/07ce9e632e6aff24d7aa3ad8e48224433bbfeb0d633fca723453f1fcae0c9f1c.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/1130403f9e46f5738a23b45ac59b57de8f360c908c713e2c0670c2cce9bd367a.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/131091b25d26b17cdca187d7542a21475c211138d900cf667682260e76ef9463.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/1f269fb2a368363c1cb2237825f1dba093b6bdd8c44cc05954fd19ec2c1fff03.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/331312668fa8bd3d7401818f4a25fa98135d7f61371cd6bfff78b18cf4fbdd92.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/40f17338fc75ae095de7d2b4d8eae0d5ca0193a7e2bcece4ee745b22a7a2f4b7.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/4de9a6942c5f1749d6cbfdd54279699975f16016b182848bc253886f52ec2ec3.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/822586ed1077201b64b954f12f25e3f9b45678c1acbabe53d8af3ca82ca71f33.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/8c5eeb02f772d02cbe8b89c05428d0dd41a97866f75f7dc1c2164a67f5a1cf98.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/8d33354bd2100c8602afc3341f131a88cc36aaeecd5a4b365ed038514708e350.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/9375a35b05d56989b2755bf72161fa707c92f28569d33765a75f91a568fda6e9.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/a158a9ffa0c9b169aee53dd4a0c44501a596755b4e4f6ede7746d65a72e2a71f.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/c7bf40bab396765f67d0fed627ed3665890608b2d0edaa3e8cb7cfc96310db45.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/d6c643278a0b28320e6f33a60f5fb64c053997cbdc39a60e53ccc574688ade9e.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/e42b72d4cc1ea412ebecbb8d040dc6c6bfee462c33903c2f1f3facb602ad742e.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/e64b7e8cedcf43d4c95deba60ec1341855c887705805bb62431693118b870c69.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/f5039ddbeb815c51491a1989532006b96fc3ae49c6c60e3c097f875b4ae915ae.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/f72f011744c4649fabddca6f7a9327861ac0a315a89b1a2e62a39774e7863845.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/feabb04b4aa507102da0a64398a40818e878fd1df9b75dda8461b3e1e7ff3f11.json +0 -1
- /xinference/web/ui/build/static/js/{main.4eb4ee80.js.LICENSE.txt → main.1eb206d1.js.LICENSE.txt} +0 -0
- {xinference-1.1.1.dist-info → xinference-1.2.0.dist-info}/LICENSE +0 -0
- {xinference-1.1.1.dist-info → xinference-1.2.0.dist-info}/WHEEL +0 -0
- {xinference-1.1.1.dist-info → xinference-1.2.0.dist-info}/entry_points.txt +0 -0
- {xinference-1.1.1.dist-info → xinference-1.2.0.dist-info}/top_level.txt +0 -0
|
@@ -0,0 +1,186 @@
|
|
|
1
|
+
{
|
|
2
|
+
"menu": {
|
|
3
|
+
"launchModel": "Launch Model",
|
|
4
|
+
"runningModels": "Running Models",
|
|
5
|
+
"registerModel": "Register Model",
|
|
6
|
+
"clusterInfo": "Cluster Information",
|
|
7
|
+
"contactUs": "Contact Us"
|
|
8
|
+
},
|
|
9
|
+
|
|
10
|
+
"model": {
|
|
11
|
+
"languageModels": "Language Models",
|
|
12
|
+
"embeddingModels": "Embedding Models",
|
|
13
|
+
"rerankModels": "Rerank Models",
|
|
14
|
+
"imageModels": "Image Models",
|
|
15
|
+
"audioModels": "Audio Models",
|
|
16
|
+
"videoModels": "Video Models",
|
|
17
|
+
"customModels": "Custom Models",
|
|
18
|
+
"flexibleModels": "Flexible Models"
|
|
19
|
+
},
|
|
20
|
+
|
|
21
|
+
"launchModel": {
|
|
22
|
+
"modelAbility": "Model Ability",
|
|
23
|
+
"generate": "generate",
|
|
24
|
+
"chat": "chat",
|
|
25
|
+
"vision": "vl-chat",
|
|
26
|
+
"status": "Status",
|
|
27
|
+
"cached": "Cached",
|
|
28
|
+
"manageCachedModels": "Manage Cached Models",
|
|
29
|
+
"favorite": "Favorite",
|
|
30
|
+
"unfavorite": "Unfavorite",
|
|
31
|
+
"search": "Search for model name and description",
|
|
32
|
+
"searchModelType": "Search for {{modelType}} model name",
|
|
33
|
+
"searchInstruction": "Type {{hotkey}} to search",
|
|
34
|
+
"clickToLaunchModel": "Click with mouse to launch the model",
|
|
35
|
+
"dimensions": "dimensions",
|
|
36
|
+
"maxTokens": "max tokens",
|
|
37
|
+
"edit": "Edit",
|
|
38
|
+
"delete": "Delete",
|
|
39
|
+
"contextLength": "context length",
|
|
40
|
+
"chatModel": "chat model",
|
|
41
|
+
"generateModel": "generate model",
|
|
42
|
+
"otherModel": "other model",
|
|
43
|
+
"confirmDeleteCustomModel": "Are you sure to delete this custom model? This behavior is irreversible.",
|
|
44
|
+
"lastConfig": "Last Config",
|
|
45
|
+
"modelEngine": "Model Engine",
|
|
46
|
+
"modelFormat": "Model Format",
|
|
47
|
+
"modelSize": "Model Size",
|
|
48
|
+
"quantization": "Quantization",
|
|
49
|
+
"nGPU": "GPU Count",
|
|
50
|
+
"nGpuLayers": "N GPU Layers",
|
|
51
|
+
"replica": "Replica",
|
|
52
|
+
"optionalConfigurations": "Optional Configurations",
|
|
53
|
+
"modelUID.optional": "(Optional) Model UID, model name by default",
|
|
54
|
+
"requestLimits.optional": "(Optional) Request Limits, the number of request limits for this model, default is None",
|
|
55
|
+
"workerIp.optional": "(Optional) Worker Ip, specify the worker ip where the model is located in a distributed scenario",
|
|
56
|
+
"workerIp": "Worker Ip, specify the worker ip where the model is located in a distributed scenario",
|
|
57
|
+
"GPUIdx.optional": "(Optional) GPU Idx, Specify the GPU index where the model is located",
|
|
58
|
+
"GPUIdx": "GPU Idx, Specify the GPU index where the model is located",
|
|
59
|
+
"downloadHub.optional": "(Optional) Download_hub",
|
|
60
|
+
"modelPath.optional": "(Optional) Model Path, For PyTorch, provide the model directory. For GGML/GGUF, provide the model file path.",
|
|
61
|
+
"GGUFQuantization.optional": "(Optional) GGUF quantization format, quantizing the Transformer part.",
|
|
62
|
+
"GGUFModelPath.optional": "(Optional) GGUF model path, should be a file ending with .gguf.",
|
|
63
|
+
"CPUOffload": "CPU Offload",
|
|
64
|
+
"CPUOffload.tip": "Unload the model to the CPU. Recommend to enable this when resources are limited or when using the GGUF option.",
|
|
65
|
+
"loraConfig": "Lora Config",
|
|
66
|
+
"loraModelConfig": "Lora Model Config",
|
|
67
|
+
"additionalParametersForInferenceEngine": "Additional parameters passed to the inference engine",
|
|
68
|
+
"enterIntegerGreaterThanZero": "Please enter an integer greater than 0.",
|
|
69
|
+
"enterCommaSeparatedNumbers": "Please enter numeric data separated by commas, for example: 0,1,2",
|
|
70
|
+
"device": "Device",
|
|
71
|
+
"loraLoadKwargsForImageModel": "Lora Load Kwargs for Image Model",
|
|
72
|
+
"loraFuseKwargsForImageModel": "Lora Fuse Kwargs for Image Model",
|
|
73
|
+
"launch": "Launch",
|
|
74
|
+
"goBack": "Go Back",
|
|
75
|
+
"copyJson": "Copy Json",
|
|
76
|
+
"cancel": "Cancel",
|
|
77
|
+
"fillCompleteParametersBeforeAdding": "Please fill in the complete parameters before adding!",
|
|
78
|
+
"model_format": "model_format",
|
|
79
|
+
"model_size_in_billions": "model_size_in_billions",
|
|
80
|
+
"quantizations": "quantizations",
|
|
81
|
+
"real_path": "real_path",
|
|
82
|
+
"path": "path",
|
|
83
|
+
"ipAddress": "IP Address",
|
|
84
|
+
"operation": "operation",
|
|
85
|
+
"copyRealPath": "Copy real_path",
|
|
86
|
+
"copyPath": "Copy path",
|
|
87
|
+
"noCacheForNow": "No cache for now!",
|
|
88
|
+
"confirmDeleteCacheFiles": "Confirm deletion of cache files? This action is irreversible."
|
|
89
|
+
},
|
|
90
|
+
|
|
91
|
+
"runningModels": {
|
|
92
|
+
"name": "Name",
|
|
93
|
+
"address": "Address",
|
|
94
|
+
"gpuIndexes": "GPU Indexes",
|
|
95
|
+
"size": "Size",
|
|
96
|
+
"quantization": "Quantization",
|
|
97
|
+
"replica": "Replica",
|
|
98
|
+
"actions": "Actions",
|
|
99
|
+
"noRunningModels": "No Running Models",
|
|
100
|
+
"noRunningModelsMatches": "No Running Models Matches"
|
|
101
|
+
},
|
|
102
|
+
|
|
103
|
+
"registerModel": {
|
|
104
|
+
"modelName": "Model Name",
|
|
105
|
+
"modelDescription": "Model Description (Optional)",
|
|
106
|
+
"contextLength": "Context Length",
|
|
107
|
+
"dimensions": "Dimensions",
|
|
108
|
+
"maxTokens": "Max Tokens",
|
|
109
|
+
"modelPath": "Model Path",
|
|
110
|
+
"modelLanguages": "Model Languages",
|
|
111
|
+
"languages": "Languages",
|
|
112
|
+
"multilingual": "Multilingual",
|
|
113
|
+
"modelAbilities": "Model Abilities",
|
|
114
|
+
"modelFamily": "Model Family",
|
|
115
|
+
"chatTemplate": "Chat Template",
|
|
116
|
+
"test": "test",
|
|
117
|
+
"testResult": "test result",
|
|
118
|
+
"noTestResults": "No test results...",
|
|
119
|
+
"stopTokenIds": "Stop Token Ids",
|
|
120
|
+
"stop": "Stop",
|
|
121
|
+
"launcher": "Launcher",
|
|
122
|
+
"launcherArguments": "Launcher Arguments (Optional)",
|
|
123
|
+
"edit": "Edit",
|
|
124
|
+
"cancel": "Cancel",
|
|
125
|
+
"registerModel": "Register Model",
|
|
126
|
+
"messagesExample": "Messages Example",
|
|
127
|
+
"JSONFormat": "JSON Format",
|
|
128
|
+
"modelSpecs": "Model Specs",
|
|
129
|
+
"modelSizeBillions": "Model Size in Billions",
|
|
130
|
+
"quantization": "Quantization",
|
|
131
|
+
"quantizationOptional": "Quantization (Optional)",
|
|
132
|
+
"delete": "Delete",
|
|
133
|
+
"controlnet": "Controlnet",
|
|
134
|
+
"more": "more",
|
|
135
|
+
"modelFormat": "Model Format",
|
|
136
|
+
"enterNumberGreaterThanZero": "Please enter a number greater than 0.",
|
|
137
|
+
"carefulQuantizationForModelRegistration": "For GPTQ/AWQ/FP8/MLX models, please be careful to fill in the quantization corresponding to the model you want to register.",
|
|
138
|
+
"quantizationCannotBeEmpty": "Quantization cannot be left empty.",
|
|
139
|
+
"enterInteger": "Please enter an integer.",
|
|
140
|
+
"enterIntegerGreaterThanZero": "Please enter an integer greater than 0.",
|
|
141
|
+
"showCustomJsonConfig": "Show custom json config used by api",
|
|
142
|
+
"packUp": "Pack up",
|
|
143
|
+
"unfold": "Unfold",
|
|
144
|
+
"copyAll": "Copy all",
|
|
145
|
+
"alphanumericWithHyphensUnderscores": "Alphanumeric characters with properly placed hyphens and underscores. Must not match any built-in model names.",
|
|
146
|
+
"chooseBuiltInOrCustomModel": "You can choose from the built-in models or input your own.",
|
|
147
|
+
"chooseOnlyBuiltInModel": "You can only choose from the built-in models.",
|
|
148
|
+
"provideModelDirectoryPath": "Provide the model directory path.",
|
|
149
|
+
"provideModelLauncher": "Provide the model launcher.",
|
|
150
|
+
"jsonArgumentsForLauncher": "A JSON-formatted dictionary representing the arguments passed to the Launcher.",
|
|
151
|
+
"provideModelDirectoryOrFilePath": "For PyTorch, provide the model directory. For GGUF, provide the model file path.",
|
|
152
|
+
"ensureChatTemplatePassesTest": "Please make sure this chat_template passes the test by clicking the TEST button on the right. Please note that this test may not cover all cases and will only be used for the most basic case.",
|
|
153
|
+
"testFailurePreventsChatWorking": "Please note that failure to pass test may prevent chats from working properly.",
|
|
154
|
+
"stopControlForChatModels": "int type, used to control the stopping of chat models",
|
|
155
|
+
"stopControlStringForChatModels": "string type, used to control the stopping of chat models",
|
|
156
|
+
"enterJsonFormattedDictionary": "Please enter the JSON-formatted dictionary."
|
|
157
|
+
},
|
|
158
|
+
|
|
159
|
+
"clusterInfo": {
|
|
160
|
+
"supervisor": "Supervisor",
|
|
161
|
+
"workers": "Workers",
|
|
162
|
+
"workerDetails": "Worker Details",
|
|
163
|
+
"count": "Count",
|
|
164
|
+
"cpuInfo": "CPU Info",
|
|
165
|
+
"usage": "Usage:",
|
|
166
|
+
"total": "Total",
|
|
167
|
+
"cpuMemoryInfo": "CPU Memory Info",
|
|
168
|
+
"version": "Version",
|
|
169
|
+
"release": "Release:",
|
|
170
|
+
"commit": "Commit:",
|
|
171
|
+
"gpuInfo": "GPU Info",
|
|
172
|
+
"gpuMemoryInfo": "GPU Memory Info",
|
|
173
|
+
"address": "Address",
|
|
174
|
+
"item": "Item",
|
|
175
|
+
"value": "Value",
|
|
176
|
+
"nodeType": "Node Type",
|
|
177
|
+
"cpuUsage": "CPU Usage",
|
|
178
|
+
"cpuTotal": "CPU Total",
|
|
179
|
+
"memUsage": "Mem Usage",
|
|
180
|
+
"memTotal": "Mem Total",
|
|
181
|
+
"gpuCount": "GPU Count",
|
|
182
|
+
"gpuMemUsage": "GPU Mem Usage",
|
|
183
|
+
"gpuMemTotal": "GPU Mem Total",
|
|
184
|
+
"worker": "Worker"
|
|
185
|
+
}
|
|
186
|
+
}
|
|
@@ -0,0 +1,186 @@
|
|
|
1
|
+
{
|
|
2
|
+
"menu": {
|
|
3
|
+
"launchModel": "启动模型",
|
|
4
|
+
"runningModels": "运行模型",
|
|
5
|
+
"registerModel": "注册模型",
|
|
6
|
+
"clusterInfo": "集群信息",
|
|
7
|
+
"contactUs": "联系我们"
|
|
8
|
+
},
|
|
9
|
+
|
|
10
|
+
"model": {
|
|
11
|
+
"languageModels": "语言模型",
|
|
12
|
+
"embeddingModels": "嵌入模型",
|
|
13
|
+
"rerankModels": "重排序模型",
|
|
14
|
+
"imageModels": "图像模型",
|
|
15
|
+
"audioModels": "音频模型",
|
|
16
|
+
"videoModels": "视频模型",
|
|
17
|
+
"customModels": "自定义模型",
|
|
18
|
+
"flexibleModels": "灵活模型"
|
|
19
|
+
},
|
|
20
|
+
|
|
21
|
+
"launchModel": {
|
|
22
|
+
"modelAbility": "模型能力",
|
|
23
|
+
"generate": "生成",
|
|
24
|
+
"chat": "聊天",
|
|
25
|
+
"vision": "视觉聊天",
|
|
26
|
+
"status": "状态",
|
|
27
|
+
"cached": "已缓存",
|
|
28
|
+
"manageCachedModels": "管理缓存模型",
|
|
29
|
+
"favorite": "收藏",
|
|
30
|
+
"unfavorite": "取消收藏",
|
|
31
|
+
"search": "搜索模型名称和描述",
|
|
32
|
+
"searchModelType": "搜索 {{modelType}} 相关的模型名称",
|
|
33
|
+
"searchInstruction": "输入 {{hotkey}} 进行搜索",
|
|
34
|
+
"clickToLaunchModel": "点击鼠标以启动模型",
|
|
35
|
+
"dimensions": "维度",
|
|
36
|
+
"maxTokens": "最大 token 数",
|
|
37
|
+
"edit": "编辑",
|
|
38
|
+
"delete": "删除",
|
|
39
|
+
"contextLength": "上下文长度",
|
|
40
|
+
"chatModel": "聊天模型",
|
|
41
|
+
"generateModel": "生成模型",
|
|
42
|
+
"otherModel": "其他模型",
|
|
43
|
+
"confirmDeleteCustomModel": "您确定要删除这个自定义模型吗?此操作无法恢复。",
|
|
44
|
+
"lastConfig": "最后配置",
|
|
45
|
+
"modelEngine": "模型引擎",
|
|
46
|
+
"modelFormat": "模型格式",
|
|
47
|
+
"modelSize": "模型大小",
|
|
48
|
+
"quantization": "量化",
|
|
49
|
+
"nGPU": "GPU 数量",
|
|
50
|
+
"nGpuLayers": "GPU 层数",
|
|
51
|
+
"replica": "副本",
|
|
52
|
+
"optionalConfigurations": "可选配置",
|
|
53
|
+
"modelUID.optional": "(可选) 模型 UID,默认是模型名称",
|
|
54
|
+
"requestLimits.optional": "(可选) 请求限制,模型的请求限制数,默认值为无",
|
|
55
|
+
"workerIp.optional": "(可选) 工作节点 IP,在分布式场景中指定模型所在的工作节点 IP",
|
|
56
|
+
"workerIp": "工作节点 IP,在分布式场景中指定模型所在的工作节点 IP",
|
|
57
|
+
"GPUIdx.optional": "(可选) GPU 索引,指定模型所在的 GPU 索引",
|
|
58
|
+
"GPUIdx": "GPU 索引,指定模型所在的 GPU 索引",
|
|
59
|
+
"downloadHub.optional": "(可选) 下载中心",
|
|
60
|
+
"modelPath.optional": "(可选) 模型路径,对于 PyTorch,提供模型目录;对于 GGML/GGUF,提供模型文件路径。",
|
|
61
|
+
"GGUFQuantization.optional": "(可选) GGUF量化格式,对Transformer部分进行量化。",
|
|
62
|
+
"GGUFModelPath.optional": "(可选) GGUF模型路径,应为以 .gguf 结尾的文件。",
|
|
63
|
+
"CPUOffload": "CPU卸载",
|
|
64
|
+
"CPUOffload.tip": "将模型卸载到CPU。当资源有限或使用GGUF选项时,建议启用此功能。",
|
|
65
|
+
"loraConfig": "Lora 配置",
|
|
66
|
+
"loraModelConfig": "Lora 模型配置",
|
|
67
|
+
"additionalParametersForInferenceEngine": "传递给推理引擎的附加参数",
|
|
68
|
+
"enterIntegerGreaterThanZero": "请输入大于 0 的整数。",
|
|
69
|
+
"enterCommaSeparatedNumbers": "请输入以逗号分隔的数字数据,例如:0,1,2",
|
|
70
|
+
"device": "设备",
|
|
71
|
+
"loraLoadKwargsForImageModel": "图像模型的 Lora 加载参数",
|
|
72
|
+
"loraFuseKwargsForImageModel": "图像模型的 Lora 融合参数",
|
|
73
|
+
"launch": "启动",
|
|
74
|
+
"goBack": "返回",
|
|
75
|
+
"copyJson": "复制 JSON",
|
|
76
|
+
"cancel": "取消",
|
|
77
|
+
"fillCompleteParametersBeforeAdding": "请在添加之前填写完整的参数!",
|
|
78
|
+
"model_format": "模型格式",
|
|
79
|
+
"model_size_in_billions": "模型大小(以十亿为单位)",
|
|
80
|
+
"quantizations": "量化方式",
|
|
81
|
+
"real_path": "真实路径",
|
|
82
|
+
"path": "路径",
|
|
83
|
+
"ipAddress": "IP 地址",
|
|
84
|
+
"operation": "操作",
|
|
85
|
+
"copyRealPath": "复制真实路径",
|
|
86
|
+
"copyPath": "复制路径",
|
|
87
|
+
"noCacheForNow": "当前没有缓存!",
|
|
88
|
+
"confirmDeleteCacheFiles": "确认删除缓存文件吗?此操作无法恢复。"
|
|
89
|
+
},
|
|
90
|
+
|
|
91
|
+
"runningModels": {
|
|
92
|
+
"name": "名称",
|
|
93
|
+
"address": "地址",
|
|
94
|
+
"gpuIndexes": "GPU 索引",
|
|
95
|
+
"size": "大小",
|
|
96
|
+
"quantization": "量化",
|
|
97
|
+
"replica": "副本",
|
|
98
|
+
"actions": "操作",
|
|
99
|
+
"noRunningModels": "没有运行中的模型",
|
|
100
|
+
"noRunningModelsMatches": "没有匹配的运行模型"
|
|
101
|
+
},
|
|
102
|
+
|
|
103
|
+
"registerModel": {
|
|
104
|
+
"modelName": "模型名称",
|
|
105
|
+
"modelDescription": "模型描述(可选)",
|
|
106
|
+
"contextLength": "上下文长度",
|
|
107
|
+
"dimensions": "维度",
|
|
108
|
+
"maxTokens": "最大 token 数",
|
|
109
|
+
"modelPath": "模型路径",
|
|
110
|
+
"modelLanguages": "模型语言",
|
|
111
|
+
"languages": "语言",
|
|
112
|
+
"multilingual": "多语言",
|
|
113
|
+
"modelAbilities": "模型能力",
|
|
114
|
+
"modelFamily": "模型系列",
|
|
115
|
+
"chatTemplate": "聊天模板",
|
|
116
|
+
"test": "测试",
|
|
117
|
+
"testResult": "测试结果",
|
|
118
|
+
"noTestResults": "没有测试结果...",
|
|
119
|
+
"stopTokenIds": "停止token ID",
|
|
120
|
+
"stop": "停止",
|
|
121
|
+
"launcher": "启动器",
|
|
122
|
+
"launcherArguments": "启动器参数(可选)",
|
|
123
|
+
"edit": "编辑",
|
|
124
|
+
"cancel": "取消",
|
|
125
|
+
"registerModel": "注册模型",
|
|
126
|
+
"messagesExample": "消息示例",
|
|
127
|
+
"JSONFormat": "JSON 格式",
|
|
128
|
+
"modelSpecs": "模型规格",
|
|
129
|
+
"modelSizeBillions": "模型大小(以十亿为单位)",
|
|
130
|
+
"quantization": "量化",
|
|
131
|
+
"quantizationOptional": "量化(可选)",
|
|
132
|
+
"delete": "删除",
|
|
133
|
+
"controlnet": "控制网",
|
|
134
|
+
"more": "更多",
|
|
135
|
+
"modelFormat": "模型格式",
|
|
136
|
+
"enterNumberGreaterThanZero": "请输入大于 0 的数字。",
|
|
137
|
+
"carefulQuantizationForModelRegistration": "对于 GPTQ/AWQ/FP8/MLX 模型,请小心填写与您要注册的模型对应的量化方式。",
|
|
138
|
+
"quantizationCannotBeEmpty": "量化方式不能为空。",
|
|
139
|
+
"enterInteger": "请输入一个整数。",
|
|
140
|
+
"enterIntegerGreaterThanZero": "请输入大于 0 的整数。",
|
|
141
|
+
"showCustomJsonConfig": "显示由 API 使用的自定义 JSON 配置",
|
|
142
|
+
"packUp": "收起",
|
|
143
|
+
"unfold": "展开",
|
|
144
|
+
"copyAll": "复制全部",
|
|
145
|
+
"alphanumericWithHyphensUnderscores": "字母数字字符,连字符和下划线应正确放置。不能与任何内置模型名称匹配。",
|
|
146
|
+
"chooseBuiltInOrCustomModel": "您可以选择内置模型或输入自定义模型。",
|
|
147
|
+
"chooseOnlyBuiltInModel": "您只能从内置模型中选择。",
|
|
148
|
+
"provideModelDirectoryPath": "提供模型目录路径。",
|
|
149
|
+
"provideModelLauncher": "提供模型启动器。",
|
|
150
|
+
"jsonArgumentsForLauncher": "一个 JSON 格式的字典,表示传递给启动器的参数。",
|
|
151
|
+
"provideModelDirectoryOrFilePath": "对于 PyTorch,提供模型目录。对于 GGUF,提供模型文件路径。",
|
|
152
|
+
"ensureChatTemplatePassesTest": "请确保通过点击右侧的测试按钮,使此聊天模板通过测试。请注意,此测试可能无法涵盖所有情况,只会用于最基本的情况。",
|
|
153
|
+
"testFailurePreventsChatWorking": "请注意,未通过测试可能会导致聊天无法正常工作。",
|
|
154
|
+
"stopControlForChatModels": "整数类型,用于控制聊天模型的停止。",
|
|
155
|
+
"stopControlStringForChatModels": "字符串类型,用于控制聊天模型的停止。",
|
|
156
|
+
"enterJsonFormattedDictionary": "请输入 JSON 格式的字典。"
|
|
157
|
+
},
|
|
158
|
+
|
|
159
|
+
"clusterInfo": {
|
|
160
|
+
"supervisor": "主管",
|
|
161
|
+
"workers": "工作节点",
|
|
162
|
+
"workerDetails": "工作节点详情",
|
|
163
|
+
"count": "数量",
|
|
164
|
+
"cpuInfo": "CPU 信息",
|
|
165
|
+
"usage": "使用率:",
|
|
166
|
+
"total": "总计",
|
|
167
|
+
"cpuMemoryInfo": "CPU 内存信息",
|
|
168
|
+
"version": "版本",
|
|
169
|
+
"release": "发布:",
|
|
170
|
+
"commit": "提交:",
|
|
171
|
+
"gpuInfo": "GPU 信息",
|
|
172
|
+
"gpuMemoryInfo": "GPU 内存信息",
|
|
173
|
+
"address": "地址",
|
|
174
|
+
"item": "项",
|
|
175
|
+
"value": "值",
|
|
176
|
+
"nodeType": "节点类型",
|
|
177
|
+
"cpuUsage": "CPU 使用率",
|
|
178
|
+
"cpuTotal": "CPU 总数",
|
|
179
|
+
"memUsage": "内存使用率",
|
|
180
|
+
"memTotal": "内存总量",
|
|
181
|
+
"gpuCount": "GPU 数量",
|
|
182
|
+
"gpuMemUsage": "GPU 内存使用率",
|
|
183
|
+
"gpuMemTotal": "GPU 内存总量",
|
|
184
|
+
"worker": "工作节点"
|
|
185
|
+
}
|
|
186
|
+
}
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.1
|
|
2
2
|
Name: xinference
|
|
3
|
-
Version: 1.
|
|
3
|
+
Version: 1.2.0
|
|
4
4
|
Summary: Model Serving Made Easy
|
|
5
5
|
Home-page: https://github.com/xorbitsai/inference
|
|
6
6
|
Author: Qin Xuye
|
|
@@ -104,8 +104,9 @@ Requires-Dist: jsonschema; extra == "all"
|
|
|
104
104
|
Requires-Dist: verovio>=4.3.1; extra == "all"
|
|
105
105
|
Requires-Dist: auto-gptq; sys_platform != "darwin" and extra == "all"
|
|
106
106
|
Requires-Dist: autoawq<0.2.6; sys_platform != "darwin" and extra == "all"
|
|
107
|
+
Requires-Dist: mlx<0.22.0; (sys_platform == "darwin" and platform_machine == "arm64") and extra == "all"
|
|
107
108
|
Requires-Dist: mlx-lm; (sys_platform == "darwin" and platform_machine == "arm64") and extra == "all"
|
|
108
|
-
Requires-Dist: mlx-vlm; (sys_platform == "darwin" and platform_machine == "arm64") and extra == "all"
|
|
109
|
+
Requires-Dist: mlx-vlm>=0.1.7; (sys_platform == "darwin" and platform_machine == "arm64") and extra == "all"
|
|
109
110
|
Requires-Dist: mlx-whisper; (sys_platform == "darwin" and platform_machine == "arm64") and extra == "all"
|
|
110
111
|
Requires-Dist: f5-tts-mlx; (sys_platform == "darwin" and platform_machine == "arm64") and extra == "all"
|
|
111
112
|
Requires-Dist: vllm>=0.2.6; sys_platform == "linux" and extra == "all"
|
|
@@ -197,8 +198,9 @@ Requires-Dist: intel-extension-for-pytorch==2.1.10+xpu; extra == "intel"
|
|
|
197
198
|
Provides-Extra: llama_cpp
|
|
198
199
|
Requires-Dist: llama-cpp-python!=0.2.58,>=0.2.25; extra == "llama-cpp"
|
|
199
200
|
Provides-Extra: mlx
|
|
201
|
+
Requires-Dist: mlx<0.22.0; extra == "mlx"
|
|
200
202
|
Requires-Dist: mlx-lm; extra == "mlx"
|
|
201
|
-
Requires-Dist: mlx-vlm; extra == "mlx"
|
|
203
|
+
Requires-Dist: mlx-vlm>=0.1.7; extra == "mlx"
|
|
202
204
|
Requires-Dist: mlx-whisper; extra == "mlx"
|
|
203
205
|
Requires-Dist: f5-tts-mlx; extra == "mlx"
|
|
204
206
|
Requires-Dist: qwen-vl-utils; extra == "mlx"
|
|
@@ -277,6 +279,7 @@ potential of cutting-edge AI models.
|
|
|
277
279
|
|
|
278
280
|
## 🔥 Hot Topics
|
|
279
281
|
### Framework Enhancements
|
|
282
|
+
- VLLM enhancement: Shared KV cache across multiple replicas: [#2732](https://github.com/xorbitsai/inference/pull/2732)
|
|
280
283
|
- Support Continuous batching for Transformers engine: [#1724](https://github.com/xorbitsai/inference/pull/1724)
|
|
281
284
|
- Support MLX backend for Apple Silicon chips: [#1765](https://github.com/xorbitsai/inference/pull/1765)
|
|
282
285
|
- Support specifying worker and GPU indexes for launching models: [#1195](https://github.com/xorbitsai/inference/pull/1195)
|
|
@@ -285,14 +288,14 @@ potential of cutting-edge AI models.
|
|
|
285
288
|
- Support speech recognition model: [#929](https://github.com/xorbitsai/inference/pull/929)
|
|
286
289
|
- Metrics support: [#906](https://github.com/xorbitsai/inference/pull/906)
|
|
287
290
|
### New Models
|
|
291
|
+
- Built-in support for [Stable Diffusion 3.5](https://huggingface.co/collections/stabilityai/stable-diffusion-35-671785cca799084f71fa2838): [#2706](https://github.com/xorbitsai/inference/pull/2706)
|
|
292
|
+
- Built-in support for [CosyVoice 2](https://huggingface.co/FunAudioLLM/CosyVoice2-0.5B): [#2684](https://github.com/xorbitsai/inference/pull/2684)
|
|
293
|
+
- Built-in support for [Fish Speech V1.5](https://huggingface.co/fishaudio/fish-speech-1.5): [#2672](https://github.com/xorbitsai/inference/pull/2672)
|
|
288
294
|
- Built-in support for [F5-TTS](https://github.com/SWivid/F5-TTS): [#2626](https://github.com/xorbitsai/inference/pull/2626)
|
|
289
295
|
- Built-in support for [GLM Edge](https://github.com/THUDM/GLM-Edge): [#2582](https://github.com/xorbitsai/inference/pull/2582)
|
|
290
296
|
- Built-in support for [QwQ-32B-Preview](https://qwenlm.github.io/blog/qwq-32b-preview/): [#2602](https://github.com/xorbitsai/inference/pull/2602)
|
|
291
297
|
- Built-in support for [Qwen 2.5 Series](https://qwenlm.github.io/blog/qwen2.5/): [#2325](https://github.com/xorbitsai/inference/pull/2325)
|
|
292
|
-
- Built-in support for [Fish Speech V1.4](https://huggingface.co/fishaudio/fish-speech-1.4): [#2295](https://github.com/xorbitsai/inference/pull/2295)
|
|
293
298
|
- Built-in support for [DeepSeek-V2.5](https://huggingface.co/deepseek-ai/DeepSeek-V2.5): [#2292](https://github.com/xorbitsai/inference/pull/2292)
|
|
294
|
-
- Built-in support for [Qwen2-Audio](https://github.com/QwenLM/Qwen2-Audio): [#2271](https://github.com/xorbitsai/inference/pull/2271)
|
|
295
|
-
- Built-in support for [Qwen2-vl-instruct](https://github.com/QwenLM/Qwen2-VL): [#2205](https://github.com/xorbitsai/inference/pull/2205)
|
|
296
299
|
### Integrations
|
|
297
300
|
- [Dify](https://docs.dify.ai/advanced/model-configuration/xinference): an LLMOps platform that enables developers (and even non-developers) to quickly build useful applications based on large language models, ensuring they are visual, operable, and improvable.
|
|
298
301
|
- [FastGPT](https://github.com/labring/FastGPT): a knowledge-based platform built on the LLM, offers out-of-the-box data processing and model invocation capabilities, allows for workflow orchestration through Flow visualization.
|