xinference 1.0.1__py3-none-any.whl → 1.2.1__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of xinference might be problematic. Click here for more details.

Files changed (343) hide show
  1. xinference/_compat.py +2 -0
  2. xinference/_version.py +3 -3
  3. xinference/api/restful_api.py +77 -71
  4. xinference/core/chat_interface.py +6 -1
  5. xinference/core/model.py +79 -19
  6. xinference/core/supervisor.py +172 -10
  7. xinference/core/utils.py +12 -8
  8. xinference/core/worker.py +102 -4
  9. xinference/deploy/cmdline.py +3 -1
  10. xinference/deploy/test/test_cmdline.py +56 -0
  11. xinference/isolation.py +24 -0
  12. xinference/model/audio/core.py +16 -0
  13. xinference/model/audio/cosyvoice.py +39 -6
  14. xinference/model/audio/f5tts.py +200 -0
  15. xinference/model/audio/f5tts_mlx.py +260 -0
  16. xinference/model/audio/fish_speech.py +36 -111
  17. xinference/model/audio/melotts.py +110 -0
  18. xinference/model/audio/model_spec.json +99 -3
  19. xinference/model/audio/model_spec_modelscope.json +27 -0
  20. xinference/model/audio/utils.py +32 -0
  21. xinference/model/audio/whisper.py +35 -10
  22. xinference/model/embedding/core.py +203 -142
  23. xinference/model/embedding/model_spec.json +7 -0
  24. xinference/model/embedding/model_spec_modelscope.json +8 -0
  25. xinference/model/image/core.py +69 -1
  26. xinference/model/image/model_spec.json +145 -4
  27. xinference/model/image/model_spec_modelscope.json +150 -4
  28. xinference/model/image/stable_diffusion/core.py +45 -13
  29. xinference/model/llm/__init__.py +4 -2
  30. xinference/model/llm/llm_family.json +536 -53
  31. xinference/model/llm/llm_family.py +15 -36
  32. xinference/model/llm/llm_family_modelscope.json +454 -20
  33. xinference/model/llm/memory.py +1 -1
  34. xinference/model/llm/mlx/core.py +248 -52
  35. xinference/model/llm/sglang/core.py +1 -0
  36. xinference/model/llm/transformers/chatglm.py +9 -5
  37. xinference/model/llm/transformers/cogagent.py +272 -0
  38. xinference/model/llm/transformers/core.py +2 -0
  39. xinference/model/llm/transformers/qwen2_vl.py +12 -1
  40. xinference/model/llm/transformers/utils.py +16 -8
  41. xinference/model/llm/utils.py +36 -4
  42. xinference/model/llm/vllm/core.py +53 -10
  43. xinference/model/llm/vllm/xavier/__init__.py +13 -0
  44. xinference/model/llm/vllm/xavier/allocator.py +74 -0
  45. xinference/model/llm/vllm/xavier/block.py +111 -0
  46. xinference/model/llm/vllm/xavier/block_manager.py +71 -0
  47. xinference/model/llm/vllm/xavier/block_tracker.py +129 -0
  48. xinference/model/llm/vllm/xavier/collective.py +74 -0
  49. xinference/model/llm/vllm/xavier/collective_manager.py +147 -0
  50. xinference/model/llm/vllm/xavier/engine.py +247 -0
  51. xinference/model/llm/vllm/xavier/executor.py +134 -0
  52. xinference/model/llm/vllm/xavier/scheduler.py +438 -0
  53. xinference/model/llm/vllm/xavier/test/__init__.py +13 -0
  54. xinference/model/llm/vllm/xavier/test/test_xavier.py +147 -0
  55. xinference/model/llm/vllm/xavier/transfer.py +319 -0
  56. xinference/model/video/diffusers.py +14 -0
  57. xinference/model/video/model_spec.json +15 -0
  58. xinference/model/video/model_spec_modelscope.json +16 -0
  59. xinference/thirdparty/cosyvoice/bin/average_model.py +92 -0
  60. xinference/thirdparty/cosyvoice/bin/export_jit.py +12 -2
  61. xinference/thirdparty/cosyvoice/bin/export_onnx.py +112 -0
  62. xinference/thirdparty/cosyvoice/bin/export_trt.sh +9 -0
  63. xinference/thirdparty/cosyvoice/bin/inference.py +5 -7
  64. xinference/thirdparty/cosyvoice/bin/spk2info.pt +0 -0
  65. xinference/thirdparty/cosyvoice/bin/train.py +42 -8
  66. xinference/thirdparty/cosyvoice/cli/cosyvoice.py +96 -25
  67. xinference/thirdparty/cosyvoice/cli/frontend.py +77 -30
  68. xinference/thirdparty/cosyvoice/cli/model.py +330 -80
  69. xinference/thirdparty/cosyvoice/dataset/dataset.py +6 -2
  70. xinference/thirdparty/cosyvoice/dataset/processor.py +76 -14
  71. xinference/thirdparty/cosyvoice/flow/decoder.py +92 -13
  72. xinference/thirdparty/cosyvoice/flow/flow.py +99 -9
  73. xinference/thirdparty/cosyvoice/flow/flow_matching.py +110 -13
  74. xinference/thirdparty/cosyvoice/flow/length_regulator.py +5 -4
  75. xinference/thirdparty/cosyvoice/hifigan/discriminator.py +140 -0
  76. xinference/thirdparty/cosyvoice/hifigan/generator.py +58 -42
  77. xinference/thirdparty/cosyvoice/hifigan/hifigan.py +67 -0
  78. xinference/thirdparty/cosyvoice/llm/llm.py +139 -6
  79. xinference/thirdparty/cosyvoice/tokenizer/assets/multilingual_zh_ja_yue_char_del.tiktoken +58836 -0
  80. xinference/thirdparty/cosyvoice/tokenizer/tokenizer.py +279 -0
  81. xinference/thirdparty/cosyvoice/transformer/embedding.py +2 -2
  82. xinference/thirdparty/cosyvoice/transformer/encoder_layer.py +7 -7
  83. xinference/thirdparty/cosyvoice/transformer/upsample_encoder.py +318 -0
  84. xinference/thirdparty/cosyvoice/utils/common.py +28 -1
  85. xinference/thirdparty/cosyvoice/utils/executor.py +69 -7
  86. xinference/thirdparty/cosyvoice/utils/file_utils.py +2 -12
  87. xinference/thirdparty/cosyvoice/utils/frontend_utils.py +9 -5
  88. xinference/thirdparty/cosyvoice/utils/losses.py +20 -0
  89. xinference/thirdparty/cosyvoice/utils/scheduler.py +1 -2
  90. xinference/thirdparty/cosyvoice/utils/train_utils.py +101 -45
  91. xinference/thirdparty/f5_tts/api.py +166 -0
  92. xinference/thirdparty/f5_tts/configs/E2TTS_Base_train.yaml +44 -0
  93. xinference/thirdparty/f5_tts/configs/E2TTS_Small_train.yaml +44 -0
  94. xinference/thirdparty/f5_tts/configs/F5TTS_Base_train.yaml +46 -0
  95. xinference/thirdparty/f5_tts/configs/F5TTS_Small_train.yaml +46 -0
  96. xinference/thirdparty/f5_tts/eval/README.md +49 -0
  97. xinference/thirdparty/f5_tts/eval/ecapa_tdnn.py +330 -0
  98. xinference/thirdparty/f5_tts/eval/eval_infer_batch.py +207 -0
  99. xinference/thirdparty/f5_tts/eval/eval_infer_batch.sh +13 -0
  100. xinference/thirdparty/f5_tts/eval/eval_librispeech_test_clean.py +84 -0
  101. xinference/thirdparty/f5_tts/eval/eval_seedtts_testset.py +84 -0
  102. xinference/thirdparty/f5_tts/eval/utils_eval.py +405 -0
  103. xinference/thirdparty/f5_tts/infer/README.md +191 -0
  104. xinference/thirdparty/f5_tts/infer/SHARED.md +74 -0
  105. xinference/thirdparty/f5_tts/infer/examples/basic/basic.toml +11 -0
  106. xinference/thirdparty/f5_tts/infer/examples/basic/basic_ref_en.wav +0 -0
  107. xinference/thirdparty/f5_tts/infer/examples/basic/basic_ref_zh.wav +0 -0
  108. xinference/thirdparty/f5_tts/infer/examples/multi/country.flac +0 -0
  109. xinference/thirdparty/f5_tts/infer/examples/multi/main.flac +0 -0
  110. xinference/thirdparty/f5_tts/infer/examples/multi/story.toml +19 -0
  111. xinference/thirdparty/f5_tts/infer/examples/multi/story.txt +1 -0
  112. xinference/thirdparty/f5_tts/infer/examples/multi/town.flac +0 -0
  113. xinference/thirdparty/f5_tts/infer/examples/vocab.txt +2545 -0
  114. xinference/thirdparty/f5_tts/infer/infer_cli.py +226 -0
  115. xinference/thirdparty/f5_tts/infer/infer_gradio.py +851 -0
  116. xinference/thirdparty/f5_tts/infer/speech_edit.py +193 -0
  117. xinference/thirdparty/f5_tts/infer/utils_infer.py +538 -0
  118. xinference/thirdparty/f5_tts/model/__init__.py +10 -0
  119. xinference/thirdparty/f5_tts/model/backbones/README.md +20 -0
  120. xinference/thirdparty/f5_tts/model/backbones/dit.py +163 -0
  121. xinference/thirdparty/f5_tts/model/backbones/mmdit.py +146 -0
  122. xinference/thirdparty/f5_tts/model/backbones/unett.py +219 -0
  123. xinference/thirdparty/f5_tts/model/cfm.py +285 -0
  124. xinference/thirdparty/f5_tts/model/dataset.py +319 -0
  125. xinference/thirdparty/f5_tts/model/modules.py +658 -0
  126. xinference/thirdparty/f5_tts/model/trainer.py +366 -0
  127. xinference/thirdparty/f5_tts/model/utils.py +185 -0
  128. xinference/thirdparty/f5_tts/scripts/count_max_epoch.py +33 -0
  129. xinference/thirdparty/f5_tts/scripts/count_params_gflops.py +39 -0
  130. xinference/thirdparty/f5_tts/socket_server.py +159 -0
  131. xinference/thirdparty/f5_tts/train/README.md +77 -0
  132. xinference/thirdparty/f5_tts/train/datasets/prepare_csv_wavs.py +139 -0
  133. xinference/thirdparty/f5_tts/train/datasets/prepare_emilia.py +230 -0
  134. xinference/thirdparty/f5_tts/train/datasets/prepare_libritts.py +92 -0
  135. xinference/thirdparty/f5_tts/train/datasets/prepare_ljspeech.py +65 -0
  136. xinference/thirdparty/f5_tts/train/datasets/prepare_wenetspeech4tts.py +125 -0
  137. xinference/thirdparty/f5_tts/train/finetune_cli.py +174 -0
  138. xinference/thirdparty/f5_tts/train/finetune_gradio.py +1846 -0
  139. xinference/thirdparty/f5_tts/train/train.py +75 -0
  140. xinference/thirdparty/fish_speech/fish_speech/conversation.py +94 -83
  141. xinference/thirdparty/fish_speech/fish_speech/models/text2semantic/llama.py +63 -20
  142. xinference/thirdparty/fish_speech/fish_speech/text/clean.py +1 -26
  143. xinference/thirdparty/fish_speech/fish_speech/text/spliter.py +1 -1
  144. xinference/thirdparty/fish_speech/fish_speech/tokenizer.py +152 -0
  145. xinference/thirdparty/fish_speech/fish_speech/train.py +2 -2
  146. xinference/thirdparty/fish_speech/fish_speech/webui/manage.py +1 -1
  147. xinference/thirdparty/fish_speech/tools/{post_api.py → api_client.py} +7 -13
  148. xinference/thirdparty/fish_speech/tools/api_server.py +98 -0
  149. xinference/thirdparty/fish_speech/tools/download_models.py +5 -5
  150. xinference/thirdparty/fish_speech/tools/fish_e2e.py +2 -2
  151. xinference/thirdparty/fish_speech/tools/inference_engine/__init__.py +192 -0
  152. xinference/thirdparty/fish_speech/tools/inference_engine/reference_loader.py +125 -0
  153. xinference/thirdparty/fish_speech/tools/inference_engine/utils.py +39 -0
  154. xinference/thirdparty/fish_speech/tools/inference_engine/vq_manager.py +57 -0
  155. xinference/thirdparty/fish_speech/tools/llama/eval_in_context.py +2 -2
  156. xinference/thirdparty/fish_speech/tools/llama/generate.py +117 -89
  157. xinference/thirdparty/fish_speech/tools/run_webui.py +104 -0
  158. xinference/thirdparty/fish_speech/tools/schema.py +11 -28
  159. xinference/thirdparty/fish_speech/tools/server/agent/__init__.py +57 -0
  160. xinference/thirdparty/fish_speech/tools/server/agent/generate.py +119 -0
  161. xinference/thirdparty/fish_speech/tools/server/agent/generation_utils.py +122 -0
  162. xinference/thirdparty/fish_speech/tools/server/agent/pre_generation_utils.py +72 -0
  163. xinference/thirdparty/fish_speech/tools/server/api_utils.py +75 -0
  164. xinference/thirdparty/fish_speech/tools/server/exception_handler.py +27 -0
  165. xinference/thirdparty/fish_speech/tools/server/inference.py +45 -0
  166. xinference/thirdparty/fish_speech/tools/server/model_manager.py +122 -0
  167. xinference/thirdparty/fish_speech/tools/server/model_utils.py +129 -0
  168. xinference/thirdparty/fish_speech/tools/server/views.py +246 -0
  169. xinference/thirdparty/fish_speech/tools/webui/__init__.py +173 -0
  170. xinference/thirdparty/fish_speech/tools/webui/inference.py +91 -0
  171. xinference/thirdparty/fish_speech/tools/webui/variables.py +14 -0
  172. xinference/thirdparty/matcha/utils/utils.py +2 -2
  173. xinference/thirdparty/melo/api.py +135 -0
  174. xinference/thirdparty/melo/app.py +61 -0
  175. xinference/thirdparty/melo/attentions.py +459 -0
  176. xinference/thirdparty/melo/commons.py +160 -0
  177. xinference/thirdparty/melo/configs/config.json +94 -0
  178. xinference/thirdparty/melo/data/example/metadata.list +20 -0
  179. xinference/thirdparty/melo/data_utils.py +413 -0
  180. xinference/thirdparty/melo/download_utils.py +67 -0
  181. xinference/thirdparty/melo/infer.py +25 -0
  182. xinference/thirdparty/melo/init_downloads.py +14 -0
  183. xinference/thirdparty/melo/losses.py +58 -0
  184. xinference/thirdparty/melo/main.py +36 -0
  185. xinference/thirdparty/melo/mel_processing.py +174 -0
  186. xinference/thirdparty/melo/models.py +1030 -0
  187. xinference/thirdparty/melo/modules.py +598 -0
  188. xinference/thirdparty/melo/monotonic_align/__init__.py +16 -0
  189. xinference/thirdparty/melo/monotonic_align/core.py +46 -0
  190. xinference/thirdparty/melo/preprocess_text.py +135 -0
  191. xinference/thirdparty/melo/split_utils.py +174 -0
  192. xinference/thirdparty/melo/text/__init__.py +35 -0
  193. xinference/thirdparty/melo/text/chinese.py +199 -0
  194. xinference/thirdparty/melo/text/chinese_bert.py +107 -0
  195. xinference/thirdparty/melo/text/chinese_mix.py +253 -0
  196. xinference/thirdparty/melo/text/cleaner.py +36 -0
  197. xinference/thirdparty/melo/text/cleaner_multiling.py +110 -0
  198. xinference/thirdparty/melo/text/cmudict.rep +129530 -0
  199. xinference/thirdparty/melo/text/cmudict_cache.pickle +0 -0
  200. xinference/thirdparty/melo/text/english.py +284 -0
  201. xinference/thirdparty/melo/text/english_bert.py +39 -0
  202. xinference/thirdparty/melo/text/english_utils/abbreviations.py +35 -0
  203. xinference/thirdparty/melo/text/english_utils/number_norm.py +97 -0
  204. xinference/thirdparty/melo/text/english_utils/time_norm.py +47 -0
  205. xinference/thirdparty/melo/text/es_phonemizer/base.py +140 -0
  206. xinference/thirdparty/melo/text/es_phonemizer/cleaner.py +109 -0
  207. xinference/thirdparty/melo/text/es_phonemizer/es_symbols.json +79 -0
  208. xinference/thirdparty/melo/text/es_phonemizer/es_symbols.txt +1 -0
  209. xinference/thirdparty/melo/text/es_phonemizer/es_symbols_v2.json +83 -0
  210. xinference/thirdparty/melo/text/es_phonemizer/es_to_ipa.py +12 -0
  211. xinference/thirdparty/melo/text/es_phonemizer/example_ipa.txt +400 -0
  212. xinference/thirdparty/melo/text/es_phonemizer/gruut_wrapper.py +253 -0
  213. xinference/thirdparty/melo/text/es_phonemizer/punctuation.py +174 -0
  214. xinference/thirdparty/melo/text/es_phonemizer/spanish_symbols.txt +1 -0
  215. xinference/thirdparty/melo/text/es_phonemizer/test.ipynb +124 -0
  216. xinference/thirdparty/melo/text/fr_phonemizer/base.py +140 -0
  217. xinference/thirdparty/melo/text/fr_phonemizer/cleaner.py +122 -0
  218. xinference/thirdparty/melo/text/fr_phonemizer/en_symbols.json +78 -0
  219. xinference/thirdparty/melo/text/fr_phonemizer/example_ipa.txt +1 -0
  220. xinference/thirdparty/melo/text/fr_phonemizer/fr_symbols.json +89 -0
  221. xinference/thirdparty/melo/text/fr_phonemizer/fr_to_ipa.py +30 -0
  222. xinference/thirdparty/melo/text/fr_phonemizer/french_abbreviations.py +48 -0
  223. xinference/thirdparty/melo/text/fr_phonemizer/french_symbols.txt +1 -0
  224. xinference/thirdparty/melo/text/fr_phonemizer/gruut_wrapper.py +258 -0
  225. xinference/thirdparty/melo/text/fr_phonemizer/punctuation.py +172 -0
  226. xinference/thirdparty/melo/text/french.py +94 -0
  227. xinference/thirdparty/melo/text/french_bert.py +39 -0
  228. xinference/thirdparty/melo/text/japanese.py +647 -0
  229. xinference/thirdparty/melo/text/japanese_bert.py +49 -0
  230. xinference/thirdparty/melo/text/ko_dictionary.py +44 -0
  231. xinference/thirdparty/melo/text/korean.py +192 -0
  232. xinference/thirdparty/melo/text/opencpop-strict.txt +429 -0
  233. xinference/thirdparty/melo/text/spanish.py +122 -0
  234. xinference/thirdparty/melo/text/spanish_bert.py +39 -0
  235. xinference/thirdparty/melo/text/symbols.py +290 -0
  236. xinference/thirdparty/melo/text/tone_sandhi.py +769 -0
  237. xinference/thirdparty/melo/train.py +635 -0
  238. xinference/thirdparty/melo/train.sh +19 -0
  239. xinference/thirdparty/melo/transforms.py +209 -0
  240. xinference/thirdparty/melo/utils.py +424 -0
  241. xinference/types.py +15 -0
  242. xinference/web/ui/build/asset-manifest.json +6 -6
  243. xinference/web/ui/build/index.html +1 -1
  244. xinference/web/ui/build/static/css/main.51a587ff.css +2 -0
  245. xinference/web/ui/build/static/css/main.51a587ff.css.map +1 -0
  246. xinference/web/ui/build/static/js/main.b0936c54.js +3 -0
  247. xinference/web/ui/build/static/js/main.b0936c54.js.map +1 -0
  248. xinference/web/ui/node_modules/.cache/babel-loader/03c4052f1b91f6ba0c5389bdcf49c43319b4076c08e4b8585dab312538ae290a.json +1 -0
  249. xinference/web/ui/node_modules/.cache/babel-loader/1786b83003b8e9605a0f5f855a185d4d16e38fc893dfb326a2a9cca206b4240a.json +1 -0
  250. xinference/web/ui/node_modules/.cache/babel-loader/17cbc181dd674b9150b80c73ed6a82656de0082d857f6e5f66d9716129ac0b38.json +1 -0
  251. xinference/web/ui/node_modules/.cache/babel-loader/185ceb8872d562e032b47e79df6a45670e06345b8ed70aad1a131e0476783c5c.json +1 -0
  252. xinference/web/ui/node_modules/.cache/babel-loader/26b8c9f34b0bed789b3a833767672e39302d1e0c09b4276f4d58d1df7b6bd93b.json +1 -0
  253. xinference/web/ui/node_modules/.cache/babel-loader/2b484da66c724d0d56a40849c109327408796a668b1381511b6e9e03baa48658.json +1 -0
  254. xinference/web/ui/node_modules/.cache/babel-loader/2cbbbce9b84df73330d4c42b82436ed881b3847628f2fbc346aa62e2859fd88c.json +1 -0
  255. xinference/web/ui/node_modules/.cache/babel-loader/2ec9b14431ed33ce6901bf9f27007be4e6e472709c99d6e22b50ce528e4b78ee.json +1 -0
  256. xinference/web/ui/node_modules/.cache/babel-loader/3b966db018f96be4a055d6ca205f0990d4d0b370e2980c17d8bca2c9a021819c.json +1 -0
  257. xinference/web/ui/node_modules/.cache/babel-loader/3eefb411b24c2b3ce053570ef50daccf154022f0e168be5ed0fec21394baf9f4.json +1 -0
  258. xinference/web/ui/node_modules/.cache/babel-loader/522b229e3cac219123f0d69673f5570e191c2d2a505dc65b312d336eae2279c0.json +1 -0
  259. xinference/web/ui/node_modules/.cache/babel-loader/52e45f17ba300580ea3fcc9f9228ccba194bb092b76f25e9255af311f8b05aab.json +1 -0
  260. xinference/web/ui/node_modules/.cache/babel-loader/5a0bc4631f936459afc1a3b1d3ec2420118b1f00e11f60ccac3e08088f3f27a8.json +1 -0
  261. xinference/web/ui/node_modules/.cache/babel-loader/611fa2c6c53b66039991d06dfb0473b5ab37fc63b4564e0f6e1718523768a045.json +1 -0
  262. xinference/web/ui/node_modules/.cache/babel-loader/6329bc76c406fe5eb305412383fbde5950f847bb5e43261f73f37622c365acb4.json +1 -0
  263. xinference/web/ui/node_modules/.cache/babel-loader/63c8e07687ea53a4f8a910ee5e42e0eb26cd1acbfbe820f3e3248a786ee51401.json +1 -0
  264. xinference/web/ui/node_modules/.cache/babel-loader/69b2d5001684174ec9da57e07914eed3eac4960018bceb6cbfa801d861301d7c.json +1 -0
  265. xinference/web/ui/node_modules/.cache/babel-loader/710c1acda69e561e30a933b98c6a56d50197868b15c21e2aad55ab6d46649eb6.json +1 -0
  266. xinference/web/ui/node_modules/.cache/babel-loader/720deca1fce5a1dc5056048fa8258fd138a82ea855f350b6613f104a73fb761f.json +1 -0
  267. xinference/web/ui/node_modules/.cache/babel-loader/76a23b92d26a499c57e61eea2b895fbc9771bd0849a72e66f8e633192017978b.json +1 -0
  268. xinference/web/ui/node_modules/.cache/babel-loader/858063f23b34dfe600254eb5afd85518b0002ec4b30b7386616c45600826e3b2.json +1 -0
  269. xinference/web/ui/node_modules/.cache/babel-loader/920b82c1c89124cf217109eeedbfcd3aae3b917be50c9dfb6bbb4ce26bdfd2e7.json +1 -0
  270. xinference/web/ui/node_modules/.cache/babel-loader/94d8b7aeb0076f2ce07db598cea0e87b13bc8d5614eb530b8d6e696c2daf6f88.json +1 -0
  271. xinference/web/ui/node_modules/.cache/babel-loader/9e917fe7022d01b2ccbe5cc0ce73d70bb72bee584ff293bad71bdff6695dee28.json +1 -0
  272. xinference/web/ui/node_modules/.cache/babel-loader/9f28fdb8399f1d0474f0aca86f1658dc94f5bf0c90f6146352de150692de8862.json +1 -0
  273. xinference/web/ui/node_modules/.cache/babel-loader/a0dfafa06b2bb7cba8cad41c482503f61944f759f4318139362602ef5cc47ccb.json +1 -0
  274. xinference/web/ui/node_modules/.cache/babel-loader/a3ff866acddf34917a7ee399e0e571a4dfd8ba66d5057db885f243e16a6eb17d.json +1 -0
  275. xinference/web/ui/node_modules/.cache/babel-loader/afb8084f539534cd594755ea2205ecd5bd1f62dddcfdf75a2eace59a28131278.json +1 -0
  276. xinference/web/ui/node_modules/.cache/babel-loader/b57b1438b77294c1f3f6cfce12ac487d8106c6f016975ba0aec94d98997e2e1e.json +1 -0
  277. xinference/web/ui/node_modules/.cache/babel-loader/b9917b0bf8e4d55ccbac1c334aa04d6ff3c5b6ed9e5d38b9ea2c687fa7d3f5a9.json +1 -0
  278. xinference/web/ui/node_modules/.cache/babel-loader/bbcc94b0149963d1d6f267ee1f4f03d3925b758392ce2f516c3fe8af0e0169fc.json +1 -0
  279. xinference/web/ui/node_modules/.cache/babel-loader/bdee44abeadc4abc17d41c52eb49c6e19a4b1a267b6e16876ce91bdeeebfc52d.json +1 -0
  280. xinference/web/ui/node_modules/.cache/babel-loader/beb112b70f4a56db95920a9e20efb6c97c37b68450716730217a9ee1a9ae92be.json +1 -0
  281. xinference/web/ui/node_modules/.cache/babel-loader/c88db97be0cdf440193b3995996e83510a04cb00048135485fc0e26d197e80b5.json +1 -0
  282. xinference/web/ui/node_modules/.cache/babel-loader/d49e5314d34310a62d01a03067ce1bec5da00abce84c5196aa9c6842fa79a430.json +1 -0
  283. xinference/web/ui/node_modules/.cache/babel-loader/d7664d18c4ddbad9c3a6a31b91f7c00fb0dde804608674a9860ee50f33e54708.json +1 -0
  284. xinference/web/ui/node_modules/.cache/babel-loader/d9072c318b819b7c90a0f7e9cc0b6413b4dbeb8e9859898e53d75ea882fcde99.json +1 -0
  285. xinference/web/ui/node_modules/.cache/babel-loader/db16a983bc08a05f0439cc61ca0840e49e1d8400eef678909f16c032a418a3d6.json +1 -0
  286. xinference/web/ui/node_modules/.cache/babel-loader/dc249829767b8abcbc3677e0b07b6d3ecbfdfe6d08cfe23a665eb33373a9aa9d.json +1 -0
  287. xinference/web/ui/node_modules/.cache/babel-loader/e242c583c2dbc2784f0fcf513523975f7d5df447e106c1c17e49e8578a6fc3ed.json +1 -0
  288. xinference/web/ui/node_modules/.cache/babel-loader/eac5f1296513e69e4b96f750ddccd4d0264e2bae4e4c449144e83274a48698d9.json +1 -0
  289. xinference/web/ui/node_modules/.cache/babel-loader/ed57202cb79649bb716400436590245547df241988fc7c8e1d85d132299542d2.json +1 -0
  290. xinference/web/ui/node_modules/.cache/babel-loader/f125bf72e773a14cdaebd0c343e80adb909d12e317ee5c00cd4a57442fbe2c62.json +1 -0
  291. xinference/web/ui/node_modules/.cache/babel-loader/f91af913d7f91c410719ab13136aaed3aaf0f8dda06652f25c42cb5231587398.json +1 -0
  292. xinference/web/ui/node_modules/.package-lock.json +67 -3
  293. xinference/web/ui/node_modules/@babel/runtime/package.json +592 -538
  294. xinference/web/ui/node_modules/html-parse-stringify/package.json +50 -0
  295. xinference/web/ui/node_modules/i18next/dist/esm/package.json +1 -0
  296. xinference/web/ui/node_modules/i18next/package.json +129 -0
  297. xinference/web/ui/node_modules/react-i18next/.eslintrc.json +74 -0
  298. xinference/web/ui/node_modules/react-i18next/dist/es/package.json +1 -0
  299. xinference/web/ui/node_modules/react-i18next/package.json +162 -0
  300. xinference/web/ui/node_modules/void-elements/package.json +34 -0
  301. xinference/web/ui/package-lock.json +69 -3
  302. xinference/web/ui/package.json +2 -0
  303. xinference/web/ui/src/locales/en.json +186 -0
  304. xinference/web/ui/src/locales/zh.json +186 -0
  305. {xinference-1.0.1.dist-info → xinference-1.2.1.dist-info}/METADATA +68 -32
  306. {xinference-1.0.1.dist-info → xinference-1.2.1.dist-info}/RECORD +316 -122
  307. xinference/thirdparty/cosyvoice/bin/export_trt.py +0 -8
  308. xinference/thirdparty/fish_speech/tools/api.py +0 -943
  309. xinference/thirdparty/fish_speech/tools/msgpack_api.py +0 -95
  310. xinference/thirdparty/fish_speech/tools/webui.py +0 -548
  311. xinference/web/ui/build/static/css/main.5061c4c3.css +0 -2
  312. xinference/web/ui/build/static/css/main.5061c4c3.css.map +0 -1
  313. xinference/web/ui/build/static/js/main.2f269bb3.js +0 -3
  314. xinference/web/ui/build/static/js/main.2f269bb3.js.map +0 -1
  315. xinference/web/ui/node_modules/.cache/babel-loader/07ce9e632e6aff24d7aa3ad8e48224433bbfeb0d633fca723453f1fcae0c9f1c.json +0 -1
  316. xinference/web/ui/node_modules/.cache/babel-loader/1130403f9e46f5738a23b45ac59b57de8f360c908c713e2c0670c2cce9bd367a.json +0 -1
  317. xinference/web/ui/node_modules/.cache/babel-loader/131091b25d26b17cdca187d7542a21475c211138d900cf667682260e76ef9463.json +0 -1
  318. xinference/web/ui/node_modules/.cache/babel-loader/1f269fb2a368363c1cb2237825f1dba093b6bdd8c44cc05954fd19ec2c1fff03.json +0 -1
  319. xinference/web/ui/node_modules/.cache/babel-loader/331312668fa8bd3d7401818f4a25fa98135d7f61371cd6bfff78b18cf4fbdd92.json +0 -1
  320. xinference/web/ui/node_modules/.cache/babel-loader/40f17338fc75ae095de7d2b4d8eae0d5ca0193a7e2bcece4ee745b22a7a2f4b7.json +0 -1
  321. xinference/web/ui/node_modules/.cache/babel-loader/4de9a6942c5f1749d6cbfdd54279699975f16016b182848bc253886f52ec2ec3.json +0 -1
  322. xinference/web/ui/node_modules/.cache/babel-loader/822586ed1077201b64b954f12f25e3f9b45678c1acbabe53d8af3ca82ca71f33.json +0 -1
  323. xinference/web/ui/node_modules/.cache/babel-loader/8d33354bd2100c8602afc3341f131a88cc36aaeecd5a4b365ed038514708e350.json +0 -1
  324. xinference/web/ui/node_modules/.cache/babel-loader/9375a35b05d56989b2755bf72161fa707c92f28569d33765a75f91a568fda6e9.json +0 -1
  325. xinference/web/ui/node_modules/.cache/babel-loader/a158a9ffa0c9b169aee53dd4a0c44501a596755b4e4f6ede7746d65a72e2a71f.json +0 -1
  326. xinference/web/ui/node_modules/.cache/babel-loader/bd6ad8159341315a1764c397621a560809f7eb7219ab5174c801fca7e969d943.json +0 -1
  327. xinference/web/ui/node_modules/.cache/babel-loader/c7bf40bab396765f67d0fed627ed3665890608b2d0edaa3e8cb7cfc96310db45.json +0 -1
  328. xinference/web/ui/node_modules/.cache/babel-loader/d6c643278a0b28320e6f33a60f5fb64c053997cbdc39a60e53ccc574688ade9e.json +0 -1
  329. xinference/web/ui/node_modules/.cache/babel-loader/e42b72d4cc1ea412ebecbb8d040dc6c6bfee462c33903c2f1f3facb602ad742e.json +0 -1
  330. xinference/web/ui/node_modules/.cache/babel-loader/e64b7e8cedcf43d4c95deba60ec1341855c887705805bb62431693118b870c69.json +0 -1
  331. xinference/web/ui/node_modules/.cache/babel-loader/f5039ddbeb815c51491a1989532006b96fc3ae49c6c60e3c097f875b4ae915ae.json +0 -1
  332. xinference/web/ui/node_modules/.cache/babel-loader/f72f011744c4649fabddca6f7a9327861ac0a315a89b1a2e62a39774e7863845.json +0 -1
  333. xinference/web/ui/node_modules/.cache/babel-loader/feabb04b4aa507102da0a64398a40818e878fd1df9b75dda8461b3e1e7ff3f11.json +0 -1
  334. /xinference/thirdparty/{cosyvoice/bin → f5_tts}/__init__.py +0 -0
  335. /xinference/thirdparty/{cosyvoice/flow → melo}/__init__.py +0 -0
  336. /xinference/thirdparty/{cosyvoice/hifigan → melo/text/english_utils}/__init__.py +0 -0
  337. /xinference/thirdparty/{cosyvoice/llm → melo/text/es_phonemizer}/__init__.py +0 -0
  338. /xinference/thirdparty/{fish_speech/tools → melo/text/fr_phonemizer}/__init__.py +0 -0
  339. /xinference/web/ui/build/static/js/{main.2f269bb3.js.LICENSE.txt → main.b0936c54.js.LICENSE.txt} +0 -0
  340. {xinference-1.0.1.dist-info → xinference-1.2.1.dist-info}/LICENSE +0 -0
  341. {xinference-1.0.1.dist-info → xinference-1.2.1.dist-info}/WHEEL +0 -0
  342. {xinference-1.0.1.dist-info → xinference-1.2.1.dist-info}/entry_points.txt +0 -0
  343. {xinference-1.0.1.dist-info → xinference-1.2.1.dist-info}/top_level.txt +0 -0
@@ -0,0 +1,186 @@
1
+ {
2
+ "menu": {
3
+ "launchModel": "Launch Model",
4
+ "runningModels": "Running Models",
5
+ "registerModel": "Register Model",
6
+ "clusterInfo": "Cluster Information",
7
+ "contactUs": "Contact Us"
8
+ },
9
+
10
+ "model": {
11
+ "languageModels": "Language Models",
12
+ "embeddingModels": "Embedding Models",
13
+ "rerankModels": "Rerank Models",
14
+ "imageModels": "Image Models",
15
+ "audioModels": "Audio Models",
16
+ "videoModels": "Video Models",
17
+ "customModels": "Custom Models",
18
+ "flexibleModels": "Flexible Models"
19
+ },
20
+
21
+ "launchModel": {
22
+ "modelAbility": "Model Ability",
23
+ "generate": "generate",
24
+ "chat": "chat",
25
+ "vision": "vl-chat",
26
+ "status": "Status",
27
+ "cached": "Cached",
28
+ "manageCachedModels": "Manage Cached Models",
29
+ "favorite": "Favorite",
30
+ "unfavorite": "Unfavorite",
31
+ "search": "Search for model name and description",
32
+ "searchModelType": "Search for {{modelType}} model name",
33
+ "searchInstruction": "Type {{hotkey}} to search",
34
+ "clickToLaunchModel": "Click with mouse to launch the model",
35
+ "dimensions": "dimensions",
36
+ "maxTokens": "max tokens",
37
+ "edit": "Edit",
38
+ "delete": "Delete",
39
+ "contextLength": "context length",
40
+ "chatModel": "chat model",
41
+ "generateModel": "generate model",
42
+ "otherModel": "other model",
43
+ "confirmDeleteCustomModel": "Are you sure to delete this custom model? This behavior is irreversible.",
44
+ "lastConfig": "Last Config",
45
+ "modelEngine": "Model Engine",
46
+ "modelFormat": "Model Format",
47
+ "modelSize": "Model Size",
48
+ "quantization": "Quantization",
49
+ "nGPU": "GPU Count",
50
+ "nGpuLayers": "N GPU Layers",
51
+ "replica": "Replica",
52
+ "optionalConfigurations": "Optional Configurations",
53
+ "modelUID.optional": "(Optional) Model UID, model name by default",
54
+ "requestLimits.optional": "(Optional) Request Limits, the number of request limits for this model, default is None",
55
+ "workerIp.optional": "(Optional) Worker Ip, specify the worker ip where the model is located in a distributed scenario",
56
+ "workerIp": "Worker Ip, specify the worker ip where the model is located in a distributed scenario",
57
+ "GPUIdx.optional": "(Optional) GPU Idx, Specify the GPU index where the model is located",
58
+ "GPUIdx": "GPU Idx, Specify the GPU index where the model is located",
59
+ "downloadHub.optional": "(Optional) Download_hub",
60
+ "modelPath.optional": "(Optional) Model Path, For PyTorch, provide the model directory. For GGML/GGUF, provide the model file path.",
61
+ "GGUFQuantization.optional": "(Optional) GGUF quantization format, quantizing the Transformer part.",
62
+ "GGUFModelPath.optional": "(Optional) GGUF model path, should be a file ending with .gguf.",
63
+ "CPUOffload": "CPU Offload",
64
+ "CPUOffload.tip": "Unload the model to the CPU. Recommend to enable this when resources are limited or when using the GGUF option.",
65
+ "loraConfig": "Lora Config",
66
+ "loraModelConfig": "Lora Model Config",
67
+ "additionalParametersForInferenceEngine": "Additional parameters passed to the inference engine",
68
+ "enterIntegerGreaterThanZero": "Please enter an integer greater than 0.",
69
+ "enterCommaSeparatedNumbers": "Please enter numeric data separated by commas, for example: 0,1,2",
70
+ "device": "Device",
71
+ "loraLoadKwargsForImageModel": "Lora Load Kwargs for Image Model",
72
+ "loraFuseKwargsForImageModel": "Lora Fuse Kwargs for Image Model",
73
+ "launch": "Launch",
74
+ "goBack": "Go Back",
75
+ "copyJson": "Copy Json",
76
+ "cancel": "Cancel",
77
+ "fillCompleteParametersBeforeAdding": "Please fill in the complete parameters before adding!",
78
+ "model_format": "model_format",
79
+ "model_size_in_billions": "model_size_in_billions",
80
+ "quantizations": "quantizations",
81
+ "real_path": "real_path",
82
+ "path": "path",
83
+ "ipAddress": "IP Address",
84
+ "operation": "operation",
85
+ "copyRealPath": "Copy real_path",
86
+ "copyPath": "Copy path",
87
+ "noCacheForNow": "No cache for now!",
88
+ "confirmDeleteCacheFiles": "Confirm deletion of cache files? This action is irreversible."
89
+ },
90
+
91
+ "runningModels": {
92
+ "name": "Name",
93
+ "address": "Address",
94
+ "gpuIndexes": "GPU Indexes",
95
+ "size": "Size",
96
+ "quantization": "Quantization",
97
+ "replica": "Replica",
98
+ "actions": "Actions",
99
+ "noRunningModels": "No Running Models",
100
+ "noRunningModelsMatches": "No Running Models Matches"
101
+ },
102
+
103
+ "registerModel": {
104
+ "modelName": "Model Name",
105
+ "modelDescription": "Model Description (Optional)",
106
+ "contextLength": "Context Length",
107
+ "dimensions": "Dimensions",
108
+ "maxTokens": "Max Tokens",
109
+ "modelPath": "Model Path",
110
+ "modelLanguages": "Model Languages",
111
+ "languages": "Languages",
112
+ "multilingual": "Multilingual",
113
+ "modelAbilities": "Model Abilities",
114
+ "modelFamily": "Model Family",
115
+ "chatTemplate": "Chat Template",
116
+ "test": "test",
117
+ "testResult": "test result",
118
+ "noTestResults": "No test results...",
119
+ "stopTokenIds": "Stop Token Ids",
120
+ "stop": "Stop",
121
+ "launcher": "Launcher",
122
+ "launcherArguments": "Launcher Arguments (Optional)",
123
+ "edit": "Edit",
124
+ "cancel": "Cancel",
125
+ "registerModel": "Register Model",
126
+ "messagesExample": "Messages Example",
127
+ "JSONFormat": "JSON Format",
128
+ "modelSpecs": "Model Specs",
129
+ "modelSizeBillions": "Model Size in Billions",
130
+ "quantization": "Quantization",
131
+ "quantizationOptional": "Quantization (Optional)",
132
+ "delete": "Delete",
133
+ "controlnet": "Controlnet",
134
+ "more": "more",
135
+ "modelFormat": "Model Format",
136
+ "enterNumberGreaterThanZero": "Please enter a number greater than 0.",
137
+ "carefulQuantizationForModelRegistration": "For GPTQ/AWQ/FP8/MLX models, please be careful to fill in the quantization corresponding to the model you want to register.",
138
+ "quantizationCannotBeEmpty": "Quantization cannot be left empty.",
139
+ "enterInteger": "Please enter an integer.",
140
+ "enterIntegerGreaterThanZero": "Please enter an integer greater than 0.",
141
+ "showCustomJsonConfig": "Show custom json config used by api",
142
+ "packUp": "Pack up",
143
+ "unfold": "Unfold",
144
+ "copyAll": "Copy all",
145
+ "alphanumericWithHyphensUnderscores": "Alphanumeric characters with properly placed hyphens and underscores. Must not match any built-in model names.",
146
+ "chooseBuiltInOrCustomModel": "You can choose from the built-in models or input your own.",
147
+ "chooseOnlyBuiltInModel": "You can only choose from the built-in models.",
148
+ "provideModelDirectoryPath": "Provide the model directory path.",
149
+ "provideModelLauncher": "Provide the model launcher.",
150
+ "jsonArgumentsForLauncher": "A JSON-formatted dictionary representing the arguments passed to the Launcher.",
151
+ "provideModelDirectoryOrFilePath": "For PyTorch, provide the model directory. For GGUF, provide the model file path.",
152
+ "ensureChatTemplatePassesTest": "Please make sure this chat_template passes the test by clicking the TEST button on the right. Please note that this test may not cover all cases and will only be used for the most basic case.",
153
+ "testFailurePreventsChatWorking": "Please note that failure to pass test may prevent chats from working properly.",
154
+ "stopControlForChatModels": "int type, used to control the stopping of chat models",
155
+ "stopControlStringForChatModels": "string type, used to control the stopping of chat models",
156
+ "enterJsonFormattedDictionary": "Please enter the JSON-formatted dictionary."
157
+ },
158
+
159
+ "clusterInfo": {
160
+ "supervisor": "Supervisor",
161
+ "workers": "Workers",
162
+ "workerDetails": "Worker Details",
163
+ "count": "Count",
164
+ "cpuInfo": "CPU Info",
165
+ "usage": "Usage:",
166
+ "total": "Total",
167
+ "cpuMemoryInfo": "CPU Memory Info",
168
+ "version": "Version",
169
+ "release": "Release:",
170
+ "commit": "Commit:",
171
+ "gpuInfo": "GPU Info",
172
+ "gpuMemoryInfo": "GPU Memory Info",
173
+ "address": "Address",
174
+ "item": "Item",
175
+ "value": "Value",
176
+ "nodeType": "Node Type",
177
+ "cpuUsage": "CPU Usage",
178
+ "cpuTotal": "CPU Total",
179
+ "memUsage": "Mem Usage",
180
+ "memTotal": "Mem Total",
181
+ "gpuCount": "GPU Count",
182
+ "gpuMemUsage": "GPU Mem Usage",
183
+ "gpuMemTotal": "GPU Mem Total",
184
+ "worker": "Worker"
185
+ }
186
+ }
@@ -0,0 +1,186 @@
1
+ {
2
+ "menu": {
3
+ "launchModel": "启动模型",
4
+ "runningModels": "运行模型",
5
+ "registerModel": "注册模型",
6
+ "clusterInfo": "集群信息",
7
+ "contactUs": "联系我们"
8
+ },
9
+
10
+ "model": {
11
+ "languageModels": "语言模型",
12
+ "embeddingModels": "嵌入模型",
13
+ "rerankModels": "重排序模型",
14
+ "imageModels": "图像模型",
15
+ "audioModels": "音频模型",
16
+ "videoModels": "视频模型",
17
+ "customModels": "自定义模型",
18
+ "flexibleModels": "灵活模型"
19
+ },
20
+
21
+ "launchModel": {
22
+ "modelAbility": "模型能力",
23
+ "generate": "生成",
24
+ "chat": "聊天",
25
+ "vision": "视觉聊天",
26
+ "status": "状态",
27
+ "cached": "已缓存",
28
+ "manageCachedModels": "管理缓存模型",
29
+ "favorite": "收藏",
30
+ "unfavorite": "取消收藏",
31
+ "search": "搜索模型名称和描述",
32
+ "searchModelType": "搜索 {{modelType}} 相关的模型名称",
33
+ "searchInstruction": "输入 {{hotkey}} 进行搜索",
34
+ "clickToLaunchModel": "点击鼠标以启动模型",
35
+ "dimensions": "维度",
36
+ "maxTokens": "最大 token 数",
37
+ "edit": "编辑",
38
+ "delete": "删除",
39
+ "contextLength": "上下文长度",
40
+ "chatModel": "聊天模型",
41
+ "generateModel": "生成模型",
42
+ "otherModel": "其他模型",
43
+ "confirmDeleteCustomModel": "您确定要删除这个自定义模型吗?此操作无法恢复。",
44
+ "lastConfig": "最后配置",
45
+ "modelEngine": "模型引擎",
46
+ "modelFormat": "模型格式",
47
+ "modelSize": "模型大小",
48
+ "quantization": "量化",
49
+ "nGPU": "GPU 数量",
50
+ "nGpuLayers": "GPU 层数",
51
+ "replica": "副本",
52
+ "optionalConfigurations": "可选配置",
53
+ "modelUID.optional": "(可选) 模型 UID,默认是模型名称",
54
+ "requestLimits.optional": "(可选) 请求限制,模型的请求限制数,默认值为无",
55
+ "workerIp.optional": "(可选) 工作节点 IP,在分布式场景中指定模型所在的工作节点 IP",
56
+ "workerIp": "工作节点 IP,在分布式场景中指定模型所在的工作节点 IP",
57
+ "GPUIdx.optional": "(可选) GPU 索引,指定模型所在的 GPU 索引",
58
+ "GPUIdx": "GPU 索引,指定模型所在的 GPU 索引",
59
+ "downloadHub.optional": "(可选) 下载中心",
60
+ "modelPath.optional": "(可选) 模型路径,对于 PyTorch,提供模型目录;对于 GGML/GGUF,提供模型文件路径。",
61
+ "GGUFQuantization.optional": "(可选) GGUF量化格式,对Transformer部分进行量化。",
62
+ "GGUFModelPath.optional": "(可选) GGUF模型路径,应为以 .gguf 结尾的文件。",
63
+ "CPUOffload": "CPU卸载",
64
+ "CPUOffload.tip": "将模型卸载到CPU。当资源有限或使用GGUF选项时,建议启用此功能。",
65
+ "loraConfig": "Lora 配置",
66
+ "loraModelConfig": "Lora 模型配置",
67
+ "additionalParametersForInferenceEngine": "传递给推理引擎的附加参数",
68
+ "enterIntegerGreaterThanZero": "请输入大于 0 的整数。",
69
+ "enterCommaSeparatedNumbers": "请输入以逗号分隔的数字数据,例如:0,1,2",
70
+ "device": "设备",
71
+ "loraLoadKwargsForImageModel": "图像模型的 Lora 加载参数",
72
+ "loraFuseKwargsForImageModel": "图像模型的 Lora 融合参数",
73
+ "launch": "启动",
74
+ "goBack": "返回",
75
+ "copyJson": "复制 JSON",
76
+ "cancel": "取消",
77
+ "fillCompleteParametersBeforeAdding": "请在添加之前填写完整的参数!",
78
+ "model_format": "模型格式",
79
+ "model_size_in_billions": "模型大小(以十亿为单位)",
80
+ "quantizations": "量化方式",
81
+ "real_path": "真实路径",
82
+ "path": "路径",
83
+ "ipAddress": "IP 地址",
84
+ "operation": "操作",
85
+ "copyRealPath": "复制真实路径",
86
+ "copyPath": "复制路径",
87
+ "noCacheForNow": "当前没有缓存!",
88
+ "confirmDeleteCacheFiles": "确认删除缓存文件吗?此操作无法恢复。"
89
+ },
90
+
91
+ "runningModels": {
92
+ "name": "名称",
93
+ "address": "地址",
94
+ "gpuIndexes": "GPU 索引",
95
+ "size": "大小",
96
+ "quantization": "量化",
97
+ "replica": "副本",
98
+ "actions": "操作",
99
+ "noRunningModels": "没有运行中的模型",
100
+ "noRunningModelsMatches": "没有匹配的运行模型"
101
+ },
102
+
103
+ "registerModel": {
104
+ "modelName": "模型名称",
105
+ "modelDescription": "模型描述(可选)",
106
+ "contextLength": "上下文长度",
107
+ "dimensions": "维度",
108
+ "maxTokens": "最大 token 数",
109
+ "modelPath": "模型路径",
110
+ "modelLanguages": "模型语言",
111
+ "languages": "语言",
112
+ "multilingual": "多语言",
113
+ "modelAbilities": "模型能力",
114
+ "modelFamily": "模型系列",
115
+ "chatTemplate": "聊天模板",
116
+ "test": "测试",
117
+ "testResult": "测试结果",
118
+ "noTestResults": "没有测试结果...",
119
+ "stopTokenIds": "停止token ID",
120
+ "stop": "停止",
121
+ "launcher": "启动器",
122
+ "launcherArguments": "启动器参数(可选)",
123
+ "edit": "编辑",
124
+ "cancel": "取消",
125
+ "registerModel": "注册模型",
126
+ "messagesExample": "消息示例",
127
+ "JSONFormat": "JSON 格式",
128
+ "modelSpecs": "模型规格",
129
+ "modelSizeBillions": "模型大小(以十亿为单位)",
130
+ "quantization": "量化",
131
+ "quantizationOptional": "量化(可选)",
132
+ "delete": "删除",
133
+ "controlnet": "控制网",
134
+ "more": "更多",
135
+ "modelFormat": "模型格式",
136
+ "enterNumberGreaterThanZero": "请输入大于 0 的数字。",
137
+ "carefulQuantizationForModelRegistration": "对于 GPTQ/AWQ/FP8/MLX 模型,请小心填写与您要注册的模型对应的量化方式。",
138
+ "quantizationCannotBeEmpty": "量化方式不能为空。",
139
+ "enterInteger": "请输入一个整数。",
140
+ "enterIntegerGreaterThanZero": "请输入大于 0 的整数。",
141
+ "showCustomJsonConfig": "显示由 API 使用的自定义 JSON 配置",
142
+ "packUp": "收起",
143
+ "unfold": "展开",
144
+ "copyAll": "复制全部",
145
+ "alphanumericWithHyphensUnderscores": "字母数字字符,连字符和下划线应正确放置。不能与任何内置模型名称匹配。",
146
+ "chooseBuiltInOrCustomModel": "您可以选择内置模型或输入自定义模型。",
147
+ "chooseOnlyBuiltInModel": "您只能从内置模型中选择。",
148
+ "provideModelDirectoryPath": "提供模型目录路径。",
149
+ "provideModelLauncher": "提供模型启动器。",
150
+ "jsonArgumentsForLauncher": "一个 JSON 格式的字典,表示传递给启动器的参数。",
151
+ "provideModelDirectoryOrFilePath": "对于 PyTorch,提供模型目录。对于 GGUF,提供模型文件路径。",
152
+ "ensureChatTemplatePassesTest": "请确保通过点击右侧的测试按钮,使此聊天模板通过测试。请注意,此测试可能无法涵盖所有情况,只会用于最基本的情况。",
153
+ "testFailurePreventsChatWorking": "请注意,未通过测试可能会导致聊天无法正常工作。",
154
+ "stopControlForChatModels": "整数类型,用于控制聊天模型的停止。",
155
+ "stopControlStringForChatModels": "字符串类型,用于控制聊天模型的停止。",
156
+ "enterJsonFormattedDictionary": "请输入 JSON 格式的字典。"
157
+ },
158
+
159
+ "clusterInfo": {
160
+ "supervisor": "主管",
161
+ "workers": "工作节点",
162
+ "workerDetails": "工作节点详情",
163
+ "count": "数量",
164
+ "cpuInfo": "CPU 信息",
165
+ "usage": "使用率:",
166
+ "total": "总计",
167
+ "cpuMemoryInfo": "CPU 内存信息",
168
+ "version": "版本",
169
+ "release": "发布:",
170
+ "commit": "提交:",
171
+ "gpuInfo": "GPU 信息",
172
+ "gpuMemoryInfo": "GPU 内存信息",
173
+ "address": "地址",
174
+ "item": "项",
175
+ "value": "值",
176
+ "nodeType": "节点类型",
177
+ "cpuUsage": "CPU 使用率",
178
+ "cpuTotal": "CPU 总数",
179
+ "memUsage": "内存使用率",
180
+ "memTotal": "内存总量",
181
+ "gpuCount": "GPU 数量",
182
+ "gpuMemUsage": "GPU 内存使用率",
183
+ "gpuMemTotal": "GPU 内存总量",
184
+ "worker": "工作节点"
185
+ }
186
+ }
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: xinference
3
- Version: 1.0.1
3
+ Version: 1.2.1
4
4
  Summary: Model Serving Made Easy
5
5
  Home-page: https://github.com/xorbitsai/inference
6
6
  Author: Qin Xuye
@@ -19,7 +19,7 @@ Classifier: Programming Language :: Python :: Implementation :: CPython
19
19
  Classifier: Topic :: Software Development :: Libraries
20
20
  Description-Content-Type: text/markdown
21
21
  License-File: LICENSE
22
- Requires-Dist: xoscar<0.4.2,>=0.3.0
22
+ Requires-Dist: xoscar>=0.4.4
23
23
  Requires-Dist: torch
24
24
  Requires-Dist: gradio
25
25
  Requires-Dist: pillow
@@ -45,7 +45,7 @@ Requires-Dist: timm
45
45
  Requires-Dist: setproctitle
46
46
  Provides-Extra: all
47
47
  Requires-Dist: llama-cpp-python!=0.2.58,>=0.2.25; extra == "all"
48
- Requires-Dist: transformers>=4.43.2; extra == "all"
48
+ Requires-Dist: transformers>=4.46.0; extra == "all"
49
49
  Requires-Dist: torch>=2.0.0; extra == "all"
50
50
  Requires-Dist: accelerate>=0.28.0; extra == "all"
51
51
  Requires-Dist: sentencepiece; extra == "all"
@@ -55,7 +55,6 @@ Requires-Dist: protobuf; extra == "all"
55
55
  Requires-Dist: einops; extra == "all"
56
56
  Requires-Dist: tiktoken>=0.6.0; extra == "all"
57
57
  Requires-Dist: sentence-transformers>=3.1.0; extra == "all"
58
- Requires-Dist: diffusers>=0.30.0; extra == "all"
59
58
  Requires-Dist: imageio-ffmpeg; extra == "all"
60
59
  Requires-Dist: controlnet-aux; extra == "all"
61
60
  Requires-Dist: orjson; extra == "all"
@@ -65,10 +64,8 @@ Requires-Dist: attrdict; extra == "all"
65
64
  Requires-Dist: timm>=0.9.16; extra == "all"
66
65
  Requires-Dist: torchvision; extra == "all"
67
66
  Requires-Dist: FlagEmbedding; extra == "all"
68
- Requires-Dist: funasr; extra == "all"
67
+ Requires-Dist: funasr<1.1.17; extra == "all"
69
68
  Requires-Dist: omegaconf~=2.3.0; extra == "all"
70
- Requires-Dist: nemo-text-processing<1.1.0; extra == "all"
71
- Requires-Dist: WeTextProcessing<1.0.4; extra == "all"
72
69
  Requires-Dist: librosa; extra == "all"
73
70
  Requires-Dist: xxhash; extra == "all"
74
71
  Requires-Dist: torchaudio; extra == "all"
@@ -77,6 +74,8 @@ Requires-Dist: lightning>=2.0.0; extra == "all"
77
74
  Requires-Dist: hydra-core>=1.3.2; extra == "all"
78
75
  Requires-Dist: inflect; extra == "all"
79
76
  Requires-Dist: conformer; extra == "all"
77
+ Requires-Dist: diffusers>=0.32.0; extra == "all"
78
+ Requires-Dist: gguf; extra == "all"
80
79
  Requires-Dist: gdown; extra == "all"
81
80
  Requires-Dist: pyarrow; extra == "all"
82
81
  Requires-Dist: HyperPyYAML; extra == "all"
@@ -91,21 +90,31 @@ Requires-Dist: loralib; extra == "all"
91
90
  Requires-Dist: ormsgpack; extra == "all"
92
91
  Requires-Dist: cachetools; extra == "all"
93
92
  Requires-Dist: silero-vad; extra == "all"
93
+ Requires-Dist: vector-quantize-pytorch<=1.17.3,>=1.14.24; extra == "all"
94
+ Requires-Dist: torchdiffeq; extra == "all"
95
+ Requires-Dist: x-transformers>=1.31.14; extra == "all"
96
+ Requires-Dist: pypinyin; extra == "all"
97
+ Requires-Dist: tomli; extra == "all"
98
+ Requires-Dist: vocos; extra == "all"
99
+ Requires-Dist: jieba; extra == "all"
100
+ Requires-Dist: soundfile; extra == "all"
94
101
  Requires-Dist: qwen-vl-utils; extra == "all"
95
102
  Requires-Dist: datamodel-code-generator; extra == "all"
96
103
  Requires-Dist: jsonschema; extra == "all"
97
104
  Requires-Dist: verovio>=4.3.1; extra == "all"
98
105
  Requires-Dist: auto-gptq; sys_platform != "darwin" and extra == "all"
99
106
  Requires-Dist: autoawq<0.2.6; sys_platform != "darwin" and extra == "all"
100
- Requires-Dist: mlx-lm; (sys_platform == "darwin" and platform_machine == "arm64") and extra == "all"
107
+ Requires-Dist: mlx-lm>=0.21.1; (sys_platform == "darwin" and platform_machine == "arm64") and extra == "all"
108
+ Requires-Dist: mlx-vlm>=0.1.11; (sys_platform == "darwin" and platform_machine == "arm64") and extra == "all"
101
109
  Requires-Dist: mlx-whisper; (sys_platform == "darwin" and platform_machine == "arm64") and extra == "all"
110
+ Requires-Dist: f5-tts-mlx; (sys_platform == "darwin" and platform_machine == "arm64") and extra == "all"
102
111
  Requires-Dist: vllm>=0.2.6; sys_platform == "linux" and extra == "all"
103
112
  Requires-Dist: sglang>=0.2.7; sys_platform == "linux" and extra == "all"
113
+ Requires-Dist: nemo-text-processing<1.1.0; sys_platform == "linux" and extra == "all"
114
+ Requires-Dist: WeTextProcessing<1.0.4; sys_platform == "linux" and extra == "all"
104
115
  Provides-Extra: audio
105
- Requires-Dist: funasr; extra == "audio"
116
+ Requires-Dist: funasr<1.1.17; extra == "audio"
106
117
  Requires-Dist: omegaconf~=2.3.0; extra == "audio"
107
- Requires-Dist: nemo-text-processing<1.1.0; extra == "audio"
108
- Requires-Dist: WeTextProcessing<1.0.4; extra == "audio"
109
118
  Requires-Dist: librosa; extra == "audio"
110
119
  Requires-Dist: xxhash; extra == "audio"
111
120
  Requires-Dist: torchaudio; extra == "audio"
@@ -116,7 +125,7 @@ Requires-Dist: lightning>=2.0.0; extra == "audio"
116
125
  Requires-Dist: hydra-core>=1.3.2; extra == "audio"
117
126
  Requires-Dist: inflect; extra == "audio"
118
127
  Requires-Dist: conformer; extra == "audio"
119
- Requires-Dist: diffusers>=0.30.0; extra == "audio"
128
+ Requires-Dist: diffusers>=0.32.0; extra == "audio"
120
129
  Requires-Dist: gdown; extra == "audio"
121
130
  Requires-Dist: pyarrow; extra == "audio"
122
131
  Requires-Dist: HyperPyYAML; extra == "audio"
@@ -127,6 +136,26 @@ Requires-Dist: loralib; extra == "audio"
127
136
  Requires-Dist: ormsgpack; extra == "audio"
128
137
  Requires-Dist: cachetools; extra == "audio"
129
138
  Requires-Dist: silero-vad; extra == "audio"
139
+ Requires-Dist: vector-quantize-pytorch<=1.17.3,>=1.14.24; extra == "audio"
140
+ Requires-Dist: torchdiffeq; extra == "audio"
141
+ Requires-Dist: x-transformers>=1.31.14; extra == "audio"
142
+ Requires-Dist: pypinyin; extra == "audio"
143
+ Requires-Dist: tomli; extra == "audio"
144
+ Requires-Dist: vocos; extra == "audio"
145
+ Requires-Dist: jieba; extra == "audio"
146
+ Requires-Dist: soundfile; extra == "audio"
147
+ Requires-Dist: cached-path; extra == "audio"
148
+ Requires-Dist: unidic-lite; extra == "audio"
149
+ Requires-Dist: cn2an; extra == "audio"
150
+ Requires-Dist: mecab-python3; extra == "audio"
151
+ Requires-Dist: num2words; extra == "audio"
152
+ Requires-Dist: pykakasi; extra == "audio"
153
+ Requires-Dist: fugashi; extra == "audio"
154
+ Requires-Dist: g2p-en; extra == "audio"
155
+ Requires-Dist: anyascii; extra == "audio"
156
+ Requires-Dist: gruut[de,es,fr]; extra == "audio"
157
+ Requires-Dist: nemo-text-processing<1.1.0; sys_platform == "linux" and extra == "audio"
158
+ Requires-Dist: WeTextProcessing<1.0.4; sys_platform == "linux" and extra == "audio"
130
159
  Provides-Extra: benchmark
131
160
  Requires-Dist: psutil; extra == "benchmark"
132
161
  Provides-Extra: dev
@@ -162,9 +191,10 @@ Requires-Dist: timm; extra == "doc"
162
191
  Provides-Extra: embedding
163
192
  Requires-Dist: sentence-transformers>=3.1.0; extra == "embedding"
164
193
  Provides-Extra: image
165
- Requires-Dist: diffusers>=0.30.0; extra == "image"
194
+ Requires-Dist: diffusers>=0.32.0; extra == "image"
166
195
  Requires-Dist: controlnet-aux; extra == "image"
167
196
  Requires-Dist: deepcache; extra == "image"
197
+ Requires-Dist: gguf; extra == "image"
168
198
  Requires-Dist: verovio>=4.3.1; extra == "image"
169
199
  Requires-Dist: transformers>=4.37.2; extra == "image"
170
200
  Requires-Dist: tiktoken>=0.6.0; extra == "image"
@@ -177,8 +207,12 @@ Requires-Dist: intel-extension-for-pytorch==2.1.10+xpu; extra == "intel"
177
207
  Provides-Extra: llama_cpp
178
208
  Requires-Dist: llama-cpp-python!=0.2.58,>=0.2.25; extra == "llama-cpp"
179
209
  Provides-Extra: mlx
180
- Requires-Dist: mlx-lm; extra == "mlx"
210
+ Requires-Dist: mlx-lm>=0.21.1; extra == "mlx"
211
+ Requires-Dist: mlx-vlm>=0.1.11; extra == "mlx"
181
212
  Requires-Dist: mlx-whisper; extra == "mlx"
213
+ Requires-Dist: f5-tts-mlx; extra == "mlx"
214
+ Requires-Dist: qwen-vl-utils; extra == "mlx"
215
+ Requires-Dist: tomli; extra == "mlx"
182
216
  Provides-Extra: rerank
183
217
  Requires-Dist: FlagEmbedding; extra == "rerank"
184
218
  Provides-Extra: sglang
@@ -186,7 +220,7 @@ Requires-Dist: outlines>=0.0.34; extra == "sglang"
186
220
  Requires-Dist: sglang>=0.2.7; sys_platform == "linux" and extra == "sglang"
187
221
  Requires-Dist: vllm>=0.5.2; sys_platform == "linux" and extra == "sglang"
188
222
  Provides-Extra: transformers
189
- Requires-Dist: transformers>=4.43.2; extra == "transformers"
223
+ Requires-Dist: transformers>=4.46.0; extra == "transformers"
190
224
  Requires-Dist: torch; extra == "transformers"
191
225
  Requires-Dist: accelerate>=0.28.0; extra == "transformers"
192
226
  Requires-Dist: sentencepiece; extra == "transformers"
@@ -208,7 +242,7 @@ Requires-Dist: jsonschema; extra == "transformers"
208
242
  Requires-Dist: auto-gptq; sys_platform != "darwin" and extra == "transformers"
209
243
  Requires-Dist: autoawq<0.2.6; sys_platform != "darwin" and extra == "transformers"
210
244
  Provides-Extra: video
211
- Requires-Dist: diffusers>=0.30.0; extra == "video"
245
+ Requires-Dist: diffusers>=0.32.0; extra == "video"
212
246
  Requires-Dist: imageio-ffmpeg; extra == "video"
213
247
  Provides-Extra: vllm
214
248
  Requires-Dist: vllm>=0.2.6; extra == "vllm"
@@ -228,7 +262,7 @@ Requires-Dist: vllm>=0.2.6; extra == "vllm"
228
262
  [![PyPI Latest Release](https://img.shields.io/pypi/v/xinference.svg?style=for-the-badge)](https://pypi.org/project/xinference/)
229
263
  [![License](https://img.shields.io/pypi/l/xinference.svg?style=for-the-badge)](https://github.com/xorbitsai/inference/blob/main/LICENSE)
230
264
  [![Build Status](https://img.shields.io/github/actions/workflow/status/xorbitsai/inference/python.yaml?branch=main&style=for-the-badge&label=GITHUB%20ACTIONS&logo=github)](https://actions-badge.atrox.dev/xorbitsai/inference/goto?ref=main)
231
- [![Slack](https://img.shields.io/badge/join_Slack-781FF5.svg?logo=slack&style=for-the-badge)](https://join.slack.com/t/xorbitsio/shared_invite/zt-1o3z9ucdh-RbfhbPVpx7prOVdM1CAuxg)
265
+ [![Discord](https://img.shields.io/badge/join_Discord-5462eb.svg?logo=discord&style=for-the-badge&logoColor=%23f5f5f5)](https://discord.gg/Xw9tszSkr5)
232
266
  [![Twitter](https://img.shields.io/twitter/follow/xorbitsio?logo=x&style=for-the-badge)](https://twitter.com/xorbitsio)
233
267
 
234
268
  <p align="center">
@@ -248,11 +282,12 @@ researcher, developer, or data scientist, Xorbits Inference empowers you to unle
248
282
  potential of cutting-edge AI models.
249
283
 
250
284
  <div align="center">
251
- <i><a href="https://join.slack.com/t/xorbitsio/shared_invite/zt-1z3zsm9ep-87yI9YZ_B79HLB2ccTq4WA">👉 Join our Slack community!</a></i>
285
+ <i><a href="https://discord.gg/Xw9tszSkr5">👉 Join our Discord community!</a></i>
252
286
  </div>
253
287
 
254
288
  ## 🔥 Hot Topics
255
289
  ### Framework Enhancements
290
+ - VLLM enhancement: Shared KV cache across multiple replicas: [#2732](https://github.com/xorbitsai/inference/pull/2732)
256
291
  - Support Continuous batching for Transformers engine: [#1724](https://github.com/xorbitsai/inference/pull/1724)
257
292
  - Support MLX backend for Apple Silicon chips: [#1765](https://github.com/xorbitsai/inference/pull/1765)
258
293
  - Support specifying worker and GPU indexes for launching models: [#1195](https://github.com/xorbitsai/inference/pull/1195)
@@ -261,19 +296,20 @@ potential of cutting-edge AI models.
261
296
  - Support speech recognition model: [#929](https://github.com/xorbitsai/inference/pull/929)
262
297
  - Metrics support: [#906](https://github.com/xorbitsai/inference/pull/906)
263
298
  ### New Models
264
- - Built-in support for [Qwen 2.5 Series](https://qwenlm.github.io/blog/qwen2.5/): [#2325](https://github.com/xorbitsai/inference/pull/2325)
265
- - Built-in support for [Fish Speech V1.4](https://huggingface.co/fishaudio/fish-speech-1.4): [#2295](https://github.com/xorbitsai/inference/pull/2295)
266
- - Built-in support for [DeepSeek-V2.5](https://huggingface.co/deepseek-ai/DeepSeek-V2.5): [#2292](https://github.com/xorbitsai/inference/pull/2292)
267
- - Built-in support for [Qwen2-Audio](https://github.com/QwenLM/Qwen2-Audio): [#2271](https://github.com/xorbitsai/inference/pull/2271)
268
- - Built-in support for [Qwen2-vl-instruct](https://github.com/QwenLM/Qwen2-VL): [#2205](https://github.com/xorbitsai/inference/pull/2205)
269
- - Built-in support for [MiniCPM3-4B](https://huggingface.co/openbmb/MiniCPM3-4B): [#2263](https://github.com/xorbitsai/inference/pull/2263)
270
- - Built-in support for [CogVideoX](https://github.com/THUDM/CogVideo): [#2049](https://github.com/xorbitsai/inference/pull/2049)
271
- - Built-in support for [flux.1-schnell & flux.1-dev](https://www.basedlabs.ai/tools/flux1): [#2007](https://github.com/xorbitsai/inference/pull/2007)
299
+ - Built-in support for [MeloTTS](https://github.com/myshell-ai/MeloTTS): [#2760](https://github.com/xorbitsai/inference/pull/2760)
300
+ - Built-in support for [CogAgent](https://github.com/THUDM/CogAgent): [#2740](https://github.com/xorbitsai/inference/pull/2740)
301
+ - Built-in support for [HunyuanVideo](https://github.com/Tencent/HunyuanVideo): [#2721](https://github.com/xorbitsai/inference/pull/2721)
302
+ - Built-in support for [HunyuanDiT](https://github.com/Tencent/HunyuanDiT): [#2727](https://github.com/xorbitsai/inference/pull/2727)
303
+ - Built-in support for [Macro-o1](https://github.com/AIDC-AI/Marco-o1): [#2749](https://github.com/xorbitsai/inference/pull/2749)
304
+ - Built-in support for [Stable Diffusion 3.5](https://huggingface.co/collections/stabilityai/stable-diffusion-35-671785cca799084f71fa2838): [#2706](https://github.com/xorbitsai/inference/pull/2706)
305
+ - Built-in support for [CosyVoice 2](https://huggingface.co/FunAudioLLM/CosyVoice2-0.5B): [#2684](https://github.com/xorbitsai/inference/pull/2684)
306
+ - Built-in support for [Fish Speech V1.5](https://huggingface.co/fishaudio/fish-speech-1.5): [#2672](https://github.com/xorbitsai/inference/pull/2672)
272
307
  ### Integrations
273
308
  - [Dify](https://docs.dify.ai/advanced/model-configuration/xinference): an LLMOps platform that enables developers (and even non-developers) to quickly build useful applications based on large language models, ensuring they are visual, operable, and improvable.
274
309
  - [FastGPT](https://github.com/labring/FastGPT): a knowledge-based platform built on the LLM, offers out-of-the-box data processing and model invocation capabilities, allows for workflow orchestration through Flow visualization.
275
- - [Chatbox](https://chatboxai.app/): a desktop client for multiple cutting-edge LLM models, available on Windows, Mac and Linux.
276
310
  - [RAGFlow](https://github.com/infiniflow/ragflow): is an open-source RAG engine based on deep document understanding.
311
+ - [MaxKB](https://github.com/1Panel-dev/MaxKB): MaxKB = Max Knowledge Base, it is a chatbot based on Large Language Models (LLM) and Retrieval-Augmented Generation (RAG).
312
+ - [Chatbox](https://chatboxai.app/): a desktop client for multiple cutting-edge LLM models, available on Windows, Mac and Linux.
277
313
 
278
314
 
279
315
  ## Key Features
@@ -389,11 +425,11 @@ Once Xinference is running, there are multiple ways you can try it: via the web
389
425
 
390
426
  ## Getting involved
391
427
 
392
- | Platform | Purpose |
393
- |-----------------------------------------------------------------------------------------------|----------------------------------------------------|
394
- | [Github Issues](https://github.com/xorbitsai/inference/issues) | Reporting bugs and filing feature requests. |
395
- | [Slack](https://join.slack.com/t/xorbitsio/shared_invite/zt-1o3z9ucdh-RbfhbPVpx7prOVdM1CAuxg) | Collaborating with other Xorbits users. |
396
- | [Twitter](https://twitter.com/xorbitsio) | Staying up-to-date on new features. |
428
+ | Platform | Purpose |
429
+ |-------------------------------------------------------------------------------------------------|---------------------------------------------|
430
+ | [Github Issues](https://github.com/xorbitsai/inference/issues) | Reporting bugs and filing feature requests. |
431
+ | [Discord](https://discord.gg/Xw9tszSkr5) | Collaborating with other Xinference users. |
432
+ | [Twitter](https://twitter.com/xorbitsio) | Staying up-to-date on new features. |
397
433
 
398
434
  ## Citation
399
435