xinference 0.14.4.post1__py3-none-any.whl → 0.15.1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of xinference might be problematic. Click here for more details.
- xinference/_compat.py +51 -0
- xinference/_version.py +3 -3
- xinference/api/restful_api.py +209 -40
- xinference/client/restful/restful_client.py +7 -26
- xinference/conftest.py +1 -1
- xinference/constants.py +5 -0
- xinference/core/cache_tracker.py +1 -1
- xinference/core/chat_interface.py +8 -14
- xinference/core/event.py +1 -1
- xinference/core/image_interface.py +28 -0
- xinference/core/model.py +110 -31
- xinference/core/scheduler.py +37 -37
- xinference/core/status_guard.py +1 -1
- xinference/core/supervisor.py +17 -10
- xinference/core/utils.py +80 -22
- xinference/core/worker.py +17 -16
- xinference/deploy/cmdline.py +8 -16
- xinference/deploy/local.py +1 -1
- xinference/deploy/supervisor.py +1 -1
- xinference/deploy/utils.py +1 -1
- xinference/deploy/worker.py +1 -1
- xinference/model/audio/cosyvoice.py +86 -41
- xinference/model/audio/fish_speech.py +9 -9
- xinference/model/audio/model_spec.json +9 -9
- xinference/model/audio/whisper.py +4 -1
- xinference/model/embedding/core.py +52 -31
- xinference/model/image/core.py +2 -1
- xinference/model/image/model_spec.json +16 -4
- xinference/model/image/model_spec_modelscope.json +16 -4
- xinference/model/image/sdapi.py +136 -0
- xinference/model/image/stable_diffusion/core.py +164 -19
- xinference/model/llm/__init__.py +29 -11
- xinference/model/llm/llama_cpp/core.py +16 -33
- xinference/model/llm/llm_family.json +1011 -1296
- xinference/model/llm/llm_family.py +34 -53
- xinference/model/llm/llm_family_csghub.json +18 -35
- xinference/model/llm/llm_family_modelscope.json +981 -1122
- xinference/model/llm/lmdeploy/core.py +56 -88
- xinference/model/llm/mlx/core.py +46 -69
- xinference/model/llm/sglang/core.py +36 -18
- xinference/model/llm/transformers/chatglm.py +168 -306
- xinference/model/llm/transformers/cogvlm2.py +36 -63
- xinference/model/llm/transformers/cogvlm2_video.py +33 -223
- xinference/model/llm/transformers/core.py +55 -50
- xinference/model/llm/transformers/deepseek_v2.py +340 -0
- xinference/model/llm/transformers/deepseek_vl.py +53 -96
- xinference/model/llm/transformers/glm4v.py +55 -111
- xinference/model/llm/transformers/intern_vl.py +39 -70
- xinference/model/llm/transformers/internlm2.py +32 -54
- xinference/model/llm/transformers/minicpmv25.py +22 -55
- xinference/model/llm/transformers/minicpmv26.py +158 -68
- xinference/model/llm/transformers/omnilmm.py +5 -28
- xinference/model/llm/transformers/qwen2_audio.py +168 -0
- xinference/model/llm/transformers/qwen2_vl.py +234 -0
- xinference/model/llm/transformers/qwen_vl.py +34 -86
- xinference/model/llm/transformers/utils.py +32 -38
- xinference/model/llm/transformers/yi_vl.py +32 -72
- xinference/model/llm/utils.py +280 -554
- xinference/model/llm/vllm/core.py +161 -100
- xinference/model/rerank/core.py +41 -8
- xinference/model/rerank/model_spec.json +7 -0
- xinference/model/rerank/model_spec_modelscope.json +7 -1
- xinference/model/utils.py +1 -31
- xinference/thirdparty/cosyvoice/bin/export_jit.py +64 -0
- xinference/thirdparty/cosyvoice/bin/export_trt.py +8 -0
- xinference/thirdparty/cosyvoice/bin/inference.py +5 -2
- xinference/thirdparty/cosyvoice/cli/cosyvoice.py +38 -22
- xinference/thirdparty/cosyvoice/cli/model.py +139 -26
- xinference/thirdparty/cosyvoice/flow/flow.py +15 -9
- xinference/thirdparty/cosyvoice/flow/length_regulator.py +20 -1
- xinference/thirdparty/cosyvoice/hifigan/generator.py +8 -4
- xinference/thirdparty/cosyvoice/llm/llm.py +14 -13
- xinference/thirdparty/cosyvoice/transformer/attention.py +7 -3
- xinference/thirdparty/cosyvoice/transformer/decoder.py +1 -1
- xinference/thirdparty/cosyvoice/transformer/embedding.py +4 -3
- xinference/thirdparty/cosyvoice/transformer/encoder.py +4 -2
- xinference/thirdparty/cosyvoice/utils/common.py +36 -0
- xinference/thirdparty/cosyvoice/utils/file_utils.py +16 -0
- xinference/thirdparty/deepseek_vl/serve/assets/Kelpy-Codos.js +100 -0
- xinference/thirdparty/deepseek_vl/serve/assets/avatar.png +0 -0
- xinference/thirdparty/deepseek_vl/serve/assets/custom.css +355 -0
- xinference/thirdparty/deepseek_vl/serve/assets/custom.js +22 -0
- xinference/thirdparty/deepseek_vl/serve/assets/favicon.ico +0 -0
- xinference/thirdparty/deepseek_vl/serve/examples/app.png +0 -0
- xinference/thirdparty/deepseek_vl/serve/examples/chart.png +0 -0
- xinference/thirdparty/deepseek_vl/serve/examples/mirror.png +0 -0
- xinference/thirdparty/deepseek_vl/serve/examples/pipeline.png +0 -0
- xinference/thirdparty/deepseek_vl/serve/examples/puzzle.png +0 -0
- xinference/thirdparty/deepseek_vl/serve/examples/rap.jpeg +0 -0
- xinference/thirdparty/fish_speech/fish_speech/configs/base.yaml +87 -0
- xinference/thirdparty/fish_speech/fish_speech/configs/firefly_gan_vq.yaml +33 -0
- xinference/thirdparty/fish_speech/fish_speech/configs/lora/r_8_alpha_16.yaml +4 -0
- xinference/thirdparty/fish_speech/fish_speech/configs/text2semantic_finetune.yaml +83 -0
- xinference/thirdparty/fish_speech/fish_speech/datasets/protos/text-data.proto +24 -0
- xinference/thirdparty/fish_speech/fish_speech/i18n/README.md +27 -0
- xinference/thirdparty/fish_speech/fish_speech/i18n/locale/en_US.json +1 -1
- xinference/thirdparty/fish_speech/fish_speech/i18n/locale/es_ES.json +1 -1
- xinference/thirdparty/fish_speech/fish_speech/i18n/locale/ja_JP.json +1 -1
- xinference/thirdparty/fish_speech/fish_speech/i18n/locale/pt_BR.json +1 -1
- xinference/thirdparty/fish_speech/fish_speech/i18n/locale/zh_CN.json +1 -1
- xinference/thirdparty/fish_speech/fish_speech/models/text2semantic/llama.py +2 -2
- xinference/thirdparty/fish_speech/fish_speech/models/vqgan/__init__.py +0 -3
- xinference/thirdparty/fish_speech/fish_speech/models/vqgan/modules/firefly.py +169 -198
- xinference/thirdparty/fish_speech/fish_speech/models/vqgan/modules/fsq.py +4 -27
- xinference/thirdparty/fish_speech/fish_speech/text/chn_text_norm/.gitignore +114 -0
- xinference/thirdparty/fish_speech/fish_speech/text/chn_text_norm/README.md +36 -0
- xinference/thirdparty/fish_speech/fish_speech/text/clean.py +9 -47
- xinference/thirdparty/fish_speech/fish_speech/text/spliter.py +2 -2
- xinference/thirdparty/fish_speech/fish_speech/train.py +2 -0
- xinference/thirdparty/fish_speech/fish_speech/webui/css/style.css +161 -0
- xinference/thirdparty/fish_speech/fish_speech/webui/html/footer.html +11 -0
- xinference/thirdparty/fish_speech/fish_speech/webui/js/animate.js +69 -0
- xinference/thirdparty/fish_speech/fish_speech/webui/manage.py +12 -10
- xinference/thirdparty/fish_speech/tools/api.py +79 -134
- xinference/thirdparty/fish_speech/tools/commons.py +35 -0
- xinference/thirdparty/fish_speech/tools/download_models.py +3 -3
- xinference/thirdparty/fish_speech/tools/file.py +17 -0
- xinference/thirdparty/fish_speech/tools/llama/build_dataset.py +1 -1
- xinference/thirdparty/fish_speech/tools/llama/generate.py +29 -24
- xinference/thirdparty/fish_speech/tools/llama/merge_lora.py +1 -1
- xinference/thirdparty/fish_speech/tools/llama/quantize.py +2 -2
- xinference/thirdparty/fish_speech/tools/msgpack_api.py +34 -0
- xinference/thirdparty/fish_speech/tools/post_api.py +85 -44
- xinference/thirdparty/fish_speech/tools/sensevoice/README.md +59 -0
- xinference/thirdparty/fish_speech/tools/sensevoice/fun_asr.py +1 -1
- xinference/thirdparty/fish_speech/tools/smart_pad.py +16 -3
- xinference/thirdparty/fish_speech/tools/vqgan/extract_vq.py +2 -2
- xinference/thirdparty/fish_speech/tools/vqgan/inference.py +4 -2
- xinference/thirdparty/fish_speech/tools/webui.py +12 -146
- xinference/thirdparty/matcha/VERSION +1 -0
- xinference/thirdparty/matcha/hifigan/LICENSE +21 -0
- xinference/thirdparty/matcha/hifigan/README.md +101 -0
- xinference/thirdparty/omnilmm/LICENSE +201 -0
- xinference/thirdparty/whisper/__init__.py +156 -0
- xinference/thirdparty/whisper/__main__.py +3 -0
- xinference/thirdparty/whisper/assets/gpt2.tiktoken +50256 -0
- xinference/thirdparty/whisper/assets/mel_filters.npz +0 -0
- xinference/thirdparty/whisper/assets/multilingual.tiktoken +50257 -0
- xinference/thirdparty/whisper/audio.py +157 -0
- xinference/thirdparty/whisper/decoding.py +826 -0
- xinference/thirdparty/whisper/model.py +314 -0
- xinference/thirdparty/whisper/normalizers/__init__.py +2 -0
- xinference/thirdparty/whisper/normalizers/basic.py +76 -0
- xinference/thirdparty/whisper/normalizers/english.json +1741 -0
- xinference/thirdparty/whisper/normalizers/english.py +550 -0
- xinference/thirdparty/whisper/timing.py +386 -0
- xinference/thirdparty/whisper/tokenizer.py +395 -0
- xinference/thirdparty/whisper/transcribe.py +605 -0
- xinference/thirdparty/whisper/triton_ops.py +109 -0
- xinference/thirdparty/whisper/utils.py +316 -0
- xinference/thirdparty/whisper/version.py +1 -0
- xinference/types.py +14 -53
- xinference/web/ui/build/asset-manifest.json +6 -6
- xinference/web/ui/build/index.html +1 -1
- xinference/web/ui/build/static/css/{main.4bafd904.css → main.5061c4c3.css} +2 -2
- xinference/web/ui/build/static/css/main.5061c4c3.css.map +1 -0
- xinference/web/ui/build/static/js/main.754740c0.js +3 -0
- xinference/web/ui/build/static/js/{main.eb13fe95.js.LICENSE.txt → main.754740c0.js.LICENSE.txt} +2 -0
- xinference/web/ui/build/static/js/main.754740c0.js.map +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/10c69dc7a296779fcffedeff9393d832dfcb0013c36824adf623d3c518b801ff.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/68bede6d95bb5ef0b35bbb3ec5b8c937eaf6862c6cdbddb5ef222a7776aaf336.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/77d50223f3e734d4485cca538cb098a8c3a7a0a1a9f01f58cdda3af42fe1adf5.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/a56d5a642409a84988891089c98ca28ad0546432dfbae8aaa51bc5a280e1cdd2.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/cd90b08d177025dfe84209596fc51878f8a86bcaa6a240848a3d2e5fd4c7ff24.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/d9ff696a3e3471f01b46c63d18af32e491eb5dc0e43cb30202c96871466df57f.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/e42b72d4cc1ea412ebecbb8d040dc6c6bfee462c33903c2f1f3facb602ad742e.json +1 -0
- xinference/web/ui/node_modules/.cache/babel-loader/f5039ddbeb815c51491a1989532006b96fc3ae49c6c60e3c097f875b4ae915ae.json +1 -0
- xinference/web/ui/node_modules/.package-lock.json +37 -0
- xinference/web/ui/node_modules/a-sync-waterfall/package.json +21 -0
- xinference/web/ui/node_modules/nunjucks/node_modules/commander/package.json +48 -0
- xinference/web/ui/node_modules/nunjucks/package.json +112 -0
- xinference/web/ui/package-lock.json +38 -0
- xinference/web/ui/package.json +1 -0
- {xinference-0.14.4.post1.dist-info → xinference-0.15.1.dist-info}/METADATA +16 -10
- {xinference-0.14.4.post1.dist-info → xinference-0.15.1.dist-info}/RECORD +179 -127
- xinference/model/llm/transformers/llama_2.py +0 -108
- xinference/thirdparty/fish_speech/fish_speech/models/vqgan/lit_module.py +0 -442
- xinference/thirdparty/fish_speech/fish_speech/models/vqgan/modules/discriminator.py +0 -44
- xinference/thirdparty/fish_speech/fish_speech/models/vqgan/modules/reference.py +0 -115
- xinference/thirdparty/fish_speech/fish_speech/models/vqgan/modules/wavenet.py +0 -225
- xinference/thirdparty/fish_speech/tools/auto_rerank.py +0 -159
- xinference/thirdparty/fish_speech/tools/gen_ref.py +0 -36
- xinference/thirdparty/fish_speech/tools/merge_asr_files.py +0 -55
- xinference/web/ui/build/static/css/main.4bafd904.css.map +0 -1
- xinference/web/ui/build/static/js/main.eb13fe95.js +0 -3
- xinference/web/ui/build/static/js/main.eb13fe95.js.map +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/0b11a5339468c13b2d31ac085e7effe4303259b2071abd46a0a8eb8529233a5e.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/213b5913e164773c2b0567455377765715f5f07225fbac77ad8e1e9dc9648a47.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/5c26a23b5eacf5b752a08531577ae3840bb247745ef9a39583dc2d05ba93a82a.json +0 -1
- xinference/web/ui/node_modules/.cache/babel-loader/978b57d1a04a701bc3fcfebc511f5f274eed6ed7eade67f6fb76c27d5fd9ecc8.json +0 -1
- {xinference-0.14.4.post1.dist-info → xinference-0.15.1.dist-info}/LICENSE +0 -0
- {xinference-0.14.4.post1.dist-info → xinference-0.15.1.dist-info}/WHEEL +0 -0
- {xinference-0.14.4.post1.dist-info → xinference-0.15.1.dist-info}/entry_points.txt +0 -0
- {xinference-0.14.4.post1.dist-info → xinference-0.15.1.dist-info}/top_level.txt +0 -0
|
@@ -46,24 +46,15 @@
|
|
|
46
46
|
"model_revision": "3cb06f589b7b1e2f8e728c77280b1114191d24de"
|
|
47
47
|
}
|
|
48
48
|
],
|
|
49
|
-
"
|
|
50
|
-
|
|
51
|
-
|
|
52
|
-
|
|
53
|
-
|
|
54
|
-
|
|
55
|
-
|
|
56
|
-
"
|
|
57
|
-
|
|
58
|
-
"stop_token_ids": [
|
|
59
|
-
70000
|
|
60
|
-
],
|
|
61
|
-
"stop": [
|
|
62
|
-
"<|endoftext|>",
|
|
63
|
-
"|||",
|
|
64
|
-
"|<end>|"
|
|
65
|
-
]
|
|
66
|
-
}
|
|
49
|
+
"chat_template": "{% for item in messages %}{% if item['role'] == 'user' %}{{ '## human: ' + item['content'] + '|<end>|' }}{% elif item['role'] == 'assistant' %}{{ '## assistant: ' + item['content'] + '|<end>|' }}{% endif %}{% endfor %}{{ '## assistant: ' }}",
|
|
50
|
+
"stop_token_ids": [
|
|
51
|
+
70000
|
|
52
|
+
],
|
|
53
|
+
"stop": [
|
|
54
|
+
"<|endoftext|>",
|
|
55
|
+
"|||",
|
|
56
|
+
"|<end>|"
|
|
57
|
+
]
|
|
67
58
|
},
|
|
68
59
|
{
|
|
69
60
|
"version": 1,
|
|
@@ -134,26 +125,17 @@
|
|
|
134
125
|
"model_revision": "ebee18c488086b396dde649f2aa6548b9b8d2404"
|
|
135
126
|
}
|
|
136
127
|
],
|
|
137
|
-
"
|
|
138
|
-
|
|
139
|
-
|
|
140
|
-
|
|
141
|
-
|
|
142
|
-
|
|
143
|
-
|
|
144
|
-
"
|
|
145
|
-
"
|
|
146
|
-
"
|
|
147
|
-
|
|
148
|
-
32001,
|
|
149
|
-
32007
|
|
150
|
-
],
|
|
151
|
-
"stop": [
|
|
152
|
-
"<|endoftext|>",
|
|
153
|
-
"<|assistant|>",
|
|
154
|
-
"<|end|>"
|
|
155
|
-
]
|
|
156
|
-
}
|
|
128
|
+
"chat_template": "{% for message in messages %}{% if message['role'] == 'system' %}{{'<|system|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ '<|endoftext|>' }}{% endif %}",
|
|
129
|
+
"stop_token_ids":[
|
|
130
|
+
32000,
|
|
131
|
+
32001,
|
|
132
|
+
32007
|
|
133
|
+
],
|
|
134
|
+
"stop": [
|
|
135
|
+
"<|endoftext|>",
|
|
136
|
+
"<|assistant|>",
|
|
137
|
+
"<|end|>"
|
|
138
|
+
]
|
|
157
139
|
},
|
|
158
140
|
{
|
|
159
141
|
"version": 1,
|
|
@@ -189,156 +171,17 @@
|
|
|
189
171
|
"model_revision": "b86bcaf57ea4dfdec5dbe12a377028b2fab0d480"
|
|
190
172
|
}
|
|
191
173
|
],
|
|
192
|
-
"
|
|
193
|
-
|
|
194
|
-
|
|
195
|
-
|
|
196
|
-
|
|
197
|
-
"assistant"
|
|
198
|
-
],
|
|
199
|
-
"intra_message_sep": "\n",
|
|
200
|
-
"inter_message_sep": "<|end|>\n",
|
|
201
|
-
"stop_token_ids":[
|
|
202
|
-
32000,
|
|
203
|
-
32001,
|
|
204
|
-
32007
|
|
205
|
-
],
|
|
206
|
-
"stop": [
|
|
207
|
-
"<|endoftext|>",
|
|
208
|
-
"<|assistant|>",
|
|
209
|
-
"<|end|>"
|
|
210
|
-
]
|
|
211
|
-
}
|
|
212
|
-
},
|
|
213
|
-
{
|
|
214
|
-
"version": 1,
|
|
215
|
-
"context_length": 8192,
|
|
216
|
-
"model_name": "chatglm3",
|
|
217
|
-
"model_lang": [
|
|
218
|
-
"en",
|
|
219
|
-
"zh"
|
|
220
|
-
],
|
|
221
|
-
"model_ability": [
|
|
222
|
-
"chat",
|
|
223
|
-
"tools"
|
|
224
|
-
],
|
|
225
|
-
"model_description": "ChatGLM3 is the third generation of ChatGLM, still open-source and trained on Chinese and English data.",
|
|
226
|
-
"model_specs": [
|
|
227
|
-
{
|
|
228
|
-
"model_format": "pytorch",
|
|
229
|
-
"model_size_in_billions": 6,
|
|
230
|
-
"quantizations": [
|
|
231
|
-
"4-bit",
|
|
232
|
-
"8-bit",
|
|
233
|
-
"none"
|
|
234
|
-
],
|
|
235
|
-
"model_id": "THUDM/chatglm3-6b",
|
|
236
|
-
"model_revision": "103caa40027ebfd8450289ca2f278eac4ff26405"
|
|
237
|
-
}
|
|
238
|
-
],
|
|
239
|
-
"prompt_style": {
|
|
240
|
-
"style_name": "CHATGLM3",
|
|
241
|
-
"system_prompt": "",
|
|
242
|
-
"roles": [
|
|
243
|
-
"user",
|
|
244
|
-
"assistant"
|
|
245
|
-
],
|
|
246
|
-
"stop_token_ids": [
|
|
247
|
-
64795,
|
|
248
|
-
64797,
|
|
249
|
-
2
|
|
250
|
-
],
|
|
251
|
-
"stop": [
|
|
252
|
-
"<|user|>",
|
|
253
|
-
"<|observation|>"
|
|
254
|
-
]
|
|
255
|
-
}
|
|
256
|
-
},
|
|
257
|
-
{
|
|
258
|
-
"version": 1,
|
|
259
|
-
"context_length": 32768,
|
|
260
|
-
"model_name": "chatglm3-32k",
|
|
261
|
-
"model_lang": [
|
|
262
|
-
"en",
|
|
263
|
-
"zh"
|
|
264
|
-
],
|
|
265
|
-
"model_ability": [
|
|
266
|
-
"chat"
|
|
267
|
-
],
|
|
268
|
-
"model_description": "ChatGLM3 is the third generation of ChatGLM, still open-source and trained on Chinese and English data.",
|
|
269
|
-
"model_specs": [
|
|
270
|
-
{
|
|
271
|
-
"model_format": "pytorch",
|
|
272
|
-
"model_size_in_billions": 6,
|
|
273
|
-
"quantizations": [
|
|
274
|
-
"4-bit",
|
|
275
|
-
"8-bit",
|
|
276
|
-
"none"
|
|
277
|
-
],
|
|
278
|
-
"model_id": "THUDM/chatglm3-6b-32k",
|
|
279
|
-
"model_revision": "339f17ff464d47b5077527c2b34e80a7719ede3e"
|
|
280
|
-
}
|
|
281
|
-
],
|
|
282
|
-
"prompt_style": {
|
|
283
|
-
"style_name": "CHATGLM3",
|
|
284
|
-
"system_prompt": "",
|
|
285
|
-
"roles": [
|
|
286
|
-
"user",
|
|
287
|
-
"assistant"
|
|
288
|
-
],
|
|
289
|
-
"stop_token_ids": [
|
|
290
|
-
64795,
|
|
291
|
-
64797,
|
|
292
|
-
2
|
|
293
|
-
],
|
|
294
|
-
"stop": [
|
|
295
|
-
"<|user|>",
|
|
296
|
-
"<|observation|>"
|
|
297
|
-
]
|
|
298
|
-
}
|
|
299
|
-
},
|
|
300
|
-
{
|
|
301
|
-
"version": 1,
|
|
302
|
-
"context_length": 131072,
|
|
303
|
-
"model_name": "chatglm3-128k",
|
|
304
|
-
"model_lang": [
|
|
305
|
-
"en",
|
|
306
|
-
"zh"
|
|
307
|
-
],
|
|
308
|
-
"model_ability": [
|
|
309
|
-
"chat"
|
|
310
|
-
],
|
|
311
|
-
"model_description": "ChatGLM3 is the third generation of ChatGLM, still open-source and trained on Chinese and English data.",
|
|
312
|
-
"model_specs": [
|
|
313
|
-
{
|
|
314
|
-
"model_format": "pytorch",
|
|
315
|
-
"model_size_in_billions": 6,
|
|
316
|
-
"quantizations": [
|
|
317
|
-
"4-bit",
|
|
318
|
-
"8-bit",
|
|
319
|
-
"none"
|
|
320
|
-
],
|
|
321
|
-
"model_id": "THUDM/chatglm3-6b-128k",
|
|
322
|
-
"model_revision": "f0afbe671009abc9e31182170cf60636d5546cda"
|
|
323
|
-
}
|
|
174
|
+
"chat_template": "{% for message in messages %}{% if message['role'] == 'system' %}{{'<|system|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ '<|endoftext|>' }}{% endif %}",
|
|
175
|
+
"stop_token_ids":[
|
|
176
|
+
32000,
|
|
177
|
+
32001,
|
|
178
|
+
32007
|
|
324
179
|
],
|
|
325
|
-
"
|
|
326
|
-
"
|
|
327
|
-
"
|
|
328
|
-
"
|
|
329
|
-
|
|
330
|
-
"assistant"
|
|
331
|
-
],
|
|
332
|
-
"stop_token_ids": [
|
|
333
|
-
64795,
|
|
334
|
-
64797,
|
|
335
|
-
2
|
|
336
|
-
],
|
|
337
|
-
"stop": [
|
|
338
|
-
"<|user|>",
|
|
339
|
-
"<|observation|>"
|
|
340
|
-
]
|
|
341
|
-
}
|
|
180
|
+
"stop": [
|
|
181
|
+
"<|endoftext|>",
|
|
182
|
+
"<|assistant|>",
|
|
183
|
+
"<|end|>"
|
|
184
|
+
]
|
|
342
185
|
},
|
|
343
186
|
{
|
|
344
187
|
"version": 1,
|
|
@@ -363,7 +206,7 @@
|
|
|
363
206
|
"none"
|
|
364
207
|
],
|
|
365
208
|
"model_id": "THUDM/glm-4-9b-chat",
|
|
366
|
-
"model_revision": "
|
|
209
|
+
"model_revision": "f6e0743b285dd808084530f070ad08e504386750"
|
|
367
210
|
},
|
|
368
211
|
{
|
|
369
212
|
"model_format": "ggufv2",
|
|
@@ -392,24 +235,17 @@
|
|
|
392
235
|
"model_revision": "0155a14edf0176863e9a003cdd78ce599e4d62c0"
|
|
393
236
|
}
|
|
394
237
|
],
|
|
395
|
-
"
|
|
396
|
-
|
|
397
|
-
|
|
398
|
-
|
|
399
|
-
|
|
400
|
-
|
|
401
|
-
|
|
402
|
-
"
|
|
403
|
-
|
|
404
|
-
|
|
405
|
-
|
|
406
|
-
],
|
|
407
|
-
"stop": [
|
|
408
|
-
"<|endoftext|>",
|
|
409
|
-
"<|user|>",
|
|
410
|
-
"<|observation|>"
|
|
411
|
-
]
|
|
412
|
-
}
|
|
238
|
+
"chat_template": "[gMASK]<sop>{% for item in messages %}{% if item['tools'] is defined %}<|system|>\n你是一个名为 ChatGLM 的人工智能助手。你是基于智谱AI训练的语言模型 GLM-4 模型开发的,你的任务是针对用户的问题和要求提供适当的答复和支持。\n\n# 可用工具{% set tools = item['tools'] %}{% for tool in tools %}{% if tool['type'] == 'function' %}\n\n## {{ tool['function']['name'] }}\n\n{{ tool['function'] | tojson(indent=4) }}\n在调用上述函数时,请使用 Json 格式表示调用的参数。{% elif tool['type'] == 'python' %}\n\n## python\n\n当你向 `python` 发送包含 Python 代码的消息时,该代码将会在一个有状态的 Jupyter notebook 环境中执行。\n`python` 返回代码执行的输出,或在执行 60 秒后返回超时。\n`/mnt/data` 将会持久化存储你的文件。在此会话中,`python` 无法访问互联网。不要使用 `python` 进行任何网络请求或者在线 API 调用,这些在线内容的访问将不会成功。{% elif tool['type'] == 'simple_browser' %}\n\n## simple_browser\n\n你可以使用 `simple_browser` 工具。该工具支持以下函数:\n`search(query: str, recency_days: int)`:使用搜索引擎进行查询并显示结果,可以使用 `recency_days` 参数控制搜索内容的时效性。\n`mclick(ids: list[int])`:获取一系列指定 id 的页面内容。每次调用时,须选择3-10个页面。选择多个角度的页面,同时尽可能选择可信任的信息来源。考虑到部分页面是无法加载的,你也可以多打开一些可能有用的页面而不用担心内容过多。\n`open_url(url: str)`:打开指定的 URL。\n\n使用 `【{引用 id}†{引用文本}】` 来引用内容。\n\n操作步骤:1. 使用 `search` 来获得信息列表; 2. 使用 `mclick` 来获取指定 ID 页面的内容; 3. 根据获得的内容进行回复。在回复中应当引用信息来源。\n 如果用户提供了 URL,也可以用 `open_url` 直接打开页面。\n如果初次搜索结果没有找到合适的信息,也可以再次使用 `search` 进行搜索。{% elif tool['type'] == 'cogview' %}\n\n## cogview\n\n如果用户的请求中包含了对图像的描述,你可以使用 `cogview` 来生成图像并展示给用户。你需要向 `cogview` 发送图像描述,规则:\n- 发送给 `cogview` 的消息必须使用英语。用户的中文描述必须完全翻译为英语。\n- 应当尽可能详细地描述图像生成的需求,需求描述约 100 英文单词。\n- 保持用户原始描述的意图。不要虚构内容或者没见过的人物。\n- 如无特殊说明,所在地为中国,持有中国立场并遵循中国社会主义价值观。{% endif %}{% endfor %}{% endif %}{% if item['content'] %}<|{{ item['role'] }}|>{{ item['metadata'] }}\n{{ item['content'] }}{% endif %}{% endfor %}{% if add_generation_prompt %}<|assistant|>{% endif %}",
|
|
239
|
+
"stop_token_ids": [
|
|
240
|
+
151329,
|
|
241
|
+
151336,
|
|
242
|
+
151338
|
|
243
|
+
],
|
|
244
|
+
"stop": [
|
|
245
|
+
"<|endoftext|>",
|
|
246
|
+
"<|user|>",
|
|
247
|
+
"<|observation|>"
|
|
248
|
+
]
|
|
413
249
|
},
|
|
414
250
|
{
|
|
415
251
|
"version": 1,
|
|
@@ -463,24 +299,17 @@
|
|
|
463
299
|
"model_revision": "782e28bd5eee3c514c07108da15e0b5e06dcf776"
|
|
464
300
|
}
|
|
465
301
|
],
|
|
466
|
-
"
|
|
467
|
-
|
|
468
|
-
|
|
469
|
-
|
|
470
|
-
|
|
471
|
-
|
|
472
|
-
|
|
473
|
-
"
|
|
474
|
-
|
|
475
|
-
|
|
476
|
-
|
|
477
|
-
],
|
|
478
|
-
"stop": [
|
|
479
|
-
"<|endoftext|>",
|
|
480
|
-
"<|user|>",
|
|
481
|
-
"<|observation|>"
|
|
482
|
-
]
|
|
483
|
-
}
|
|
302
|
+
"chat_template": "[gMASK]<sop>{% for item in messages %}{% if item['tools'] is defined %}<|system|>\n你是一个名为 GLM-4 的人工智能助手。你是基于智谱AI训练的语言模型 GLM-4 模型开发的,你的任务是针对用户的问题和要求提供适当的答复和支持。\n\n# 可用工具{% set tools = item['tools'] %}{% for tool in tools %}{% if tool['type'] == 'function' %}\n\n## {{ tool['function']['name'] }}\n\n{{ tool['function'] | tojson(indent=4) }}\n在调用上述函数时,请使用 Json 格式表示调用的参数。{% elif tool['type'] == 'python' %}\n\n## python\n\n当你向 `python` 发送包含 Python 代码的消息时,该代码将会在一个有状态的 Jupyter notebook 环境中执行。\n`python` 返回代码执行的输出,或在执行 60 秒后返回超时。\n`/mnt/data` 将会持久化存储你的文件。在此会话中,`python` 无法访问互联网。不要使用 `python` 进行任何网络请求或者在线 API 调用,这些在线内容的访问将不会成功。{% elif tool['type'] == 'simple_browser' %}\n\n## simple_browser\n\n你可以使用 `simple_browser` 工具。该工具支持以下函数:\n`search(query: str, recency_days: int)`:使用搜索引擎进行查询并显示结果,可以使用 `recency_days` 参数控制搜索内容的时效性。\n`mclick(ids: list[int])`:获取一系列指定 id 的页面内容。每次调用时,须选择3-10个页面。选择多个角度的页面,同时尽可能选择可信任的信息来源。考虑到部分页面是无法加载的,你也可以多打开一些可能有用的页面而不用担心内容过多。\n`open_url(url: str)`:打开指定的 URL。\n\n使用 `【{引用 id}†{引用文本}】` 来引用内容。\n\n操作步骤:1. 使用 `search` 来获得信息列表; 2. 使用 `mclick` 来获取指定 ID 页面的内容; 3. 根据获得的内容进行回复。在回复中应当引用信息来源。\n 如果用户提供了 URL,也可以用 `open_url` 直接打开页面。\n如果初次搜索结果没有找到合适的信息,也可以再次使用 `search` 进行搜索。{% elif tool['type'] == 'cogview' %}\n\n## cogview\n\n如果用户的请求中包含了对图像的描述,你可以使用 `cogview` 来生成图像并展示给用户。你需要向 `cogview` 发送图像描述,规则:\n- 发送给 `cogview` 的消息必须使用英语。用户的中文描述必须完全翻译为英语。\n- 应当尽可能详细地描述图像生成的需求,需求描述约 100 英文单词。\n- 保持用户原始描述的意图。不要虚构内容或者没见过的人物。\n- 如无特殊说明,所在地为中国,持有中国立场并遵循中国社会主义价值观。{% endif %}{% endfor %}{% endif %}{% if item['content'] %}<|{{ item['role'] }}|>{{ item['metadata'] }}\n{{ item['content'] }}{% endif %}{% endfor %}{% if add_generation_prompt %}<|assistant|>{% endif %}",
|
|
303
|
+
"stop_token_ids": [
|
|
304
|
+
151329,
|
|
305
|
+
151336,
|
|
306
|
+
151338
|
|
307
|
+
],
|
|
308
|
+
"stop": [
|
|
309
|
+
"<|endoftext|>",
|
|
310
|
+
"<|user|>",
|
|
311
|
+
"<|observation|>"
|
|
312
|
+
]
|
|
484
313
|
},
|
|
485
314
|
{
|
|
486
315
|
"version": 1,
|
|
@@ -505,27 +334,20 @@
|
|
|
505
334
|
"none"
|
|
506
335
|
],
|
|
507
336
|
"model_id": "THUDM/glm-4v-9b",
|
|
508
|
-
"model_revision": "
|
|
337
|
+
"model_revision": "01328faefe122fe605c1c127b62e6031d3ffebf7"
|
|
509
338
|
}
|
|
510
339
|
],
|
|
511
|
-
"
|
|
512
|
-
|
|
513
|
-
|
|
514
|
-
|
|
515
|
-
|
|
516
|
-
|
|
517
|
-
|
|
518
|
-
"
|
|
519
|
-
|
|
520
|
-
|
|
521
|
-
|
|
522
|
-
],
|
|
523
|
-
"stop": [
|
|
524
|
-
"<|endoftext|>",
|
|
525
|
-
"<|user|>",
|
|
526
|
-
"<|observation|>"
|
|
527
|
-
]
|
|
528
|
-
}
|
|
340
|
+
"chat_template": "",
|
|
341
|
+
"stop_token_ids": [
|
|
342
|
+
151329,
|
|
343
|
+
151336,
|
|
344
|
+
151338
|
|
345
|
+
],
|
|
346
|
+
"stop": [
|
|
347
|
+
"<|endoftext|>",
|
|
348
|
+
"<|user|>",
|
|
349
|
+
"<|observation|>"
|
|
350
|
+
]
|
|
529
351
|
},
|
|
530
352
|
{
|
|
531
353
|
"version": 1,
|
|
@@ -567,24 +389,17 @@
|
|
|
567
389
|
"model_revision": "6a04071c54c943949826d4815ee00717ed8cf153"
|
|
568
390
|
}
|
|
569
391
|
],
|
|
570
|
-
"
|
|
571
|
-
|
|
572
|
-
|
|
573
|
-
|
|
574
|
-
|
|
575
|
-
|
|
576
|
-
|
|
577
|
-
"
|
|
578
|
-
|
|
579
|
-
|
|
580
|
-
|
|
581
|
-
],
|
|
582
|
-
"stop": [
|
|
583
|
-
"<|endoftext|>",
|
|
584
|
-
"<|user|>",
|
|
585
|
-
"<|observation|>"
|
|
586
|
-
]
|
|
587
|
-
}
|
|
392
|
+
"chat_template": "{% for item in messages %}{% if loop.first and item['role'] == 'system' %}{{ '<|system|>\n' + item['content'] }}{% elif loop.first %}{{ '<|system|>\n你是一位智能编程助手,你叫CodeGeeX。你会为用户回答关于编程、代码、计算机方面的任何问题,并提供格式规范、可以执行、准确安全的代码,并在必要时提供详细的解释。' }}{% endif %}{% if item['role'] == 'user' %}{{ '<|user|>\n' + item['content'] }}{% elif item['role'] == 'assistant' %}{{ '<|assistant|>\n' + item['content'] }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% endif %}",
|
|
393
|
+
"stop_token_ids": [
|
|
394
|
+
151329,
|
|
395
|
+
151336,
|
|
396
|
+
151338
|
|
397
|
+
],
|
|
398
|
+
"stop": [
|
|
399
|
+
"<|endoftext|>",
|
|
400
|
+
"<|user|>",
|
|
401
|
+
"<|observation|>"
|
|
402
|
+
]
|
|
588
403
|
},
|
|
589
404
|
{
|
|
590
405
|
"version": 1,
|
|
@@ -622,14 +437,13 @@
|
|
|
622
437
|
"model_revision": "1e4944aaa1d8c8d0cdca28bb8e3a003303d0781b"
|
|
623
438
|
}
|
|
624
439
|
],
|
|
625
|
-
"
|
|
626
|
-
|
|
627
|
-
|
|
628
|
-
|
|
629
|
-
|
|
630
|
-
|
|
631
|
-
|
|
632
|
-
}
|
|
440
|
+
"chat_template": "{% for item in messages %}{% if loop.first and item['role'] == 'system' %}{{ '<|system|> \n' + item['content'] }}{% endif %}{% if item['role'] == 'user' %}{{ '<|user|> \n' + item['content'] }}{% elif item['role'] == 'assistant' %}{{ '<|assistant|> \n' + item['content'] }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>' }}{% endif %}",
|
|
441
|
+
"stop_token_ids": [
|
|
442
|
+
3
|
|
443
|
+
],
|
|
444
|
+
"stop": [
|
|
445
|
+
"<|endoftext|>"
|
|
446
|
+
]
|
|
633
447
|
},
|
|
634
448
|
{
|
|
635
449
|
"version": 1,
|
|
@@ -842,22 +656,11 @@
|
|
|
842
656
|
"model_revision": "36d9a7388cc80e5f4b3e9701ca2f250d21a96c30"
|
|
843
657
|
}
|
|
844
658
|
],
|
|
845
|
-
"
|
|
846
|
-
|
|
847
|
-
"system_prompt": "<s>[INST] <<SYS>>\nYou are a helpful AI assistant.\n<</SYS>>\n\n",
|
|
848
|
-
"roles": [
|
|
849
|
-
"[INST]",
|
|
850
|
-
"[/INST]"
|
|
851
|
-
],
|
|
852
|
-
"intra_message_sep": " ",
|
|
853
|
-
"inter_message_sep": " </s><s>",
|
|
854
|
-
"stop_token_ids": [
|
|
659
|
+
"chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = '<<SYS>>\n' + messages[0]['content'] | trim + '\n<</SYS>>\n\n' %}{% set messages = messages[1:] %}{% else %}{% set system_message = '' %}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if loop.index0 == 0 %}{% set content = system_message + message['content'] %}{% else %}{% set content = message['content'] %}{% endif %}{% if message['role'] == 'user' %}{{ '<s>' + '[INST] ' + content | trim + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + content | trim + ' ' + '</s>' }}{% endif %}{% endfor %}",
|
|
660
|
+
"stop_token_ids": [
|
|
855
661
|
2
|
|
856
|
-
|
|
857
|
-
|
|
858
|
-
"</s>"
|
|
859
|
-
]
|
|
860
|
-
}
|
|
662
|
+
],
|
|
663
|
+
"stop": []
|
|
861
664
|
},
|
|
862
665
|
{
|
|
863
666
|
"version": 1,
|
|
@@ -1210,24 +1013,15 @@
|
|
|
1210
1013
|
"model_id": "TechxGenus/Meta-Llama-3-70B-Instruct-GPTQ"
|
|
1211
1014
|
}
|
|
1212
1015
|
],
|
|
1213
|
-
"
|
|
1214
|
-
|
|
1215
|
-
|
|
1216
|
-
|
|
1217
|
-
|
|
1218
|
-
|
|
1219
|
-
|
|
1220
|
-
"
|
|
1221
|
-
|
|
1222
|
-
"stop_token_ids": [
|
|
1223
|
-
128001,
|
|
1224
|
-
128009
|
|
1225
|
-
],
|
|
1226
|
-
"stop": [
|
|
1227
|
-
"<|end_of_text|>",
|
|
1228
|
-
"<|eot_id|>"
|
|
1229
|
-
]
|
|
1230
|
-
}
|
|
1016
|
+
"chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = '<|begin_of_text|>' + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
|
|
1017
|
+
"stop_token_ids": [
|
|
1018
|
+
128001,
|
|
1019
|
+
128009
|
|
1020
|
+
],
|
|
1021
|
+
"stop": [
|
|
1022
|
+
"<|end_of_text|>",
|
|
1023
|
+
"<|eot_id|>"
|
|
1024
|
+
]
|
|
1231
1025
|
},
|
|
1232
1026
|
{
|
|
1233
1027
|
"version": 1,
|
|
@@ -1505,24 +1299,15 @@
|
|
|
1505
1299
|
"model_id": "hugging-quants/Meta-Llama-3.1-405B-Instruct-AWQ-INT4"
|
|
1506
1300
|
}
|
|
1507
1301
|
],
|
|
1508
|
-
"
|
|
1509
|
-
|
|
1510
|
-
|
|
1511
|
-
|
|
1512
|
-
|
|
1513
|
-
|
|
1514
|
-
|
|
1515
|
-
"
|
|
1516
|
-
|
|
1517
|
-
"stop_token_ids": [
|
|
1518
|
-
128001,
|
|
1519
|
-
128009
|
|
1520
|
-
],
|
|
1521
|
-
"stop": [
|
|
1522
|
-
"<|end_of_text|>",
|
|
1523
|
-
"<|eot_id|>"
|
|
1524
|
-
]
|
|
1525
|
-
}
|
|
1302
|
+
"chat_template": "{{- '<|begin_of_text|>' }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\n\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\n\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\n\" }}\n{{- \"Today Date: \" + date_string + \"\n\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\n\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\n\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\n\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\n\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\n\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\n\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\n\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' }}\n{%- endif %}\n",
|
|
1303
|
+
"stop_token_ids": [
|
|
1304
|
+
128001,
|
|
1305
|
+
128009
|
|
1306
|
+
],
|
|
1307
|
+
"stop": [
|
|
1308
|
+
"<|end_of_text|>",
|
|
1309
|
+
"<|eot_id|>"
|
|
1310
|
+
]
|
|
1526
1311
|
},
|
|
1527
1312
|
{
|
|
1528
1313
|
"version": 1,
|
|
@@ -1558,8 +1343,7 @@
|
|
|
1558
1343
|
"zh"
|
|
1559
1344
|
],
|
|
1560
1345
|
"model_ability": [
|
|
1561
|
-
"chat"
|
|
1562
|
-
"tools"
|
|
1346
|
+
"chat"
|
|
1563
1347
|
],
|
|
1564
1348
|
"model_description": "Qwen-chat is a fine-tuned version of the Qwen LLM trained with alignment techniques, specializing in chatting.",
|
|
1565
1349
|
"model_specs": [
|
|
@@ -1662,25 +1446,17 @@
|
|
|
1662
1446
|
"model_id": "Qwen/Qwen-72B-Chat-{quantization}"
|
|
1663
1447
|
}
|
|
1664
1448
|
],
|
|
1665
|
-
"
|
|
1666
|
-
|
|
1667
|
-
|
|
1668
|
-
|
|
1669
|
-
|
|
1670
|
-
|
|
1671
|
-
|
|
1672
|
-
"
|
|
1673
|
-
"
|
|
1674
|
-
|
|
1675
|
-
|
|
1676
|
-
151645
|
|
1677
|
-
],
|
|
1678
|
-
"stop": [
|
|
1679
|
-
"<|endoftext|>",
|
|
1680
|
-
"<|im_start|>",
|
|
1681
|
-
"<|im_end|>"
|
|
1682
|
-
]
|
|
1683
|
-
}
|
|
1449
|
+
"chat_template": "{% for item in messages %}{% if loop.first and item['role'] == 'system' %}{{ '<|im_start|>system\n' + item['content'] + '<|im_end|>\n' }}{% elif loop.first %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{% if item['role'] == 'user' %}{{ '<|im_start|>user\n' + item['content'] + '<|im_end|>' }}{% elif item['role'] == 'assistant' %}{{ '<|im_start|>assistant\n' + item['content'] + '<|im_end|>' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
|
1450
|
+
"stop_token_ids": [
|
|
1451
|
+
151643,
|
|
1452
|
+
151644,
|
|
1453
|
+
151645
|
|
1454
|
+
],
|
|
1455
|
+
"stop": [
|
|
1456
|
+
"<|endoftext|>",
|
|
1457
|
+
"<|im_start|>",
|
|
1458
|
+
"<|im_end|>"
|
|
1459
|
+
]
|
|
1684
1460
|
},
|
|
1685
1461
|
{
|
|
1686
1462
|
"version": 1,
|
|
@@ -2025,25 +1801,17 @@
|
|
|
2025
1801
|
}
|
|
2026
1802
|
}
|
|
2027
1803
|
],
|
|
2028
|
-
"
|
|
2029
|
-
|
|
2030
|
-
|
|
2031
|
-
|
|
2032
|
-
|
|
2033
|
-
|
|
2034
|
-
|
|
2035
|
-
"
|
|
2036
|
-
"
|
|
2037
|
-
|
|
2038
|
-
|
|
2039
|
-
151645
|
|
2040
|
-
],
|
|
2041
|
-
"stop": [
|
|
2042
|
-
"<|endoftext|>",
|
|
2043
|
-
"<|im_start|>",
|
|
2044
|
-
"<|im_end|>"
|
|
2045
|
-
]
|
|
2046
|
-
}
|
|
1804
|
+
"chat_template": "{%- macro json_to_python_type(json_spec) %}\n {%- set basic_type_map = {\n \"string\": \"str\",\n \"number\": \"float\",\n \"integer\": \"int\",\n \"boolean\": \"bool\"\n} %}\n {%- if basic_type_map[json_spec.type] is defined %}\n {{- basic_type_map[json_spec.type] }}\n {%- elif json_spec.type == \"array\" %}\n {{- \"list[\" + json_to_python_type(json_spec|items) + \"]\" }}\n {%- elif json_spec.type == \"object\" %}\n {%- if json_spec.additionalProperties is defined %}\n {{- \"dict[str, \" + json_to_python_type(json_spec.additionalProperties) + ']' }}\n {%- else %}\n {{- \"dict\" }}\n {%- endif %}\n {%- elif json_spec.type is iterable %}\n {{- \"Union[\" }}\n {%- for t in json_spec.type %}\n {{- json_to_python_type({\"type\": t}) }}\n {%- if not loop.last %}\n {{- \",\" }}\n {%- endif %}\n {%- endfor %}\n {{- \"]\" }}\n {%- else %}\n {{- \"Any\" }}\n {%- endif %}\n{%- endmacro %}\n\n{%- if tools %}\n {{- '<|im_start|>system\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] + '\n\n' }}\n {%- endif %}\n {{- '# Tools\n\n' }}\n {{- \"You are a function calling AI model. You are provided with function signatures within <tools></tools> XML tags. You may call one or more functions to assist with the user query. Don't make assumptions about what values to plug into functions. Here are the available tools: <tools> \" }}\n {%- for tool in tools %}\n {%- if tool.function is defined %}\n {%- set tool = tool.function %}\n {%- endif %}\n {{- '{\"type\": \"function\", \"function\": ' }}\n {{- '{\"name\": ' + tool.name + '\", ' }}\n {{- '\"description\": \"' + tool.name + '(' }}\n {%- for param_name, param_fields in tool.parameters.properties|items %}\n {{- param_name + \": \" + json_to_python_type(param_fields) }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- if tool.return is defined %}\n {{- \" -> \" + json_to_python_type(tool.return) }}\n {%- endif %}\n {{- \" - \" + tool.description + \"\n\n\" }}\n {%- for param_name, param_fields in tool.parameters.properties|items %}\n {%- if loop.first %}\n {{- \" Args:\n\" }}\n {%- endif %}\n {{- \" \" + param_name + \"(\" + json_to_python_type(param_fields) + \"): \" + param_fields.description|trim }}\n {%- endfor %}\n {%- if tool.return is defined and tool.return.description is defined %}\n {{- \"\n Returns:\n \" + tool.return.description }}\n {%- endif %}\n {{- '\"' }}\n {{- ', \"parameters\": ' }}\n {%- if tool.parameters.properties | length == 0 %}\n {{- \"{}\" }}\n {%- else %}\n {{- tool.parameters|tojson }}\n {%- endif %}\n {{- \"}\" }}\n {%- if not loop.last %}\n {{- \"\n\" }}\n {%- endif %}\n {%- endfor %}\n {{- \" </tools>\" }}\n {{- 'Use the following pydantic model json schema for each tool call you will make: {\"properties\": {\"arguments\": {\"title\": \"Arguments\", \"type\": \"object\"}, \"name\": {\"title\": \"Name\", \"type\": \"string\"}}, \"required\": [\"arguments\", \"name\"], \"title\": \"FunctionCall\", \"type\": \"object\"}\n' }}\n {{- \"For each function call return a json object with function name and arguments within <tool_call></tool_call> XML tags as follows:\n\" }}\n {{- \"<tool_call>\n\" }}\n {{- '{\"name\": <function-name>, \"arguments\": <args-json-object>}\n' }}\n {{- '</tool_call><|im_end|>\n' }}\n{%- else %}\n {%- if messages[0]['role'] != 'system' %}\n {{- '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}\n {%- else %}\n {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if message.role == \"user\" or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and message.tool_calls is not defined) %}\n {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role + '\n<tool_call>\n' }}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '{' }}\n {{- '\"name\": \"' }}\n {{- tool_call.name }}\n {%- if tool_call.arguments is defined %}\n {{- ', ' }}\n {{- '\"arguments\": ' }}\n {{- tool_call.arguments|tojson }}\n {%- endif %}\n {{- '\"}' }}\n {{- '\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if not message.name is defined %}\n {{- raise_exception(\"Tool response dicts require a 'name' key indicating the name of the called function!\") }}\n {%- endif %}\n {{- '<|im_start|>user\n<tool_response>\n' }}\n {{- '{\"name\": \"' }}\n {{- message.name }}\n {{- '\", \"content\": ' }}\n {{- message.content|tojson + '}' }}\n {{- '\n</tool_response><|im_end|>\n' }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\n' }}\n{%- endif %}",
|
|
1805
|
+
"stop_token_ids": [
|
|
1806
|
+
151643,
|
|
1807
|
+
151644,
|
|
1808
|
+
151645
|
|
1809
|
+
],
|
|
1810
|
+
"stop": [
|
|
1811
|
+
"<|endoftext|>",
|
|
1812
|
+
"<|im_start|>",
|
|
1813
|
+
"<|im_end|>"
|
|
1814
|
+
]
|
|
2047
1815
|
},
|
|
2048
1816
|
{
|
|
2049
1817
|
"version": 1,
|
|
@@ -2078,25 +1846,17 @@
|
|
|
2078
1846
|
"model_id": "Qwen/Qwen1.5-MoE-A2.7B-Chat-GPTQ-Int4"
|
|
2079
1847
|
}
|
|
2080
1848
|
],
|
|
2081
|
-
"
|
|
2082
|
-
|
|
2083
|
-
|
|
2084
|
-
|
|
2085
|
-
|
|
2086
|
-
|
|
2087
|
-
|
|
2088
|
-
"
|
|
2089
|
-
"
|
|
2090
|
-
|
|
2091
|
-
|
|
2092
|
-
151645
|
|
2093
|
-
],
|
|
2094
|
-
"stop": [
|
|
2095
|
-
"<|endoftext|>",
|
|
2096
|
-
"<|im_start|>",
|
|
2097
|
-
"<|im_end|>"
|
|
2098
|
-
]
|
|
2099
|
-
}
|
|
1849
|
+
"chat_template": "{%- macro json_to_python_type(json_spec) %}\n {%- set basic_type_map = {\n \"string\": \"str\",\n \"number\": \"float\",\n \"integer\": \"int\",\n \"boolean\": \"bool\"\n} %}\n {%- if basic_type_map[json_spec.type] is defined %}\n {{- basic_type_map[json_spec.type] }}\n {%- elif json_spec.type == \"array\" %}\n {{- \"list[\" + json_to_python_type(json_spec|items) + \"]\" }}\n {%- elif json_spec.type == \"object\" %}\n {%- if json_spec.additionalProperties is defined %}\n {{- \"dict[str, \" + json_to_python_type(json_spec.additionalProperties) + ']' }}\n {%- else %}\n {{- \"dict\" }}\n {%- endif %}\n {%- elif json_spec.type is iterable %}\n {{- \"Union[\" }}\n {%- for t in json_spec.type %}\n {{- json_to_python_type({\"type\": t}) }}\n {%- if not loop.last %}\n {{- \",\" }}\n {%- endif %}\n {%- endfor %}\n {{- \"]\" }}\n {%- else %}\n {{- \"Any\" }}\n {%- endif %}\n{%- endmacro %}\n\n{%- if tools %}\n {{- '<|im_start|>system\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] + '\n\n' }}\n {%- endif %}\n {{- '# Tools\n\n' }}\n {{- \"You are a function calling AI model. You are provided with function signatures within <tools></tools> XML tags. You may call one or more functions to assist with the user query. Don't make assumptions about what values to plug into functions. Here are the available tools: <tools> \" }}\n {%- for tool in tools %}\n {%- if tool.function is defined %}\n {%- set tool = tool.function %}\n {%- endif %}\n {{- '{\"type\": \"function\", \"function\": ' }}\n {{- '{\"name\": ' + tool.name + '\", ' }}\n {{- '\"description\": \"' + tool.name + '(' }}\n {%- for param_name, param_fields in tool.parameters.properties|items %}\n {{- param_name + \": \" + json_to_python_type(param_fields) }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- if tool.return is defined %}\n {{- \" -> \" + json_to_python_type(tool.return) }}\n {%- endif %}\n {{- \" - \" + tool.description + \"\n\n\" }}\n {%- for param_name, param_fields in tool.parameters.properties|items %}\n {%- if loop.first %}\n {{- \" Args:\n\" }}\n {%- endif %}\n {{- \" \" + param_name + \"(\" + json_to_python_type(param_fields) + \"): \" + param_fields.description|trim }}\n {%- endfor %}\n {%- if tool.return is defined and tool.return.description is defined %}\n {{- \"\n Returns:\n \" + tool.return.description }}\n {%- endif %}\n {{- '\"' }}\n {{- ', \"parameters\": ' }}\n {%- if tool.parameters.properties | length == 0 %}\n {{- \"{}\" }}\n {%- else %}\n {{- tool.parameters|tojson }}\n {%- endif %}\n {{- \"}\" }}\n {%- if not loop.last %}\n {{- \"\n\" }}\n {%- endif %}\n {%- endfor %}\n {{- \" </tools>\" }}\n {{- 'Use the following pydantic model json schema for each tool call you will make: {\"properties\": {\"arguments\": {\"title\": \"Arguments\", \"type\": \"object\"}, \"name\": {\"title\": \"Name\", \"type\": \"string\"}}, \"required\": [\"arguments\", \"name\"], \"title\": \"FunctionCall\", \"type\": \"object\"}\n' }}\n {{- \"For each function call return a json object with function name and arguments within <tool_call></tool_call> XML tags as follows:\n\" }}\n {{- \"<tool_call>\n\" }}\n {{- '{\"name\": <function-name>, \"arguments\": <args-json-object>}\n' }}\n {{- '</tool_call><|im_end|>\n' }}\n{%- else %}\n {%- if messages[0]['role'] != 'system' %}\n {{- '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}\n {%- else %}\n {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if message.role == \"user\" or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and message.tool_calls is not defined) %}\n {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role + '\n<tool_call>\n' }}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '{' }}\n {{- '\"name\": \"' }}\n {{- tool_call.name }}\n {%- if tool_call.arguments is defined %}\n {{- ', ' }}\n {{- '\"arguments\": ' }}\n {{- tool_call.arguments|tojson }}\n {%- endif %}\n {{- '\"}' }}\n {{- '\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if not message.name is defined %}\n {{- raise_exception(\"Tool response dicts require a 'name' key indicating the name of the called function!\") }}\n {%- endif %}\n {{- '<|im_start|>user\n<tool_response>\n' }}\n {{- '{\"name\": \"' }}\n {{- message.name }}\n {{- '\", \"content\": ' }}\n {{- message.content|tojson + '}' }}\n {{- '\n</tool_response><|im_end|>\n' }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\n' }}\n{%- endif %}",
|
|
1850
|
+
"stop_token_ids": [
|
|
1851
|
+
151643,
|
|
1852
|
+
151644,
|
|
1853
|
+
151645
|
|
1854
|
+
],
|
|
1855
|
+
"stop": [
|
|
1856
|
+
"<|endoftext|>",
|
|
1857
|
+
"<|im_start|>",
|
|
1858
|
+
"<|im_end|>"
|
|
1859
|
+
]
|
|
2100
1860
|
},
|
|
2101
1861
|
{
|
|
2102
1862
|
"version": 1,
|
|
@@ -2171,25 +1931,17 @@
|
|
|
2171
1931
|
"model_id": "Qwen/CodeQwen1.5-7B-Chat-AWQ"
|
|
2172
1932
|
}
|
|
2173
1933
|
],
|
|
2174
|
-
"
|
|
2175
|
-
|
|
2176
|
-
|
|
2177
|
-
|
|
2178
|
-
|
|
2179
|
-
|
|
2180
|
-
|
|
2181
|
-
"
|
|
2182
|
-
"
|
|
2183
|
-
|
|
2184
|
-
|
|
2185
|
-
151645
|
|
2186
|
-
],
|
|
2187
|
-
"stop": [
|
|
2188
|
-
"<|endoftext|>",
|
|
2189
|
-
"<|im_start|>",
|
|
2190
|
-
"<|im_end|>"
|
|
2191
|
-
]
|
|
2192
|
-
}
|
|
1934
|
+
"chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
|
1935
|
+
"stop_token_ids": [
|
|
1936
|
+
151643,
|
|
1937
|
+
151644,
|
|
1938
|
+
151645
|
|
1939
|
+
],
|
|
1940
|
+
"stop": [
|
|
1941
|
+
"<|endoftext|>",
|
|
1942
|
+
"<|im_start|>",
|
|
1943
|
+
"<|im_end|>"
|
|
1944
|
+
]
|
|
2193
1945
|
},
|
|
2194
1946
|
{
|
|
2195
1947
|
"version": 1,
|
|
@@ -2479,25 +2231,17 @@
|
|
|
2479
2231
|
}
|
|
2480
2232
|
}
|
|
2481
2233
|
],
|
|
2482
|
-
"
|
|
2483
|
-
|
|
2484
|
-
|
|
2485
|
-
|
|
2486
|
-
|
|
2487
|
-
|
|
2488
|
-
|
|
2489
|
-
"
|
|
2490
|
-
"
|
|
2491
|
-
|
|
2492
|
-
|
|
2493
|
-
151645
|
|
2494
|
-
],
|
|
2495
|
-
"stop": [
|
|
2496
|
-
"<|endoftext|>",
|
|
2497
|
-
"<|im_start|>",
|
|
2498
|
-
"<|im_end|>"
|
|
2499
|
-
]
|
|
2500
|
-
}
|
|
2234
|
+
"chat_template": "{%- macro json_to_python_type(json_spec) %}\n {%- set basic_type_map = {\n \"string\": \"str\",\n \"number\": \"float\",\n \"integer\": \"int\",\n \"boolean\": \"bool\"\n} %}\n {%- if basic_type_map[json_spec.type] is defined %}\n {{- basic_type_map[json_spec.type] }}\n {%- elif json_spec.type == \"array\" %}\n {{- \"list[\" + json_to_python_type(json_spec|items) + \"]\" }}\n {%- elif json_spec.type == \"object\" %}\n {%- if json_spec.additionalProperties is defined %}\n {{- \"dict[str, \" + json_to_python_type(json_spec.additionalProperties) + ']' }}\n {%- else %}\n {{- \"dict\" }}\n {%- endif %}\n {%- elif json_spec.type is iterable %}\n {{- \"Union[\" }}\n {%- for t in json_spec.type %}\n {{- json_to_python_type({\"type\": t}) }}\n {%- if not loop.last %}\n {{- \",\" }}\n {%- endif %}\n {%- endfor %}\n {{- \"]\" }}\n {%- else %}\n {{- \"Any\" }}\n {%- endif %}\n{%- endmacro %}\n\n{%- if tools %}\n {{- '<|im_start|>system\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] + '\n\n' }}\n {%- endif %}\n {{- '# Tools\n\n' }}\n {{- \"You are a function calling AI model. You are provided with function signatures within <tools></tools> XML tags. You may call one or more functions to assist with the user query. Don't make assumptions about what values to plug into functions. Here are the available tools: <tools> \" }}\n {%- for tool in tools %}\n {%- if tool.function is defined %}\n {%- set tool = tool.function %}\n {%- endif %}\n {{- '{\"type\": \"function\", \"function\": ' }}\n {{- '{\"name\": ' + tool.name + '\", ' }}\n {{- '\"description\": \"' + tool.name + '(' }}\n {%- for param_name, param_fields in tool.parameters.properties|items %}\n {{- param_name + \": \" + json_to_python_type(param_fields) }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- if tool.return is defined %}\n {{- \" -> \" + json_to_python_type(tool.return) }}\n {%- endif %}\n {{- \" - \" + tool.description + \"\n\n\" }}\n {%- for param_name, param_fields in tool.parameters.properties|items %}\n {%- if loop.first %}\n {{- \" Args:\n\" }}\n {%- endif %}\n {{- \" \" + param_name + \"(\" + json_to_python_type(param_fields) + \"): \" + param_fields.description|trim }}\n {%- endfor %}\n {%- if tool.return is defined and tool.return.description is defined %}\n {{- \"\n Returns:\n \" + tool.return.description }}\n {%- endif %}\n {{- '\"' }}\n {{- ', \"parameters\": ' }}\n {%- if tool.parameters.properties | length == 0 %}\n {{- \"{}\" }}\n {%- else %}\n {{- tool.parameters|tojson }}\n {%- endif %}\n {{- \"}\" }}\n {%- if not loop.last %}\n {{- \"\n\" }}\n {%- endif %}\n {%- endfor %}\n {{- \" </tools>\" }}\n {{- 'Use the following pydantic model json schema for each tool call you will make: {\"properties\": {\"arguments\": {\"title\": \"Arguments\", \"type\": \"object\"}, \"name\": {\"title\": \"Name\", \"type\": \"string\"}}, \"required\": [\"arguments\", \"name\"], \"title\": \"FunctionCall\", \"type\": \"object\"}\n' }}\n {{- \"For each function call return a json object with function name and arguments within <tool_call></tool_call> XML tags as follows:\n\" }}\n {{- \"<tool_call>\n\" }}\n {{- '{\"name\": <function-name>, \"arguments\": <args-json-object>}\n' }}\n {{- '</tool_call><|im_end|>\n' }}\n{%- else %}\n {%- if messages[0]['role'] != 'system' %}\n {{- '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}\n {%- else %}\n {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if message.role == \"user\" or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and message.tool_calls is not defined) %}\n {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role + '\n<tool_call>\n' }}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '{' }}\n {{- '\"name\": \"' }}\n {{- tool_call.name }}\n {%- if tool_call.arguments is defined %}\n {{- ', ' }}\n {{- '\"arguments\": ' }}\n {{- tool_call.arguments|tojson }}\n {%- endif %}\n {{- '\"}' }}\n {{- '\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if not message.name is defined %}\n {{- raise_exception(\"Tool response dicts require a 'name' key indicating the name of the called function!\") }}\n {%- endif %}\n {{- '<|im_start|>user\n<tool_response>\n' }}\n {{- '{\"name\": \"' }}\n {{- message.name }}\n {{- '\", \"content\": ' }}\n {{- message.content|tojson + '}' }}\n {{- '\n</tool_response><|im_end|>\n' }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\n' }}\n{%- endif %}",
|
|
2235
|
+
"stop_token_ids": [
|
|
2236
|
+
151643,
|
|
2237
|
+
151644,
|
|
2238
|
+
151645
|
|
2239
|
+
],
|
|
2240
|
+
"stop": [
|
|
2241
|
+
"<|endoftext|>",
|
|
2242
|
+
"<|im_start|>",
|
|
2243
|
+
"<|im_end|>"
|
|
2244
|
+
]
|
|
2501
2245
|
},
|
|
2502
2246
|
{
|
|
2503
2247
|
"version": 1,
|
|
@@ -2560,25 +2304,17 @@
|
|
|
2560
2304
|
}
|
|
2561
2305
|
}
|
|
2562
2306
|
],
|
|
2563
|
-
"
|
|
2564
|
-
|
|
2565
|
-
|
|
2566
|
-
|
|
2567
|
-
|
|
2568
|
-
|
|
2569
|
-
|
|
2570
|
-
"
|
|
2571
|
-
"
|
|
2572
|
-
|
|
2573
|
-
|
|
2574
|
-
151645
|
|
2575
|
-
],
|
|
2576
|
-
"stop": [
|
|
2577
|
-
"<|endoftext|>",
|
|
2578
|
-
"<|im_start|>",
|
|
2579
|
-
"<|im_end|>"
|
|
2580
|
-
]
|
|
2581
|
-
}
|
|
2307
|
+
"chat_template": "{%- macro json_to_python_type(json_spec) %}\n {%- set basic_type_map = {\n \"string\": \"str\",\n \"number\": \"float\",\n \"integer\": \"int\",\n \"boolean\": \"bool\"\n} %}\n {%- if basic_type_map[json_spec.type] is defined %}\n {{- basic_type_map[json_spec.type] }}\n {%- elif json_spec.type == \"array\" %}\n {{- \"list[\" + json_to_python_type(json_spec|items) + \"]\" }}\n {%- elif json_spec.type == \"object\" %}\n {%- if json_spec.additionalProperties is defined %}\n {{- \"dict[str, \" + json_to_python_type(json_spec.additionalProperties) + ']' }}\n {%- else %}\n {{- \"dict\" }}\n {%- endif %}\n {%- elif json_spec.type is iterable %}\n {{- \"Union[\" }}\n {%- for t in json_spec.type %}\n {{- json_to_python_type({\"type\": t}) }}\n {%- if not loop.last %}\n {{- \",\" }}\n {%- endif %}\n {%- endfor %}\n {{- \"]\" }}\n {%- else %}\n {{- \"Any\" }}\n {%- endif %}\n{%- endmacro %}\n\n{%- if tools %}\n {{- '<|im_start|>system\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] + '\n\n' }}\n {%- endif %}\n {{- '# Tools\n\n' }}\n {{- \"You are a function calling AI model. You are provided with function signatures within <tools></tools> XML tags. You may call one or more functions to assist with the user query. Don't make assumptions about what values to plug into functions. Here are the available tools: <tools> \" }}\n {%- for tool in tools %}\n {%- if tool.function is defined %}\n {%- set tool = tool.function %}\n {%- endif %}\n {{- '{\"type\": \"function\", \"function\": ' }}\n {{- '{\"name\": ' + tool.name + '\", ' }}\n {{- '\"description\": \"' + tool.name + '(' }}\n {%- for param_name, param_fields in tool.parameters.properties|items %}\n {{- param_name + \": \" + json_to_python_type(param_fields) }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- if tool.return is defined %}\n {{- \" -> \" + json_to_python_type(tool.return) }}\n {%- endif %}\n {{- \" - \" + tool.description + \"\n\n\" }}\n {%- for param_name, param_fields in tool.parameters.properties|items %}\n {%- if loop.first %}\n {{- \" Args:\n\" }}\n {%- endif %}\n {{- \" \" + param_name + \"(\" + json_to_python_type(param_fields) + \"): \" + param_fields.description|trim }}\n {%- endfor %}\n {%- if tool.return is defined and tool.return.description is defined %}\n {{- \"\n Returns:\n \" + tool.return.description }}\n {%- endif %}\n {{- '\"' }}\n {{- ', \"parameters\": ' }}\n {%- if tool.parameters.properties | length == 0 %}\n {{- \"{}\" }}\n {%- else %}\n {{- tool.parameters|tojson }}\n {%- endif %}\n {{- \"}\" }}\n {%- if not loop.last %}\n {{- \"\n\" }}\n {%- endif %}\n {%- endfor %}\n {{- \" </tools>\" }}\n {{- 'Use the following pydantic model json schema for each tool call you will make: {\"properties\": {\"arguments\": {\"title\": \"Arguments\", \"type\": \"object\"}, \"name\": {\"title\": \"Name\", \"type\": \"string\"}}, \"required\": [\"arguments\", \"name\"], \"title\": \"FunctionCall\", \"type\": \"object\"}\n' }}\n {{- \"For each function call return a json object with function name and arguments within <tool_call></tool_call> XML tags as follows:\n\" }}\n {{- \"<tool_call>\n\" }}\n {{- '{\"name\": <function-name>, \"arguments\": <args-json-object>}\n' }}\n {{- '</tool_call><|im_end|>\n' }}\n{%- else %}\n {%- if messages[0]['role'] != 'system' %}\n {{- '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}\n {%- else %}\n {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if message.role == \"user\" or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and message.tool_calls is not defined) %}\n {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role + '\n<tool_call>\n' }}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '{' }}\n {{- '\"name\": \"' }}\n {{- tool_call.name }}\n {%- if tool_call.arguments is defined %}\n {{- ', ' }}\n {{- '\"arguments\": ' }}\n {{- tool_call.arguments|tojson }}\n {%- endif %}\n {{- '\"}' }}\n {{- '\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if not message.name is defined %}\n {{- raise_exception(\"Tool response dicts require a 'name' key indicating the name of the called function!\") }}\n {%- endif %}\n {{- '<|im_start|>user\n<tool_response>\n' }}\n {{- '{\"name\": \"' }}\n {{- message.name }}\n {{- '\", \"content\": ' }}\n {{- message.content|tojson + '}' }}\n {{- '\n</tool_response><|im_end|>\n' }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\n' }}\n{%- endif %}",
|
|
2308
|
+
"stop_token_ids": [
|
|
2309
|
+
151643,
|
|
2310
|
+
151644,
|
|
2311
|
+
151645
|
|
2312
|
+
],
|
|
2313
|
+
"stop": [
|
|
2314
|
+
"<|endoftext|>",
|
|
2315
|
+
"<|im_start|>",
|
|
2316
|
+
"<|im_end|>"
|
|
2317
|
+
]
|
|
2582
2318
|
},
|
|
2583
2319
|
{
|
|
2584
2320
|
"version": 1,
|
|
@@ -2623,19 +2359,8 @@
|
|
|
2623
2359
|
"8-bit",
|
|
2624
2360
|
"none"
|
|
2625
2361
|
],
|
|
2626
|
-
"model_id": "
|
|
2627
|
-
"model_revision": "
|
|
2628
|
-
},
|
|
2629
|
-
{
|
|
2630
|
-
"model_format": "pytorch",
|
|
2631
|
-
"model_size_in_billions": 13,
|
|
2632
|
-
"quantizations": [
|
|
2633
|
-
"4-bit",
|
|
2634
|
-
"8-bit",
|
|
2635
|
-
"none"
|
|
2636
|
-
],
|
|
2637
|
-
"model_id": "WizardLM/WizardMath-13B-V1.0",
|
|
2638
|
-
"model_revision": "ef95532e96e634c634992dab891a17032dc71c8d"
|
|
2362
|
+
"model_id": "WizardLMTeam/WizardMath-7B-V1.0",
|
|
2363
|
+
"model_revision": "825a586f260d6c583b8aa9ceab6cdfaa3d9a4ddc"
|
|
2639
2364
|
},
|
|
2640
2365
|
{
|
|
2641
2366
|
"model_format": "pytorch",
|
|
@@ -2645,19 +2370,17 @@
|
|
|
2645
2370
|
"8-bit",
|
|
2646
2371
|
"none"
|
|
2647
2372
|
],
|
|
2648
|
-
"model_id": "
|
|
2649
|
-
"model_revision": "
|
|
2373
|
+
"model_id": "WizardLMTeam/WizardMath-70B-V1.0",
|
|
2374
|
+
"model_revision": "4dd9f3fcd8c056561d67ec59ae011f7c146aebd2"
|
|
2650
2375
|
}
|
|
2651
2376
|
],
|
|
2652
|
-
"
|
|
2653
|
-
|
|
2654
|
-
|
|
2655
|
-
|
|
2656
|
-
|
|
2657
|
-
|
|
2658
|
-
|
|
2659
|
-
"intra_message_sep": "\n\n### "
|
|
2660
|
-
}
|
|
2377
|
+
"chat_template": "{% for item in messages %}{% if loop.first and item['role'] == 'system' %}{{ item['content'] + '\n\n### ' }}{% elif loop.first %}{{ 'Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n### ' }}{% endif %}{% if item['role'] == 'user' %}{{ 'Instruction: ' + item['content'] + '\n\n### ' }}{% elif item['role'] == 'assistant' %}{{ 'Response: ' + item['content'] + '\n\n### ' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'Response: Let\\'s think step by step.' }}{% endif %}",
|
|
2378
|
+
"stop_token_ids": [
|
|
2379
|
+
2
|
|
2380
|
+
],
|
|
2381
|
+
"stop": [
|
|
2382
|
+
"</s>"
|
|
2383
|
+
]
|
|
2661
2384
|
},
|
|
2662
2385
|
{
|
|
2663
2386
|
"version": 1,
|
|
@@ -2979,22 +2702,13 @@
|
|
|
2979
2702
|
"model_file_name_template": "codellama-34b-instruct.{quantization}.gguf"
|
|
2980
2703
|
}
|
|
2981
2704
|
],
|
|
2982
|
-
"
|
|
2983
|
-
|
|
2984
|
-
"system_prompt": "<s>[INST] <<SYS>>\nWrite code to solve the following coding problem that obeys the constraints and passes the example test cases. Please wrap your code answer using ```:\n<</SYS>>\n\n",
|
|
2985
|
-
"roles": [
|
|
2986
|
-
"[INST]",
|
|
2987
|
-
"[/INST]"
|
|
2988
|
-
],
|
|
2989
|
-
"intra_message_sep": " ",
|
|
2990
|
-
"inter_message_sep": " </s><s>",
|
|
2991
|
-
"stop_token_ids": [
|
|
2705
|
+
"chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = '<<SYS>>\n' + messages[0]['content'] | trim + '\n<</SYS>>\n\n' %}{% set messages = messages[1:] %}{% else %}{% set system_message = '' %}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if loop.index0 == 0 %}{% set content = system_message + message['content'] %}{% else %}{% set content = message['content'] %}{% endif %}{% if message['role'] == 'user' %}{{ '<s>' + '[INST] ' + content | trim + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + content | trim + ' ' + '</s>' }}{% endif %}{% endfor %}",
|
|
2706
|
+
"stop_token_ids": [
|
|
2992
2707
|
2
|
|
2993
|
-
|
|
2994
|
-
|
|
2995
|
-
|
|
2996
|
-
|
|
2997
|
-
}
|
|
2708
|
+
],
|
|
2709
|
+
"stop": [
|
|
2710
|
+
"</s>"
|
|
2711
|
+
]
|
|
2998
2712
|
},
|
|
2999
2713
|
{
|
|
3000
2714
|
"version": 1,
|
|
@@ -3032,20 +2746,12 @@
|
|
|
3032
2746
|
"model_revision": "a56c793eb7a721ab6c270f779024e0375e8afd4a"
|
|
3033
2747
|
}
|
|
3034
2748
|
],
|
|
3035
|
-
"
|
|
3036
|
-
|
|
3037
|
-
|
|
3038
|
-
|
|
3039
|
-
|
|
3040
|
-
|
|
3041
|
-
],
|
|
3042
|
-
"intra_message_sep": "",
|
|
3043
|
-
"inter_message_sep": "</s>",
|
|
3044
|
-
"stop_token_ids": [
|
|
3045
|
-
2,
|
|
3046
|
-
195
|
|
3047
|
-
]
|
|
3048
|
-
}
|
|
2749
|
+
"chat_template": "{{ (messages|selectattr('role', 'equalto', 'system')|list|last).content|trim if (messages|selectattr('role', 'equalto', 'system')|list) else '' }}\n\n{% for message in messages %}\n{% if message['role'] == 'user' %}\n<reserved_106>\n{{ message['content']|trim -}}\n{% if not loop.last %}\n\n\n{% endif %}\n{% elif message['role'] == 'assistant' %}\n<reserved_107>\n{{ message['content']|trim -}}\n{% if not loop.last %}\n\n\n{% endif %}\n{% endif %}\n{% endfor %}\n{% if add_generation_prompt and messages[-1]['role'] != 'assistant' %}\n<reserved_107>\n{% endif %}",
|
|
2750
|
+
"stop_token_ids": [
|
|
2751
|
+
2,
|
|
2752
|
+
195
|
|
2753
|
+
],
|
|
2754
|
+
"stop": []
|
|
3049
2755
|
},
|
|
3050
2756
|
{
|
|
3051
2757
|
"version": 1,
|
|
@@ -3189,22 +2895,13 @@
|
|
|
3189
2895
|
"model_file_name_template": "mistral-7b-instruct-v0.1.{quantization}.gguf"
|
|
3190
2896
|
}
|
|
3191
2897
|
],
|
|
3192
|
-
"
|
|
3193
|
-
|
|
3194
|
-
|
|
3195
|
-
|
|
3196
|
-
|
|
3197
|
-
|
|
3198
|
-
|
|
3199
|
-
"intra_message_sep": " ",
|
|
3200
|
-
"inter_message_sep": "<s>",
|
|
3201
|
-
"stop_token_ids": [
|
|
3202
|
-
2
|
|
3203
|
-
],
|
|
3204
|
-
"stop": [
|
|
3205
|
-
"</s>"
|
|
3206
|
-
]
|
|
3207
|
-
}
|
|
2898
|
+
"chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n\n{{- '<s>' }}\n{%- for message in loop_messages %}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}\n {{- raise_exception('After the optional system message, conversation roles must alternate user/assistant/user/assistant/...') }}\n {%- endif %}\n {%- if message['role'] == 'user' %}\n {%- if loop.first and system_message is defined %}\n {{- ' [INST] ' + system_message + '\n\n' + message['content'] + ' [/INST]' }}\n {%- else %}\n {{- ' [INST] ' + message['content'] + ' [/INST]' }}\n {%- endif %}\n {%- elif message['role'] == 'assistant' %}\n {{- ' ' + message['content'] + '</s>'}}\n {%- else %}\n {{- raise_exception('Only user and assistant roles are supported, with the exception of an initial optional system message!') }}\n {%- endif %}\n{%- endfor %}\n",
|
|
2899
|
+
"stop_token_ids": [
|
|
2900
|
+
2
|
|
2901
|
+
],
|
|
2902
|
+
"stop": [
|
|
2903
|
+
"</s>"
|
|
2904
|
+
]
|
|
3208
2905
|
},
|
|
3209
2906
|
{
|
|
3210
2907
|
"version": 1,
|
|
@@ -3266,22 +2963,13 @@
|
|
|
3266
2963
|
"model_file_name_template": "mistral-7b-instruct-v0.2.{quantization}.gguf"
|
|
3267
2964
|
}
|
|
3268
2965
|
],
|
|
3269
|
-
"
|
|
3270
|
-
|
|
3271
|
-
|
|
3272
|
-
|
|
3273
|
-
|
|
3274
|
-
|
|
3275
|
-
|
|
3276
|
-
"intra_message_sep": " ",
|
|
3277
|
-
"inter_message_sep": "<s>",
|
|
3278
|
-
"stop_token_ids": [
|
|
3279
|
-
2
|
|
3280
|
-
],
|
|
3281
|
-
"stop": [
|
|
3282
|
-
"</s>"
|
|
3283
|
-
]
|
|
3284
|
-
}
|
|
2966
|
+
"chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n\n{{- '<s>' }}\n{%- for message in loop_messages %}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}\n {{- raise_exception('After the optional system message, conversation roles must alternate user/assistant/user/assistant/...') }}\n {%- endif %}\n {%- if message['role'] == 'user' %}\n {%- if loop.first and system_message is defined %}\n {{- ' [INST] ' + system_message + '\n\n' + message['content'] + ' [/INST]' }}\n {%- else %}\n {{- ' [INST] ' + message['content'] + ' [/INST]' }}\n {%- endif %}\n {%- elif message['role'] == 'assistant' %}\n {{- ' ' + message['content'] + '</s>'}}\n {%- else %}\n {{- raise_exception('Only user and assistant roles are supported, with the exception of an initial optional system message!') }}\n {%- endif %}\n{%- endfor %}\n",
|
|
2967
|
+
"stop_token_ids": [
|
|
2968
|
+
2
|
|
2969
|
+
],
|
|
2970
|
+
"stop": [
|
|
2971
|
+
"</s>"
|
|
2972
|
+
]
|
|
3285
2973
|
},
|
|
3286
2974
|
{
|
|
3287
2975
|
"version": 1,
|
|
@@ -3342,22 +3030,13 @@
|
|
|
3342
3030
|
"model_file_name_template": "Mistral-7B-Instruct-v0.3.{quantization}.gguf"
|
|
3343
3031
|
}
|
|
3344
3032
|
],
|
|
3345
|
-
"
|
|
3346
|
-
|
|
3347
|
-
|
|
3348
|
-
|
|
3349
|
-
|
|
3350
|
-
|
|
3351
|
-
|
|
3352
|
-
"intra_message_sep": " ",
|
|
3353
|
-
"inter_message_sep": "<s>",
|
|
3354
|
-
"stop_token_ids": [
|
|
3355
|
-
2
|
|
3356
|
-
],
|
|
3357
|
-
"stop": [
|
|
3358
|
-
"</s>"
|
|
3359
|
-
]
|
|
3360
|
-
}
|
|
3033
|
+
"chat_template": "{%- if messages[0][\"role\"] == \"system\" %}\n {%- set system_message = messages[0][\"content\"] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n{%- set user_messages = loop_messages | selectattr(\"role\", \"equalto\", \"user\") | list %}\n\n{#- This block checks for alternating user/assistant messages, skipping tool calling messages #}\n{%- set ns = namespace() %}\n{%- set ns.index = 0 %}\n{%- for message in loop_messages %}\n {%- if not (message.role == \"tool\" or message.role == \"tool_results\" or (message.tool_calls is defined and message.tool_calls is not none)) %}\n {%- if (message[\"role\"] == \"user\") != (ns.index % 2 == 0) %}\n {{- raise_exception(\"After the optional system message, conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif %}\n {%- set ns.index = ns.index + 1 %}\n {%- endif %}\n{%- endfor %}\n\n{{- '<s>' }}\n{%- for message in loop_messages %}\n {%- if message[\"role\"] == \"user\" %}\n {%- if tools is not none and (message == user_messages[-1]) %}\n {{- \"[AVAILABLE_TOOLS] [\" }}\n {%- for tool in tools %}\n {%- set tool = tool.function %}\n {{- '{\"type\": \"function\", \"function\": {' }}\n {%- for key, val in tool.items() if key != \"return\" %}\n {%- if val is string %}\n {{- '\"' + key + '\": \"' + val + '\"' }}\n {%- else %}\n {{- '\"' + key + '\": ' + val|tojson }}\n {%- endif %}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \"}}\" }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- else %}\n {{- \"]\" }}\n {%- endif %}\n {%- endfor %}\n {{- \"[/AVAILABLE_TOOLS]\" }}\n {%- endif %}\n {%- if loop.last and system_message is defined %}\n {{- \"[INST] \" + system_message + \"\n\n\" + message[\"content\"] + \"[/INST]\" }}\n {%- else %}\n {{- \"[INST] \" + message[\"content\"] + \"[/INST]\" }}\n {%- endif %}\n {%- elif message.tool_calls is defined and message.tool_calls is not none %}\n {{- \"[TOOL_CALLS] [\" }}\n {%- for tool_call in message.tool_calls %}\n {%- set out = tool_call.function|tojson %}\n {{- out[:-1] }}\n {%- if not tool_call.id is defined or tool_call.id|length != 9 %}\n {{- raise_exception(\"Tool call IDs should be alphanumeric strings with length 9!\") }}\n {%- endif %}\n {{- ', \"id\": \"' + tool_call.id + '\"}' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- else %}\n {{- \"]\" + '</s>' }}\n {%- endif %}\n {%- endfor %}\n {%- elif message[\"role\"] == \"assistant\" %}\n {{- \" \" + message[\"content\"]|trim + '</s>'}}\n {%- elif message[\"role\"] == \"tool_results\" or message[\"role\"] == \"tool\" %}\n {%- if message.content is defined and message.content.content is defined %}\n {%- set content = message.content.content %}\n {%- else %}\n {%- set content = message.content %}\n {%- endif %}\n {{- '[TOOL_RESULTS] {\"content\": ' + content|string + \", \" }}\n {%- if not message.tool_call_id is defined or message.tool_call_id|length != 9 %}\n {{- raise_exception(\"Tool call IDs should be alphanumeric strings with length 9!\") }}\n {%- endif %}\n {{- '\"call_id\": \"' + message.tool_call_id + '\"}[/TOOL_RESULTS]' }}\n {%- else %}\n {{- raise_exception(\"Only user and assistant roles are supported, with the exception of an initial optional system message!\") }}\n {%- endif %}\n{%- endfor %}\n",
|
|
3034
|
+
"stop_token_ids": [
|
|
3035
|
+
2
|
|
3036
|
+
],
|
|
3037
|
+
"stop": [
|
|
3038
|
+
"</s>"
|
|
3039
|
+
]
|
|
3361
3040
|
},
|
|
3362
3041
|
{
|
|
3363
3042
|
"version": 1,
|
|
@@ -3466,22 +3145,13 @@
|
|
|
3466
3145
|
"model_id": "mlx-community/Mistral-Nemo-Instruct-2407-8bit"
|
|
3467
3146
|
}
|
|
3468
3147
|
],
|
|
3469
|
-
"
|
|
3470
|
-
|
|
3471
|
-
|
|
3472
|
-
|
|
3473
|
-
|
|
3474
|
-
|
|
3475
|
-
|
|
3476
|
-
"intra_message_sep": "",
|
|
3477
|
-
"inter_message_sep": "</s>",
|
|
3478
|
-
"stop_token_ids": [
|
|
3479
|
-
2
|
|
3480
|
-
],
|
|
3481
|
-
"stop": [
|
|
3482
|
-
"</s>"
|
|
3483
|
-
]
|
|
3484
|
-
}
|
|
3148
|
+
"chat_template": "{%- if messages[0][\"role\"] == \"system\" %}\n {%- set system_message = messages[0][\"content\"] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n{%- set user_messages = loop_messages | selectattr(\"role\", \"equalto\", \"user\") | list %}\n\n{#- This block checks for alternating user/assistant messages, skipping tool calling messages #}\n{%- set ns = namespace() %}\n{%- set ns.index = 0 %}\n{%- for message in loop_messages %}\n {%- if not (message.role == \"tool\" or message.role == \"tool_results\" or (message.tool_calls is defined and message.tool_calls is not none)) %}\n {%- if (message[\"role\"] == \"user\") != (ns.index % 2 == 0) %}\n {{- raise_exception(\"After the optional system message, conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif %}\n {%- set ns.index = ns.index + 1 %}\n {%- endif %}\n{%- endfor %}\n\n{{- '<s>' }}\n{%- for message in loop_messages %}\n {%- if message[\"role\"] == \"user\" %}\n {%- if tools is not none and (message == user_messages[-1]) %}\n {{- \"[AVAILABLE_TOOLS][\" }}\n {%- for tool in tools %}\n {%- set tool = tool.function %}\n {{- '{\"type\": \"function\", \"function\": {' }}\n {%- for key, val in tool.items() if key != \"return\" %}\n {%- if val is string %}\n {{- '\"' + key + '\": \"' + val + '\"' }}\n {%- else %}\n {{- '\"' + key + '\": ' + val|tojson }}\n {%- endif %}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \"}}\" }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- else %}\n {{- \"]\" }}\n {%- endif %}\n {%- endfor %}\n {{- \"[/AVAILABLE_TOOLS]\" }}\n {%- endif %}\n {%- if loop.last and system_message is defined %}\n {{- \"[INST]\" + system_message + \"\n\n\" + message[\"content\"] + \"[/INST]\" }}\n {%- else %}\n {{- \"[INST]\" + message[\"content\"] + \"[/INST]\" }}\n {%- endif %}\n {%- elif (message.tool_calls is defined and message.tool_calls is not none) %}\n {{- \"[TOOL_CALLS][\" }}\n {%- for tool_call in message.tool_calls %}\n {%- set out = tool_call.function|tojson %}\n {{- out[:-1] }}\n {%- if not tool_call.id is defined or tool_call.id|length != 9 %}\n {{- raise_exception(\"Tool call IDs should be alphanumeric strings with length 9!\") }}\n {%- endif %}\n {{- ', \"id\": \"' + tool_call.id + '\"}' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- else %}\n {{- \"]\" + '</s>' }}\n {%- endif %}\n {%- endfor %}\n {%- elif message[\"role\"] == \"assistant\" %}\n {{- message[\"content\"] + '</s>'}}\n {%- elif message[\"role\"] == \"tool_results\" or message[\"role\"] == \"tool\" %}\n {%- if message.content is defined and message.content.content is defined %}\n {%- set content = message.content.content %}\n {%- else %}\n {%- set content = message.content %}\n {%- endif %}\n {{- '[TOOL_RESULTS]{\"content\": ' + content|string + \", \" }}\n {%- if not message.tool_call_id is defined or message.tool_call_id|length != 9 %}\n {{- raise_exception(\"Tool call IDs should be alphanumeric strings with length 9!\") }}\n {%- endif %}\n {{- '\"call_id\": \"' + message.tool_call_id + '\"}[/TOOL_RESULTS]' }}\n {%- else %}\n {{- raise_exception(\"Only user and assistant roles are supported, with the exception of an initial optional system message!\") }}\n {%- endif %}\n{%- endfor %}\n",
|
|
3149
|
+
"stop_token_ids": [
|
|
3150
|
+
2
|
|
3151
|
+
],
|
|
3152
|
+
"stop": [
|
|
3153
|
+
"</s>"
|
|
3154
|
+
]
|
|
3485
3155
|
},
|
|
3486
3156
|
{
|
|
3487
3157
|
"version": 1,
|
|
@@ -3623,22 +3293,13 @@
|
|
|
3623
3293
|
"model_id": "mlx-community/Mistral-Large-Instruct-2407-8bit"
|
|
3624
3294
|
}
|
|
3625
3295
|
],
|
|
3626
|
-
"
|
|
3627
|
-
|
|
3628
|
-
|
|
3629
|
-
|
|
3630
|
-
|
|
3631
|
-
|
|
3632
|
-
|
|
3633
|
-
"intra_message_sep": "",
|
|
3634
|
-
"inter_message_sep": "</s>",
|
|
3635
|
-
"stop_token_ids": [
|
|
3636
|
-
2
|
|
3637
|
-
],
|
|
3638
|
-
"stop": [
|
|
3639
|
-
"</s>"
|
|
3640
|
-
]
|
|
3641
|
-
}
|
|
3296
|
+
"chat_template": "{%- if messages[0][\"role\"] == \"system\" %}\n {%- set system_message = messages[0][\"content\"] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n{%- set user_messages = loop_messages | selectattr(\"role\", \"equalto\", \"user\") | list %}\n\n{#- This block checks for alternating user/assistant messages, skipping tool calling messages #}\n{%- set ns = namespace() %}\n{%- set ns.index = 0 %}\n{%- for message in loop_messages %}\n {%- if not (message.role == \"tool\" or message.role == \"tool_results\" or (message.tool_calls is defined and message.tool_calls is not none)) %}\n {%- if (message[\"role\"] == \"user\") != (ns.index % 2 == 0) %}\n {{- raise_exception(\"After the optional system message, conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif %}\n {%- set ns.index = ns.index + 1 %}\n {%- endif %}\n{%- endfor %}\n\n{{- '<s>' }}\n{%- for message in loop_messages %}\n {%- if message[\"role\"] == \"user\" %}\n {%- if tools is not none and (message == user_messages[-1]) %}\n {{- \"[AVAILABLE_TOOLS][\" }}\n {%- for tool in tools %}\n {%- set tool = tool.function %}\n {{- '{\"type\": \"function\", \"function\": {' }}\n {%- for key, val in tool.items() if key != \"return\" %}\n {%- if val is string %}\n {{- '\"' + key + '\": \"' + val + '\"' }}\n {%- else %}\n {{- '\"' + key + '\": ' + val|tojson }}\n {%- endif %}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \"}}\" }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- else %}\n {{- \"]\" }}\n {%- endif %}\n {%- endfor %}\n {{- \"[/AVAILABLE_TOOLS]\" }}\n {%- endif %}\n {%- if loop.last and system_message is defined %}\n {{- \"[INST]\" + system_message + \"\n\n\" + message[\"content\"] + \"[/INST]\" }}\n {%- else %}\n {{- \"[INST]\" + message[\"content\"] + \"[/INST]\" }}\n {%- endif %}\n {%- elif (message.tool_calls is defined and message.tool_calls is not none) %}\n {{- \"[TOOL_CALLS][\" }}\n {%- for tool_call in message.tool_calls %}\n {%- set out = tool_call.function|tojson %}\n {{- out[:-1] }}\n {%- if not tool_call.id is defined or tool_call.id|length != 9 %}\n {{- raise_exception(\"Tool call IDs should be alphanumeric strings with length 9!\") }}\n {%- endif %}\n {{- ', \"id\": \"' + tool_call.id + '\"}' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- else %}\n {{- \"]\" + '</s>' }}\n {%- endif %}\n {%- endfor %}\n {%- elif message[\"role\"] == \"assistant\" %}\n {{- message[\"content\"] + '</s>'}}\n {%- elif message[\"role\"] == \"tool_results\" or message[\"role\"] == \"tool\" %}\n {%- if message.content is defined and message.content.content is defined %}\n {%- set content = message.content.content %}\n {%- else %}\n {%- set content = message.content %}\n {%- endif %}\n {{- '[TOOL_RESULTS]{\"content\": ' + content|string + \", \" }}\n {%- if not message.tool_call_id is defined or message.tool_call_id|length != 9 %}\n {{- raise_exception(\"Tool call IDs should be alphanumeric strings with length 9!\") }}\n {%- endif %}\n {{- '\"call_id\": \"' + message.tool_call_id + '\"}[/TOOL_RESULTS]' }}\n {%- else %}\n {{- raise_exception(\"Only user and assistant roles are supported, with the exception of an initial optional system message!\") }}\n {%- endif %}\n{%- endfor %}\n",
|
|
3297
|
+
"stop_token_ids": [
|
|
3298
|
+
2
|
|
3299
|
+
],
|
|
3300
|
+
"stop": [
|
|
3301
|
+
"</s>"
|
|
3302
|
+
]
|
|
3642
3303
|
},
|
|
3643
3304
|
{
|
|
3644
3305
|
"version": 1,
|
|
@@ -3745,22 +3406,13 @@
|
|
|
3745
3406
|
"model_file_name_template": "openhermes-2.5-mistral-7b.{quantization}.gguf"
|
|
3746
3407
|
}
|
|
3747
3408
|
],
|
|
3748
|
-
"
|
|
3749
|
-
|
|
3750
|
-
|
|
3751
|
-
|
|
3752
|
-
|
|
3753
|
-
|
|
3754
|
-
|
|
3755
|
-
"intra_message_sep": "<|im_end|>\n<|im_start|>",
|
|
3756
|
-
"inter_message_sep": "",
|
|
3757
|
-
"stop_token_ids": [
|
|
3758
|
-
32000
|
|
3759
|
-
],
|
|
3760
|
-
"stop": [
|
|
3761
|
-
"<|im_end|>"
|
|
3762
|
-
]
|
|
3763
|
-
}
|
|
3409
|
+
"chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
|
3410
|
+
"stop_token_ids": [
|
|
3411
|
+
32000
|
|
3412
|
+
],
|
|
3413
|
+
"stop": [
|
|
3414
|
+
"<|im_end|>"
|
|
3415
|
+
]
|
|
3764
3416
|
},
|
|
3765
3417
|
{
|
|
3766
3418
|
"version": 1,
|
|
@@ -3909,16 +3561,13 @@
|
|
|
3909
3561
|
"model_file_name_template": "mixtral-8x7b-instruct-v0.1.{quantization}.gguf"
|
|
3910
3562
|
}
|
|
3911
3563
|
],
|
|
3912
|
-
"
|
|
3913
|
-
|
|
3914
|
-
|
|
3915
|
-
|
|
3916
|
-
|
|
3917
|
-
|
|
3918
|
-
|
|
3919
|
-
"intra_message_sep": "",
|
|
3920
|
-
"inter_message_sep": ""
|
|
3921
|
-
}
|
|
3564
|
+
"chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n\n{{- '<s>' }}\n{%- for message in loop_messages %}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}\n {{- raise_exception('After the optional system message, conversation roles must alternate user/assistant/user/assistant/...') }}\n {%- endif %}\n {%- if message['role'] == 'user' %}\n {%- if loop.first and system_message is defined %}\n {{- ' [INST] ' + system_message + '\n\n' + message['content'] + ' [/INST]' }}\n {%- else %}\n {{- ' [INST] ' + message['content'] + ' [/INST]' }}\n {%- endif %}\n {%- elif message['role'] == 'assistant' %}\n {{- ' ' + message['content'] + '</s>'}}\n {%- else %}\n {{- raise_exception('Only user and assistant roles are supported, with the exception of an initial optional system message!') }}\n {%- endif %}\n{%- endfor %}\n",
|
|
3565
|
+
"stop_token_ids": [
|
|
3566
|
+
2
|
|
3567
|
+
],
|
|
3568
|
+
"stop": [
|
|
3569
|
+
"</s>"
|
|
3570
|
+
]
|
|
3922
3571
|
},
|
|
3923
3572
|
{
|
|
3924
3573
|
"version": 1,
|
|
@@ -4045,16 +3694,13 @@
|
|
|
4045
3694
|
}
|
|
4046
3695
|
}
|
|
4047
3696
|
],
|
|
4048
|
-
"
|
|
4049
|
-
|
|
4050
|
-
|
|
4051
|
-
|
|
4052
|
-
|
|
4053
|
-
|
|
4054
|
-
|
|
4055
|
-
"intra_message_sep": "",
|
|
4056
|
-
"inter_message_sep": ""
|
|
4057
|
-
}
|
|
3697
|
+
"chat_template": "{%- if messages[0][\"role\"] == \"system\" %}\n {%- set system_message = messages[0][\"content\"] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n{%- set user_messages = loop_messages | selectattr(\"role\", \"equalto\", \"user\") | list %}\n\n{#- This block checks for alternating user/assistant messages, skipping tool calling messages #}\n{%- set ns = namespace() %}\n{%- set ns.index = 0 %}\n{%- for message in loop_messages %}\n {%- if not (message.role == \"tool\" or message.role == \"tool_results\" or (message.tool_calls is defined and message.tool_calls is not none)) %}\n {%- if (message[\"role\"] == \"user\") != (ns.index % 2 == 0) %}\n {{- raise_exception(\"After the optional system message, conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif %}\n {%- set ns.index = ns.index + 1 %}\n {%- endif %}\n{%- endfor %}\n\n{{- '<s>' }}\n{%- for message in loop_messages %}\n {%- if message[\"role\"] == \"user\" %}\n {%- if tools is not none and (message == user_messages[-1]) %}\n {{- \"[AVAILABLE_TOOLS] [\" }}\n {%- for tool in tools %}\n {%- set tool = tool.function %}\n {{- '{\"type\": \"function\", \"function\": {' }}\n {%- for key, val in tool.items() if key != \"return\" %}\n {%- if val is string %}\n {{- '\"' + key + '\": \"' + val + '\"' }}\n {%- else %}\n {{- '\"' + key + '\": ' + val|tojson }}\n {%- endif %}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \"}}\" }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- else %}\n {{- \"]\" }}\n {%- endif %}\n {%- endfor %}\n {{- \"[/AVAILABLE_TOOLS]\" }}\n {%- endif %}\n {%- if loop.last and system_message is defined %}\n {{- \"[INST] \" + system_message + \"\n\n\" + message[\"content\"] + \"[/INST]\" }}\n {%- else %}\n {{- \"[INST] \" + message[\"content\"] + \"[/INST]\" }}\n {%- endif %}\n {%- elif message.tool_calls is defined and message.tool_calls is not none %}\n {{- \"[TOOL_CALLS] [\" }}\n {%- for tool_call in message.tool_calls %}\n {%- set out = tool_call.function|tojson %}\n {{- out[:-1] }}\n {%- if not tool_call.id is defined or tool_call.id|length != 9 %}\n {{- raise_exception(\"Tool call IDs should be alphanumeric strings with length 9!\") }}\n {%- endif %}\n {{- ', \"id\": \"' + tool_call.id + '\"}' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- else %}\n {{- \"]\" + '</s>' }}\n {%- endif %}\n {%- endfor %}\n {%- elif message[\"role\"] == \"assistant\" %}\n {{- \" \" + message[\"content\"]|trim + '</s>'}}\n {%- elif message[\"role\"] == \"tool_results\" or message[\"role\"] == \"tool\" %}\n {%- if message.content is defined and message.content.content is defined %}\n {%- set content = message.content.content %}\n {%- else %}\n {%- set content = message.content %}\n {%- endif %}\n {{- '[TOOL_RESULTS] {\"content\": ' + content|string + \", \" }}\n {%- if not message.tool_call_id is defined or message.tool_call_id|length != 9 %}\n {{- raise_exception(\"Tool call IDs should be alphanumeric strings with length 9!\") }}\n {%- endif %}\n {{- '\"call_id\": \"' + message.tool_call_id + '\"}[/TOOL_RESULTS]' }}\n {%- else %}\n {{- raise_exception(\"Only user and assistant roles are supported, with the exception of an initial optional system message!\") }}\n {%- endif %}\n{%- endfor %}\n",
|
|
3698
|
+
"stop_token_ids": [
|
|
3699
|
+
2
|
|
3700
|
+
],
|
|
3701
|
+
"stop": [
|
|
3702
|
+
"</s>"
|
|
3703
|
+
]
|
|
4058
3704
|
},
|
|
4059
3705
|
{
|
|
4060
3706
|
"version": 1,
|
|
@@ -4225,28 +3871,19 @@
|
|
|
4225
3871
|
"model_file_name_template": "yi-34b-chat.{quantization}.gguf"
|
|
4226
3872
|
}
|
|
4227
3873
|
],
|
|
4228
|
-
"
|
|
4229
|
-
|
|
4230
|
-
|
|
4231
|
-
|
|
4232
|
-
|
|
4233
|
-
|
|
4234
|
-
|
|
4235
|
-
|
|
4236
|
-
"
|
|
4237
|
-
"
|
|
4238
|
-
|
|
4239
|
-
|
|
4240
|
-
|
|
4241
|
-
8
|
|
4242
|
-
],
|
|
4243
|
-
"stop": [
|
|
4244
|
-
"<|endoftext|>",
|
|
4245
|
-
"<|im_start|>",
|
|
4246
|
-
"<|im_end|>",
|
|
4247
|
-
"<|im_sep|>"
|
|
4248
|
-
]
|
|
4249
|
-
}
|
|
3874
|
+
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
|
3875
|
+
"stop_token_ids": [
|
|
3876
|
+
2,
|
|
3877
|
+
6,
|
|
3878
|
+
7,
|
|
3879
|
+
8
|
|
3880
|
+
],
|
|
3881
|
+
"stop": [
|
|
3882
|
+
"<|endoftext|>",
|
|
3883
|
+
"<|im_start|>",
|
|
3884
|
+
"<|im_end|>",
|
|
3885
|
+
"<|im_sep|>"
|
|
3886
|
+
]
|
|
4250
3887
|
},
|
|
4251
3888
|
{
|
|
4252
3889
|
"version": 1,
|
|
@@ -4494,28 +4131,19 @@
|
|
|
4494
4131
|
"model_revision": "3c12761a2c6663f216caab6dff84b0dd29b472ac"
|
|
4495
4132
|
}
|
|
4496
4133
|
],
|
|
4497
|
-
"
|
|
4498
|
-
|
|
4499
|
-
|
|
4500
|
-
|
|
4501
|
-
|
|
4502
|
-
|
|
4503
|
-
|
|
4504
|
-
|
|
4505
|
-
"
|
|
4506
|
-
"
|
|
4507
|
-
|
|
4508
|
-
|
|
4509
|
-
|
|
4510
|
-
8
|
|
4511
|
-
],
|
|
4512
|
-
"stop": [
|
|
4513
|
-
"<|endoftext|>",
|
|
4514
|
-
"<|im_start|>",
|
|
4515
|
-
"<|im_end|>",
|
|
4516
|
-
"<|im_sep|>"
|
|
4517
|
-
]
|
|
4518
|
-
}
|
|
4134
|
+
"chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\n' + content + '<|im_end|>\n<|im_start|>assistant\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\n' }}{% endif %}{% endfor %}",
|
|
4135
|
+
"stop_token_ids": [
|
|
4136
|
+
2,
|
|
4137
|
+
6,
|
|
4138
|
+
7,
|
|
4139
|
+
8
|
|
4140
|
+
],
|
|
4141
|
+
"stop": [
|
|
4142
|
+
"<|endoftext|>",
|
|
4143
|
+
"<|im_start|>",
|
|
4144
|
+
"<|im_end|>",
|
|
4145
|
+
"<|im_sep|>"
|
|
4146
|
+
]
|
|
4519
4147
|
},
|
|
4520
4148
|
{
|
|
4521
4149
|
"version": 1,
|
|
@@ -4593,28 +4221,19 @@
|
|
|
4593
4221
|
"model_file_name_template": "Yi-1.5-34B-Chat-16K-{quantization}.gguf"
|
|
4594
4222
|
}
|
|
4595
4223
|
],
|
|
4596
|
-
"
|
|
4597
|
-
|
|
4598
|
-
|
|
4599
|
-
|
|
4600
|
-
|
|
4601
|
-
|
|
4602
|
-
|
|
4603
|
-
|
|
4604
|
-
"
|
|
4605
|
-
"
|
|
4606
|
-
|
|
4607
|
-
|
|
4608
|
-
|
|
4609
|
-
8
|
|
4610
|
-
],
|
|
4611
|
-
"stop": [
|
|
4612
|
-
"<|endoftext|>",
|
|
4613
|
-
"<|im_start|>",
|
|
4614
|
-
"<|im_end|>",
|
|
4615
|
-
"<|im_sep|>"
|
|
4616
|
-
]
|
|
4617
|
-
}
|
|
4224
|
+
"chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\n' + content + '<|im_end|>\n<|im_start|>assistant\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\n' }}{% endif %}{% endfor %}",
|
|
4225
|
+
"stop_token_ids": [
|
|
4226
|
+
2,
|
|
4227
|
+
6,
|
|
4228
|
+
7,
|
|
4229
|
+
8
|
|
4230
|
+
],
|
|
4231
|
+
"stop": [
|
|
4232
|
+
"<|endoftext|>",
|
|
4233
|
+
"<|im_start|>",
|
|
4234
|
+
"<|im_end|>",
|
|
4235
|
+
"<|im_sep|>"
|
|
4236
|
+
]
|
|
4618
4237
|
},
|
|
4619
4238
|
{
|
|
4620
4239
|
"version": 1,
|
|
@@ -4627,17 +4246,6 @@
|
|
|
4627
4246
|
"chat"
|
|
4628
4247
|
],
|
|
4629
4248
|
"model_specs": [
|
|
4630
|
-
{
|
|
4631
|
-
"model_format": "pytorch",
|
|
4632
|
-
"model_size_in_billions": 7,
|
|
4633
|
-
"quantizations": [
|
|
4634
|
-
"4-bit",
|
|
4635
|
-
"8-bit",
|
|
4636
|
-
"none"
|
|
4637
|
-
],
|
|
4638
|
-
"model_id": "WizardLM/WizardCoder-Python-7B-V1.0",
|
|
4639
|
-
"model_revision": "e40673a27a4aefcff2c6d2b3b1e0681a38703e4e"
|
|
4640
|
-
},
|
|
4641
4249
|
{
|
|
4642
4250
|
"model_format": "pytorch",
|
|
4643
4251
|
"model_size_in_billions": 13,
|
|
@@ -4646,8 +4254,8 @@
|
|
|
4646
4254
|
"8-bit",
|
|
4647
4255
|
"none"
|
|
4648
4256
|
],
|
|
4649
|
-
"model_id": "
|
|
4650
|
-
"model_revision": "
|
|
4257
|
+
"model_id": "WizardLMTeam/WizardCoder-Python-13B-V1.0",
|
|
4258
|
+
"model_revision": "5ac6748b1f5a4c282107ddc7d3b69fdc4a686d75"
|
|
4651
4259
|
},
|
|
4652
4260
|
{
|
|
4653
4261
|
"model_format": "pytorch",
|
|
@@ -4657,8 +4265,8 @@
|
|
|
4657
4265
|
"8-bit",
|
|
4658
4266
|
"none"
|
|
4659
4267
|
],
|
|
4660
|
-
"model_id": "
|
|
4661
|
-
"model_revision": "
|
|
4268
|
+
"model_id": "WizardLMTeam/WizardCoder-Python-34B-V1.0",
|
|
4269
|
+
"model_revision": "897fc6d9e12136c68c441b2350d015902c144b20"
|
|
4662
4270
|
},
|
|
4663
4271
|
{
|
|
4664
4272
|
"model_format": "ggufv2",
|
|
@@ -4721,157 +4329,13 @@
|
|
|
4721
4329
|
"model_file_name_template": "wizardcoder-python-34b-v1.0.{quantization}.gguf"
|
|
4722
4330
|
}
|
|
4723
4331
|
],
|
|
4724
|
-
"
|
|
4725
|
-
|
|
4726
|
-
|
|
4727
|
-
"roles": [
|
|
4728
|
-
"Instruction",
|
|
4729
|
-
"Response"
|
|
4730
|
-
],
|
|
4731
|
-
"intra_message_sep": "\n\n### ",
|
|
4732
|
-
"stop": [
|
|
4733
|
-
"</s>"
|
|
4734
|
-
]
|
|
4735
|
-
}
|
|
4736
|
-
},
|
|
4737
|
-
{
|
|
4738
|
-
"version": 1,
|
|
4739
|
-
"context_length": 8192,
|
|
4740
|
-
"model_name": "zephyr-7b-alpha",
|
|
4741
|
-
"model_lang": [
|
|
4742
|
-
"en"
|
|
4743
|
-
],
|
|
4744
|
-
"model_ability": [
|
|
4745
|
-
"chat"
|
|
4746
|
-
],
|
|
4747
|
-
"model_description": "Zephyr-7B-α is the first model in the series, and is a fine-tuned version of mistralai/Mistral-7B-v0.1.",
|
|
4748
|
-
"model_specs": [
|
|
4749
|
-
{
|
|
4750
|
-
"model_format": "pytorch",
|
|
4751
|
-
"model_size_in_billions": 7,
|
|
4752
|
-
"quantizations": [
|
|
4753
|
-
"4-bit",
|
|
4754
|
-
"8-bit",
|
|
4755
|
-
"none"
|
|
4756
|
-
],
|
|
4757
|
-
"model_id": "HuggingFaceH4/zephyr-7b-alpha",
|
|
4758
|
-
"model_revision": "f28e1c0e5a1af475bcd7bdf6554e69abc6c0c7ee"
|
|
4759
|
-
}
|
|
4760
|
-
],
|
|
4761
|
-
"prompt_style": {
|
|
4762
|
-
"style_name": "NO_COLON_TWO",
|
|
4763
|
-
"system_prompt": "<|system|>\nYou are a friendly chatbot.</s>\n",
|
|
4764
|
-
"roles": [
|
|
4765
|
-
"<|user|>\n",
|
|
4766
|
-
"<|assistant|>\n"
|
|
4767
|
-
],
|
|
4768
|
-
"intra_message_sep": "</s>\n",
|
|
4769
|
-
"inter_message_sep": "</s>\n",
|
|
4770
|
-
"stop_token_ids": [
|
|
4771
|
-
2
|
|
4772
|
-
],
|
|
4773
|
-
"stop": [
|
|
4774
|
-
"</s>"
|
|
4775
|
-
]
|
|
4776
|
-
}
|
|
4777
|
-
},
|
|
4778
|
-
{
|
|
4779
|
-
"version": 1,
|
|
4780
|
-
"context_length": 8192,
|
|
4781
|
-
"model_name": "zephyr-7b-beta",
|
|
4782
|
-
"model_lang": [
|
|
4783
|
-
"en"
|
|
4784
|
-
],
|
|
4785
|
-
"model_ability": [
|
|
4786
|
-
"chat"
|
|
4787
|
-
],
|
|
4788
|
-
"model_description": "Zephyr-7B-β is the second model in the series, and is a fine-tuned version of mistralai/Mistral-7B-v0.1",
|
|
4789
|
-
"model_specs": [
|
|
4790
|
-
{
|
|
4791
|
-
"model_format": "pytorch",
|
|
4792
|
-
"model_size_in_billions": 7,
|
|
4793
|
-
"quantizations": [
|
|
4794
|
-
"4-bit",
|
|
4795
|
-
"8-bit",
|
|
4796
|
-
"none"
|
|
4797
|
-
],
|
|
4798
|
-
"model_id": "HuggingFaceH4/zephyr-7b-beta",
|
|
4799
|
-
"model_revision": "3bac358730f8806e5c3dc7c7e19eb36e045bf720"
|
|
4800
|
-
}
|
|
4801
|
-
],
|
|
4802
|
-
"prompt_style": {
|
|
4803
|
-
"style_name": "NO_COLON_TWO",
|
|
4804
|
-
"system_prompt": "<|system|>\nYou are a friendly chatbot.</s>\n",
|
|
4805
|
-
"roles": [
|
|
4806
|
-
"<|user|>\n",
|
|
4807
|
-
"<|assistant|>\n"
|
|
4808
|
-
],
|
|
4809
|
-
"intra_message_sep": "</s>\n",
|
|
4810
|
-
"inter_message_sep": "</s>\n",
|
|
4811
|
-
"stop_token_ids": [
|
|
4812
|
-
2
|
|
4813
|
-
],
|
|
4814
|
-
"stop": [
|
|
4815
|
-
"</s>"
|
|
4816
|
-
]
|
|
4817
|
-
}
|
|
4818
|
-
},
|
|
4819
|
-
{
|
|
4820
|
-
"version": 1,
|
|
4821
|
-
"context_length": 4096,
|
|
4822
|
-
"model_name": "gorilla-openfunctions-v1",
|
|
4823
|
-
"model_lang": [
|
|
4824
|
-
"en"
|
|
4825
|
-
],
|
|
4826
|
-
"model_ability": [
|
|
4827
|
-
"chat"
|
|
4828
|
-
],
|
|
4829
|
-
"model_description": "OpenFunctions is designed to extend Large Language Model (LLM) Chat Completion feature to formulate executable APIs call given natural language instructions and API context.",
|
|
4830
|
-
"model_specs": [
|
|
4831
|
-
{
|
|
4832
|
-
"model_format": "pytorch",
|
|
4833
|
-
"model_size_in_billions": 7,
|
|
4834
|
-
"quantizations": [
|
|
4835
|
-
"4-bit",
|
|
4836
|
-
"8-bit",
|
|
4837
|
-
"none"
|
|
4838
|
-
],
|
|
4839
|
-
"model_id": "gorilla-llm/gorilla-openfunctions-v1",
|
|
4840
|
-
"model_revision": "74615f614ee845eab114e71541fd5098d1709958"
|
|
4841
|
-
},
|
|
4842
|
-
{
|
|
4843
|
-
"model_format": "ggufv2",
|
|
4844
|
-
"model_size_in_billions": 7,
|
|
4845
|
-
"quantizations": [
|
|
4846
|
-
"Q2_K",
|
|
4847
|
-
"Q3_K_L",
|
|
4848
|
-
"Q3_K_M",
|
|
4849
|
-
"Q3_K_S",
|
|
4850
|
-
"Q4_0",
|
|
4851
|
-
"Q4_K_M",
|
|
4852
|
-
"Q4_K_S",
|
|
4853
|
-
"Q5_0",
|
|
4854
|
-
"Q5_K_M",
|
|
4855
|
-
"Q5_K_S",
|
|
4856
|
-
"Q6_K",
|
|
4857
|
-
"Q8_0"
|
|
4858
|
-
],
|
|
4859
|
-
"model_id": "TheBloke/gorilla-openfunctions-v1-GGUF",
|
|
4860
|
-
"model_file_name_template": "gorilla-openfunctions-v1.{quantization}.gguf"
|
|
4861
|
-
}
|
|
4332
|
+
"chat_template": "{% for item in messages %}{% if loop.first and item['role'] == 'system' %}{{ item['content'] + '\n\n### ' }}{% elif loop.first %}{{ 'Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n### ' }}{% endif %}{% if item['role'] == 'user' %}{{ 'Instruction: ' + item['content'] + '\n\n### ' }}{% elif item['role'] == 'assistant' %}{{ 'Response: ' + item['content'] + '\n\n### ' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'Response: Let\\'s think step by step.' }}{% endif %}",
|
|
4333
|
+
"stop_token_ids": [
|
|
4334
|
+
2
|
|
4862
4335
|
],
|
|
4863
|
-
"
|
|
4864
|
-
"
|
|
4865
|
-
|
|
4866
|
-
"roles": [
|
|
4867
|
-
"",
|
|
4868
|
-
""
|
|
4869
|
-
],
|
|
4870
|
-
"intra_message_sep": "\n",
|
|
4871
|
-
"inter_message_sep": "\n",
|
|
4872
|
-
"stop_token_ids": [],
|
|
4873
|
-
"stop": []
|
|
4874
|
-
}
|
|
4336
|
+
"stop": [
|
|
4337
|
+
"</s>"
|
|
4338
|
+
]
|
|
4875
4339
|
},
|
|
4876
4340
|
{
|
|
4877
4341
|
"version": 1,
|
|
@@ -4913,18 +4377,15 @@
|
|
|
4913
4377
|
"model_file_name_template": "gorilla-openfunctions-v2.{quantization}.gguf"
|
|
4914
4378
|
}
|
|
4915
4379
|
],
|
|
4916
|
-
"
|
|
4917
|
-
|
|
4918
|
-
|
|
4919
|
-
|
|
4920
|
-
|
|
4921
|
-
|
|
4922
|
-
|
|
4923
|
-
"
|
|
4924
|
-
|
|
4925
|
-
"stop_token_ids": [],
|
|
4926
|
-
"stop": []
|
|
4927
|
-
}
|
|
4380
|
+
"chat_template": "{% if not add_generation_prompt is defined %}\n{% set add_generation_prompt = false %}\n{% endif %}\n{%- set ns = namespace(found=false) -%}\n{%- for message in messages -%}\n {%- if message['role'] == 'system' -%}\n {%- set ns.found = true -%}\n {%- endif -%}\n{%- endfor -%}\n{{'<|begin▁of▁sentence|>'}}{%- if not ns.found -%}\n{{'You are an AI programming assistant, utilizing the Gorilla LLM model, developed by Gorilla LLM, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer\n'}}\n{%- endif %}\n{%- for message in messages %}\n {%- if message['role'] == 'system' %}\n{{ message['content'] }}\n {%- else %}\n {%- if message['role'] == 'user' %}\n{{'### Instruction:\n' + message['content'] + '\n'}}\n {%- else %}\n{{'### Response:\n' + message['content'] + '\n<|EOT|>\n'}}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{% if add_generation_prompt %}\n{{'### Response:'}}\n{% endif %}",
|
|
4381
|
+
"stop_token_ids": [
|
|
4382
|
+
100015,
|
|
4383
|
+
100001
|
|
4384
|
+
],
|
|
4385
|
+
"stop": [
|
|
4386
|
+
"<|EOT|>",
|
|
4387
|
+
"<|end▁of▁sentence|>"
|
|
4388
|
+
]
|
|
4928
4389
|
},
|
|
4929
4390
|
{
|
|
4930
4391
|
"version": 1,
|
|
@@ -4959,19 +4420,13 @@
|
|
|
4959
4420
|
"model_revision": "6f16f00805f45b5249f709ce21820122eeb43556"
|
|
4960
4421
|
}
|
|
4961
4422
|
],
|
|
4962
|
-
"
|
|
4963
|
-
|
|
4964
|
-
|
|
4965
|
-
|
|
4966
|
-
|
|
4967
|
-
|
|
4968
|
-
|
|
4969
|
-
"intra_message_sep": "\n\n",
|
|
4970
|
-
"inter_message_sep": "<|end▁of▁sentence|>",
|
|
4971
|
-
"stop": [
|
|
4972
|
-
"<|end▁of▁sentence|>"
|
|
4973
|
-
]
|
|
4974
|
-
}
|
|
4423
|
+
"chat_template": "",
|
|
4424
|
+
"stop_token_ids": [
|
|
4425
|
+
100001
|
|
4426
|
+
],
|
|
4427
|
+
"stop": [
|
|
4428
|
+
"<|end▁of▁sentence|>"
|
|
4429
|
+
]
|
|
4975
4430
|
},
|
|
4976
4431
|
{
|
|
4977
4432
|
"version": 1,
|
|
@@ -5126,19 +4581,13 @@
|
|
|
5126
4581
|
"model_file_name_template": "deepseek-llm-67b-chat.{quantization}.gguf"
|
|
5127
4582
|
}
|
|
5128
4583
|
],
|
|
5129
|
-
"
|
|
5130
|
-
|
|
5131
|
-
|
|
5132
|
-
|
|
5133
|
-
|
|
5134
|
-
|
|
5135
|
-
|
|
5136
|
-
"intra_message_sep": "\n\n",
|
|
5137
|
-
"inter_message_sep": "<|end▁of▁sentence|>",
|
|
5138
|
-
"stop": [
|
|
5139
|
-
"<|end▁of▁sentence|>"
|
|
5140
|
-
]
|
|
5141
|
-
}
|
|
4584
|
+
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{{ '<|begin▁of▁sentence|>' }}{% for message in messages %}{% if message['role'] == 'user' %}{{ 'User: ' + message['content'] + '\n\n' }}{% elif message['role'] == 'assistant' %}{{ 'Assistant: ' + message['content'] + '<|end▁of▁sentence|>' }}{% elif message['role'] == 'system' %}{{ message['content'] + '\n\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'Assistant:' }}{% endif %}",
|
|
4585
|
+
"stop_token_ids": [
|
|
4586
|
+
100001
|
|
4587
|
+
],
|
|
4588
|
+
"stop": [
|
|
4589
|
+
"<|end▁of▁sentence|>"
|
|
4590
|
+
]
|
|
5142
4591
|
},
|
|
5143
4592
|
{
|
|
5144
4593
|
"version": 1,
|
|
@@ -5523,18 +4972,13 @@
|
|
|
5523
4972
|
"model_revision": "c40b499bac2712cd3c445cf1b05d2c6558ab0d29"
|
|
5524
4973
|
}
|
|
5525
4974
|
],
|
|
5526
|
-
"
|
|
5527
|
-
|
|
5528
|
-
|
|
5529
|
-
|
|
5530
|
-
|
|
5531
|
-
|
|
5532
|
-
|
|
5533
|
-
"inter_message_sep": "\n",
|
|
5534
|
-
"stop": [
|
|
5535
|
-
"<|EOT|>"
|
|
5536
|
-
]
|
|
5537
|
-
}
|
|
4975
|
+
"chat_template": "{% if not add_generation_prompt is defined %}\n{% set add_generation_prompt = false %}\n{% endif %}\n{%- set ns = namespace(found=false) -%}\n{%- for message in messages -%}\n {%- if message['role'] == 'system' -%}\n {%- set ns.found = true -%}\n {%- endif -%}\n{%- endfor -%}\n{{'<|begin▁of▁sentence|>'}}{%- if not ns.found -%}\n{{'You are an AI programming assistant, utilizing the Deepseek Coder model, developed by Deepseek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer\n'}}\n{%- endif %}\n{%- for message in messages %}\n {%- if message['role'] == 'system' %}\n{{ message['content'] }}\n {%- else %}\n {%- if message['role'] == 'user' %}\n{{'### Instruction:\n' + message['content'] + '\n'}}\n {%- else %}\n{{'### Response:\n' + message['content'] + '\n<|EOT|>\n'}}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{% if add_generation_prompt %}\n{{'### Response:'}}\n{% endif %}",
|
|
4976
|
+
"stop_token_ids": [
|
|
4977
|
+
32021
|
|
4978
|
+
],
|
|
4979
|
+
"stop": [
|
|
4980
|
+
"<|EOT|>"
|
|
4981
|
+
]
|
|
5538
4982
|
},
|
|
5539
4983
|
{
|
|
5540
4984
|
"version": 1,
|
|
@@ -5618,23 +5062,15 @@
|
|
|
5618
5062
|
"model_revision": "b666125047cd98c5a7c85ca28720b44a06aed124"
|
|
5619
5063
|
}
|
|
5620
5064
|
],
|
|
5621
|
-
"
|
|
5622
|
-
|
|
5623
|
-
|
|
5624
|
-
|
|
5625
|
-
|
|
5626
|
-
|
|
5627
|
-
|
|
5628
|
-
"
|
|
5629
|
-
|
|
5630
|
-
2,
|
|
5631
|
-
92542
|
|
5632
|
-
],
|
|
5633
|
-
"stop": [
|
|
5634
|
-
"</s>",
|
|
5635
|
-
"<|im_end|>"
|
|
5636
|
-
]
|
|
5637
|
-
}
|
|
5065
|
+
"chat_template": "{{ '<s>' }}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
|
5066
|
+
"stop_token_ids": [
|
|
5067
|
+
2,
|
|
5068
|
+
92542
|
|
5069
|
+
],
|
|
5070
|
+
"stop": [
|
|
5071
|
+
"</s>",
|
|
5072
|
+
"<|im_end|>"
|
|
5073
|
+
]
|
|
5638
5074
|
},
|
|
5639
5075
|
{
|
|
5640
5076
|
"version": 1,
|
|
@@ -5755,23 +5191,15 @@
|
|
|
5755
5191
|
"model_revision": "0ec94d61d30ab161b49c69f9bf92ec2b9986d234"
|
|
5756
5192
|
}
|
|
5757
5193
|
],
|
|
5758
|
-
"
|
|
5759
|
-
|
|
5760
|
-
|
|
5761
|
-
|
|
5762
|
-
|
|
5763
|
-
|
|
5764
|
-
|
|
5765
|
-
"
|
|
5766
|
-
|
|
5767
|
-
2,
|
|
5768
|
-
92542
|
|
5769
|
-
],
|
|
5770
|
-
"stop": [
|
|
5771
|
-
"</s>",
|
|
5772
|
-
"<|im_end|>"
|
|
5773
|
-
]
|
|
5774
|
-
}
|
|
5194
|
+
"chat_template": "{{ '<s>' }}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
|
5195
|
+
"stop_token_ids": [
|
|
5196
|
+
2,
|
|
5197
|
+
92542
|
|
5198
|
+
],
|
|
5199
|
+
"stop": [
|
|
5200
|
+
"</s>",
|
|
5201
|
+
"<|im_end|>"
|
|
5202
|
+
]
|
|
5775
5203
|
},
|
|
5776
5204
|
{
|
|
5777
5205
|
"version": 1,
|
|
@@ -5822,23 +5250,15 @@
|
|
|
5822
5250
|
"model_file_name_template": "internlm2_5-7b-chat-1m-{quantization}.gguf"
|
|
5823
5251
|
}
|
|
5824
5252
|
],
|
|
5825
|
-
"
|
|
5826
|
-
|
|
5827
|
-
|
|
5828
|
-
|
|
5829
|
-
|
|
5830
|
-
|
|
5831
|
-
|
|
5832
|
-
"
|
|
5833
|
-
|
|
5834
|
-
2,
|
|
5835
|
-
92542
|
|
5836
|
-
],
|
|
5837
|
-
"stop": [
|
|
5838
|
-
"</s>",
|
|
5839
|
-
"<|im_end|>"
|
|
5840
|
-
]
|
|
5841
|
-
}
|
|
5253
|
+
"chat_template": "{{ '<s>' }}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
|
5254
|
+
"stop_token_ids": [
|
|
5255
|
+
2,
|
|
5256
|
+
92542
|
|
5257
|
+
],
|
|
5258
|
+
"stop": [
|
|
5259
|
+
"</s>",
|
|
5260
|
+
"<|im_end|>"
|
|
5261
|
+
]
|
|
5842
5262
|
},
|
|
5843
5263
|
{
|
|
5844
5264
|
"version":1,
|
|
@@ -5873,14 +5293,13 @@
|
|
|
5873
5293
|
"model_revision":"ef62bae5af34be653b9801037cd613e05ab24fdc"
|
|
5874
5294
|
}
|
|
5875
5295
|
],
|
|
5876
|
-
"
|
|
5877
|
-
|
|
5878
|
-
|
|
5879
|
-
|
|
5880
|
-
|
|
5881
|
-
|
|
5882
|
-
|
|
5883
|
-
}
|
|
5296
|
+
"chat_template": "",
|
|
5297
|
+
"stop_token_ids": [
|
|
5298
|
+
2
|
|
5299
|
+
],
|
|
5300
|
+
"stop": [
|
|
5301
|
+
"</s>"
|
|
5302
|
+
]
|
|
5884
5303
|
},
|
|
5885
5304
|
{
|
|
5886
5305
|
"version":1,
|
|
@@ -5915,14 +5334,13 @@
|
|
|
5915
5334
|
"model_revision":"f92aff28552de35de3be204e8fe292dd4824e544"
|
|
5916
5335
|
}
|
|
5917
5336
|
],
|
|
5918
|
-
"
|
|
5919
|
-
|
|
5920
|
-
|
|
5921
|
-
|
|
5922
|
-
|
|
5923
|
-
|
|
5924
|
-
|
|
5925
|
-
}
|
|
5337
|
+
"chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = '<|begin_of_text|>' + content %}{% endif %}{{ content }}{% endfor %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}",
|
|
5338
|
+
"stop_token_ids": [
|
|
5339
|
+
128001
|
|
5340
|
+
],
|
|
5341
|
+
"stop": [
|
|
5342
|
+
"<|end_of_text|>"
|
|
5343
|
+
]
|
|
5926
5344
|
},
|
|
5927
5345
|
{
|
|
5928
5346
|
"version":1,
|
|
@@ -5957,18 +5375,15 @@
|
|
|
5957
5375
|
"model_revision":"051e2df6505f1fc4305f2c9bd42ed90db8bf4874"
|
|
5958
5376
|
}
|
|
5959
5377
|
],
|
|
5960
|
-
"
|
|
5961
|
-
|
|
5962
|
-
|
|
5963
|
-
|
|
5964
|
-
|
|
5965
|
-
|
|
5966
|
-
|
|
5967
|
-
"
|
|
5968
|
-
|
|
5969
|
-
"<|endoftext|>"
|
|
5970
|
-
]
|
|
5971
|
-
}
|
|
5378
|
+
"chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
|
5379
|
+
"stop_token_ids": [
|
|
5380
|
+
151645,
|
|
5381
|
+
151643
|
|
5382
|
+
],
|
|
5383
|
+
"stop": [
|
|
5384
|
+
"<|im_end|>",
|
|
5385
|
+
"<|endoftext|>"
|
|
5386
|
+
]
|
|
5972
5387
|
},
|
|
5973
5388
|
{
|
|
5974
5389
|
"version": 1,
|
|
@@ -6003,24 +5418,17 @@
|
|
|
6003
5418
|
"model_revision": "5d3a5aa033ed2c502300d426c81cc5b13bcd1409"
|
|
6004
5419
|
}
|
|
6005
5420
|
],
|
|
6006
|
-
"
|
|
6007
|
-
|
|
6008
|
-
|
|
6009
|
-
|
|
6010
|
-
|
|
6011
|
-
|
|
6012
|
-
|
|
6013
|
-
"
|
|
6014
|
-
|
|
6015
|
-
|
|
6016
|
-
|
|
6017
|
-
],
|
|
6018
|
-
"stop": [
|
|
6019
|
-
"<|endoftext|>",
|
|
6020
|
-
"<|im_start|>",
|
|
6021
|
-
"<|im_end|>"
|
|
6022
|
-
]
|
|
6023
|
-
}
|
|
5421
|
+
"chat_template": "",
|
|
5422
|
+
"stop_token_ids": [
|
|
5423
|
+
151643,
|
|
5424
|
+
151644,
|
|
5425
|
+
151645
|
|
5426
|
+
],
|
|
5427
|
+
"stop": [
|
|
5428
|
+
"<|endoftext|>",
|
|
5429
|
+
"<|im_start|>",
|
|
5430
|
+
"<|im_end|>"
|
|
5431
|
+
]
|
|
6024
5432
|
},
|
|
6025
5433
|
{
|
|
6026
5434
|
"version": 1,
|
|
@@ -6055,18 +5463,17 @@
|
|
|
6055
5463
|
"model_id": "OrionStarAI/Orion-14B-Chat-{quantization}"
|
|
6056
5464
|
}
|
|
6057
5465
|
],
|
|
6058
|
-
"
|
|
6059
|
-
|
|
6060
|
-
|
|
6061
|
-
|
|
6062
|
-
|
|
6063
|
-
|
|
6064
|
-
|
|
6065
|
-
|
|
6066
|
-
|
|
6067
|
-
|
|
6068
|
-
|
|
6069
|
-
}
|
|
5466
|
+
"chat_template": "{% for message in messages %}{% if loop.first %}{{ '<s>' }}{% endif %}{% if message['role'] == 'user' %}{{ 'Human: ' + message['content'] + '\n\nAssistant: ' + '</s>' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + '</s>' }}{% endif %}{% endfor %}",
|
|
5467
|
+
"stop_token_ids": [
|
|
5468
|
+
1,
|
|
5469
|
+
2,
|
|
5470
|
+
0
|
|
5471
|
+
],
|
|
5472
|
+
"stop": [
|
|
5473
|
+
"<s>",
|
|
5474
|
+
"</s>",
|
|
5475
|
+
"<unk>"
|
|
5476
|
+
]
|
|
6070
5477
|
},
|
|
6071
5478
|
{
|
|
6072
5479
|
"version": 1,
|
|
@@ -6093,18 +5500,17 @@
|
|
|
6093
5500
|
"model_revision": "eba2e20808407fb431a76b90d5d506e04a0325f2"
|
|
6094
5501
|
}
|
|
6095
5502
|
],
|
|
6096
|
-
"
|
|
6097
|
-
|
|
6098
|
-
|
|
6099
|
-
|
|
6100
|
-
|
|
6101
|
-
|
|
6102
|
-
|
|
6103
|
-
|
|
6104
|
-
|
|
6105
|
-
|
|
6106
|
-
|
|
6107
|
-
}
|
|
5503
|
+
"chat_template": "{% for message in messages %}{% if loop.first %}{{ '<s>' }}{% endif %}{% if message['role'] == 'user' %}{{ 'Human: ' + message['content'] + '\n\nAssistant: ' + '</s>' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + '</s>' }}{% endif %}{% endfor %}",
|
|
5504
|
+
"stop_token_ids": [
|
|
5505
|
+
1,
|
|
5506
|
+
2,
|
|
5507
|
+
0
|
|
5508
|
+
],
|
|
5509
|
+
"stop": [
|
|
5510
|
+
"<s>",
|
|
5511
|
+
"</s>",
|
|
5512
|
+
"<unk>"
|
|
5513
|
+
]
|
|
6108
5514
|
},
|
|
6109
5515
|
{
|
|
6110
5516
|
"version": 1,
|
|
@@ -6139,28 +5545,19 @@
|
|
|
6139
5545
|
"model_revision": "ea29a9a430f27893e780366dae81d4ca5ebab561"
|
|
6140
5546
|
}
|
|
6141
5547
|
],
|
|
6142
|
-
"
|
|
6143
|
-
|
|
6144
|
-
|
|
6145
|
-
|
|
6146
|
-
|
|
6147
|
-
|
|
6148
|
-
|
|
6149
|
-
|
|
6150
|
-
"
|
|
6151
|
-
"
|
|
6152
|
-
|
|
6153
|
-
|
|
6154
|
-
|
|
6155
|
-
8
|
|
6156
|
-
],
|
|
6157
|
-
"stop": [
|
|
6158
|
-
"<|endoftext|>",
|
|
6159
|
-
"<|im_start|>",
|
|
6160
|
-
"<|im_end|>",
|
|
6161
|
-
"<|im_sep|>"
|
|
6162
|
-
]
|
|
6163
|
-
}
|
|
5548
|
+
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
|
5549
|
+
"stop_token_ids": [
|
|
5550
|
+
2,
|
|
5551
|
+
6,
|
|
5552
|
+
7,
|
|
5553
|
+
8
|
|
5554
|
+
],
|
|
5555
|
+
"stop": [
|
|
5556
|
+
"<|endoftext|>",
|
|
5557
|
+
"<|im_start|>",
|
|
5558
|
+
"<|im_end|>",
|
|
5559
|
+
"<|im_sep|>"
|
|
5560
|
+
]
|
|
6164
5561
|
},
|
|
6165
5562
|
{
|
|
6166
5563
|
"version": 1,
|
|
@@ -6195,17 +5592,17 @@
|
|
|
6195
5592
|
"model_id": "google/gemma-7b-it"
|
|
6196
5593
|
}
|
|
6197
5594
|
],
|
|
6198
|
-
"
|
|
6199
|
-
|
|
6200
|
-
|
|
6201
|
-
|
|
6202
|
-
|
|
6203
|
-
|
|
6204
|
-
|
|
6205
|
-
|
|
6206
|
-
|
|
6207
|
-
|
|
6208
|
-
|
|
5595
|
+
"chat_template": "{{ '<bos>' }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}",
|
|
5596
|
+
"stop_token_ids": [
|
|
5597
|
+
1,
|
|
5598
|
+
106,
|
|
5599
|
+
107
|
|
5600
|
+
],
|
|
5601
|
+
"stop": [
|
|
5602
|
+
"<eos>",
|
|
5603
|
+
"<end_of_turn>",
|
|
5604
|
+
"<start_of_turn>"
|
|
5605
|
+
]
|
|
6209
5606
|
},
|
|
6210
5607
|
{
|
|
6211
5608
|
"version": 1,
|
|
@@ -6385,17 +5782,17 @@
|
|
|
6385
5782
|
"model_id": "mlx-community/gemma-2-27b-it-fp16"
|
|
6386
5783
|
}
|
|
6387
5784
|
],
|
|
6388
|
-
"
|
|
6389
|
-
|
|
6390
|
-
|
|
6391
|
-
|
|
6392
|
-
|
|
6393
|
-
|
|
6394
|
-
|
|
6395
|
-
|
|
6396
|
-
|
|
6397
|
-
|
|
6398
|
-
|
|
5785
|
+
"chat_template": "{{ '<bos>' }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}",
|
|
5786
|
+
"stop_token_ids": [
|
|
5787
|
+
1,
|
|
5788
|
+
106,
|
|
5789
|
+
107
|
|
5790
|
+
],
|
|
5791
|
+
"stop": [
|
|
5792
|
+
"<eos>",
|
|
5793
|
+
"<end_of_turn>",
|
|
5794
|
+
"<start_of_turn>"
|
|
5795
|
+
]
|
|
6399
5796
|
},
|
|
6400
5797
|
{
|
|
6401
5798
|
"version": 1,
|
|
@@ -6539,23 +5936,15 @@
|
|
|
6539
5936
|
"model_revision": "0df19b6e10f1a19ca663f7cc1141aae10f1825f4"
|
|
6540
5937
|
}
|
|
6541
5938
|
],
|
|
6542
|
-
"
|
|
6543
|
-
|
|
6544
|
-
|
|
6545
|
-
|
|
6546
|
-
|
|
6547
|
-
|
|
6548
|
-
|
|
6549
|
-
|
|
6550
|
-
|
|
6551
|
-
100006,
|
|
6552
|
-
100007
|
|
6553
|
-
],
|
|
6554
|
-
"stop": [
|
|
6555
|
-
"[CLS]",
|
|
6556
|
-
"</s>"
|
|
6557
|
-
]
|
|
6558
|
-
}
|
|
5939
|
+
"chat_template": "{% for item in messages %}{% if loop.first and item['role'] == 'system' %}{{ item['content'] + '\n' }}{% endif %}{% if item['role'] == 'user' %}{{ 'USER: ' + item['content'] + '\n' }}{% elif item['role'] == 'assistant' %}{{ 'ASSISTANT: ' + item['content'] + '\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'ASSISTANT: ' }}{% endif %}",
|
|
5940
|
+
"stop_token_ids": [
|
|
5941
|
+
100006,
|
|
5942
|
+
100007
|
|
5943
|
+
],
|
|
5944
|
+
"stop": [
|
|
5945
|
+
"[CLS]",
|
|
5946
|
+
"</s>"
|
|
5947
|
+
]
|
|
6559
5948
|
},
|
|
6560
5949
|
{
|
|
6561
5950
|
"version": 1,
|
|
@@ -6626,23 +6015,15 @@
|
|
|
6626
6015
|
"model_revision": "a06fd164c7170714924d2881c61c8348425ebc94"
|
|
6627
6016
|
}
|
|
6628
6017
|
],
|
|
6629
|
-
"
|
|
6630
|
-
|
|
6631
|
-
|
|
6632
|
-
|
|
6633
|
-
|
|
6634
|
-
|
|
6635
|
-
|
|
6636
|
-
|
|
6637
|
-
|
|
6638
|
-
100006,
|
|
6639
|
-
100007
|
|
6640
|
-
],
|
|
6641
|
-
"stop": [
|
|
6642
|
-
"[CLS]",
|
|
6643
|
-
"</s>"
|
|
6644
|
-
]
|
|
6645
|
-
}
|
|
6018
|
+
"chat_template": "{% for item in messages %}{% if loop.first and item['role'] == 'system' %}{{ item['content'] + '\n' }}{% endif %}{% if item['role'] == 'user' %}{{ 'USER: ' + item['content'] + '\n' }}{% elif item['role'] == 'assistant' %}{{ 'ASSISTANT: ' + item['content'] + '\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'ASSISTANT: ' }}{% endif %}",
|
|
6019
|
+
"stop_token_ids": [
|
|
6020
|
+
100006,
|
|
6021
|
+
100007
|
|
6022
|
+
],
|
|
6023
|
+
"stop": [
|
|
6024
|
+
"[CLS]",
|
|
6025
|
+
"</s>"
|
|
6026
|
+
]
|
|
6646
6027
|
},
|
|
6647
6028
|
{
|
|
6648
6029
|
"version": 1,
|
|
@@ -6666,22 +6047,15 @@
|
|
|
6666
6047
|
"model_revision": "fe1d74027ebdd81cef5f815fa3a2d432a6b5de2a"
|
|
6667
6048
|
}
|
|
6668
6049
|
],
|
|
6669
|
-
"
|
|
6670
|
-
|
|
6671
|
-
|
|
6672
|
-
|
|
6673
|
-
|
|
6674
|
-
|
|
6675
|
-
|
|
6676
|
-
"
|
|
6677
|
-
|
|
6678
|
-
2
|
|
6679
|
-
],
|
|
6680
|
-
"stop": [
|
|
6681
|
-
"<s>",
|
|
6682
|
-
"</s>"
|
|
6683
|
-
]
|
|
6684
|
-
}
|
|
6050
|
+
"chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<用户>' + message['content'].strip() + '<AI>'}}{% else %}{{message['content'].strip()}}{% endif %}{% endfor %}",
|
|
6051
|
+
"stop_token_ids": [
|
|
6052
|
+
1,
|
|
6053
|
+
2
|
|
6054
|
+
],
|
|
6055
|
+
"stop": [
|
|
6056
|
+
"<s>",
|
|
6057
|
+
"</s>"
|
|
6058
|
+
]
|
|
6685
6059
|
},
|
|
6686
6060
|
{
|
|
6687
6061
|
"version": 1,
|
|
@@ -6705,22 +6079,15 @@
|
|
|
6705
6079
|
"model_revision": "35b90dd57d977b6e5bc4907986fa5b77aa15a82e"
|
|
6706
6080
|
}
|
|
6707
6081
|
],
|
|
6708
|
-
"
|
|
6709
|
-
|
|
6710
|
-
|
|
6711
|
-
|
|
6712
|
-
|
|
6713
|
-
|
|
6714
|
-
|
|
6715
|
-
"
|
|
6716
|
-
|
|
6717
|
-
2
|
|
6718
|
-
],
|
|
6719
|
-
"stop": [
|
|
6720
|
-
"<s>",
|
|
6721
|
-
"</s>"
|
|
6722
|
-
]
|
|
6723
|
-
}
|
|
6082
|
+
"chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<用户>' + message['content'].strip() + '<AI>'}}{% else %}{{message['content'].strip()}}{% endif %}{% endfor %}",
|
|
6083
|
+
"stop_token_ids": [
|
|
6084
|
+
1,
|
|
6085
|
+
2
|
|
6086
|
+
],
|
|
6087
|
+
"stop": [
|
|
6088
|
+
"<s>",
|
|
6089
|
+
"</s>"
|
|
6090
|
+
]
|
|
6724
6091
|
},
|
|
6725
6092
|
{
|
|
6726
6093
|
"version": 1,
|
|
@@ -6744,22 +6111,15 @@
|
|
|
6744
6111
|
"model_revision": "f4a3ba49f3f18695945c2a7c12400d4da99da498"
|
|
6745
6112
|
}
|
|
6746
6113
|
],
|
|
6747
|
-
"
|
|
6748
|
-
|
|
6749
|
-
|
|
6750
|
-
|
|
6751
|
-
|
|
6752
|
-
|
|
6753
|
-
|
|
6754
|
-
"
|
|
6755
|
-
|
|
6756
|
-
2
|
|
6757
|
-
],
|
|
6758
|
-
"stop": [
|
|
6759
|
-
"<s>",
|
|
6760
|
-
"</s>"
|
|
6761
|
-
]
|
|
6762
|
-
}
|
|
6114
|
+
"chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<用户>' + message['content'].strip() + '<AI>'}}{% else %}{{message['content'].strip()}}{% endif %}{% endfor %}",
|
|
6115
|
+
"stop_token_ids": [
|
|
6116
|
+
1,
|
|
6117
|
+
2
|
|
6118
|
+
],
|
|
6119
|
+
"stop": [
|
|
6120
|
+
"<s>",
|
|
6121
|
+
"</s>"
|
|
6122
|
+
]
|
|
6763
6123
|
},
|
|
6764
6124
|
{
|
|
6765
6125
|
"version": 1,
|
|
@@ -6783,22 +6143,15 @@
|
|
|
6783
6143
|
"model_revision": "e7a50289e4f839674cf8d4a5a2ce032ccacf64ac"
|
|
6784
6144
|
}
|
|
6785
6145
|
],
|
|
6786
|
-
"
|
|
6787
|
-
|
|
6788
|
-
|
|
6789
|
-
|
|
6790
|
-
|
|
6791
|
-
|
|
6792
|
-
|
|
6793
|
-
"
|
|
6794
|
-
|
|
6795
|
-
2
|
|
6796
|
-
],
|
|
6797
|
-
"stop": [
|
|
6798
|
-
"<s>",
|
|
6799
|
-
"</s>"
|
|
6800
|
-
]
|
|
6801
|
-
}
|
|
6146
|
+
"chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<用户>' + message['content'].strip() + '<AI>'}}{% else %}{{message['content'].strip()}}{% endif %}{% endfor %}",
|
|
6147
|
+
"stop_token_ids": [
|
|
6148
|
+
1,
|
|
6149
|
+
2
|
|
6150
|
+
],
|
|
6151
|
+
"stop": [
|
|
6152
|
+
"<s>",
|
|
6153
|
+
"</s>"
|
|
6154
|
+
]
|
|
6802
6155
|
},
|
|
6803
6156
|
{
|
|
6804
6157
|
"version": 1,
|
|
@@ -6822,22 +6175,15 @@
|
|
|
6822
6175
|
"model_revision": "b560a1593779b735a84a6daf72fba96ae38da288"
|
|
6823
6176
|
}
|
|
6824
6177
|
],
|
|
6825
|
-
"
|
|
6826
|
-
|
|
6827
|
-
|
|
6828
|
-
|
|
6829
|
-
|
|
6830
|
-
|
|
6831
|
-
|
|
6832
|
-
"
|
|
6833
|
-
|
|
6834
|
-
2
|
|
6835
|
-
],
|
|
6836
|
-
"stop": [
|
|
6837
|
-
"<s>",
|
|
6838
|
-
"</s>"
|
|
6839
|
-
]
|
|
6840
|
-
}
|
|
6178
|
+
"chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<用户>' + message['content'].strip() + '<AI>'}}{% else %}{{message['content'].strip()}}{% endif %}{% endfor %}",
|
|
6179
|
+
"stop_token_ids": [
|
|
6180
|
+
1,
|
|
6181
|
+
2
|
|
6182
|
+
],
|
|
6183
|
+
"stop": [
|
|
6184
|
+
"<s>",
|
|
6185
|
+
"</s>"
|
|
6186
|
+
]
|
|
6841
6187
|
},
|
|
6842
6188
|
{
|
|
6843
6189
|
"version": 1,
|
|
@@ -7010,20 +6356,15 @@
|
|
|
7010
6356
|
"model_revision": "35febfc08f723ac0df32480eb4af349a7d08656e"
|
|
7011
6357
|
}
|
|
7012
6358
|
],
|
|
7013
|
-
"
|
|
7014
|
-
|
|
7015
|
-
|
|
7016
|
-
|
|
7017
|
-
|
|
7018
|
-
|
|
7019
|
-
|
|
7020
|
-
"
|
|
7021
|
-
|
|
7022
|
-
"stop_token_ids": [
|
|
7023
|
-
6,
|
|
7024
|
-
255001
|
|
7025
|
-
]
|
|
7026
|
-
}
|
|
6359
|
+
"chat_template": "{{ '<BOS_TOKEN>' }}{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% elif false == true %}{% set loop_messages = messages %}{% set system_message = 'You are Command-R, a brilliant, sophisticated, AI-assistant trained to assist human users by providing thorough responses. You are trained by Cohere.' %}{% else %}{% set loop_messages = messages %}{% set system_message = false %}{% endif %}{% if system_message != false %}{{ '<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>' + system_message + '<|END_OF_TURN_TOKEN|>' }}{% endif %}{% for message in loop_messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|START_OF_TURN_TOKEN|><|USER_TOKEN|>' + content.strip() + '<|END_OF_TURN_TOKEN|>' }}{% elif message['role'] == 'assistant' %}{{ '<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>' + content.strip() + '<|END_OF_TURN_TOKEN|>' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>' }}{% endif %}",
|
|
6360
|
+
"stop_token_ids": [
|
|
6361
|
+
6,
|
|
6362
|
+
255001
|
|
6363
|
+
],
|
|
6364
|
+
"stop": [
|
|
6365
|
+
"<EOS_TOKEN>",
|
|
6366
|
+
"<|END_OF_TURN_TOKEN|>"
|
|
6367
|
+
]
|
|
7027
6368
|
},
|
|
7028
6369
|
{
|
|
7029
6370
|
"version": 1,
|
|
@@ -7050,20 +6391,15 @@
|
|
|
7050
6391
|
"model_revision": "1dddf3b95bc1391f6307299eb1c162c194bde9bd"
|
|
7051
6392
|
}
|
|
7052
6393
|
],
|
|
7053
|
-
"
|
|
7054
|
-
|
|
7055
|
-
|
|
7056
|
-
|
|
7057
|
-
|
|
7058
|
-
|
|
7059
|
-
|
|
7060
|
-
"
|
|
7061
|
-
|
|
7062
|
-
"stop_token_ids": [
|
|
7063
|
-
2,
|
|
7064
|
-
32000
|
|
7065
|
-
]
|
|
7066
|
-
}
|
|
6394
|
+
"chat_template": "ssage in messages %}{{ 'GPT4 Correct ' + message['role'].title() + ': ' + message['content'] + '<|end_of_turn|>'}}{% endfor %}{% if add_generation_prompt %}{{ 'GPT4 Correct Assistant:' }}{% endif %}",
|
|
6395
|
+
"stop_token_ids": [
|
|
6396
|
+
2,
|
|
6397
|
+
32000
|
|
6398
|
+
],
|
|
6399
|
+
"stop": [
|
|
6400
|
+
"</s>",
|
|
6401
|
+
"<|end_of_turn|>"
|
|
6402
|
+
]
|
|
7067
6403
|
},
|
|
7068
6404
|
{
|
|
7069
6405
|
"version": 1,
|
|
@@ -7113,25 +6449,17 @@
|
|
|
7113
6449
|
"model_revision": "9db32d9127cac0c85961e169d75da57a18a847b1"
|
|
7114
6450
|
}
|
|
7115
6451
|
],
|
|
7116
|
-
"
|
|
7117
|
-
|
|
7118
|
-
|
|
7119
|
-
|
|
7120
|
-
|
|
7121
|
-
|
|
7122
|
-
|
|
7123
|
-
|
|
7124
|
-
|
|
7125
|
-
|
|
7126
|
-
|
|
7127
|
-
92542
|
|
7128
|
-
],
|
|
7129
|
-
"stop": [
|
|
7130
|
-
"</s>",
|
|
7131
|
-
"<|im_end|>",
|
|
7132
|
-
"<|im_start|>"
|
|
7133
|
-
]
|
|
7134
|
-
}
|
|
6452
|
+
"chat_template": "{{ '<s>' }}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
|
6453
|
+
"stop_token_ids": [
|
|
6454
|
+
2,
|
|
6455
|
+
92542,
|
|
6456
|
+
92543
|
|
6457
|
+
],
|
|
6458
|
+
"stop": [
|
|
6459
|
+
"</s>",
|
|
6460
|
+
"<|im_end|>",
|
|
6461
|
+
"<|im_start|>"
|
|
6462
|
+
]
|
|
7135
6463
|
},
|
|
7136
6464
|
{
|
|
7137
6465
|
"version": 1,
|
|
@@ -7270,25 +6598,9 @@
|
|
|
7270
6598
|
"model_revision": "1bc796bf80f2ebc7d6a14c15f55217a4600d50a4"
|
|
7271
6599
|
}
|
|
7272
6600
|
],
|
|
7273
|
-
"
|
|
7274
|
-
|
|
7275
|
-
|
|
7276
|
-
"roles": [
|
|
7277
|
-
"<|im_start|>user",
|
|
7278
|
-
"<|im_start|>assistant"
|
|
7279
|
-
],
|
|
7280
|
-
"intra_message_sep": "<|im_end|>",
|
|
7281
|
-
"stop_token_ids": [
|
|
7282
|
-
2,
|
|
7283
|
-
92543,
|
|
7284
|
-
92542
|
|
7285
|
-
],
|
|
7286
|
-
"stop": [
|
|
7287
|
-
"</s>",
|
|
7288
|
-
"<|im_end|>",
|
|
7289
|
-
"<|im_start|>"
|
|
7290
|
-
]
|
|
7291
|
-
}
|
|
6601
|
+
"chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
|
6602
|
+
"stop_token_ids": [],
|
|
6603
|
+
"stop": []
|
|
7292
6604
|
},
|
|
7293
6605
|
{
|
|
7294
6606
|
"version": 1,
|
|
@@ -7323,24 +6635,15 @@
|
|
|
7323
6635
|
"model_revision": "7863e362174f4718c2fe9cba4befd0b580a3194f"
|
|
7324
6636
|
}
|
|
7325
6637
|
],
|
|
7326
|
-
"
|
|
7327
|
-
|
|
7328
|
-
|
|
7329
|
-
|
|
7330
|
-
|
|
7331
|
-
|
|
7332
|
-
|
|
7333
|
-
"
|
|
7334
|
-
|
|
7335
|
-
"stop_token_ids": [
|
|
7336
|
-
128001,
|
|
7337
|
-
128009
|
|
7338
|
-
],
|
|
7339
|
-
"stop": [
|
|
7340
|
-
"<|end_of_text|>",
|
|
7341
|
-
"<|eot_id|>"
|
|
7342
|
-
]
|
|
7343
|
-
}
|
|
6638
|
+
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = '<|begin_of_text|>' + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% else %}{{ '<|end_of_text|>' }}{% endif %}",
|
|
6639
|
+
"stop_token_ids": [
|
|
6640
|
+
128001,
|
|
6641
|
+
128009
|
|
6642
|
+
],
|
|
6643
|
+
"stop": [
|
|
6644
|
+
"<|end_of_text|>",
|
|
6645
|
+
"<|eot_id|>"
|
|
6646
|
+
]
|
|
7344
6647
|
},
|
|
7345
6648
|
{
|
|
7346
6649
|
"version": 1,
|
|
@@ -7368,24 +6671,15 @@
|
|
|
7368
6671
|
"model_revision": "f375ead7d8202ebe2c3d09f1068abdddeb2929fa"
|
|
7369
6672
|
}
|
|
7370
6673
|
],
|
|
7371
|
-
"
|
|
7372
|
-
|
|
7373
|
-
|
|
7374
|
-
|
|
7375
|
-
|
|
7376
|
-
|
|
7377
|
-
|
|
7378
|
-
"
|
|
7379
|
-
|
|
7380
|
-
"stop_token_ids": [
|
|
7381
|
-
128001,
|
|
7382
|
-
128009
|
|
7383
|
-
],
|
|
7384
|
-
"stop": [
|
|
7385
|
-
"<|end_of_text|>",
|
|
7386
|
-
"<|eot_id|>"
|
|
7387
|
-
]
|
|
7388
|
-
}
|
|
6674
|
+
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = '<|begin_of_text|>' + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% else %}{{ '<|end_of_text|>' }}{% endif %}",
|
|
6675
|
+
"stop_token_ids": [
|
|
6676
|
+
128001,
|
|
6677
|
+
128009
|
|
6678
|
+
],
|
|
6679
|
+
"stop": [
|
|
6680
|
+
"<|end_of_text|>",
|
|
6681
|
+
"<|eot_id|>"
|
|
6682
|
+
]
|
|
7389
6683
|
},
|
|
7390
6684
|
{
|
|
7391
6685
|
"version": 1,
|
|
@@ -7449,24 +6743,15 @@
|
|
|
7449
6743
|
"model_id": "Tele-AI/TeleChat-52B"
|
|
7450
6744
|
}
|
|
7451
6745
|
],
|
|
7452
|
-
"
|
|
7453
|
-
|
|
7454
|
-
"
|
|
7455
|
-
"
|
|
7456
|
-
|
|
7457
|
-
|
|
7458
|
-
|
|
7459
|
-
|
|
7460
|
-
|
|
7461
|
-
"stop": [
|
|
7462
|
-
"<_end>",
|
|
7463
|
-
"<_start>"
|
|
7464
|
-
],
|
|
7465
|
-
"stop_token_ids": [
|
|
7466
|
-
160133,
|
|
7467
|
-
160132
|
|
7468
|
-
]
|
|
7469
|
-
}
|
|
6746
|
+
"chat_template": "{{ (messages|selectattr('role', 'equalto', 'system')|list|last).content|trim if (messages|selectattr('role', 'equalto', 'system')|list) else '' }}{%- for message in messages -%}{%- if message['role'] == 'user' -%}{{- '<_user>' + message['content'] +'<_bot>' -}}{%- elif message['role'] == 'assistant' -%}{{- message['content'] + '<_end>' -}}{%- endif -%}{%- endfor -%}",
|
|
6747
|
+
"stop": [
|
|
6748
|
+
"<_end>",
|
|
6749
|
+
"<_start>"
|
|
6750
|
+
],
|
|
6751
|
+
"stop_token_ids": [
|
|
6752
|
+
160133,
|
|
6753
|
+
160132
|
|
6754
|
+
]
|
|
7470
6755
|
},
|
|
7471
6756
|
{
|
|
7472
6757
|
"version": 1,
|
|
@@ -7513,21 +6798,451 @@
|
|
|
7513
6798
|
"model_file_name_template": "csg-wukong-1B-chat-v0.1.{quantization}.gguf"
|
|
7514
6799
|
}
|
|
7515
6800
|
],
|
|
7516
|
-
"
|
|
7517
|
-
|
|
7518
|
-
|
|
7519
|
-
|
|
7520
|
-
|
|
7521
|
-
|
|
6801
|
+
"chat_template": "{% for item in messages %}{% if loop.first and item['role'] == 'system' %}{{ item['content'] + '\n' }}{% elif loop.first %}{{ '<|system|>\nYou are a creative super artificial intelligence assistant, possessing all the knowledge of humankind. Your name is csg-wukong, developed by OpenCSG. You need to understand and infer the true intentions of users based on the topics discussed in the chat history, and respond to user questions correctly as required. You enjoy responding to users with accurate and insightful answers. Please pay attention to the appropriate style and format when replying, try to avoid repetitive words and sentences, and keep your responses as concise and profound as possible. You carefully consider the context of the discussion when replying to users. When the user says \"continue,\" please proceed with the continuation of the previous assistant\\'s response.</s>\n' }}{% endif %}{% if item['role'] == 'user' %}{{ '<|user|>\n' + item['content'] + '</s>\n' }}{% elif item['role'] == 'assistant' %}{{ '<|assistant|>\n' + item['content'] + '</s>\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% endif %}",
|
|
6802
|
+
"stop_token_ids": [
|
|
6803
|
+
2
|
|
6804
|
+
],
|
|
6805
|
+
"stop": [
|
|
6806
|
+
"</s>"
|
|
6807
|
+
]
|
|
6808
|
+
},
|
|
6809
|
+
{
|
|
6810
|
+
"version":1,
|
|
6811
|
+
"context_length":32768,
|
|
6812
|
+
"model_name":"qwen2-vl-instruct",
|
|
6813
|
+
"model_lang":[
|
|
6814
|
+
"en",
|
|
6815
|
+
"zh"
|
|
6816
|
+
],
|
|
6817
|
+
"model_ability":[
|
|
6818
|
+
"chat",
|
|
6819
|
+
"vision"
|
|
6820
|
+
],
|
|
6821
|
+
"model_description":"Qwen2-VL: To See the World More Clearly.Qwen2-VL is the latest version of the vision language models in the Qwen model familities.",
|
|
6822
|
+
"model_specs":[
|
|
6823
|
+
{
|
|
6824
|
+
"model_format":"pytorch",
|
|
6825
|
+
"model_size_in_billions":2,
|
|
6826
|
+
"quantizations":[
|
|
6827
|
+
"none"
|
|
6828
|
+
],
|
|
6829
|
+
"model_id":"Qwen/Qwen2-VL-2B-Instruct",
|
|
6830
|
+
"model_revision":"096da3b96240e3d66d35be0e5ccbe282eea8d6b1"
|
|
6831
|
+
},
|
|
6832
|
+
{
|
|
6833
|
+
"model_format":"gptq",
|
|
6834
|
+
"model_size_in_billions":2,
|
|
6835
|
+
"quantizations":[
|
|
6836
|
+
"Int8"
|
|
6837
|
+
],
|
|
6838
|
+
"model_id":"Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int8",
|
|
6839
|
+
"model_revision":"d15fb11857ccc566903e2e71341f9db7babb567b"
|
|
6840
|
+
},
|
|
6841
|
+
{
|
|
6842
|
+
"model_format":"gptq",
|
|
6843
|
+
"model_size_in_billions":2,
|
|
6844
|
+
"quantizations":[
|
|
6845
|
+
"Int4"
|
|
6846
|
+
],
|
|
6847
|
+
"model_id":"Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int4",
|
|
6848
|
+
"model_revision":"800d396518c82960ce6d231adecd07bbc474f0a9"
|
|
6849
|
+
},
|
|
6850
|
+
{
|
|
6851
|
+
"model_format":"awq",
|
|
6852
|
+
"model_size_in_billions":2,
|
|
6853
|
+
"quantizations":[
|
|
6854
|
+
"Int4"
|
|
6855
|
+
],
|
|
6856
|
+
"model_id":"Qwen/Qwen2-VL-2B-Instruct-AWQ",
|
|
6857
|
+
"model_revision":"ea8c5854c0044e28626719292de0d9b1a671f6fc"
|
|
6858
|
+
},
|
|
6859
|
+
{
|
|
6860
|
+
"model_format":"pytorch",
|
|
6861
|
+
"model_size_in_billions":7,
|
|
6862
|
+
"quantizations":[
|
|
6863
|
+
"none"
|
|
6864
|
+
],
|
|
6865
|
+
"model_id":"Qwen/Qwen2-VL-7B-Instruct",
|
|
6866
|
+
"model_revision":"6010982c1010c3b222fa98afc81575f124aa9bd6"
|
|
6867
|
+
},
|
|
6868
|
+
{
|
|
6869
|
+
"model_format":"gptq",
|
|
6870
|
+
"model_size_in_billions":7,
|
|
6871
|
+
"quantizations":[
|
|
6872
|
+
"Int8"
|
|
6873
|
+
],
|
|
6874
|
+
"model_id":"Qwen/Qwen2-VL-7B-Instruct-GPTQ-Int8",
|
|
6875
|
+
"model_revision":"3d152a77eaccfd72d59baedb0b183a1b8fd56e48"
|
|
6876
|
+
},
|
|
6877
|
+
{
|
|
6878
|
+
"model_format":"gptq",
|
|
6879
|
+
"model_size_in_billions":7,
|
|
6880
|
+
"quantizations":[
|
|
6881
|
+
"Int4"
|
|
6882
|
+
],
|
|
6883
|
+
"model_id":"Qwen/Qwen2-VL-7B-Instruct-GPTQ-Int4",
|
|
6884
|
+
"model_revision":"5ab897112fa83b9699826be8753ef9184585c77d"
|
|
6885
|
+
},
|
|
6886
|
+
{
|
|
6887
|
+
"model_format":"awq",
|
|
6888
|
+
"model_size_in_billions":7,
|
|
6889
|
+
"quantizations":[
|
|
6890
|
+
"Int4"
|
|
6891
|
+
],
|
|
6892
|
+
"model_id":"Qwen/Qwen2-VL-7B-Instruct-AWQ",
|
|
6893
|
+
"model_revision":"f94216e8b513933bccd567bcd9b7350199f32538"
|
|
6894
|
+
}
|
|
6895
|
+
],
|
|
6896
|
+
"prompt_style":{
|
|
6897
|
+
"style_name":"QWEN",
|
|
6898
|
+
"system_prompt":"You are a helpful assistant",
|
|
6899
|
+
"roles":[
|
|
6900
|
+
"user",
|
|
6901
|
+
"assistant"
|
|
7522
6902
|
],
|
|
7523
|
-
"
|
|
7524
|
-
|
|
7525
|
-
|
|
7526
|
-
|
|
6903
|
+
"stop": [
|
|
6904
|
+
"<|im_end|>",
|
|
6905
|
+
"<|endoftext|>"
|
|
6906
|
+
]
|
|
6907
|
+
}
|
|
6908
|
+
},
|
|
6909
|
+
{
|
|
6910
|
+
"version": 1,
|
|
6911
|
+
"context_length": 32768,
|
|
6912
|
+
"model_name": "minicpm3-4b",
|
|
6913
|
+
"model_lang": [
|
|
6914
|
+
"zh"
|
|
6915
|
+
],
|
|
6916
|
+
"model_ability": [
|
|
6917
|
+
"chat"
|
|
6918
|
+
],
|
|
6919
|
+
"model_description": "MiniCPM3-4B is the 3rd generation of MiniCPM series. The overall performance of MiniCPM3-4B surpasses Phi-3.5-mini-Instruct and GPT-3.5-Turbo-0125, being comparable with many recent 7B~9B models.",
|
|
6920
|
+
"model_specs": [
|
|
6921
|
+
{
|
|
6922
|
+
"model_format": "pytorch",
|
|
6923
|
+
"model_size_in_billions": 4,
|
|
6924
|
+
"quantizations": [
|
|
6925
|
+
"none"
|
|
6926
|
+
],
|
|
6927
|
+
"model_id": "openbmb/MiniCPM3-4B",
|
|
6928
|
+
"model_revision": "75f9f1097d9d66d11f37fff49210bf940455f8ac"
|
|
6929
|
+
},
|
|
6930
|
+
{
|
|
6931
|
+
"model_format": "gptq",
|
|
6932
|
+
"model_size_in_billions": 4,
|
|
6933
|
+
"quantizations": [
|
|
6934
|
+
"none"
|
|
6935
|
+
],
|
|
6936
|
+
"model_id": "openbmb/MiniCPM3-4B-GPTQ-Int4",
|
|
6937
|
+
"model_revision": "97a66a62f7d09c1ee35b087b42694716a8113dce"
|
|
6938
|
+
}
|
|
6939
|
+
],
|
|
6940
|
+
"chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
|
6941
|
+
"stop_token_ids": [
|
|
6942
|
+
1,
|
|
6943
|
+
2
|
|
6944
|
+
],
|
|
6945
|
+
"stop": [
|
|
6946
|
+
"<s>",
|
|
6947
|
+
"</s>"
|
|
6948
|
+
]
|
|
6949
|
+
},
|
|
6950
|
+
{
|
|
6951
|
+
"version":1,
|
|
6952
|
+
"context_length":32768,
|
|
6953
|
+
"model_name":"qwen2-audio-instruct",
|
|
6954
|
+
"model_lang":[
|
|
6955
|
+
"en",
|
|
6956
|
+
"zh"
|
|
6957
|
+
],
|
|
6958
|
+
"model_ability":[
|
|
6959
|
+
"chat",
|
|
6960
|
+
"audio"
|
|
6961
|
+
],
|
|
6962
|
+
"model_description":"Qwen2-Audio: A large-scale audio-language model which is capable of accepting various audio signal inputs and performing audio analysis or direct textual responses with regard to speech instructions.",
|
|
6963
|
+
"model_specs":[
|
|
6964
|
+
{
|
|
6965
|
+
"model_format":"pytorch",
|
|
6966
|
+
"model_size_in_billions":7,
|
|
6967
|
+
"quantizations":[
|
|
6968
|
+
"none"
|
|
6969
|
+
],
|
|
6970
|
+
"model_id":"Qwen/Qwen2-Audio-7B-Instruct",
|
|
6971
|
+
"model_revision":"bac62d2c6808845904c709c17a0402d817558c64"
|
|
6972
|
+
}
|
|
6973
|
+
],
|
|
6974
|
+
"prompt_style":{
|
|
6975
|
+
"style_name":"QWEN",
|
|
6976
|
+
"system_prompt":"You are a helpful assistant",
|
|
6977
|
+
"roles":[
|
|
6978
|
+
"user",
|
|
6979
|
+
"assistant"
|
|
6980
|
+
],
|
|
6981
|
+
"stop": [
|
|
6982
|
+
"<|im_end|>",
|
|
6983
|
+
"<|endoftext|>"
|
|
6984
|
+
]
|
|
6985
|
+
}
|
|
6986
|
+
},
|
|
6987
|
+
{
|
|
6988
|
+
"version":1,
|
|
6989
|
+
"context_length":32768,
|
|
6990
|
+
"model_name":"qwen2-audio",
|
|
6991
|
+
"model_lang":[
|
|
6992
|
+
"en",
|
|
6993
|
+
"zh"
|
|
6994
|
+
],
|
|
6995
|
+
"model_ability":[
|
|
6996
|
+
"chat",
|
|
6997
|
+
"audio"
|
|
6998
|
+
],
|
|
6999
|
+
"model_description":"Qwen2-Audio: A large-scale audio-language model which is capable of accepting various audio signal inputs and performing audio analysis or direct textual responses with regard to speech instructions.",
|
|
7000
|
+
"model_specs":[
|
|
7001
|
+
{
|
|
7002
|
+
"model_format":"pytorch",
|
|
7003
|
+
"model_size_in_billions":7,
|
|
7004
|
+
"quantizations":[
|
|
7005
|
+
"none"
|
|
7006
|
+
],
|
|
7007
|
+
"model_id":"Qwen/Qwen2-Audio-7B",
|
|
7008
|
+
"model_revision":"8577bc71d330c8fa32ffe9f8a1374100759f2466"
|
|
7009
|
+
}
|
|
7010
|
+
],
|
|
7011
|
+
"prompt_style":{
|
|
7012
|
+
"style_name":"QWEN",
|
|
7013
|
+
"system_prompt":"You are a helpful assistant",
|
|
7014
|
+
"roles":[
|
|
7015
|
+
"user",
|
|
7016
|
+
"assistant"
|
|
7527
7017
|
],
|
|
7528
7018
|
"stop": [
|
|
7529
|
-
"
|
|
7019
|
+
"<|im_end|>",
|
|
7020
|
+
"<|endoftext|>"
|
|
7530
7021
|
]
|
|
7531
7022
|
}
|
|
7023
|
+
},
|
|
7024
|
+
{
|
|
7025
|
+
"version": 1,
|
|
7026
|
+
"context_length": 128000,
|
|
7027
|
+
"model_name": "deepseek-v2",
|
|
7028
|
+
"model_lang": [
|
|
7029
|
+
"en",
|
|
7030
|
+
"zh"
|
|
7031
|
+
],
|
|
7032
|
+
"model_ability": [
|
|
7033
|
+
"generate"
|
|
7034
|
+
],
|
|
7035
|
+
"model_description": "DeepSeek-V2, a strong Mixture-of-Experts (MoE) language model characterized by economical training and efficient inference. ",
|
|
7036
|
+
"model_specs": [
|
|
7037
|
+
{
|
|
7038
|
+
"model_format": "pytorch",
|
|
7039
|
+
"model_size_in_billions": 16,
|
|
7040
|
+
"quantizations": [
|
|
7041
|
+
"4-bit",
|
|
7042
|
+
"8-bit",
|
|
7043
|
+
"none"
|
|
7044
|
+
],
|
|
7045
|
+
"model_id": "deepseek-ai/DeepSeek-V2-Lite",
|
|
7046
|
+
"model_revision": "604d5664dddd88a0433dbae533b7fe9472482de0"
|
|
7047
|
+
},
|
|
7048
|
+
{
|
|
7049
|
+
"model_format": "pytorch",
|
|
7050
|
+
"model_size_in_billions": 236,
|
|
7051
|
+
"quantizations": [
|
|
7052
|
+
"4-bit",
|
|
7053
|
+
"8-bit",
|
|
7054
|
+
"none"
|
|
7055
|
+
],
|
|
7056
|
+
"model_id": "deepseek-ai/DeepSeek-V2",
|
|
7057
|
+
"model_revision": "4461458f186c35188585855f28f77af5661ad489"
|
|
7058
|
+
}
|
|
7059
|
+
]
|
|
7060
|
+
},
|
|
7061
|
+
{
|
|
7062
|
+
"version": 1,
|
|
7063
|
+
"context_length": 128000,
|
|
7064
|
+
"model_name": "deepseek-v2-chat",
|
|
7065
|
+
"model_lang": [
|
|
7066
|
+
"en",
|
|
7067
|
+
"zh"
|
|
7068
|
+
],
|
|
7069
|
+
"model_ability": [
|
|
7070
|
+
"chat"
|
|
7071
|
+
],
|
|
7072
|
+
"model_description": "DeepSeek-V2, a strong Mixture-of-Experts (MoE) language model characterized by economical training and efficient inference. ",
|
|
7073
|
+
"model_specs": [
|
|
7074
|
+
{
|
|
7075
|
+
"model_format": "pytorch",
|
|
7076
|
+
"model_size_in_billions": 16,
|
|
7077
|
+
"quantizations": [
|
|
7078
|
+
"4-bit",
|
|
7079
|
+
"8-bit",
|
|
7080
|
+
"none"
|
|
7081
|
+
],
|
|
7082
|
+
"model_id": "deepseek-ai/DeepSeek-V2-Lite-Chat",
|
|
7083
|
+
"model_revision": "85864749cd611b4353ce1decdb286193298f64c7"
|
|
7084
|
+
},
|
|
7085
|
+
{
|
|
7086
|
+
"model_format": "pytorch",
|
|
7087
|
+
"model_size_in_billions": 236,
|
|
7088
|
+
"quantizations": [
|
|
7089
|
+
"4-bit",
|
|
7090
|
+
"8-bit",
|
|
7091
|
+
"none"
|
|
7092
|
+
],
|
|
7093
|
+
"model_id": "deepseek-ai/DeepSeek-V2-Chat",
|
|
7094
|
+
"model_revision": "8e3f5f6c2226787e41ba3e9283a06389d178c926"
|
|
7095
|
+
}
|
|
7096
|
+
],
|
|
7097
|
+
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{{ '<|begin▁of▁sentence|>' }}{% for message in messages %}{% if message['role'] == 'user' %}{{ 'User: ' + message['content'] + '\n\n' }}{% elif message['role'] == 'assistant' %}{{ 'Assistant: ' + message['content'] + '<|end▁of▁sentence|>' }}{% elif message['role'] == 'system' %}{{ message['content'] + '\n\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'Assistant:' }}{% endif %}",
|
|
7098
|
+
"stop_token_ids": [
|
|
7099
|
+
100001
|
|
7100
|
+
],
|
|
7101
|
+
"stop": [
|
|
7102
|
+
"<|end▁of▁sentence|>"
|
|
7103
|
+
]
|
|
7104
|
+
},
|
|
7105
|
+
{
|
|
7106
|
+
"version": 1,
|
|
7107
|
+
"context_length": 128000,
|
|
7108
|
+
"model_name": "deepseek-v2-chat-0628",
|
|
7109
|
+
"model_lang": [
|
|
7110
|
+
"en",
|
|
7111
|
+
"zh"
|
|
7112
|
+
],
|
|
7113
|
+
"model_ability": [
|
|
7114
|
+
"chat"
|
|
7115
|
+
],
|
|
7116
|
+
"model_description": "DeepSeek-V2-Chat-0628 is an improved version of DeepSeek-V2-Chat. ",
|
|
7117
|
+
"model_specs": [
|
|
7118
|
+
{
|
|
7119
|
+
"model_format": "pytorch",
|
|
7120
|
+
"model_size_in_billions": 236,
|
|
7121
|
+
"quantizations": [
|
|
7122
|
+
"4-bit",
|
|
7123
|
+
"8-bit",
|
|
7124
|
+
"none"
|
|
7125
|
+
],
|
|
7126
|
+
"model_id": "deepseek-ai/DeepSeek-V2-Chat-0628",
|
|
7127
|
+
"model_revision": "5d09e272c2b223830f4e84359cd9dd047a5d7c78"
|
|
7128
|
+
}
|
|
7129
|
+
],
|
|
7130
|
+
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{{ '<|begin▁of▁sentence|>' }}{% for message in messages %}{% if message['role'] == 'user' %}{{ '<|User|>' + message['content'] }}{% elif message['role'] == 'assistant' %}{{ '<|Assistant|>' + message['content'] + '<|end▁of▁sentence|>' }}{% elif message['role'] == 'system' %}{{ message['content'] + '\n\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|Assistant|>' }}{% endif %}",
|
|
7131
|
+
"stop_token_ids": [
|
|
7132
|
+
100001
|
|
7133
|
+
],
|
|
7134
|
+
"stop": [
|
|
7135
|
+
"<|end▁of▁sentence|>"
|
|
7136
|
+
]
|
|
7137
|
+
},
|
|
7138
|
+
{
|
|
7139
|
+
"version": 1,
|
|
7140
|
+
"context_length": 128000,
|
|
7141
|
+
"model_name": "deepseek-v2.5",
|
|
7142
|
+
"model_lang": [
|
|
7143
|
+
"en",
|
|
7144
|
+
"zh"
|
|
7145
|
+
],
|
|
7146
|
+
"model_ability": [
|
|
7147
|
+
"chat"
|
|
7148
|
+
],
|
|
7149
|
+
"model_description": "DeepSeek-V2.5 is an upgraded version that combines DeepSeek-V2-Chat and DeepSeek-Coder-V2-Instruct. The new model integrates the general and coding abilities of the two previous versions.",
|
|
7150
|
+
"model_specs": [
|
|
7151
|
+
{
|
|
7152
|
+
"model_format": "pytorch",
|
|
7153
|
+
"model_size_in_billions": 236,
|
|
7154
|
+
"quantizations": [
|
|
7155
|
+
"4-bit",
|
|
7156
|
+
"8-bit",
|
|
7157
|
+
"none"
|
|
7158
|
+
],
|
|
7159
|
+
"model_id": "deepseek-ai/DeepSeek-V2.5",
|
|
7160
|
+
"model_revision": "24b08cb750e0c2757de112d2e16327cb21ed4833"
|
|
7161
|
+
}
|
|
7162
|
+
],
|
|
7163
|
+
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set ns = namespace(is_first=false, is_tool=false, is_output_first=true, system_prompt='') %}{%- for message in messages %} {%- if message['role'] == 'system' %} {% set ns.system_prompt = message['content'] %} {%- endif %}{%- endfor %}{{'<|begin▁of▁sentence|>'}}{{ns.system_prompt}}{%- for message in messages %} {%- if message['role'] == 'user' %} {%- set ns.is_tool = false -%}{{'<|User|>' + message['content']}} {%- endif %} {%- if message['role'] == 'assistant' and message['content'] is none %} {%- set ns.is_tool = false -%} {%- for tool in message['tool_calls']%} {%- if not ns.is_first %}{{'<|Assistant|><|tool▁calls▁begin|><|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\\n' + '```json' + '\\n' + tool['function']['arguments'] + '\\n' + '```' + '<|tool▁call▁end|>'}} {%- set ns.is_first = true -%} {%- else %}{{'\\n' + '<|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\\n' + '```json' + '\\n' + tool['function']['arguments'] + '\\n' + '```' + '<|tool▁call▁end|>'}}{{'<|tool▁calls▁end|><|end▁of▁sentence|>'}} {%- endif %} {%- endfor %} {%- endif %} {%- if message['role'] == 'assistant' and message['content'] is not none %} {%- if ns.is_tool %}{{'<|tool▁outputs▁end|>' + message['content'] + '<|end▁of▁sentence|>'}} {%- set ns.is_tool = false -%} {%- else %}{{'<|Assistant|>' + message['content'] + '<|end▁of▁sentence|>'}} {%- endif %} {%- endif %} {%- if message['role'] == 'tool' %} {%- set ns.is_tool = true -%} {%- if ns.is_output_first %}{{'<|tool▁outputs▁begin|><|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}} {%- set ns.is_output_first = false %} {%- else %}{{'\\n<|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}} {%- endif %} {%- endif %}{%- endfor -%}{% if ns.is_tool %}{{'<|tool▁outputs▁end|>'}}{% endif %}{% if add_generation_prompt and not ns.is_tool %}{{'<|Assistant|>'}}{% endif %}",
|
|
7164
|
+
"stop_token_ids": [
|
|
7165
|
+
100001
|
|
7166
|
+
],
|
|
7167
|
+
"stop": [
|
|
7168
|
+
"<|end▁of▁sentence|>"
|
|
7169
|
+
]
|
|
7170
|
+
},
|
|
7171
|
+
{
|
|
7172
|
+
"version": 1,
|
|
7173
|
+
"context_length": 131072,
|
|
7174
|
+
"model_name": "yi-coder-chat",
|
|
7175
|
+
"model_lang": [
|
|
7176
|
+
"en"
|
|
7177
|
+
],
|
|
7178
|
+
"model_ability": [
|
|
7179
|
+
"chat"
|
|
7180
|
+
],
|
|
7181
|
+
"model_description": "Yi-Coder is a series of open-source code language models that delivers state-of-the-art coding performance with fewer than 10 billion parameters.Excelling in long-context understanding with a maximum context length of 128K tokens.Supporting 52 major programming languages, including popular ones such as Java, Python, JavaScript, and C++.",
|
|
7182
|
+
"model_specs": [
|
|
7183
|
+
{
|
|
7184
|
+
"model_format": "pytorch",
|
|
7185
|
+
"model_size_in_billions": 9,
|
|
7186
|
+
"quantizations": [
|
|
7187
|
+
"none"
|
|
7188
|
+
],
|
|
7189
|
+
"model_id": "01ai/Yi-Coder-9B-Chat",
|
|
7190
|
+
"model_revision": "356a1f8d4e4a606d0b879e54191ca809918576b8"
|
|
7191
|
+
},
|
|
7192
|
+
{
|
|
7193
|
+
"model_format": "pytorch",
|
|
7194
|
+
"model_size_in_billions": "1_5",
|
|
7195
|
+
"quantizations": [
|
|
7196
|
+
"none"
|
|
7197
|
+
],
|
|
7198
|
+
"model_id": "01ai/Yi-Coder-1.5B-Chat",
|
|
7199
|
+
"model_revision": "92fdd1b2f1539ac990e7f4a921db5601da2f0299"
|
|
7200
|
+
}
|
|
7201
|
+
],
|
|
7202
|
+
"chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ '<|im_start|>system\n' + system_message + '<|im_end|>\n' }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\n' + content + '<|im_end|>\n<|im_start|>assistant\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\n' }}{% endif %}{% endfor %}",
|
|
7203
|
+
"stop_token_ids": [
|
|
7204
|
+
1,
|
|
7205
|
+
2,
|
|
7206
|
+
6,
|
|
7207
|
+
7
|
|
7208
|
+
],
|
|
7209
|
+
"stop": [
|
|
7210
|
+
"<|startoftext|>",
|
|
7211
|
+
"<|endoftext|>",
|
|
7212
|
+
"<|im_start|>",
|
|
7213
|
+
"<|im_end|>"
|
|
7214
|
+
]
|
|
7215
|
+
},
|
|
7216
|
+
{
|
|
7217
|
+
"version": 1,
|
|
7218
|
+
"context_length": 131072,
|
|
7219
|
+
"model_name": "yi-coder",
|
|
7220
|
+
"model_lang": [
|
|
7221
|
+
"en"
|
|
7222
|
+
],
|
|
7223
|
+
"model_ability": [
|
|
7224
|
+
"generate"
|
|
7225
|
+
],
|
|
7226
|
+
"model_description": "Yi-Coder is a series of open-source code language models that delivers state-of-the-art coding performance with fewer than 10 billion parameters.Excelling in long-context understanding with a maximum context length of 128K tokens.Supporting 52 major programming languages, including popular ones such as Java, Python, JavaScript, and C++.",
|
|
7227
|
+
"model_specs": [
|
|
7228
|
+
{
|
|
7229
|
+
"model_format": "pytorch",
|
|
7230
|
+
"model_size_in_billions": 9,
|
|
7231
|
+
"quantizations": [
|
|
7232
|
+
"none"
|
|
7233
|
+
],
|
|
7234
|
+
"model_id": "01-ai/Yi-Coder-9B",
|
|
7235
|
+
"model_revision": "e20f8087a9507ac8bce409dc5db5d0c608124238"
|
|
7236
|
+
},
|
|
7237
|
+
{
|
|
7238
|
+
"model_format": "pytorch",
|
|
7239
|
+
"model_size_in_billions": "1_5",
|
|
7240
|
+
"quantizations": [
|
|
7241
|
+
"none"
|
|
7242
|
+
],
|
|
7243
|
+
"model_id": "01-ai/Yi-Coder-1.5B",
|
|
7244
|
+
"model_revision": "00e59e64f47d3c78e4cfbdd345888479797e8109"
|
|
7245
|
+
}
|
|
7246
|
+
]
|
|
7532
7247
|
}
|
|
7533
7248
|
]
|