@lobehub/chat 1.65.1 → 1.66.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +51 -0
- package/changelog/v1.json +18 -0
- package/docker-compose/local/docker-compose.yml +14 -0
- package/docker-compose/local/searxng-settings.yml +2582 -0
- package/docker-compose/setup.sh +3 -1
- package/docs/self-hosting/advanced/model-list.mdx +4 -2
- package/docs/self-hosting/advanced/model-list.zh-CN.mdx +4 -2
- package/package.json +7 -7
- package/src/config/aiModels/ai360.ts +4 -0
- package/src/config/aiModels/baichuan.ts +16 -0
- package/src/config/aiModels/google.ts +3 -3
- package/src/config/aiModels/groq.ts +10 -0
- package/src/config/aiModels/hunyuan.ts +40 -0
- package/src/config/aiModels/moonshot.ts +32 -0
- package/src/config/aiModels/qwen.ts +43 -26
- package/src/config/aiModels/spark.ts +30 -0
- package/src/config/aiModels/stepfun.ts +44 -0
- package/src/config/aiModels/zhipu.ts +36 -0
- package/src/features/Conversation/Messages/Assistant/Tool/Render/CustomRender.tsx +7 -7
- package/src/features/MobileSwitchLoading/index.tsx +0 -1
- package/src/libs/agent-runtime/ai360/index.ts +15 -2
- package/src/libs/agent-runtime/baichuan/index.ts +14 -4
- package/src/libs/agent-runtime/hunyuan/index.ts +18 -0
- package/src/libs/agent-runtime/minimax/index.ts +2 -11
- package/src/libs/agent-runtime/moonshot/index.ts +13 -4
- package/src/libs/agent-runtime/qwen/index.ts +4 -0
- package/src/libs/agent-runtime/stepfun/index.ts +15 -2
- package/src/libs/agent-runtime/zhipu/index.ts +19 -7
- package/src/utils/parseModels.test.ts +14 -0
- package/src/utils/parseModels.ts +4 -0
package/docker-compose/setup.sh
CHANGED
@@ -394,6 +394,7 @@ SUB_DIR="docker-compose/local"
|
|
394
394
|
FILES=(
|
395
395
|
"$SUB_DIR/docker-compose.yml"
|
396
396
|
"$SUB_DIR/init_data.json"
|
397
|
+
"$SUB_DIR/searxng-settings.yml"
|
397
398
|
)
|
398
399
|
ENV_EXAMPLES=(
|
399
400
|
"$SUB_DIR/.env.zh-CN.example"
|
@@ -434,6 +435,7 @@ section_download_files(){
|
|
434
435
|
|
435
436
|
download_file "$SOURCE_URL/${FILES[0]}" "docker-compose.yml"
|
436
437
|
download_file "$SOURCE_URL/${FILES[1]}" "init_data.json"
|
438
|
+
download_file "$SOURCE_URL/${FILES[2]}" "searxng-settings.yml"
|
437
439
|
|
438
440
|
# Download .env.example with the specified language
|
439
441
|
if [ "$LANGUAGE" = "zh_CN" ]; then
|
@@ -657,4 +659,4 @@ section_display_configurated_report() {
|
|
657
659
|
printf "\n%s" "$(show_message "tips_show_documentation")"
|
658
660
|
printf "%s\n" $(show_message "tips_show_documentation_url")
|
659
661
|
}
|
660
|
-
section_display_configurated_report
|
662
|
+
section_display_configurated_report
|
@@ -17,7 +17,7 @@ LobeChat supports customizing the model list during deployment. This configurati
|
|
17
17
|
You can use `+` to add a model, `-` to hide a model, and use `model name=display name<extension configuration>` to customize the display name of a model, separated by English commas. The basic syntax is as follows:
|
18
18
|
|
19
19
|
```text
|
20
|
-
id=displayName<maxToken:vision:reasoning:fc:file>,model2,model3
|
20
|
+
id=displayName<maxToken:vision:reasoning:search:fc:file>,model2,model3
|
21
21
|
```
|
22
22
|
|
23
23
|
For example: `+qwen-7b-chat,+glm-6b,-gpt-3.5-turbo,gpt-4-0125-preview=gpt-4-turbo`
|
@@ -29,7 +29,7 @@ In the above example, it adds `qwen-7b-chat` and `glm-6b` to the model list, rem
|
|
29
29
|
Considering the diversity of model capabilities, we started to add extension configuration in version `0.147.8`, with the following rules:
|
30
30
|
|
31
31
|
```shell
|
32
|
-
id=displayName<maxToken:vision:reasoning:fc:file>
|
32
|
+
id=displayName<maxToken:vision:reasoning:search:fc:file>
|
33
33
|
```
|
34
34
|
|
35
35
|
The first value in angle brackets is designated as the `maxToken` for this model. The second value and beyond are the model's extension capabilities, separated by colons `:`, and the order is not important.
|
@@ -40,6 +40,7 @@ Examples are as follows:
|
|
40
40
|
- `spark-v3.5=讯飞星火 v3.5<8192:fc>`: Xunfei Spark 3.5 model, maximum context of 8k, supports Function Call;
|
41
41
|
- `gemini-1.5-flash-latest=Gemini 1.5 Flash<16000:vision>`: Google Vision model, maximum context of 16k, supports image recognition;
|
42
42
|
- `o3-mini=OpenAI o3-mini<200000:reasoning:fc>`: OpenAI o3-mini model, maximum context of 200k, supports reasoning and Function Call;
|
43
|
+
- `qwen-max-latest=Qwen Max<32768:search:fc>`: Qwen 2.5 Max model, maximum context of 32k, supports web search and Function Call;
|
43
44
|
- `gpt-4-all=ChatGPT Plus<128000:fc:vision:file>`, hacked version of ChatGPT Plus web, context of 128k, supports image recognition, Function Call, file upload.
|
44
45
|
|
45
46
|
Currently supported extension capabilities are:
|
@@ -49,4 +50,5 @@ Currently supported extension capabilities are:
|
|
49
50
|
| `fc` | Function Calling |
|
50
51
|
| `vision` | Image Recognition |
|
51
52
|
| `reasoning` | Support Reasoning |
|
53
|
+
| `search` | Support Web Search |
|
52
54
|
| `file` | File Upload (a bit hacky, not recommended for daily use) |
|
@@ -16,7 +16,7 @@ LobeChat 支持在部署时自定义模型列表,详情请参考 [模型提供
|
|
16
16
|
你可以使用 `+` 增加一个模型,使用 `-` 来隐藏一个模型,使用 `模型名=展示名<扩展配置>` 来自定义模型的展示名,用英文逗号隔开。通过 `<>` 来添加扩展配置。基本语法如下:
|
17
17
|
|
18
18
|
```text
|
19
|
-
id=displayName<maxToken:vision:reasoning:fc:file>,model2,model3
|
19
|
+
id=displayName<maxToken:vision:reasoning:search:fc:file>,model2,model3
|
20
20
|
```
|
21
21
|
|
22
22
|
例如: `+qwen-7b-chat,+glm-6b,-gpt-3.5-turbo,gpt-4-0125-preview=gpt-4-turbo`
|
@@ -28,7 +28,7 @@ id=displayName<maxToken:vision:reasoning:fc:file>,model2,model3
|
|
28
28
|
考虑到模型的能力多样性,我们在 `0.147.8` 版本开始增加扩展性配置,它的规则如下:
|
29
29
|
|
30
30
|
```shell
|
31
|
-
id=displayName<maxToken:vision:reasoning:fc:file>
|
31
|
+
id=displayName<maxToken:vision:reasoning:search:fc:file>
|
32
32
|
```
|
33
33
|
|
34
34
|
尖括号第一个值约定为这个模型的 `maxToken` 。第二个及以后作为模型的扩展能力,能力与能力之间用冒号 `:` 作为分隔符,顺序不重要。
|
@@ -39,6 +39,7 @@ id=displayName<maxToken:vision:reasoning:fc:file>
|
|
39
39
|
- `spark-v3.5=讯飞星火 v3.5<8192:fc>`:讯飞星火 3.5 模型,最大上下文 8k,支持 Function Call;
|
40
40
|
- `gemini-1.5-flash-latest=Gemini 1.5 Flash<16000:vision>`:Google 视觉模型,最大上下文 16k,支持图像识别;
|
41
41
|
- `o3-mini=OpenAI o3-mini<200000:reasoning:fc>`:OpenAI o3-mini 模型,最大上下文 200k,支持推理及 Function Call;
|
42
|
+
- `qwen-max-latest=Qwen Max<32768:search:fc>`:通义千问 2.5 Max 模型,最大上下文 32k,支持联网搜索及 Function Call;
|
42
43
|
- `gpt-4-all=ChatGPT Plus<128000:fc:vision:file>`,hack 的 ChatGPT Plus 网页版,上下 128k ,支持图像识别、Function Call、文件上传
|
43
44
|
|
44
45
|
目前支持的扩展能力有:
|
@@ -48,4 +49,5 @@ id=displayName<maxToken:vision:reasoning:fc:file>
|
|
48
49
|
| `fc` | 函数调用(function calling) |
|
49
50
|
| `vision` | 视觉识别 |
|
50
51
|
| `reasoning` | 支持推理 |
|
52
|
+
| `search` | 支持联网搜索 |
|
51
53
|
| `file` | 文件上传(比较 hack,不建议日常使用) |
|
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@lobehub/chat",
|
3
|
-
"version": "1.
|
3
|
+
"version": "1.66.0",
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
5
5
|
"keywords": [
|
6
6
|
"framework",
|
@@ -131,7 +131,7 @@
|
|
131
131
|
"@lobehub/tts": "^1.28.0",
|
132
132
|
"@lobehub/ui": "^1.165.2",
|
133
133
|
"@neondatabase/serverless": "^0.10.4",
|
134
|
-
"@next/third-parties": "^15.
|
134
|
+
"@next/third-parties": "^15.2.0",
|
135
135
|
"@react-spring/web": "^9.7.5",
|
136
136
|
"@sentry/nextjs": "^7.120.2",
|
137
137
|
"@serwist/next": "^9.0.11",
|
@@ -179,7 +179,7 @@
|
|
179
179
|
"mdast-util-to-markdown": "^2.1.2",
|
180
180
|
"modern-screenshot": "^4.5.5",
|
181
181
|
"nanoid": "^5.0.9",
|
182
|
-
"next": "^15.
|
182
|
+
"next": "^15.2.0",
|
183
183
|
"next-auth": "beta",
|
184
184
|
"next-mdx-remote": "^5.0.0",
|
185
185
|
"nextjs-toploader": "^3.7.15",
|
@@ -250,8 +250,8 @@
|
|
250
250
|
"@lobehub/i18n-cli": "^1.20.3",
|
251
251
|
"@lobehub/lint": "^1.25.5",
|
252
252
|
"@lobehub/seo-cli": "^1.4.3",
|
253
|
-
"@next/bundle-analyzer": "^15.
|
254
|
-
"@next/eslint-plugin-next": "^15.
|
253
|
+
"@next/bundle-analyzer": "^15.2.0",
|
254
|
+
"@next/eslint-plugin-next": "^15.2.0",
|
255
255
|
"@peculiar/webcrypto": "^1.5.0",
|
256
256
|
"@semantic-release/exec": "^6.0.3",
|
257
257
|
"@testing-library/jest-dom": "^6.6.3",
|
@@ -269,8 +269,8 @@
|
|
269
269
|
"@types/node": "^22.10.5",
|
270
270
|
"@types/numeral": "^2.0.5",
|
271
271
|
"@types/pg": "^8.11.10",
|
272
|
-
"@types/react": "^19.0.
|
273
|
-
"@types/react-dom": "^19.0.
|
272
|
+
"@types/react": "^19.0.10",
|
273
|
+
"@types/react-dom": "^19.0.4",
|
274
274
|
"@types/rtl-detect": "^1.0.3",
|
275
275
|
"@types/semver": "^7.5.8",
|
276
276
|
"@types/systemjs": "^6.15.1",
|
@@ -52,6 +52,7 @@ const ai360ChatModels: AIChatModelCard[] = [
|
|
52
52
|
{
|
53
53
|
abilities: {
|
54
54
|
functionCall: true,
|
55
|
+
search: true,
|
55
56
|
},
|
56
57
|
contextWindowTokens: 8000,
|
57
58
|
description:
|
@@ -64,6 +65,9 @@ const ai360ChatModels: AIChatModelCard[] = [
|
|
64
65
|
input: 2,
|
65
66
|
output: 5,
|
66
67
|
},
|
68
|
+
settings: {
|
69
|
+
searchImpl: 'params',
|
70
|
+
},
|
67
71
|
type: 'chat',
|
68
72
|
},
|
69
73
|
{
|
@@ -4,6 +4,7 @@ const baichuanChatModels: AIChatModelCard[] = [
|
|
4
4
|
{
|
5
5
|
abilities: {
|
6
6
|
functionCall: true,
|
7
|
+
search: true,
|
7
8
|
},
|
8
9
|
contextWindowTokens: 32_768,
|
9
10
|
description:
|
@@ -17,11 +18,15 @@ const baichuanChatModels: AIChatModelCard[] = [
|
|
17
18
|
input: 100,
|
18
19
|
output: 100,
|
19
20
|
},
|
21
|
+
settings: {
|
22
|
+
searchImpl: 'params',
|
23
|
+
},
|
20
24
|
type: 'chat',
|
21
25
|
},
|
22
26
|
{
|
23
27
|
abilities: {
|
24
28
|
functionCall: true,
|
29
|
+
search: true,
|
25
30
|
},
|
26
31
|
contextWindowTokens: 32_768,
|
27
32
|
description:
|
@@ -35,11 +40,15 @@ const baichuanChatModels: AIChatModelCard[] = [
|
|
35
40
|
input: 15,
|
36
41
|
output: 15,
|
37
42
|
},
|
43
|
+
settings: {
|
44
|
+
searchImpl: 'params',
|
45
|
+
},
|
38
46
|
type: 'chat',
|
39
47
|
},
|
40
48
|
{
|
41
49
|
abilities: {
|
42
50
|
functionCall: true,
|
51
|
+
search: true,
|
43
52
|
},
|
44
53
|
contextWindowTokens: 32_768,
|
45
54
|
description:
|
@@ -53,11 +62,15 @@ const baichuanChatModels: AIChatModelCard[] = [
|
|
53
62
|
input: 0.98,
|
54
63
|
output: 0.98,
|
55
64
|
},
|
65
|
+
settings: {
|
66
|
+
searchImpl: 'params',
|
67
|
+
},
|
56
68
|
type: 'chat',
|
57
69
|
},
|
58
70
|
{
|
59
71
|
abilities: {
|
60
72
|
functionCall: true,
|
73
|
+
search: true,
|
61
74
|
},
|
62
75
|
contextWindowTokens: 32_768,
|
63
76
|
description:
|
@@ -70,6 +83,9 @@ const baichuanChatModels: AIChatModelCard[] = [
|
|
70
83
|
input: 12,
|
71
84
|
output: 12,
|
72
85
|
},
|
86
|
+
settings: {
|
87
|
+
searchImpl: 'params',
|
88
|
+
},
|
73
89
|
type: 'chat',
|
74
90
|
},
|
75
91
|
{
|
@@ -80,9 +80,9 @@ const googleChatModels: AIChatModelCard[] = [
|
|
80
80
|
vision: true,
|
81
81
|
},
|
82
82
|
contextWindowTokens: 1_048_576 + 8192,
|
83
|
-
description: '
|
84
|
-
displayName: 'Gemini 2.0 Flash-Lite
|
85
|
-
id: 'gemini-2.0-flash-lite-
|
83
|
+
description: 'Gemini 2.0 Flash 模型变体,针对成本效益和低延迟等目标进行了优化。',
|
84
|
+
displayName: 'Gemini 2.0 Flash-Lite 001',
|
85
|
+
id: 'gemini-2.0-flash-lite-001',
|
86
86
|
maxOutput: 8192,
|
87
87
|
pricing: {
|
88
88
|
cachedInput: 0.018_75,
|
@@ -176,6 +176,16 @@ const groqChatModels: AIChatModelCard[] = [
|
|
176
176
|
},
|
177
177
|
type: 'chat',
|
178
178
|
},
|
179
|
+
{
|
180
|
+
contextWindowTokens: 32_768,
|
181
|
+
displayName: 'Mixtral Saba 24B',
|
182
|
+
id: 'mistral-saba-24b',
|
183
|
+
pricing: {
|
184
|
+
input: 0.79,
|
185
|
+
output: 0.79,
|
186
|
+
},
|
187
|
+
type: 'chat',
|
188
|
+
},
|
179
189
|
{
|
180
190
|
abilities: {
|
181
191
|
functionCall: true,
|
@@ -20,6 +20,9 @@ const hunyuanChatModels: AIChatModelCard[] = [
|
|
20
20
|
type: 'chat',
|
21
21
|
},
|
22
22
|
{
|
23
|
+
abilities: {
|
24
|
+
search: true,
|
25
|
+
},
|
23
26
|
contextWindowTokens: 32_000,
|
24
27
|
description:
|
25
28
|
'采用更优的路由策略,同时缓解了负载均衡和专家趋同的问题。长文方面,大海捞针指标达到99.9%。MOE-32K 性价比相对更高,在平衡效果、价格的同时,可对实现对长文本输入的处理。',
|
@@ -33,9 +36,15 @@ const hunyuanChatModels: AIChatModelCard[] = [
|
|
33
36
|
output: 2,
|
34
37
|
},
|
35
38
|
releasedAt: '2025-02-10',
|
39
|
+
settings: {
|
40
|
+
searchImpl: 'params',
|
41
|
+
},
|
36
42
|
type: 'chat',
|
37
43
|
},
|
38
44
|
{
|
45
|
+
abilities: {
|
46
|
+
search: true,
|
47
|
+
},
|
39
48
|
contextWindowTokens: 256_000,
|
40
49
|
description:
|
41
50
|
'采用更优的路由策略,同时缓解了负载均衡和专家趋同的问题。长文方面,大海捞针指标达到99.9%。MOE-256K 在长度和效果上进一步突破,极大的扩展了可输入长度。',
|
@@ -49,11 +58,15 @@ const hunyuanChatModels: AIChatModelCard[] = [
|
|
49
58
|
output: 2,
|
50
59
|
},
|
51
60
|
releasedAt: '2025-02-10',
|
61
|
+
settings: {
|
62
|
+
searchImpl: 'params',
|
63
|
+
},
|
52
64
|
type: 'chat',
|
53
65
|
},
|
54
66
|
{
|
55
67
|
abilities: {
|
56
68
|
functionCall: true,
|
69
|
+
search: true,
|
57
70
|
},
|
58
71
|
contextWindowTokens: 32_000,
|
59
72
|
description:
|
@@ -68,11 +81,15 @@ const hunyuanChatModels: AIChatModelCard[] = [
|
|
68
81
|
output: 50,
|
69
82
|
},
|
70
83
|
releasedAt: '2025-01-10',
|
84
|
+
settings: {
|
85
|
+
searchImpl: 'params',
|
86
|
+
},
|
71
87
|
type: 'chat',
|
72
88
|
},
|
73
89
|
{
|
74
90
|
abilities: {
|
75
91
|
functionCall: true,
|
92
|
+
search: true,
|
76
93
|
},
|
77
94
|
contextWindowTokens: 32_000,
|
78
95
|
description:
|
@@ -86,11 +103,15 @@ const hunyuanChatModels: AIChatModelCard[] = [
|
|
86
103
|
output: 50,
|
87
104
|
},
|
88
105
|
releasedAt: '2025-01-10',
|
106
|
+
settings: {
|
107
|
+
searchImpl: 'params',
|
108
|
+
},
|
89
109
|
type: 'chat',
|
90
110
|
},
|
91
111
|
{
|
92
112
|
abilities: {
|
93
113
|
functionCall: true,
|
114
|
+
search: true,
|
94
115
|
},
|
95
116
|
contextWindowTokens: 32_000,
|
96
117
|
description:
|
@@ -104,11 +125,15 @@ const hunyuanChatModels: AIChatModelCard[] = [
|
|
104
125
|
output: 50,
|
105
126
|
},
|
106
127
|
releasedAt: '2025-01-10',
|
128
|
+
settings: {
|
129
|
+
searchImpl: 'params',
|
130
|
+
},
|
107
131
|
type: 'chat',
|
108
132
|
},
|
109
133
|
{
|
110
134
|
abilities: {
|
111
135
|
functionCall: true,
|
136
|
+
search: true,
|
112
137
|
},
|
113
138
|
contextWindowTokens: 32_000,
|
114
139
|
description:
|
@@ -122,9 +147,15 @@ const hunyuanChatModels: AIChatModelCard[] = [
|
|
122
147
|
output: 50,
|
123
148
|
},
|
124
149
|
releasedAt: '2024-11-20',
|
150
|
+
settings: {
|
151
|
+
searchImpl: 'params',
|
152
|
+
},
|
125
153
|
type: 'chat',
|
126
154
|
},
|
127
155
|
{
|
156
|
+
abilities: {
|
157
|
+
search: true,
|
158
|
+
},
|
128
159
|
contextWindowTokens: 32_000,
|
129
160
|
description:
|
130
161
|
'Hunyuan-large 模型总参数量约 389B,激活参数量约 52B,是当前业界参数规模最大、效果最好的 Transformer 架构的开源 MoE 模型。',
|
@@ -138,9 +169,15 @@ const hunyuanChatModels: AIChatModelCard[] = [
|
|
138
169
|
output: 12,
|
139
170
|
},
|
140
171
|
releasedAt: '2025-02-10',
|
172
|
+
settings: {
|
173
|
+
searchImpl: 'params',
|
174
|
+
},
|
141
175
|
type: 'chat',
|
142
176
|
},
|
143
177
|
{
|
178
|
+
abilities: {
|
179
|
+
search: true,
|
180
|
+
},
|
144
181
|
contextWindowTokens: 134_000,
|
145
182
|
description:
|
146
183
|
'擅长处理长文任务如文档摘要和文档问答等,同时也具备处理通用文本生成任务的能力。在长文本的分析和生成上表现优异,能有效应对复杂和详尽的长文内容处理需求。',
|
@@ -154,6 +191,9 @@ const hunyuanChatModels: AIChatModelCard[] = [
|
|
154
191
|
output: 18,
|
155
192
|
},
|
156
193
|
releasedAt: '2024-12-18',
|
194
|
+
settings: {
|
195
|
+
searchImpl: 'params',
|
196
|
+
},
|
157
197
|
type: 'chat',
|
158
198
|
},
|
159
199
|
{
|
@@ -4,6 +4,7 @@ const moonshotChatModels: AIChatModelCard[] = [
|
|
4
4
|
{
|
5
5
|
abilities: {
|
6
6
|
functionCall: true,
|
7
|
+
search: true,
|
7
8
|
vision: true,
|
8
9
|
},
|
9
10
|
contextWindowTokens: 131_072,
|
@@ -19,11 +20,15 @@ const moonshotChatModels: AIChatModelCard[] = [
|
|
19
20
|
output: 60,
|
20
21
|
},
|
21
22
|
releasedAt: '2025-02-17',
|
23
|
+
settings: {
|
24
|
+
searchImpl: 'params',
|
25
|
+
},
|
22
26
|
type: 'chat',
|
23
27
|
},
|
24
28
|
{
|
25
29
|
abilities: {
|
26
30
|
functionCall: true,
|
31
|
+
search: true,
|
27
32
|
},
|
28
33
|
contextWindowTokens: 131_072,
|
29
34
|
description:
|
@@ -35,11 +40,15 @@ const moonshotChatModels: AIChatModelCard[] = [
|
|
35
40
|
input: 60,
|
36
41
|
output: 60,
|
37
42
|
},
|
43
|
+
settings: {
|
44
|
+
searchImpl: 'params',
|
45
|
+
},
|
38
46
|
type: 'chat',
|
39
47
|
},
|
40
48
|
{
|
41
49
|
abilities: {
|
42
50
|
functionCall: true,
|
51
|
+
search: true,
|
43
52
|
},
|
44
53
|
contextWindowTokens: 8192,
|
45
54
|
description:
|
@@ -51,11 +60,15 @@ const moonshotChatModels: AIChatModelCard[] = [
|
|
51
60
|
input: 12,
|
52
61
|
output: 12,
|
53
62
|
},
|
63
|
+
settings: {
|
64
|
+
searchImpl: 'params',
|
65
|
+
},
|
54
66
|
type: 'chat',
|
55
67
|
},
|
56
68
|
{
|
57
69
|
abilities: {
|
58
70
|
functionCall: true,
|
71
|
+
search: true,
|
59
72
|
},
|
60
73
|
contextWindowTokens: 32_768,
|
61
74
|
description:
|
@@ -67,11 +80,15 @@ const moonshotChatModels: AIChatModelCard[] = [
|
|
67
80
|
input: 24,
|
68
81
|
output: 24,
|
69
82
|
},
|
83
|
+
settings: {
|
84
|
+
searchImpl: 'params',
|
85
|
+
},
|
70
86
|
type: 'chat',
|
71
87
|
},
|
72
88
|
{
|
73
89
|
abilities: {
|
74
90
|
functionCall: true,
|
91
|
+
search: true,
|
75
92
|
},
|
76
93
|
contextWindowTokens: 131_072,
|
77
94
|
description:
|
@@ -83,11 +100,15 @@ const moonshotChatModels: AIChatModelCard[] = [
|
|
83
100
|
input: 60,
|
84
101
|
output: 60,
|
85
102
|
},
|
103
|
+
settings: {
|
104
|
+
searchImpl: 'params',
|
105
|
+
},
|
86
106
|
type: 'chat',
|
87
107
|
},
|
88
108
|
{
|
89
109
|
abilities: {
|
90
110
|
functionCall: true,
|
111
|
+
search: true,
|
91
112
|
vision: true,
|
92
113
|
},
|
93
114
|
contextWindowTokens: 8192,
|
@@ -101,11 +122,15 @@ const moonshotChatModels: AIChatModelCard[] = [
|
|
101
122
|
output: 12,
|
102
123
|
},
|
103
124
|
releasedAt: '2025-01-14',
|
125
|
+
settings: {
|
126
|
+
searchImpl: 'params',
|
127
|
+
},
|
104
128
|
type: 'chat',
|
105
129
|
},
|
106
130
|
{
|
107
131
|
abilities: {
|
108
132
|
functionCall: true,
|
133
|
+
search: true,
|
109
134
|
vision: true,
|
110
135
|
},
|
111
136
|
contextWindowTokens: 32_768,
|
@@ -119,11 +144,15 @@ const moonshotChatModels: AIChatModelCard[] = [
|
|
119
144
|
output: 24,
|
120
145
|
},
|
121
146
|
releasedAt: '2025-01-14',
|
147
|
+
settings: {
|
148
|
+
searchImpl: 'params',
|
149
|
+
},
|
122
150
|
type: 'chat',
|
123
151
|
},
|
124
152
|
{
|
125
153
|
abilities: {
|
126
154
|
functionCall: true,
|
155
|
+
search: true,
|
127
156
|
vision: true,
|
128
157
|
},
|
129
158
|
contextWindowTokens: 131_072,
|
@@ -137,6 +166,9 @@ const moonshotChatModels: AIChatModelCard[] = [
|
|
137
166
|
output: 60,
|
138
167
|
},
|
139
168
|
releasedAt: '2025-01-14',
|
169
|
+
settings: {
|
170
|
+
searchImpl: 'params',
|
171
|
+
},
|
140
172
|
type: 'chat',
|
141
173
|
},
|
142
174
|
];
|
@@ -62,8 +62,8 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
62
62
|
organization: 'Qwen',
|
63
63
|
pricing: {
|
64
64
|
currency: 'CNY',
|
65
|
-
input:
|
66
|
-
output:
|
65
|
+
input: 2.4,
|
66
|
+
output: 9.6,
|
67
67
|
},
|
68
68
|
settings: {
|
69
69
|
searchImpl: 'params',
|
@@ -410,6 +410,23 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
410
410
|
},
|
411
411
|
type: 'chat',
|
412
412
|
},
|
413
|
+
{
|
414
|
+
abilities: {
|
415
|
+
vision: true,
|
416
|
+
},
|
417
|
+
contextWindowTokens: 32_768,
|
418
|
+
description: 'Qwen-Omni 系列模型支持输入多种模态的数据,包括视频、音频、图片、文本,并输出音频与文本。',
|
419
|
+
displayName: 'Qwen Omni Turbo',
|
420
|
+
id: 'qwen-omni-turbo-latest',
|
421
|
+
maxOutput: 2048,
|
422
|
+
organization: 'Qwen',
|
423
|
+
pricing: {
|
424
|
+
currency: 'CNY',
|
425
|
+
input: 0,
|
426
|
+
output: 0,
|
427
|
+
},
|
428
|
+
type: 'chat',
|
429
|
+
},
|
413
430
|
{
|
414
431
|
abilities: {
|
415
432
|
vision: true,
|
@@ -452,7 +469,7 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
452
469
|
abilities: {
|
453
470
|
reasoning: true,
|
454
471
|
},
|
455
|
-
contextWindowTokens:
|
472
|
+
contextWindowTokens: 65_792,
|
456
473
|
description:
|
457
474
|
'DeepSeek-R1 在后训练阶段大规模使用了强化学习技术,在仅有极少标注数据的情况下,极大提升了模型推理能力。在数学、代码、自然语言推理等任务上,性能较高,能力较强。',
|
458
475
|
displayName: 'DeepSeek R1',
|
@@ -462,8 +479,8 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
462
479
|
organization: 'DeepSeek',
|
463
480
|
pricing: {
|
464
481
|
currency: 'CNY',
|
465
|
-
input:
|
466
|
-
output:
|
482
|
+
input: 4,
|
483
|
+
output: 16,
|
467
484
|
},
|
468
485
|
releasedAt: '2025-01-27',
|
469
486
|
type: 'chat',
|
@@ -472,7 +489,7 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
472
489
|
abilities: {
|
473
490
|
functionCall: true,
|
474
491
|
},
|
475
|
-
contextWindowTokens:
|
492
|
+
contextWindowTokens: 65_792,
|
476
493
|
description:
|
477
494
|
'DeepSeek-V3 为自研 MoE 模型,671B 参数,激活 37B,在 14.8T token 上进行了预训练,在长文本、代码、数学、百科、中文能力上表现优秀。',
|
478
495
|
displayName: 'DeepSeek V3',
|
@@ -482,8 +499,8 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
482
499
|
organization: 'DeepSeek',
|
483
500
|
pricing: {
|
484
501
|
currency: 'CNY',
|
485
|
-
input:
|
486
|
-
output:
|
502
|
+
input: 2,
|
503
|
+
output: 8,
|
487
504
|
},
|
488
505
|
releasedAt: '2025-01-27',
|
489
506
|
type: 'chat',
|
@@ -492,12 +509,12 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
492
509
|
abilities: {
|
493
510
|
reasoning: true,
|
494
511
|
},
|
495
|
-
contextWindowTokens:
|
512
|
+
contextWindowTokens: 32_768,
|
496
513
|
description:
|
497
514
|
'DeepSeek-R1-Distill-Qwen-1.5B 是一个基于 Qwen2.5-Math-1.5B 的蒸馏大型语言模型,使用了 DeepSeek R1 的输出。',
|
498
515
|
displayName: 'DeepSeek R1 Distill Qwen 1.5B',
|
499
516
|
id: 'deepseek-r1-distill-qwen-1.5b',
|
500
|
-
maxOutput:
|
517
|
+
maxOutput: 16_384,
|
501
518
|
organization: 'DeepSeek',
|
502
519
|
pricing: {
|
503
520
|
currency: 'CNY',
|
@@ -510,17 +527,17 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
510
527
|
abilities: {
|
511
528
|
reasoning: true,
|
512
529
|
},
|
513
|
-
contextWindowTokens:
|
530
|
+
contextWindowTokens: 32_768,
|
514
531
|
description:
|
515
532
|
'DeepSeek-R1-Distill-Qwen-7B 是一个基于 Qwen2.5-Math-7B 的蒸馏大型语言模型,使用了 DeepSeek R1 的输出。',
|
516
533
|
displayName: 'DeepSeek R1 Distill Qwen 7B',
|
517
534
|
id: 'deepseek-r1-distill-qwen-7b',
|
518
|
-
maxOutput:
|
535
|
+
maxOutput: 16_384,
|
519
536
|
organization: 'DeepSeek',
|
520
537
|
pricing: {
|
521
538
|
currency: 'CNY',
|
522
|
-
input: 0,
|
523
|
-
output:
|
539
|
+
input: 0.5,
|
540
|
+
output: 1,
|
524
541
|
},
|
525
542
|
type: 'chat',
|
526
543
|
},
|
@@ -528,17 +545,17 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
528
545
|
abilities: {
|
529
546
|
reasoning: true,
|
530
547
|
},
|
531
|
-
contextWindowTokens:
|
548
|
+
contextWindowTokens: 32_768,
|
532
549
|
description:
|
533
550
|
'DeepSeek-R1-Distill-Qwen-14B 是一个基于 Qwen2.5-14B 的蒸馏大型语言模型,使用了 DeepSeek R1 的输出。',
|
534
551
|
displayName: 'DeepSeek R1 Distill Qwen 14B',
|
535
552
|
id: 'deepseek-r1-distill-qwen-14b',
|
536
|
-
maxOutput:
|
553
|
+
maxOutput: 16_384,
|
537
554
|
organization: 'DeepSeek',
|
538
555
|
pricing: {
|
539
556
|
currency: 'CNY',
|
540
|
-
input:
|
541
|
-
output:
|
557
|
+
input: 1,
|
558
|
+
output: 3,
|
542
559
|
},
|
543
560
|
type: 'chat',
|
544
561
|
},
|
@@ -546,17 +563,17 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
546
563
|
abilities: {
|
547
564
|
reasoning: true,
|
548
565
|
},
|
549
|
-
contextWindowTokens:
|
566
|
+
contextWindowTokens: 32_768,
|
550
567
|
description:
|
551
568
|
'DeepSeek-R1-Distill-Qwen-32B 是一个基于 Qwen2.5-32B 的蒸馏大型语言模型,使用了 DeepSeek R1 的输出。',
|
552
569
|
displayName: 'DeepSeek R1 Distill Qwen 32B',
|
553
570
|
id: 'deepseek-r1-distill-qwen-32b',
|
554
|
-
maxOutput:
|
571
|
+
maxOutput: 16_384,
|
555
572
|
organization: 'DeepSeek',
|
556
573
|
pricing: {
|
557
574
|
currency: 'CNY',
|
558
|
-
input:
|
559
|
-
output:
|
575
|
+
input: 2,
|
576
|
+
output: 6,
|
560
577
|
},
|
561
578
|
type: 'chat',
|
562
579
|
},
|
@@ -564,12 +581,12 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
564
581
|
abilities: {
|
565
582
|
reasoning: true,
|
566
583
|
},
|
567
|
-
contextWindowTokens:
|
584
|
+
contextWindowTokens: 32_768,
|
568
585
|
description:
|
569
586
|
'DeepSeek-R1-Distill-Llama-8B 是一个基于 Llama-3.1-8B 的蒸馏大型语言模型,使用了 DeepSeek R1 的输出。',
|
570
587
|
displayName: 'DeepSeek R1 Distill Llama 8B',
|
571
588
|
id: 'deepseek-r1-distill-llama-8b',
|
572
|
-
maxOutput:
|
589
|
+
maxOutput: 16_384,
|
573
590
|
organization: 'DeepSeek',
|
574
591
|
pricing: {
|
575
592
|
currency: 'CNY',
|
@@ -582,12 +599,12 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
582
599
|
abilities: {
|
583
600
|
reasoning: true,
|
584
601
|
},
|
585
|
-
contextWindowTokens:
|
602
|
+
contextWindowTokens: 32_768,
|
586
603
|
description:
|
587
604
|
'DeepSeek-R1-Distill-Llama-70B 是一个基于 Llama-3.3-70B-Instruct 的蒸馏大型语言模型,使用了 DeepSeek R1 的输出。',
|
588
605
|
displayName: 'DeepSeek R1 Distill Llama 70B',
|
589
606
|
id: 'deepseek-r1-distill-llama-70b',
|
590
|
-
maxOutput:
|
607
|
+
maxOutput: 16_384,
|
591
608
|
organization: 'DeepSeek',
|
592
609
|
pricing: {
|
593
610
|
currency: 'CNY',
|