@lobehub/chat 1.65.1 → 1.66.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -394,6 +394,7 @@ SUB_DIR="docker-compose/local"
394
394
  FILES=(
395
395
  "$SUB_DIR/docker-compose.yml"
396
396
  "$SUB_DIR/init_data.json"
397
+ "$SUB_DIR/searxng-settings.yml"
397
398
  )
398
399
  ENV_EXAMPLES=(
399
400
  "$SUB_DIR/.env.zh-CN.example"
@@ -434,6 +435,7 @@ section_download_files(){
434
435
 
435
436
  download_file "$SOURCE_URL/${FILES[0]}" "docker-compose.yml"
436
437
  download_file "$SOURCE_URL/${FILES[1]}" "init_data.json"
438
+ download_file "$SOURCE_URL/${FILES[2]}" "searxng-settings.yml"
437
439
 
438
440
  # Download .env.example with the specified language
439
441
  if [ "$LANGUAGE" = "zh_CN" ]; then
@@ -657,4 +659,4 @@ section_display_configurated_report() {
657
659
  printf "\n%s" "$(show_message "tips_show_documentation")"
658
660
  printf "%s\n" $(show_message "tips_show_documentation_url")
659
661
  }
660
- section_display_configurated_report
662
+ section_display_configurated_report
@@ -17,7 +17,7 @@ LobeChat supports customizing the model list during deployment. This configurati
17
17
  You can use `+` to add a model, `-` to hide a model, and use `model name=display name<extension configuration>` to customize the display name of a model, separated by English commas. The basic syntax is as follows:
18
18
 
19
19
  ```text
20
- id=displayName<maxToken:vision:reasoning:fc:file>,model2,model3
20
+ id=displayName<maxToken:vision:reasoning:search:fc:file>,model2,model3
21
21
  ```
22
22
 
23
23
  For example: `+qwen-7b-chat,+glm-6b,-gpt-3.5-turbo,gpt-4-0125-preview=gpt-4-turbo`
@@ -29,7 +29,7 @@ In the above example, it adds `qwen-7b-chat` and `glm-6b` to the model list, rem
29
29
  Considering the diversity of model capabilities, we started to add extension configuration in version `0.147.8`, with the following rules:
30
30
 
31
31
  ```shell
32
- id=displayName<maxToken:vision:reasoning:fc:file>
32
+ id=displayName<maxToken:vision:reasoning:search:fc:file>
33
33
  ```
34
34
 
35
35
  The first value in angle brackets is designated as the `maxToken` for this model. The second value and beyond are the model's extension capabilities, separated by colons `:`, and the order is not important.
@@ -40,6 +40,7 @@ Examples are as follows:
40
40
  - `spark-v3.5=讯飞星火 v3.5<8192:fc>`: Xunfei Spark 3.5 model, maximum context of 8k, supports Function Call;
41
41
  - `gemini-1.5-flash-latest=Gemini 1.5 Flash<16000:vision>`: Google Vision model, maximum context of 16k, supports image recognition;
42
42
  - `o3-mini=OpenAI o3-mini<200000:reasoning:fc>`: OpenAI o3-mini model, maximum context of 200k, supports reasoning and Function Call;
43
+ - `qwen-max-latest=Qwen Max<32768:search:fc>`: Qwen 2.5 Max model, maximum context of 32k, supports web search and Function Call;
43
44
  - `gpt-4-all=ChatGPT Plus<128000:fc:vision:file>`, hacked version of ChatGPT Plus web, context of 128k, supports image recognition, Function Call, file upload.
44
45
 
45
46
  Currently supported extension capabilities are:
@@ -49,4 +50,5 @@ Currently supported extension capabilities are:
49
50
  | `fc` | Function Calling |
50
51
  | `vision` | Image Recognition |
51
52
  | `reasoning` | Support Reasoning |
53
+ | `search` | Support Web Search |
52
54
  | `file` | File Upload (a bit hacky, not recommended for daily use) |
@@ -16,7 +16,7 @@ LobeChat 支持在部署时自定义模型列表,详情请参考 [模型提供
16
16
  你可以使用 `+` 增加一个模型,使用 `-` 来隐藏一个模型,使用 `模型名=展示名<扩展配置>` 来自定义模型的展示名,用英文逗号隔开。通过 `<>` 来添加扩展配置。基本语法如下:
17
17
 
18
18
  ```text
19
- id=displayName<maxToken:vision:reasoning:fc:file>,model2,model3
19
+ id=displayName<maxToken:vision:reasoning:search:fc:file>,model2,model3
20
20
  ```
21
21
 
22
22
  例如: `+qwen-7b-chat,+glm-6b,-gpt-3.5-turbo,gpt-4-0125-preview=gpt-4-turbo`
@@ -28,7 +28,7 @@ id=displayName<maxToken:vision:reasoning:fc:file>,model2,model3
28
28
  考虑到模型的能力多样性,我们在 `0.147.8` 版本开始增加扩展性配置,它的规则如下:
29
29
 
30
30
  ```shell
31
- id=displayName<maxToken:vision:reasoning:fc:file>
31
+ id=displayName<maxToken:vision:reasoning:search:fc:file>
32
32
  ```
33
33
 
34
34
  尖括号第一个值约定为这个模型的 `maxToken` 。第二个及以后作为模型的扩展能力,能力与能力之间用冒号 `:` 作为分隔符,顺序不重要。
@@ -39,6 +39,7 @@ id=displayName<maxToken:vision:reasoning:fc:file>
39
39
  - `spark-v3.5=讯飞星火 v3.5<8192:fc>`:讯飞星火 3.5 模型,最大上下文 8k,支持 Function Call;
40
40
  - `gemini-1.5-flash-latest=Gemini 1.5 Flash<16000:vision>`:Google 视觉模型,最大上下文 16k,支持图像识别;
41
41
  - `o3-mini=OpenAI o3-mini<200000:reasoning:fc>`:OpenAI o3-mini 模型,最大上下文 200k,支持推理及 Function Call;
42
+ - `qwen-max-latest=Qwen Max<32768:search:fc>`:通义千问 2.5 Max 模型,最大上下文 32k,支持联网搜索及 Function Call;
42
43
  - `gpt-4-all=ChatGPT Plus<128000:fc:vision:file>`,hack 的 ChatGPT Plus 网页版,上下 128k ,支持图像识别、Function Call、文件上传
43
44
 
44
45
  目前支持的扩展能力有:
@@ -48,4 +49,5 @@ id=displayName<maxToken:vision:reasoning:fc:file>
48
49
  | `fc` | 函数调用(function calling) |
49
50
  | `vision` | 视觉识别 |
50
51
  | `reasoning` | 支持推理 |
52
+ | `search` | 支持联网搜索 |
51
53
  | `file` | 文件上传(比较 hack,不建议日常使用) |
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@lobehub/chat",
3
- "version": "1.65.1",
3
+ "version": "1.66.0",
4
4
  "description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
5
5
  "keywords": [
6
6
  "framework",
@@ -131,7 +131,7 @@
131
131
  "@lobehub/tts": "^1.28.0",
132
132
  "@lobehub/ui": "^1.165.2",
133
133
  "@neondatabase/serverless": "^0.10.4",
134
- "@next/third-parties": "^15.1.4",
134
+ "@next/third-parties": "^15.2.0",
135
135
  "@react-spring/web": "^9.7.5",
136
136
  "@sentry/nextjs": "^7.120.2",
137
137
  "@serwist/next": "^9.0.11",
@@ -179,7 +179,7 @@
179
179
  "mdast-util-to-markdown": "^2.1.2",
180
180
  "modern-screenshot": "^4.5.5",
181
181
  "nanoid": "^5.0.9",
182
- "next": "^15.1.4",
182
+ "next": "^15.2.0",
183
183
  "next-auth": "beta",
184
184
  "next-mdx-remote": "^5.0.0",
185
185
  "nextjs-toploader": "^3.7.15",
@@ -250,8 +250,8 @@
250
250
  "@lobehub/i18n-cli": "^1.20.3",
251
251
  "@lobehub/lint": "^1.25.5",
252
252
  "@lobehub/seo-cli": "^1.4.3",
253
- "@next/bundle-analyzer": "^15.1.4",
254
- "@next/eslint-plugin-next": "^15.1.4",
253
+ "@next/bundle-analyzer": "^15.2.0",
254
+ "@next/eslint-plugin-next": "^15.2.0",
255
255
  "@peculiar/webcrypto": "^1.5.0",
256
256
  "@semantic-release/exec": "^6.0.3",
257
257
  "@testing-library/jest-dom": "^6.6.3",
@@ -269,8 +269,8 @@
269
269
  "@types/node": "^22.10.5",
270
270
  "@types/numeral": "^2.0.5",
271
271
  "@types/pg": "^8.11.10",
272
- "@types/react": "^19.0.3",
273
- "@types/react-dom": "^19.0.2",
272
+ "@types/react": "^19.0.10",
273
+ "@types/react-dom": "^19.0.4",
274
274
  "@types/rtl-detect": "^1.0.3",
275
275
  "@types/semver": "^7.5.8",
276
276
  "@types/systemjs": "^6.15.1",
@@ -52,6 +52,7 @@ const ai360ChatModels: AIChatModelCard[] = [
52
52
  {
53
53
  abilities: {
54
54
  functionCall: true,
55
+ search: true,
55
56
  },
56
57
  contextWindowTokens: 8000,
57
58
  description:
@@ -64,6 +65,9 @@ const ai360ChatModels: AIChatModelCard[] = [
64
65
  input: 2,
65
66
  output: 5,
66
67
  },
68
+ settings: {
69
+ searchImpl: 'params',
70
+ },
67
71
  type: 'chat',
68
72
  },
69
73
  {
@@ -4,6 +4,7 @@ const baichuanChatModels: AIChatModelCard[] = [
4
4
  {
5
5
  abilities: {
6
6
  functionCall: true,
7
+ search: true,
7
8
  },
8
9
  contextWindowTokens: 32_768,
9
10
  description:
@@ -17,11 +18,15 @@ const baichuanChatModels: AIChatModelCard[] = [
17
18
  input: 100,
18
19
  output: 100,
19
20
  },
21
+ settings: {
22
+ searchImpl: 'params',
23
+ },
20
24
  type: 'chat',
21
25
  },
22
26
  {
23
27
  abilities: {
24
28
  functionCall: true,
29
+ search: true,
25
30
  },
26
31
  contextWindowTokens: 32_768,
27
32
  description:
@@ -35,11 +40,15 @@ const baichuanChatModels: AIChatModelCard[] = [
35
40
  input: 15,
36
41
  output: 15,
37
42
  },
43
+ settings: {
44
+ searchImpl: 'params',
45
+ },
38
46
  type: 'chat',
39
47
  },
40
48
  {
41
49
  abilities: {
42
50
  functionCall: true,
51
+ search: true,
43
52
  },
44
53
  contextWindowTokens: 32_768,
45
54
  description:
@@ -53,11 +62,15 @@ const baichuanChatModels: AIChatModelCard[] = [
53
62
  input: 0.98,
54
63
  output: 0.98,
55
64
  },
65
+ settings: {
66
+ searchImpl: 'params',
67
+ },
56
68
  type: 'chat',
57
69
  },
58
70
  {
59
71
  abilities: {
60
72
  functionCall: true,
73
+ search: true,
61
74
  },
62
75
  contextWindowTokens: 32_768,
63
76
  description:
@@ -70,6 +83,9 @@ const baichuanChatModels: AIChatModelCard[] = [
70
83
  input: 12,
71
84
  output: 12,
72
85
  },
86
+ settings: {
87
+ searchImpl: 'params',
88
+ },
73
89
  type: 'chat',
74
90
  },
75
91
  {
@@ -80,9 +80,9 @@ const googleChatModels: AIChatModelCard[] = [
80
80
  vision: true,
81
81
  },
82
82
  contextWindowTokens: 1_048_576 + 8192,
83
- description: '一个 Gemini 2.0 Flash 模型,针对成本效益和低延迟等目标进行了优化。',
84
- displayName: 'Gemini 2.0 Flash-Lite Preview 02-05',
85
- id: 'gemini-2.0-flash-lite-preview-02-05',
83
+ description: 'Gemini 2.0 Flash 模型变体,针对成本效益和低延迟等目标进行了优化。',
84
+ displayName: 'Gemini 2.0 Flash-Lite 001',
85
+ id: 'gemini-2.0-flash-lite-001',
86
86
  maxOutput: 8192,
87
87
  pricing: {
88
88
  cachedInput: 0.018_75,
@@ -176,6 +176,16 @@ const groqChatModels: AIChatModelCard[] = [
176
176
  },
177
177
  type: 'chat',
178
178
  },
179
+ {
180
+ contextWindowTokens: 32_768,
181
+ displayName: 'Mixtral Saba 24B',
182
+ id: 'mistral-saba-24b',
183
+ pricing: {
184
+ input: 0.79,
185
+ output: 0.79,
186
+ },
187
+ type: 'chat',
188
+ },
179
189
  {
180
190
  abilities: {
181
191
  functionCall: true,
@@ -20,6 +20,9 @@ const hunyuanChatModels: AIChatModelCard[] = [
20
20
  type: 'chat',
21
21
  },
22
22
  {
23
+ abilities: {
24
+ search: true,
25
+ },
23
26
  contextWindowTokens: 32_000,
24
27
  description:
25
28
  '采用更优的路由策略,同时缓解了负载均衡和专家趋同的问题。长文方面,大海捞针指标达到99.9%。MOE-32K 性价比相对更高,在平衡效果、价格的同时,可对实现对长文本输入的处理。',
@@ -33,9 +36,15 @@ const hunyuanChatModels: AIChatModelCard[] = [
33
36
  output: 2,
34
37
  },
35
38
  releasedAt: '2025-02-10',
39
+ settings: {
40
+ searchImpl: 'params',
41
+ },
36
42
  type: 'chat',
37
43
  },
38
44
  {
45
+ abilities: {
46
+ search: true,
47
+ },
39
48
  contextWindowTokens: 256_000,
40
49
  description:
41
50
  '采用更优的路由策略,同时缓解了负载均衡和专家趋同的问题。长文方面,大海捞针指标达到99.9%。MOE-256K 在长度和效果上进一步突破,极大的扩展了可输入长度。',
@@ -49,11 +58,15 @@ const hunyuanChatModels: AIChatModelCard[] = [
49
58
  output: 2,
50
59
  },
51
60
  releasedAt: '2025-02-10',
61
+ settings: {
62
+ searchImpl: 'params',
63
+ },
52
64
  type: 'chat',
53
65
  },
54
66
  {
55
67
  abilities: {
56
68
  functionCall: true,
69
+ search: true,
57
70
  },
58
71
  contextWindowTokens: 32_000,
59
72
  description:
@@ -68,11 +81,15 @@ const hunyuanChatModels: AIChatModelCard[] = [
68
81
  output: 50,
69
82
  },
70
83
  releasedAt: '2025-01-10',
84
+ settings: {
85
+ searchImpl: 'params',
86
+ },
71
87
  type: 'chat',
72
88
  },
73
89
  {
74
90
  abilities: {
75
91
  functionCall: true,
92
+ search: true,
76
93
  },
77
94
  contextWindowTokens: 32_000,
78
95
  description:
@@ -86,11 +103,15 @@ const hunyuanChatModels: AIChatModelCard[] = [
86
103
  output: 50,
87
104
  },
88
105
  releasedAt: '2025-01-10',
106
+ settings: {
107
+ searchImpl: 'params',
108
+ },
89
109
  type: 'chat',
90
110
  },
91
111
  {
92
112
  abilities: {
93
113
  functionCall: true,
114
+ search: true,
94
115
  },
95
116
  contextWindowTokens: 32_000,
96
117
  description:
@@ -104,11 +125,15 @@ const hunyuanChatModels: AIChatModelCard[] = [
104
125
  output: 50,
105
126
  },
106
127
  releasedAt: '2025-01-10',
128
+ settings: {
129
+ searchImpl: 'params',
130
+ },
107
131
  type: 'chat',
108
132
  },
109
133
  {
110
134
  abilities: {
111
135
  functionCall: true,
136
+ search: true,
112
137
  },
113
138
  contextWindowTokens: 32_000,
114
139
  description:
@@ -122,9 +147,15 @@ const hunyuanChatModels: AIChatModelCard[] = [
122
147
  output: 50,
123
148
  },
124
149
  releasedAt: '2024-11-20',
150
+ settings: {
151
+ searchImpl: 'params',
152
+ },
125
153
  type: 'chat',
126
154
  },
127
155
  {
156
+ abilities: {
157
+ search: true,
158
+ },
128
159
  contextWindowTokens: 32_000,
129
160
  description:
130
161
  'Hunyuan-large 模型总参数量约 389B,激活参数量约 52B,是当前业界参数规模最大、效果最好的 Transformer 架构的开源 MoE 模型。',
@@ -138,9 +169,15 @@ const hunyuanChatModels: AIChatModelCard[] = [
138
169
  output: 12,
139
170
  },
140
171
  releasedAt: '2025-02-10',
172
+ settings: {
173
+ searchImpl: 'params',
174
+ },
141
175
  type: 'chat',
142
176
  },
143
177
  {
178
+ abilities: {
179
+ search: true,
180
+ },
144
181
  contextWindowTokens: 134_000,
145
182
  description:
146
183
  '擅长处理长文任务如文档摘要和文档问答等,同时也具备处理通用文本生成任务的能力。在长文本的分析和生成上表现优异,能有效应对复杂和详尽的长文内容处理需求。',
@@ -154,6 +191,9 @@ const hunyuanChatModels: AIChatModelCard[] = [
154
191
  output: 18,
155
192
  },
156
193
  releasedAt: '2024-12-18',
194
+ settings: {
195
+ searchImpl: 'params',
196
+ },
157
197
  type: 'chat',
158
198
  },
159
199
  {
@@ -4,6 +4,7 @@ const moonshotChatModels: AIChatModelCard[] = [
4
4
  {
5
5
  abilities: {
6
6
  functionCall: true,
7
+ search: true,
7
8
  vision: true,
8
9
  },
9
10
  contextWindowTokens: 131_072,
@@ -19,11 +20,15 @@ const moonshotChatModels: AIChatModelCard[] = [
19
20
  output: 60,
20
21
  },
21
22
  releasedAt: '2025-02-17',
23
+ settings: {
24
+ searchImpl: 'params',
25
+ },
22
26
  type: 'chat',
23
27
  },
24
28
  {
25
29
  abilities: {
26
30
  functionCall: true,
31
+ search: true,
27
32
  },
28
33
  contextWindowTokens: 131_072,
29
34
  description:
@@ -35,11 +40,15 @@ const moonshotChatModels: AIChatModelCard[] = [
35
40
  input: 60,
36
41
  output: 60,
37
42
  },
43
+ settings: {
44
+ searchImpl: 'params',
45
+ },
38
46
  type: 'chat',
39
47
  },
40
48
  {
41
49
  abilities: {
42
50
  functionCall: true,
51
+ search: true,
43
52
  },
44
53
  contextWindowTokens: 8192,
45
54
  description:
@@ -51,11 +60,15 @@ const moonshotChatModels: AIChatModelCard[] = [
51
60
  input: 12,
52
61
  output: 12,
53
62
  },
63
+ settings: {
64
+ searchImpl: 'params',
65
+ },
54
66
  type: 'chat',
55
67
  },
56
68
  {
57
69
  abilities: {
58
70
  functionCall: true,
71
+ search: true,
59
72
  },
60
73
  contextWindowTokens: 32_768,
61
74
  description:
@@ -67,11 +80,15 @@ const moonshotChatModels: AIChatModelCard[] = [
67
80
  input: 24,
68
81
  output: 24,
69
82
  },
83
+ settings: {
84
+ searchImpl: 'params',
85
+ },
70
86
  type: 'chat',
71
87
  },
72
88
  {
73
89
  abilities: {
74
90
  functionCall: true,
91
+ search: true,
75
92
  },
76
93
  contextWindowTokens: 131_072,
77
94
  description:
@@ -83,11 +100,15 @@ const moonshotChatModels: AIChatModelCard[] = [
83
100
  input: 60,
84
101
  output: 60,
85
102
  },
103
+ settings: {
104
+ searchImpl: 'params',
105
+ },
86
106
  type: 'chat',
87
107
  },
88
108
  {
89
109
  abilities: {
90
110
  functionCall: true,
111
+ search: true,
91
112
  vision: true,
92
113
  },
93
114
  contextWindowTokens: 8192,
@@ -101,11 +122,15 @@ const moonshotChatModels: AIChatModelCard[] = [
101
122
  output: 12,
102
123
  },
103
124
  releasedAt: '2025-01-14',
125
+ settings: {
126
+ searchImpl: 'params',
127
+ },
104
128
  type: 'chat',
105
129
  },
106
130
  {
107
131
  abilities: {
108
132
  functionCall: true,
133
+ search: true,
109
134
  vision: true,
110
135
  },
111
136
  contextWindowTokens: 32_768,
@@ -119,11 +144,15 @@ const moonshotChatModels: AIChatModelCard[] = [
119
144
  output: 24,
120
145
  },
121
146
  releasedAt: '2025-01-14',
147
+ settings: {
148
+ searchImpl: 'params',
149
+ },
122
150
  type: 'chat',
123
151
  },
124
152
  {
125
153
  abilities: {
126
154
  functionCall: true,
155
+ search: true,
127
156
  vision: true,
128
157
  },
129
158
  contextWindowTokens: 131_072,
@@ -137,6 +166,9 @@ const moonshotChatModels: AIChatModelCard[] = [
137
166
  output: 60,
138
167
  },
139
168
  releasedAt: '2025-01-14',
169
+ settings: {
170
+ searchImpl: 'params',
171
+ },
140
172
  type: 'chat',
141
173
  },
142
174
  ];
@@ -62,8 +62,8 @@ const qwenChatModels: AIChatModelCard[] = [
62
62
  organization: 'Qwen',
63
63
  pricing: {
64
64
  currency: 'CNY',
65
- input: 20,
66
- output: 60,
65
+ input: 2.4,
66
+ output: 9.6,
67
67
  },
68
68
  settings: {
69
69
  searchImpl: 'params',
@@ -410,6 +410,23 @@ const qwenChatModels: AIChatModelCard[] = [
410
410
  },
411
411
  type: 'chat',
412
412
  },
413
+ {
414
+ abilities: {
415
+ vision: true,
416
+ },
417
+ contextWindowTokens: 32_768,
418
+ description: 'Qwen-Omni 系列模型支持输入多种模态的数据,包括视频、音频、图片、文本,并输出音频与文本。',
419
+ displayName: 'Qwen Omni Turbo',
420
+ id: 'qwen-omni-turbo-latest',
421
+ maxOutput: 2048,
422
+ organization: 'Qwen',
423
+ pricing: {
424
+ currency: 'CNY',
425
+ input: 0,
426
+ output: 0,
427
+ },
428
+ type: 'chat',
429
+ },
413
430
  {
414
431
  abilities: {
415
432
  vision: true,
@@ -452,7 +469,7 @@ const qwenChatModels: AIChatModelCard[] = [
452
469
  abilities: {
453
470
  reasoning: true,
454
471
  },
455
- contextWindowTokens: 131_072,
472
+ contextWindowTokens: 65_792,
456
473
  description:
457
474
  'DeepSeek-R1 在后训练阶段大规模使用了强化学习技术,在仅有极少标注数据的情况下,极大提升了模型推理能力。在数学、代码、自然语言推理等任务上,性能较高,能力较强。',
458
475
  displayName: 'DeepSeek R1',
@@ -462,8 +479,8 @@ const qwenChatModels: AIChatModelCard[] = [
462
479
  organization: 'DeepSeek',
463
480
  pricing: {
464
481
  currency: 'CNY',
465
- input: 0,
466
- output: 0,
482
+ input: 4,
483
+ output: 16,
467
484
  },
468
485
  releasedAt: '2025-01-27',
469
486
  type: 'chat',
@@ -472,7 +489,7 @@ const qwenChatModels: AIChatModelCard[] = [
472
489
  abilities: {
473
490
  functionCall: true,
474
491
  },
475
- contextWindowTokens: 131_072,
492
+ contextWindowTokens: 65_792,
476
493
  description:
477
494
  'DeepSeek-V3 为自研 MoE 模型,671B 参数,激活 37B,在 14.8T token 上进行了预训练,在长文本、代码、数学、百科、中文能力上表现优秀。',
478
495
  displayName: 'DeepSeek V3',
@@ -482,8 +499,8 @@ const qwenChatModels: AIChatModelCard[] = [
482
499
  organization: 'DeepSeek',
483
500
  pricing: {
484
501
  currency: 'CNY',
485
- input: 0,
486
- output: 0,
502
+ input: 2,
503
+ output: 8,
487
504
  },
488
505
  releasedAt: '2025-01-27',
489
506
  type: 'chat',
@@ -492,12 +509,12 @@ const qwenChatModels: AIChatModelCard[] = [
492
509
  abilities: {
493
510
  reasoning: true,
494
511
  },
495
- contextWindowTokens: 131_072,
512
+ contextWindowTokens: 32_768,
496
513
  description:
497
514
  'DeepSeek-R1-Distill-Qwen-1.5B 是一个基于 Qwen2.5-Math-1.5B 的蒸馏大型语言模型,使用了 DeepSeek R1 的输出。',
498
515
  displayName: 'DeepSeek R1 Distill Qwen 1.5B',
499
516
  id: 'deepseek-r1-distill-qwen-1.5b',
500
- maxOutput: 8192,
517
+ maxOutput: 16_384,
501
518
  organization: 'DeepSeek',
502
519
  pricing: {
503
520
  currency: 'CNY',
@@ -510,17 +527,17 @@ const qwenChatModels: AIChatModelCard[] = [
510
527
  abilities: {
511
528
  reasoning: true,
512
529
  },
513
- contextWindowTokens: 131_072,
530
+ contextWindowTokens: 32_768,
514
531
  description:
515
532
  'DeepSeek-R1-Distill-Qwen-7B 是一个基于 Qwen2.5-Math-7B 的蒸馏大型语言模型,使用了 DeepSeek R1 的输出。',
516
533
  displayName: 'DeepSeek R1 Distill Qwen 7B',
517
534
  id: 'deepseek-r1-distill-qwen-7b',
518
- maxOutput: 8192,
535
+ maxOutput: 16_384,
519
536
  organization: 'DeepSeek',
520
537
  pricing: {
521
538
  currency: 'CNY',
522
- input: 0,
523
- output: 0,
539
+ input: 0.5,
540
+ output: 1,
524
541
  },
525
542
  type: 'chat',
526
543
  },
@@ -528,17 +545,17 @@ const qwenChatModels: AIChatModelCard[] = [
528
545
  abilities: {
529
546
  reasoning: true,
530
547
  },
531
- contextWindowTokens: 131_072,
548
+ contextWindowTokens: 32_768,
532
549
  description:
533
550
  'DeepSeek-R1-Distill-Qwen-14B 是一个基于 Qwen2.5-14B 的蒸馏大型语言模型,使用了 DeepSeek R1 的输出。',
534
551
  displayName: 'DeepSeek R1 Distill Qwen 14B',
535
552
  id: 'deepseek-r1-distill-qwen-14b',
536
- maxOutput: 8192,
553
+ maxOutput: 16_384,
537
554
  organization: 'DeepSeek',
538
555
  pricing: {
539
556
  currency: 'CNY',
540
- input: 0,
541
- output: 0,
557
+ input: 1,
558
+ output: 3,
542
559
  },
543
560
  type: 'chat',
544
561
  },
@@ -546,17 +563,17 @@ const qwenChatModels: AIChatModelCard[] = [
546
563
  abilities: {
547
564
  reasoning: true,
548
565
  },
549
- contextWindowTokens: 131_072,
566
+ contextWindowTokens: 32_768,
550
567
  description:
551
568
  'DeepSeek-R1-Distill-Qwen-32B 是一个基于 Qwen2.5-32B 的蒸馏大型语言模型,使用了 DeepSeek R1 的输出。',
552
569
  displayName: 'DeepSeek R1 Distill Qwen 32B',
553
570
  id: 'deepseek-r1-distill-qwen-32b',
554
- maxOutput: 8192,
571
+ maxOutput: 16_384,
555
572
  organization: 'DeepSeek',
556
573
  pricing: {
557
574
  currency: 'CNY',
558
- input: 0,
559
- output: 0,
575
+ input: 2,
576
+ output: 6,
560
577
  },
561
578
  type: 'chat',
562
579
  },
@@ -564,12 +581,12 @@ const qwenChatModels: AIChatModelCard[] = [
564
581
  abilities: {
565
582
  reasoning: true,
566
583
  },
567
- contextWindowTokens: 131_072,
584
+ contextWindowTokens: 32_768,
568
585
  description:
569
586
  'DeepSeek-R1-Distill-Llama-8B 是一个基于 Llama-3.1-8B 的蒸馏大型语言模型,使用了 DeepSeek R1 的输出。',
570
587
  displayName: 'DeepSeek R1 Distill Llama 8B',
571
588
  id: 'deepseek-r1-distill-llama-8b',
572
- maxOutput: 8192,
589
+ maxOutput: 16_384,
573
590
  organization: 'DeepSeek',
574
591
  pricing: {
575
592
  currency: 'CNY',
@@ -582,12 +599,12 @@ const qwenChatModels: AIChatModelCard[] = [
582
599
  abilities: {
583
600
  reasoning: true,
584
601
  },
585
- contextWindowTokens: 131_072,
602
+ contextWindowTokens: 32_768,
586
603
  description:
587
604
  'DeepSeek-R1-Distill-Llama-70B 是一个基于 Llama-3.3-70B-Instruct 的蒸馏大型语言模型,使用了 DeepSeek R1 的输出。',
588
605
  displayName: 'DeepSeek R1 Distill Llama 70B',
589
606
  id: 'deepseek-r1-distill-llama-70b',
590
- maxOutput: 8192,
607
+ maxOutput: 16_384,
591
608
  organization: 'DeepSeek',
592
609
  pricing: {
593
610
  currency: 'CNY',