@lobehub/lobehub 2.0.0-next.186 → 2.0.0-next.187
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +25 -0
- package/changelog/v1.json +9 -0
- package/package.json +1 -1
- package/packages/fetch-sse/src/__tests__/request.test.ts +608 -0
- package/packages/model-bank/src/aiModels/aihubmix.ts +44 -8
- package/packages/model-bank/src/aiModels/google.ts +49 -17
- package/packages/model-bank/src/aiModels/hunyuan.ts +20 -0
- package/packages/model-bank/src/aiModels/infiniai.ts +48 -7
- package/packages/model-bank/src/aiModels/lobehub.ts +13 -11
- package/packages/model-bank/src/aiModels/minimax.ts +46 -2
- package/packages/model-bank/src/aiModels/ollamacloud.ts +40 -5
- package/packages/model-bank/src/aiModels/openai.ts +6 -3
- package/packages/model-bank/src/aiModels/qwen.ts +1 -1
- package/packages/model-bank/src/aiModels/siliconcloud.ts +60 -0
- package/packages/model-bank/src/aiModels/vertexai.ts +77 -44
- package/packages/model-bank/src/aiModels/volcengine.ts +111 -2
- package/packages/model-bank/src/aiModels/zenmux.ts +19 -13
- package/packages/model-bank/src/aiModels/zhipu.ts +64 -2
- package/packages/model-bank/src/types/aiModel.ts +3 -0
- package/packages/model-runtime/src/providers/volcengine/index.ts +2 -1
- package/packages/model-runtime/src/providers/zhipu/index.test.ts +0 -27
- package/packages/model-runtime/src/providers/zhipu/index.ts +1 -1
- package/packages/model-runtime/src/utils/modelParse.ts +26 -21
- package/packages/types/src/agent/chatConfig.ts +6 -2
- package/src/features/ChatInput/ActionBar/Model/ControlsForm.tsx +40 -1
- package/src/features/ChatInput/ActionBar/Model/GPT52ProReasoningEffortSlider.tsx +59 -0
- package/src/features/ChatInput/ActionBar/Model/GPT52ReasoningEffortSlider.tsx +61 -0
- package/src/features/ChatInput/ActionBar/Model/TextVerbositySlider.tsx +1 -1
- package/src/features/ChatInput/ActionBar/Model/ThinkingLevel2Slider.tsx +58 -0
- package/src/features/ChatInput/ActionBar/Model/ThinkingLevelSlider.tsx +10 -8
- package/src/services/chat/mecha/modelParamsResolver.ts +11 -0
|
@@ -2,6 +2,66 @@ import { AIChatModelCard, AIImageModelCard } from '../types/aiModel';
|
|
|
2
2
|
|
|
3
3
|
// https://siliconflow.cn/zh-cn/models
|
|
4
4
|
const siliconcloudChatModels: AIChatModelCard[] = [
|
|
5
|
+
{
|
|
6
|
+
abilities: {
|
|
7
|
+
functionCall: true,
|
|
8
|
+
reasoning: true,
|
|
9
|
+
},
|
|
10
|
+
contextWindowTokens: 200_000,
|
|
11
|
+
description:
|
|
12
|
+
'GLM-4.7 是智谱新一代旗舰模型,总参数量 355B,激活参数量 32B,在通用对话、推理和智能体能力方面全面升级。GLM-4.7 增强了 Interleaved Thinking(交错思考),并引入了 Preserved Thinking(保留思考)和 Turn-level Thinking(轮级思考)。',
|
|
13
|
+
displayName: 'GLM-4.7 (Pro)',
|
|
14
|
+
id: 'Pro/zai-org/glm-4.7',
|
|
15
|
+
pricing: {
|
|
16
|
+
currency: 'CNY',
|
|
17
|
+
units: [
|
|
18
|
+
{
|
|
19
|
+
lookup: {
|
|
20
|
+
prices: {
|
|
21
|
+
'[0, 0.032]_[0, 0.0002]': 0.4,
|
|
22
|
+
'[0, 0.032]_[0.0002, infinity]': 0.6,
|
|
23
|
+
'[0.032, infinity]_[0, infinity]': 0.8,
|
|
24
|
+
},
|
|
25
|
+
pricingParams: ['textInput', 'textOutput'],
|
|
26
|
+
},
|
|
27
|
+
name: 'textInput_cacheRead',
|
|
28
|
+
strategy: 'lookup',
|
|
29
|
+
unit: 'millionTokens',
|
|
30
|
+
},
|
|
31
|
+
{
|
|
32
|
+
lookup: {
|
|
33
|
+
prices: {
|
|
34
|
+
'[0, 0.032]_[0, 0.0002]': 2,
|
|
35
|
+
'[0, 0.032]_[0.0002, infinity]': 3,
|
|
36
|
+
'[0.032, infinity]_[0, infinity]': 4,
|
|
37
|
+
},
|
|
38
|
+
pricingParams: ['textInput', 'textOutput'],
|
|
39
|
+
},
|
|
40
|
+
name: 'textInput',
|
|
41
|
+
strategy: 'lookup',
|
|
42
|
+
unit: 'millionTokens',
|
|
43
|
+
},
|
|
44
|
+
{
|
|
45
|
+
lookup: {
|
|
46
|
+
prices: {
|
|
47
|
+
'[0, 0.032]_[0, 0.0002]': 8,
|
|
48
|
+
'[0, 0.032]_[0.0002, infinity]': 14,
|
|
49
|
+
'[0.032, infinity]_[0, infinity]': 16,
|
|
50
|
+
},
|
|
51
|
+
pricingParams: ['textInput', 'textOutput'],
|
|
52
|
+
},
|
|
53
|
+
name: 'textOutput',
|
|
54
|
+
strategy: 'lookup',
|
|
55
|
+
unit: 'millionTokens',
|
|
56
|
+
},
|
|
57
|
+
],
|
|
58
|
+
},
|
|
59
|
+
releasedAt: '2025-12-22',
|
|
60
|
+
settings: {
|
|
61
|
+
extendParams: ['enableReasoning', 'reasoningBudgetToken'],
|
|
62
|
+
},
|
|
63
|
+
type: 'chat',
|
|
64
|
+
},
|
|
5
65
|
{
|
|
6
66
|
abilities: {
|
|
7
67
|
functionCall: true,
|
|
@@ -3,35 +3,6 @@ import { imagenGenParameters, nanoBananaParameters } from './google';
|
|
|
3
3
|
|
|
4
4
|
// ref: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models
|
|
5
5
|
const vertexaiChatModels: AIChatModelCard[] = [
|
|
6
|
-
{
|
|
7
|
-
abilities: {
|
|
8
|
-
imageOutput: true,
|
|
9
|
-
reasoning: true,
|
|
10
|
-
search: true,
|
|
11
|
-
vision: true,
|
|
12
|
-
},
|
|
13
|
-
contextWindowTokens: 131_072 + 32_768,
|
|
14
|
-
description:
|
|
15
|
-
"Gemini 3 Pro Image (Nano Banana Pro) is Google's image generation model and also supports multimodal chat.",
|
|
16
|
-
displayName: 'Nano Banana Pro',
|
|
17
|
-
enabled: true,
|
|
18
|
-
id: 'gemini-3-pro-image-preview',
|
|
19
|
-
maxOutput: 32_768,
|
|
20
|
-
pricing: {
|
|
21
|
-
approximatePricePerImage: 0.134,
|
|
22
|
-
units: [
|
|
23
|
-
{ name: 'imageOutput', rate: 120, strategy: 'fixed', unit: 'millionTokens' },
|
|
24
|
-
{ name: 'textInput', rate: 2, strategy: 'fixed', unit: 'millionTokens' },
|
|
25
|
-
{ name: 'textOutput', rate: 12, strategy: 'fixed', unit: 'millionTokens' },
|
|
26
|
-
],
|
|
27
|
-
},
|
|
28
|
-
releasedAt: '2025-11-20',
|
|
29
|
-
settings: {
|
|
30
|
-
searchImpl: 'params',
|
|
31
|
-
searchProvider: 'google',
|
|
32
|
-
},
|
|
33
|
-
type: 'chat',
|
|
34
|
-
},
|
|
35
6
|
{
|
|
36
7
|
abilities: {
|
|
37
8
|
functionCall: true,
|
|
@@ -42,7 +13,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
|
42
13
|
},
|
|
43
14
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
44
15
|
description:
|
|
45
|
-
|
|
16
|
+
'Gemini 3 Pro is Google’s most powerful agent and vibe-coding model, delivering richer visuals and deeper interaction on top of state-of-the-art reasoning.',
|
|
46
17
|
displayName: 'Gemini 3 Pro Preview',
|
|
47
18
|
enabled: true,
|
|
48
19
|
id: 'gemini-3-pro-preview',
|
|
@@ -85,6 +56,41 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
|
85
56
|
],
|
|
86
57
|
},
|
|
87
58
|
releasedAt: '2025-11-18',
|
|
59
|
+
settings: {
|
|
60
|
+
extendParams: ['thinkingLevel2', 'urlContext'],
|
|
61
|
+
searchImpl: 'params',
|
|
62
|
+
searchProvider: 'google',
|
|
63
|
+
},
|
|
64
|
+
type: 'chat',
|
|
65
|
+
},
|
|
66
|
+
{
|
|
67
|
+
abilities: {
|
|
68
|
+
functionCall: true,
|
|
69
|
+
reasoning: true,
|
|
70
|
+
search: true,
|
|
71
|
+
video: true,
|
|
72
|
+
vision: true,
|
|
73
|
+
},
|
|
74
|
+
contextWindowTokens: 1_048_576 + 65_536,
|
|
75
|
+
description: 'Gemini 3 Flash 是为速度而打造的最智能的模型,将前沿智能与卓越的搜索接地相结合。',
|
|
76
|
+
displayName: 'Gemini 3 Flash Preview',
|
|
77
|
+
enabled: true,
|
|
78
|
+
id: 'gemini-3-flash-preview',
|
|
79
|
+
maxOutput: 65_536,
|
|
80
|
+
pricing: {
|
|
81
|
+
units: [
|
|
82
|
+
{ name: 'textInput_cacheRead', rate: 0.05, strategy: 'fixed', unit: 'millionTokens' },
|
|
83
|
+
{ name: 'textInput', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
84
|
+
{ name: 'textOutput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
|
|
85
|
+
{
|
|
86
|
+
lookup: { prices: { '1h': 1 }, pricingParams: ['ttl'] },
|
|
87
|
+
name: 'textInput_cacheWrite',
|
|
88
|
+
strategy: 'lookup',
|
|
89
|
+
unit: 'millionTokens',
|
|
90
|
+
},
|
|
91
|
+
],
|
|
92
|
+
},
|
|
93
|
+
releasedAt: '2025-12-17',
|
|
88
94
|
settings: {
|
|
89
95
|
extendParams: ['thinkingLevel', 'urlContext'],
|
|
90
96
|
searchImpl: 'params',
|
|
@@ -92,6 +98,35 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
|
92
98
|
},
|
|
93
99
|
type: 'chat',
|
|
94
100
|
},
|
|
101
|
+
{
|
|
102
|
+
abilities: {
|
|
103
|
+
imageOutput: true,
|
|
104
|
+
reasoning: true,
|
|
105
|
+
search: true,
|
|
106
|
+
vision: true,
|
|
107
|
+
},
|
|
108
|
+
contextWindowTokens: 131_072 + 32_768,
|
|
109
|
+
description:
|
|
110
|
+
'Gemini 3 Pro Image(Nano Banana Pro)是 Google 的图像生成模型,同时支持多模态对话。',
|
|
111
|
+
displayName: 'Nano Banana Pro',
|
|
112
|
+
enabled: true,
|
|
113
|
+
id: 'gemini-3-pro-image-preview',
|
|
114
|
+
maxOutput: 32_768,
|
|
115
|
+
pricing: {
|
|
116
|
+
approximatePricePerImage: 0.134,
|
|
117
|
+
units: [
|
|
118
|
+
{ name: 'imageOutput', rate: 120, strategy: 'fixed', unit: 'millionTokens' },
|
|
119
|
+
{ name: 'textInput', rate: 2, strategy: 'fixed', unit: 'millionTokens' },
|
|
120
|
+
{ name: 'textOutput', rate: 12, strategy: 'fixed', unit: 'millionTokens' },
|
|
121
|
+
],
|
|
122
|
+
},
|
|
123
|
+
releasedAt: '2025-11-20',
|
|
124
|
+
settings: {
|
|
125
|
+
searchImpl: 'params',
|
|
126
|
+
searchProvider: 'google',
|
|
127
|
+
},
|
|
128
|
+
type: 'chat',
|
|
129
|
+
},
|
|
95
130
|
{
|
|
96
131
|
abilities: {
|
|
97
132
|
functionCall: true,
|
|
@@ -101,9 +136,8 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
|
101
136
|
},
|
|
102
137
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
103
138
|
description:
|
|
104
|
-
|
|
139
|
+
'Gemini 2.5 Pro is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
|
|
105
140
|
displayName: 'Gemini 2.5 Pro',
|
|
106
|
-
enabled: true,
|
|
107
141
|
id: 'gemini-2.5-pro',
|
|
108
142
|
maxOutput: 65_536,
|
|
109
143
|
pricing: {
|
|
@@ -129,7 +163,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
|
129
163
|
},
|
|
130
164
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
131
165
|
description:
|
|
132
|
-
|
|
166
|
+
'Gemini 2.5 Pro Preview is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
|
|
133
167
|
displayName: 'Gemini 2.5 Pro Preview 05-06',
|
|
134
168
|
id: 'gemini-2.5-pro-preview-05-06',
|
|
135
169
|
maxOutput: 65_536,
|
|
@@ -150,7 +184,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
|
150
184
|
},
|
|
151
185
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
152
186
|
description:
|
|
153
|
-
|
|
187
|
+
'Gemini 2.5 Pro Preview is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
|
|
154
188
|
displayName: 'Gemini 2.5 Pro Preview 03-25',
|
|
155
189
|
id: 'gemini-2.5-pro-preview-03-25',
|
|
156
190
|
maxOutput: 65_536,
|
|
@@ -171,9 +205,8 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
|
171
205
|
vision: true,
|
|
172
206
|
},
|
|
173
207
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
174
|
-
description:
|
|
208
|
+
description: 'Gemini 2.5 Flash is Google’s best-value model with full capabilities.',
|
|
175
209
|
displayName: 'Gemini 2.5 Flash',
|
|
176
|
-
enabled: true,
|
|
177
210
|
id: 'gemini-2.5-flash',
|
|
178
211
|
maxOutput: 65_536,
|
|
179
212
|
pricing: {
|
|
@@ -198,7 +231,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
|
198
231
|
vision: true,
|
|
199
232
|
},
|
|
200
233
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
201
|
-
description:
|
|
234
|
+
description: 'Gemini 2.5 Flash Preview is Google’s best-value model with full capabilities.',
|
|
202
235
|
displayName: 'Gemini 2.5 Flash Preview 04-17',
|
|
203
236
|
id: 'gemini-2.5-flash-preview-04-17',
|
|
204
237
|
maxOutput: 65_536,
|
|
@@ -218,7 +251,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
|
218
251
|
},
|
|
219
252
|
contextWindowTokens: 32_768 + 8192,
|
|
220
253
|
description:
|
|
221
|
-
|
|
254
|
+
'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
|
|
222
255
|
displayName: 'Nano Banana',
|
|
223
256
|
enabled: true,
|
|
224
257
|
id: 'gemini-2.5-flash-image',
|
|
@@ -243,7 +276,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
|
243
276
|
},
|
|
244
277
|
contextWindowTokens: 1_000_000 + 64_000,
|
|
245
278
|
description:
|
|
246
|
-
|
|
279
|
+
'Gemini 2.5 Flash-Lite is Google’s smallest, best-value model, designed for large-scale use.',
|
|
247
280
|
displayName: 'Gemini 2.5 Flash-Lite',
|
|
248
281
|
id: 'gemini-2.5-flash-lite',
|
|
249
282
|
maxOutput: 64_000,
|
|
@@ -270,7 +303,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
|
270
303
|
},
|
|
271
304
|
contextWindowTokens: 1_000_000 + 64_000,
|
|
272
305
|
description:
|
|
273
|
-
|
|
306
|
+
'Gemini 2.5 Flash-Lite Preview is Google’s smallest, best-value model, designed for large-scale use.',
|
|
274
307
|
displayName: 'Gemini 2.5 Flash-Lite Preview 06-17',
|
|
275
308
|
id: 'gemini-2.5-flash-lite-preview-06-17',
|
|
276
309
|
maxOutput: 64_000,
|
|
@@ -315,8 +348,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
|
315
348
|
vision: true,
|
|
316
349
|
},
|
|
317
350
|
contextWindowTokens: 1_048_576 + 8192,
|
|
318
|
-
description:
|
|
319
|
-
'A Gemini 2.0 Flash variant optimized for cost efficiency and low latency.',
|
|
351
|
+
description: 'A Gemini 2.0 Flash variant optimized for cost efficiency and low latency.',
|
|
320
352
|
displayName: 'Gemini 2.0 Flash-Lite',
|
|
321
353
|
id: 'gemini-2.0-flash-lite',
|
|
322
354
|
maxOutput: 8192,
|
|
@@ -336,7 +368,8 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
|
336
368
|
vision: true,
|
|
337
369
|
},
|
|
338
370
|
contextWindowTokens: 1_000_000 + 8192,
|
|
339
|
-
description:
|
|
371
|
+
description:
|
|
372
|
+
'Gemini 1.5 Flash 002 is an efficient multimodal model built for broad deployment.',
|
|
340
373
|
displayName: 'Gemini 1.5 Flash 002',
|
|
341
374
|
id: 'gemini-1.5-flash-002',
|
|
342
375
|
maxOutput: 8192,
|
|
@@ -379,7 +412,7 @@ const vertexaiImageModels: AIImageModelCard[] = [
|
|
|
379
412
|
enabled: true,
|
|
380
413
|
type: 'image',
|
|
381
414
|
description:
|
|
382
|
-
|
|
415
|
+
'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
|
|
383
416
|
releasedAt: '2025-08-26',
|
|
384
417
|
parameters: nanoBananaParameters,
|
|
385
418
|
pricing: {
|
|
@@ -3,6 +3,66 @@ import { AIChatModelCard, AIImageModelCard } from '../types/aiModel';
|
|
|
3
3
|
// https://www.volcengine.com/docs/82379/1330310
|
|
4
4
|
|
|
5
5
|
const doubaoChatModels: AIChatModelCard[] = [
|
|
6
|
+
{
|
|
7
|
+
abilities: {
|
|
8
|
+
functionCall: true,
|
|
9
|
+
reasoning: true,
|
|
10
|
+
vision: true,
|
|
11
|
+
},
|
|
12
|
+
config: {
|
|
13
|
+
deploymentName: 'doubao-seed-1-8-251215',
|
|
14
|
+
},
|
|
15
|
+
contextWindowTokens: 256_000,
|
|
16
|
+
description:
|
|
17
|
+
'Doubao-Seed-1.8 有着更强的多模态理解能力和 Agent 能力,支持文本/图片/视频输入与上下文缓存,可在复杂任务中提供更出色的表现。',
|
|
18
|
+
displayName: 'Doubao Seed 1.8',
|
|
19
|
+
id: 'doubao-seed-1.8',
|
|
20
|
+
maxOutput: 64_000,
|
|
21
|
+
pricing: {
|
|
22
|
+
currency: 'CNY',
|
|
23
|
+
units: [
|
|
24
|
+
{
|
|
25
|
+
lookup: {
|
|
26
|
+
prices: {
|
|
27
|
+
'[0, 0.032]': 0.8,
|
|
28
|
+
'[0.032, 0.128]': 1.2,
|
|
29
|
+
'[0.128, 0.256]': 2.4,
|
|
30
|
+
},
|
|
31
|
+
pricingParams: ['textInputRange'],
|
|
32
|
+
},
|
|
33
|
+
name: 'textInput',
|
|
34
|
+
strategy: 'lookup',
|
|
35
|
+
unit: 'millionTokens',
|
|
36
|
+
},
|
|
37
|
+
{
|
|
38
|
+
lookup: {
|
|
39
|
+
prices: {
|
|
40
|
+
'[0, 0.032]_[0, 0.0002]': 2,
|
|
41
|
+
'[0, 0.032]_[0.0002, infinity]': 8,
|
|
42
|
+
'[0.032, 0.128]_[0, infinity]': 16,
|
|
43
|
+
'[0.128, 0.256]_[0, infinity]': 24,
|
|
44
|
+
},
|
|
45
|
+
pricingParams: ['textInputRange', 'textOutputRange'],
|
|
46
|
+
},
|
|
47
|
+
name: 'textOutput',
|
|
48
|
+
strategy: 'lookup',
|
|
49
|
+
unit: 'millionTokens',
|
|
50
|
+
},
|
|
51
|
+
{ name: 'textInput_cacheRead', rate: 0.16, strategy: 'fixed', unit: 'millionTokens' },
|
|
52
|
+
{
|
|
53
|
+
lookup: { prices: { '1h': 0.017 }, pricingParams: ['ttl'] },
|
|
54
|
+
name: 'textInput_cacheWrite',
|
|
55
|
+
strategy: 'lookup',
|
|
56
|
+
unit: 'millionTokens',
|
|
57
|
+
},
|
|
58
|
+
],
|
|
59
|
+
},
|
|
60
|
+
releasedAt: '2025-12-18',
|
|
61
|
+
settings: {
|
|
62
|
+
extendParams: ['gpt5ReasoningEffort'],
|
|
63
|
+
},
|
|
64
|
+
type: 'chat',
|
|
65
|
+
},
|
|
6
66
|
{
|
|
7
67
|
abilities: {
|
|
8
68
|
functionCall: true,
|
|
@@ -56,6 +116,57 @@ const doubaoChatModels: AIChatModelCard[] = [
|
|
|
56
116
|
},
|
|
57
117
|
type: 'chat',
|
|
58
118
|
},
|
|
119
|
+
{
|
|
120
|
+
abilities: {
|
|
121
|
+
functionCall: true,
|
|
122
|
+
reasoning: true,
|
|
123
|
+
},
|
|
124
|
+
config: {
|
|
125
|
+
deploymentName: 'deepseek-v3-2-251201',
|
|
126
|
+
},
|
|
127
|
+
contextWindowTokens: 131_072,
|
|
128
|
+
description:
|
|
129
|
+
'DeepSeek-V3.2 是深度求索推出的首个将思考融入工具使用的混合推理模型,用高效架构省算力、大规模强化学习提能力、大规模合成任务数据强泛化,三者结合性能媲美 GPT-5-High,输出长度大幅降低,显著减少了计算开销与用户等待时间。',
|
|
130
|
+
displayName: 'DeepSeek V3.2',
|
|
131
|
+
enabled: true,
|
|
132
|
+
id: 'deepseek-v3.2',
|
|
133
|
+
maxOutput: 32_768,
|
|
134
|
+
pricing: {
|
|
135
|
+
currency: 'CNY',
|
|
136
|
+
units: [
|
|
137
|
+
{
|
|
138
|
+
lookup: {
|
|
139
|
+
prices: {
|
|
140
|
+
'[0, 0.032]': 2,
|
|
141
|
+
'[0.032, 0.128]': 4,
|
|
142
|
+
},
|
|
143
|
+
pricingParams: ['textInputRange'],
|
|
144
|
+
},
|
|
145
|
+
name: 'textInput',
|
|
146
|
+
strategy: 'lookup',
|
|
147
|
+
unit: 'millionTokens',
|
|
148
|
+
},
|
|
149
|
+
{
|
|
150
|
+
lookup: {
|
|
151
|
+
prices: {
|
|
152
|
+
'[0, 0.032]': 3,
|
|
153
|
+
'[0.032, 0.128]': 6,
|
|
154
|
+
},
|
|
155
|
+
pricingParams: ['textInputRange'],
|
|
156
|
+
},
|
|
157
|
+
name: 'textOutput',
|
|
158
|
+
strategy: 'lookup',
|
|
159
|
+
unit: 'millionTokens',
|
|
160
|
+
},
|
|
161
|
+
{ name: 'textInput_cacheRead', rate: 0.4, strategy: 'fixed', unit: 'millionTokens' },
|
|
162
|
+
{ name: 'textInput_cacheWrite', rate: 0.017, strategy: 'fixed', unit: 'millionTokens' },
|
|
163
|
+
],
|
|
164
|
+
},
|
|
165
|
+
settings: {
|
|
166
|
+
extendParams: ['enableReasoning'],
|
|
167
|
+
},
|
|
168
|
+
type: 'chat',
|
|
169
|
+
},
|
|
59
170
|
{
|
|
60
171
|
abilities: {
|
|
61
172
|
functionCall: true,
|
|
@@ -171,7 +282,6 @@ const doubaoChatModels: AIChatModelCard[] = [
|
|
|
171
282
|
description:
|
|
172
283
|
'Doubao-Seed-1.6-thinking significantly strengthens reasoning, further improving core abilities in coding, math, and logical reasoning over Doubao-1.5-thinking-pro, while adding vision understanding. It supports a 256k context window and up to 16k output tokens.',
|
|
173
284
|
displayName: 'Doubao Seed 1.6 Thinking',
|
|
174
|
-
enabled: true,
|
|
175
285
|
id: 'doubao-seed-1.6-thinking',
|
|
176
286
|
maxOutput: 32_000,
|
|
177
287
|
pricing: {
|
|
@@ -221,7 +331,6 @@ const doubaoChatModels: AIChatModelCard[] = [
|
|
|
221
331
|
description:
|
|
222
332
|
'Doubao-Seed-1.6 is a new multimodal deep-reasoning model with auto, thinking, and non-thinking modes. In non-thinking mode, it significantly outperforms Doubao-1.5-pro/250115. It supports a 256k context window and up to 16k output tokens.',
|
|
223
333
|
displayName: 'Doubao Seed 1.6',
|
|
224
|
-
enabled: true,
|
|
225
334
|
id: 'doubao-seed-1.6',
|
|
226
335
|
maxOutput: 32_000,
|
|
227
336
|
pricing: {
|
|
@@ -32,7 +32,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
32
32
|
},
|
|
33
33
|
releasedAt: '2025-12-11',
|
|
34
34
|
settings: {
|
|
35
|
-
extendParams: ['
|
|
35
|
+
extendParams: ['gpt5_2ReasoningEffort', 'textVerbosity'],
|
|
36
36
|
searchImpl: 'params',
|
|
37
37
|
},
|
|
38
38
|
type: 'chat',
|
|
@@ -147,6 +147,11 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
147
147
|
],
|
|
148
148
|
},
|
|
149
149
|
releasedAt: '2025-11-20',
|
|
150
|
+
settings: {
|
|
151
|
+
extendParams: ['thinkingLevel2', 'urlContext'],
|
|
152
|
+
searchImpl: 'params',
|
|
153
|
+
searchProvider: 'google',
|
|
154
|
+
},
|
|
150
155
|
type: 'chat',
|
|
151
156
|
},
|
|
152
157
|
{
|
|
@@ -170,7 +175,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
170
175
|
],
|
|
171
176
|
},
|
|
172
177
|
settings: {
|
|
173
|
-
extendParams: ['
|
|
178
|
+
extendParams: ['thinkingLevel2', 'urlContext'],
|
|
174
179
|
searchImpl: 'params',
|
|
175
180
|
searchProvider: 'google',
|
|
176
181
|
},
|
|
@@ -348,7 +353,8 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
348
353
|
reasoning: true,
|
|
349
354
|
},
|
|
350
355
|
contextWindowTokens: 256_000,
|
|
351
|
-
description:
|
|
356
|
+
description:
|
|
357
|
+
"Grok 4 is xAI's flagship reasoning model with strong reasoning and multimodal capability.",
|
|
352
358
|
displayName: 'Grok 4',
|
|
353
359
|
id: 'x-ai/grok-4',
|
|
354
360
|
maxOutput: 256_000,
|
|
@@ -707,7 +713,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
707
713
|
},
|
|
708
714
|
contextWindowTokens: 1_050_000,
|
|
709
715
|
description:
|
|
710
|
-
|
|
716
|
+
'Gemini 2.5 Pro is Google’s flagship reasoning model with long context support for complex tasks.',
|
|
711
717
|
displayName: 'Gemini 2.5 Pro',
|
|
712
718
|
id: 'google/gemini-2.5-pro',
|
|
713
719
|
maxOutput: 65_530,
|
|
@@ -759,7 +765,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
759
765
|
},
|
|
760
766
|
contextWindowTokens: 1_050_000,
|
|
761
767
|
description:
|
|
762
|
-
|
|
768
|
+
'Gemini 2.5 Flash (Lite/Pro/Flash) is Google’s family spanning low latency to high-performance reasoning.',
|
|
763
769
|
displayName: 'Gemini 2.5 Flash',
|
|
764
770
|
id: 'google/gemini-2.5-flash',
|
|
765
771
|
maxOutput: 65_530,
|
|
@@ -931,8 +937,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
931
937
|
reasoning: true,
|
|
932
938
|
},
|
|
933
939
|
contextWindowTokens: 1_050_000,
|
|
934
|
-
description:
|
|
935
|
-
'GPT-4.1 Mini offers lower latency and better value for mid-context workloads.',
|
|
940
|
+
description: 'GPT-4.1 Mini offers lower latency and better value for mid-context workloads.',
|
|
936
941
|
displayName: 'GPT-4.1 Mini',
|
|
937
942
|
id: 'openai/gpt-4.1-mini',
|
|
938
943
|
maxOutput: 32_770,
|
|
@@ -1076,7 +1081,8 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
1076
1081
|
reasoning: true,
|
|
1077
1082
|
},
|
|
1078
1083
|
contextWindowTokens: 200_000,
|
|
1079
|
-
description:
|
|
1084
|
+
description:
|
|
1085
|
+
"GLM 4.6 is Z.AI's flagship model with extended context length and coding capability.",
|
|
1080
1086
|
displayName: 'GLM 4.6',
|
|
1081
1087
|
id: 'z-ai/glm-4.6',
|
|
1082
1088
|
maxOutput: 128_000,
|
|
@@ -1125,7 +1131,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
1125
1131
|
},
|
|
1126
1132
|
contextWindowTokens: 200_000,
|
|
1127
1133
|
description:
|
|
1128
|
-
|
|
1134
|
+
'Opus 4.1 is Anthropic’s high-end model optimized for programming, complex reasoning, and long-running tasks.',
|
|
1129
1135
|
displayName: 'Claude Opus 4.1',
|
|
1130
1136
|
id: 'anthropic/claude-opus-4.1',
|
|
1131
1137
|
maxOutput: 32_000,
|
|
@@ -1150,7 +1156,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
1150
1156
|
},
|
|
1151
1157
|
contextWindowTokens: 200_000,
|
|
1152
1158
|
description:
|
|
1153
|
-
|
|
1159
|
+
'Opus 4 is Anthropic’s flagship model designed for complex tasks and enterprise applications.',
|
|
1154
1160
|
displayName: 'Claude Opus 4',
|
|
1155
1161
|
id: 'anthropic/claude-opus-4',
|
|
1156
1162
|
maxOutput: 32_000,
|
|
@@ -1218,7 +1224,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
1218
1224
|
},
|
|
1219
1225
|
contextWindowTokens: 200_000,
|
|
1220
1226
|
description:
|
|
1221
|
-
|
|
1227
|
+
'Claude Sonnet 4.5 is Anthropic’s latest hybrid reasoning model optimized for complex reasoning and coding.',
|
|
1222
1228
|
displayName: 'Claude Sonnet 4.5',
|
|
1223
1229
|
id: 'anthropic/claude-sonnet-4.5',
|
|
1224
1230
|
maxOutput: 64_000,
|
|
@@ -1259,7 +1265,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
1259
1265
|
reasoning: true,
|
|
1260
1266
|
},
|
|
1261
1267
|
contextWindowTokens: 200_000,
|
|
1262
|
-
description:
|
|
1268
|
+
description: 'OpenAI o4-mini is a small, efficient reasoning model for low-latency scenarios.',
|
|
1263
1269
|
displayName: 'o4 Mini',
|
|
1264
1270
|
id: 'openai/o4-mini',
|
|
1265
1271
|
maxOutput: 100_000,
|
|
@@ -1542,7 +1548,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
1542
1548
|
abilities: { reasoning: true },
|
|
1543
1549
|
contextWindowTokens: 128_000,
|
|
1544
1550
|
description:
|
|
1545
|
-
|
|
1551
|
+
'GLM 4.5 is Z.AI’s flagship model with hybrid reasoning optimized for engineering and long-context tasks.',
|
|
1546
1552
|
displayName: 'GLM 4.5',
|
|
1547
1553
|
id: 'z-ai/glm-4.5',
|
|
1548
1554
|
maxOutput: 96_000,
|
|
@@ -12,9 +12,71 @@ const zhipuChatModels: AIChatModelCard[] = [
|
|
|
12
12
|
},
|
|
13
13
|
contextWindowTokens: 200_000,
|
|
14
14
|
description:
|
|
15
|
-
'
|
|
16
|
-
displayName: 'GLM-4.
|
|
15
|
+
'GLM-4.7 是智谱最新旗舰模型,面向 Agentic Coding 场景强化了编码能力、长程任务规划与工具协同,并在多个公开基准的当期榜单中取得开源模型中的领先表现。通用能力提升,回复更简洁自然,写作更具沉浸感。在执行复杂智能体任务,在工具调用时指令遵循更强,Artifacts 与 Agentic Coding 的前端美感和长程任务完成效率进一步提升。',
|
|
16
|
+
displayName: 'GLM-4.7',
|
|
17
17
|
enabled: true,
|
|
18
|
+
id: 'glm-4.7',
|
|
19
|
+
maxOutput: 131_072,
|
|
20
|
+
pricing: {
|
|
21
|
+
currency: 'CNY',
|
|
22
|
+
units: [
|
|
23
|
+
{
|
|
24
|
+
lookup: {
|
|
25
|
+
prices: {
|
|
26
|
+
'[0, 0.032]_[0, 0.0002]': 0.4,
|
|
27
|
+
'[0, 0.032]_[0.0002, infinity]': 0.6,
|
|
28
|
+
'[0.032, 0.2]': 0.8,
|
|
29
|
+
},
|
|
30
|
+
pricingParams: ['textInput', 'textOutput'],
|
|
31
|
+
},
|
|
32
|
+
name: 'textInput_cacheRead',
|
|
33
|
+
strategy: 'lookup',
|
|
34
|
+
unit: 'millionTokens',
|
|
35
|
+
},
|
|
36
|
+
{
|
|
37
|
+
lookup: {
|
|
38
|
+
prices: {
|
|
39
|
+
'[0, 0.032]_[0, 0.0002]': 2,
|
|
40
|
+
'[0, 0.032]_[0.0002, infinity]': 3,
|
|
41
|
+
'[0.032, 0.2]': 4,
|
|
42
|
+
},
|
|
43
|
+
pricingParams: ['textInput', 'textOutput'],
|
|
44
|
+
},
|
|
45
|
+
name: 'textInput',
|
|
46
|
+
strategy: 'lookup',
|
|
47
|
+
unit: 'millionTokens',
|
|
48
|
+
},
|
|
49
|
+
{
|
|
50
|
+
lookup: {
|
|
51
|
+
prices: {
|
|
52
|
+
'[0, 0.032]_[0, 0.0002]': 8,
|
|
53
|
+
'[0, 0.032]_[0.0002, infinity]': 14,
|
|
54
|
+
'[0.032, 0.2]': 16,
|
|
55
|
+
},
|
|
56
|
+
pricingParams: ['textInput', 'textOutput'],
|
|
57
|
+
},
|
|
58
|
+
name: 'textOutput',
|
|
59
|
+
strategy: 'lookup',
|
|
60
|
+
unit: 'millionTokens',
|
|
61
|
+
},
|
|
62
|
+
],
|
|
63
|
+
},
|
|
64
|
+
settings: {
|
|
65
|
+
extendParams: ['enableReasoning'],
|
|
66
|
+
searchImpl: 'params',
|
|
67
|
+
},
|
|
68
|
+
type: 'chat',
|
|
69
|
+
},
|
|
70
|
+
{
|
|
71
|
+
abilities: {
|
|
72
|
+
functionCall: true,
|
|
73
|
+
reasoning: true,
|
|
74
|
+
search: true,
|
|
75
|
+
},
|
|
76
|
+
contextWindowTokens: 200_000,
|
|
77
|
+
description:
|
|
78
|
+
'智谱最新旗舰模型 GLM-4.6 (355B) 在高级编码、长文本处理、推理与智能体能力上全面超越前代,尤其在编程能力上对齐 Claude Sonnet 4,成为国内顶尖的 Coding 模型。',
|
|
79
|
+
displayName: 'GLM-4.6',
|
|
18
80
|
id: 'glm-4.6',
|
|
19
81
|
maxOutput: 131_072,
|
|
20
82
|
pricing: {
|
|
@@ -239,10 +239,13 @@ export type ExtendParamsType =
|
|
|
239
239
|
| 'reasoningEffort'
|
|
240
240
|
| 'gpt5ReasoningEffort'
|
|
241
241
|
| 'gpt5_1ReasoningEffort'
|
|
242
|
+
| 'gpt5_2ReasoningEffort'
|
|
243
|
+
| 'gpt5_2ProReasoningEffort'
|
|
242
244
|
| 'textVerbosity'
|
|
243
245
|
| 'thinking'
|
|
244
246
|
| 'thinkingBudget'
|
|
245
247
|
| 'thinkingLevel'
|
|
248
|
+
| 'thinkingLevel2'
|
|
246
249
|
| 'imageAspectRatio'
|
|
247
250
|
| 'imageResolution'
|
|
248
251
|
| 'urlContext';
|
|
@@ -7,12 +7,13 @@ export const LobeVolcengineAI = createOpenAICompatibleRuntime({
|
|
|
7
7
|
baseURL: 'https://ark.cn-beijing.volces.com/api/v3',
|
|
8
8
|
chatCompletion: {
|
|
9
9
|
handlePayload: (payload) => {
|
|
10
|
-
const { model, thinking, ...rest } = payload;
|
|
10
|
+
const { model, thinking, reasoning_effort, ...rest } = payload;
|
|
11
11
|
|
|
12
12
|
return {
|
|
13
13
|
...rest,
|
|
14
14
|
model,
|
|
15
15
|
...(thinking?.type && { thinking: { type: thinking.type } }),
|
|
16
|
+
...(reasoning_effort && { reasoning_effort }),
|
|
16
17
|
} as any;
|
|
17
18
|
},
|
|
18
19
|
},
|