@lobehub/lobehub 2.0.0-next.185 → 2.0.0-next.187
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +50 -0
- package/changelog/v1.json +18 -0
- package/e2e/src/support/webServer.ts +0 -1
- package/next.config.ts +14 -1
- package/package.json +2 -1
- package/packages/fetch-sse/src/__tests__/request.test.ts +608 -0
- package/packages/model-bank/src/aiModels/aihubmix.ts +44 -8
- package/packages/model-bank/src/aiModels/google.ts +49 -17
- package/packages/model-bank/src/aiModels/hunyuan.ts +20 -0
- package/packages/model-bank/src/aiModels/infiniai.ts +48 -7
- package/packages/model-bank/src/aiModels/lobehub.ts +13 -11
- package/packages/model-bank/src/aiModels/minimax.ts +46 -2
- package/packages/model-bank/src/aiModels/ollamacloud.ts +40 -5
- package/packages/model-bank/src/aiModels/openai.ts +6 -3
- package/packages/model-bank/src/aiModels/qwen.ts +1 -1
- package/packages/model-bank/src/aiModels/siliconcloud.ts +60 -0
- package/packages/model-bank/src/aiModels/vertexai.ts +77 -44
- package/packages/model-bank/src/aiModels/volcengine.ts +111 -2
- package/packages/model-bank/src/aiModels/zenmux.ts +19 -13
- package/packages/model-bank/src/aiModels/zhipu.ts +64 -2
- package/packages/model-bank/src/types/aiModel.ts +3 -0
- package/packages/model-runtime/src/providers/volcengine/index.ts +2 -1
- package/packages/model-runtime/src/providers/zhipu/index.test.ts +0 -27
- package/packages/model-runtime/src/providers/zhipu/index.ts +1 -1
- package/packages/model-runtime/src/utils/modelParse.ts +26 -21
- package/packages/types/src/agent/chatConfig.ts +6 -2
- package/packages/types/src/message/common/base.ts +1 -1
- package/packages/types/src/message/ui/chat.ts +3 -3
- package/packages/types/src/message/ui/extra.ts +2 -2
- package/packages/types/src/openai/plugin.ts +1 -1
- package/packages/types/src/session/agentSession.ts +1 -1
- package/packages/types/src/user/settings/index.ts +1 -1
- package/src/app/(backend)/oidc/[...oidc]/route.ts +2 -2
- package/src/app/[variants]/(auth)/oauth/consent/[uid]/page.tsx +2 -2
- package/src/envs/auth.ts +2 -0
- package/src/features/ChatInput/ActionBar/Model/ControlsForm.tsx +40 -1
- package/src/features/ChatInput/ActionBar/Model/GPT52ProReasoningEffortSlider.tsx +59 -0
- package/src/features/ChatInput/ActionBar/Model/GPT52ReasoningEffortSlider.tsx +61 -0
- package/src/features/ChatInput/ActionBar/Model/TextVerbositySlider.tsx +1 -1
- package/src/features/ChatInput/ActionBar/Model/ThinkingLevel2Slider.tsx +58 -0
- package/src/features/ChatInput/ActionBar/Model/ThinkingLevelSlider.tsx +10 -8
- package/src/libs/next/config/define-config.ts +19 -10
- package/src/libs/next/proxy/define-config.ts +4 -5
- package/src/libs/trpc/lambda/context.ts +9 -11
- package/src/server/services/oidc/oidcProvider.ts +2 -2
- package/src/services/chat/mecha/modelParamsResolver.ts +11 -0
- package/vercel.json +1 -1
- package/src/envs/oidc.ts +0 -18
|
@@ -25,7 +25,7 @@ const aihubmixModels: AIChatModelCard[] = [
|
|
|
25
25
|
},
|
|
26
26
|
releasedAt: '2025-12-11',
|
|
27
27
|
settings: {
|
|
28
|
-
extendParams: ['
|
|
28
|
+
extendParams: ['gpt5_2ReasoningEffort', 'textVerbosity'],
|
|
29
29
|
searchImpl: 'params',
|
|
30
30
|
},
|
|
31
31
|
type: 'chat',
|
|
@@ -51,6 +51,7 @@ const aihubmixModels: AIChatModelCard[] = [
|
|
|
51
51
|
},
|
|
52
52
|
releasedAt: '2025-12-11',
|
|
53
53
|
settings: {
|
|
54
|
+
extendParams: ['gpt5_2ProReasoningEffort', 'textVerbosity'],
|
|
54
55
|
searchImpl: 'params',
|
|
55
56
|
},
|
|
56
57
|
type: 'chat',
|
|
@@ -189,7 +190,8 @@ const aihubmixModels: AIChatModelCard[] = [
|
|
|
189
190
|
vision: true,
|
|
190
191
|
},
|
|
191
192
|
contextWindowTokens: 400_000,
|
|
192
|
-
description:
|
|
193
|
+
description:
|
|
194
|
+
'GPT-5 pro uses more compute to think deeper and consistently deliver better answers.',
|
|
193
195
|
displayName: 'GPT-5 pro',
|
|
194
196
|
id: 'gpt-5-pro',
|
|
195
197
|
maxOutput: 272_000,
|
|
@@ -478,7 +480,8 @@ const aihubmixModels: AIChatModelCard[] = [
|
|
|
478
480
|
vision: true,
|
|
479
481
|
},
|
|
480
482
|
contextWindowTokens: 1_047_576,
|
|
481
|
-
description:
|
|
483
|
+
description:
|
|
484
|
+
'GPT-4.1 is our flagship model for complex tasks and cross-domain problem solving.',
|
|
482
485
|
displayName: 'GPT-4.1',
|
|
483
486
|
id: 'gpt-4.1',
|
|
484
487
|
maxOutput: 32_768,
|
|
@@ -502,8 +505,7 @@ const aihubmixModels: AIChatModelCard[] = [
|
|
|
502
505
|
vision: true,
|
|
503
506
|
},
|
|
504
507
|
contextWindowTokens: 1_047_576,
|
|
505
|
-
description:
|
|
506
|
-
'GPT-4.1 mini balances intelligence, speed, and cost for many use cases.',
|
|
508
|
+
description: 'GPT-4.1 mini balances intelligence, speed, and cost for many use cases.',
|
|
507
509
|
displayName: 'GPT-4.1 mini',
|
|
508
510
|
id: 'gpt-4.1-mini',
|
|
509
511
|
maxOutput: 32_768,
|
|
@@ -1137,6 +1139,41 @@ const aihubmixModels: AIChatModelCard[] = [
|
|
|
1137
1139
|
],
|
|
1138
1140
|
},
|
|
1139
1141
|
releasedAt: '2025-11-18',
|
|
1142
|
+
settings: {
|
|
1143
|
+
extendParams: ['thinkingLevel2', 'urlContext'],
|
|
1144
|
+
searchImpl: 'params',
|
|
1145
|
+
searchProvider: 'google',
|
|
1146
|
+
},
|
|
1147
|
+
type: 'chat',
|
|
1148
|
+
},
|
|
1149
|
+
{
|
|
1150
|
+
abilities: {
|
|
1151
|
+
functionCall: true,
|
|
1152
|
+
reasoning: true,
|
|
1153
|
+
search: true,
|
|
1154
|
+
video: true,
|
|
1155
|
+
vision: true,
|
|
1156
|
+
},
|
|
1157
|
+
contextWindowTokens: 1_048_576 + 65_536,
|
|
1158
|
+
description: 'Gemini 3 Flash 是为速度而打造的最智能的模型,将前沿智能与卓越的搜索接地相结合。',
|
|
1159
|
+
displayName: 'Gemini 3 Flash Preview',
|
|
1160
|
+
enabled: true,
|
|
1161
|
+
id: 'gemini-3-flash-preview',
|
|
1162
|
+
maxOutput: 65_536,
|
|
1163
|
+
pricing: {
|
|
1164
|
+
units: [
|
|
1165
|
+
{ name: 'textInput_cacheRead', rate: 0.05, strategy: 'fixed', unit: 'millionTokens' },
|
|
1166
|
+
{ name: 'textInput', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
1167
|
+
{ name: 'textOutput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
|
|
1168
|
+
{
|
|
1169
|
+
lookup: { prices: { '1h': 1 }, pricingParams: ['ttl'] },
|
|
1170
|
+
name: 'textInput_cacheWrite',
|
|
1171
|
+
strategy: 'lookup',
|
|
1172
|
+
unit: 'millionTokens',
|
|
1173
|
+
},
|
|
1174
|
+
],
|
|
1175
|
+
},
|
|
1176
|
+
releasedAt: '2025-12-17',
|
|
1140
1177
|
settings: {
|
|
1141
1178
|
extendParams: ['thinkingLevel', 'urlContext'],
|
|
1142
1179
|
searchImpl: 'params',
|
|
@@ -1210,7 +1247,7 @@ const aihubmixModels: AIChatModelCard[] = [
|
|
|
1210
1247
|
vision: true,
|
|
1211
1248
|
},
|
|
1212
1249
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
1213
|
-
description:
|
|
1250
|
+
description: 'Gemini 2.5 Flash is Google’s best-value model with full capabilities.',
|
|
1214
1251
|
displayName: 'Gemini 2.5 Flash',
|
|
1215
1252
|
id: 'gemini-2.5-flash',
|
|
1216
1253
|
maxOutput: 65_536,
|
|
@@ -1238,7 +1275,6 @@ const aihubmixModels: AIChatModelCard[] = [
|
|
|
1238
1275
|
description:
|
|
1239
1276
|
'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, allowing image generation and editing through conversation.',
|
|
1240
1277
|
displayName: 'Nano Banana',
|
|
1241
|
-
enabled: true,
|
|
1242
1278
|
id: 'gemini-2.5-flash-image',
|
|
1243
1279
|
maxOutput: 8192,
|
|
1244
1280
|
pricing: {
|
|
@@ -1262,7 +1298,7 @@ const aihubmixModels: AIChatModelCard[] = [
|
|
|
1262
1298
|
},
|
|
1263
1299
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
1264
1300
|
description:
|
|
1265
|
-
|
|
1301
|
+
'Gemini 2.5 Flash-Lite is Google’s smallest, best-value model, designed for large-scale use.',
|
|
1266
1302
|
displayName: 'Gemini 2.5 Flash-Lite',
|
|
1267
1303
|
id: 'gemini-2.5-flash-lite',
|
|
1268
1304
|
maxOutput: 65_536,
|
|
@@ -123,7 +123,7 @@ const googleChatModels: AIChatModelCard[] = [
|
|
|
123
123
|
},
|
|
124
124
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
125
125
|
description:
|
|
126
|
-
|
|
126
|
+
'Gemini 3 Pro is Google’s most powerful agent and vibe-coding model, delivering richer visuals and deeper interaction on top of state-of-the-art reasoning.',
|
|
127
127
|
displayName: 'Gemini 3 Pro Preview',
|
|
128
128
|
enabled: true,
|
|
129
129
|
id: 'gemini-3-pro-preview',
|
|
@@ -166,6 +166,41 @@ const googleChatModels: AIChatModelCard[] = [
|
|
|
166
166
|
],
|
|
167
167
|
},
|
|
168
168
|
releasedAt: '2025-11-18',
|
|
169
|
+
settings: {
|
|
170
|
+
extendParams: ['thinkingLevel2', 'urlContext'],
|
|
171
|
+
searchImpl: 'params',
|
|
172
|
+
searchProvider: 'google',
|
|
173
|
+
},
|
|
174
|
+
type: 'chat',
|
|
175
|
+
},
|
|
176
|
+
{
|
|
177
|
+
abilities: {
|
|
178
|
+
functionCall: true,
|
|
179
|
+
reasoning: true,
|
|
180
|
+
search: true,
|
|
181
|
+
video: true,
|
|
182
|
+
vision: true,
|
|
183
|
+
},
|
|
184
|
+
contextWindowTokens: 1_048_576 + 65_536,
|
|
185
|
+
description: 'Gemini 3 Flash 是为速度而打造的最智能的模型,将前沿智能与卓越的搜索接地相结合。',
|
|
186
|
+
displayName: 'Gemini 3 Flash Preview',
|
|
187
|
+
enabled: true,
|
|
188
|
+
id: 'gemini-3-flash-preview',
|
|
189
|
+
maxOutput: 65_536,
|
|
190
|
+
pricing: {
|
|
191
|
+
units: [
|
|
192
|
+
{ name: 'textInput_cacheRead', rate: 0.05, strategy: 'fixed', unit: 'millionTokens' },
|
|
193
|
+
{ name: 'textInput', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
194
|
+
{ name: 'textOutput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
|
|
195
|
+
{
|
|
196
|
+
lookup: { prices: { '1h': 1 }, pricingParams: ['ttl'] },
|
|
197
|
+
name: 'textInput_cacheWrite',
|
|
198
|
+
strategy: 'lookup',
|
|
199
|
+
unit: 'millionTokens',
|
|
200
|
+
},
|
|
201
|
+
],
|
|
202
|
+
},
|
|
203
|
+
releasedAt: '2025-12-17',
|
|
169
204
|
settings: {
|
|
170
205
|
extendParams: ['thinkingLevel', 'urlContext'],
|
|
171
206
|
searchImpl: 'params',
|
|
@@ -182,7 +217,7 @@ const googleChatModels: AIChatModelCard[] = [
|
|
|
182
217
|
},
|
|
183
218
|
contextWindowTokens: 131_072 + 32_768,
|
|
184
219
|
description:
|
|
185
|
-
|
|
220
|
+
'Gemini 3 Pro Image (Nano Banana Pro) is Google’s image generation model and also supports multimodal chat.',
|
|
186
221
|
displayName: 'Nano Banana Pro',
|
|
187
222
|
enabled: true,
|
|
188
223
|
id: 'gemini-3-pro-image-preview',
|
|
@@ -213,9 +248,8 @@ const googleChatModels: AIChatModelCard[] = [
|
|
|
213
248
|
},
|
|
214
249
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
215
250
|
description:
|
|
216
|
-
|
|
251
|
+
'Gemini 2.5 Pro is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
|
|
217
252
|
displayName: 'Gemini 2.5 Pro',
|
|
218
|
-
enabled: true,
|
|
219
253
|
id: 'gemini-2.5-pro',
|
|
220
254
|
maxOutput: 65_536,
|
|
221
255
|
pricing: {
|
|
@@ -273,7 +307,7 @@ const googleChatModels: AIChatModelCard[] = [
|
|
|
273
307
|
},
|
|
274
308
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
275
309
|
description:
|
|
276
|
-
|
|
310
|
+
'Gemini 2.5 Pro Preview is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
|
|
277
311
|
displayName: 'Gemini 2.5 Pro Preview 06-05',
|
|
278
312
|
id: 'gemini-2.5-pro-preview-06-05',
|
|
279
313
|
maxOutput: 65_536,
|
|
@@ -326,7 +360,7 @@ const googleChatModels: AIChatModelCard[] = [
|
|
|
326
360
|
},
|
|
327
361
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
328
362
|
description:
|
|
329
|
-
|
|
363
|
+
'Gemini 2.5 Pro Preview is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
|
|
330
364
|
displayName: 'Gemini 2.5 Pro Preview 05-06',
|
|
331
365
|
id: 'gemini-2.5-pro-preview-05-06',
|
|
332
366
|
maxOutput: 65_536,
|
|
@@ -377,9 +411,8 @@ const googleChatModels: AIChatModelCard[] = [
|
|
|
377
411
|
vision: true,
|
|
378
412
|
},
|
|
379
413
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
380
|
-
description:
|
|
414
|
+
description: 'Gemini 2.5 Flash is Google’s best-value model with full capabilities.',
|
|
381
415
|
displayName: 'Gemini 2.5 Flash',
|
|
382
|
-
enabled: true,
|
|
383
416
|
id: 'gemini-2.5-flash',
|
|
384
417
|
maxOutput: 65_536,
|
|
385
418
|
pricing: {
|
|
@@ -432,9 +465,8 @@ const googleChatModels: AIChatModelCard[] = [
|
|
|
432
465
|
},
|
|
433
466
|
contextWindowTokens: 32_768 + 8192,
|
|
434
467
|
description:
|
|
435
|
-
|
|
468
|
+
'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
|
|
436
469
|
displayName: 'Nano Banana',
|
|
437
|
-
enabled: true,
|
|
438
470
|
id: 'gemini-2.5-flash-image',
|
|
439
471
|
maxOutput: 8192,
|
|
440
472
|
pricing: {
|
|
@@ -459,7 +491,7 @@ const googleChatModels: AIChatModelCard[] = [
|
|
|
459
491
|
},
|
|
460
492
|
contextWindowTokens: 32_768 + 8192,
|
|
461
493
|
description:
|
|
462
|
-
|
|
494
|
+
'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
|
|
463
495
|
displayName: 'Nano Banana (Preview)',
|
|
464
496
|
id: 'gemini-2.5-flash-image-preview',
|
|
465
497
|
maxOutput: 8192,
|
|
@@ -488,7 +520,7 @@ const googleChatModels: AIChatModelCard[] = [
|
|
|
488
520
|
},
|
|
489
521
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
490
522
|
description:
|
|
491
|
-
|
|
523
|
+
'Gemini 2.5 Flash-Lite is Google’s smallest, best-value model, designed for large-scale use.',
|
|
492
524
|
displayName: 'Gemini 2.5 Flash-Lite',
|
|
493
525
|
id: 'gemini-2.5-flash-lite',
|
|
494
526
|
maxOutput: 65_536,
|
|
@@ -706,7 +738,8 @@ const googleChatModels: AIChatModelCard[] = [
|
|
|
706
738
|
vision: true,
|
|
707
739
|
},
|
|
708
740
|
contextWindowTokens: 1_008_192,
|
|
709
|
-
description:
|
|
741
|
+
description:
|
|
742
|
+
'Gemini 1.5 Flash 002 is an efficient multimodal model built for broad deployment.',
|
|
710
743
|
displayName: 'Gemini 1.5 Flash 002',
|
|
711
744
|
id: 'gemini-1.5-flash-002', // Deprecated on 2025-09-24
|
|
712
745
|
maxOutput: 8192,
|
|
@@ -903,7 +936,7 @@ const googleImageModels: AIImageModelCard[] = [
|
|
|
903
936
|
type: 'image',
|
|
904
937
|
enabled: true,
|
|
905
938
|
description:
|
|
906
|
-
|
|
939
|
+
'Gemini 3 Pro Image (Nano Banana Pro) is Google’s image generation model and also supports multimodal chat.',
|
|
907
940
|
releasedAt: '2025-11-18',
|
|
908
941
|
parameters: nanoBananaProParameters,
|
|
909
942
|
pricing: {
|
|
@@ -918,10 +951,9 @@ const googleImageModels: AIImageModelCard[] = [
|
|
|
918
951
|
{
|
|
919
952
|
displayName: 'Nano Banana',
|
|
920
953
|
id: 'gemini-2.5-flash-image:image',
|
|
921
|
-
enabled: true,
|
|
922
954
|
type: 'image',
|
|
923
955
|
description:
|
|
924
|
-
|
|
956
|
+
'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
|
|
925
957
|
releasedAt: '2025-08-26',
|
|
926
958
|
parameters: nanoBananaParameters,
|
|
927
959
|
pricing: {
|
|
@@ -938,7 +970,7 @@ const googleImageModels: AIImageModelCard[] = [
|
|
|
938
970
|
id: 'gemini-2.5-flash-image-preview:image',
|
|
939
971
|
type: 'image',
|
|
940
972
|
description:
|
|
941
|
-
|
|
973
|
+
'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
|
|
942
974
|
releasedAt: '2025-08-26',
|
|
943
975
|
parameters: CHAT_MODEL_IMAGE_GENERATION_PARAMS,
|
|
944
976
|
pricing: {
|
|
@@ -446,6 +446,26 @@ const hunyuanChatModels: AIChatModelCard[] = [
|
|
|
446
446
|
},
|
|
447
447
|
type: 'chat',
|
|
448
448
|
},
|
|
449
|
+
{
|
|
450
|
+
abilities: {
|
|
451
|
+
vision: true,
|
|
452
|
+
},
|
|
453
|
+
contextWindowTokens: 24_000,
|
|
454
|
+
description:
|
|
455
|
+
'基于文本 TurboS 基座生产的图生文快思考模型,相比上一版本在图像基础识别、图像分析推理等维度都有明显的效果提升。',
|
|
456
|
+
displayName: '混元图生文',
|
|
457
|
+
id: 'hunyuan-vision-1.5-instruct',
|
|
458
|
+
maxOutput: 16_000,
|
|
459
|
+
pricing: {
|
|
460
|
+
currency: 'CNY',
|
|
461
|
+
units: [
|
|
462
|
+
{ name: 'textInput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
|
|
463
|
+
{ name: 'textOutput', rate: 9, strategy: 'fixed', unit: 'millionTokens' },
|
|
464
|
+
],
|
|
465
|
+
},
|
|
466
|
+
releasedAt: '2025-12-17',
|
|
467
|
+
type: 'chat',
|
|
468
|
+
},
|
|
449
469
|
{
|
|
450
470
|
abilities: {
|
|
451
471
|
vision: true,
|
|
@@ -51,9 +51,29 @@ const infiniaiChatModels: AIChatModelCard[] = [
|
|
|
51
51
|
},
|
|
52
52
|
contextWindowTokens: 200_000,
|
|
53
53
|
description:
|
|
54
|
-
'MiniMax-M2
|
|
55
|
-
displayName: 'MiniMax M2',
|
|
54
|
+
'MiniMax-M2.1 是 MiniMax 系列的最新版本,专为多语言编程和真实世界复杂任务优化。作为一款 AI 原生模型,MiniMax-M2.1 在模型性能、智能体框架支持以及多场景适配方面实现了显著提升,旨在帮助企业和个人更快地找到 AI 原生的工作与生活方式。',
|
|
55
|
+
displayName: 'MiniMax M2.1',
|
|
56
56
|
enabled: true,
|
|
57
|
+
id: 'minimax-m2.1',
|
|
58
|
+
maxOutput: 200_000,
|
|
59
|
+
pricing: {
|
|
60
|
+
currency: 'CNY',
|
|
61
|
+
units: [
|
|
62
|
+
{ name: 'textInput', rate: 2.1, strategy: 'fixed', unit: 'millionTokens' },
|
|
63
|
+
{ name: 'textOutput', rate: 8.4, strategy: 'fixed', unit: 'millionTokens' },
|
|
64
|
+
],
|
|
65
|
+
},
|
|
66
|
+
type: 'chat',
|
|
67
|
+
},
|
|
68
|
+
{
|
|
69
|
+
abilities: {
|
|
70
|
+
functionCall: true,
|
|
71
|
+
reasoning: true,
|
|
72
|
+
},
|
|
73
|
+
contextWindowTokens: 200_000,
|
|
74
|
+
description:
|
|
75
|
+
'MiniMax-M2 是一款专为编码与智能体工作流优化的专家混合(MoE)语言模型,具有约 230B 总参数与约 10B 活跃参数。它在保持强通用智能的同时,针对多文件编辑、代码-运行-修复闭环、测试校验修复等开发者场景进行深度增强,在终端、IDE 与 CI 等真实环境中表现稳定、高效。',
|
|
76
|
+
displayName: 'MiniMax M2',
|
|
57
77
|
id: 'minimax-m2',
|
|
58
78
|
maxOutput: 200_000,
|
|
59
79
|
pricing: {
|
|
@@ -71,16 +91,38 @@ const infiniaiChatModels: AIChatModelCard[] = [
|
|
|
71
91
|
reasoning: true,
|
|
72
92
|
},
|
|
73
93
|
contextWindowTokens: 131_072,
|
|
74
|
-
description: 'GLM-4.
|
|
75
|
-
displayName: 'GLM-4.
|
|
94
|
+
description: 'GLM-4.7 是智谱AI推出的最新大语言模型,具备更强的推理和生成能力。',
|
|
95
|
+
displayName: 'GLM-4.7',
|
|
76
96
|
enabled: true,
|
|
97
|
+
id: 'glm-4.7',
|
|
98
|
+
maxOutput: 4096,
|
|
99
|
+
pricing: {
|
|
100
|
+
currency: 'CNY',
|
|
101
|
+
units: [
|
|
102
|
+
{ name: 'textInput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
|
|
103
|
+
{ name: 'textOutput', rate: 14, strategy: 'fixed', unit: 'millionTokens' },
|
|
104
|
+
],
|
|
105
|
+
},
|
|
106
|
+
settings: {
|
|
107
|
+
extendParams: ['enableReasoning'],
|
|
108
|
+
},
|
|
109
|
+
type: 'chat',
|
|
110
|
+
},
|
|
111
|
+
{
|
|
112
|
+
abilities: {
|
|
113
|
+
functionCall: true,
|
|
114
|
+
reasoning: true,
|
|
115
|
+
},
|
|
116
|
+
contextWindowTokens: 131_072,
|
|
117
|
+
description: 'GLM-4.6 是智谱AI推出的最新大语言模型,具备更强的推理和生成能力。',
|
|
118
|
+
displayName: 'GLM-4.6',
|
|
77
119
|
id: 'glm-4.6',
|
|
78
120
|
maxOutput: 4096,
|
|
79
121
|
pricing: {
|
|
80
122
|
currency: 'CNY',
|
|
81
123
|
units: [
|
|
82
|
-
{ name: 'textInput', rate:
|
|
83
|
-
{ name: 'textOutput', rate:
|
|
124
|
+
{ name: 'textInput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
|
|
125
|
+
{ name: 'textOutput', rate: 14, strategy: 'fixed', unit: 'millionTokens' },
|
|
84
126
|
],
|
|
85
127
|
},
|
|
86
128
|
settings: {
|
|
@@ -96,7 +138,6 @@ const infiniaiChatModels: AIChatModelCard[] = [
|
|
|
96
138
|
contextWindowTokens: 131_072,
|
|
97
139
|
description: 'DeepSeek-V3.2-Exp is an experimental DeepSeek LLM with stronger reasoning and generation.',
|
|
98
140
|
displayName: 'DeepSeek V3.2 Exp',
|
|
99
|
-
enabled: true,
|
|
100
141
|
id: 'deepseek-v3.2-exp',
|
|
101
142
|
maxOutput: 65_536,
|
|
102
143
|
pricing: {
|
|
@@ -12,7 +12,8 @@ const lobehubChatModels: AIChatModelCard[] = [
|
|
|
12
12
|
vision: true,
|
|
13
13
|
},
|
|
14
14
|
contextWindowTokens: 400_000,
|
|
15
|
-
description:
|
|
15
|
+
description:
|
|
16
|
+
"GPT-5.2 — OpenAI's flagship model optimized for coding and agent tasks across industries.",
|
|
16
17
|
displayName: 'GPT-5.2',
|
|
17
18
|
enabled: true,
|
|
18
19
|
id: 'gpt-5.2',
|
|
@@ -26,7 +27,7 @@ const lobehubChatModels: AIChatModelCard[] = [
|
|
|
26
27
|
},
|
|
27
28
|
releasedAt: '2025-12-11',
|
|
28
29
|
settings: {
|
|
29
|
-
extendParams: ['
|
|
30
|
+
extendParams: ['gpt5_2ReasoningEffort', 'textVerbosity'],
|
|
30
31
|
searchImpl: 'params',
|
|
31
32
|
},
|
|
32
33
|
type: 'chat',
|
|
@@ -164,7 +165,8 @@ const lobehubChatModels: AIChatModelCard[] = [
|
|
|
164
165
|
vision: true,
|
|
165
166
|
},
|
|
166
167
|
contextWindowTokens: 1_047_576,
|
|
167
|
-
description:
|
|
168
|
+
description:
|
|
169
|
+
'GPT-4.1 is our flagship model for complex tasks and cross-domain problem solving.',
|
|
168
170
|
displayName: 'GPT-4.1',
|
|
169
171
|
enabled: true,
|
|
170
172
|
id: 'gpt-4.1',
|
|
@@ -628,7 +630,7 @@ const lobehubChatModels: AIChatModelCard[] = [
|
|
|
628
630
|
},
|
|
629
631
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
630
632
|
description:
|
|
631
|
-
|
|
633
|
+
'Gemini 3 Pro is Google’s most powerful agent and vibe-coding model, delivering richer visuals and deeper interaction on top of state-of-the-art reasoning.',
|
|
632
634
|
displayName: 'Gemini 3 Pro Preview',
|
|
633
635
|
enabled: true,
|
|
634
636
|
id: 'gemini-3-pro-preview',
|
|
@@ -672,7 +674,7 @@ const lobehubChatModels: AIChatModelCard[] = [
|
|
|
672
674
|
},
|
|
673
675
|
releasedAt: '2025-11-18',
|
|
674
676
|
settings: {
|
|
675
|
-
extendParams: ['
|
|
677
|
+
extendParams: ['thinkingLevel2', 'urlContext'],
|
|
676
678
|
searchImpl: 'params',
|
|
677
679
|
searchProvider: 'google',
|
|
678
680
|
},
|
|
@@ -728,7 +730,7 @@ const lobehubChatModels: AIChatModelCard[] = [
|
|
|
728
730
|
},
|
|
729
731
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
730
732
|
description:
|
|
731
|
-
|
|
733
|
+
'Gemini 2.5 Pro is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
|
|
732
734
|
displayName: 'Gemini 2.5 Pro',
|
|
733
735
|
enabled: true,
|
|
734
736
|
id: 'gemini-2.5-pro',
|
|
@@ -780,7 +782,7 @@ const lobehubChatModels: AIChatModelCard[] = [
|
|
|
780
782
|
vision: true,
|
|
781
783
|
},
|
|
782
784
|
contextWindowTokens: 1_048_576 + 65_536,
|
|
783
|
-
description:
|
|
785
|
+
description: 'Gemini 2.5 Flash is Google’s best-value model with full capabilities.',
|
|
784
786
|
displayName: 'Gemini 2.5 Flash',
|
|
785
787
|
enabled: true,
|
|
786
788
|
id: 'gemini-2.5-flash',
|
|
@@ -819,7 +821,7 @@ const lobehubChatModels: AIChatModelCard[] = [
|
|
|
819
821
|
},
|
|
820
822
|
contextWindowTokens: 131_072 + 32_768,
|
|
821
823
|
description:
|
|
822
|
-
|
|
824
|
+
'Gemini 3 Pro Image (Nano Banana Pro) is Google’s image generation model and also supports multimodal chat.',
|
|
823
825
|
displayName: 'Nano Banana Pro',
|
|
824
826
|
enabled: true,
|
|
825
827
|
id: 'gemini-3-pro-image-preview',
|
|
@@ -845,7 +847,7 @@ const lobehubChatModels: AIChatModelCard[] = [
|
|
|
845
847
|
},
|
|
846
848
|
contextWindowTokens: 32_768 + 32_768,
|
|
847
849
|
description:
|
|
848
|
-
|
|
850
|
+
'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
|
|
849
851
|
displayName: 'Nano Banana',
|
|
850
852
|
enabled: true,
|
|
851
853
|
id: 'gemini-2.5-flash-image-preview',
|
|
@@ -1039,7 +1041,7 @@ export const nanoBananaProParameters: ModelParamsSchema = {
|
|
|
1039
1041
|
const lobehubImageModels: AIImageModelCard[] = [
|
|
1040
1042
|
{
|
|
1041
1043
|
description:
|
|
1042
|
-
|
|
1044
|
+
'Gemini 3 Pro Image (Nano Banana Pro) is Google’s image generation model and also supports multimodal chat.',
|
|
1043
1045
|
displayName: 'Nano Banana Pro',
|
|
1044
1046
|
enabled: true,
|
|
1045
1047
|
id: 'gemini-3-pro-image-preview:image',
|
|
@@ -1057,7 +1059,7 @@ const lobehubImageModels: AIImageModelCard[] = [
|
|
|
1057
1059
|
},
|
|
1058
1060
|
{
|
|
1059
1061
|
description:
|
|
1060
|
-
|
|
1062
|
+
'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
|
|
1061
1063
|
displayName: 'Nano Banana',
|
|
1062
1064
|
enabled: true,
|
|
1063
1065
|
id: 'gemini-2.5-flash-image-preview:image',
|
|
@@ -7,9 +7,53 @@ const minimaxChatModels: AIChatModelCard[] = [
|
|
|
7
7
|
reasoning: true,
|
|
8
8
|
},
|
|
9
9
|
contextWindowTokens: 204_800,
|
|
10
|
-
description: '
|
|
11
|
-
displayName: 'MiniMax M2',
|
|
10
|
+
description: '强大多语言编程实力,全面升级编程体验',
|
|
11
|
+
displayName: 'MiniMax M2.1',
|
|
12
12
|
enabled: true,
|
|
13
|
+
id: 'MiniMax-M2.1',
|
|
14
|
+
maxOutput: 131_072,
|
|
15
|
+
pricing: {
|
|
16
|
+
currency: 'CNY',
|
|
17
|
+
units: [
|
|
18
|
+
{ name: 'textInput_cacheRead', rate: 0.21, strategy: 'fixed', unit: 'millionTokens' },
|
|
19
|
+
{ name: 'textInput_cacheWrite', rate: 2.625, strategy: 'fixed', unit: 'millionTokens' },
|
|
20
|
+
{ name: 'textInput', rate: 2.1, strategy: 'fixed', unit: 'millionTokens' },
|
|
21
|
+
{ name: 'textOutput', rate: 8.4, strategy: 'fixed', unit: 'millionTokens' },
|
|
22
|
+
],
|
|
23
|
+
},
|
|
24
|
+
releasedAt: '2025-12-23',
|
|
25
|
+
type: 'chat',
|
|
26
|
+
},
|
|
27
|
+
{
|
|
28
|
+
abilities: {
|
|
29
|
+
functionCall: true,
|
|
30
|
+
reasoning: true,
|
|
31
|
+
},
|
|
32
|
+
contextWindowTokens: 204_800,
|
|
33
|
+
description: '强大多语言编程实力,全面升级编程体验。更快,更高效。',
|
|
34
|
+
displayName: 'MiniMax M2.1 Lightning',
|
|
35
|
+
id: 'MiniMax-M2.1-Lightning',
|
|
36
|
+
maxOutput: 131_072,
|
|
37
|
+
pricing: {
|
|
38
|
+
currency: 'CNY',
|
|
39
|
+
units: [
|
|
40
|
+
{ name: 'textInput_cacheRead', rate: 0.21, strategy: 'fixed', unit: 'millionTokens' },
|
|
41
|
+
{ name: 'textInput_cacheWrite', rate: 2.625, strategy: 'fixed', unit: 'millionTokens' },
|
|
42
|
+
{ name: 'textInput', rate: 2.1, strategy: 'fixed', unit: 'millionTokens' },
|
|
43
|
+
{ name: 'textOutput', rate: 16.8, strategy: 'fixed', unit: 'millionTokens' },
|
|
44
|
+
],
|
|
45
|
+
},
|
|
46
|
+
releasedAt: '2025-12-23',
|
|
47
|
+
type: 'chat',
|
|
48
|
+
},
|
|
49
|
+
{
|
|
50
|
+
abilities: {
|
|
51
|
+
functionCall: true,
|
|
52
|
+
reasoning: true,
|
|
53
|
+
},
|
|
54
|
+
contextWindowTokens: 204_800,
|
|
55
|
+
description: '专为高效编码与Agent工作流而生',
|
|
56
|
+
displayName: 'MiniMax M2',
|
|
13
57
|
id: 'MiniMax-M2',
|
|
14
58
|
maxOutput: 131_072,
|
|
15
59
|
pricing: {
|
|
@@ -42,11 +42,34 @@ const ollamaCloudModels: AIChatModelCard[] = [
|
|
|
42
42
|
abilities: {
|
|
43
43
|
functionCall: true,
|
|
44
44
|
reasoning: true,
|
|
45
|
+
vision: true,
|
|
45
46
|
},
|
|
46
|
-
contextWindowTokens:
|
|
47
|
-
description: '
|
|
48
|
-
displayName: '
|
|
47
|
+
contextWindowTokens: 1_048_576,
|
|
48
|
+
description: 'Gemini 3 Flash 是为速度而打造的最智能的模型,将前沿智能与卓越的搜索接地相结合。',
|
|
49
|
+
displayName: 'Gemini 3 Flash Preview',
|
|
50
|
+
id: 'gemini-3-flash-preview',
|
|
51
|
+
releasedAt: '2025-12-17',
|
|
52
|
+
type: 'chat',
|
|
53
|
+
},
|
|
54
|
+
{
|
|
55
|
+
abilities: {
|
|
56
|
+
functionCall: true,
|
|
57
|
+
reasoning: true,
|
|
58
|
+
},
|
|
59
|
+
contextWindowTokens: 204_800,
|
|
60
|
+
displayName: 'MiniMax M2.1',
|
|
49
61
|
enabled: true,
|
|
62
|
+
id: 'minimax-m2.1',
|
|
63
|
+
type: 'chat',
|
|
64
|
+
},
|
|
65
|
+
{
|
|
66
|
+
abilities: {
|
|
67
|
+
functionCall: true,
|
|
68
|
+
reasoning: true,
|
|
69
|
+
},
|
|
70
|
+
contextWindowTokens: 204_800,
|
|
71
|
+
description: 'MiniMax M2 是专为编码和代理工作流程构建的高效大型语言模型。',
|
|
72
|
+
displayName: 'MiniMax M2',
|
|
50
73
|
id: 'minimax-m2',
|
|
51
74
|
type: 'chat',
|
|
52
75
|
},
|
|
@@ -57,9 +80,21 @@ const ollamaCloudModels: AIChatModelCard[] = [
|
|
|
57
80
|
},
|
|
58
81
|
contextWindowTokens: 200_000,
|
|
59
82
|
description:
|
|
60
|
-
'
|
|
61
|
-
displayName: 'GLM-4.
|
|
83
|
+
'GLM-4.7 是智谱最新旗舰模型,面向 Agentic Coding 场景强化了编码能力、长程任务规划与工具协同,并在多个公开基准的当期榜单中取得开源模型中的领先表现。通用能力提升,回复更简洁自然,写作更具沉浸感。在执行复杂智能体任务,在工具调用时指令遵循更强,Artifacts 与 Agentic Coding 的前端美感和长程任务完成效率进一步提升。',
|
|
84
|
+
displayName: 'GLM-4.7',
|
|
62
85
|
enabled: true,
|
|
86
|
+
id: 'glm-4.7',
|
|
87
|
+
type: 'chat',
|
|
88
|
+
},
|
|
89
|
+
{
|
|
90
|
+
abilities: {
|
|
91
|
+
functionCall: true,
|
|
92
|
+
reasoning: true,
|
|
93
|
+
},
|
|
94
|
+
contextWindowTokens: 200_000,
|
|
95
|
+
description:
|
|
96
|
+
'智谱旗舰模型 GLM-4.6 (355B) 在高级编码、长文本处理、推理与智能体能力上全面超越前代,尤其在编程能力上对齐 Claude Sonnet 4,成为国内顶尖的 Coding 模型。',
|
|
97
|
+
displayName: 'GLM-4.6',
|
|
63
98
|
id: 'glm-4.6',
|
|
64
99
|
type: 'chat',
|
|
65
100
|
},
|
|
@@ -42,7 +42,7 @@ export const openaiChatModels: AIChatModelCard[] = [
|
|
|
42
42
|
},
|
|
43
43
|
releasedAt: '2025-12-11',
|
|
44
44
|
settings: {
|
|
45
|
-
extendParams: ['
|
|
45
|
+
extendParams: ['gpt5_2ReasoningEffort', 'textVerbosity'],
|
|
46
46
|
searchImpl: 'params',
|
|
47
47
|
},
|
|
48
48
|
type: 'chat',
|
|
@@ -68,6 +68,7 @@ export const openaiChatModels: AIChatModelCard[] = [
|
|
|
68
68
|
},
|
|
69
69
|
releasedAt: '2025-12-11',
|
|
70
70
|
settings: {
|
|
71
|
+
extendParams: ['gpt5_2ProReasoningEffort'],
|
|
71
72
|
searchImpl: 'params',
|
|
72
73
|
},
|
|
73
74
|
type: 'chat',
|
|
@@ -203,7 +204,8 @@ export const openaiChatModels: AIChatModelCard[] = [
|
|
|
203
204
|
vision: true,
|
|
204
205
|
},
|
|
205
206
|
contextWindowTokens: 400_000,
|
|
206
|
-
description:
|
|
207
|
+
description:
|
|
208
|
+
'GPT-5 pro uses more compute to think deeper and consistently deliver better answers.',
|
|
207
209
|
displayName: 'GPT-5 pro',
|
|
208
210
|
id: 'gpt-5-pro',
|
|
209
211
|
maxOutput: 272_000,
|
|
@@ -563,7 +565,8 @@ export const openaiChatModels: AIChatModelCard[] = [
|
|
|
563
565
|
vision: true,
|
|
564
566
|
},
|
|
565
567
|
contextWindowTokens: 1_047_576,
|
|
566
|
-
description:
|
|
568
|
+
description:
|
|
569
|
+
'GPT-4.1 is our flagship model for complex tasks and cross-domain problem solving.',
|
|
567
570
|
displayName: 'GPT-4.1',
|
|
568
571
|
id: 'gpt-4.1',
|
|
569
572
|
maxOutput: 32_768,
|