@lobehub/lobehub 2.0.0-next.186 → 2.0.0-next.187

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (31) hide show
  1. package/CHANGELOG.md +25 -0
  2. package/changelog/v1.json +9 -0
  3. package/package.json +1 -1
  4. package/packages/fetch-sse/src/__tests__/request.test.ts +608 -0
  5. package/packages/model-bank/src/aiModels/aihubmix.ts +44 -8
  6. package/packages/model-bank/src/aiModels/google.ts +49 -17
  7. package/packages/model-bank/src/aiModels/hunyuan.ts +20 -0
  8. package/packages/model-bank/src/aiModels/infiniai.ts +48 -7
  9. package/packages/model-bank/src/aiModels/lobehub.ts +13 -11
  10. package/packages/model-bank/src/aiModels/minimax.ts +46 -2
  11. package/packages/model-bank/src/aiModels/ollamacloud.ts +40 -5
  12. package/packages/model-bank/src/aiModels/openai.ts +6 -3
  13. package/packages/model-bank/src/aiModels/qwen.ts +1 -1
  14. package/packages/model-bank/src/aiModels/siliconcloud.ts +60 -0
  15. package/packages/model-bank/src/aiModels/vertexai.ts +77 -44
  16. package/packages/model-bank/src/aiModels/volcengine.ts +111 -2
  17. package/packages/model-bank/src/aiModels/zenmux.ts +19 -13
  18. package/packages/model-bank/src/aiModels/zhipu.ts +64 -2
  19. package/packages/model-bank/src/types/aiModel.ts +3 -0
  20. package/packages/model-runtime/src/providers/volcengine/index.ts +2 -1
  21. package/packages/model-runtime/src/providers/zhipu/index.test.ts +0 -27
  22. package/packages/model-runtime/src/providers/zhipu/index.ts +1 -1
  23. package/packages/model-runtime/src/utils/modelParse.ts +26 -21
  24. package/packages/types/src/agent/chatConfig.ts +6 -2
  25. package/src/features/ChatInput/ActionBar/Model/ControlsForm.tsx +40 -1
  26. package/src/features/ChatInput/ActionBar/Model/GPT52ProReasoningEffortSlider.tsx +59 -0
  27. package/src/features/ChatInput/ActionBar/Model/GPT52ReasoningEffortSlider.tsx +61 -0
  28. package/src/features/ChatInput/ActionBar/Model/TextVerbositySlider.tsx +1 -1
  29. package/src/features/ChatInput/ActionBar/Model/ThinkingLevel2Slider.tsx +58 -0
  30. package/src/features/ChatInput/ActionBar/Model/ThinkingLevelSlider.tsx +10 -8
  31. package/src/services/chat/mecha/modelParamsResolver.ts +11 -0
@@ -25,7 +25,7 @@ const aihubmixModels: AIChatModelCard[] = [
25
25
  },
26
26
  releasedAt: '2025-12-11',
27
27
  settings: {
28
- extendParams: ['gpt5_1ReasoningEffort', 'textVerbosity'],
28
+ extendParams: ['gpt5_2ReasoningEffort', 'textVerbosity'],
29
29
  searchImpl: 'params',
30
30
  },
31
31
  type: 'chat',
@@ -51,6 +51,7 @@ const aihubmixModels: AIChatModelCard[] = [
51
51
  },
52
52
  releasedAt: '2025-12-11',
53
53
  settings: {
54
+ extendParams: ['gpt5_2ProReasoningEffort', 'textVerbosity'],
54
55
  searchImpl: 'params',
55
56
  },
56
57
  type: 'chat',
@@ -189,7 +190,8 @@ const aihubmixModels: AIChatModelCard[] = [
189
190
  vision: true,
190
191
  },
191
192
  contextWindowTokens: 400_000,
192
- description: 'GPT-5 pro uses more compute to think deeper and consistently deliver better answers.',
193
+ description:
194
+ 'GPT-5 pro uses more compute to think deeper and consistently deliver better answers.',
193
195
  displayName: 'GPT-5 pro',
194
196
  id: 'gpt-5-pro',
195
197
  maxOutput: 272_000,
@@ -478,7 +480,8 @@ const aihubmixModels: AIChatModelCard[] = [
478
480
  vision: true,
479
481
  },
480
482
  contextWindowTokens: 1_047_576,
481
- description: 'GPT-4.1 is our flagship model for complex tasks and cross-domain problem solving.',
483
+ description:
484
+ 'GPT-4.1 is our flagship model for complex tasks and cross-domain problem solving.',
482
485
  displayName: 'GPT-4.1',
483
486
  id: 'gpt-4.1',
484
487
  maxOutput: 32_768,
@@ -502,8 +505,7 @@ const aihubmixModels: AIChatModelCard[] = [
502
505
  vision: true,
503
506
  },
504
507
  contextWindowTokens: 1_047_576,
505
- description:
506
- 'GPT-4.1 mini balances intelligence, speed, and cost for many use cases.',
508
+ description: 'GPT-4.1 mini balances intelligence, speed, and cost for many use cases.',
507
509
  displayName: 'GPT-4.1 mini',
508
510
  id: 'gpt-4.1-mini',
509
511
  maxOutput: 32_768,
@@ -1137,6 +1139,41 @@ const aihubmixModels: AIChatModelCard[] = [
1137
1139
  ],
1138
1140
  },
1139
1141
  releasedAt: '2025-11-18',
1142
+ settings: {
1143
+ extendParams: ['thinkingLevel2', 'urlContext'],
1144
+ searchImpl: 'params',
1145
+ searchProvider: 'google',
1146
+ },
1147
+ type: 'chat',
1148
+ },
1149
+ {
1150
+ abilities: {
1151
+ functionCall: true,
1152
+ reasoning: true,
1153
+ search: true,
1154
+ video: true,
1155
+ vision: true,
1156
+ },
1157
+ contextWindowTokens: 1_048_576 + 65_536,
1158
+ description: 'Gemini 3 Flash 是为速度而打造的最智能的模型,将前沿智能与卓越的搜索接地相结合。',
1159
+ displayName: 'Gemini 3 Flash Preview',
1160
+ enabled: true,
1161
+ id: 'gemini-3-flash-preview',
1162
+ maxOutput: 65_536,
1163
+ pricing: {
1164
+ units: [
1165
+ { name: 'textInput_cacheRead', rate: 0.05, strategy: 'fixed', unit: 'millionTokens' },
1166
+ { name: 'textInput', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
1167
+ { name: 'textOutput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
1168
+ {
1169
+ lookup: { prices: { '1h': 1 }, pricingParams: ['ttl'] },
1170
+ name: 'textInput_cacheWrite',
1171
+ strategy: 'lookup',
1172
+ unit: 'millionTokens',
1173
+ },
1174
+ ],
1175
+ },
1176
+ releasedAt: '2025-12-17',
1140
1177
  settings: {
1141
1178
  extendParams: ['thinkingLevel', 'urlContext'],
1142
1179
  searchImpl: 'params',
@@ -1210,7 +1247,7 @@ const aihubmixModels: AIChatModelCard[] = [
1210
1247
  vision: true,
1211
1248
  },
1212
1249
  contextWindowTokens: 1_048_576 + 65_536,
1213
- description: "Gemini 2.5 Flash is Google’s best-value model with full capabilities.",
1250
+ description: 'Gemini 2.5 Flash is Google’s best-value model with full capabilities.',
1214
1251
  displayName: 'Gemini 2.5 Flash',
1215
1252
  id: 'gemini-2.5-flash',
1216
1253
  maxOutput: 65_536,
@@ -1238,7 +1275,6 @@ const aihubmixModels: AIChatModelCard[] = [
1238
1275
  description:
1239
1276
  'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, allowing image generation and editing through conversation.',
1240
1277
  displayName: 'Nano Banana',
1241
- enabled: true,
1242
1278
  id: 'gemini-2.5-flash-image',
1243
1279
  maxOutput: 8192,
1244
1280
  pricing: {
@@ -1262,7 +1298,7 @@ const aihubmixModels: AIChatModelCard[] = [
1262
1298
  },
1263
1299
  contextWindowTokens: 1_048_576 + 65_536,
1264
1300
  description:
1265
- "Gemini 2.5 Flash-Lite is Google’s smallest, best-value model, designed for large-scale use.",
1301
+ 'Gemini 2.5 Flash-Lite is Google’s smallest, best-value model, designed for large-scale use.',
1266
1302
  displayName: 'Gemini 2.5 Flash-Lite',
1267
1303
  id: 'gemini-2.5-flash-lite',
1268
1304
  maxOutput: 65_536,
@@ -123,7 +123,7 @@ const googleChatModels: AIChatModelCard[] = [
123
123
  },
124
124
  contextWindowTokens: 1_048_576 + 65_536,
125
125
  description:
126
- "Gemini 3 Pro is Google’s most powerful agent and vibe-coding model, delivering richer visuals and deeper interaction on top of state-of-the-art reasoning.",
126
+ 'Gemini 3 Pro is Google’s most powerful agent and vibe-coding model, delivering richer visuals and deeper interaction on top of state-of-the-art reasoning.',
127
127
  displayName: 'Gemini 3 Pro Preview',
128
128
  enabled: true,
129
129
  id: 'gemini-3-pro-preview',
@@ -166,6 +166,41 @@ const googleChatModels: AIChatModelCard[] = [
166
166
  ],
167
167
  },
168
168
  releasedAt: '2025-11-18',
169
+ settings: {
170
+ extendParams: ['thinkingLevel2', 'urlContext'],
171
+ searchImpl: 'params',
172
+ searchProvider: 'google',
173
+ },
174
+ type: 'chat',
175
+ },
176
+ {
177
+ abilities: {
178
+ functionCall: true,
179
+ reasoning: true,
180
+ search: true,
181
+ video: true,
182
+ vision: true,
183
+ },
184
+ contextWindowTokens: 1_048_576 + 65_536,
185
+ description: 'Gemini 3 Flash 是为速度而打造的最智能的模型,将前沿智能与卓越的搜索接地相结合。',
186
+ displayName: 'Gemini 3 Flash Preview',
187
+ enabled: true,
188
+ id: 'gemini-3-flash-preview',
189
+ maxOutput: 65_536,
190
+ pricing: {
191
+ units: [
192
+ { name: 'textInput_cacheRead', rate: 0.05, strategy: 'fixed', unit: 'millionTokens' },
193
+ { name: 'textInput', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
194
+ { name: 'textOutput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
195
+ {
196
+ lookup: { prices: { '1h': 1 }, pricingParams: ['ttl'] },
197
+ name: 'textInput_cacheWrite',
198
+ strategy: 'lookup',
199
+ unit: 'millionTokens',
200
+ },
201
+ ],
202
+ },
203
+ releasedAt: '2025-12-17',
169
204
  settings: {
170
205
  extendParams: ['thinkingLevel', 'urlContext'],
171
206
  searchImpl: 'params',
@@ -182,7 +217,7 @@ const googleChatModels: AIChatModelCard[] = [
182
217
  },
183
218
  contextWindowTokens: 131_072 + 32_768,
184
219
  description:
185
- "Gemini 3 Pro Image (Nano Banana Pro) is Google’s image generation model and also supports multimodal chat.",
220
+ 'Gemini 3 Pro Image (Nano Banana Pro) is Google’s image generation model and also supports multimodal chat.',
186
221
  displayName: 'Nano Banana Pro',
187
222
  enabled: true,
188
223
  id: 'gemini-3-pro-image-preview',
@@ -213,9 +248,8 @@ const googleChatModels: AIChatModelCard[] = [
213
248
  },
214
249
  contextWindowTokens: 1_048_576 + 65_536,
215
250
  description:
216
- "Gemini 2.5 Pro is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.",
251
+ 'Gemini 2.5 Pro is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
217
252
  displayName: 'Gemini 2.5 Pro',
218
- enabled: true,
219
253
  id: 'gemini-2.5-pro',
220
254
  maxOutput: 65_536,
221
255
  pricing: {
@@ -273,7 +307,7 @@ const googleChatModels: AIChatModelCard[] = [
273
307
  },
274
308
  contextWindowTokens: 1_048_576 + 65_536,
275
309
  description:
276
- "Gemini 2.5 Pro Preview is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.",
310
+ 'Gemini 2.5 Pro Preview is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
277
311
  displayName: 'Gemini 2.5 Pro Preview 06-05',
278
312
  id: 'gemini-2.5-pro-preview-06-05',
279
313
  maxOutput: 65_536,
@@ -326,7 +360,7 @@ const googleChatModels: AIChatModelCard[] = [
326
360
  },
327
361
  contextWindowTokens: 1_048_576 + 65_536,
328
362
  description:
329
- "Gemini 2.5 Pro Preview is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.",
363
+ 'Gemini 2.5 Pro Preview is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
330
364
  displayName: 'Gemini 2.5 Pro Preview 05-06',
331
365
  id: 'gemini-2.5-pro-preview-05-06',
332
366
  maxOutput: 65_536,
@@ -377,9 +411,8 @@ const googleChatModels: AIChatModelCard[] = [
377
411
  vision: true,
378
412
  },
379
413
  contextWindowTokens: 1_048_576 + 65_536,
380
- description: "Gemini 2.5 Flash is Google’s best-value model with full capabilities.",
414
+ description: 'Gemini 2.5 Flash is Google’s best-value model with full capabilities.',
381
415
  displayName: 'Gemini 2.5 Flash',
382
- enabled: true,
383
416
  id: 'gemini-2.5-flash',
384
417
  maxOutput: 65_536,
385
418
  pricing: {
@@ -432,9 +465,8 @@ const googleChatModels: AIChatModelCard[] = [
432
465
  },
433
466
  contextWindowTokens: 32_768 + 8192,
434
467
  description:
435
- "Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.",
468
+ 'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
436
469
  displayName: 'Nano Banana',
437
- enabled: true,
438
470
  id: 'gemini-2.5-flash-image',
439
471
  maxOutput: 8192,
440
472
  pricing: {
@@ -459,7 +491,7 @@ const googleChatModels: AIChatModelCard[] = [
459
491
  },
460
492
  contextWindowTokens: 32_768 + 8192,
461
493
  description:
462
- "Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.",
494
+ 'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
463
495
  displayName: 'Nano Banana (Preview)',
464
496
  id: 'gemini-2.5-flash-image-preview',
465
497
  maxOutput: 8192,
@@ -488,7 +520,7 @@ const googleChatModels: AIChatModelCard[] = [
488
520
  },
489
521
  contextWindowTokens: 1_048_576 + 65_536,
490
522
  description:
491
- "Gemini 2.5 Flash-Lite is Google’s smallest, best-value model, designed for large-scale use.",
523
+ 'Gemini 2.5 Flash-Lite is Google’s smallest, best-value model, designed for large-scale use.',
492
524
  displayName: 'Gemini 2.5 Flash-Lite',
493
525
  id: 'gemini-2.5-flash-lite',
494
526
  maxOutput: 65_536,
@@ -706,7 +738,8 @@ const googleChatModels: AIChatModelCard[] = [
706
738
  vision: true,
707
739
  },
708
740
  contextWindowTokens: 1_008_192,
709
- description: 'Gemini 1.5 Flash 002 is an efficient multimodal model built for broad deployment.',
741
+ description:
742
+ 'Gemini 1.5 Flash 002 is an efficient multimodal model built for broad deployment.',
710
743
  displayName: 'Gemini 1.5 Flash 002',
711
744
  id: 'gemini-1.5-flash-002', // Deprecated on 2025-09-24
712
745
  maxOutput: 8192,
@@ -903,7 +936,7 @@ const googleImageModels: AIImageModelCard[] = [
903
936
  type: 'image',
904
937
  enabled: true,
905
938
  description:
906
- "Gemini 3 Pro Image (Nano Banana Pro) is Google’s image generation model and also supports multimodal chat.",
939
+ 'Gemini 3 Pro Image (Nano Banana Pro) is Google’s image generation model and also supports multimodal chat.',
907
940
  releasedAt: '2025-11-18',
908
941
  parameters: nanoBananaProParameters,
909
942
  pricing: {
@@ -918,10 +951,9 @@ const googleImageModels: AIImageModelCard[] = [
918
951
  {
919
952
  displayName: 'Nano Banana',
920
953
  id: 'gemini-2.5-flash-image:image',
921
- enabled: true,
922
954
  type: 'image',
923
955
  description:
924
- "Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.",
956
+ 'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
925
957
  releasedAt: '2025-08-26',
926
958
  parameters: nanoBananaParameters,
927
959
  pricing: {
@@ -938,7 +970,7 @@ const googleImageModels: AIImageModelCard[] = [
938
970
  id: 'gemini-2.5-flash-image-preview:image',
939
971
  type: 'image',
940
972
  description:
941
- "Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.",
973
+ 'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
942
974
  releasedAt: '2025-08-26',
943
975
  parameters: CHAT_MODEL_IMAGE_GENERATION_PARAMS,
944
976
  pricing: {
@@ -446,6 +446,26 @@ const hunyuanChatModels: AIChatModelCard[] = [
446
446
  },
447
447
  type: 'chat',
448
448
  },
449
+ {
450
+ abilities: {
451
+ vision: true,
452
+ },
453
+ contextWindowTokens: 24_000,
454
+ description:
455
+ '基于文本 TurboS 基座生产的图生文快思考模型,相比上一版本在图像基础识别、图像分析推理等维度都有明显的效果提升。',
456
+ displayName: '混元图生文',
457
+ id: 'hunyuan-vision-1.5-instruct',
458
+ maxOutput: 16_000,
459
+ pricing: {
460
+ currency: 'CNY',
461
+ units: [
462
+ { name: 'textInput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
463
+ { name: 'textOutput', rate: 9, strategy: 'fixed', unit: 'millionTokens' },
464
+ ],
465
+ },
466
+ releasedAt: '2025-12-17',
467
+ type: 'chat',
468
+ },
449
469
  {
450
470
  abilities: {
451
471
  vision: true,
@@ -51,9 +51,29 @@ const infiniaiChatModels: AIChatModelCard[] = [
51
51
  },
52
52
  contextWindowTokens: 200_000,
53
53
  description:
54
- 'MiniMax-M2 is a MoE language model optimized for coding and agent workflows, with about 230B total parameters and around 10B active parameters. It strengthens developer scenarios such as multi-file editing, code-run-fix loops, and test verification/repair, while delivering stable, efficient performance in real environments like terminals, IDEs, and CI.',
55
- displayName: 'MiniMax M2',
54
+ 'MiniMax-M2.1 MiniMax 系列的最新版本,专为多语言编程和真实世界复杂任务优化。作为一款 AI 原生模型,MiniMax-M2.1 在模型性能、智能体框架支持以及多场景适配方面实现了显著提升,旨在帮助企业和个人更快地找到 AI 原生的工作与生活方式。',
55
+ displayName: 'MiniMax M2.1',
56
56
  enabled: true,
57
+ id: 'minimax-m2.1',
58
+ maxOutput: 200_000,
59
+ pricing: {
60
+ currency: 'CNY',
61
+ units: [
62
+ { name: 'textInput', rate: 2.1, strategy: 'fixed', unit: 'millionTokens' },
63
+ { name: 'textOutput', rate: 8.4, strategy: 'fixed', unit: 'millionTokens' },
64
+ ],
65
+ },
66
+ type: 'chat',
67
+ },
68
+ {
69
+ abilities: {
70
+ functionCall: true,
71
+ reasoning: true,
72
+ },
73
+ contextWindowTokens: 200_000,
74
+ description:
75
+ 'MiniMax-M2 是一款专为编码与智能体工作流优化的专家混合(MoE)语言模型,具有约 230B 总参数与约 10B 活跃参数。它在保持强通用智能的同时,针对多文件编辑、代码-运行-修复闭环、测试校验修复等开发者场景进行深度增强,在终端、IDE 与 CI 等真实环境中表现稳定、高效。',
76
+ displayName: 'MiniMax M2',
57
77
  id: 'minimax-m2',
58
78
  maxOutput: 200_000,
59
79
  pricing: {
@@ -71,16 +91,38 @@ const infiniaiChatModels: AIChatModelCard[] = [
71
91
  reasoning: true,
72
92
  },
73
93
  contextWindowTokens: 131_072,
74
- description: 'GLM-4.6 is Zhipu AI’s latest LLM, with stronger reasoning and generation.',
75
- displayName: 'GLM-4.6',
94
+ description: 'GLM-4.7 是智谱AI推出的最新大语言模型,具备更强的推理和生成能力。',
95
+ displayName: 'GLM-4.7',
76
96
  enabled: true,
97
+ id: 'glm-4.7',
98
+ maxOutput: 4096,
99
+ pricing: {
100
+ currency: 'CNY',
101
+ units: [
102
+ { name: 'textInput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
103
+ { name: 'textOutput', rate: 14, strategy: 'fixed', unit: 'millionTokens' },
104
+ ],
105
+ },
106
+ settings: {
107
+ extendParams: ['enableReasoning'],
108
+ },
109
+ type: 'chat',
110
+ },
111
+ {
112
+ abilities: {
113
+ functionCall: true,
114
+ reasoning: true,
115
+ },
116
+ contextWindowTokens: 131_072,
117
+ description: 'GLM-4.6 是智谱AI推出的最新大语言模型,具备更强的推理和生成能力。',
118
+ displayName: 'GLM-4.6',
77
119
  id: 'glm-4.6',
78
120
  maxOutput: 4096,
79
121
  pricing: {
80
122
  currency: 'CNY',
81
123
  units: [
82
- { name: 'textInput', rate: 4, strategy: 'fixed', unit: 'millionTokens' },
83
- { name: 'textOutput', rate: 16, strategy: 'fixed', unit: 'millionTokens' },
124
+ { name: 'textInput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
125
+ { name: 'textOutput', rate: 14, strategy: 'fixed', unit: 'millionTokens' },
84
126
  ],
85
127
  },
86
128
  settings: {
@@ -96,7 +138,6 @@ const infiniaiChatModels: AIChatModelCard[] = [
96
138
  contextWindowTokens: 131_072,
97
139
  description: 'DeepSeek-V3.2-Exp is an experimental DeepSeek LLM with stronger reasoning and generation.',
98
140
  displayName: 'DeepSeek V3.2 Exp',
99
- enabled: true,
100
141
  id: 'deepseek-v3.2-exp',
101
142
  maxOutput: 65_536,
102
143
  pricing: {
@@ -12,7 +12,8 @@ const lobehubChatModels: AIChatModelCard[] = [
12
12
  vision: true,
13
13
  },
14
14
  contextWindowTokens: 400_000,
15
- description: "GPT-5.2 — OpenAI's flagship model optimized for coding and agent tasks across industries.",
15
+ description:
16
+ "GPT-5.2 — OpenAI's flagship model optimized for coding and agent tasks across industries.",
16
17
  displayName: 'GPT-5.2',
17
18
  enabled: true,
18
19
  id: 'gpt-5.2',
@@ -26,7 +27,7 @@ const lobehubChatModels: AIChatModelCard[] = [
26
27
  },
27
28
  releasedAt: '2025-12-11',
28
29
  settings: {
29
- extendParams: ['gpt5_1ReasoningEffort', 'textVerbosity'],
30
+ extendParams: ['gpt5_2ReasoningEffort', 'textVerbosity'],
30
31
  searchImpl: 'params',
31
32
  },
32
33
  type: 'chat',
@@ -164,7 +165,8 @@ const lobehubChatModels: AIChatModelCard[] = [
164
165
  vision: true,
165
166
  },
166
167
  contextWindowTokens: 1_047_576,
167
- description: 'GPT-4.1 is our flagship model for complex tasks and cross-domain problem solving.',
168
+ description:
169
+ 'GPT-4.1 is our flagship model for complex tasks and cross-domain problem solving.',
168
170
  displayName: 'GPT-4.1',
169
171
  enabled: true,
170
172
  id: 'gpt-4.1',
@@ -628,7 +630,7 @@ const lobehubChatModels: AIChatModelCard[] = [
628
630
  },
629
631
  contextWindowTokens: 1_048_576 + 65_536,
630
632
  description:
631
- "Gemini 3 Pro is Google’s most powerful agent and vibe-coding model, delivering richer visuals and deeper interaction on top of state-of-the-art reasoning.",
633
+ 'Gemini 3 Pro is Google’s most powerful agent and vibe-coding model, delivering richer visuals and deeper interaction on top of state-of-the-art reasoning.',
632
634
  displayName: 'Gemini 3 Pro Preview',
633
635
  enabled: true,
634
636
  id: 'gemini-3-pro-preview',
@@ -672,7 +674,7 @@ const lobehubChatModels: AIChatModelCard[] = [
672
674
  },
673
675
  releasedAt: '2025-11-18',
674
676
  settings: {
675
- extendParams: ['thinkingLevel', 'urlContext'],
677
+ extendParams: ['thinkingLevel2', 'urlContext'],
676
678
  searchImpl: 'params',
677
679
  searchProvider: 'google',
678
680
  },
@@ -728,7 +730,7 @@ const lobehubChatModels: AIChatModelCard[] = [
728
730
  },
729
731
  contextWindowTokens: 1_048_576 + 65_536,
730
732
  description:
731
- "Gemini 2.5 Pro is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.",
733
+ 'Gemini 2.5 Pro is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
732
734
  displayName: 'Gemini 2.5 Pro',
733
735
  enabled: true,
734
736
  id: 'gemini-2.5-pro',
@@ -780,7 +782,7 @@ const lobehubChatModels: AIChatModelCard[] = [
780
782
  vision: true,
781
783
  },
782
784
  contextWindowTokens: 1_048_576 + 65_536,
783
- description: "Gemini 2.5 Flash is Google’s best-value model with full capabilities.",
785
+ description: 'Gemini 2.5 Flash is Google’s best-value model with full capabilities.',
784
786
  displayName: 'Gemini 2.5 Flash',
785
787
  enabled: true,
786
788
  id: 'gemini-2.5-flash',
@@ -819,7 +821,7 @@ const lobehubChatModels: AIChatModelCard[] = [
819
821
  },
820
822
  contextWindowTokens: 131_072 + 32_768,
821
823
  description:
822
- "Gemini 3 Pro Image (Nano Banana Pro) is Google’s image generation model and also supports multimodal chat.",
824
+ 'Gemini 3 Pro Image (Nano Banana Pro) is Google’s image generation model and also supports multimodal chat.',
823
825
  displayName: 'Nano Banana Pro',
824
826
  enabled: true,
825
827
  id: 'gemini-3-pro-image-preview',
@@ -845,7 +847,7 @@ const lobehubChatModels: AIChatModelCard[] = [
845
847
  },
846
848
  contextWindowTokens: 32_768 + 32_768,
847
849
  description:
848
- "Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.",
850
+ 'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
849
851
  displayName: 'Nano Banana',
850
852
  enabled: true,
851
853
  id: 'gemini-2.5-flash-image-preview',
@@ -1039,7 +1041,7 @@ export const nanoBananaProParameters: ModelParamsSchema = {
1039
1041
  const lobehubImageModels: AIImageModelCard[] = [
1040
1042
  {
1041
1043
  description:
1042
- "Gemini 3 Pro Image (Nano Banana Pro) is Google’s image generation model and also supports multimodal chat.",
1044
+ 'Gemini 3 Pro Image (Nano Banana Pro) is Google’s image generation model and also supports multimodal chat.',
1043
1045
  displayName: 'Nano Banana Pro',
1044
1046
  enabled: true,
1045
1047
  id: 'gemini-3-pro-image-preview:image',
@@ -1057,7 +1059,7 @@ const lobehubImageModels: AIImageModelCard[] = [
1057
1059
  },
1058
1060
  {
1059
1061
  description:
1060
- "Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.",
1062
+ 'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
1061
1063
  displayName: 'Nano Banana',
1062
1064
  enabled: true,
1063
1065
  id: 'gemini-2.5-flash-image-preview:image',
@@ -7,9 +7,53 @@ const minimaxChatModels: AIChatModelCard[] = [
7
7
  reasoning: true,
8
8
  },
9
9
  contextWindowTokens: 204_800,
10
- description: 'Built for efficient coding and agent workflows.',
11
- displayName: 'MiniMax M2',
10
+ description: '强大多语言编程实力,全面升级编程体验',
11
+ displayName: 'MiniMax M2.1',
12
12
  enabled: true,
13
+ id: 'MiniMax-M2.1',
14
+ maxOutput: 131_072,
15
+ pricing: {
16
+ currency: 'CNY',
17
+ units: [
18
+ { name: 'textInput_cacheRead', rate: 0.21, strategy: 'fixed', unit: 'millionTokens' },
19
+ { name: 'textInput_cacheWrite', rate: 2.625, strategy: 'fixed', unit: 'millionTokens' },
20
+ { name: 'textInput', rate: 2.1, strategy: 'fixed', unit: 'millionTokens' },
21
+ { name: 'textOutput', rate: 8.4, strategy: 'fixed', unit: 'millionTokens' },
22
+ ],
23
+ },
24
+ releasedAt: '2025-12-23',
25
+ type: 'chat',
26
+ },
27
+ {
28
+ abilities: {
29
+ functionCall: true,
30
+ reasoning: true,
31
+ },
32
+ contextWindowTokens: 204_800,
33
+ description: '强大多语言编程实力,全面升级编程体验。更快,更高效。',
34
+ displayName: 'MiniMax M2.1 Lightning',
35
+ id: 'MiniMax-M2.1-Lightning',
36
+ maxOutput: 131_072,
37
+ pricing: {
38
+ currency: 'CNY',
39
+ units: [
40
+ { name: 'textInput_cacheRead', rate: 0.21, strategy: 'fixed', unit: 'millionTokens' },
41
+ { name: 'textInput_cacheWrite', rate: 2.625, strategy: 'fixed', unit: 'millionTokens' },
42
+ { name: 'textInput', rate: 2.1, strategy: 'fixed', unit: 'millionTokens' },
43
+ { name: 'textOutput', rate: 16.8, strategy: 'fixed', unit: 'millionTokens' },
44
+ ],
45
+ },
46
+ releasedAt: '2025-12-23',
47
+ type: 'chat',
48
+ },
49
+ {
50
+ abilities: {
51
+ functionCall: true,
52
+ reasoning: true,
53
+ },
54
+ contextWindowTokens: 204_800,
55
+ description: '专为高效编码与Agent工作流而生',
56
+ displayName: 'MiniMax M2',
13
57
  id: 'MiniMax-M2',
14
58
  maxOutput: 131_072,
15
59
  pricing: {
@@ -42,11 +42,34 @@ const ollamaCloudModels: AIChatModelCard[] = [
42
42
  abilities: {
43
43
  functionCall: true,
44
44
  reasoning: true,
45
+ vision: true,
45
46
  },
46
- contextWindowTokens: 200_000,
47
- description: 'MiniMax M2 is an efficient LLM built for coding and agent workflows.',
48
- displayName: 'MiniMax M2',
47
+ contextWindowTokens: 1_048_576,
48
+ description: 'Gemini 3 Flash 是为速度而打造的最智能的模型,将前沿智能与卓越的搜索接地相结合。',
49
+ displayName: 'Gemini 3 Flash Preview',
50
+ id: 'gemini-3-flash-preview',
51
+ releasedAt: '2025-12-17',
52
+ type: 'chat',
53
+ },
54
+ {
55
+ abilities: {
56
+ functionCall: true,
57
+ reasoning: true,
58
+ },
59
+ contextWindowTokens: 204_800,
60
+ displayName: 'MiniMax M2.1',
49
61
  enabled: true,
62
+ id: 'minimax-m2.1',
63
+ type: 'chat',
64
+ },
65
+ {
66
+ abilities: {
67
+ functionCall: true,
68
+ reasoning: true,
69
+ },
70
+ contextWindowTokens: 204_800,
71
+ description: 'MiniMax M2 是专为编码和代理工作流程构建的高效大型语言模型。',
72
+ displayName: 'MiniMax M2',
50
73
  id: 'minimax-m2',
51
74
  type: 'chat',
52
75
  },
@@ -57,9 +80,21 @@ const ollamaCloudModels: AIChatModelCard[] = [
57
80
  },
58
81
  contextWindowTokens: 200_000,
59
82
  description:
60
- 'Zhipu’s latest flagship GLM-4.6 (355B) surpasses prior versions in advanced coding, long-text handling, reasoning, and agent capabilities, aligning with Claude Sonnet 4 in programming performance and ranking among the top coding models in China.',
61
- displayName: 'GLM-4.6',
83
+ 'GLM-4.7 是智谱最新旗舰模型,面向 Agentic Coding 场景强化了编码能力、长程任务规划与工具协同,并在多个公开基准的当期榜单中取得开源模型中的领先表现。通用能力提升,回复更简洁自然,写作更具沉浸感。在执行复杂智能体任务,在工具调用时指令遵循更强,Artifacts Agentic Coding 的前端美感和长程任务完成效率进一步提升。',
84
+ displayName: 'GLM-4.7',
62
85
  enabled: true,
86
+ id: 'glm-4.7',
87
+ type: 'chat',
88
+ },
89
+ {
90
+ abilities: {
91
+ functionCall: true,
92
+ reasoning: true,
93
+ },
94
+ contextWindowTokens: 200_000,
95
+ description:
96
+ '智谱旗舰模型 GLM-4.6 (355B) 在高级编码、长文本处理、推理与智能体能力上全面超越前代,尤其在编程能力上对齐 Claude Sonnet 4,成为国内顶尖的 Coding 模型。',
97
+ displayName: 'GLM-4.6',
63
98
  id: 'glm-4.6',
64
99
  type: 'chat',
65
100
  },
@@ -42,7 +42,7 @@ export const openaiChatModels: AIChatModelCard[] = [
42
42
  },
43
43
  releasedAt: '2025-12-11',
44
44
  settings: {
45
- extendParams: ['gpt5_1ReasoningEffort', 'textVerbosity'],
45
+ extendParams: ['gpt5_2ReasoningEffort', 'textVerbosity'],
46
46
  searchImpl: 'params',
47
47
  },
48
48
  type: 'chat',
@@ -68,6 +68,7 @@ export const openaiChatModels: AIChatModelCard[] = [
68
68
  },
69
69
  releasedAt: '2025-12-11',
70
70
  settings: {
71
+ extendParams: ['gpt5_2ProReasoningEffort'],
71
72
  searchImpl: 'params',
72
73
  },
73
74
  type: 'chat',
@@ -203,7 +204,8 @@ export const openaiChatModels: AIChatModelCard[] = [
203
204
  vision: true,
204
205
  },
205
206
  contextWindowTokens: 400_000,
206
- description: 'GPT-5 pro uses more compute to think deeper and consistently deliver better answers.',
207
+ description:
208
+ 'GPT-5 pro uses more compute to think deeper and consistently deliver better answers.',
207
209
  displayName: 'GPT-5 pro',
208
210
  id: 'gpt-5-pro',
209
211
  maxOutput: 272_000,
@@ -563,7 +565,8 @@ export const openaiChatModels: AIChatModelCard[] = [
563
565
  vision: true,
564
566
  },
565
567
  contextWindowTokens: 1_047_576,
566
- description: 'GPT-4.1 is our flagship model for complex tasks and cross-domain problem solving.',
568
+ description:
569
+ 'GPT-4.1 is our flagship model for complex tasks and cross-domain problem solving.',
567
570
  displayName: 'GPT-4.1',
568
571
  id: 'gpt-4.1',
569
572
  maxOutput: 32_768,
@@ -9,7 +9,7 @@ const qwenChatModels: AIChatModelCard[] = [
9
9
  vision: true,
10
10
  },
11
11
  config: {
12
- deploymentName: 'qwen3-vl-plus',
12
+ deploymentName: 'qwen3-vl-plus-2025-12-19',
13
13
  },
14
14
  contextWindowTokens: 262_144,
15
15
  description: