@lobehub/lobehub 2.0.0-next.186 → 2.0.0-next.188

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (140) hide show
  1. package/CHANGELOG.md +50 -0
  2. package/changelog/v1.json +18 -0
  3. package/locales/ar/models.json +89 -5
  4. package/locales/ar/plugin.json +5 -0
  5. package/locales/ar/providers.json +1 -0
  6. package/locales/bg-BG/models.json +68 -0
  7. package/locales/bg-BG/plugin.json +5 -0
  8. package/locales/bg-BG/providers.json +1 -0
  9. package/locales/de-DE/models.json +85 -0
  10. package/locales/de-DE/plugin.json +5 -0
  11. package/locales/de-DE/providers.json +1 -0
  12. package/locales/en-US/models.json +11 -10
  13. package/locales/en-US/plugin.json +5 -0
  14. package/locales/en-US/providers.json +1 -0
  15. package/locales/es-ES/models.json +72 -0
  16. package/locales/es-ES/plugin.json +5 -0
  17. package/locales/es-ES/providers.json +1 -0
  18. package/locales/fa-IR/models.json +86 -0
  19. package/locales/fa-IR/plugin.json +5 -0
  20. package/locales/fa-IR/providers.json +1 -0
  21. package/locales/fr-FR/models.json +49 -0
  22. package/locales/fr-FR/plugin.json +5 -0
  23. package/locales/fr-FR/providers.json +1 -0
  24. package/locales/it-IT/models.json +82 -0
  25. package/locales/it-IT/plugin.json +5 -0
  26. package/locales/it-IT/providers.json +1 -0
  27. package/locales/ja-JP/models.json +42 -5
  28. package/locales/ja-JP/plugin.json +5 -0
  29. package/locales/ja-JP/providers.json +1 -0
  30. package/locales/ko-KR/models.json +54 -0
  31. package/locales/ko-KR/plugin.json +5 -0
  32. package/locales/ko-KR/providers.json +1 -0
  33. package/locales/nl-NL/models.json +12 -1
  34. package/locales/nl-NL/plugin.json +5 -0
  35. package/locales/nl-NL/providers.json +1 -0
  36. package/locales/pl-PL/models.json +46 -0
  37. package/locales/pl-PL/plugin.json +5 -0
  38. package/locales/pl-PL/providers.json +1 -0
  39. package/locales/pt-BR/models.json +59 -0
  40. package/locales/pt-BR/plugin.json +5 -0
  41. package/locales/pt-BR/providers.json +1 -0
  42. package/locales/ru-RU/models.json +85 -0
  43. package/locales/ru-RU/plugin.json +5 -0
  44. package/locales/ru-RU/providers.json +1 -0
  45. package/locales/tr-TR/models.json +81 -0
  46. package/locales/tr-TR/plugin.json +5 -0
  47. package/locales/tr-TR/providers.json +1 -0
  48. package/locales/vi-VN/models.json +54 -0
  49. package/locales/vi-VN/plugin.json +5 -0
  50. package/locales/vi-VN/providers.json +1 -0
  51. package/locales/zh-CN/models.json +42 -5
  52. package/locales/zh-CN/plugin.json +5 -0
  53. package/locales/zh-CN/providers.json +1 -0
  54. package/locales/zh-TW/models.json +85 -0
  55. package/locales/zh-TW/plugin.json +5 -0
  56. package/locales/zh-TW/providers.json +1 -0
  57. package/package.json +1 -1
  58. package/packages/builtin-tool-gtd/src/manifest.ts +13 -8
  59. package/packages/builtin-tool-gtd/src/systemRole.ts +54 -19
  60. package/packages/builtin-tool-knowledge-base/package.json +1 -0
  61. package/packages/builtin-tool-knowledge-base/src/client/Inspector/ReadKnowledge/index.tsx +97 -0
  62. package/packages/builtin-tool-knowledge-base/src/client/Inspector/SearchKnowledgeBase/index.tsx +75 -0
  63. package/packages/builtin-tool-knowledge-base/src/client/Inspector/index.ts +11 -0
  64. package/packages/builtin-tool-knowledge-base/src/client/Render/ReadKnowledge/FileCard.tsx +12 -12
  65. package/packages/builtin-tool-knowledge-base/src/client/Render/ReadKnowledge/index.tsx +16 -25
  66. package/packages/builtin-tool-knowledge-base/src/client/Render/SearchKnowledgeBase/Item/index.tsx +21 -47
  67. package/packages/builtin-tool-knowledge-base/src/client/Render/SearchKnowledgeBase/index.tsx +19 -31
  68. package/packages/builtin-tool-knowledge-base/src/client/Render/index.ts +0 -5
  69. package/packages/builtin-tool-knowledge-base/src/client/index.ts +5 -1
  70. package/packages/builtin-tool-knowledge-base/src/executor/index.ts +119 -0
  71. package/packages/builtin-tool-local-system/package.json +1 -0
  72. package/packages/builtin-tool-local-system/src/client/Inspector/EditLocalFile/index.tsx +44 -29
  73. package/packages/builtin-tool-local-system/src/client/Inspector/GrepContent/index.tsx +20 -18
  74. package/packages/builtin-tool-local-system/src/client/Inspector/ListLocalFiles/index.tsx +76 -0
  75. package/packages/builtin-tool-local-system/src/client/Inspector/ReadLocalFile/index.tsx +8 -32
  76. package/packages/builtin-tool-local-system/src/client/Inspector/RenameLocalFile/index.tsx +62 -0
  77. package/packages/builtin-tool-local-system/src/client/Inspector/SearchLocalFiles/index.tsx +17 -11
  78. package/packages/builtin-tool-local-system/src/client/Inspector/WriteLocalFile/index.tsx +61 -0
  79. package/packages/builtin-tool-local-system/src/client/Inspector/index.ts +6 -0
  80. package/packages/builtin-tool-local-system/src/client/Render/EditLocalFile/index.tsx +6 -1
  81. package/packages/builtin-tool-local-system/src/client/Render/SearchFiles/SearchQuery/SearchView.tsx +19 -31
  82. package/packages/builtin-tool-local-system/src/client/Render/SearchFiles/SearchQuery/index.tsx +2 -42
  83. package/packages/builtin-tool-local-system/src/client/Render/index.ts +0 -2
  84. package/packages/builtin-tool-local-system/src/client/components/FilePathDisplay.tsx +56 -0
  85. package/packages/builtin-tool-local-system/src/client/components/index.ts +2 -0
  86. package/packages/builtin-tool-local-system/src/executor/index.ts +435 -0
  87. package/packages/builtin-tool-web-browsing/src/client/Inspector/Search/index.tsx +32 -5
  88. package/packages/fetch-sse/src/__tests__/request.test.ts +608 -0
  89. package/packages/model-bank/src/aiModels/aihubmix.ts +44 -8
  90. package/packages/model-bank/src/aiModels/google.ts +49 -17
  91. package/packages/model-bank/src/aiModels/hunyuan.ts +20 -0
  92. package/packages/model-bank/src/aiModels/infiniai.ts +48 -7
  93. package/packages/model-bank/src/aiModels/lobehub.ts +13 -11
  94. package/packages/model-bank/src/aiModels/minimax.ts +46 -2
  95. package/packages/model-bank/src/aiModels/ollamacloud.ts +40 -5
  96. package/packages/model-bank/src/aiModels/openai.ts +6 -3
  97. package/packages/model-bank/src/aiModels/qwen.ts +1 -1
  98. package/packages/model-bank/src/aiModels/siliconcloud.ts +60 -0
  99. package/packages/model-bank/src/aiModels/vertexai.ts +77 -44
  100. package/packages/model-bank/src/aiModels/volcengine.ts +111 -2
  101. package/packages/model-bank/src/aiModels/zenmux.ts +19 -13
  102. package/packages/model-bank/src/aiModels/zhipu.ts +64 -2
  103. package/packages/model-bank/src/types/aiModel.ts +3 -0
  104. package/packages/model-runtime/src/core/contextBuilders/google.test.ts +84 -0
  105. package/packages/model-runtime/src/core/contextBuilders/google.ts +37 -1
  106. package/packages/model-runtime/src/providers/volcengine/index.ts +2 -1
  107. package/packages/model-runtime/src/providers/zhipu/index.test.ts +0 -27
  108. package/packages/model-runtime/src/providers/zhipu/index.ts +1 -1
  109. package/packages/model-runtime/src/utils/modelParse.ts +26 -21
  110. package/packages/types/src/agent/chatConfig.ts +6 -2
  111. package/src/features/ChatInput/ActionBar/Model/ControlsForm.tsx +40 -1
  112. package/src/features/ChatInput/ActionBar/Model/GPT52ProReasoningEffortSlider.tsx +59 -0
  113. package/src/features/ChatInput/ActionBar/Model/GPT52ReasoningEffortSlider.tsx +61 -0
  114. package/src/features/ChatInput/ActionBar/Model/TextVerbositySlider.tsx +1 -1
  115. package/src/features/ChatInput/ActionBar/Model/ThinkingLevel2Slider.tsx +58 -0
  116. package/src/features/ChatInput/ActionBar/Model/ThinkingLevelSlider.tsx +10 -8
  117. package/src/helpers/toolEngineering/index.ts +1 -1
  118. package/src/locales/default/plugin.ts +6 -0
  119. package/src/server/modules/Mecha/AgentToolsEngine/__tests__/index.test.ts +1 -1
  120. package/src/server/modules/Mecha/AgentToolsEngine/index.ts +1 -1
  121. package/src/services/chat/mecha/modelParamsResolver.ts +11 -0
  122. package/src/store/chat/slices/builtinTool/actions/index.ts +1 -11
  123. package/src/store/tool/slices/builtin/executors/index.ts +4 -0
  124. package/src/styles/text.ts +1 -1
  125. package/src/tools/executionRuntimes.ts +3 -8
  126. package/src/tools/identifiers.ts +1 -1
  127. package/src/tools/index.ts +1 -1
  128. package/src/tools/inspectors.ts +5 -0
  129. package/src/tools/renders.ts +6 -12
  130. package/packages/builtin-tool-local-system/src/client/Render/RenameLocalFile/index.tsx +0 -37
  131. package/src/store/chat/slices/builtinTool/actions/__tests__/localSystem.test.ts +0 -201
  132. package/src/store/chat/slices/builtinTool/actions/knowledgeBase.ts +0 -163
  133. package/src/store/chat/slices/builtinTool/actions/localSystem.ts +0 -241
  134. package/src/tools/knowledge-base/ExecutionRuntime/index.ts +0 -25
  135. package/src/tools/knowledge-base/Render/ReadKnowledge/index.tsx +0 -29
  136. package/src/tools/knowledge-base/Render/SearchKnowledgeBase/index.tsx +0 -29
  137. package/src/tools/knowledge-base/Render/index.ts +0 -7
  138. package/src/tools/knowledge-base/index.ts +0 -12
  139. package/src/tools/local-system/ExecutionRuntime/index.ts +0 -9
  140. package/src/tools/local-system/systemRole.ts +0 -1
@@ -2,6 +2,66 @@ import { AIChatModelCard, AIImageModelCard } from '../types/aiModel';
2
2
 
3
3
  // https://siliconflow.cn/zh-cn/models
4
4
  const siliconcloudChatModels: AIChatModelCard[] = [
5
+ {
6
+ abilities: {
7
+ functionCall: true,
8
+ reasoning: true,
9
+ },
10
+ contextWindowTokens: 200_000,
11
+ description:
12
+ 'GLM-4.7 是智谱新一代旗舰模型,总参数量 355B,激活参数量 32B,在通用对话、推理和智能体能力方面全面升级。GLM-4.7 增强了 Interleaved Thinking(交错思考),并引入了 Preserved Thinking(保留思考)和 Turn-level Thinking(轮级思考)。',
13
+ displayName: 'GLM-4.7 (Pro)',
14
+ id: 'Pro/zai-org/glm-4.7',
15
+ pricing: {
16
+ currency: 'CNY',
17
+ units: [
18
+ {
19
+ lookup: {
20
+ prices: {
21
+ '[0, 0.032]_[0, 0.0002]': 0.4,
22
+ '[0, 0.032]_[0.0002, infinity]': 0.6,
23
+ '[0.032, infinity]_[0, infinity]': 0.8,
24
+ },
25
+ pricingParams: ['textInput', 'textOutput'],
26
+ },
27
+ name: 'textInput_cacheRead',
28
+ strategy: 'lookup',
29
+ unit: 'millionTokens',
30
+ },
31
+ {
32
+ lookup: {
33
+ prices: {
34
+ '[0, 0.032]_[0, 0.0002]': 2,
35
+ '[0, 0.032]_[0.0002, infinity]': 3,
36
+ '[0.032, infinity]_[0, infinity]': 4,
37
+ },
38
+ pricingParams: ['textInput', 'textOutput'],
39
+ },
40
+ name: 'textInput',
41
+ strategy: 'lookup',
42
+ unit: 'millionTokens',
43
+ },
44
+ {
45
+ lookup: {
46
+ prices: {
47
+ '[0, 0.032]_[0, 0.0002]': 8,
48
+ '[0, 0.032]_[0.0002, infinity]': 14,
49
+ '[0.032, infinity]_[0, infinity]': 16,
50
+ },
51
+ pricingParams: ['textInput', 'textOutput'],
52
+ },
53
+ name: 'textOutput',
54
+ strategy: 'lookup',
55
+ unit: 'millionTokens',
56
+ },
57
+ ],
58
+ },
59
+ releasedAt: '2025-12-22',
60
+ settings: {
61
+ extendParams: ['enableReasoning', 'reasoningBudgetToken'],
62
+ },
63
+ type: 'chat',
64
+ },
5
65
  {
6
66
  abilities: {
7
67
  functionCall: true,
@@ -3,35 +3,6 @@ import { imagenGenParameters, nanoBananaParameters } from './google';
3
3
 
4
4
  // ref: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models
5
5
  const vertexaiChatModels: AIChatModelCard[] = [
6
- {
7
- abilities: {
8
- imageOutput: true,
9
- reasoning: true,
10
- search: true,
11
- vision: true,
12
- },
13
- contextWindowTokens: 131_072 + 32_768,
14
- description:
15
- "Gemini 3 Pro Image (Nano Banana Pro) is Google's image generation model and also supports multimodal chat.",
16
- displayName: 'Nano Banana Pro',
17
- enabled: true,
18
- id: 'gemini-3-pro-image-preview',
19
- maxOutput: 32_768,
20
- pricing: {
21
- approximatePricePerImage: 0.134,
22
- units: [
23
- { name: 'imageOutput', rate: 120, strategy: 'fixed', unit: 'millionTokens' },
24
- { name: 'textInput', rate: 2, strategy: 'fixed', unit: 'millionTokens' },
25
- { name: 'textOutput', rate: 12, strategy: 'fixed', unit: 'millionTokens' },
26
- ],
27
- },
28
- releasedAt: '2025-11-20',
29
- settings: {
30
- searchImpl: 'params',
31
- searchProvider: 'google',
32
- },
33
- type: 'chat',
34
- },
35
6
  {
36
7
  abilities: {
37
8
  functionCall: true,
@@ -42,7 +13,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
42
13
  },
43
14
  contextWindowTokens: 1_048_576 + 65_536,
44
15
  description:
45
- "Gemini 3 Pro is Google’s most powerful agent and vibe-coding model, delivering richer visuals and deeper interaction on top of state-of-the-art reasoning.",
16
+ 'Gemini 3 Pro is Google’s most powerful agent and vibe-coding model, delivering richer visuals and deeper interaction on top of state-of-the-art reasoning.',
46
17
  displayName: 'Gemini 3 Pro Preview',
47
18
  enabled: true,
48
19
  id: 'gemini-3-pro-preview',
@@ -85,6 +56,41 @@ const vertexaiChatModels: AIChatModelCard[] = [
85
56
  ],
86
57
  },
87
58
  releasedAt: '2025-11-18',
59
+ settings: {
60
+ extendParams: ['thinkingLevel2', 'urlContext'],
61
+ searchImpl: 'params',
62
+ searchProvider: 'google',
63
+ },
64
+ type: 'chat',
65
+ },
66
+ {
67
+ abilities: {
68
+ functionCall: true,
69
+ reasoning: true,
70
+ search: true,
71
+ video: true,
72
+ vision: true,
73
+ },
74
+ contextWindowTokens: 1_048_576 + 65_536,
75
+ description: 'Gemini 3 Flash 是为速度而打造的最智能的模型,将前沿智能与卓越的搜索接地相结合。',
76
+ displayName: 'Gemini 3 Flash Preview',
77
+ enabled: true,
78
+ id: 'gemini-3-flash-preview',
79
+ maxOutput: 65_536,
80
+ pricing: {
81
+ units: [
82
+ { name: 'textInput_cacheRead', rate: 0.05, strategy: 'fixed', unit: 'millionTokens' },
83
+ { name: 'textInput', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
84
+ { name: 'textOutput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
85
+ {
86
+ lookup: { prices: { '1h': 1 }, pricingParams: ['ttl'] },
87
+ name: 'textInput_cacheWrite',
88
+ strategy: 'lookup',
89
+ unit: 'millionTokens',
90
+ },
91
+ ],
92
+ },
93
+ releasedAt: '2025-12-17',
88
94
  settings: {
89
95
  extendParams: ['thinkingLevel', 'urlContext'],
90
96
  searchImpl: 'params',
@@ -92,6 +98,35 @@ const vertexaiChatModels: AIChatModelCard[] = [
92
98
  },
93
99
  type: 'chat',
94
100
  },
101
+ {
102
+ abilities: {
103
+ imageOutput: true,
104
+ reasoning: true,
105
+ search: true,
106
+ vision: true,
107
+ },
108
+ contextWindowTokens: 131_072 + 32_768,
109
+ description:
110
+ 'Gemini 3 Pro Image(Nano Banana Pro)是 Google 的图像生成模型,同时支持多模态对话。',
111
+ displayName: 'Nano Banana Pro',
112
+ enabled: true,
113
+ id: 'gemini-3-pro-image-preview',
114
+ maxOutput: 32_768,
115
+ pricing: {
116
+ approximatePricePerImage: 0.134,
117
+ units: [
118
+ { name: 'imageOutput', rate: 120, strategy: 'fixed', unit: 'millionTokens' },
119
+ { name: 'textInput', rate: 2, strategy: 'fixed', unit: 'millionTokens' },
120
+ { name: 'textOutput', rate: 12, strategy: 'fixed', unit: 'millionTokens' },
121
+ ],
122
+ },
123
+ releasedAt: '2025-11-20',
124
+ settings: {
125
+ searchImpl: 'params',
126
+ searchProvider: 'google',
127
+ },
128
+ type: 'chat',
129
+ },
95
130
  {
96
131
  abilities: {
97
132
  functionCall: true,
@@ -101,9 +136,8 @@ const vertexaiChatModels: AIChatModelCard[] = [
101
136
  },
102
137
  contextWindowTokens: 1_048_576 + 65_536,
103
138
  description:
104
- "Gemini 2.5 Pro is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.",
139
+ 'Gemini 2.5 Pro is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
105
140
  displayName: 'Gemini 2.5 Pro',
106
- enabled: true,
107
141
  id: 'gemini-2.5-pro',
108
142
  maxOutput: 65_536,
109
143
  pricing: {
@@ -129,7 +163,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
129
163
  },
130
164
  contextWindowTokens: 1_048_576 + 65_536,
131
165
  description:
132
- "Gemini 2.5 Pro Preview is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.",
166
+ 'Gemini 2.5 Pro Preview is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
133
167
  displayName: 'Gemini 2.5 Pro Preview 05-06',
134
168
  id: 'gemini-2.5-pro-preview-05-06',
135
169
  maxOutput: 65_536,
@@ -150,7 +184,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
150
184
  },
151
185
  contextWindowTokens: 1_048_576 + 65_536,
152
186
  description:
153
- "Gemini 2.5 Pro Preview is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.",
187
+ 'Gemini 2.5 Pro Preview is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
154
188
  displayName: 'Gemini 2.5 Pro Preview 03-25',
155
189
  id: 'gemini-2.5-pro-preview-03-25',
156
190
  maxOutput: 65_536,
@@ -171,9 +205,8 @@ const vertexaiChatModels: AIChatModelCard[] = [
171
205
  vision: true,
172
206
  },
173
207
  contextWindowTokens: 1_048_576 + 65_536,
174
- description: "Gemini 2.5 Flash is Google’s best-value model with full capabilities.",
208
+ description: 'Gemini 2.5 Flash is Google’s best-value model with full capabilities.',
175
209
  displayName: 'Gemini 2.5 Flash',
176
- enabled: true,
177
210
  id: 'gemini-2.5-flash',
178
211
  maxOutput: 65_536,
179
212
  pricing: {
@@ -198,7 +231,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
198
231
  vision: true,
199
232
  },
200
233
  contextWindowTokens: 1_048_576 + 65_536,
201
- description: "Gemini 2.5 Flash Preview is Google’s best-value model with full capabilities.",
234
+ description: 'Gemini 2.5 Flash Preview is Google’s best-value model with full capabilities.',
202
235
  displayName: 'Gemini 2.5 Flash Preview 04-17',
203
236
  id: 'gemini-2.5-flash-preview-04-17',
204
237
  maxOutput: 65_536,
@@ -218,7 +251,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
218
251
  },
219
252
  contextWindowTokens: 32_768 + 8192,
220
253
  description:
221
- "Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.",
254
+ 'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
222
255
  displayName: 'Nano Banana',
223
256
  enabled: true,
224
257
  id: 'gemini-2.5-flash-image',
@@ -243,7 +276,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
243
276
  },
244
277
  contextWindowTokens: 1_000_000 + 64_000,
245
278
  description:
246
- "Gemini 2.5 Flash-Lite is Google’s smallest, best-value model, designed for large-scale use.",
279
+ 'Gemini 2.5 Flash-Lite is Google’s smallest, best-value model, designed for large-scale use.',
247
280
  displayName: 'Gemini 2.5 Flash-Lite',
248
281
  id: 'gemini-2.5-flash-lite',
249
282
  maxOutput: 64_000,
@@ -270,7 +303,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
270
303
  },
271
304
  contextWindowTokens: 1_000_000 + 64_000,
272
305
  description:
273
- "Gemini 2.5 Flash-Lite Preview is Google’s smallest, best-value model, designed for large-scale use.",
306
+ 'Gemini 2.5 Flash-Lite Preview is Google’s smallest, best-value model, designed for large-scale use.',
274
307
  displayName: 'Gemini 2.5 Flash-Lite Preview 06-17',
275
308
  id: 'gemini-2.5-flash-lite-preview-06-17',
276
309
  maxOutput: 64_000,
@@ -315,8 +348,7 @@ const vertexaiChatModels: AIChatModelCard[] = [
315
348
  vision: true,
316
349
  },
317
350
  contextWindowTokens: 1_048_576 + 8192,
318
- description:
319
- 'A Gemini 2.0 Flash variant optimized for cost efficiency and low latency.',
351
+ description: 'A Gemini 2.0 Flash variant optimized for cost efficiency and low latency.',
320
352
  displayName: 'Gemini 2.0 Flash-Lite',
321
353
  id: 'gemini-2.0-flash-lite',
322
354
  maxOutput: 8192,
@@ -336,7 +368,8 @@ const vertexaiChatModels: AIChatModelCard[] = [
336
368
  vision: true,
337
369
  },
338
370
  contextWindowTokens: 1_000_000 + 8192,
339
- description: 'Gemini 1.5 Flash 002 is an efficient multimodal model built for broad deployment.',
371
+ description:
372
+ 'Gemini 1.5 Flash 002 is an efficient multimodal model built for broad deployment.',
340
373
  displayName: 'Gemini 1.5 Flash 002',
341
374
  id: 'gemini-1.5-flash-002',
342
375
  maxOutput: 8192,
@@ -379,7 +412,7 @@ const vertexaiImageModels: AIImageModelCard[] = [
379
412
  enabled: true,
380
413
  type: 'image',
381
414
  description:
382
- "Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.",
415
+ 'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
383
416
  releasedAt: '2025-08-26',
384
417
  parameters: nanoBananaParameters,
385
418
  pricing: {
@@ -3,6 +3,66 @@ import { AIChatModelCard, AIImageModelCard } from '../types/aiModel';
3
3
  // https://www.volcengine.com/docs/82379/1330310
4
4
 
5
5
  const doubaoChatModels: AIChatModelCard[] = [
6
+ {
7
+ abilities: {
8
+ functionCall: true,
9
+ reasoning: true,
10
+ vision: true,
11
+ },
12
+ config: {
13
+ deploymentName: 'doubao-seed-1-8-251215',
14
+ },
15
+ contextWindowTokens: 256_000,
16
+ description:
17
+ 'Doubao-Seed-1.8 有着更强的多模态理解能力和 Agent 能力,支持文本/图片/视频输入与上下文缓存,可在复杂任务中提供更出色的表现。',
18
+ displayName: 'Doubao Seed 1.8',
19
+ id: 'doubao-seed-1.8',
20
+ maxOutput: 64_000,
21
+ pricing: {
22
+ currency: 'CNY',
23
+ units: [
24
+ {
25
+ lookup: {
26
+ prices: {
27
+ '[0, 0.032]': 0.8,
28
+ '[0.032, 0.128]': 1.2,
29
+ '[0.128, 0.256]': 2.4,
30
+ },
31
+ pricingParams: ['textInputRange'],
32
+ },
33
+ name: 'textInput',
34
+ strategy: 'lookup',
35
+ unit: 'millionTokens',
36
+ },
37
+ {
38
+ lookup: {
39
+ prices: {
40
+ '[0, 0.032]_[0, 0.0002]': 2,
41
+ '[0, 0.032]_[0.0002, infinity]': 8,
42
+ '[0.032, 0.128]_[0, infinity]': 16,
43
+ '[0.128, 0.256]_[0, infinity]': 24,
44
+ },
45
+ pricingParams: ['textInputRange', 'textOutputRange'],
46
+ },
47
+ name: 'textOutput',
48
+ strategy: 'lookup',
49
+ unit: 'millionTokens',
50
+ },
51
+ { name: 'textInput_cacheRead', rate: 0.16, strategy: 'fixed', unit: 'millionTokens' },
52
+ {
53
+ lookup: { prices: { '1h': 0.017 }, pricingParams: ['ttl'] },
54
+ name: 'textInput_cacheWrite',
55
+ strategy: 'lookup',
56
+ unit: 'millionTokens',
57
+ },
58
+ ],
59
+ },
60
+ releasedAt: '2025-12-18',
61
+ settings: {
62
+ extendParams: ['gpt5ReasoningEffort'],
63
+ },
64
+ type: 'chat',
65
+ },
6
66
  {
7
67
  abilities: {
8
68
  functionCall: true,
@@ -56,6 +116,57 @@ const doubaoChatModels: AIChatModelCard[] = [
56
116
  },
57
117
  type: 'chat',
58
118
  },
119
+ {
120
+ abilities: {
121
+ functionCall: true,
122
+ reasoning: true,
123
+ },
124
+ config: {
125
+ deploymentName: 'deepseek-v3-2-251201',
126
+ },
127
+ contextWindowTokens: 131_072,
128
+ description:
129
+ 'DeepSeek-V3.2 是深度求索推出的首个将思考融入工具使用的混合推理模型,用高效架构省算力、大规模强化学习提能力、大规模合成任务数据强泛化,三者结合性能媲美 GPT-5-High,输出长度大幅降低,显著减少了计算开销与用户等待时间。',
130
+ displayName: 'DeepSeek V3.2',
131
+ enabled: true,
132
+ id: 'deepseek-v3.2',
133
+ maxOutput: 32_768,
134
+ pricing: {
135
+ currency: 'CNY',
136
+ units: [
137
+ {
138
+ lookup: {
139
+ prices: {
140
+ '[0, 0.032]': 2,
141
+ '[0.032, 0.128]': 4,
142
+ },
143
+ pricingParams: ['textInputRange'],
144
+ },
145
+ name: 'textInput',
146
+ strategy: 'lookup',
147
+ unit: 'millionTokens',
148
+ },
149
+ {
150
+ lookup: {
151
+ prices: {
152
+ '[0, 0.032]': 3,
153
+ '[0.032, 0.128]': 6,
154
+ },
155
+ pricingParams: ['textInputRange'],
156
+ },
157
+ name: 'textOutput',
158
+ strategy: 'lookup',
159
+ unit: 'millionTokens',
160
+ },
161
+ { name: 'textInput_cacheRead', rate: 0.4, strategy: 'fixed', unit: 'millionTokens' },
162
+ { name: 'textInput_cacheWrite', rate: 0.017, strategy: 'fixed', unit: 'millionTokens' },
163
+ ],
164
+ },
165
+ settings: {
166
+ extendParams: ['enableReasoning'],
167
+ },
168
+ type: 'chat',
169
+ },
59
170
  {
60
171
  abilities: {
61
172
  functionCall: true,
@@ -171,7 +282,6 @@ const doubaoChatModels: AIChatModelCard[] = [
171
282
  description:
172
283
  'Doubao-Seed-1.6-thinking significantly strengthens reasoning, further improving core abilities in coding, math, and logical reasoning over Doubao-1.5-thinking-pro, while adding vision understanding. It supports a 256k context window and up to 16k output tokens.',
173
284
  displayName: 'Doubao Seed 1.6 Thinking',
174
- enabled: true,
175
285
  id: 'doubao-seed-1.6-thinking',
176
286
  maxOutput: 32_000,
177
287
  pricing: {
@@ -221,7 +331,6 @@ const doubaoChatModels: AIChatModelCard[] = [
221
331
  description:
222
332
  'Doubao-Seed-1.6 is a new multimodal deep-reasoning model with auto, thinking, and non-thinking modes. In non-thinking mode, it significantly outperforms Doubao-1.5-pro/250115. It supports a 256k context window and up to 16k output tokens.',
223
333
  displayName: 'Doubao Seed 1.6',
224
- enabled: true,
225
334
  id: 'doubao-seed-1.6',
226
335
  maxOutput: 32_000,
227
336
  pricing: {
@@ -32,7 +32,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
32
32
  },
33
33
  releasedAt: '2025-12-11',
34
34
  settings: {
35
- extendParams: ['gpt5_1ReasoningEffort', 'textVerbosity'],
35
+ extendParams: ['gpt5_2ReasoningEffort', 'textVerbosity'],
36
36
  searchImpl: 'params',
37
37
  },
38
38
  type: 'chat',
@@ -147,6 +147,11 @@ const zenmuxChatModels: AIChatModelCard[] = [
147
147
  ],
148
148
  },
149
149
  releasedAt: '2025-11-20',
150
+ settings: {
151
+ extendParams: ['thinkingLevel2', 'urlContext'],
152
+ searchImpl: 'params',
153
+ searchProvider: 'google',
154
+ },
150
155
  type: 'chat',
151
156
  },
152
157
  {
@@ -170,7 +175,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
170
175
  ],
171
176
  },
172
177
  settings: {
173
- extendParams: ['thinkingLevel', 'urlContext'],
178
+ extendParams: ['thinkingLevel2', 'urlContext'],
174
179
  searchImpl: 'params',
175
180
  searchProvider: 'google',
176
181
  },
@@ -348,7 +353,8 @@ const zenmuxChatModels: AIChatModelCard[] = [
348
353
  reasoning: true,
349
354
  },
350
355
  contextWindowTokens: 256_000,
351
- description: "Grok 4 is xAI's flagship reasoning model with strong reasoning and multimodal capability.",
356
+ description:
357
+ "Grok 4 is xAI's flagship reasoning model with strong reasoning and multimodal capability.",
352
358
  displayName: 'Grok 4',
353
359
  id: 'x-ai/grok-4',
354
360
  maxOutput: 256_000,
@@ -707,7 +713,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
707
713
  },
708
714
  contextWindowTokens: 1_050_000,
709
715
  description:
710
- "Gemini 2.5 Pro is Google’s flagship reasoning model with long context support for complex tasks.",
716
+ 'Gemini 2.5 Pro is Google’s flagship reasoning model with long context support for complex tasks.',
711
717
  displayName: 'Gemini 2.5 Pro',
712
718
  id: 'google/gemini-2.5-pro',
713
719
  maxOutput: 65_530,
@@ -759,7 +765,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
759
765
  },
760
766
  contextWindowTokens: 1_050_000,
761
767
  description:
762
- "Gemini 2.5 Flash (Lite/Pro/Flash) is Google’s family spanning low latency to high-performance reasoning.",
768
+ 'Gemini 2.5 Flash (Lite/Pro/Flash) is Google’s family spanning low latency to high-performance reasoning.',
763
769
  displayName: 'Gemini 2.5 Flash',
764
770
  id: 'google/gemini-2.5-flash',
765
771
  maxOutput: 65_530,
@@ -931,8 +937,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
931
937
  reasoning: true,
932
938
  },
933
939
  contextWindowTokens: 1_050_000,
934
- description:
935
- 'GPT-4.1 Mini offers lower latency and better value for mid-context workloads.',
940
+ description: 'GPT-4.1 Mini offers lower latency and better value for mid-context workloads.',
936
941
  displayName: 'GPT-4.1 Mini',
937
942
  id: 'openai/gpt-4.1-mini',
938
943
  maxOutput: 32_770,
@@ -1076,7 +1081,8 @@ const zenmuxChatModels: AIChatModelCard[] = [
1076
1081
  reasoning: true,
1077
1082
  },
1078
1083
  contextWindowTokens: 200_000,
1079
- description: "GLM 4.6 is Z.AI's flagship model with extended context length and coding capability.",
1084
+ description:
1085
+ "GLM 4.6 is Z.AI's flagship model with extended context length and coding capability.",
1080
1086
  displayName: 'GLM 4.6',
1081
1087
  id: 'z-ai/glm-4.6',
1082
1088
  maxOutput: 128_000,
@@ -1125,7 +1131,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
1125
1131
  },
1126
1132
  contextWindowTokens: 200_000,
1127
1133
  description:
1128
- "Opus 4.1 is Anthropic’s high-end model optimized for programming, complex reasoning, and long-running tasks.",
1134
+ 'Opus 4.1 is Anthropic’s high-end model optimized for programming, complex reasoning, and long-running tasks.',
1129
1135
  displayName: 'Claude Opus 4.1',
1130
1136
  id: 'anthropic/claude-opus-4.1',
1131
1137
  maxOutput: 32_000,
@@ -1150,7 +1156,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
1150
1156
  },
1151
1157
  contextWindowTokens: 200_000,
1152
1158
  description:
1153
- "Opus 4 is Anthropic’s flagship model designed for complex tasks and enterprise applications.",
1159
+ 'Opus 4 is Anthropic’s flagship model designed for complex tasks and enterprise applications.',
1154
1160
  displayName: 'Claude Opus 4',
1155
1161
  id: 'anthropic/claude-opus-4',
1156
1162
  maxOutput: 32_000,
@@ -1218,7 +1224,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
1218
1224
  },
1219
1225
  contextWindowTokens: 200_000,
1220
1226
  description:
1221
- "Claude Sonnet 4.5 is Anthropic’s latest hybrid reasoning model optimized for complex reasoning and coding.",
1227
+ 'Claude Sonnet 4.5 is Anthropic’s latest hybrid reasoning model optimized for complex reasoning and coding.',
1222
1228
  displayName: 'Claude Sonnet 4.5',
1223
1229
  id: 'anthropic/claude-sonnet-4.5',
1224
1230
  maxOutput: 64_000,
@@ -1259,7 +1265,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
1259
1265
  reasoning: true,
1260
1266
  },
1261
1267
  contextWindowTokens: 200_000,
1262
- description: "OpenAI o4-mini is a small, efficient reasoning model for low-latency scenarios.",
1268
+ description: 'OpenAI o4-mini is a small, efficient reasoning model for low-latency scenarios.',
1263
1269
  displayName: 'o4 Mini',
1264
1270
  id: 'openai/o4-mini',
1265
1271
  maxOutput: 100_000,
@@ -1542,7 +1548,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
1542
1548
  abilities: { reasoning: true },
1543
1549
  contextWindowTokens: 128_000,
1544
1550
  description:
1545
- "GLM 4.5 is Z.AI’s flagship model with hybrid reasoning optimized for engineering and long-context tasks.",
1551
+ 'GLM 4.5 is Z.AI’s flagship model with hybrid reasoning optimized for engineering and long-context tasks.',
1546
1552
  displayName: 'GLM 4.5',
1547
1553
  id: 'z-ai/glm-4.5',
1548
1554
  maxOutput: 96_000,
@@ -12,9 +12,71 @@ const zhipuChatModels: AIChatModelCard[] = [
12
12
  },
13
13
  contextWindowTokens: 200_000,
14
14
  description:
15
- 'Zhipu’s latest flagship GLM-4.6 (355B) surpasses the prior generation in advanced coding, long-context handling, reasoning, and agent capabilities. Its coding performance is on par with Claude Sonnet 4, making it a top-tier coding model in China.',
16
- displayName: 'GLM-4.6',
15
+ 'GLM-4.7 是智谱最新旗舰模型,面向 Agentic Coding 场景强化了编码能力、长程任务规划与工具协同,并在多个公开基准的当期榜单中取得开源模型中的领先表现。通用能力提升,回复更简洁自然,写作更具沉浸感。在执行复杂智能体任务,在工具调用时指令遵循更强,Artifacts Agentic Coding 的前端美感和长程任务完成效率进一步提升。',
16
+ displayName: 'GLM-4.7',
17
17
  enabled: true,
18
+ id: 'glm-4.7',
19
+ maxOutput: 131_072,
20
+ pricing: {
21
+ currency: 'CNY',
22
+ units: [
23
+ {
24
+ lookup: {
25
+ prices: {
26
+ '[0, 0.032]_[0, 0.0002]': 0.4,
27
+ '[0, 0.032]_[0.0002, infinity]': 0.6,
28
+ '[0.032, 0.2]': 0.8,
29
+ },
30
+ pricingParams: ['textInput', 'textOutput'],
31
+ },
32
+ name: 'textInput_cacheRead',
33
+ strategy: 'lookup',
34
+ unit: 'millionTokens',
35
+ },
36
+ {
37
+ lookup: {
38
+ prices: {
39
+ '[0, 0.032]_[0, 0.0002]': 2,
40
+ '[0, 0.032]_[0.0002, infinity]': 3,
41
+ '[0.032, 0.2]': 4,
42
+ },
43
+ pricingParams: ['textInput', 'textOutput'],
44
+ },
45
+ name: 'textInput',
46
+ strategy: 'lookup',
47
+ unit: 'millionTokens',
48
+ },
49
+ {
50
+ lookup: {
51
+ prices: {
52
+ '[0, 0.032]_[0, 0.0002]': 8,
53
+ '[0, 0.032]_[0.0002, infinity]': 14,
54
+ '[0.032, 0.2]': 16,
55
+ },
56
+ pricingParams: ['textInput', 'textOutput'],
57
+ },
58
+ name: 'textOutput',
59
+ strategy: 'lookup',
60
+ unit: 'millionTokens',
61
+ },
62
+ ],
63
+ },
64
+ settings: {
65
+ extendParams: ['enableReasoning'],
66
+ searchImpl: 'params',
67
+ },
68
+ type: 'chat',
69
+ },
70
+ {
71
+ abilities: {
72
+ functionCall: true,
73
+ reasoning: true,
74
+ search: true,
75
+ },
76
+ contextWindowTokens: 200_000,
77
+ description:
78
+ '智谱最新旗舰模型 GLM-4.6 (355B) 在高级编码、长文本处理、推理与智能体能力上全面超越前代,尤其在编程能力上对齐 Claude Sonnet 4,成为国内顶尖的 Coding 模型。',
79
+ displayName: 'GLM-4.6',
18
80
  id: 'glm-4.6',
19
81
  maxOutput: 131_072,
20
82
  pricing: {
@@ -239,10 +239,13 @@ export type ExtendParamsType =
239
239
  | 'reasoningEffort'
240
240
  | 'gpt5ReasoningEffort'
241
241
  | 'gpt5_1ReasoningEffort'
242
+ | 'gpt5_2ReasoningEffort'
243
+ | 'gpt5_2ProReasoningEffort'
242
244
  | 'textVerbosity'
243
245
  | 'thinking'
244
246
  | 'thinkingBudget'
245
247
  | 'thinkingLevel'
248
+ | 'thinkingLevel2'
246
249
  | 'imageAspectRatio'
247
250
  | 'imageResolution'
248
251
  | 'urlContext';