@lobehub/lobehub 2.0.0-next.322 → 2.0.0-next.324
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +80 -0
- package/apps/desktop/src/main/core/infrastructure/UpdaterManager.ts +9 -76
- package/apps/desktop/src/main/core/infrastructure/__tests__/UpdaterManager.test.ts +0 -1
- package/apps/desktop/src/main/modules/updater/configs.ts +0 -4
- package/changelog/v1.json +24 -0
- package/e2e/src/mocks/llm/index.ts +3 -3
- package/locales/ar/common.json +5 -0
- package/locales/ar/error.json +10 -1
- package/locales/bg-BG/common.json +5 -0
- package/locales/bg-BG/error.json +10 -1
- package/locales/de-DE/common.json +5 -0
- package/locales/de-DE/error.json +10 -1
- package/locales/en-US/common.json +5 -0
- package/locales/es-ES/common.json +5 -0
- package/locales/es-ES/error.json +10 -1
- package/locales/fa-IR/common.json +5 -0
- package/locales/fa-IR/error.json +10 -1
- package/locales/fr-FR/common.json +5 -0
- package/locales/fr-FR/error.json +10 -1
- package/locales/it-IT/common.json +5 -0
- package/locales/it-IT/error.json +10 -1
- package/locales/ja-JP/common.json +5 -0
- package/locales/ja-JP/error.json +10 -1
- package/locales/ko-KR/common.json +5 -0
- package/locales/ko-KR/error.json +10 -1
- package/locales/nl-NL/common.json +5 -0
- package/locales/nl-NL/error.json +10 -1
- package/locales/pl-PL/common.json +5 -0
- package/locales/pl-PL/error.json +10 -1
- package/locales/pt-BR/common.json +5 -0
- package/locales/pt-BR/error.json +10 -1
- package/locales/ru-RU/common.json +5 -0
- package/locales/ru-RU/error.json +10 -1
- package/locales/tr-TR/common.json +5 -0
- package/locales/tr-TR/error.json +10 -1
- package/locales/vi-VN/common.json +5 -0
- package/locales/vi-VN/error.json +10 -1
- package/locales/zh-CN/common.json +5 -0
- package/locales/zh-TW/common.json +5 -0
- package/locales/zh-TW/error.json +10 -1
- package/package.json +2 -2
- package/packages/business/const/src/branding.ts +1 -0
- package/packages/business/const/src/llm.ts +2 -1
- package/packages/const/src/settings/llm.ts +2 -1
- package/packages/const/src/settings/systemAgent.ts +12 -7
- package/packages/database/src/models/agent.ts +18 -1
- package/packages/database/src/models/chatGroup.ts +18 -1
- package/packages/database/src/types/chatGroup.ts +1 -0
- package/packages/model-bank/package.json +1 -1
- package/packages/model-bank/src/aiModels/index.ts +2 -2
- package/packages/model-bank/src/aiModels/lobehub/chat/anthropic.ts +256 -0
- package/packages/model-bank/src/aiModels/lobehub/chat/deepseek.ts +45 -0
- package/packages/model-bank/src/aiModels/lobehub/chat/google.ts +267 -0
- package/packages/model-bank/src/aiModels/lobehub/chat/index.ts +26 -0
- package/packages/model-bank/src/aiModels/lobehub/chat/minimax.ts +75 -0
- package/packages/model-bank/src/aiModels/lobehub/chat/moonshot.ts +28 -0
- package/packages/model-bank/src/aiModels/lobehub/chat/openai.ts +345 -0
- package/packages/model-bank/src/aiModels/lobehub/chat/xai.ts +32 -0
- package/packages/model-bank/src/aiModels/lobehub/image.ts +240 -0
- package/packages/model-bank/src/aiModels/lobehub/index.ts +10 -0
- package/packages/model-bank/src/aiModels/lobehub/utils.ts +58 -0
- package/packages/model-bank/src/modelProviders/index.ts +10 -10
- package/packages/model-runtime/src/core/streams/qwen.test.ts +320 -0
- package/packages/model-runtime/src/core/streams/qwen.ts +19 -10
- package/packages/types/package.json +1 -1
- package/packages/types/src/agentGroup/index.ts +2 -0
- package/packages/types/src/discover/assistants.ts +9 -0
- package/packages/types/src/discover/fork.ts +163 -0
- package/packages/types/src/discover/groupAgents.ts +13 -4
- package/packages/types/src/discover/index.ts +9 -0
- package/src/app/[variants]/(auth)/_layout/index.tsx +2 -1
- package/src/app/[variants]/(auth)/auth-error/page.tsx +5 -5
- package/src/app/[variants]/(desktop)/desktop-onboarding/components/LobeMessage.tsx +5 -0
- package/src/app/[variants]/(desktop)/desktop-onboarding/features/WelcomeStep.tsx +3 -1
- package/src/app/[variants]/(main)/agent/_layout/Sidebar/Topic/List/Item/index.tsx +1 -2
- package/src/app/[variants]/(main)/community/(detail)/agent/features/Header.tsx +37 -0
- package/src/app/[variants]/(main)/community/(detail)/agent/features/Sidebar/ActionButton/ForkAndChat.tsx +133 -0
- package/src/app/[variants]/(main)/community/(detail)/agent/features/Sidebar/ActionButton/index.tsx +2 -2
- package/src/app/[variants]/(main)/community/(detail)/group_agent/features/Details/index.tsx +7 -10
- package/src/app/[variants]/(main)/community/(detail)/group_agent/features/Sidebar/ActionButton/ForkGroupAndChat.tsx +208 -0
- package/src/app/[variants]/(main)/community/(detail)/group_agent/features/Sidebar/ActionButton/index.tsx +2 -2
- package/src/app/[variants]/(main)/community/(detail)/user/features/DetailProvider.tsx +2 -0
- package/src/app/[variants]/(main)/community/(detail)/user/features/UserContent.tsx +7 -0
- package/src/app/[variants]/(main)/community/(detail)/user/features/UserForkedAgentGroups.tsx +63 -0
- package/src/app/[variants]/(main)/community/(detail)/user/features/UserForkedAgents.tsx +61 -0
- package/src/app/[variants]/(main)/community/(detail)/user/index.tsx +3 -1
- package/src/app/[variants]/(main)/group/_layout/Sidebar/Topic/List/Item/index.tsx +1 -2
- package/src/app/[variants]/(main)/home/features/WelcomeText/index.tsx +3 -1
- package/src/app/[variants]/(main)/settings/profile/index.tsx +92 -68
- package/src/app/[variants]/(mobile)/chat/features/Topic/index.tsx +2 -1
- package/src/app/[variants]/onboarding/components/LobeMessage.tsx +5 -0
- package/src/app/[variants]/onboarding/features/TelemetryStep.tsx +3 -1
- package/src/features/CommandMenu/AskAgentCommands.tsx +105 -0
- package/src/features/CommandMenu/CommandMenuContext.tsx +57 -38
- package/src/features/CommandMenu/components/CommandInput.tsx +43 -9
- package/src/features/CommandMenu/index.tsx +89 -27
- package/src/features/CommandMenu/types.ts +6 -0
- package/src/features/CommandMenu/useCommandMenu.ts +62 -39
- package/src/locales/default/common.ts +5 -0
- package/src/locales/default/discover.ts +371 -0
- package/src/server/globalConfig/parseMemoryExtractionConfig.ts +7 -8
- package/src/server/routers/lambda/agent.ts +14 -0
- package/src/server/routers/lambda/agentGroup.ts +19 -3
- package/src/server/routers/lambda/market/agent.ts +234 -26
- package/src/server/routers/lambda/market/agentGroup.ts +204 -1
- package/src/server/services/discover/index.ts +52 -2
- package/src/services/agent.ts +8 -0
- package/src/services/chatGroup/index.ts +8 -0
- package/src/services/marketApi.ts +78 -0
- package/src/store/user/slices/settings/selectors/__snapshots__/settings.test.ts.snap +12 -12
- package/packages/model-bank/src/aiModels/lobehub.ts +0 -1315
|
@@ -1,1315 +0,0 @@
|
|
|
1
|
-
import { ModelParamsSchema } from 'model-bank';
|
|
2
|
-
|
|
3
|
-
import { AIChatModelCard, AIImageModelCard } from '../types/aiModel';
|
|
4
|
-
import { huanyuanImageParamsSchema, qwenEditParamsSchema, qwenImageParamsSchema } from './fal';
|
|
5
|
-
|
|
6
|
-
const lobehubChatModels: AIChatModelCard[] = [
|
|
7
|
-
{
|
|
8
|
-
abilities: {
|
|
9
|
-
functionCall: true,
|
|
10
|
-
reasoning: true,
|
|
11
|
-
search: true,
|
|
12
|
-
vision: true,
|
|
13
|
-
},
|
|
14
|
-
contextWindowTokens: 400_000,
|
|
15
|
-
description:
|
|
16
|
-
"GPT-5.2 — OpenAI's flagship model optimized for coding and agent tasks across industries.",
|
|
17
|
-
displayName: 'GPT-5.2',
|
|
18
|
-
enabled: true,
|
|
19
|
-
id: 'gpt-5.2',
|
|
20
|
-
maxOutput: 128_000,
|
|
21
|
-
pricing: {
|
|
22
|
-
units: [
|
|
23
|
-
{ name: 'textInput', rate: 1.75, strategy: 'fixed', unit: 'millionTokens' },
|
|
24
|
-
{ name: 'textInput_cacheRead', rate: 0.175, strategy: 'fixed', unit: 'millionTokens' },
|
|
25
|
-
{ name: 'textOutput', rate: 14, strategy: 'fixed', unit: 'millionTokens' },
|
|
26
|
-
],
|
|
27
|
-
},
|
|
28
|
-
releasedAt: '2025-12-11',
|
|
29
|
-
settings: {
|
|
30
|
-
extendParams: ['gpt5_2ReasoningEffort', 'textVerbosity'],
|
|
31
|
-
},
|
|
32
|
-
type: 'chat',
|
|
33
|
-
},
|
|
34
|
-
{
|
|
35
|
-
abilities: {
|
|
36
|
-
functionCall: true,
|
|
37
|
-
reasoning: true,
|
|
38
|
-
search: true,
|
|
39
|
-
vision: true,
|
|
40
|
-
},
|
|
41
|
-
contextWindowTokens: 400_000,
|
|
42
|
-
description:
|
|
43
|
-
'GPT-5.1 — a flagship model optimized for coding and agent tasks with configurable reasoning effort and longer context.',
|
|
44
|
-
displayName: 'GPT-5.1',
|
|
45
|
-
enabled: true,
|
|
46
|
-
id: 'gpt-5.1',
|
|
47
|
-
maxOutput: 128_000,
|
|
48
|
-
pricing: {
|
|
49
|
-
units: [
|
|
50
|
-
{ name: 'textInput', rate: 1.25, strategy: 'fixed', unit: 'millionTokens' },
|
|
51
|
-
{ name: 'textInput_cacheRead', rate: 0.125, strategy: 'fixed', unit: 'millionTokens' },
|
|
52
|
-
{ name: 'textOutput', rate: 10, strategy: 'fixed', unit: 'millionTokens' },
|
|
53
|
-
],
|
|
54
|
-
},
|
|
55
|
-
releasedAt: '2025-11-13',
|
|
56
|
-
settings: {
|
|
57
|
-
extendParams: ['gpt5_1ReasoningEffort', 'textVerbosity'],
|
|
58
|
-
},
|
|
59
|
-
type: 'chat',
|
|
60
|
-
},
|
|
61
|
-
{
|
|
62
|
-
abilities: {
|
|
63
|
-
functionCall: true,
|
|
64
|
-
reasoning: true,
|
|
65
|
-
search: true,
|
|
66
|
-
vision: true,
|
|
67
|
-
},
|
|
68
|
-
contextWindowTokens: 400_000,
|
|
69
|
-
description:
|
|
70
|
-
'The best model for cross-domain coding and agent tasks. GPT-5 leaps in accuracy, speed, reasoning, context awareness, structured thinking, and problem solving.',
|
|
71
|
-
displayName: 'GPT-5',
|
|
72
|
-
enabled: true,
|
|
73
|
-
id: 'gpt-5',
|
|
74
|
-
maxOutput: 128_000,
|
|
75
|
-
pricing: {
|
|
76
|
-
units: [
|
|
77
|
-
{ name: 'textInput', rate: 1.25, strategy: 'fixed', unit: 'millionTokens' },
|
|
78
|
-
{ name: 'textOutput', rate: 10, strategy: 'fixed', unit: 'millionTokens' },
|
|
79
|
-
{ name: 'textInput_cacheRead', rate: 0.13, strategy: 'fixed', unit: 'millionTokens' },
|
|
80
|
-
],
|
|
81
|
-
},
|
|
82
|
-
releasedAt: '2025-08-07',
|
|
83
|
-
settings: {
|
|
84
|
-
extendParams: ['reasoningEffort'],
|
|
85
|
-
},
|
|
86
|
-
type: 'chat',
|
|
87
|
-
},
|
|
88
|
-
{
|
|
89
|
-
abilities: {
|
|
90
|
-
functionCall: true,
|
|
91
|
-
reasoning: true,
|
|
92
|
-
search: true,
|
|
93
|
-
vision: true,
|
|
94
|
-
},
|
|
95
|
-
contextWindowTokens: 400_000,
|
|
96
|
-
description:
|
|
97
|
-
'A faster, more cost-efficient GPT-5 variant for well-defined tasks, delivering quicker responses while maintaining quality.',
|
|
98
|
-
displayName: 'GPT-5 mini',
|
|
99
|
-
enabled: true,
|
|
100
|
-
id: 'gpt-5-mini',
|
|
101
|
-
maxOutput: 128_000,
|
|
102
|
-
pricing: {
|
|
103
|
-
units: [
|
|
104
|
-
{ name: 'textInput', rate: 0.25, strategy: 'fixed', unit: 'millionTokens' },
|
|
105
|
-
{ name: 'textOutput', rate: 2, strategy: 'fixed', unit: 'millionTokens' },
|
|
106
|
-
{ name: 'textInput_cacheRead', rate: 0.03, strategy: 'fixed', unit: 'millionTokens' },
|
|
107
|
-
],
|
|
108
|
-
},
|
|
109
|
-
releasedAt: '2025-08-07',
|
|
110
|
-
settings: {
|
|
111
|
-
extendParams: ['reasoningEffort'],
|
|
112
|
-
searchImpl: 'params',
|
|
113
|
-
},
|
|
114
|
-
type: 'chat',
|
|
115
|
-
},
|
|
116
|
-
{
|
|
117
|
-
abilities: {
|
|
118
|
-
functionCall: true,
|
|
119
|
-
reasoning: true,
|
|
120
|
-
vision: true,
|
|
121
|
-
},
|
|
122
|
-
contextWindowTokens: 400_000,
|
|
123
|
-
description:
|
|
124
|
-
'The fastest and most cost-effective GPT-5 variant, ideal for latency- and cost-sensitive applications.',
|
|
125
|
-
displayName: 'GPT-5 nano',
|
|
126
|
-
id: 'gpt-5-nano',
|
|
127
|
-
maxOutput: 128_000,
|
|
128
|
-
pricing: {
|
|
129
|
-
units: [
|
|
130
|
-
{ name: 'textInput', rate: 0.05, strategy: 'fixed', unit: 'millionTokens' },
|
|
131
|
-
{ name: 'textOutput', rate: 0.4, strategy: 'fixed', unit: 'millionTokens' },
|
|
132
|
-
{ name: 'textInput_cacheRead', rate: 0.01, strategy: 'fixed', unit: 'millionTokens' },
|
|
133
|
-
],
|
|
134
|
-
},
|
|
135
|
-
releasedAt: '2025-08-07',
|
|
136
|
-
type: 'chat',
|
|
137
|
-
},
|
|
138
|
-
{
|
|
139
|
-
abilities: {
|
|
140
|
-
vision: true,
|
|
141
|
-
},
|
|
142
|
-
contextWindowTokens: 400_000,
|
|
143
|
-
description:
|
|
144
|
-
'The GPT-5 model used in ChatGPT, combining strong understanding and generation for conversational applications.',
|
|
145
|
-
displayName: 'GPT-5 Chat',
|
|
146
|
-
enabled: true,
|
|
147
|
-
id: 'gpt-5-chat-latest',
|
|
148
|
-
maxOutput: 128_000,
|
|
149
|
-
pricing: {
|
|
150
|
-
units: [
|
|
151
|
-
{ name: 'textInput', rate: 1.25, strategy: 'fixed', unit: 'millionTokens' },
|
|
152
|
-
{ name: 'textOutput', rate: 10, strategy: 'fixed', unit: 'millionTokens' },
|
|
153
|
-
{ name: 'textInput_cacheRead', rate: 0.13, strategy: 'fixed', unit: 'millionTokens' },
|
|
154
|
-
],
|
|
155
|
-
},
|
|
156
|
-
releasedAt: '2025-08-07',
|
|
157
|
-
type: 'chat',
|
|
158
|
-
},
|
|
159
|
-
{
|
|
160
|
-
abilities: {
|
|
161
|
-
functionCall: true,
|
|
162
|
-
vision: true,
|
|
163
|
-
},
|
|
164
|
-
contextWindowTokens: 1_047_576,
|
|
165
|
-
description:
|
|
166
|
-
'GPT-4.1 is our flagship model for complex tasks and cross-domain problem solving.',
|
|
167
|
-
displayName: 'GPT-4.1',
|
|
168
|
-
enabled: true,
|
|
169
|
-
id: 'gpt-4.1',
|
|
170
|
-
maxOutput: 32_768,
|
|
171
|
-
pricing: {
|
|
172
|
-
units: [
|
|
173
|
-
{ name: 'textInput', rate: 2, strategy: 'fixed', unit: 'millionTokens' },
|
|
174
|
-
{ name: 'textOutput', rate: 8, strategy: 'fixed', unit: 'millionTokens' },
|
|
175
|
-
{ name: 'textInput_cacheRead', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
176
|
-
],
|
|
177
|
-
},
|
|
178
|
-
releasedAt: '2025-04-14',
|
|
179
|
-
type: 'chat',
|
|
180
|
-
},
|
|
181
|
-
{
|
|
182
|
-
abilities: {
|
|
183
|
-
functionCall: true,
|
|
184
|
-
vision: true,
|
|
185
|
-
},
|
|
186
|
-
contextWindowTokens: 1_047_576,
|
|
187
|
-
description:
|
|
188
|
-
'GPT-4.1 mini balances intelligence, speed, and cost, making it attractive for many use cases.',
|
|
189
|
-
displayName: 'GPT-4.1 mini',
|
|
190
|
-
id: 'gpt-4.1-mini',
|
|
191
|
-
maxOutput: 32_768,
|
|
192
|
-
pricing: {
|
|
193
|
-
units: [
|
|
194
|
-
{ name: 'textInput', rate: 0.4, strategy: 'fixed', unit: 'millionTokens' },
|
|
195
|
-
{ name: 'textOutput', rate: 1.6, strategy: 'fixed', unit: 'millionTokens' },
|
|
196
|
-
{ name: 'textInput_cacheRead', rate: 0.1, strategy: 'fixed', unit: 'millionTokens' },
|
|
197
|
-
],
|
|
198
|
-
},
|
|
199
|
-
releasedAt: '2025-04-14',
|
|
200
|
-
type: 'chat',
|
|
201
|
-
},
|
|
202
|
-
{
|
|
203
|
-
abilities: {
|
|
204
|
-
functionCall: true,
|
|
205
|
-
vision: true,
|
|
206
|
-
},
|
|
207
|
-
contextWindowTokens: 1_047_576,
|
|
208
|
-
description:
|
|
209
|
-
'GPT-4.1 mini balances intelligence, speed, and cost, making it attractive for many use cases.',
|
|
210
|
-
displayName: 'GPT-4.1 nano',
|
|
211
|
-
id: 'gpt-4.1-nano',
|
|
212
|
-
maxOutput: 32_768,
|
|
213
|
-
pricing: {
|
|
214
|
-
units: [
|
|
215
|
-
{ name: 'textInput', rate: 0.1, strategy: 'fixed', unit: 'millionTokens' },
|
|
216
|
-
{ name: 'textOutput', rate: 0.4, strategy: 'fixed', unit: 'millionTokens' },
|
|
217
|
-
{ name: 'textInput_cacheRead', rate: 0.025, strategy: 'fixed', unit: 'millionTokens' },
|
|
218
|
-
],
|
|
219
|
-
},
|
|
220
|
-
releasedAt: '2025-04-14',
|
|
221
|
-
type: 'chat',
|
|
222
|
-
},
|
|
223
|
-
{
|
|
224
|
-
abilities: {
|
|
225
|
-
functionCall: true,
|
|
226
|
-
vision: true,
|
|
227
|
-
},
|
|
228
|
-
contextWindowTokens: 128_000,
|
|
229
|
-
description:
|
|
230
|
-
'GPT-4o mini is OpenAI’s latest model after GPT-4 Omni, supporting text+image input with text output. It is their most advanced small model, far cheaper than recent frontier models and over 60% cheaper than GPT-3.5 Turbo, while maintaining top-tier intelligence (82% MMLU).',
|
|
231
|
-
displayName: 'GPT-4o mini',
|
|
232
|
-
id: 'gpt-4o-mini',
|
|
233
|
-
maxOutput: 16_385,
|
|
234
|
-
pricing: {
|
|
235
|
-
units: [
|
|
236
|
-
{ name: 'textInput', rate: 0.15, strategy: 'fixed', unit: 'millionTokens' },
|
|
237
|
-
{ name: 'textOutput', rate: 0.6, strategy: 'fixed', unit: 'millionTokens' },
|
|
238
|
-
{ name: 'textInput_cacheRead', rate: 0.075, strategy: 'fixed', unit: 'millionTokens' },
|
|
239
|
-
],
|
|
240
|
-
},
|
|
241
|
-
releasedAt: '2024-07-18',
|
|
242
|
-
type: 'chat',
|
|
243
|
-
},
|
|
244
|
-
{
|
|
245
|
-
abilities: {
|
|
246
|
-
functionCall: true,
|
|
247
|
-
vision: true,
|
|
248
|
-
},
|
|
249
|
-
contextWindowTokens: 128_000,
|
|
250
|
-
description:
|
|
251
|
-
'ChatGPT-4o is a dynamic model updated in real time. It combines strong understanding and generation for large-scale use cases like customer support, education, and technical support.',
|
|
252
|
-
displayName: 'GPT-4o',
|
|
253
|
-
id: 'gpt-4o',
|
|
254
|
-
pricing: {
|
|
255
|
-
units: [
|
|
256
|
-
{ name: 'textInput', rate: 2.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
257
|
-
{ name: 'textOutput', rate: 10, strategy: 'fixed', unit: 'millionTokens' },
|
|
258
|
-
{ name: 'textInput_cacheRead', rate: 1.25, strategy: 'fixed', unit: 'millionTokens' },
|
|
259
|
-
],
|
|
260
|
-
},
|
|
261
|
-
releasedAt: '2024-05-13',
|
|
262
|
-
type: 'chat',
|
|
263
|
-
},
|
|
264
|
-
{
|
|
265
|
-
abilities: {
|
|
266
|
-
vision: true,
|
|
267
|
-
},
|
|
268
|
-
contextWindowTokens: 128_000,
|
|
269
|
-
description:
|
|
270
|
-
'ChatGPT-4o is a dynamic model updated in real time. It combines strong understanding and generation for large-scale use cases like customer support, education, and technical support.',
|
|
271
|
-
displayName: 'ChatGPT-4o',
|
|
272
|
-
id: 'chatgpt-4o-latest',
|
|
273
|
-
pricing: {
|
|
274
|
-
units: [
|
|
275
|
-
{ name: 'textInput', rate: 5, strategy: 'fixed', unit: 'millionTokens' },
|
|
276
|
-
{ name: 'textOutput', rate: 15, strategy: 'fixed', unit: 'millionTokens' },
|
|
277
|
-
],
|
|
278
|
-
},
|
|
279
|
-
releasedAt: '2024-08-14',
|
|
280
|
-
type: 'chat',
|
|
281
|
-
},
|
|
282
|
-
{
|
|
283
|
-
abilities: {
|
|
284
|
-
functionCall: true,
|
|
285
|
-
vision: true,
|
|
286
|
-
},
|
|
287
|
-
contextWindowTokens: 128_000,
|
|
288
|
-
description:
|
|
289
|
-
'The latest GPT-4 Turbo adds vision. Visual requests support JSON mode and function calling. It is a cost-effective multimodal model that balances accuracy and efficiency for real-time applications.',
|
|
290
|
-
displayName: 'GPT-4 Turbo',
|
|
291
|
-
id: 'gpt-4-turbo',
|
|
292
|
-
pricing: {
|
|
293
|
-
units: [
|
|
294
|
-
{ name: 'textInput', rate: 10, strategy: 'fixed', unit: 'millionTokens' },
|
|
295
|
-
{ name: 'textOutput', rate: 30, strategy: 'fixed', unit: 'millionTokens' },
|
|
296
|
-
],
|
|
297
|
-
},
|
|
298
|
-
type: 'chat',
|
|
299
|
-
},
|
|
300
|
-
{
|
|
301
|
-
abilities: {
|
|
302
|
-
functionCall: true,
|
|
303
|
-
reasoning: true,
|
|
304
|
-
vision: true,
|
|
305
|
-
},
|
|
306
|
-
contextWindowTokens: 200_000,
|
|
307
|
-
description:
|
|
308
|
-
'o3 is a strong all-round model that sets a new bar for math, science, programming, and visual reasoning. It excels at technical writing and instruction following, and can analyze text, code, and images for multi-step problems.',
|
|
309
|
-
displayName: 'o3',
|
|
310
|
-
enabled: true,
|
|
311
|
-
id: 'o3',
|
|
312
|
-
maxOutput: 100_000,
|
|
313
|
-
pricing: {
|
|
314
|
-
units: [
|
|
315
|
-
{ name: 'textInput', rate: 2, strategy: 'fixed', unit: 'millionTokens' },
|
|
316
|
-
{ name: 'textOutput', rate: 8, strategy: 'fixed', unit: 'millionTokens' },
|
|
317
|
-
{ name: 'textInput_cacheRead', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
318
|
-
],
|
|
319
|
-
},
|
|
320
|
-
releasedAt: '2025-04-17',
|
|
321
|
-
settings: {
|
|
322
|
-
extendParams: ['reasoningEffort'],
|
|
323
|
-
},
|
|
324
|
-
type: 'chat',
|
|
325
|
-
},
|
|
326
|
-
{
|
|
327
|
-
abilities: {
|
|
328
|
-
functionCall: true,
|
|
329
|
-
reasoning: true,
|
|
330
|
-
vision: true,
|
|
331
|
-
},
|
|
332
|
-
contextWindowTokens: 200_000,
|
|
333
|
-
description:
|
|
334
|
-
'o4-mini is the latest small o-series model, optimized for fast, effective reasoning with high efficiency in coding and vision tasks.',
|
|
335
|
-
displayName: 'o4-mini',
|
|
336
|
-
enabled: true,
|
|
337
|
-
id: 'o4-mini',
|
|
338
|
-
maxOutput: 100_000,
|
|
339
|
-
pricing: {
|
|
340
|
-
units: [
|
|
341
|
-
{ name: 'textInput', rate: 1.1, strategy: 'fixed', unit: 'millionTokens' },
|
|
342
|
-
{ name: 'textOutput', rate: 4.4, strategy: 'fixed', unit: 'millionTokens' },
|
|
343
|
-
{ name: 'textInput_cacheRead', rate: 0.275, strategy: 'fixed', unit: 'millionTokens' },
|
|
344
|
-
],
|
|
345
|
-
},
|
|
346
|
-
releasedAt: '2025-04-17',
|
|
347
|
-
settings: {
|
|
348
|
-
extendParams: ['reasoningEffort'],
|
|
349
|
-
},
|
|
350
|
-
type: 'chat',
|
|
351
|
-
},
|
|
352
|
-
// {
|
|
353
|
-
// contextWindowTokens: 128_000,
|
|
354
|
-
// description:
|
|
355
|
-
// 'o1 is OpenAI’s new reasoning model for complex tasks requiring broad general knowledge. It has a 128K context window and an October 2023 knowledge cutoff.',
|
|
356
|
-
// displayName: 'OpenAI o1-preview',
|
|
357
|
-
// enabled: true,
|
|
358
|
-
// id: 'o1-preview',
|
|
359
|
-
// maxOutput: 32_768,
|
|
360
|
-
// pricing: {
|
|
361
|
-
// input: 15,
|
|
362
|
-
// output: 60,
|
|
363
|
-
// },
|
|
364
|
-
// releasedAt: '2024-09-12',
|
|
365
|
-
// type: 'chat',
|
|
366
|
-
// },
|
|
367
|
-
{
|
|
368
|
-
abilities: {
|
|
369
|
-
functionCall: true,
|
|
370
|
-
reasoning: true,
|
|
371
|
-
search: true,
|
|
372
|
-
vision: true,
|
|
373
|
-
},
|
|
374
|
-
contextWindowTokens: 200_000,
|
|
375
|
-
description: 'Claude Sonnet 4.5 is Anthropic’s most intelligent model to date.',
|
|
376
|
-
displayName: 'Claude Sonnet 4.5',
|
|
377
|
-
enabled: true,
|
|
378
|
-
id: 'claude-sonnet-4-5-20250929',
|
|
379
|
-
maxOutput: 8192,
|
|
380
|
-
pricing: {
|
|
381
|
-
units: [
|
|
382
|
-
{ name: 'textInput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
|
|
383
|
-
{ name: 'textOutput', rate: 15, strategy: 'fixed', unit: 'millionTokens' },
|
|
384
|
-
{ name: 'textInput_cacheRead', rate: 0.3, strategy: 'fixed', unit: 'millionTokens' },
|
|
385
|
-
{ name: 'textInput_cacheWrite', rate: 3.75, strategy: 'fixed', unit: 'millionTokens' },
|
|
386
|
-
],
|
|
387
|
-
},
|
|
388
|
-
releasedAt: '2025-09-30',
|
|
389
|
-
settings: {
|
|
390
|
-
extendParams: ['disableContextCaching', 'enableReasoning', 'reasoningBudgetToken'],
|
|
391
|
-
},
|
|
392
|
-
type: 'chat',
|
|
393
|
-
},
|
|
394
|
-
{
|
|
395
|
-
abilities: {
|
|
396
|
-
functionCall: true,
|
|
397
|
-
reasoning: true,
|
|
398
|
-
search: true,
|
|
399
|
-
vision: true,
|
|
400
|
-
},
|
|
401
|
-
contextWindowTokens: 200_000,
|
|
402
|
-
description:
|
|
403
|
-
'Claude Sonnet 4 is Anthropic’s most intelligent model to date, offering near-instant responses or extended step-by-step thinking with fine-grained control for API users.',
|
|
404
|
-
displayName: 'Claude Sonnet 4',
|
|
405
|
-
id: 'claude-sonnet-4-20250514',
|
|
406
|
-
maxOutput: 8192,
|
|
407
|
-
pricing: {
|
|
408
|
-
units: [
|
|
409
|
-
{ name: 'textInput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
|
|
410
|
-
{ name: 'textOutput', rate: 15, strategy: 'fixed', unit: 'millionTokens' },
|
|
411
|
-
{ name: 'textInput_cacheRead', rate: 0.3, strategy: 'fixed', unit: 'millionTokens' },
|
|
412
|
-
{
|
|
413
|
-
lookup: { prices: { '1h': 6, '5m': 3.75 }, pricingParams: ['ttl'] },
|
|
414
|
-
name: 'textInput_cacheWrite',
|
|
415
|
-
strategy: 'lookup',
|
|
416
|
-
unit: 'millionTokens',
|
|
417
|
-
},
|
|
418
|
-
],
|
|
419
|
-
},
|
|
420
|
-
releasedAt: '2025-05-23',
|
|
421
|
-
settings: {
|
|
422
|
-
extendParams: ['disableContextCaching', 'enableReasoning', 'reasoningBudgetToken'],
|
|
423
|
-
},
|
|
424
|
-
type: 'chat',
|
|
425
|
-
},
|
|
426
|
-
{
|
|
427
|
-
abilities: {
|
|
428
|
-
functionCall: true,
|
|
429
|
-
reasoning: true,
|
|
430
|
-
search: true,
|
|
431
|
-
vision: true,
|
|
432
|
-
},
|
|
433
|
-
contextWindowTokens: 200_000,
|
|
434
|
-
description:
|
|
435
|
-
'Claude Sonnet 3.7 is Anthropic’s most intelligent model and the first hybrid reasoning model on the market, supporting near-instant responses or extended thinking with fine-grained control.',
|
|
436
|
-
displayName: 'Claude Sonnet 3.7',
|
|
437
|
-
id: 'claude-3-7-sonnet-20250219',
|
|
438
|
-
maxOutput: 8192,
|
|
439
|
-
pricing: {
|
|
440
|
-
units: [
|
|
441
|
-
{ name: 'textInput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
|
|
442
|
-
{ name: 'textOutput', rate: 15, strategy: 'fixed', unit: 'millionTokens' },
|
|
443
|
-
{ name: 'textInput_cacheRead', rate: 0.3, strategy: 'fixed', unit: 'millionTokens' },
|
|
444
|
-
{
|
|
445
|
-
lookup: { prices: { '1h': 6, '5m': 3.75 }, pricingParams: ['ttl'] },
|
|
446
|
-
name: 'textInput_cacheWrite',
|
|
447
|
-
strategy: 'lookup',
|
|
448
|
-
unit: 'millionTokens',
|
|
449
|
-
},
|
|
450
|
-
],
|
|
451
|
-
},
|
|
452
|
-
settings: {
|
|
453
|
-
extendParams: ['disableContextCaching', 'enableReasoning', 'reasoningBudgetToken'],
|
|
454
|
-
},
|
|
455
|
-
type: 'chat',
|
|
456
|
-
},
|
|
457
|
-
{
|
|
458
|
-
abilities: {
|
|
459
|
-
functionCall: true,
|
|
460
|
-
reasoning: true,
|
|
461
|
-
search: true,
|
|
462
|
-
structuredOutput: true,
|
|
463
|
-
vision: true,
|
|
464
|
-
},
|
|
465
|
-
contextWindowTokens: 200_000,
|
|
466
|
-
description:
|
|
467
|
-
'Claude Opus 4.5 is Anthropic’s flagship model, combining excellent intelligence and scalable performance for the highest-quality reasoning tasks.',
|
|
468
|
-
displayName: 'Claude Opus 4.5',
|
|
469
|
-
enabled: true,
|
|
470
|
-
id: 'claude-opus-4-5-20251101',
|
|
471
|
-
maxOutput: 64_000,
|
|
472
|
-
pricing: {
|
|
473
|
-
units: [
|
|
474
|
-
{ name: 'textInput_cacheRead', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
475
|
-
{ name: 'textInput', rate: 5, strategy: 'fixed', unit: 'millionTokens' },
|
|
476
|
-
{ name: 'textOutput', rate: 25, strategy: 'fixed', unit: 'millionTokens' },
|
|
477
|
-
{
|
|
478
|
-
lookup: { prices: { '1h': 10, '5m': 6.25 }, pricingParams: ['ttl'] },
|
|
479
|
-
name: 'textInput_cacheWrite',
|
|
480
|
-
strategy: 'lookup',
|
|
481
|
-
unit: 'millionTokens',
|
|
482
|
-
},
|
|
483
|
-
],
|
|
484
|
-
},
|
|
485
|
-
releasedAt: '2025-11-24',
|
|
486
|
-
settings: {
|
|
487
|
-
extendParams: ['disableContextCaching', 'enableReasoning', 'reasoningBudgetToken'],
|
|
488
|
-
},
|
|
489
|
-
type: 'chat',
|
|
490
|
-
},
|
|
491
|
-
{
|
|
492
|
-
abilities: {
|
|
493
|
-
functionCall: true,
|
|
494
|
-
reasoning: true,
|
|
495
|
-
search: true,
|
|
496
|
-
vision: true,
|
|
497
|
-
},
|
|
498
|
-
contextWindowTokens: 200_000,
|
|
499
|
-
description:
|
|
500
|
-
'Claude Opus 4.1 is Anthropic’s latest and most capable model for highly complex tasks, excelling in performance, intelligence, fluency, and understanding.',
|
|
501
|
-
displayName: 'Claude Opus 4.1',
|
|
502
|
-
id: 'claude-opus-4-1-20250805',
|
|
503
|
-
maxOutput: 32_000,
|
|
504
|
-
pricing: {
|
|
505
|
-
units: [
|
|
506
|
-
{ name: 'textInput_cacheRead', rate: 1.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
507
|
-
{ name: 'textInput', rate: 15, strategy: 'fixed', unit: 'millionTokens' },
|
|
508
|
-
{ name: 'textOutput', rate: 75, strategy: 'fixed', unit: 'millionTokens' },
|
|
509
|
-
{
|
|
510
|
-
lookup: { prices: { '1h': 30, '5m': 18.75 }, pricingParams: ['ttl'] },
|
|
511
|
-
name: 'textInput_cacheWrite',
|
|
512
|
-
strategy: 'lookup',
|
|
513
|
-
unit: 'millionTokens',
|
|
514
|
-
},
|
|
515
|
-
],
|
|
516
|
-
},
|
|
517
|
-
releasedAt: '2025-08-05',
|
|
518
|
-
settings: {
|
|
519
|
-
extendParams: ['disableContextCaching', 'enableReasoning', 'reasoningBudgetToken'],
|
|
520
|
-
},
|
|
521
|
-
type: 'chat',
|
|
522
|
-
},
|
|
523
|
-
{
|
|
524
|
-
abilities: {
|
|
525
|
-
functionCall: true,
|
|
526
|
-
reasoning: true,
|
|
527
|
-
search: true,
|
|
528
|
-
vision: true,
|
|
529
|
-
},
|
|
530
|
-
contextWindowTokens: 200_000,
|
|
531
|
-
description:
|
|
532
|
-
'Claude Opus 4 is Anthropic’s most powerful model for highly complex tasks, excelling in performance, intelligence, fluency, and understanding.',
|
|
533
|
-
displayName: 'Claude Opus 4',
|
|
534
|
-
id: 'claude-opus-4-20250514',
|
|
535
|
-
maxOutput: 32_000,
|
|
536
|
-
pricing: {
|
|
537
|
-
units: [
|
|
538
|
-
{ name: 'textInput_cacheRead', rate: 1.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
539
|
-
{ name: 'textInput', rate: 15, strategy: 'fixed', unit: 'millionTokens' },
|
|
540
|
-
{ name: 'textOutput', rate: 75, strategy: 'fixed', unit: 'millionTokens' },
|
|
541
|
-
{
|
|
542
|
-
lookup: { prices: { '1h': 30, '5m': 18.75 }, pricingParams: ['ttl'] },
|
|
543
|
-
name: 'textInput_cacheWrite',
|
|
544
|
-
strategy: 'lookup',
|
|
545
|
-
unit: 'millionTokens',
|
|
546
|
-
},
|
|
547
|
-
],
|
|
548
|
-
},
|
|
549
|
-
releasedAt: '2025-05-23',
|
|
550
|
-
settings: {
|
|
551
|
-
extendParams: ['disableContextCaching', 'enableReasoning', 'reasoningBudgetToken'],
|
|
552
|
-
},
|
|
553
|
-
type: 'chat',
|
|
554
|
-
},
|
|
555
|
-
|
|
556
|
-
{
|
|
557
|
-
abilities: {
|
|
558
|
-
functionCall: true,
|
|
559
|
-
reasoning: true,
|
|
560
|
-
search: true,
|
|
561
|
-
structuredOutput: true,
|
|
562
|
-
vision: true,
|
|
563
|
-
},
|
|
564
|
-
contextWindowTokens: 200_000,
|
|
565
|
-
description:
|
|
566
|
-
'Claude Haiku 4.5 is Anthropic’s fastest and most intelligent Haiku model, with lightning speed and extended thinking.',
|
|
567
|
-
displayName: 'Claude Haiku 4.5',
|
|
568
|
-
enabled: true,
|
|
569
|
-
id: 'claude-haiku-4-5-20251001',
|
|
570
|
-
maxOutput: 64_000,
|
|
571
|
-
pricing: {
|
|
572
|
-
units: [
|
|
573
|
-
{ name: 'textInput', rate: 1, strategy: 'fixed', unit: 'millionTokens' },
|
|
574
|
-
{ name: 'textOutput', rate: 5, strategy: 'fixed', unit: 'millionTokens' },
|
|
575
|
-
{ name: 'textInput_cacheRead', rate: 0.1, strategy: 'fixed', unit: 'millionTokens' },
|
|
576
|
-
{
|
|
577
|
-
lookup: { prices: { '1h': 2, '5m': 1.25 }, pricingParams: ['ttl'] },
|
|
578
|
-
name: 'textInput_cacheWrite',
|
|
579
|
-
strategy: 'lookup',
|
|
580
|
-
unit: 'millionTokens',
|
|
581
|
-
},
|
|
582
|
-
],
|
|
583
|
-
},
|
|
584
|
-
releasedAt: '2025-10-16',
|
|
585
|
-
settings: {
|
|
586
|
-
extendParams: ['disableContextCaching', 'enableReasoning', 'reasoningBudgetToken'],
|
|
587
|
-
},
|
|
588
|
-
type: 'chat',
|
|
589
|
-
},
|
|
590
|
-
{
|
|
591
|
-
abilities: {
|
|
592
|
-
functionCall: true,
|
|
593
|
-
vision: true,
|
|
594
|
-
},
|
|
595
|
-
contextWindowTokens: 200_000,
|
|
596
|
-
description:
|
|
597
|
-
'Claude 3.5 Haiku is Anthropic’s fastest next-gen model, improving across skills and surpassing the previous flagship Claude 3 Opus on many benchmarks.',
|
|
598
|
-
displayName: 'Claude 3.5 Haiku',
|
|
599
|
-
id: 'claude-3-5-haiku-20241022',
|
|
600
|
-
maxOutput: 8192,
|
|
601
|
-
pricing: {
|
|
602
|
-
units: [
|
|
603
|
-
{ name: 'textInput_cacheRead', rate: 0.08, strategy: 'fixed', unit: 'millionTokens' },
|
|
604
|
-
{ name: 'textInput', rate: 0.8, strategy: 'fixed', unit: 'millionTokens' },
|
|
605
|
-
{ name: 'textOutput', rate: 4, strategy: 'fixed', unit: 'millionTokens' },
|
|
606
|
-
{
|
|
607
|
-
lookup: { prices: { '1h': 1.6, '5m': 1 }, pricingParams: ['ttl'] },
|
|
608
|
-
name: 'textInput_cacheWrite',
|
|
609
|
-
strategy: 'lookup',
|
|
610
|
-
unit: 'millionTokens',
|
|
611
|
-
},
|
|
612
|
-
],
|
|
613
|
-
},
|
|
614
|
-
releasedAt: '2024-11-05',
|
|
615
|
-
settings: {
|
|
616
|
-
extendParams: ['disableContextCaching'],
|
|
617
|
-
},
|
|
618
|
-
type: 'chat',
|
|
619
|
-
},
|
|
620
|
-
{
|
|
621
|
-
abilities: {
|
|
622
|
-
functionCall: true,
|
|
623
|
-
reasoning: true,
|
|
624
|
-
search: true,
|
|
625
|
-
video: true,
|
|
626
|
-
vision: true,
|
|
627
|
-
},
|
|
628
|
-
contextWindowTokens: 1_048_576 + 65_536,
|
|
629
|
-
description:
|
|
630
|
-
'Gemini 3 Pro is Google’s most powerful agent and vibe-coding model, delivering richer visuals and deeper interaction on top of state-of-the-art reasoning.',
|
|
631
|
-
displayName: 'Gemini 3 Pro Preview',
|
|
632
|
-
enabled: true,
|
|
633
|
-
id: 'gemini-3-pro-preview',
|
|
634
|
-
maxOutput: 65_536,
|
|
635
|
-
pricing: {
|
|
636
|
-
units: [
|
|
637
|
-
{
|
|
638
|
-
name: 'textInput_cacheRead',
|
|
639
|
-
strategy: 'tiered',
|
|
640
|
-
tiers: [
|
|
641
|
-
{ rate: 0.2, upTo: 200_000 },
|
|
642
|
-
{ rate: 0.4, upTo: 'infinity' },
|
|
643
|
-
],
|
|
644
|
-
unit: 'millionTokens',
|
|
645
|
-
},
|
|
646
|
-
{
|
|
647
|
-
name: 'textInput',
|
|
648
|
-
strategy: 'tiered',
|
|
649
|
-
tiers: [
|
|
650
|
-
{ rate: 2, upTo: 200_000 },
|
|
651
|
-
{ rate: 4, upTo: 'infinity' },
|
|
652
|
-
],
|
|
653
|
-
unit: 'millionTokens',
|
|
654
|
-
},
|
|
655
|
-
{
|
|
656
|
-
name: 'textOutput',
|
|
657
|
-
strategy: 'tiered',
|
|
658
|
-
tiers: [
|
|
659
|
-
{ rate: 12, upTo: 200_000 },
|
|
660
|
-
{ rate: 18, upTo: 'infinity' },
|
|
661
|
-
],
|
|
662
|
-
unit: 'millionTokens',
|
|
663
|
-
},
|
|
664
|
-
{
|
|
665
|
-
lookup: { prices: { '1h': 4.5 }, pricingParams: ['ttl'] },
|
|
666
|
-
name: 'textInput_cacheWrite',
|
|
667
|
-
strategy: 'lookup',
|
|
668
|
-
unit: 'millionTokens',
|
|
669
|
-
},
|
|
670
|
-
],
|
|
671
|
-
},
|
|
672
|
-
releasedAt: '2025-11-18',
|
|
673
|
-
settings: {
|
|
674
|
-
extendParams: ['thinkingLevel2', 'urlContext'],
|
|
675
|
-
searchImpl: 'params',
|
|
676
|
-
searchProvider: 'google',
|
|
677
|
-
},
|
|
678
|
-
type: 'chat',
|
|
679
|
-
},
|
|
680
|
-
{
|
|
681
|
-
abilities: {
|
|
682
|
-
functionCall: true,
|
|
683
|
-
reasoning: true,
|
|
684
|
-
search: true,
|
|
685
|
-
video: true,
|
|
686
|
-
vision: true,
|
|
687
|
-
},
|
|
688
|
-
contextWindowTokens: 1_048_576 + 65_536,
|
|
689
|
-
description:
|
|
690
|
-
'Gemini 3 Flash Preview is Google’s latest best-value model, improving on Gemini 2.5 Flash.',
|
|
691
|
-
displayName: 'Gemini 3 Flash Preview',
|
|
692
|
-
enabled: true,
|
|
693
|
-
id: 'gemini-3-flash-preview',
|
|
694
|
-
maxOutput: 65_536,
|
|
695
|
-
pricing: {
|
|
696
|
-
units: [
|
|
697
|
-
{
|
|
698
|
-
name: 'textInput',
|
|
699
|
-
rate: 0.5,
|
|
700
|
-
strategy: 'fixed',
|
|
701
|
-
unit: 'millionTokens',
|
|
702
|
-
},
|
|
703
|
-
{
|
|
704
|
-
name: 'textInput_cacheRead',
|
|
705
|
-
rate: 0.05,
|
|
706
|
-
strategy: 'fixed',
|
|
707
|
-
unit: 'millionTokens',
|
|
708
|
-
},
|
|
709
|
-
{ name: 'textOutput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
|
|
710
|
-
],
|
|
711
|
-
},
|
|
712
|
-
releasedAt: '2025-12-17',
|
|
713
|
-
settings: {
|
|
714
|
-
extendParams: ['thinkingLevel', 'urlContext'],
|
|
715
|
-
searchImpl: 'params',
|
|
716
|
-
searchProvider: 'google',
|
|
717
|
-
},
|
|
718
|
-
type: 'chat',
|
|
719
|
-
},
|
|
720
|
-
{
|
|
721
|
-
abilities: {
|
|
722
|
-
functionCall: true,
|
|
723
|
-
reasoning: true,
|
|
724
|
-
search: true,
|
|
725
|
-
video: true,
|
|
726
|
-
vision: true,
|
|
727
|
-
},
|
|
728
|
-
contextWindowTokens: 1_048_576 + 65_536,
|
|
729
|
-
description:
|
|
730
|
-
'Gemini 2.5 Pro is Google’s most advanced reasoning model, able to reason over code, math, and STEM problems and analyze large datasets, codebases, and documents with long context.',
|
|
731
|
-
displayName: 'Gemini 2.5 Pro',
|
|
732
|
-
enabled: true,
|
|
733
|
-
id: 'gemini-2.5-pro',
|
|
734
|
-
maxOutput: 65_536,
|
|
735
|
-
pricing: {
|
|
736
|
-
units: [
|
|
737
|
-
{
|
|
738
|
-
name: 'textInput_cacheRead',
|
|
739
|
-
strategy: 'tiered',
|
|
740
|
-
tiers: [
|
|
741
|
-
{ rate: 0.31, upTo: 200_000 },
|
|
742
|
-
{ rate: 0.625, upTo: 'infinity' },
|
|
743
|
-
],
|
|
744
|
-
unit: 'millionTokens',
|
|
745
|
-
},
|
|
746
|
-
{
|
|
747
|
-
name: 'textInput',
|
|
748
|
-
strategy: 'tiered',
|
|
749
|
-
tiers: [
|
|
750
|
-
{ rate: 1.25, upTo: 200_000 },
|
|
751
|
-
{ rate: 2.5, upTo: 'infinity' },
|
|
752
|
-
],
|
|
753
|
-
unit: 'millionTokens',
|
|
754
|
-
},
|
|
755
|
-
{
|
|
756
|
-
name: 'textOutput',
|
|
757
|
-
strategy: 'tiered',
|
|
758
|
-
tiers: [
|
|
759
|
-
{ rate: 10, upTo: 200_000 },
|
|
760
|
-
{ rate: 15, upTo: 'infinity' },
|
|
761
|
-
],
|
|
762
|
-
unit: 'millionTokens',
|
|
763
|
-
},
|
|
764
|
-
],
|
|
765
|
-
},
|
|
766
|
-
releasedAt: '2025-06-17',
|
|
767
|
-
settings: {
|
|
768
|
-
searchImpl: 'params',
|
|
769
|
-
searchProvider: 'google',
|
|
770
|
-
},
|
|
771
|
-
type: 'chat',
|
|
772
|
-
},
|
|
773
|
-
{
|
|
774
|
-
abilities: {
|
|
775
|
-
functionCall: true,
|
|
776
|
-
reasoning: true,
|
|
777
|
-
search: true,
|
|
778
|
-
video: true,
|
|
779
|
-
vision: true,
|
|
780
|
-
},
|
|
781
|
-
contextWindowTokens: 1_048_576 + 65_536,
|
|
782
|
-
description: 'Gemini 2.5 Flash is Google’s best-value model with full capabilities.',
|
|
783
|
-
displayName: 'Gemini 2.5 Flash',
|
|
784
|
-
enabled: true,
|
|
785
|
-
id: 'gemini-2.5-flash',
|
|
786
|
-
maxOutput: 65_536,
|
|
787
|
-
pricing: {
|
|
788
|
-
units: [
|
|
789
|
-
{
|
|
790
|
-
name: 'textInput_cacheRead',
|
|
791
|
-
rate: 0.03,
|
|
792
|
-
strategy: 'fixed',
|
|
793
|
-
unit: 'millionTokens',
|
|
794
|
-
},
|
|
795
|
-
{
|
|
796
|
-
name: 'textInput',
|
|
797
|
-
rate: 0.3,
|
|
798
|
-
strategy: 'fixed',
|
|
799
|
-
unit: 'millionTokens',
|
|
800
|
-
},
|
|
801
|
-
{ name: 'textOutput', rate: 2.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
802
|
-
],
|
|
803
|
-
},
|
|
804
|
-
releasedAt: '2025-06-17',
|
|
805
|
-
settings: {
|
|
806
|
-
extendParams: ['enableReasoning', 'reasoningBudgetToken'],
|
|
807
|
-
searchImpl: 'params',
|
|
808
|
-
searchProvider: 'google',
|
|
809
|
-
},
|
|
810
|
-
type: 'chat',
|
|
811
|
-
},
|
|
812
|
-
{
|
|
813
|
-
abilities: {
|
|
814
|
-
imageOutput: true,
|
|
815
|
-
reasoning: true,
|
|
816
|
-
search: true,
|
|
817
|
-
vision: true,
|
|
818
|
-
},
|
|
819
|
-
contextWindowTokens: 131_072 + 32_768,
|
|
820
|
-
description:
|
|
821
|
-
'Gemini 3 Pro Image (Nano Banana Pro) is Google’s image generation model and also supports multimodal chat.',
|
|
822
|
-
displayName: 'Nano Banana Pro',
|
|
823
|
-
enabled: true,
|
|
824
|
-
id: 'gemini-3-pro-image-preview',
|
|
825
|
-
maxOutput: 32_768,
|
|
826
|
-
pricing: {
|
|
827
|
-
approximatePricePerImage: 0.134,
|
|
828
|
-
units: [
|
|
829
|
-
{ name: 'imageOutput', rate: 120, strategy: 'fixed', unit: 'millionTokens' },
|
|
830
|
-
{ name: 'textInput', rate: 2, strategy: 'fixed', unit: 'millionTokens' },
|
|
831
|
-
{ name: 'textOutput', rate: 12, strategy: 'fixed', unit: 'millionTokens' },
|
|
832
|
-
],
|
|
833
|
-
},
|
|
834
|
-
settings: {
|
|
835
|
-
searchImpl: 'params',
|
|
836
|
-
searchProvider: 'google',
|
|
837
|
-
},
|
|
838
|
-
type: 'chat',
|
|
839
|
-
},
|
|
840
|
-
{
|
|
841
|
-
abilities: {
|
|
842
|
-
imageOutput: true,
|
|
843
|
-
vision: true,
|
|
844
|
-
},
|
|
845
|
-
contextWindowTokens: 32_768 + 32_768,
|
|
846
|
-
description:
|
|
847
|
-
'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
|
|
848
|
-
displayName: 'Nano Banana',
|
|
849
|
-
enabled: true,
|
|
850
|
-
id: 'gemini-2.5-flash-image-preview',
|
|
851
|
-
maxOutput: 32_768,
|
|
852
|
-
pricing: {
|
|
853
|
-
units: [
|
|
854
|
-
{ name: 'textInput', rate: 0.3, strategy: 'fixed', unit: 'millionTokens' },
|
|
855
|
-
{ name: 'imageInput', rate: 0.3, strategy: 'fixed', unit: 'millionTokens' },
|
|
856
|
-
{ name: 'textOutput', rate: 2.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
857
|
-
{ name: 'imageOutput', rate: 30, strategy: 'fixed', unit: 'millionTokens' },
|
|
858
|
-
],
|
|
859
|
-
},
|
|
860
|
-
releasedAt: '2025-08-26',
|
|
861
|
-
type: 'chat',
|
|
862
|
-
},
|
|
863
|
-
{
|
|
864
|
-
abilities: {
|
|
865
|
-
functionCall: true,
|
|
866
|
-
reasoning: true,
|
|
867
|
-
search: true,
|
|
868
|
-
vision: true,
|
|
869
|
-
},
|
|
870
|
-
contextWindowTokens: 256_000,
|
|
871
|
-
description:
|
|
872
|
-
'Our newest and strongest flagship model, excelling in NLP, math, and reasoning—an ideal all-rounder.',
|
|
873
|
-
displayName: 'Grok 4',
|
|
874
|
-
enabled: true,
|
|
875
|
-
id: 'grok-4',
|
|
876
|
-
pricing: {
|
|
877
|
-
units: [
|
|
878
|
-
{ name: 'textInput_cacheRead', rate: 0.75, strategy: 'fixed', unit: 'millionTokens' },
|
|
879
|
-
{ name: 'textInput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
|
|
880
|
-
{ name: 'textOutput', rate: 15, strategy: 'fixed', unit: 'millionTokens' },
|
|
881
|
-
],
|
|
882
|
-
},
|
|
883
|
-
releasedAt: '2025-07-09',
|
|
884
|
-
settings: {
|
|
885
|
-
// reasoning_effort is not supported by grok-4. Specifying reasoning_effort parameter will get an error response.
|
|
886
|
-
// extendParams: ['reasoningEffort'],
|
|
887
|
-
searchImpl: 'params',
|
|
888
|
-
},
|
|
889
|
-
type: 'chat',
|
|
890
|
-
},
|
|
891
|
-
{
|
|
892
|
-
abilities: {
|
|
893
|
-
imageOutput: true,
|
|
894
|
-
vision: true,
|
|
895
|
-
},
|
|
896
|
-
contextWindowTokens: 32_768,
|
|
897
|
-
description: 'Gemini 2.0 Flash experimental model with image generation support.',
|
|
898
|
-
displayName: 'Gemini 2.0 Flash (Image Generation) Experimental',
|
|
899
|
-
enabled: true,
|
|
900
|
-
id: 'gemini-2.0-flash-exp-image-generation',
|
|
901
|
-
maxOutput: 8192,
|
|
902
|
-
pricing: {
|
|
903
|
-
units: [
|
|
904
|
-
{ name: 'textInput', rate: 0.1, strategy: 'fixed', unit: 'millionTokens' },
|
|
905
|
-
{ name: 'imageInput', rate: 0.1, strategy: 'fixed', unit: 'millionTokens' },
|
|
906
|
-
{ name: 'audioInput', rate: 0.7, strategy: 'fixed', unit: 'millionTokens' },
|
|
907
|
-
{ name: 'textOutput', rate: 0.4, strategy: 'fixed', unit: 'millionTokens' },
|
|
908
|
-
{ name: 'imageOutput', rate: 30, strategy: 'fixed', unit: 'millionTokens' },
|
|
909
|
-
],
|
|
910
|
-
},
|
|
911
|
-
releasedAt: '2025-03-14',
|
|
912
|
-
type: 'chat',
|
|
913
|
-
},
|
|
914
|
-
{
|
|
915
|
-
abilities: {
|
|
916
|
-
functionCall: true,
|
|
917
|
-
},
|
|
918
|
-
contextWindowTokens: 65_536,
|
|
919
|
-
description:
|
|
920
|
-
'DeepSeek V3.2 balances reasoning and output length for daily QA and agent tasks. Public benchmarks reach GPT-5 levels, and it is the first to integrate thinking into tool use, leading open-source agent evaluations.',
|
|
921
|
-
displayName: 'DeepSeek V3.2',
|
|
922
|
-
enabled: true,
|
|
923
|
-
id: 'deepseek-chat',
|
|
924
|
-
pricing: {
|
|
925
|
-
units: [
|
|
926
|
-
{ name: 'textInput', rate: 0.56, strategy: 'fixed', unit: 'millionTokens' },
|
|
927
|
-
{ name: 'textInput_cacheRead', rate: 0.07, strategy: 'fixed', unit: 'millionTokens' },
|
|
928
|
-
{ name: 'textOutput', rate: 1.68, strategy: 'fixed', unit: 'millionTokens' },
|
|
929
|
-
],
|
|
930
|
-
},
|
|
931
|
-
releasedAt: '2025-12-01',
|
|
932
|
-
type: 'chat',
|
|
933
|
-
},
|
|
934
|
-
{
|
|
935
|
-
abilities: {
|
|
936
|
-
functionCall: true,
|
|
937
|
-
reasoning: true,
|
|
938
|
-
},
|
|
939
|
-
contextWindowTokens: 65_536,
|
|
940
|
-
description:
|
|
941
|
-
'DeepSeek V3.2 Thinking is a deep reasoning model that generates chain-of-thought before outputs for higher accuracy, with top competition results and reasoning comparable to Gemini-3.0-Pro.',
|
|
942
|
-
displayName: 'DeepSeek V3.2 Thinking',
|
|
943
|
-
enabled: true,
|
|
944
|
-
id: 'deepseek-reasoner',
|
|
945
|
-
pricing: {
|
|
946
|
-
units: [
|
|
947
|
-
{ name: 'textInput', rate: 0.55, strategy: 'fixed', unit: 'millionTokens' },
|
|
948
|
-
{ name: 'textOutput', rate: 2.19, strategy: 'fixed', unit: 'millionTokens' },
|
|
949
|
-
{ name: 'textInput_cacheRead', rate: 0.14, strategy: 'fixed', unit: 'millionTokens' },
|
|
950
|
-
],
|
|
951
|
-
},
|
|
952
|
-
releasedAt: '2025-12-01',
|
|
953
|
-
type: 'chat',
|
|
954
|
-
},
|
|
955
|
-
{
|
|
956
|
-
abilities: {
|
|
957
|
-
functionCall: true,
|
|
958
|
-
reasoning: true,
|
|
959
|
-
search: true,
|
|
960
|
-
},
|
|
961
|
-
contextWindowTokens: 204_800,
|
|
962
|
-
description: 'Built for efficient coding and agent workflows.',
|
|
963
|
-
displayName: 'MiniMax M2',
|
|
964
|
-
enabled: true,
|
|
965
|
-
id: 'MiniMax-M2-Stable',
|
|
966
|
-
maxOutput: 131_072,
|
|
967
|
-
pricing: {
|
|
968
|
-
units: [
|
|
969
|
-
{ name: 'textInput', rate: 0.3, strategy: 'fixed', unit: 'millionTokens' },
|
|
970
|
-
{
|
|
971
|
-
name: 'textInput_cacheWrite',
|
|
972
|
-
rate: 0.375,
|
|
973
|
-
strategy: 'fixed',
|
|
974
|
-
unit: 'millionTokens',
|
|
975
|
-
},
|
|
976
|
-
{
|
|
977
|
-
name: 'textInput_cacheRead',
|
|
978
|
-
rate: 0.03,
|
|
979
|
-
strategy: 'fixed',
|
|
980
|
-
unit: 'millionTokens',
|
|
981
|
-
},
|
|
982
|
-
|
|
983
|
-
{ name: 'textOutput', rate: 1.2, strategy: 'fixed', unit: 'millionTokens' },
|
|
984
|
-
],
|
|
985
|
-
},
|
|
986
|
-
releasedAt: '2025-10-27',
|
|
987
|
-
settings: {
|
|
988
|
-
searchImpl: 'params',
|
|
989
|
-
},
|
|
990
|
-
type: 'chat',
|
|
991
|
-
},
|
|
992
|
-
{
|
|
993
|
-
abilities: {
|
|
994
|
-
functionCall: true,
|
|
995
|
-
search: true,
|
|
996
|
-
},
|
|
997
|
-
contextWindowTokens: 131_072,
|
|
998
|
-
description:
|
|
999
|
-
'kimi-k2 is an MoE foundation model with strong coding and agent capabilities (1T total params, 32B active), outperforming other mainstream open models across reasoning, programming, math, and agent benchmarks.',
|
|
1000
|
-
displayName: 'Kimi K2',
|
|
1001
|
-
enabled: true,
|
|
1002
|
-
id: 'kimi-k2-0711-preview',
|
|
1003
|
-
pricing: {
|
|
1004
|
-
units: [
|
|
1005
|
-
{ name: 'textInput', rate: 0.6, strategy: 'fixed', unit: 'millionTokens' },
|
|
1006
|
-
{ name: 'textOutput', rate: 2.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
1007
|
-
{ name: 'textInput_cacheRead', rate: 0.15, strategy: 'fixed', unit: 'millionTokens' },
|
|
1008
|
-
],
|
|
1009
|
-
},
|
|
1010
|
-
releasedAt: '2025-07-11',
|
|
1011
|
-
settings: {
|
|
1012
|
-
searchImpl: 'params',
|
|
1013
|
-
},
|
|
1014
|
-
type: 'chat',
|
|
1015
|
-
},
|
|
1016
|
-
];
|
|
1017
|
-
|
|
1018
|
-
// Common parameters for Imagen models
|
|
1019
|
-
const imagenBaseParameters: ModelParamsSchema = {
|
|
1020
|
-
aspectRatio: {
|
|
1021
|
-
default: '1:1',
|
|
1022
|
-
enum: ['1:1', '16:9', '9:16', '3:4', '4:3'],
|
|
1023
|
-
},
|
|
1024
|
-
prompt: { default: '' },
|
|
1025
|
-
};
|
|
1026
|
-
|
|
1027
|
-
const NANO_BANANA_ASPECT_RATIOS = [
|
|
1028
|
-
'1:1', // 1024x1024 / 2048x2048 / 4096x4096
|
|
1029
|
-
'2:3', // 848x1264 / 1696x2528 / 3392x5056
|
|
1030
|
-
'3:2', // 1264x848 / 2528x1696 / 5056x3392
|
|
1031
|
-
'3:4', // 896x1200 / 1792x2400 / 3584x4800
|
|
1032
|
-
'4:3', // 1200x896 / 2400x1792 / 4800x3584
|
|
1033
|
-
'4:5', // 928x1152 / 1856x2304 / 3712x4608
|
|
1034
|
-
'5:4', // 1152x928 / 2304x1856 / 4608x3712
|
|
1035
|
-
'9:16', // 768x1376 / 1536x2752 / 3072x5504
|
|
1036
|
-
'16:9', // 1376x768 / 2752x1536 / 5504x3072
|
|
1037
|
-
'21:9', // 1584x672 / 3168x1344 / 6336x2688
|
|
1038
|
-
];
|
|
1039
|
-
|
|
1040
|
-
export const nanoBananaParameters: ModelParamsSchema = {
|
|
1041
|
-
aspectRatio: {
|
|
1042
|
-
default: '1:1',
|
|
1043
|
-
enum: NANO_BANANA_ASPECT_RATIOS,
|
|
1044
|
-
},
|
|
1045
|
-
imageUrls: {
|
|
1046
|
-
default: [],
|
|
1047
|
-
},
|
|
1048
|
-
prompt: { default: '' },
|
|
1049
|
-
};
|
|
1050
|
-
|
|
1051
|
-
export const nanoBananaProParameters: ModelParamsSchema = {
|
|
1052
|
-
aspectRatio: {
|
|
1053
|
-
default: '1:1',
|
|
1054
|
-
enum: NANO_BANANA_ASPECT_RATIOS,
|
|
1055
|
-
},
|
|
1056
|
-
imageUrls: {
|
|
1057
|
-
default: [],
|
|
1058
|
-
},
|
|
1059
|
-
prompt: { default: '' },
|
|
1060
|
-
resolution: {
|
|
1061
|
-
default: '1K',
|
|
1062
|
-
enum: ['1K', '2K', '4K'],
|
|
1063
|
-
},
|
|
1064
|
-
};
|
|
1065
|
-
|
|
1066
|
-
const gptImage1Schema = {
|
|
1067
|
-
imageUrls: { default: [], maxCount: 1, maxFileSize: 5 },
|
|
1068
|
-
prompt: { default: '' },
|
|
1069
|
-
size: {
|
|
1070
|
-
default: 'auto',
|
|
1071
|
-
enum: ['auto', '1024x1024', '1536x1024', '1024x1536'],
|
|
1072
|
-
},
|
|
1073
|
-
};
|
|
1074
|
-
const lobehubImageModels: AIImageModelCard[] = [
|
|
1075
|
-
{
|
|
1076
|
-
description:
|
|
1077
|
-
'Gemini 3 Pro Image (Nano Banana Pro) is Google’s image generation model and also supports multimodal chat.',
|
|
1078
|
-
displayName: 'Nano Banana Pro',
|
|
1079
|
-
enabled: true,
|
|
1080
|
-
id: 'gemini-3-pro-image-preview:image',
|
|
1081
|
-
parameters: nanoBananaProParameters,
|
|
1082
|
-
pricing: {
|
|
1083
|
-
approximatePricePerImage: 0.134,
|
|
1084
|
-
units: [
|
|
1085
|
-
{ name: 'imageOutput', rate: 120, strategy: 'fixed', unit: 'millionTokens' },
|
|
1086
|
-
{ name: 'textInput', rate: 2, strategy: 'fixed', unit: 'millionTokens' },
|
|
1087
|
-
{ name: 'textOutput', rate: 12, strategy: 'fixed', unit: 'millionTokens' },
|
|
1088
|
-
],
|
|
1089
|
-
},
|
|
1090
|
-
releasedAt: '2025-11-18',
|
|
1091
|
-
type: 'image',
|
|
1092
|
-
},
|
|
1093
|
-
{
|
|
1094
|
-
description:
|
|
1095
|
-
'Nano Banana is Google’s newest, fastest, and most efficient native multimodal model, enabling conversational image generation and editing.',
|
|
1096
|
-
displayName: 'Nano Banana',
|
|
1097
|
-
enabled: true,
|
|
1098
|
-
id: 'gemini-2.5-flash-image-preview:image',
|
|
1099
|
-
parameters: nanoBananaParameters,
|
|
1100
|
-
pricing: {
|
|
1101
|
-
approximatePricePerImage: 0.039,
|
|
1102
|
-
units: [
|
|
1103
|
-
{ name: 'textInput', rate: 0.3, strategy: 'fixed', unit: 'millionTokens' },
|
|
1104
|
-
{ name: 'imageInput', rate: 0.3, strategy: 'fixed', unit: 'millionTokens' },
|
|
1105
|
-
{ name: 'textOutput', rate: 2.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
1106
|
-
{ name: 'imageOutput', rate: 30, strategy: 'fixed', unit: 'millionTokens' },
|
|
1107
|
-
],
|
|
1108
|
-
},
|
|
1109
|
-
releasedAt: '2025-08-26',
|
|
1110
|
-
type: 'image',
|
|
1111
|
-
},
|
|
1112
|
-
{
|
|
1113
|
-
description: 'Imagen 4th generation text-to-image model series',
|
|
1114
|
-
displayName: 'Imagen 4 Fast',
|
|
1115
|
-
enabled: true,
|
|
1116
|
-
id: 'imagen-4.0-fast-generate-001',
|
|
1117
|
-
organization: 'Deepmind',
|
|
1118
|
-
parameters: imagenBaseParameters,
|
|
1119
|
-
pricing: {
|
|
1120
|
-
units: [{ name: 'imageGeneration', rate: 0.02, strategy: 'fixed', unit: 'image' }],
|
|
1121
|
-
},
|
|
1122
|
-
releasedAt: '2025-08-15',
|
|
1123
|
-
type: 'image',
|
|
1124
|
-
},
|
|
1125
|
-
{
|
|
1126
|
-
description: 'Imagen 4th generation text-to-image model series',
|
|
1127
|
-
displayName: 'Imagen 4',
|
|
1128
|
-
enabled: true,
|
|
1129
|
-
id: 'imagen-4.0-generate-001',
|
|
1130
|
-
organization: 'Deepmind',
|
|
1131
|
-
parameters: imagenBaseParameters,
|
|
1132
|
-
pricing: {
|
|
1133
|
-
units: [{ name: 'imageGeneration', rate: 0.04, strategy: 'fixed', unit: 'image' }],
|
|
1134
|
-
},
|
|
1135
|
-
releasedAt: '2025-08-15',
|
|
1136
|
-
type: 'image',
|
|
1137
|
-
},
|
|
1138
|
-
{
|
|
1139
|
-
description: 'Imagen 4th generation text-to-image model series Ultra version',
|
|
1140
|
-
displayName: 'Imagen 4 Ultra',
|
|
1141
|
-
enabled: true,
|
|
1142
|
-
id: 'imagen-4.0-ultra-generate-001',
|
|
1143
|
-
organization: 'Deepmind',
|
|
1144
|
-
parameters: imagenBaseParameters,
|
|
1145
|
-
pricing: {
|
|
1146
|
-
units: [{ name: 'imageGeneration', rate: 0.06, strategy: 'fixed', unit: 'image' }],
|
|
1147
|
-
},
|
|
1148
|
-
releasedAt: '2025-08-15',
|
|
1149
|
-
type: 'image',
|
|
1150
|
-
},
|
|
1151
|
-
{
|
|
1152
|
-
description:
|
|
1153
|
-
'An enhanced GPT Image 1 model with 4× faster generation, more precise editing, and improved text rendering.',
|
|
1154
|
-
displayName: 'GPT Image 1.5',
|
|
1155
|
-
enabled: true,
|
|
1156
|
-
id: 'gpt-image-1.5',
|
|
1157
|
-
parameters: gptImage1Schema,
|
|
1158
|
-
pricing: {
|
|
1159
|
-
approximatePricePerImage: 0.034,
|
|
1160
|
-
units: [
|
|
1161
|
-
{ name: 'textInput', rate: 5, strategy: 'fixed', unit: 'millionTokens' },
|
|
1162
|
-
{ name: 'textInput_cacheRead', rate: 1.25, strategy: 'fixed', unit: 'millionTokens' },
|
|
1163
|
-
{ name: 'imageInput', rate: 8, strategy: 'fixed', unit: 'millionTokens' },
|
|
1164
|
-
{ name: 'imageInput_cacheRead', rate: 2, strategy: 'fixed', unit: 'millionTokens' },
|
|
1165
|
-
{ name: 'imageOutput', rate: 32, strategy: 'fixed', unit: 'millionTokens' },
|
|
1166
|
-
],
|
|
1167
|
-
},
|
|
1168
|
-
releasedAt: '2025-12-16',
|
|
1169
|
-
type: 'image',
|
|
1170
|
-
},
|
|
1171
|
-
{
|
|
1172
|
-
description: 'ChatGPT native multimodal image generation model.',
|
|
1173
|
-
displayName: 'GPT Image 1',
|
|
1174
|
-
enabled: true,
|
|
1175
|
-
id: 'gpt-image-1',
|
|
1176
|
-
parameters: gptImage1Schema,
|
|
1177
|
-
pricing: {
|
|
1178
|
-
approximatePricePerImage: 0.042,
|
|
1179
|
-
units: [
|
|
1180
|
-
{ name: 'textInput', rate: 5, strategy: 'fixed', unit: 'millionTokens' },
|
|
1181
|
-
{ name: 'textInput_cacheRead', rate: 1.25, strategy: 'fixed', unit: 'millionTokens' },
|
|
1182
|
-
{ name: 'imageInput', rate: 10, strategy: 'fixed', unit: 'millionTokens' },
|
|
1183
|
-
{ name: 'imageInput_cacheRead', rate: 2.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
1184
|
-
{ name: 'imageOutput', rate: 40, strategy: 'fixed', unit: 'millionTokens' },
|
|
1185
|
-
],
|
|
1186
|
-
},
|
|
1187
|
-
type: 'image',
|
|
1188
|
-
},
|
|
1189
|
-
{
|
|
1190
|
-
description:
|
|
1191
|
-
'The latest DALL·E model, released in November 2023, supports more realistic, accurate image generation with stronger detail.',
|
|
1192
|
-
displayName: 'DALL·E 3',
|
|
1193
|
-
enabled: true,
|
|
1194
|
-
id: 'dall-e-3',
|
|
1195
|
-
parameters: {
|
|
1196
|
-
prompt: { default: '' },
|
|
1197
|
-
quality: {
|
|
1198
|
-
default: 'standard',
|
|
1199
|
-
enum: ['standard', 'hd'],
|
|
1200
|
-
},
|
|
1201
|
-
size: {
|
|
1202
|
-
default: '1024x1024',
|
|
1203
|
-
enum: ['1024x1024', '1792x1024', '1024x1792'],
|
|
1204
|
-
},
|
|
1205
|
-
},
|
|
1206
|
-
pricing: {
|
|
1207
|
-
approximatePricePerImage: 0.004,
|
|
1208
|
-
units: [
|
|
1209
|
-
{
|
|
1210
|
-
lookup: {
|
|
1211
|
-
prices: {
|
|
1212
|
-
hd_1024x1024: 0.08,
|
|
1213
|
-
hd_1024x1792: 0.12,
|
|
1214
|
-
hd_1792x1024: 0.12,
|
|
1215
|
-
standard_1024x1024: 0.04,
|
|
1216
|
-
standard_1024x1792: 0.08,
|
|
1217
|
-
standard_1792x1024: 0.08,
|
|
1218
|
-
},
|
|
1219
|
-
pricingParams: ['quality', 'size'],
|
|
1220
|
-
},
|
|
1221
|
-
name: 'imageGeneration',
|
|
1222
|
-
strategy: 'lookup',
|
|
1223
|
-
unit: 'image',
|
|
1224
|
-
},
|
|
1225
|
-
],
|
|
1226
|
-
},
|
|
1227
|
-
type: 'image',
|
|
1228
|
-
},
|
|
1229
|
-
{
|
|
1230
|
-
description:
|
|
1231
|
-
'Seedream 4.5, built by ByteDance Seed team, supports multi-image editing and composition. Features enhanced subject consistency, precise instruction following, spatial logic understanding, aesthetic expression, poster layout and logo design with high-precision text-image rendering.',
|
|
1232
|
-
displayName: 'Seedream 4.5',
|
|
1233
|
-
enabled: true,
|
|
1234
|
-
id: 'fal-ai/bytedance/seedream/v4.5',
|
|
1235
|
-
parameters: {
|
|
1236
|
-
height: { default: 2048, max: 4096, min: 1920, step: 1 },
|
|
1237
|
-
imageUrls: { default: [], maxCount: 10, maxFileSize: 10 * 1024 * 1024 },
|
|
1238
|
-
prompt: { default: '' },
|
|
1239
|
-
seed: { default: null },
|
|
1240
|
-
width: { default: 2048, max: 4096, min: 1920, step: 1 },
|
|
1241
|
-
},
|
|
1242
|
-
pricing: {
|
|
1243
|
-
units: [{ name: 'imageGeneration', rate: 0.04, strategy: 'fixed', unit: 'image' }],
|
|
1244
|
-
},
|
|
1245
|
-
releasedAt: '2025-12-04',
|
|
1246
|
-
type: 'image',
|
|
1247
|
-
},
|
|
1248
|
-
{
|
|
1249
|
-
description:
|
|
1250
|
-
'Seedream 4.0, built by ByteDance Seed, supports text and image inputs for highly controllable, high-quality image generation from prompts.',
|
|
1251
|
-
displayName: 'Seedream 4.0',
|
|
1252
|
-
enabled: false,
|
|
1253
|
-
id: 'fal-ai/bytedance/seedream/v4',
|
|
1254
|
-
parameters: {
|
|
1255
|
-
height: { default: 1024, max: 4096, min: 1024, step: 1 },
|
|
1256
|
-
imageUrls: { default: [], maxCount: 10, maxFileSize: 10 * 1024 * 1024 },
|
|
1257
|
-
prompt: {
|
|
1258
|
-
default: '',
|
|
1259
|
-
},
|
|
1260
|
-
seed: { default: null },
|
|
1261
|
-
width: { default: 1024, max: 4096, min: 1024, step: 1 },
|
|
1262
|
-
},
|
|
1263
|
-
pricing: {
|
|
1264
|
-
units: [{ name: 'imageGeneration', rate: 0.03, strategy: 'fixed', unit: 'image' }],
|
|
1265
|
-
},
|
|
1266
|
-
releasedAt: '2025-09-09',
|
|
1267
|
-
type: 'image',
|
|
1268
|
-
},
|
|
1269
|
-
{
|
|
1270
|
-
description: 'A powerful native multimodal image generation model.',
|
|
1271
|
-
displayName: 'HunyuanImage 3.0',
|
|
1272
|
-
enabled: true,
|
|
1273
|
-
id: 'fal-ai/hunyuan-image/v3',
|
|
1274
|
-
parameters: huanyuanImageParamsSchema,
|
|
1275
|
-
pricing: {
|
|
1276
|
-
// Original price: 0.1 x 1024 x 1024 / 100_0000 = 0.1048576$
|
|
1277
|
-
units: [{ name: 'imageGeneration', rate: 0.11, strategy: 'fixed', unit: 'image' }],
|
|
1278
|
-
},
|
|
1279
|
-
releasedAt: '2025-09-28',
|
|
1280
|
-
type: 'image',
|
|
1281
|
-
},
|
|
1282
|
-
{
|
|
1283
|
-
description:
|
|
1284
|
-
'A professional image editing model from the Qwen team, supporting semantic and appearance edits, precise Chinese/English text editing, style transfer, rotation, and more.',
|
|
1285
|
-
displayName: 'Qwen Edit',
|
|
1286
|
-
enabled: true,
|
|
1287
|
-
id: 'fal-ai/qwen-image-edit',
|
|
1288
|
-
parameters: qwenEditParamsSchema,
|
|
1289
|
-
pricing: {
|
|
1290
|
-
// https://fal.ai/models/fal-ai/qwen-image-edit
|
|
1291
|
-
// Original price: 0.03 x 1328 x 1328 / 100_0000 = 0.05290752
|
|
1292
|
-
units: [{ name: 'imageGeneration', rate: 0.06, strategy: 'fixed', unit: 'image' }],
|
|
1293
|
-
},
|
|
1294
|
-
releasedAt: '2025-08-19',
|
|
1295
|
-
type: 'image',
|
|
1296
|
-
},
|
|
1297
|
-
{
|
|
1298
|
-
description:
|
|
1299
|
-
'A powerful image generation model from the Qwen team with strong Chinese text rendering and diverse visual styles.',
|
|
1300
|
-
displayName: 'Qwen Image',
|
|
1301
|
-
enabled: true,
|
|
1302
|
-
id: 'fal-ai/qwen-image',
|
|
1303
|
-
parameters: qwenImageParamsSchema,
|
|
1304
|
-
pricing: {
|
|
1305
|
-
// Original price: 0.02 x 1328 x 1328 / 100_0000 = 0.03527168
|
|
1306
|
-
units: [{ name: 'imageGeneration', rate: 0.04, strategy: 'fixed', unit: 'image' }],
|
|
1307
|
-
},
|
|
1308
|
-
releasedAt: '2025-08-04',
|
|
1309
|
-
type: 'image',
|
|
1310
|
-
},
|
|
1311
|
-
];
|
|
1312
|
-
|
|
1313
|
-
export const allModels = [...lobehubChatModels, ...lobehubImageModels];
|
|
1314
|
-
|
|
1315
|
-
export default allModels;
|