lsd-pi 1.1.2 → 1.1.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +2 -1
- package/dist/bedrock-auth.d.ts +25 -0
- package/dist/bedrock-auth.js +59 -0
- package/dist/headless.js +8 -3
- package/dist/loader.js +1 -0
- package/dist/onboarding-llm.d.ts +37 -0
- package/dist/onboarding-llm.js +64 -0
- package/dist/onboarding.d.ts +2 -14
- package/dist/onboarding.js +146 -71
- package/dist/pi-migration.js +1 -0
- package/dist/resources/extensions/memory/auto-extract.js +21 -3
- package/dist/resources/extensions/memory/dream.js +703 -0
- package/dist/resources/extensions/memory/extension-manifest.json +2 -2
- package/dist/resources/extensions/memory/index.js +115 -8
- package/dist/resources/extensions/slash-commands/extension-manifest.json +10 -10
- package/dist/resources/extensions/slash-commands/index.js +0 -4
- package/dist/resources/extensions/slash-commands/plan.js +181 -45
- package/dist/resources/extensions/subagent/agents.js +14 -1
- package/dist/resources/extensions/subagent/configured-model.js +3 -2
- package/dist/resources/extensions/subagent/index.js +34 -28
- package/dist/resources/extensions/subagent/launch-helpers.js +24 -0
- package/dist/resources/extensions/subagent/model-resolution.js +41 -3
- package/dist/resources/extensions/usage/extension-manifest.json +11 -0
- package/dist/resources/extensions/usage/index.js +346 -0
- package/{src/resources/skills/create-gsd-extension → dist/resources/skills/create-lsd-extension}/SKILL.md +6 -6
- package/{src/resources/skills/create-gsd-extension → dist/resources/skills/create-lsd-extension}/references/custom-tools.md +1 -1
- package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/extension-lifecycle.md +2 -2
- package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/extensioncontext-reference.md +1 -1
- package/{src/resources/skills/create-gsd-extension → dist/resources/skills/create-lsd-extension}/references/key-rules-gotchas.md +4 -4
- package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/packaging-distribution.md +6 -6
- package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/workflows/create-extension.md +3 -3
- package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/workflows/debug-extension.md +5 -5
- package/dist/resources/skills/teams-debug/SKILL.md +5 -6
- package/dist/resources/skills/teams-document/SKILL.md +1 -2
- package/dist/resources/skills/teams-plan/SKILL.md +3 -4
- package/dist/resources/skills/teams-run/SKILL.md +3 -4
- package/dist/resources/skills/teams-verify/SKILL.md +4 -5
- package/dist/startup-model-validation.js +1 -0
- package/dist/welcome-screen.js +13 -11
- package/dist/wizard.js +12 -0
- package/package.json +1 -1
- package/packages/pi-ai/dist/models.generated.d.ts +688 -409
- package/packages/pi-ai/dist/models.generated.d.ts.map +1 -1
- package/packages/pi-ai/dist/models.generated.js +761 -488
- package/packages/pi-ai/dist/models.generated.js.map +1 -1
- package/packages/pi-ai/scripts/generate-models.ts +40 -18
- package/packages/pi-ai/src/models.generated.ts +759 -486
- package/packages/pi-coding-agent/dist/cli/config-selector.js +1 -1
- package/packages/pi-coding-agent/dist/cli/config-selector.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/agent-session.d.ts +1 -2
- package/packages/pi-coding-agent/dist/core/agent-session.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/agent-session.js +6 -30
- package/packages/pi-coding-agent/dist/core/agent-session.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/settings-manager.d.ts +6 -0
- package/packages/pi-coding-agent/dist/core/settings-manager.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/settings-manager.js +44 -1
- package/packages/pi-coding-agent/dist/core/settings-manager.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/skill-tool.test.js +9 -5
- package/packages/pi-coding-agent/dist/core/skill-tool.test.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/skills.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/skills.js +3 -2
- package/packages/pi-coding-agent/dist/core/skills.js.map +1 -1
- package/packages/pi-coding-agent/dist/index.d.ts +1 -1
- package/packages/pi-coding-agent/dist/index.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/index.js +1 -1
- package/packages/pi-coding-agent/dist/index.js.map +1 -1
- package/packages/pi-coding-agent/dist/main.js +1 -1
- package/packages/pi-coding-agent/dist/main.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/bash-execution.js +2 -2
- package/packages/pi-coding-agent/dist/modes/interactive/components/bash-execution.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/footer.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/footer.js +15 -12
- package/packages/pi-coding-agent/dist/modes/interactive/components/footer.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/settings-selector.d.ts +6 -0
- package/packages/pi-coding-agent/dist/modes/interactive/components/settings-selector.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/settings-selector.js +25 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/settings-selector.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/tool-execution.js +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/tool-execution.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/controllers/chat-controller.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/controllers/chat-controller.js +10 -0
- package/packages/pi-coding-agent/dist/modes/interactive/controllers/chat-controller.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/controllers/extension-ui-controller.js +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/controllers/extension-ui-controller.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.js +31 -22
- package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/theme/theme.d.ts +18 -5
- package/packages/pi-coding-agent/dist/modes/interactive/theme/theme.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/theme/theme.js +139 -20
- package/packages/pi-coding-agent/dist/modes/interactive/theme/theme.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/theme/themes.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/theme/themes.js +4 -0
- package/packages/pi-coding-agent/dist/modes/interactive/theme/themes.js.map +1 -1
- package/packages/pi-coding-agent/package.json +1 -1
- package/packages/pi-coding-agent/src/cli/config-selector.ts +1 -1
- package/packages/pi-coding-agent/src/core/agent-session.ts +5 -28
- package/packages/pi-coding-agent/src/core/settings-manager.ts +52 -1
- package/packages/pi-coding-agent/src/core/skill-tool.test.ts +18 -5
- package/packages/pi-coding-agent/src/core/skills.ts +3 -2
- package/packages/pi-coding-agent/src/index.ts +1 -1
- package/packages/pi-coding-agent/src/main.ts +1 -1
- package/packages/pi-coding-agent/src/modes/interactive/components/bash-execution.ts +2 -2
- package/packages/pi-coding-agent/src/modes/interactive/components/footer.ts +12 -13
- package/packages/pi-coding-agent/src/modes/interactive/components/settings-selector.ts +39 -1
- package/packages/pi-coding-agent/src/modes/interactive/components/tool-execution.ts +1 -1
- package/packages/pi-coding-agent/src/modes/interactive/controllers/chat-controller.ts +10 -0
- package/packages/pi-coding-agent/src/modes/interactive/controllers/extension-ui-controller.ts +1 -1
- package/packages/pi-coding-agent/src/modes/interactive/interactive-mode.ts +46 -20
- package/packages/pi-coding-agent/src/modes/interactive/theme/theme.ts +171 -20
- package/packages/pi-coding-agent/src/modes/interactive/theme/themes.ts +4 -0
- package/packages/pi-tui/dist/components/editor.d.ts +1 -0
- package/packages/pi-tui/dist/components/editor.d.ts.map +1 -1
- package/packages/pi-tui/dist/components/editor.js +23 -0
- package/packages/pi-tui/dist/components/editor.js.map +1 -1
- package/packages/pi-tui/src/components/editor.ts +23 -0
- package/pkg/dist/modes/interactive/theme/theme.d.ts +18 -5
- package/pkg/dist/modes/interactive/theme/theme.d.ts.map +1 -1
- package/pkg/dist/modes/interactive/theme/theme.js +139 -20
- package/pkg/dist/modes/interactive/theme/theme.js.map +1 -1
- package/pkg/dist/modes/interactive/theme/themes.d.ts.map +1 -1
- package/pkg/dist/modes/interactive/theme/themes.js +4 -0
- package/pkg/dist/modes/interactive/theme/themes.js.map +1 -1
- package/pkg/package.json +1 -1
- package/src/resources/extensions/memory/auto-extract.ts +23 -3
- package/src/resources/extensions/memory/dream.ts +814 -0
- package/src/resources/extensions/memory/extension-manifest.json +2 -2
- package/src/resources/extensions/memory/index.ts +134 -13
- package/src/resources/extensions/memory/tests/auto-extract.test.ts +10 -2
- package/src/resources/extensions/memory/tests/dream.test.ts +142 -0
- package/src/resources/extensions/slash-commands/extension-manifest.json +10 -10
- package/src/resources/extensions/slash-commands/index.ts +3 -7
- package/src/resources/extensions/slash-commands/plan.ts +192 -46
- package/src/resources/extensions/subagent/agents.ts +11 -1
- package/src/resources/extensions/subagent/configured-model.ts +3 -2
- package/src/resources/extensions/subagent/index.ts +38 -30
- package/src/resources/extensions/subagent/launch-helpers.ts +30 -0
- package/src/resources/extensions/subagent/model-resolution.ts +40 -3
- package/src/resources/extensions/usage/extension-manifest.json +11 -0
- package/src/resources/extensions/usage/index.ts +441 -0
- package/{dist/resources/skills/create-gsd-extension → src/resources/skills/create-lsd-extension}/SKILL.md +6 -6
- package/{dist/resources/skills/create-gsd-extension → src/resources/skills/create-lsd-extension}/references/custom-tools.md +1 -1
- package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/extension-lifecycle.md +2 -2
- package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/extensioncontext-reference.md +1 -1
- package/{dist/resources/skills/create-gsd-extension → src/resources/skills/create-lsd-extension}/references/key-rules-gotchas.md +4 -4
- package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/packaging-distribution.md +6 -6
- package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/workflows/create-extension.md +3 -3
- package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/workflows/debug-extension.md +5 -5
- package/src/resources/skills/teams-debug/SKILL.md +5 -6
- package/src/resources/skills/teams-document/SKILL.md +1 -2
- package/src/resources/skills/teams-plan/SKILL.md +3 -4
- package/src/resources/skills/teams-run/SKILL.md +3 -4
- package/src/resources/skills/teams-verify/SKILL.md +4 -5
- package/dist/resources/extensions/slash-commands/create-extension.js +0 -264
- package/dist/resources/extensions/slash-commands/create-slash-command.js +0 -208
- package/src/resources/extensions/slash-commands/create-extension.ts +0 -297
- package/src/resources/extensions/slash-commands/create-slash-command.ts +0 -234
- /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/compaction-session-control.md +0 -0
- /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/custom-commands.md +0 -0
- /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/custom-rendering.md +0 -0
- /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/custom-ui.md +0 -0
- /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/events-reference.md +0 -0
- /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/extensionapi-reference.md +0 -0
- /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/mode-behavior.md +0 -0
- /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/model-provider-management.md +0 -0
- /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/remote-execution-overrides.md +0 -0
- /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/state-management.md +0 -0
- /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/system-prompt-modification.md +0 -0
- /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/workflows/add-capability.md +0 -0
- /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/compaction-session-control.md +0 -0
- /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/custom-commands.md +0 -0
- /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/custom-rendering.md +0 -0
- /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/custom-ui.md +0 -0
- /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/events-reference.md +0 -0
- /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/extensionapi-reference.md +0 -0
- /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/mode-behavior.md +0 -0
- /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/model-provider-management.md +0 -0
- /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/remote-execution-overrides.md +0 -0
- /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/state-management.md +0 -0
- /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/system-prompt-modification.md +0 -0
- /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/templates/extension-skeleton.ts +0 -0
- /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/templates/stateful-tool-skeleton.ts +0 -0
- /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/workflows/add-capability.md +0 -0
|
@@ -104,6 +104,23 @@ export const MODELS = {
|
|
|
104
104
|
contextWindow: 200000,
|
|
105
105
|
maxTokens: 8192,
|
|
106
106
|
},
|
|
107
|
+
"anthropic.claude-3-5-sonnet-20240620-v1:0": {
|
|
108
|
+
id: "anthropic.claude-3-5-sonnet-20240620-v1:0",
|
|
109
|
+
name: "Claude Sonnet 3.5",
|
|
110
|
+
api: "bedrock-converse-stream",
|
|
111
|
+
provider: "amazon-bedrock",
|
|
112
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
113
|
+
reasoning: false,
|
|
114
|
+
input: ["text", "image"],
|
|
115
|
+
cost: {
|
|
116
|
+
input: 3,
|
|
117
|
+
output: 15,
|
|
118
|
+
cacheRead: 0.3,
|
|
119
|
+
cacheWrite: 3.75,
|
|
120
|
+
},
|
|
121
|
+
contextWindow: 200000,
|
|
122
|
+
maxTokens: 8192,
|
|
123
|
+
},
|
|
107
124
|
"anthropic.claude-3-5-sonnet-20241022-v2:0": {
|
|
108
125
|
id: "anthropic.claude-3-5-sonnet-20241022-v2:0",
|
|
109
126
|
name: "Claude Sonnet 3.5 v2",
|
|
@@ -138,6 +155,23 @@ export const MODELS = {
|
|
|
138
155
|
contextWindow: 200000,
|
|
139
156
|
maxTokens: 8192,
|
|
140
157
|
},
|
|
158
|
+
"anthropic.claude-3-haiku-20240307-v1:0": {
|
|
159
|
+
id: "anthropic.claude-3-haiku-20240307-v1:0",
|
|
160
|
+
name: "Claude Haiku 3",
|
|
161
|
+
api: "bedrock-converse-stream",
|
|
162
|
+
provider: "amazon-bedrock",
|
|
163
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
164
|
+
reasoning: false,
|
|
165
|
+
input: ["text", "image"],
|
|
166
|
+
cost: {
|
|
167
|
+
input: 0.25,
|
|
168
|
+
output: 1.25,
|
|
169
|
+
cacheRead: 0,
|
|
170
|
+
cacheWrite: 0,
|
|
171
|
+
},
|
|
172
|
+
contextWindow: 200000,
|
|
173
|
+
maxTokens: 4096,
|
|
174
|
+
},
|
|
141
175
|
"anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
142
176
|
id: "anthropic.claude-haiku-4-5-20251001-v1:0",
|
|
143
177
|
name: "Claude Haiku 4.5",
|
|
@@ -767,23 +801,6 @@ export const MODELS = {
|
|
|
767
801
|
contextWindow: 204800,
|
|
768
802
|
maxTokens: 131072,
|
|
769
803
|
},
|
|
770
|
-
"minimax.minimax-m2.5": {
|
|
771
|
-
id: "minimax.minimax-m2.5",
|
|
772
|
-
name: "MiniMax M2.5",
|
|
773
|
-
api: "bedrock-converse-stream",
|
|
774
|
-
provider: "amazon-bedrock",
|
|
775
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
776
|
-
reasoning: true,
|
|
777
|
-
input: ["text"],
|
|
778
|
-
cost: {
|
|
779
|
-
input: 0.3,
|
|
780
|
-
output: 1.2,
|
|
781
|
-
cacheRead: 0,
|
|
782
|
-
cacheWrite: 0,
|
|
783
|
-
},
|
|
784
|
-
contextWindow: 196608,
|
|
785
|
-
maxTokens: 98304,
|
|
786
|
-
},
|
|
787
804
|
"mistral.devstral-2-123b": {
|
|
788
805
|
id: "mistral.devstral-2-123b",
|
|
789
806
|
name: "Devstral 2 123B",
|
|
@@ -1022,23 +1039,6 @@ export const MODELS = {
|
|
|
1022
1039
|
contextWindow: 128000,
|
|
1023
1040
|
maxTokens: 4096,
|
|
1024
1041
|
},
|
|
1025
|
-
"nvidia.nemotron-super-3-120b": {
|
|
1026
|
-
id: "nvidia.nemotron-super-3-120b",
|
|
1027
|
-
name: "NVIDIA Nemotron 3 Super 120B A12B",
|
|
1028
|
-
api: "bedrock-converse-stream",
|
|
1029
|
-
provider: "amazon-bedrock",
|
|
1030
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
1031
|
-
reasoning: true,
|
|
1032
|
-
input: ["text"],
|
|
1033
|
-
cost: {
|
|
1034
|
-
input: 0.15,
|
|
1035
|
-
output: 0.65,
|
|
1036
|
-
cacheRead: 0,
|
|
1037
|
-
cacheWrite: 0,
|
|
1038
|
-
},
|
|
1039
|
-
contextWindow: 262144,
|
|
1040
|
-
maxTokens: 131072,
|
|
1041
|
-
},
|
|
1042
1042
|
"openai.gpt-oss-120b-1:0": {
|
|
1043
1043
|
id: "openai.gpt-oss-120b-1:0",
|
|
1044
1044
|
name: "gpt-oss-120b",
|
|
@@ -1413,23 +1413,6 @@ export const MODELS = {
|
|
|
1413
1413
|
contextWindow: 200000,
|
|
1414
1414
|
maxTokens: 131072,
|
|
1415
1415
|
},
|
|
1416
|
-
"zai.glm-5": {
|
|
1417
|
-
id: "zai.glm-5",
|
|
1418
|
-
name: "GLM-5",
|
|
1419
|
-
api: "bedrock-converse-stream",
|
|
1420
|
-
provider: "amazon-bedrock",
|
|
1421
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
1422
|
-
reasoning: true,
|
|
1423
|
-
input: ["text"],
|
|
1424
|
-
cost: {
|
|
1425
|
-
input: 1,
|
|
1426
|
-
output: 3.2,
|
|
1427
|
-
cacheRead: 0,
|
|
1428
|
-
cacheWrite: 0,
|
|
1429
|
-
},
|
|
1430
|
-
contextWindow: 202752,
|
|
1431
|
-
maxTokens: 101376,
|
|
1432
|
-
},
|
|
1433
1416
|
},
|
|
1434
1417
|
"anthropic": {
|
|
1435
1418
|
"claude-3-5-haiku-20241022": {
|
|
@@ -1466,6 +1449,23 @@ export const MODELS = {
|
|
|
1466
1449
|
contextWindow: 200000,
|
|
1467
1450
|
maxTokens: 8192,
|
|
1468
1451
|
},
|
|
1452
|
+
"claude-3-5-sonnet-20240620": {
|
|
1453
|
+
id: "claude-3-5-sonnet-20240620",
|
|
1454
|
+
name: "Claude Sonnet 3.5",
|
|
1455
|
+
api: "anthropic-messages",
|
|
1456
|
+
provider: "anthropic",
|
|
1457
|
+
baseUrl: "https://api.anthropic.com",
|
|
1458
|
+
reasoning: false,
|
|
1459
|
+
input: ["text", "image"],
|
|
1460
|
+
cost: {
|
|
1461
|
+
input: 3,
|
|
1462
|
+
output: 15,
|
|
1463
|
+
cacheRead: 0.3,
|
|
1464
|
+
cacheWrite: 3.75,
|
|
1465
|
+
},
|
|
1466
|
+
contextWindow: 200000,
|
|
1467
|
+
maxTokens: 8192,
|
|
1468
|
+
},
|
|
1469
1469
|
"claude-3-5-sonnet-20241022": {
|
|
1470
1470
|
id: "claude-3-5-sonnet-20241022",
|
|
1471
1471
|
name: "Claude Sonnet 3.5 v2",
|
|
@@ -1517,6 +1517,57 @@ export const MODELS = {
|
|
|
1517
1517
|
contextWindow: 200000,
|
|
1518
1518
|
maxTokens: 64000,
|
|
1519
1519
|
},
|
|
1520
|
+
"claude-3-haiku-20240307": {
|
|
1521
|
+
id: "claude-3-haiku-20240307",
|
|
1522
|
+
name: "Claude Haiku 3",
|
|
1523
|
+
api: "anthropic-messages",
|
|
1524
|
+
provider: "anthropic",
|
|
1525
|
+
baseUrl: "https://api.anthropic.com",
|
|
1526
|
+
reasoning: false,
|
|
1527
|
+
input: ["text", "image"],
|
|
1528
|
+
cost: {
|
|
1529
|
+
input: 0.25,
|
|
1530
|
+
output: 1.25,
|
|
1531
|
+
cacheRead: 0.03,
|
|
1532
|
+
cacheWrite: 0.3,
|
|
1533
|
+
},
|
|
1534
|
+
contextWindow: 200000,
|
|
1535
|
+
maxTokens: 4096,
|
|
1536
|
+
},
|
|
1537
|
+
"claude-3-opus-20240229": {
|
|
1538
|
+
id: "claude-3-opus-20240229",
|
|
1539
|
+
name: "Claude Opus 3",
|
|
1540
|
+
api: "anthropic-messages",
|
|
1541
|
+
provider: "anthropic",
|
|
1542
|
+
baseUrl: "https://api.anthropic.com",
|
|
1543
|
+
reasoning: false,
|
|
1544
|
+
input: ["text", "image"],
|
|
1545
|
+
cost: {
|
|
1546
|
+
input: 15,
|
|
1547
|
+
output: 75,
|
|
1548
|
+
cacheRead: 1.5,
|
|
1549
|
+
cacheWrite: 18.75,
|
|
1550
|
+
},
|
|
1551
|
+
contextWindow: 200000,
|
|
1552
|
+
maxTokens: 4096,
|
|
1553
|
+
},
|
|
1554
|
+
"claude-3-sonnet-20240229": {
|
|
1555
|
+
id: "claude-3-sonnet-20240229",
|
|
1556
|
+
name: "Claude Sonnet 3",
|
|
1557
|
+
api: "anthropic-messages",
|
|
1558
|
+
provider: "anthropic",
|
|
1559
|
+
baseUrl: "https://api.anthropic.com",
|
|
1560
|
+
reasoning: false,
|
|
1561
|
+
input: ["text", "image"],
|
|
1562
|
+
cost: {
|
|
1563
|
+
input: 3,
|
|
1564
|
+
output: 15,
|
|
1565
|
+
cacheRead: 0.3,
|
|
1566
|
+
cacheWrite: 0.3,
|
|
1567
|
+
},
|
|
1568
|
+
contextWindow: 200000,
|
|
1569
|
+
maxTokens: 4096,
|
|
1570
|
+
},
|
|
1520
1571
|
"claude-haiku-4-5": {
|
|
1521
1572
|
id: "claude-haiku-4-5",
|
|
1522
1573
|
name: "Claude Haiku 4.5 (latest)",
|
|
@@ -1774,6 +1825,40 @@ export const MODELS = {
|
|
|
1774
1825
|
contextWindow: 200000,
|
|
1775
1826
|
maxTokens: 100000,
|
|
1776
1827
|
},
|
|
1828
|
+
"gpt-4": {
|
|
1829
|
+
id: "gpt-4",
|
|
1830
|
+
name: "GPT-4",
|
|
1831
|
+
api: "azure-openai-responses",
|
|
1832
|
+
provider: "azure-openai-responses",
|
|
1833
|
+
baseUrl: "",
|
|
1834
|
+
reasoning: false,
|
|
1835
|
+
input: ["text"],
|
|
1836
|
+
cost: {
|
|
1837
|
+
input: 30,
|
|
1838
|
+
output: 60,
|
|
1839
|
+
cacheRead: 0,
|
|
1840
|
+
cacheWrite: 0,
|
|
1841
|
+
},
|
|
1842
|
+
contextWindow: 8192,
|
|
1843
|
+
maxTokens: 8192,
|
|
1844
|
+
},
|
|
1845
|
+
"gpt-4-turbo": {
|
|
1846
|
+
id: "gpt-4-turbo",
|
|
1847
|
+
name: "GPT-4 Turbo",
|
|
1848
|
+
api: "azure-openai-responses",
|
|
1849
|
+
provider: "azure-openai-responses",
|
|
1850
|
+
baseUrl: "",
|
|
1851
|
+
reasoning: false,
|
|
1852
|
+
input: ["text", "image"],
|
|
1853
|
+
cost: {
|
|
1854
|
+
input: 10,
|
|
1855
|
+
output: 30,
|
|
1856
|
+
cacheRead: 0,
|
|
1857
|
+
cacheWrite: 0,
|
|
1858
|
+
},
|
|
1859
|
+
contextWindow: 128000,
|
|
1860
|
+
maxTokens: 4096,
|
|
1861
|
+
},
|
|
1777
1862
|
"gpt-4.1": {
|
|
1778
1863
|
id: "gpt-4.1",
|
|
1779
1864
|
name: "GPT-4.1",
|
|
@@ -1825,6 +1910,91 @@ export const MODELS = {
|
|
|
1825
1910
|
contextWindow: 1047576,
|
|
1826
1911
|
maxTokens: 32768,
|
|
1827
1912
|
},
|
|
1913
|
+
"gpt-4o": {
|
|
1914
|
+
id: "gpt-4o",
|
|
1915
|
+
name: "GPT-4o",
|
|
1916
|
+
api: "azure-openai-responses",
|
|
1917
|
+
provider: "azure-openai-responses",
|
|
1918
|
+
baseUrl: "",
|
|
1919
|
+
reasoning: false,
|
|
1920
|
+
input: ["text", "image"],
|
|
1921
|
+
cost: {
|
|
1922
|
+
input: 2.5,
|
|
1923
|
+
output: 10,
|
|
1924
|
+
cacheRead: 1.25,
|
|
1925
|
+
cacheWrite: 0,
|
|
1926
|
+
},
|
|
1927
|
+
contextWindow: 128000,
|
|
1928
|
+
maxTokens: 16384,
|
|
1929
|
+
},
|
|
1930
|
+
"gpt-4o-2024-05-13": {
|
|
1931
|
+
id: "gpt-4o-2024-05-13",
|
|
1932
|
+
name: "GPT-4o (2024-05-13)",
|
|
1933
|
+
api: "azure-openai-responses",
|
|
1934
|
+
provider: "azure-openai-responses",
|
|
1935
|
+
baseUrl: "",
|
|
1936
|
+
reasoning: false,
|
|
1937
|
+
input: ["text", "image"],
|
|
1938
|
+
cost: {
|
|
1939
|
+
input: 5,
|
|
1940
|
+
output: 15,
|
|
1941
|
+
cacheRead: 0,
|
|
1942
|
+
cacheWrite: 0,
|
|
1943
|
+
},
|
|
1944
|
+
contextWindow: 128000,
|
|
1945
|
+
maxTokens: 4096,
|
|
1946
|
+
},
|
|
1947
|
+
"gpt-4o-2024-08-06": {
|
|
1948
|
+
id: "gpt-4o-2024-08-06",
|
|
1949
|
+
name: "GPT-4o (2024-08-06)",
|
|
1950
|
+
api: "azure-openai-responses",
|
|
1951
|
+
provider: "azure-openai-responses",
|
|
1952
|
+
baseUrl: "",
|
|
1953
|
+
reasoning: false,
|
|
1954
|
+
input: ["text", "image"],
|
|
1955
|
+
cost: {
|
|
1956
|
+
input: 2.5,
|
|
1957
|
+
output: 10,
|
|
1958
|
+
cacheRead: 1.25,
|
|
1959
|
+
cacheWrite: 0,
|
|
1960
|
+
},
|
|
1961
|
+
contextWindow: 128000,
|
|
1962
|
+
maxTokens: 16384,
|
|
1963
|
+
},
|
|
1964
|
+
"gpt-4o-2024-11-20": {
|
|
1965
|
+
id: "gpt-4o-2024-11-20",
|
|
1966
|
+
name: "GPT-4o (2024-11-20)",
|
|
1967
|
+
api: "azure-openai-responses",
|
|
1968
|
+
provider: "azure-openai-responses",
|
|
1969
|
+
baseUrl: "",
|
|
1970
|
+
reasoning: false,
|
|
1971
|
+
input: ["text", "image"],
|
|
1972
|
+
cost: {
|
|
1973
|
+
input: 2.5,
|
|
1974
|
+
output: 10,
|
|
1975
|
+
cacheRead: 1.25,
|
|
1976
|
+
cacheWrite: 0,
|
|
1977
|
+
},
|
|
1978
|
+
contextWindow: 128000,
|
|
1979
|
+
maxTokens: 16384,
|
|
1980
|
+
},
|
|
1981
|
+
"gpt-4o-mini": {
|
|
1982
|
+
id: "gpt-4o-mini",
|
|
1983
|
+
name: "GPT-4o mini",
|
|
1984
|
+
api: "azure-openai-responses",
|
|
1985
|
+
provider: "azure-openai-responses",
|
|
1986
|
+
baseUrl: "",
|
|
1987
|
+
reasoning: false,
|
|
1988
|
+
input: ["text", "image"],
|
|
1989
|
+
cost: {
|
|
1990
|
+
input: 0.15,
|
|
1991
|
+
output: 0.6,
|
|
1992
|
+
cacheRead: 0.08,
|
|
1993
|
+
cacheWrite: 0,
|
|
1994
|
+
},
|
|
1995
|
+
contextWindow: 128000,
|
|
1996
|
+
maxTokens: 16384,
|
|
1997
|
+
},
|
|
1828
1998
|
"gpt-5": {
|
|
1829
1999
|
id: "gpt-5",
|
|
1830
2000
|
name: "GPT-5",
|
|
@@ -2080,23 +2250,6 @@ export const MODELS = {
|
|
|
2080
2250
|
contextWindow: 400000,
|
|
2081
2251
|
maxTokens: 128000,
|
|
2082
2252
|
},
|
|
2083
|
-
"gpt-5.3-chat-latest": {
|
|
2084
|
-
id: "gpt-5.3-chat-latest",
|
|
2085
|
-
name: "GPT-5.3 Chat (latest)",
|
|
2086
|
-
api: "azure-openai-responses",
|
|
2087
|
-
provider: "azure-openai-responses",
|
|
2088
|
-
baseUrl: "",
|
|
2089
|
-
reasoning: false,
|
|
2090
|
-
input: ["text", "image"],
|
|
2091
|
-
cost: {
|
|
2092
|
-
input: 1.75,
|
|
2093
|
-
output: 14,
|
|
2094
|
-
cacheRead: 0.175,
|
|
2095
|
-
cacheWrite: 0,
|
|
2096
|
-
},
|
|
2097
|
-
contextWindow: 128000,
|
|
2098
|
-
maxTokens: 16384,
|
|
2099
|
-
},
|
|
2100
2253
|
"gpt-5.3-codex": {
|
|
2101
2254
|
id: "gpt-5.3-codex",
|
|
2102
2255
|
name: "GPT-5.3 Codex",
|
|
@@ -2610,6 +2763,25 @@ export const MODELS = {
|
|
|
2610
2763
|
contextWindow: 128000,
|
|
2611
2764
|
maxTokens: 16384,
|
|
2612
2765
|
},
|
|
2766
|
+
"gpt-4o": {
|
|
2767
|
+
id: "gpt-4o",
|
|
2768
|
+
name: "GPT-4o",
|
|
2769
|
+
api: "openai-completions",
|
|
2770
|
+
provider: "github-copilot",
|
|
2771
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
2772
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
2773
|
+
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false },
|
|
2774
|
+
reasoning: false,
|
|
2775
|
+
input: ["text", "image"],
|
|
2776
|
+
cost: {
|
|
2777
|
+
input: 0,
|
|
2778
|
+
output: 0,
|
|
2779
|
+
cacheRead: 0,
|
|
2780
|
+
cacheWrite: 0,
|
|
2781
|
+
},
|
|
2782
|
+
contextWindow: 128000,
|
|
2783
|
+
maxTokens: 4096,
|
|
2784
|
+
},
|
|
2613
2785
|
"gpt-5": {
|
|
2614
2786
|
id: "gpt-5",
|
|
2615
2787
|
name: "GPT-5",
|
|
@@ -2829,7 +3001,58 @@ export const MODELS = {
|
|
|
2829
3001
|
},
|
|
2830
3002
|
},
|
|
2831
3003
|
"google": {
|
|
2832
|
-
"gemini-
|
|
3004
|
+
"gemini-1.5-flash": {
|
|
3005
|
+
id: "gemini-1.5-flash",
|
|
3006
|
+
name: "Gemini 1.5 Flash",
|
|
3007
|
+
api: "google-generative-ai",
|
|
3008
|
+
provider: "google",
|
|
3009
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
3010
|
+
reasoning: false,
|
|
3011
|
+
input: ["text", "image"],
|
|
3012
|
+
cost: {
|
|
3013
|
+
input: 0.075,
|
|
3014
|
+
output: 0.3,
|
|
3015
|
+
cacheRead: 0.01875,
|
|
3016
|
+
cacheWrite: 0,
|
|
3017
|
+
},
|
|
3018
|
+
contextWindow: 1000000,
|
|
3019
|
+
maxTokens: 8192,
|
|
3020
|
+
},
|
|
3021
|
+
"gemini-1.5-flash-8b": {
|
|
3022
|
+
id: "gemini-1.5-flash-8b",
|
|
3023
|
+
name: "Gemini 1.5 Flash-8B",
|
|
3024
|
+
api: "google-generative-ai",
|
|
3025
|
+
provider: "google",
|
|
3026
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
3027
|
+
reasoning: false,
|
|
3028
|
+
input: ["text", "image"],
|
|
3029
|
+
cost: {
|
|
3030
|
+
input: 0.0375,
|
|
3031
|
+
output: 0.15,
|
|
3032
|
+
cacheRead: 0.01,
|
|
3033
|
+
cacheWrite: 0,
|
|
3034
|
+
},
|
|
3035
|
+
contextWindow: 1000000,
|
|
3036
|
+
maxTokens: 8192,
|
|
3037
|
+
},
|
|
3038
|
+
"gemini-1.5-pro": {
|
|
3039
|
+
id: "gemini-1.5-pro",
|
|
3040
|
+
name: "Gemini 1.5 Pro",
|
|
3041
|
+
api: "google-generative-ai",
|
|
3042
|
+
provider: "google",
|
|
3043
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
3044
|
+
reasoning: false,
|
|
3045
|
+
input: ["text", "image"],
|
|
3046
|
+
cost: {
|
|
3047
|
+
input: 1.25,
|
|
3048
|
+
output: 5,
|
|
3049
|
+
cacheRead: 0.3125,
|
|
3050
|
+
cacheWrite: 0,
|
|
3051
|
+
},
|
|
3052
|
+
contextWindow: 1000000,
|
|
3053
|
+
maxTokens: 8192,
|
|
3054
|
+
},
|
|
3055
|
+
"gemini-2.0-flash": {
|
|
2833
3056
|
id: "gemini-2.0-flash",
|
|
2834
3057
|
name: "Gemini 2.0 Flash",
|
|
2835
3058
|
api: "google-generative-ai",
|
|
@@ -3186,23 +3409,6 @@ export const MODELS = {
|
|
|
3186
3409
|
contextWindow: 131072,
|
|
3187
3410
|
maxTokens: 65536,
|
|
3188
3411
|
},
|
|
3189
|
-
"gemma-3-27b-it": {
|
|
3190
|
-
id: "gemma-3-27b-it",
|
|
3191
|
-
name: "Gemma 3 27B",
|
|
3192
|
-
api: "google-generative-ai",
|
|
3193
|
-
provider: "google",
|
|
3194
|
-
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
3195
|
-
reasoning: false,
|
|
3196
|
-
input: ["text", "image"],
|
|
3197
|
-
cost: {
|
|
3198
|
-
input: 0,
|
|
3199
|
-
output: 0,
|
|
3200
|
-
cacheRead: 0,
|
|
3201
|
-
cacheWrite: 0,
|
|
3202
|
-
},
|
|
3203
|
-
contextWindow: 131072,
|
|
3204
|
-
maxTokens: 8192,
|
|
3205
|
-
},
|
|
3206
3412
|
},
|
|
3207
3413
|
"google-antigravity": {
|
|
3208
3414
|
"claude-opus-4-5-thinking": {
|
|
@@ -3464,6 +3670,57 @@ export const MODELS = {
|
|
|
3464
3670
|
},
|
|
3465
3671
|
},
|
|
3466
3672
|
"google-vertex": {
|
|
3673
|
+
"gemini-1.5-flash": {
|
|
3674
|
+
id: "gemini-1.5-flash",
|
|
3675
|
+
name: "Gemini 1.5 Flash (Vertex)",
|
|
3676
|
+
api: "google-vertex",
|
|
3677
|
+
provider: "google-vertex",
|
|
3678
|
+
baseUrl: "https://{location}-aiplatform.googleapis.com",
|
|
3679
|
+
reasoning: false,
|
|
3680
|
+
input: ["text", "image"],
|
|
3681
|
+
cost: {
|
|
3682
|
+
input: 0.075,
|
|
3683
|
+
output: 0.3,
|
|
3684
|
+
cacheRead: 0.01875,
|
|
3685
|
+
cacheWrite: 0,
|
|
3686
|
+
},
|
|
3687
|
+
contextWindow: 1000000,
|
|
3688
|
+
maxTokens: 8192,
|
|
3689
|
+
},
|
|
3690
|
+
"gemini-1.5-flash-8b": {
|
|
3691
|
+
id: "gemini-1.5-flash-8b",
|
|
3692
|
+
name: "Gemini 1.5 Flash-8B (Vertex)",
|
|
3693
|
+
api: "google-vertex",
|
|
3694
|
+
provider: "google-vertex",
|
|
3695
|
+
baseUrl: "https://{location}-aiplatform.googleapis.com",
|
|
3696
|
+
reasoning: false,
|
|
3697
|
+
input: ["text", "image"],
|
|
3698
|
+
cost: {
|
|
3699
|
+
input: 0.0375,
|
|
3700
|
+
output: 0.15,
|
|
3701
|
+
cacheRead: 0.01,
|
|
3702
|
+
cacheWrite: 0,
|
|
3703
|
+
},
|
|
3704
|
+
contextWindow: 1000000,
|
|
3705
|
+
maxTokens: 8192,
|
|
3706
|
+
},
|
|
3707
|
+
"gemini-1.5-pro": {
|
|
3708
|
+
id: "gemini-1.5-pro",
|
|
3709
|
+
name: "Gemini 1.5 Pro (Vertex)",
|
|
3710
|
+
api: "google-vertex",
|
|
3711
|
+
provider: "google-vertex",
|
|
3712
|
+
baseUrl: "https://{location}-aiplatform.googleapis.com",
|
|
3713
|
+
reasoning: false,
|
|
3714
|
+
input: ["text", "image"],
|
|
3715
|
+
cost: {
|
|
3716
|
+
input: 1.25,
|
|
3717
|
+
output: 5,
|
|
3718
|
+
cacheRead: 0.3125,
|
|
3719
|
+
cacheWrite: 0,
|
|
3720
|
+
},
|
|
3721
|
+
contextWindow: 1000000,
|
|
3722
|
+
maxTokens: 8192,
|
|
3723
|
+
},
|
|
3467
3724
|
"gemini-2.0-flash": {
|
|
3468
3725
|
id: "gemini-2.0-flash",
|
|
3469
3726
|
name: "Gemini 2.0 Flash (Vertex)",
|
|
@@ -3653,40 +3910,6 @@ export const MODELS = {
|
|
|
3653
3910
|
contextWindow: 8192,
|
|
3654
3911
|
maxTokens: 8192,
|
|
3655
3912
|
},
|
|
3656
|
-
"groq/compound": {
|
|
3657
|
-
id: "groq/compound",
|
|
3658
|
-
name: "Compound",
|
|
3659
|
-
api: "openai-completions",
|
|
3660
|
-
provider: "groq",
|
|
3661
|
-
baseUrl: "https://api.groq.com/openai/v1",
|
|
3662
|
-
reasoning: true,
|
|
3663
|
-
input: ["text"],
|
|
3664
|
-
cost: {
|
|
3665
|
-
input: 0,
|
|
3666
|
-
output: 0,
|
|
3667
|
-
cacheRead: 0,
|
|
3668
|
-
cacheWrite: 0,
|
|
3669
|
-
},
|
|
3670
|
-
contextWindow: 131072,
|
|
3671
|
-
maxTokens: 8192,
|
|
3672
|
-
},
|
|
3673
|
-
"groq/compound-mini": {
|
|
3674
|
-
id: "groq/compound-mini",
|
|
3675
|
-
name: "Compound Mini",
|
|
3676
|
-
api: "openai-completions",
|
|
3677
|
-
provider: "groq",
|
|
3678
|
-
baseUrl: "https://api.groq.com/openai/v1",
|
|
3679
|
-
reasoning: true,
|
|
3680
|
-
input: ["text"],
|
|
3681
|
-
cost: {
|
|
3682
|
-
input: 0,
|
|
3683
|
-
output: 0,
|
|
3684
|
-
cacheRead: 0,
|
|
3685
|
-
cacheWrite: 0,
|
|
3686
|
-
},
|
|
3687
|
-
contextWindow: 131072,
|
|
3688
|
-
maxTokens: 8192,
|
|
3689
|
-
},
|
|
3690
3913
|
"llama-3.1-8b-instant": {
|
|
3691
3914
|
id: "llama-3.1-8b-instant",
|
|
3692
3915
|
name: "Llama 3.1 8B Instant",
|
|
@@ -3874,23 +4097,6 @@ export const MODELS = {
|
|
|
3874
4097
|
contextWindow: 131072,
|
|
3875
4098
|
maxTokens: 65536,
|
|
3876
4099
|
},
|
|
3877
|
-
"openai/gpt-oss-safeguard-20b": {
|
|
3878
|
-
id: "openai/gpt-oss-safeguard-20b",
|
|
3879
|
-
name: "Safety GPT OSS 20B",
|
|
3880
|
-
api: "openai-completions",
|
|
3881
|
-
provider: "groq",
|
|
3882
|
-
baseUrl: "https://api.groq.com/openai/v1",
|
|
3883
|
-
reasoning: true,
|
|
3884
|
-
input: ["text"],
|
|
3885
|
-
cost: {
|
|
3886
|
-
input: 0.075,
|
|
3887
|
-
output: 0.3,
|
|
3888
|
-
cacheRead: 0.037,
|
|
3889
|
-
cacheWrite: 0,
|
|
3890
|
-
},
|
|
3891
|
-
contextWindow: 131072,
|
|
3892
|
-
maxTokens: 65536,
|
|
3893
|
-
},
|
|
3894
4100
|
"qwen-qwq-32b": {
|
|
3895
4101
|
id: "qwen-qwq-32b",
|
|
3896
4102
|
name: "Qwen QwQ 32B",
|
|
@@ -3923,7 +4129,7 @@ export const MODELS = {
|
|
|
3923
4129
|
cacheWrite: 0,
|
|
3924
4130
|
},
|
|
3925
4131
|
contextWindow: 131072,
|
|
3926
|
-
maxTokens:
|
|
4132
|
+
maxTokens: 16384,
|
|
3927
4133
|
},
|
|
3928
4134
|
},
|
|
3929
4135
|
"huggingface": {
|
|
@@ -4820,39 +5026,22 @@ export const MODELS = {
|
|
|
4820
5026
|
contextWindow: 128000,
|
|
4821
5027
|
maxTokens: 16384,
|
|
4822
5028
|
},
|
|
4823
|
-
"mistral-small-2603": {
|
|
4824
|
-
id: "mistral-small-2603",
|
|
4825
|
-
name: "Mistral Small 4",
|
|
4826
|
-
api: "mistral-conversations",
|
|
4827
|
-
provider: "mistral",
|
|
4828
|
-
baseUrl: "https://api.mistral.ai",
|
|
4829
|
-
reasoning: true,
|
|
4830
|
-
input: ["text", "image"],
|
|
4831
|
-
cost: {
|
|
4832
|
-
input: 0.15,
|
|
4833
|
-
output: 0.6,
|
|
4834
|
-
cacheRead: 0,
|
|
4835
|
-
cacheWrite: 0,
|
|
4836
|
-
},
|
|
4837
|
-
contextWindow: 256000,
|
|
4838
|
-
maxTokens: 256000,
|
|
4839
|
-
},
|
|
4840
5029
|
"mistral-small-latest": {
|
|
4841
5030
|
id: "mistral-small-latest",
|
|
4842
5031
|
name: "Mistral Small (latest)",
|
|
4843
5032
|
api: "mistral-conversations",
|
|
4844
5033
|
provider: "mistral",
|
|
4845
5034
|
baseUrl: "https://api.mistral.ai",
|
|
4846
|
-
reasoning:
|
|
5035
|
+
reasoning: false,
|
|
4847
5036
|
input: ["text", "image"],
|
|
4848
5037
|
cost: {
|
|
4849
|
-
input: 0.
|
|
4850
|
-
output: 0.
|
|
5038
|
+
input: 0.1,
|
|
5039
|
+
output: 0.3,
|
|
4851
5040
|
cacheRead: 0,
|
|
4852
5041
|
cacheWrite: 0,
|
|
4853
5042
|
},
|
|
4854
|
-
contextWindow:
|
|
4855
|
-
maxTokens:
|
|
5043
|
+
contextWindow: 128000,
|
|
5044
|
+
maxTokens: 16384,
|
|
4856
5045
|
},
|
|
4857
5046
|
"open-mistral-7b": {
|
|
4858
5047
|
id: "open-mistral-7b",
|
|
@@ -4931,83 +5120,202 @@ export const MODELS = {
|
|
|
4931
5120
|
reasoning: false,
|
|
4932
5121
|
input: ["text", "image"],
|
|
4933
5122
|
cost: {
|
|
4934
|
-
input: 2,
|
|
4935
|
-
output: 6,
|
|
4936
|
-
cacheRead: 0,
|
|
5123
|
+
input: 2,
|
|
5124
|
+
output: 6,
|
|
5125
|
+
cacheRead: 0,
|
|
5126
|
+
cacheWrite: 0,
|
|
5127
|
+
},
|
|
5128
|
+
contextWindow: 128000,
|
|
5129
|
+
maxTokens: 128000,
|
|
5130
|
+
},
|
|
5131
|
+
},
|
|
5132
|
+
"openai": {
|
|
5133
|
+
"codex-mini-latest": {
|
|
5134
|
+
id: "codex-mini-latest",
|
|
5135
|
+
name: "Codex Mini",
|
|
5136
|
+
api: "openai-responses",
|
|
5137
|
+
provider: "openai",
|
|
5138
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5139
|
+
reasoning: true,
|
|
5140
|
+
input: ["text"],
|
|
5141
|
+
cost: {
|
|
5142
|
+
input: 1.5,
|
|
5143
|
+
output: 6,
|
|
5144
|
+
cacheRead: 0.375,
|
|
5145
|
+
cacheWrite: 0,
|
|
5146
|
+
},
|
|
5147
|
+
contextWindow: 200000,
|
|
5148
|
+
maxTokens: 100000,
|
|
5149
|
+
},
|
|
5150
|
+
"gpt-4": {
|
|
5151
|
+
id: "gpt-4",
|
|
5152
|
+
name: "GPT-4",
|
|
5153
|
+
api: "openai-responses",
|
|
5154
|
+
provider: "openai",
|
|
5155
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5156
|
+
reasoning: false,
|
|
5157
|
+
input: ["text"],
|
|
5158
|
+
cost: {
|
|
5159
|
+
input: 30,
|
|
5160
|
+
output: 60,
|
|
5161
|
+
cacheRead: 0,
|
|
5162
|
+
cacheWrite: 0,
|
|
5163
|
+
},
|
|
5164
|
+
contextWindow: 8192,
|
|
5165
|
+
maxTokens: 8192,
|
|
5166
|
+
},
|
|
5167
|
+
"gpt-4-turbo": {
|
|
5168
|
+
id: "gpt-4-turbo",
|
|
5169
|
+
name: "GPT-4 Turbo",
|
|
5170
|
+
api: "openai-responses",
|
|
5171
|
+
provider: "openai",
|
|
5172
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5173
|
+
reasoning: false,
|
|
5174
|
+
input: ["text", "image"],
|
|
5175
|
+
cost: {
|
|
5176
|
+
input: 10,
|
|
5177
|
+
output: 30,
|
|
5178
|
+
cacheRead: 0,
|
|
5179
|
+
cacheWrite: 0,
|
|
5180
|
+
},
|
|
5181
|
+
contextWindow: 128000,
|
|
5182
|
+
maxTokens: 4096,
|
|
5183
|
+
},
|
|
5184
|
+
"gpt-4.1": {
|
|
5185
|
+
id: "gpt-4.1",
|
|
5186
|
+
name: "GPT-4.1",
|
|
5187
|
+
api: "openai-responses",
|
|
5188
|
+
provider: "openai",
|
|
5189
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5190
|
+
reasoning: false,
|
|
5191
|
+
input: ["text", "image"],
|
|
5192
|
+
cost: {
|
|
5193
|
+
input: 2,
|
|
5194
|
+
output: 8,
|
|
5195
|
+
cacheRead: 0.5,
|
|
5196
|
+
cacheWrite: 0,
|
|
5197
|
+
},
|
|
5198
|
+
contextWindow: 1047576,
|
|
5199
|
+
maxTokens: 32768,
|
|
5200
|
+
},
|
|
5201
|
+
"gpt-4.1-mini": {
|
|
5202
|
+
id: "gpt-4.1-mini",
|
|
5203
|
+
name: "GPT-4.1 mini",
|
|
5204
|
+
api: "openai-responses",
|
|
5205
|
+
provider: "openai",
|
|
5206
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5207
|
+
reasoning: false,
|
|
5208
|
+
input: ["text", "image"],
|
|
5209
|
+
cost: {
|
|
5210
|
+
input: 0.4,
|
|
5211
|
+
output: 1.6,
|
|
5212
|
+
cacheRead: 0.1,
|
|
5213
|
+
cacheWrite: 0,
|
|
5214
|
+
},
|
|
5215
|
+
contextWindow: 1047576,
|
|
5216
|
+
maxTokens: 32768,
|
|
5217
|
+
},
|
|
5218
|
+
"gpt-4.1-nano": {
|
|
5219
|
+
id: "gpt-4.1-nano",
|
|
5220
|
+
name: "GPT-4.1 nano",
|
|
5221
|
+
api: "openai-responses",
|
|
5222
|
+
provider: "openai",
|
|
5223
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5224
|
+
reasoning: false,
|
|
5225
|
+
input: ["text", "image"],
|
|
5226
|
+
cost: {
|
|
5227
|
+
input: 0.1,
|
|
5228
|
+
output: 0.4,
|
|
5229
|
+
cacheRead: 0.03,
|
|
5230
|
+
cacheWrite: 0,
|
|
5231
|
+
},
|
|
5232
|
+
contextWindow: 1047576,
|
|
5233
|
+
maxTokens: 32768,
|
|
5234
|
+
},
|
|
5235
|
+
"gpt-4o": {
|
|
5236
|
+
id: "gpt-4o",
|
|
5237
|
+
name: "GPT-4o",
|
|
5238
|
+
api: "openai-responses",
|
|
5239
|
+
provider: "openai",
|
|
5240
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5241
|
+
reasoning: false,
|
|
5242
|
+
input: ["text", "image"],
|
|
5243
|
+
cost: {
|
|
5244
|
+
input: 2.5,
|
|
5245
|
+
output: 10,
|
|
5246
|
+
cacheRead: 1.25,
|
|
4937
5247
|
cacheWrite: 0,
|
|
4938
5248
|
},
|
|
4939
5249
|
contextWindow: 128000,
|
|
4940
|
-
maxTokens:
|
|
5250
|
+
maxTokens: 16384,
|
|
4941
5251
|
},
|
|
4942
|
-
|
|
4943
|
-
|
|
4944
|
-
|
|
4945
|
-
id: "codex-mini-latest",
|
|
4946
|
-
name: "Codex Mini",
|
|
5252
|
+
"gpt-4o-2024-05-13": {
|
|
5253
|
+
id: "gpt-4o-2024-05-13",
|
|
5254
|
+
name: "GPT-4o (2024-05-13)",
|
|
4947
5255
|
api: "openai-responses",
|
|
4948
5256
|
provider: "openai",
|
|
4949
5257
|
baseUrl: "https://api.openai.com/v1",
|
|
4950
|
-
reasoning:
|
|
4951
|
-
input: ["text"],
|
|
5258
|
+
reasoning: false,
|
|
5259
|
+
input: ["text", "image"],
|
|
4952
5260
|
cost: {
|
|
4953
|
-
input:
|
|
4954
|
-
output:
|
|
4955
|
-
cacheRead: 0
|
|
5261
|
+
input: 5,
|
|
5262
|
+
output: 15,
|
|
5263
|
+
cacheRead: 0,
|
|
4956
5264
|
cacheWrite: 0,
|
|
4957
5265
|
},
|
|
4958
|
-
contextWindow:
|
|
4959
|
-
maxTokens:
|
|
5266
|
+
contextWindow: 128000,
|
|
5267
|
+
maxTokens: 4096,
|
|
4960
5268
|
},
|
|
4961
|
-
"gpt-
|
|
4962
|
-
id: "gpt-
|
|
4963
|
-
name: "GPT-
|
|
5269
|
+
"gpt-4o-2024-08-06": {
|
|
5270
|
+
id: "gpt-4o-2024-08-06",
|
|
5271
|
+
name: "GPT-4o (2024-08-06)",
|
|
4964
5272
|
api: "openai-responses",
|
|
4965
5273
|
provider: "openai",
|
|
4966
5274
|
baseUrl: "https://api.openai.com/v1",
|
|
4967
5275
|
reasoning: false,
|
|
4968
5276
|
input: ["text", "image"],
|
|
4969
5277
|
cost: {
|
|
4970
|
-
input: 2,
|
|
4971
|
-
output:
|
|
4972
|
-
cacheRead:
|
|
5278
|
+
input: 2.5,
|
|
5279
|
+
output: 10,
|
|
5280
|
+
cacheRead: 1.25,
|
|
4973
5281
|
cacheWrite: 0,
|
|
4974
5282
|
},
|
|
4975
|
-
contextWindow:
|
|
4976
|
-
maxTokens:
|
|
5283
|
+
contextWindow: 128000,
|
|
5284
|
+
maxTokens: 16384,
|
|
4977
5285
|
},
|
|
4978
|
-
"gpt-
|
|
4979
|
-
id: "gpt-
|
|
4980
|
-
name: "GPT-
|
|
5286
|
+
"gpt-4o-2024-11-20": {
|
|
5287
|
+
id: "gpt-4o-2024-11-20",
|
|
5288
|
+
name: "GPT-4o (2024-11-20)",
|
|
4981
5289
|
api: "openai-responses",
|
|
4982
5290
|
provider: "openai",
|
|
4983
5291
|
baseUrl: "https://api.openai.com/v1",
|
|
4984
5292
|
reasoning: false,
|
|
4985
5293
|
input: ["text", "image"],
|
|
4986
5294
|
cost: {
|
|
4987
|
-
input:
|
|
4988
|
-
output:
|
|
4989
|
-
cacheRead:
|
|
5295
|
+
input: 2.5,
|
|
5296
|
+
output: 10,
|
|
5297
|
+
cacheRead: 1.25,
|
|
4990
5298
|
cacheWrite: 0,
|
|
4991
5299
|
},
|
|
4992
|
-
contextWindow:
|
|
4993
|
-
maxTokens:
|
|
5300
|
+
contextWindow: 128000,
|
|
5301
|
+
maxTokens: 16384,
|
|
4994
5302
|
},
|
|
4995
|
-
"gpt-
|
|
4996
|
-
id: "gpt-
|
|
4997
|
-
name: "GPT-
|
|
5303
|
+
"gpt-4o-mini": {
|
|
5304
|
+
id: "gpt-4o-mini",
|
|
5305
|
+
name: "GPT-4o mini",
|
|
4998
5306
|
api: "openai-responses",
|
|
4999
5307
|
provider: "openai",
|
|
5000
5308
|
baseUrl: "https://api.openai.com/v1",
|
|
5001
5309
|
reasoning: false,
|
|
5002
5310
|
input: ["text", "image"],
|
|
5003
5311
|
cost: {
|
|
5004
|
-
input: 0.
|
|
5005
|
-
output: 0.
|
|
5006
|
-
cacheRead: 0.
|
|
5312
|
+
input: 0.15,
|
|
5313
|
+
output: 0.6,
|
|
5314
|
+
cacheRead: 0.08,
|
|
5007
5315
|
cacheWrite: 0,
|
|
5008
5316
|
},
|
|
5009
|
-
contextWindow:
|
|
5010
|
-
maxTokens:
|
|
5317
|
+
contextWindow: 128000,
|
|
5318
|
+
maxTokens: 16384,
|
|
5011
5319
|
},
|
|
5012
5320
|
"gpt-5": {
|
|
5013
5321
|
id: "gpt-5",
|
|
@@ -5264,23 +5572,6 @@ export const MODELS = {
|
|
|
5264
5572
|
contextWindow: 400000,
|
|
5265
5573
|
maxTokens: 128000,
|
|
5266
5574
|
},
|
|
5267
|
-
"gpt-5.3-chat-latest": {
|
|
5268
|
-
id: "gpt-5.3-chat-latest",
|
|
5269
|
-
name: "GPT-5.3 Chat (latest)",
|
|
5270
|
-
api: "openai-responses",
|
|
5271
|
-
provider: "openai",
|
|
5272
|
-
baseUrl: "https://api.openai.com/v1",
|
|
5273
|
-
reasoning: false,
|
|
5274
|
-
input: ["text", "image"],
|
|
5275
|
-
cost: {
|
|
5276
|
-
input: 1.75,
|
|
5277
|
-
output: 14,
|
|
5278
|
-
cacheRead: 0.175,
|
|
5279
|
-
cacheWrite: 0,
|
|
5280
|
-
},
|
|
5281
|
-
contextWindow: 128000,
|
|
5282
|
-
maxTokens: 16384,
|
|
5283
|
-
},
|
|
5284
5575
|
"gpt-5.3-codex": {
|
|
5285
5576
|
id: "gpt-5.3-codex",
|
|
5286
5577
|
name: "GPT-5.3 Codex",
|
|
@@ -6203,23 +6494,6 @@ export const MODELS = {
|
|
|
6203
6494
|
contextWindow: 1000000,
|
|
6204
6495
|
maxTokens: 128000,
|
|
6205
6496
|
},
|
|
6206
|
-
"qwen3.6-plus-free": {
|
|
6207
|
-
id: "qwen3.6-plus-free",
|
|
6208
|
-
name: "Qwen3.6 Plus Free",
|
|
6209
|
-
api: "openai-completions",
|
|
6210
|
-
provider: "opencode",
|
|
6211
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6212
|
-
reasoning: true,
|
|
6213
|
-
input: ["text"],
|
|
6214
|
-
cost: {
|
|
6215
|
-
input: 0,
|
|
6216
|
-
output: 0,
|
|
6217
|
-
cacheRead: 0,
|
|
6218
|
-
cacheWrite: 0,
|
|
6219
|
-
},
|
|
6220
|
-
contextWindow: 1048576,
|
|
6221
|
-
maxTokens: 64000,
|
|
6222
|
-
},
|
|
6223
6497
|
},
|
|
6224
6498
|
"opencode-go": {
|
|
6225
6499
|
"glm-5": {
|
|
@@ -6428,6 +6702,23 @@ export const MODELS = {
|
|
|
6428
6702
|
contextWindow: 300000,
|
|
6429
6703
|
maxTokens: 5120,
|
|
6430
6704
|
},
|
|
6705
|
+
"anthropic/claude-3-haiku": {
|
|
6706
|
+
id: "anthropic/claude-3-haiku",
|
|
6707
|
+
name: "Anthropic: Claude 3 Haiku",
|
|
6708
|
+
api: "openai-completions",
|
|
6709
|
+
provider: "openrouter",
|
|
6710
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
6711
|
+
reasoning: false,
|
|
6712
|
+
input: ["text", "image"],
|
|
6713
|
+
cost: {
|
|
6714
|
+
input: 0.25,
|
|
6715
|
+
output: 1.25,
|
|
6716
|
+
cacheRead: 0.03,
|
|
6717
|
+
cacheWrite: 0.3,
|
|
6718
|
+
},
|
|
6719
|
+
contextWindow: 200000,
|
|
6720
|
+
maxTokens: 4096,
|
|
6721
|
+
},
|
|
6431
6722
|
"anthropic/claude-3.5-haiku": {
|
|
6432
6723
|
id: "anthropic/claude-3.5-haiku",
|
|
6433
6724
|
name: "Anthropic: Claude 3.5 Haiku",
|
|
@@ -7159,6 +7450,23 @@ export const MODELS = {
|
|
|
7159
7450
|
contextWindow: 1048576,
|
|
7160
7451
|
maxTokens: 65536,
|
|
7161
7452
|
},
|
|
7453
|
+
"google/gemini-3-pro-preview": {
|
|
7454
|
+
id: "google/gemini-3-pro-preview",
|
|
7455
|
+
name: "Google: Gemini 3 Pro Preview",
|
|
7456
|
+
api: "openai-completions",
|
|
7457
|
+
provider: "openrouter",
|
|
7458
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7459
|
+
reasoning: true,
|
|
7460
|
+
input: ["text", "image"],
|
|
7461
|
+
cost: {
|
|
7462
|
+
input: 2,
|
|
7463
|
+
output: 12,
|
|
7464
|
+
cacheRead: 0.19999999999999998,
|
|
7465
|
+
cacheWrite: 0.375,
|
|
7466
|
+
},
|
|
7467
|
+
contextWindow: 1048576,
|
|
7468
|
+
maxTokens: 65536,
|
|
7469
|
+
},
|
|
7162
7470
|
"google/gemini-3.1-flash-lite-preview": {
|
|
7163
7471
|
id: "google/gemini-3.1-flash-lite-preview",
|
|
7164
7472
|
name: "Google: Gemini 3.1 Flash Lite Preview",
|
|
@@ -7261,22 +7569,22 @@ export const MODELS = {
|
|
|
7261
7569
|
contextWindow: 128000,
|
|
7262
7570
|
maxTokens: 32000,
|
|
7263
7571
|
},
|
|
7264
|
-
"kwaipilot/kat-coder-pro
|
|
7265
|
-
id: "kwaipilot/kat-coder-pro
|
|
7266
|
-
name: "Kwaipilot: KAT-Coder-Pro
|
|
7572
|
+
"kwaipilot/kat-coder-pro": {
|
|
7573
|
+
id: "kwaipilot/kat-coder-pro",
|
|
7574
|
+
name: "Kwaipilot: KAT-Coder-Pro V1",
|
|
7267
7575
|
api: "openai-completions",
|
|
7268
7576
|
provider: "openrouter",
|
|
7269
7577
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7270
7578
|
reasoning: false,
|
|
7271
7579
|
input: ["text"],
|
|
7272
7580
|
cost: {
|
|
7273
|
-
input: 0.
|
|
7274
|
-
output:
|
|
7275
|
-
cacheRead: 0.
|
|
7581
|
+
input: 0.207,
|
|
7582
|
+
output: 0.828,
|
|
7583
|
+
cacheRead: 0.0414,
|
|
7276
7584
|
cacheWrite: 0,
|
|
7277
7585
|
},
|
|
7278
7586
|
contextWindow: 256000,
|
|
7279
|
-
maxTokens:
|
|
7587
|
+
maxTokens: 128000,
|
|
7280
7588
|
},
|
|
7281
7589
|
"meituan/longcat-flash-chat": {
|
|
7282
7590
|
id: "meituan/longcat-flash-chat",
|
|
@@ -7474,9 +7782,9 @@ export const MODELS = {
|
|
|
7474
7782
|
reasoning: true,
|
|
7475
7783
|
input: ["text"],
|
|
7476
7784
|
cost: {
|
|
7477
|
-
input: 0.
|
|
7478
|
-
output: 1.
|
|
7479
|
-
cacheRead: 0.
|
|
7785
|
+
input: 0.19999999999999998,
|
|
7786
|
+
output: 1.17,
|
|
7787
|
+
cacheRead: 0.09999999999999999,
|
|
7480
7788
|
cacheWrite: 0,
|
|
7481
7789
|
},
|
|
7482
7790
|
contextWindow: 196608,
|
|
@@ -7805,6 +8113,23 @@ export const MODELS = {
|
|
|
7805
8113
|
contextWindow: 262144,
|
|
7806
8114
|
maxTokens: 4096,
|
|
7807
8115
|
},
|
|
8116
|
+
"mistralai/mistral-small-3.1-24b-instruct:free": {
|
|
8117
|
+
id: "mistralai/mistral-small-3.1-24b-instruct:free",
|
|
8118
|
+
name: "Mistral: Mistral Small 3.1 24B (free)",
|
|
8119
|
+
api: "openai-completions",
|
|
8120
|
+
provider: "openrouter",
|
|
8121
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8122
|
+
reasoning: false,
|
|
8123
|
+
input: ["text", "image"],
|
|
8124
|
+
cost: {
|
|
8125
|
+
input: 0,
|
|
8126
|
+
output: 0,
|
|
8127
|
+
cacheRead: 0,
|
|
8128
|
+
cacheWrite: 0,
|
|
8129
|
+
},
|
|
8130
|
+
contextWindow: 128000,
|
|
8131
|
+
maxTokens: 4096,
|
|
8132
|
+
},
|
|
7808
8133
|
"mistralai/mistral-small-3.2-24b-instruct": {
|
|
7809
8134
|
id: "mistralai/mistral-small-3.2-24b-instruct",
|
|
7810
8135
|
name: "Mistral: Mistral Small 3.2 24B",
|
|
@@ -7916,13 +8241,13 @@ export const MODELS = {
|
|
|
7916
8241
|
reasoning: false,
|
|
7917
8242
|
input: ["text"],
|
|
7918
8243
|
cost: {
|
|
7919
|
-
input: 0.
|
|
7920
|
-
output: 2.
|
|
8244
|
+
input: 0.55,
|
|
8245
|
+
output: 2.2,
|
|
7921
8246
|
cacheRead: 0,
|
|
7922
8247
|
cacheWrite: 0,
|
|
7923
8248
|
},
|
|
7924
|
-
contextWindow:
|
|
7925
|
-
maxTokens:
|
|
8249
|
+
contextWindow: 131000,
|
|
8250
|
+
maxTokens: 4096,
|
|
7926
8251
|
},
|
|
7927
8252
|
"moonshotai/kimi-k2-0905": {
|
|
7928
8253
|
id: "moonshotai/kimi-k2-0905",
|
|
@@ -7984,8 +8309,8 @@ export const MODELS = {
|
|
|
7984
8309
|
reasoning: false,
|
|
7985
8310
|
input: ["text"],
|
|
7986
8311
|
cost: {
|
|
7987
|
-
input: 0.
|
|
7988
|
-
output:
|
|
8312
|
+
input: 0.27,
|
|
8313
|
+
output: 1,
|
|
7989
8314
|
cacheRead: 0,
|
|
7990
8315
|
cacheWrite: 0,
|
|
7991
8316
|
},
|
|
@@ -8071,7 +8396,7 @@ export const MODELS = {
|
|
|
8071
8396
|
cost: {
|
|
8072
8397
|
input: 0.09999999999999999,
|
|
8073
8398
|
output: 0.5,
|
|
8074
|
-
cacheRead: 0.
|
|
8399
|
+
cacheRead: 0.04,
|
|
8075
8400
|
cacheWrite: 0,
|
|
8076
8401
|
},
|
|
8077
8402
|
contextWindow: 262144,
|
|
@@ -8145,6 +8470,57 @@ export const MODELS = {
|
|
|
8145
8470
|
contextWindow: 128000,
|
|
8146
8471
|
maxTokens: 4096,
|
|
8147
8472
|
},
|
|
8473
|
+
"openai/gpt-3.5-turbo": {
|
|
8474
|
+
id: "openai/gpt-3.5-turbo",
|
|
8475
|
+
name: "OpenAI: GPT-3.5 Turbo",
|
|
8476
|
+
api: "openai-completions",
|
|
8477
|
+
provider: "openrouter",
|
|
8478
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8479
|
+
reasoning: false,
|
|
8480
|
+
input: ["text"],
|
|
8481
|
+
cost: {
|
|
8482
|
+
input: 0.5,
|
|
8483
|
+
output: 1.5,
|
|
8484
|
+
cacheRead: 0,
|
|
8485
|
+
cacheWrite: 0,
|
|
8486
|
+
},
|
|
8487
|
+
contextWindow: 16385,
|
|
8488
|
+
maxTokens: 4096,
|
|
8489
|
+
},
|
|
8490
|
+
"openai/gpt-3.5-turbo-0613": {
|
|
8491
|
+
id: "openai/gpt-3.5-turbo-0613",
|
|
8492
|
+
name: "OpenAI: GPT-3.5 Turbo (older v0613)",
|
|
8493
|
+
api: "openai-completions",
|
|
8494
|
+
provider: "openrouter",
|
|
8495
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8496
|
+
reasoning: false,
|
|
8497
|
+
input: ["text"],
|
|
8498
|
+
cost: {
|
|
8499
|
+
input: 1,
|
|
8500
|
+
output: 2,
|
|
8501
|
+
cacheRead: 0,
|
|
8502
|
+
cacheWrite: 0,
|
|
8503
|
+
},
|
|
8504
|
+
contextWindow: 4095,
|
|
8505
|
+
maxTokens: 4096,
|
|
8506
|
+
},
|
|
8507
|
+
"openai/gpt-3.5-turbo-16k": {
|
|
8508
|
+
id: "openai/gpt-3.5-turbo-16k",
|
|
8509
|
+
name: "OpenAI: GPT-3.5 Turbo 16k",
|
|
8510
|
+
api: "openai-completions",
|
|
8511
|
+
provider: "openrouter",
|
|
8512
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8513
|
+
reasoning: false,
|
|
8514
|
+
input: ["text"],
|
|
8515
|
+
cost: {
|
|
8516
|
+
input: 3,
|
|
8517
|
+
output: 4,
|
|
8518
|
+
cacheRead: 0,
|
|
8519
|
+
cacheWrite: 0,
|
|
8520
|
+
},
|
|
8521
|
+
contextWindow: 16385,
|
|
8522
|
+
maxTokens: 4096,
|
|
8523
|
+
},
|
|
8148
8524
|
"openai/gpt-4": {
|
|
8149
8525
|
id: "openai/gpt-4",
|
|
8150
8526
|
name: "OpenAI: GPT-4",
|
|
@@ -8196,6 +8572,40 @@ export const MODELS = {
|
|
|
8196
8572
|
contextWindow: 128000,
|
|
8197
8573
|
maxTokens: 4096,
|
|
8198
8574
|
},
|
|
8575
|
+
"openai/gpt-4-turbo": {
|
|
8576
|
+
id: "openai/gpt-4-turbo",
|
|
8577
|
+
name: "OpenAI: GPT-4 Turbo",
|
|
8578
|
+
api: "openai-completions",
|
|
8579
|
+
provider: "openrouter",
|
|
8580
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8581
|
+
reasoning: false,
|
|
8582
|
+
input: ["text", "image"],
|
|
8583
|
+
cost: {
|
|
8584
|
+
input: 10,
|
|
8585
|
+
output: 30,
|
|
8586
|
+
cacheRead: 0,
|
|
8587
|
+
cacheWrite: 0,
|
|
8588
|
+
},
|
|
8589
|
+
contextWindow: 128000,
|
|
8590
|
+
maxTokens: 4096,
|
|
8591
|
+
},
|
|
8592
|
+
"openai/gpt-4-turbo-preview": {
|
|
8593
|
+
id: "openai/gpt-4-turbo-preview",
|
|
8594
|
+
name: "OpenAI: GPT-4 Turbo Preview",
|
|
8595
|
+
api: "openai-completions",
|
|
8596
|
+
provider: "openrouter",
|
|
8597
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8598
|
+
reasoning: false,
|
|
8599
|
+
input: ["text"],
|
|
8600
|
+
cost: {
|
|
8601
|
+
input: 10,
|
|
8602
|
+
output: 30,
|
|
8603
|
+
cacheRead: 0,
|
|
8604
|
+
cacheWrite: 0,
|
|
8605
|
+
},
|
|
8606
|
+
contextWindow: 128000,
|
|
8607
|
+
maxTokens: 4096,
|
|
8608
|
+
},
|
|
8199
8609
|
"openai/gpt-4.1": {
|
|
8200
8610
|
id: "openai/gpt-4.1",
|
|
8201
8611
|
name: "OpenAI: GPT-4.1",
|
|
@@ -8741,55 +9151,21 @@ export const MODELS = {
|
|
|
8741
9151
|
maxTokens: 128000,
|
|
8742
9152
|
},
|
|
8743
9153
|
"openai/gpt-5.4-pro": {
|
|
8744
|
-
id: "openai/gpt-5.4-pro",
|
|
8745
|
-
name: "OpenAI: GPT-5.4 Pro",
|
|
8746
|
-
api: "openai-completions",
|
|
8747
|
-
provider: "openrouter",
|
|
8748
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8749
|
-
reasoning: true,
|
|
8750
|
-
input: ["text", "image"],
|
|
8751
|
-
cost: {
|
|
8752
|
-
input: 30,
|
|
8753
|
-
output: 180,
|
|
8754
|
-
cacheRead: 0,
|
|
8755
|
-
cacheWrite: 0,
|
|
8756
|
-
},
|
|
8757
|
-
contextWindow: 1050000,
|
|
8758
|
-
maxTokens: 128000,
|
|
8759
|
-
},
|
|
8760
|
-
"openai/gpt-audio": {
|
|
8761
|
-
id: "openai/gpt-audio",
|
|
8762
|
-
name: "OpenAI: GPT Audio",
|
|
8763
|
-
api: "openai-completions",
|
|
8764
|
-
provider: "openrouter",
|
|
8765
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8766
|
-
reasoning: false,
|
|
8767
|
-
input: ["text"],
|
|
8768
|
-
cost: {
|
|
8769
|
-
input: 2.5,
|
|
8770
|
-
output: 10,
|
|
8771
|
-
cacheRead: 0,
|
|
8772
|
-
cacheWrite: 0,
|
|
8773
|
-
},
|
|
8774
|
-
contextWindow: 128000,
|
|
8775
|
-
maxTokens: 16384,
|
|
8776
|
-
},
|
|
8777
|
-
"openai/gpt-audio-mini": {
|
|
8778
|
-
id: "openai/gpt-audio-mini",
|
|
8779
|
-
name: "OpenAI: GPT Audio Mini",
|
|
9154
|
+
id: "openai/gpt-5.4-pro",
|
|
9155
|
+
name: "OpenAI: GPT-5.4 Pro",
|
|
8780
9156
|
api: "openai-completions",
|
|
8781
9157
|
provider: "openrouter",
|
|
8782
9158
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
8783
|
-
reasoning:
|
|
8784
|
-
input: ["text"],
|
|
9159
|
+
reasoning: true,
|
|
9160
|
+
input: ["text", "image"],
|
|
8785
9161
|
cost: {
|
|
8786
|
-
input:
|
|
8787
|
-
output:
|
|
9162
|
+
input: 30,
|
|
9163
|
+
output: 180,
|
|
8788
9164
|
cacheRead: 0,
|
|
8789
9165
|
cacheWrite: 0,
|
|
8790
9166
|
},
|
|
8791
|
-
contextWindow:
|
|
8792
|
-
maxTokens:
|
|
9167
|
+
contextWindow: 1050000,
|
|
9168
|
+
maxTokens: 128000,
|
|
8793
9169
|
},
|
|
8794
9170
|
"openai/gpt-oss-120b": {
|
|
8795
9171
|
id: "openai/gpt-oss-120b",
|
|
@@ -9112,7 +9488,7 @@ export const MODELS = {
|
|
|
9112
9488
|
cacheWrite: 0,
|
|
9113
9489
|
},
|
|
9114
9490
|
contextWindow: 32768,
|
|
9115
|
-
maxTokens:
|
|
9491
|
+
maxTokens: 4096,
|
|
9116
9492
|
},
|
|
9117
9493
|
"qwen/qwen-max": {
|
|
9118
9494
|
id: "qwen/qwen-max",
|
|
@@ -9352,6 +9728,23 @@ export const MODELS = {
|
|
|
9352
9728
|
contextWindow: 40960,
|
|
9353
9729
|
maxTokens: 40960,
|
|
9354
9730
|
},
|
|
9731
|
+
"qwen/qwen3-4b:free": {
|
|
9732
|
+
id: "qwen/qwen3-4b:free",
|
|
9733
|
+
name: "Qwen: Qwen3 4B (free)",
|
|
9734
|
+
api: "openai-completions",
|
|
9735
|
+
provider: "openrouter",
|
|
9736
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9737
|
+
reasoning: true,
|
|
9738
|
+
input: ["text"],
|
|
9739
|
+
cost: {
|
|
9740
|
+
input: 0,
|
|
9741
|
+
output: 0,
|
|
9742
|
+
cacheRead: 0,
|
|
9743
|
+
cacheWrite: 0,
|
|
9744
|
+
},
|
|
9745
|
+
contextWindow: 40960,
|
|
9746
|
+
maxTokens: 4096,
|
|
9747
|
+
},
|
|
9355
9748
|
"qwen/qwen3-8b": {
|
|
9356
9749
|
id: "qwen/qwen3-8b",
|
|
9357
9750
|
name: "Qwen: Qwen3 8B",
|
|
@@ -9758,7 +10151,7 @@ export const MODELS = {
|
|
|
9758
10151
|
cacheWrite: 0,
|
|
9759
10152
|
},
|
|
9760
10153
|
contextWindow: 256000,
|
|
9761
|
-
maxTokens:
|
|
10154
|
+
maxTokens: 4096,
|
|
9762
10155
|
},
|
|
9763
10156
|
"qwen/qwen3.5-flash-02-23": {
|
|
9764
10157
|
id: "qwen/qwen3.5-flash-02-23",
|
|
@@ -9794,23 +10187,6 @@ export const MODELS = {
|
|
|
9794
10187
|
contextWindow: 1000000,
|
|
9795
10188
|
maxTokens: 65536,
|
|
9796
10189
|
},
|
|
9797
|
-
"qwen/qwen3.6-plus-preview:free": {
|
|
9798
|
-
id: "qwen/qwen3.6-plus-preview:free",
|
|
9799
|
-
name: "Qwen: Qwen3.6 Plus Preview (free)",
|
|
9800
|
-
api: "openai-completions",
|
|
9801
|
-
provider: "openrouter",
|
|
9802
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9803
|
-
reasoning: true,
|
|
9804
|
-
input: ["text"],
|
|
9805
|
-
cost: {
|
|
9806
|
-
input: 0,
|
|
9807
|
-
output: 0,
|
|
9808
|
-
cacheRead: 0,
|
|
9809
|
-
cacheWrite: 0,
|
|
9810
|
-
},
|
|
9811
|
-
contextWindow: 1000000,
|
|
9812
|
-
maxTokens: 32000,
|
|
9813
|
-
},
|
|
9814
10190
|
"qwen/qwq-32b": {
|
|
9815
10191
|
id: "qwen/qwq-32b",
|
|
9816
10192
|
name: "Qwen: QwQ 32B",
|
|
@@ -9828,23 +10204,6 @@ export const MODELS = {
|
|
|
9828
10204
|
contextWindow: 131072,
|
|
9829
10205
|
maxTokens: 131072,
|
|
9830
10206
|
},
|
|
9831
|
-
"reka/reka-edge": {
|
|
9832
|
-
id: "reka/reka-edge",
|
|
9833
|
-
name: "Reka Edge",
|
|
9834
|
-
api: "openai-completions",
|
|
9835
|
-
provider: "openrouter",
|
|
9836
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9837
|
-
reasoning: false,
|
|
9838
|
-
input: ["text", "image"],
|
|
9839
|
-
cost: {
|
|
9840
|
-
input: 0.09999999999999999,
|
|
9841
|
-
output: 0.09999999999999999,
|
|
9842
|
-
cacheRead: 0,
|
|
9843
|
-
cacheWrite: 0,
|
|
9844
|
-
},
|
|
9845
|
-
contextWindow: 16384,
|
|
9846
|
-
maxTokens: 16384,
|
|
9847
|
-
},
|
|
9848
10207
|
"relace/relace-search": {
|
|
9849
10208
|
id: "relace/relace-search",
|
|
9850
10209
|
name: "Relace: Relace Search",
|
|
@@ -9907,11 +10266,11 @@ export const MODELS = {
|
|
|
9907
10266
|
cost: {
|
|
9908
10267
|
input: 0.09999999999999999,
|
|
9909
10268
|
output: 0.3,
|
|
9910
|
-
cacheRead: 0,
|
|
10269
|
+
cacheRead: 0.02,
|
|
9911
10270
|
cacheWrite: 0,
|
|
9912
10271
|
},
|
|
9913
|
-
contextWindow:
|
|
9914
|
-
maxTokens:
|
|
10272
|
+
contextWindow: 256000,
|
|
10273
|
+
maxTokens: 256000,
|
|
9915
10274
|
},
|
|
9916
10275
|
"stepfun/step-3.5-flash:free": {
|
|
9917
10276
|
id: "stepfun/step-3.5-flash:free",
|
|
@@ -10117,23 +10476,6 @@ export const MODELS = {
|
|
|
10117
10476
|
contextWindow: 2000000,
|
|
10118
10477
|
maxTokens: 30000,
|
|
10119
10478
|
},
|
|
10120
|
-
"x-ai/grok-4.20": {
|
|
10121
|
-
id: "x-ai/grok-4.20",
|
|
10122
|
-
name: "xAI: Grok 4.20",
|
|
10123
|
-
api: "openai-completions",
|
|
10124
|
-
provider: "openrouter",
|
|
10125
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
10126
|
-
reasoning: true,
|
|
10127
|
-
input: ["text", "image"],
|
|
10128
|
-
cost: {
|
|
10129
|
-
input: 2,
|
|
10130
|
-
output: 6,
|
|
10131
|
-
cacheRead: 0.19999999999999998,
|
|
10132
|
-
cacheWrite: 0,
|
|
10133
|
-
},
|
|
10134
|
-
contextWindow: 2000000,
|
|
10135
|
-
maxTokens: 4096,
|
|
10136
|
-
},
|
|
10137
10479
|
"x-ai/grok-4.20-beta": {
|
|
10138
10480
|
id: "x-ai/grok-4.20-beta",
|
|
10139
10481
|
name: "xAI: Grok 4.20 Beta",
|
|
@@ -10398,9 +10740,9 @@ export const MODELS = {
|
|
|
10398
10740
|
reasoning: true,
|
|
10399
10741
|
input: ["text"],
|
|
10400
10742
|
cost: {
|
|
10401
|
-
input:
|
|
10402
|
-
output:
|
|
10403
|
-
cacheRead: 0.
|
|
10743
|
+
input: 0.96,
|
|
10744
|
+
output: 3.1999999999999997,
|
|
10745
|
+
cacheRead: 0.192,
|
|
10404
10746
|
cacheWrite: 0,
|
|
10405
10747
|
},
|
|
10406
10748
|
contextWindow: 202752,
|
|
@@ -10434,12 +10776,12 @@ export const MODELS = {
|
|
|
10434
10776
|
reasoning: false,
|
|
10435
10777
|
input: ["text"],
|
|
10436
10778
|
cost: {
|
|
10437
|
-
input: 0.
|
|
10438
|
-
output: 0.
|
|
10439
|
-
cacheRead: 0
|
|
10779
|
+
input: 0.071,
|
|
10780
|
+
output: 0.463,
|
|
10781
|
+
cacheRead: 0,
|
|
10440
10782
|
cacheWrite: 0,
|
|
10441
10783
|
},
|
|
10442
|
-
contextWindow:
|
|
10784
|
+
contextWindow: 40960,
|
|
10443
10785
|
maxTokens: 16384,
|
|
10444
10786
|
},
|
|
10445
10787
|
"alibaba/qwen-3-30b": {
|
|
@@ -10468,13 +10810,13 @@ export const MODELS = {
|
|
|
10468
10810
|
reasoning: true,
|
|
10469
10811
|
input: ["text"],
|
|
10470
10812
|
cost: {
|
|
10471
|
-
input: 0.
|
|
10472
|
-
output: 0.
|
|
10473
|
-
cacheRead: 0,
|
|
10813
|
+
input: 0.29,
|
|
10814
|
+
output: 0.59,
|
|
10815
|
+
cacheRead: 0.145,
|
|
10474
10816
|
cacheWrite: 0,
|
|
10475
10817
|
},
|
|
10476
|
-
contextWindow:
|
|
10477
|
-
maxTokens:
|
|
10818
|
+
contextWindow: 131072,
|
|
10819
|
+
maxTokens: 40960,
|
|
10478
10820
|
},
|
|
10479
10821
|
"alibaba/qwen3-235b-a22b-thinking": {
|
|
10480
10822
|
id: "alibaba/qwen3-235b-a22b-thinking",
|
|
@@ -10502,13 +10844,13 @@ export const MODELS = {
|
|
|
10502
10844
|
reasoning: false,
|
|
10503
10845
|
input: ["text"],
|
|
10504
10846
|
cost: {
|
|
10505
|
-
input:
|
|
10506
|
-
output:
|
|
10507
|
-
cacheRead: 0.
|
|
10847
|
+
input: 0.39999999999999997,
|
|
10848
|
+
output: 1.5999999999999999,
|
|
10849
|
+
cacheRead: 0.022,
|
|
10508
10850
|
cacheWrite: 0,
|
|
10509
10851
|
},
|
|
10510
10852
|
contextWindow: 262144,
|
|
10511
|
-
maxTokens:
|
|
10853
|
+
maxTokens: 66536,
|
|
10512
10854
|
},
|
|
10513
10855
|
"alibaba/qwen3-coder-30b-a3b": {
|
|
10514
10856
|
id: "alibaba/qwen3-coder-30b-a3b",
|
|
@@ -10663,6 +11005,23 @@ export const MODELS = {
|
|
|
10663
11005
|
contextWindow: 1000000,
|
|
10664
11006
|
maxTokens: 64000,
|
|
10665
11007
|
},
|
|
11008
|
+
"anthropic/claude-3-haiku": {
|
|
11009
|
+
id: "anthropic/claude-3-haiku",
|
|
11010
|
+
name: "Claude 3 Haiku",
|
|
11011
|
+
api: "anthropic-messages",
|
|
11012
|
+
provider: "vercel-ai-gateway",
|
|
11013
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11014
|
+
reasoning: false,
|
|
11015
|
+
input: ["text", "image"],
|
|
11016
|
+
cost: {
|
|
11017
|
+
input: 0.25,
|
|
11018
|
+
output: 1.25,
|
|
11019
|
+
cacheRead: 0.03,
|
|
11020
|
+
cacheWrite: 0.3,
|
|
11021
|
+
},
|
|
11022
|
+
contextWindow: 200000,
|
|
11023
|
+
maxTokens: 4096,
|
|
11024
|
+
},
|
|
10666
11025
|
"anthropic/claude-3.5-haiku": {
|
|
10667
11026
|
id: "anthropic/claude-3.5-haiku",
|
|
10668
11027
|
name: "Claude 3.5 Haiku",
|
|
@@ -10961,13 +11320,13 @@ export const MODELS = {
|
|
|
10961
11320
|
reasoning: true,
|
|
10962
11321
|
input: ["text"],
|
|
10963
11322
|
cost: {
|
|
10964
|
-
input: 0.
|
|
10965
|
-
output: 1.
|
|
10966
|
-
cacheRead: 0
|
|
11323
|
+
input: 0.5,
|
|
11324
|
+
output: 1.5,
|
|
11325
|
+
cacheRead: 0,
|
|
10967
11326
|
cacheWrite: 0,
|
|
10968
11327
|
},
|
|
10969
11328
|
contextWindow: 163840,
|
|
10970
|
-
maxTokens:
|
|
11329
|
+
maxTokens: 16384,
|
|
10971
11330
|
},
|
|
10972
11331
|
"deepseek/deepseek-v3.1-terminus": {
|
|
10973
11332
|
id: "deepseek/deepseek-v3.1-terminus",
|
|
@@ -11207,23 +11566,6 @@ export const MODELS = {
|
|
|
11207
11566
|
contextWindow: 32000,
|
|
11208
11567
|
maxTokens: 16384,
|
|
11209
11568
|
},
|
|
11210
|
-
"kwaipilot/kat-coder-pro-v2": {
|
|
11211
|
-
id: "kwaipilot/kat-coder-pro-v2",
|
|
11212
|
-
name: "Kat Coder Pro V2",
|
|
11213
|
-
api: "anthropic-messages",
|
|
11214
|
-
provider: "vercel-ai-gateway",
|
|
11215
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11216
|
-
reasoning: true,
|
|
11217
|
-
input: ["text"],
|
|
11218
|
-
cost: {
|
|
11219
|
-
input: 0.3,
|
|
11220
|
-
output: 1.2,
|
|
11221
|
-
cacheRead: 0.06,
|
|
11222
|
-
cacheWrite: 0,
|
|
11223
|
-
},
|
|
11224
|
-
contextWindow: 256000,
|
|
11225
|
-
maxTokens: 256000,
|
|
11226
|
-
},
|
|
11227
11569
|
"meituan/longcat-flash-chat": {
|
|
11228
11570
|
id: "meituan/longcat-flash-chat",
|
|
11229
11571
|
name: "LongCat Flash Chat",
|
|
@@ -11239,7 +11581,7 @@ export const MODELS = {
|
|
|
11239
11581
|
cacheWrite: 0,
|
|
11240
11582
|
},
|
|
11241
11583
|
contextWindow: 128000,
|
|
11242
|
-
maxTokens:
|
|
11584
|
+
maxTokens: 100000,
|
|
11243
11585
|
},
|
|
11244
11586
|
"meituan/longcat-flash-thinking": {
|
|
11245
11587
|
id: "meituan/longcat-flash-thinking",
|
|
@@ -11284,13 +11626,13 @@ export const MODELS = {
|
|
|
11284
11626
|
reasoning: false,
|
|
11285
11627
|
input: ["text"],
|
|
11286
11628
|
cost: {
|
|
11287
|
-
input: 0.
|
|
11288
|
-
output: 0.
|
|
11289
|
-
cacheRead: 0,
|
|
11629
|
+
input: 0.09999999999999999,
|
|
11630
|
+
output: 0.09999999999999999,
|
|
11631
|
+
cacheRead: 0.09999999999999999,
|
|
11290
11632
|
cacheWrite: 0,
|
|
11291
11633
|
},
|
|
11292
11634
|
contextWindow: 128000,
|
|
11293
|
-
maxTokens:
|
|
11635
|
+
maxTokens: 16384,
|
|
11294
11636
|
},
|
|
11295
11637
|
"meta/llama-3.2-11b": {
|
|
11296
11638
|
id: "meta/llama-3.2-11b",
|
|
@@ -11352,12 +11694,12 @@ export const MODELS = {
|
|
|
11352
11694
|
reasoning: false,
|
|
11353
11695
|
input: ["text", "image"],
|
|
11354
11696
|
cost: {
|
|
11355
|
-
input: 0.
|
|
11356
|
-
output:
|
|
11697
|
+
input: 0.24,
|
|
11698
|
+
output: 0.9700000000000001,
|
|
11357
11699
|
cacheRead: 0,
|
|
11358
11700
|
cacheWrite: 0,
|
|
11359
11701
|
},
|
|
11360
|
-
contextWindow:
|
|
11702
|
+
contextWindow: 128000,
|
|
11361
11703
|
maxTokens: 8192,
|
|
11362
11704
|
},
|
|
11363
11705
|
"meta/llama-4-scout": {
|
|
@@ -11802,6 +12144,23 @@ export const MODELS = {
|
|
|
11802
12144
|
contextWindow: 131072,
|
|
11803
12145
|
maxTokens: 131072,
|
|
11804
12146
|
},
|
|
12147
|
+
"openai/gpt-4-turbo": {
|
|
12148
|
+
id: "openai/gpt-4-turbo",
|
|
12149
|
+
name: "GPT-4 Turbo",
|
|
12150
|
+
api: "anthropic-messages",
|
|
12151
|
+
provider: "vercel-ai-gateway",
|
|
12152
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12153
|
+
reasoning: false,
|
|
12154
|
+
input: ["text", "image"],
|
|
12155
|
+
cost: {
|
|
12156
|
+
input: 10,
|
|
12157
|
+
output: 30,
|
|
12158
|
+
cacheRead: 0,
|
|
12159
|
+
cacheWrite: 0,
|
|
12160
|
+
},
|
|
12161
|
+
contextWindow: 128000,
|
|
12162
|
+
maxTokens: 4096,
|
|
12163
|
+
},
|
|
11805
12164
|
"openai/gpt-4.1": {
|
|
11806
12165
|
id: "openai/gpt-4.1",
|
|
11807
12166
|
name: "GPT-4.1",
|
|
@@ -12244,23 +12603,6 @@ export const MODELS = {
|
|
|
12244
12603
|
contextWindow: 1050000,
|
|
12245
12604
|
maxTokens: 128000,
|
|
12246
12605
|
},
|
|
12247
|
-
"openai/gpt-oss-120b": {
|
|
12248
|
-
id: "openai/gpt-oss-120b",
|
|
12249
|
-
name: "gpt-oss-120b",
|
|
12250
|
-
api: "anthropic-messages",
|
|
12251
|
-
provider: "vercel-ai-gateway",
|
|
12252
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12253
|
-
reasoning: true,
|
|
12254
|
-
input: ["text"],
|
|
12255
|
-
cost: {
|
|
12256
|
-
input: 0.15,
|
|
12257
|
-
output: 0.6,
|
|
12258
|
-
cacheRead: 0,
|
|
12259
|
-
cacheWrite: 0,
|
|
12260
|
-
},
|
|
12261
|
-
contextWindow: 128000,
|
|
12262
|
-
maxTokens: 8192,
|
|
12263
|
-
},
|
|
12264
12606
|
"openai/gpt-oss-20b": {
|
|
12265
12607
|
id: "openai/gpt-oss-20b",
|
|
12266
12608
|
name: "gpt-oss-20b",
|
|
@@ -12406,8 +12748,8 @@ export const MODELS = {
|
|
|
12406
12748
|
reasoning: false,
|
|
12407
12749
|
input: ["text", "image"],
|
|
12408
12750
|
cost: {
|
|
12409
|
-
input:
|
|
12410
|
-
output:
|
|
12751
|
+
input: 1,
|
|
12752
|
+
output: 1,
|
|
12411
12753
|
cacheRead: 0,
|
|
12412
12754
|
cacheWrite: 0,
|
|
12413
12755
|
},
|
|
@@ -12423,8 +12765,8 @@ export const MODELS = {
|
|
|
12423
12765
|
reasoning: false,
|
|
12424
12766
|
input: ["text", "image"],
|
|
12425
12767
|
cost: {
|
|
12426
|
-
input:
|
|
12427
|
-
output:
|
|
12768
|
+
input: 3,
|
|
12769
|
+
output: 15,
|
|
12428
12770
|
cacheRead: 0,
|
|
12429
12771
|
cacheWrite: 0,
|
|
12430
12772
|
},
|
|
@@ -12618,23 +12960,6 @@ export const MODELS = {
|
|
|
12618
12960
|
contextWindow: 2000000,
|
|
12619
12961
|
maxTokens: 30000,
|
|
12620
12962
|
},
|
|
12621
|
-
"xai/grok-4.20-multi-agent": {
|
|
12622
|
-
id: "xai/grok-4.20-multi-agent",
|
|
12623
|
-
name: "Grok 4.20 Multi-Agent",
|
|
12624
|
-
api: "anthropic-messages",
|
|
12625
|
-
provider: "vercel-ai-gateway",
|
|
12626
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12627
|
-
reasoning: true,
|
|
12628
|
-
input: ["text"],
|
|
12629
|
-
cost: {
|
|
12630
|
-
input: 2,
|
|
12631
|
-
output: 6,
|
|
12632
|
-
cacheRead: 0.19999999999999998,
|
|
12633
|
-
cacheWrite: 0,
|
|
12634
|
-
},
|
|
12635
|
-
contextWindow: 2000000,
|
|
12636
|
-
maxTokens: 2000000,
|
|
12637
|
-
},
|
|
12638
12963
|
"xai/grok-4.20-multi-agent-beta": {
|
|
12639
12964
|
id: "xai/grok-4.20-multi-agent-beta",
|
|
12640
12965
|
name: "Grok 4.20 Multi Agent Beta",
|
|
@@ -12652,23 +12977,6 @@ export const MODELS = {
|
|
|
12652
12977
|
contextWindow: 2000000,
|
|
12653
12978
|
maxTokens: 2000000,
|
|
12654
12979
|
},
|
|
12655
|
-
"xai/grok-4.20-non-reasoning": {
|
|
12656
|
-
id: "xai/grok-4.20-non-reasoning",
|
|
12657
|
-
name: "Grok 4.20 Non-Reasoning",
|
|
12658
|
-
api: "anthropic-messages",
|
|
12659
|
-
provider: "vercel-ai-gateway",
|
|
12660
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12661
|
-
reasoning: false,
|
|
12662
|
-
input: ["text", "image"],
|
|
12663
|
-
cost: {
|
|
12664
|
-
input: 2,
|
|
12665
|
-
output: 6,
|
|
12666
|
-
cacheRead: 0.19999999999999998,
|
|
12667
|
-
cacheWrite: 0,
|
|
12668
|
-
},
|
|
12669
|
-
contextWindow: 2000000,
|
|
12670
|
-
maxTokens: 2000000,
|
|
12671
|
-
},
|
|
12672
12980
|
"xai/grok-4.20-non-reasoning-beta": {
|
|
12673
12981
|
id: "xai/grok-4.20-non-reasoning-beta",
|
|
12674
12982
|
name: "Grok 4.20 Beta Non-Reasoning",
|
|
@@ -12686,23 +12994,6 @@ export const MODELS = {
|
|
|
12686
12994
|
contextWindow: 2000000,
|
|
12687
12995
|
maxTokens: 2000000,
|
|
12688
12996
|
},
|
|
12689
|
-
"xai/grok-4.20-reasoning": {
|
|
12690
|
-
id: "xai/grok-4.20-reasoning",
|
|
12691
|
-
name: "Grok 4.20 Reasoning",
|
|
12692
|
-
api: "anthropic-messages",
|
|
12693
|
-
provider: "vercel-ai-gateway",
|
|
12694
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12695
|
-
reasoning: true,
|
|
12696
|
-
input: ["text", "image"],
|
|
12697
|
-
cost: {
|
|
12698
|
-
input: 2,
|
|
12699
|
-
output: 6,
|
|
12700
|
-
cacheRead: 0.19999999999999998,
|
|
12701
|
-
cacheWrite: 0,
|
|
12702
|
-
},
|
|
12703
|
-
contextWindow: 2000000,
|
|
12704
|
-
maxTokens: 2000000,
|
|
12705
|
-
},
|
|
12706
12997
|
"xai/grok-4.20-reasoning-beta": {
|
|
12707
12998
|
id: "xai/grok-4.20-reasoning-beta",
|
|
12708
12999
|
name: "Grok 4.20 Beta Reasoning",
|
|
@@ -12746,9 +13037,9 @@ export const MODELS = {
|
|
|
12746
13037
|
reasoning: true,
|
|
12747
13038
|
input: ["text"],
|
|
12748
13039
|
cost: {
|
|
12749
|
-
input: 0.
|
|
12750
|
-
output: 0.
|
|
12751
|
-
cacheRead: 0.
|
|
13040
|
+
input: 0.09999999999999999,
|
|
13041
|
+
output: 0.3,
|
|
13042
|
+
cacheRead: 0.02,
|
|
12752
13043
|
cacheWrite: 0,
|
|
12753
13044
|
},
|
|
12754
13045
|
contextWindow: 262144,
|
|
@@ -12884,7 +13175,7 @@ export const MODELS = {
|
|
|
12884
13175
|
cost: {
|
|
12885
13176
|
input: 0.6,
|
|
12886
13177
|
output: 2.2,
|
|
12887
|
-
cacheRead: 0
|
|
13178
|
+
cacheRead: 0,
|
|
12888
13179
|
cacheWrite: 0,
|
|
12889
13180
|
},
|
|
12890
13181
|
contextWindow: 200000,
|
|
@@ -13514,24 +13805,6 @@ export const MODELS = {
|
|
|
13514
13805
|
contextWindow: 200000,
|
|
13515
13806
|
maxTokens: 131072,
|
|
13516
13807
|
},
|
|
13517
|
-
"glm-4.7-flashx": {
|
|
13518
|
-
id: "glm-4.7-flashx",
|
|
13519
|
-
name: "GLM-4.7-FlashX",
|
|
13520
|
-
api: "openai-completions",
|
|
13521
|
-
provider: "zai",
|
|
13522
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13523
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
13524
|
-
reasoning: true,
|
|
13525
|
-
input: ["text"],
|
|
13526
|
-
cost: {
|
|
13527
|
-
input: 0.07,
|
|
13528
|
-
output: 0.4,
|
|
13529
|
-
cacheRead: 0.01,
|
|
13530
|
-
cacheWrite: 0,
|
|
13531
|
-
},
|
|
13532
|
-
contextWindow: 200000,
|
|
13533
|
-
maxTokens: 131072,
|
|
13534
|
-
},
|
|
13535
13808
|
"glm-5": {
|
|
13536
13809
|
id: "glm-5",
|
|
13537
13810
|
name: "GLM-5",
|