lsd-pi 1.1.2 → 1.1.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (183) hide show
  1. package/README.md +2 -1
  2. package/dist/bedrock-auth.d.ts +25 -0
  3. package/dist/bedrock-auth.js +59 -0
  4. package/dist/headless.js +8 -3
  5. package/dist/loader.js +1 -0
  6. package/dist/onboarding-llm.d.ts +37 -0
  7. package/dist/onboarding-llm.js +64 -0
  8. package/dist/onboarding.d.ts +2 -14
  9. package/dist/onboarding.js +146 -71
  10. package/dist/pi-migration.js +1 -0
  11. package/dist/resources/extensions/memory/auto-extract.js +21 -3
  12. package/dist/resources/extensions/memory/dream.js +703 -0
  13. package/dist/resources/extensions/memory/extension-manifest.json +2 -2
  14. package/dist/resources/extensions/memory/index.js +115 -8
  15. package/dist/resources/extensions/slash-commands/extension-manifest.json +10 -10
  16. package/dist/resources/extensions/slash-commands/index.js +0 -4
  17. package/dist/resources/extensions/slash-commands/plan.js +181 -45
  18. package/dist/resources/extensions/subagent/agents.js +14 -1
  19. package/dist/resources/extensions/subagent/configured-model.js +3 -2
  20. package/dist/resources/extensions/subagent/index.js +34 -28
  21. package/dist/resources/extensions/subagent/launch-helpers.js +24 -0
  22. package/dist/resources/extensions/subagent/model-resolution.js +41 -3
  23. package/dist/resources/extensions/usage/extension-manifest.json +11 -0
  24. package/dist/resources/extensions/usage/index.js +346 -0
  25. package/{src/resources/skills/create-gsd-extension → dist/resources/skills/create-lsd-extension}/SKILL.md +6 -6
  26. package/{src/resources/skills/create-gsd-extension → dist/resources/skills/create-lsd-extension}/references/custom-tools.md +1 -1
  27. package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/extension-lifecycle.md +2 -2
  28. package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/extensioncontext-reference.md +1 -1
  29. package/{src/resources/skills/create-gsd-extension → dist/resources/skills/create-lsd-extension}/references/key-rules-gotchas.md +4 -4
  30. package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/packaging-distribution.md +6 -6
  31. package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/workflows/create-extension.md +3 -3
  32. package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/workflows/debug-extension.md +5 -5
  33. package/dist/resources/skills/teams-debug/SKILL.md +5 -6
  34. package/dist/resources/skills/teams-document/SKILL.md +1 -2
  35. package/dist/resources/skills/teams-plan/SKILL.md +3 -4
  36. package/dist/resources/skills/teams-run/SKILL.md +3 -4
  37. package/dist/resources/skills/teams-verify/SKILL.md +4 -5
  38. package/dist/startup-model-validation.js +1 -0
  39. package/dist/welcome-screen.js +13 -11
  40. package/dist/wizard.js +12 -0
  41. package/package.json +1 -1
  42. package/packages/pi-ai/dist/models.generated.d.ts +688 -409
  43. package/packages/pi-ai/dist/models.generated.d.ts.map +1 -1
  44. package/packages/pi-ai/dist/models.generated.js +761 -488
  45. package/packages/pi-ai/dist/models.generated.js.map +1 -1
  46. package/packages/pi-ai/scripts/generate-models.ts +40 -18
  47. package/packages/pi-ai/src/models.generated.ts +759 -486
  48. package/packages/pi-coding-agent/dist/cli/config-selector.js +1 -1
  49. package/packages/pi-coding-agent/dist/cli/config-selector.js.map +1 -1
  50. package/packages/pi-coding-agent/dist/core/agent-session.d.ts +1 -2
  51. package/packages/pi-coding-agent/dist/core/agent-session.d.ts.map +1 -1
  52. package/packages/pi-coding-agent/dist/core/agent-session.js +6 -30
  53. package/packages/pi-coding-agent/dist/core/agent-session.js.map +1 -1
  54. package/packages/pi-coding-agent/dist/core/settings-manager.d.ts +6 -0
  55. package/packages/pi-coding-agent/dist/core/settings-manager.d.ts.map +1 -1
  56. package/packages/pi-coding-agent/dist/core/settings-manager.js +44 -1
  57. package/packages/pi-coding-agent/dist/core/settings-manager.js.map +1 -1
  58. package/packages/pi-coding-agent/dist/core/skill-tool.test.js +9 -5
  59. package/packages/pi-coding-agent/dist/core/skill-tool.test.js.map +1 -1
  60. package/packages/pi-coding-agent/dist/core/skills.d.ts.map +1 -1
  61. package/packages/pi-coding-agent/dist/core/skills.js +3 -2
  62. package/packages/pi-coding-agent/dist/core/skills.js.map +1 -1
  63. package/packages/pi-coding-agent/dist/index.d.ts +1 -1
  64. package/packages/pi-coding-agent/dist/index.d.ts.map +1 -1
  65. package/packages/pi-coding-agent/dist/index.js +1 -1
  66. package/packages/pi-coding-agent/dist/index.js.map +1 -1
  67. package/packages/pi-coding-agent/dist/main.js +1 -1
  68. package/packages/pi-coding-agent/dist/main.js.map +1 -1
  69. package/packages/pi-coding-agent/dist/modes/interactive/components/bash-execution.js +2 -2
  70. package/packages/pi-coding-agent/dist/modes/interactive/components/bash-execution.js.map +1 -1
  71. package/packages/pi-coding-agent/dist/modes/interactive/components/footer.d.ts.map +1 -1
  72. package/packages/pi-coding-agent/dist/modes/interactive/components/footer.js +15 -12
  73. package/packages/pi-coding-agent/dist/modes/interactive/components/footer.js.map +1 -1
  74. package/packages/pi-coding-agent/dist/modes/interactive/components/settings-selector.d.ts +6 -0
  75. package/packages/pi-coding-agent/dist/modes/interactive/components/settings-selector.d.ts.map +1 -1
  76. package/packages/pi-coding-agent/dist/modes/interactive/components/settings-selector.js +25 -1
  77. package/packages/pi-coding-agent/dist/modes/interactive/components/settings-selector.js.map +1 -1
  78. package/packages/pi-coding-agent/dist/modes/interactive/components/tool-execution.js +1 -1
  79. package/packages/pi-coding-agent/dist/modes/interactive/components/tool-execution.js.map +1 -1
  80. package/packages/pi-coding-agent/dist/modes/interactive/controllers/chat-controller.d.ts.map +1 -1
  81. package/packages/pi-coding-agent/dist/modes/interactive/controllers/chat-controller.js +10 -0
  82. package/packages/pi-coding-agent/dist/modes/interactive/controllers/chat-controller.js.map +1 -1
  83. package/packages/pi-coding-agent/dist/modes/interactive/controllers/extension-ui-controller.js +1 -1
  84. package/packages/pi-coding-agent/dist/modes/interactive/controllers/extension-ui-controller.js.map +1 -1
  85. package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.d.ts.map +1 -1
  86. package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.js +31 -22
  87. package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.js.map +1 -1
  88. package/packages/pi-coding-agent/dist/modes/interactive/theme/theme.d.ts +18 -5
  89. package/packages/pi-coding-agent/dist/modes/interactive/theme/theme.d.ts.map +1 -1
  90. package/packages/pi-coding-agent/dist/modes/interactive/theme/theme.js +139 -20
  91. package/packages/pi-coding-agent/dist/modes/interactive/theme/theme.js.map +1 -1
  92. package/packages/pi-coding-agent/dist/modes/interactive/theme/themes.d.ts.map +1 -1
  93. package/packages/pi-coding-agent/dist/modes/interactive/theme/themes.js +4 -0
  94. package/packages/pi-coding-agent/dist/modes/interactive/theme/themes.js.map +1 -1
  95. package/packages/pi-coding-agent/package.json +1 -1
  96. package/packages/pi-coding-agent/src/cli/config-selector.ts +1 -1
  97. package/packages/pi-coding-agent/src/core/agent-session.ts +5 -28
  98. package/packages/pi-coding-agent/src/core/settings-manager.ts +52 -1
  99. package/packages/pi-coding-agent/src/core/skill-tool.test.ts +18 -5
  100. package/packages/pi-coding-agent/src/core/skills.ts +3 -2
  101. package/packages/pi-coding-agent/src/index.ts +1 -1
  102. package/packages/pi-coding-agent/src/main.ts +1 -1
  103. package/packages/pi-coding-agent/src/modes/interactive/components/bash-execution.ts +2 -2
  104. package/packages/pi-coding-agent/src/modes/interactive/components/footer.ts +12 -13
  105. package/packages/pi-coding-agent/src/modes/interactive/components/settings-selector.ts +39 -1
  106. package/packages/pi-coding-agent/src/modes/interactive/components/tool-execution.ts +1 -1
  107. package/packages/pi-coding-agent/src/modes/interactive/controllers/chat-controller.ts +10 -0
  108. package/packages/pi-coding-agent/src/modes/interactive/controllers/extension-ui-controller.ts +1 -1
  109. package/packages/pi-coding-agent/src/modes/interactive/interactive-mode.ts +46 -20
  110. package/packages/pi-coding-agent/src/modes/interactive/theme/theme.ts +171 -20
  111. package/packages/pi-coding-agent/src/modes/interactive/theme/themes.ts +4 -0
  112. package/packages/pi-tui/dist/components/editor.d.ts +1 -0
  113. package/packages/pi-tui/dist/components/editor.d.ts.map +1 -1
  114. package/packages/pi-tui/dist/components/editor.js +23 -0
  115. package/packages/pi-tui/dist/components/editor.js.map +1 -1
  116. package/packages/pi-tui/src/components/editor.ts +23 -0
  117. package/pkg/dist/modes/interactive/theme/theme.d.ts +18 -5
  118. package/pkg/dist/modes/interactive/theme/theme.d.ts.map +1 -1
  119. package/pkg/dist/modes/interactive/theme/theme.js +139 -20
  120. package/pkg/dist/modes/interactive/theme/theme.js.map +1 -1
  121. package/pkg/dist/modes/interactive/theme/themes.d.ts.map +1 -1
  122. package/pkg/dist/modes/interactive/theme/themes.js +4 -0
  123. package/pkg/dist/modes/interactive/theme/themes.js.map +1 -1
  124. package/pkg/package.json +1 -1
  125. package/src/resources/extensions/memory/auto-extract.ts +23 -3
  126. package/src/resources/extensions/memory/dream.ts +814 -0
  127. package/src/resources/extensions/memory/extension-manifest.json +2 -2
  128. package/src/resources/extensions/memory/index.ts +134 -13
  129. package/src/resources/extensions/memory/tests/auto-extract.test.ts +10 -2
  130. package/src/resources/extensions/memory/tests/dream.test.ts +142 -0
  131. package/src/resources/extensions/slash-commands/extension-manifest.json +10 -10
  132. package/src/resources/extensions/slash-commands/index.ts +3 -7
  133. package/src/resources/extensions/slash-commands/plan.ts +192 -46
  134. package/src/resources/extensions/subagent/agents.ts +11 -1
  135. package/src/resources/extensions/subagent/configured-model.ts +3 -2
  136. package/src/resources/extensions/subagent/index.ts +38 -30
  137. package/src/resources/extensions/subagent/launch-helpers.ts +30 -0
  138. package/src/resources/extensions/subagent/model-resolution.ts +40 -3
  139. package/src/resources/extensions/usage/extension-manifest.json +11 -0
  140. package/src/resources/extensions/usage/index.ts +441 -0
  141. package/{dist/resources/skills/create-gsd-extension → src/resources/skills/create-lsd-extension}/SKILL.md +6 -6
  142. package/{dist/resources/skills/create-gsd-extension → src/resources/skills/create-lsd-extension}/references/custom-tools.md +1 -1
  143. package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/extension-lifecycle.md +2 -2
  144. package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/extensioncontext-reference.md +1 -1
  145. package/{dist/resources/skills/create-gsd-extension → src/resources/skills/create-lsd-extension}/references/key-rules-gotchas.md +4 -4
  146. package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/packaging-distribution.md +6 -6
  147. package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/workflows/create-extension.md +3 -3
  148. package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/workflows/debug-extension.md +5 -5
  149. package/src/resources/skills/teams-debug/SKILL.md +5 -6
  150. package/src/resources/skills/teams-document/SKILL.md +1 -2
  151. package/src/resources/skills/teams-plan/SKILL.md +3 -4
  152. package/src/resources/skills/teams-run/SKILL.md +3 -4
  153. package/src/resources/skills/teams-verify/SKILL.md +4 -5
  154. package/dist/resources/extensions/slash-commands/create-extension.js +0 -264
  155. package/dist/resources/extensions/slash-commands/create-slash-command.js +0 -208
  156. package/src/resources/extensions/slash-commands/create-extension.ts +0 -297
  157. package/src/resources/extensions/slash-commands/create-slash-command.ts +0 -234
  158. /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/compaction-session-control.md +0 -0
  159. /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/custom-commands.md +0 -0
  160. /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/custom-rendering.md +0 -0
  161. /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/custom-ui.md +0 -0
  162. /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/events-reference.md +0 -0
  163. /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/extensionapi-reference.md +0 -0
  164. /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/mode-behavior.md +0 -0
  165. /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/model-provider-management.md +0 -0
  166. /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/remote-execution-overrides.md +0 -0
  167. /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/state-management.md +0 -0
  168. /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/references/system-prompt-modification.md +0 -0
  169. /package/dist/resources/skills/{create-gsd-extension → create-lsd-extension}/workflows/add-capability.md +0 -0
  170. /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/compaction-session-control.md +0 -0
  171. /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/custom-commands.md +0 -0
  172. /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/custom-rendering.md +0 -0
  173. /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/custom-ui.md +0 -0
  174. /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/events-reference.md +0 -0
  175. /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/extensionapi-reference.md +0 -0
  176. /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/mode-behavior.md +0 -0
  177. /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/model-provider-management.md +0 -0
  178. /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/remote-execution-overrides.md +0 -0
  179. /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/state-management.md +0 -0
  180. /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/references/system-prompt-modification.md +0 -0
  181. /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/templates/extension-skeleton.ts +0 -0
  182. /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/templates/stateful-tool-skeleton.ts +0 -0
  183. /package/src/resources/skills/{create-gsd-extension → create-lsd-extension}/workflows/add-capability.md +0 -0
@@ -104,6 +104,23 @@ export const MODELS = {
104
104
  contextWindow: 200000,
105
105
  maxTokens: 8192,
106
106
  },
107
+ "anthropic.claude-3-5-sonnet-20240620-v1:0": {
108
+ id: "anthropic.claude-3-5-sonnet-20240620-v1:0",
109
+ name: "Claude Sonnet 3.5",
110
+ api: "bedrock-converse-stream",
111
+ provider: "amazon-bedrock",
112
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
113
+ reasoning: false,
114
+ input: ["text", "image"],
115
+ cost: {
116
+ input: 3,
117
+ output: 15,
118
+ cacheRead: 0.3,
119
+ cacheWrite: 3.75,
120
+ },
121
+ contextWindow: 200000,
122
+ maxTokens: 8192,
123
+ },
107
124
  "anthropic.claude-3-5-sonnet-20241022-v2:0": {
108
125
  id: "anthropic.claude-3-5-sonnet-20241022-v2:0",
109
126
  name: "Claude Sonnet 3.5 v2",
@@ -138,6 +155,23 @@ export const MODELS = {
138
155
  contextWindow: 200000,
139
156
  maxTokens: 8192,
140
157
  },
158
+ "anthropic.claude-3-haiku-20240307-v1:0": {
159
+ id: "anthropic.claude-3-haiku-20240307-v1:0",
160
+ name: "Claude Haiku 3",
161
+ api: "bedrock-converse-stream",
162
+ provider: "amazon-bedrock",
163
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
164
+ reasoning: false,
165
+ input: ["text", "image"],
166
+ cost: {
167
+ input: 0.25,
168
+ output: 1.25,
169
+ cacheRead: 0,
170
+ cacheWrite: 0,
171
+ },
172
+ contextWindow: 200000,
173
+ maxTokens: 4096,
174
+ },
141
175
  "anthropic.claude-haiku-4-5-20251001-v1:0": {
142
176
  id: "anthropic.claude-haiku-4-5-20251001-v1:0",
143
177
  name: "Claude Haiku 4.5",
@@ -767,23 +801,6 @@ export const MODELS = {
767
801
  contextWindow: 204800,
768
802
  maxTokens: 131072,
769
803
  },
770
- "minimax.minimax-m2.5": {
771
- id: "minimax.minimax-m2.5",
772
- name: "MiniMax M2.5",
773
- api: "bedrock-converse-stream",
774
- provider: "amazon-bedrock",
775
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
776
- reasoning: true,
777
- input: ["text"],
778
- cost: {
779
- input: 0.3,
780
- output: 1.2,
781
- cacheRead: 0,
782
- cacheWrite: 0,
783
- },
784
- contextWindow: 196608,
785
- maxTokens: 98304,
786
- },
787
804
  "mistral.devstral-2-123b": {
788
805
  id: "mistral.devstral-2-123b",
789
806
  name: "Devstral 2 123B",
@@ -1022,23 +1039,6 @@ export const MODELS = {
1022
1039
  contextWindow: 128000,
1023
1040
  maxTokens: 4096,
1024
1041
  },
1025
- "nvidia.nemotron-super-3-120b": {
1026
- id: "nvidia.nemotron-super-3-120b",
1027
- name: "NVIDIA Nemotron 3 Super 120B A12B",
1028
- api: "bedrock-converse-stream",
1029
- provider: "amazon-bedrock",
1030
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
1031
- reasoning: true,
1032
- input: ["text"],
1033
- cost: {
1034
- input: 0.15,
1035
- output: 0.65,
1036
- cacheRead: 0,
1037
- cacheWrite: 0,
1038
- },
1039
- contextWindow: 262144,
1040
- maxTokens: 131072,
1041
- },
1042
1042
  "openai.gpt-oss-120b-1:0": {
1043
1043
  id: "openai.gpt-oss-120b-1:0",
1044
1044
  name: "gpt-oss-120b",
@@ -1413,23 +1413,6 @@ export const MODELS = {
1413
1413
  contextWindow: 200000,
1414
1414
  maxTokens: 131072,
1415
1415
  },
1416
- "zai.glm-5": {
1417
- id: "zai.glm-5",
1418
- name: "GLM-5",
1419
- api: "bedrock-converse-stream",
1420
- provider: "amazon-bedrock",
1421
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
1422
- reasoning: true,
1423
- input: ["text"],
1424
- cost: {
1425
- input: 1,
1426
- output: 3.2,
1427
- cacheRead: 0,
1428
- cacheWrite: 0,
1429
- },
1430
- contextWindow: 202752,
1431
- maxTokens: 101376,
1432
- },
1433
1416
  },
1434
1417
  "anthropic": {
1435
1418
  "claude-3-5-haiku-20241022": {
@@ -1466,6 +1449,23 @@ export const MODELS = {
1466
1449
  contextWindow: 200000,
1467
1450
  maxTokens: 8192,
1468
1451
  },
1452
+ "claude-3-5-sonnet-20240620": {
1453
+ id: "claude-3-5-sonnet-20240620",
1454
+ name: "Claude Sonnet 3.5",
1455
+ api: "anthropic-messages",
1456
+ provider: "anthropic",
1457
+ baseUrl: "https://api.anthropic.com",
1458
+ reasoning: false,
1459
+ input: ["text", "image"],
1460
+ cost: {
1461
+ input: 3,
1462
+ output: 15,
1463
+ cacheRead: 0.3,
1464
+ cacheWrite: 3.75,
1465
+ },
1466
+ contextWindow: 200000,
1467
+ maxTokens: 8192,
1468
+ },
1469
1469
  "claude-3-5-sonnet-20241022": {
1470
1470
  id: "claude-3-5-sonnet-20241022",
1471
1471
  name: "Claude Sonnet 3.5 v2",
@@ -1517,6 +1517,57 @@ export const MODELS = {
1517
1517
  contextWindow: 200000,
1518
1518
  maxTokens: 64000,
1519
1519
  },
1520
+ "claude-3-haiku-20240307": {
1521
+ id: "claude-3-haiku-20240307",
1522
+ name: "Claude Haiku 3",
1523
+ api: "anthropic-messages",
1524
+ provider: "anthropic",
1525
+ baseUrl: "https://api.anthropic.com",
1526
+ reasoning: false,
1527
+ input: ["text", "image"],
1528
+ cost: {
1529
+ input: 0.25,
1530
+ output: 1.25,
1531
+ cacheRead: 0.03,
1532
+ cacheWrite: 0.3,
1533
+ },
1534
+ contextWindow: 200000,
1535
+ maxTokens: 4096,
1536
+ },
1537
+ "claude-3-opus-20240229": {
1538
+ id: "claude-3-opus-20240229",
1539
+ name: "Claude Opus 3",
1540
+ api: "anthropic-messages",
1541
+ provider: "anthropic",
1542
+ baseUrl: "https://api.anthropic.com",
1543
+ reasoning: false,
1544
+ input: ["text", "image"],
1545
+ cost: {
1546
+ input: 15,
1547
+ output: 75,
1548
+ cacheRead: 1.5,
1549
+ cacheWrite: 18.75,
1550
+ },
1551
+ contextWindow: 200000,
1552
+ maxTokens: 4096,
1553
+ },
1554
+ "claude-3-sonnet-20240229": {
1555
+ id: "claude-3-sonnet-20240229",
1556
+ name: "Claude Sonnet 3",
1557
+ api: "anthropic-messages",
1558
+ provider: "anthropic",
1559
+ baseUrl: "https://api.anthropic.com",
1560
+ reasoning: false,
1561
+ input: ["text", "image"],
1562
+ cost: {
1563
+ input: 3,
1564
+ output: 15,
1565
+ cacheRead: 0.3,
1566
+ cacheWrite: 0.3,
1567
+ },
1568
+ contextWindow: 200000,
1569
+ maxTokens: 4096,
1570
+ },
1520
1571
  "claude-haiku-4-5": {
1521
1572
  id: "claude-haiku-4-5",
1522
1573
  name: "Claude Haiku 4.5 (latest)",
@@ -1774,6 +1825,40 @@ export const MODELS = {
1774
1825
  contextWindow: 200000,
1775
1826
  maxTokens: 100000,
1776
1827
  },
1828
+ "gpt-4": {
1829
+ id: "gpt-4",
1830
+ name: "GPT-4",
1831
+ api: "azure-openai-responses",
1832
+ provider: "azure-openai-responses",
1833
+ baseUrl: "",
1834
+ reasoning: false,
1835
+ input: ["text"],
1836
+ cost: {
1837
+ input: 30,
1838
+ output: 60,
1839
+ cacheRead: 0,
1840
+ cacheWrite: 0,
1841
+ },
1842
+ contextWindow: 8192,
1843
+ maxTokens: 8192,
1844
+ },
1845
+ "gpt-4-turbo": {
1846
+ id: "gpt-4-turbo",
1847
+ name: "GPT-4 Turbo",
1848
+ api: "azure-openai-responses",
1849
+ provider: "azure-openai-responses",
1850
+ baseUrl: "",
1851
+ reasoning: false,
1852
+ input: ["text", "image"],
1853
+ cost: {
1854
+ input: 10,
1855
+ output: 30,
1856
+ cacheRead: 0,
1857
+ cacheWrite: 0,
1858
+ },
1859
+ contextWindow: 128000,
1860
+ maxTokens: 4096,
1861
+ },
1777
1862
  "gpt-4.1": {
1778
1863
  id: "gpt-4.1",
1779
1864
  name: "GPT-4.1",
@@ -1825,6 +1910,91 @@ export const MODELS = {
1825
1910
  contextWindow: 1047576,
1826
1911
  maxTokens: 32768,
1827
1912
  },
1913
+ "gpt-4o": {
1914
+ id: "gpt-4o",
1915
+ name: "GPT-4o",
1916
+ api: "azure-openai-responses",
1917
+ provider: "azure-openai-responses",
1918
+ baseUrl: "",
1919
+ reasoning: false,
1920
+ input: ["text", "image"],
1921
+ cost: {
1922
+ input: 2.5,
1923
+ output: 10,
1924
+ cacheRead: 1.25,
1925
+ cacheWrite: 0,
1926
+ },
1927
+ contextWindow: 128000,
1928
+ maxTokens: 16384,
1929
+ },
1930
+ "gpt-4o-2024-05-13": {
1931
+ id: "gpt-4o-2024-05-13",
1932
+ name: "GPT-4o (2024-05-13)",
1933
+ api: "azure-openai-responses",
1934
+ provider: "azure-openai-responses",
1935
+ baseUrl: "",
1936
+ reasoning: false,
1937
+ input: ["text", "image"],
1938
+ cost: {
1939
+ input: 5,
1940
+ output: 15,
1941
+ cacheRead: 0,
1942
+ cacheWrite: 0,
1943
+ },
1944
+ contextWindow: 128000,
1945
+ maxTokens: 4096,
1946
+ },
1947
+ "gpt-4o-2024-08-06": {
1948
+ id: "gpt-4o-2024-08-06",
1949
+ name: "GPT-4o (2024-08-06)",
1950
+ api: "azure-openai-responses",
1951
+ provider: "azure-openai-responses",
1952
+ baseUrl: "",
1953
+ reasoning: false,
1954
+ input: ["text", "image"],
1955
+ cost: {
1956
+ input: 2.5,
1957
+ output: 10,
1958
+ cacheRead: 1.25,
1959
+ cacheWrite: 0,
1960
+ },
1961
+ contextWindow: 128000,
1962
+ maxTokens: 16384,
1963
+ },
1964
+ "gpt-4o-2024-11-20": {
1965
+ id: "gpt-4o-2024-11-20",
1966
+ name: "GPT-4o (2024-11-20)",
1967
+ api: "azure-openai-responses",
1968
+ provider: "azure-openai-responses",
1969
+ baseUrl: "",
1970
+ reasoning: false,
1971
+ input: ["text", "image"],
1972
+ cost: {
1973
+ input: 2.5,
1974
+ output: 10,
1975
+ cacheRead: 1.25,
1976
+ cacheWrite: 0,
1977
+ },
1978
+ contextWindow: 128000,
1979
+ maxTokens: 16384,
1980
+ },
1981
+ "gpt-4o-mini": {
1982
+ id: "gpt-4o-mini",
1983
+ name: "GPT-4o mini",
1984
+ api: "azure-openai-responses",
1985
+ provider: "azure-openai-responses",
1986
+ baseUrl: "",
1987
+ reasoning: false,
1988
+ input: ["text", "image"],
1989
+ cost: {
1990
+ input: 0.15,
1991
+ output: 0.6,
1992
+ cacheRead: 0.08,
1993
+ cacheWrite: 0,
1994
+ },
1995
+ contextWindow: 128000,
1996
+ maxTokens: 16384,
1997
+ },
1828
1998
  "gpt-5": {
1829
1999
  id: "gpt-5",
1830
2000
  name: "GPT-5",
@@ -2080,23 +2250,6 @@ export const MODELS = {
2080
2250
  contextWindow: 400000,
2081
2251
  maxTokens: 128000,
2082
2252
  },
2083
- "gpt-5.3-chat-latest": {
2084
- id: "gpt-5.3-chat-latest",
2085
- name: "GPT-5.3 Chat (latest)",
2086
- api: "azure-openai-responses",
2087
- provider: "azure-openai-responses",
2088
- baseUrl: "",
2089
- reasoning: false,
2090
- input: ["text", "image"],
2091
- cost: {
2092
- input: 1.75,
2093
- output: 14,
2094
- cacheRead: 0.175,
2095
- cacheWrite: 0,
2096
- },
2097
- contextWindow: 128000,
2098
- maxTokens: 16384,
2099
- },
2100
2253
  "gpt-5.3-codex": {
2101
2254
  id: "gpt-5.3-codex",
2102
2255
  name: "GPT-5.3 Codex",
@@ -2610,6 +2763,25 @@ export const MODELS = {
2610
2763
  contextWindow: 128000,
2611
2764
  maxTokens: 16384,
2612
2765
  },
2766
+ "gpt-4o": {
2767
+ id: "gpt-4o",
2768
+ name: "GPT-4o",
2769
+ api: "openai-completions",
2770
+ provider: "github-copilot",
2771
+ baseUrl: "https://api.individual.githubcopilot.com",
2772
+ headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
2773
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false },
2774
+ reasoning: false,
2775
+ input: ["text", "image"],
2776
+ cost: {
2777
+ input: 0,
2778
+ output: 0,
2779
+ cacheRead: 0,
2780
+ cacheWrite: 0,
2781
+ },
2782
+ contextWindow: 128000,
2783
+ maxTokens: 4096,
2784
+ },
2613
2785
  "gpt-5": {
2614
2786
  id: "gpt-5",
2615
2787
  name: "GPT-5",
@@ -2829,7 +3001,58 @@ export const MODELS = {
2829
3001
  },
2830
3002
  },
2831
3003
  "google": {
2832
- "gemini-2.0-flash": {
3004
+ "gemini-1.5-flash": {
3005
+ id: "gemini-1.5-flash",
3006
+ name: "Gemini 1.5 Flash",
3007
+ api: "google-generative-ai",
3008
+ provider: "google",
3009
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
3010
+ reasoning: false,
3011
+ input: ["text", "image"],
3012
+ cost: {
3013
+ input: 0.075,
3014
+ output: 0.3,
3015
+ cacheRead: 0.01875,
3016
+ cacheWrite: 0,
3017
+ },
3018
+ contextWindow: 1000000,
3019
+ maxTokens: 8192,
3020
+ },
3021
+ "gemini-1.5-flash-8b": {
3022
+ id: "gemini-1.5-flash-8b",
3023
+ name: "Gemini 1.5 Flash-8B",
3024
+ api: "google-generative-ai",
3025
+ provider: "google",
3026
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
3027
+ reasoning: false,
3028
+ input: ["text", "image"],
3029
+ cost: {
3030
+ input: 0.0375,
3031
+ output: 0.15,
3032
+ cacheRead: 0.01,
3033
+ cacheWrite: 0,
3034
+ },
3035
+ contextWindow: 1000000,
3036
+ maxTokens: 8192,
3037
+ },
3038
+ "gemini-1.5-pro": {
3039
+ id: "gemini-1.5-pro",
3040
+ name: "Gemini 1.5 Pro",
3041
+ api: "google-generative-ai",
3042
+ provider: "google",
3043
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
3044
+ reasoning: false,
3045
+ input: ["text", "image"],
3046
+ cost: {
3047
+ input: 1.25,
3048
+ output: 5,
3049
+ cacheRead: 0.3125,
3050
+ cacheWrite: 0,
3051
+ },
3052
+ contextWindow: 1000000,
3053
+ maxTokens: 8192,
3054
+ },
3055
+ "gemini-2.0-flash": {
2833
3056
  id: "gemini-2.0-flash",
2834
3057
  name: "Gemini 2.0 Flash",
2835
3058
  api: "google-generative-ai",
@@ -3186,23 +3409,6 @@ export const MODELS = {
3186
3409
  contextWindow: 131072,
3187
3410
  maxTokens: 65536,
3188
3411
  },
3189
- "gemma-3-27b-it": {
3190
- id: "gemma-3-27b-it",
3191
- name: "Gemma 3 27B",
3192
- api: "google-generative-ai",
3193
- provider: "google",
3194
- baseUrl: "https://generativelanguage.googleapis.com/v1beta",
3195
- reasoning: false,
3196
- input: ["text", "image"],
3197
- cost: {
3198
- input: 0,
3199
- output: 0,
3200
- cacheRead: 0,
3201
- cacheWrite: 0,
3202
- },
3203
- contextWindow: 131072,
3204
- maxTokens: 8192,
3205
- },
3206
3412
  },
3207
3413
  "google-antigravity": {
3208
3414
  "claude-opus-4-5-thinking": {
@@ -3464,6 +3670,57 @@ export const MODELS = {
3464
3670
  },
3465
3671
  },
3466
3672
  "google-vertex": {
3673
+ "gemini-1.5-flash": {
3674
+ id: "gemini-1.5-flash",
3675
+ name: "Gemini 1.5 Flash (Vertex)",
3676
+ api: "google-vertex",
3677
+ provider: "google-vertex",
3678
+ baseUrl: "https://{location}-aiplatform.googleapis.com",
3679
+ reasoning: false,
3680
+ input: ["text", "image"],
3681
+ cost: {
3682
+ input: 0.075,
3683
+ output: 0.3,
3684
+ cacheRead: 0.01875,
3685
+ cacheWrite: 0,
3686
+ },
3687
+ contextWindow: 1000000,
3688
+ maxTokens: 8192,
3689
+ },
3690
+ "gemini-1.5-flash-8b": {
3691
+ id: "gemini-1.5-flash-8b",
3692
+ name: "Gemini 1.5 Flash-8B (Vertex)",
3693
+ api: "google-vertex",
3694
+ provider: "google-vertex",
3695
+ baseUrl: "https://{location}-aiplatform.googleapis.com",
3696
+ reasoning: false,
3697
+ input: ["text", "image"],
3698
+ cost: {
3699
+ input: 0.0375,
3700
+ output: 0.15,
3701
+ cacheRead: 0.01,
3702
+ cacheWrite: 0,
3703
+ },
3704
+ contextWindow: 1000000,
3705
+ maxTokens: 8192,
3706
+ },
3707
+ "gemini-1.5-pro": {
3708
+ id: "gemini-1.5-pro",
3709
+ name: "Gemini 1.5 Pro (Vertex)",
3710
+ api: "google-vertex",
3711
+ provider: "google-vertex",
3712
+ baseUrl: "https://{location}-aiplatform.googleapis.com",
3713
+ reasoning: false,
3714
+ input: ["text", "image"],
3715
+ cost: {
3716
+ input: 1.25,
3717
+ output: 5,
3718
+ cacheRead: 0.3125,
3719
+ cacheWrite: 0,
3720
+ },
3721
+ contextWindow: 1000000,
3722
+ maxTokens: 8192,
3723
+ },
3467
3724
  "gemini-2.0-flash": {
3468
3725
  id: "gemini-2.0-flash",
3469
3726
  name: "Gemini 2.0 Flash (Vertex)",
@@ -3653,40 +3910,6 @@ export const MODELS = {
3653
3910
  contextWindow: 8192,
3654
3911
  maxTokens: 8192,
3655
3912
  },
3656
- "groq/compound": {
3657
- id: "groq/compound",
3658
- name: "Compound",
3659
- api: "openai-completions",
3660
- provider: "groq",
3661
- baseUrl: "https://api.groq.com/openai/v1",
3662
- reasoning: true,
3663
- input: ["text"],
3664
- cost: {
3665
- input: 0,
3666
- output: 0,
3667
- cacheRead: 0,
3668
- cacheWrite: 0,
3669
- },
3670
- contextWindow: 131072,
3671
- maxTokens: 8192,
3672
- },
3673
- "groq/compound-mini": {
3674
- id: "groq/compound-mini",
3675
- name: "Compound Mini",
3676
- api: "openai-completions",
3677
- provider: "groq",
3678
- baseUrl: "https://api.groq.com/openai/v1",
3679
- reasoning: true,
3680
- input: ["text"],
3681
- cost: {
3682
- input: 0,
3683
- output: 0,
3684
- cacheRead: 0,
3685
- cacheWrite: 0,
3686
- },
3687
- contextWindow: 131072,
3688
- maxTokens: 8192,
3689
- },
3690
3913
  "llama-3.1-8b-instant": {
3691
3914
  id: "llama-3.1-8b-instant",
3692
3915
  name: "Llama 3.1 8B Instant",
@@ -3874,23 +4097,6 @@ export const MODELS = {
3874
4097
  contextWindow: 131072,
3875
4098
  maxTokens: 65536,
3876
4099
  },
3877
- "openai/gpt-oss-safeguard-20b": {
3878
- id: "openai/gpt-oss-safeguard-20b",
3879
- name: "Safety GPT OSS 20B",
3880
- api: "openai-completions",
3881
- provider: "groq",
3882
- baseUrl: "https://api.groq.com/openai/v1",
3883
- reasoning: true,
3884
- input: ["text"],
3885
- cost: {
3886
- input: 0.075,
3887
- output: 0.3,
3888
- cacheRead: 0.037,
3889
- cacheWrite: 0,
3890
- },
3891
- contextWindow: 131072,
3892
- maxTokens: 65536,
3893
- },
3894
4100
  "qwen-qwq-32b": {
3895
4101
  id: "qwen-qwq-32b",
3896
4102
  name: "Qwen QwQ 32B",
@@ -3923,7 +4129,7 @@ export const MODELS = {
3923
4129
  cacheWrite: 0,
3924
4130
  },
3925
4131
  contextWindow: 131072,
3926
- maxTokens: 40960,
4132
+ maxTokens: 16384,
3927
4133
  },
3928
4134
  },
3929
4135
  "huggingface": {
@@ -4820,39 +5026,22 @@ export const MODELS = {
4820
5026
  contextWindow: 128000,
4821
5027
  maxTokens: 16384,
4822
5028
  },
4823
- "mistral-small-2603": {
4824
- id: "mistral-small-2603",
4825
- name: "Mistral Small 4",
4826
- api: "mistral-conversations",
4827
- provider: "mistral",
4828
- baseUrl: "https://api.mistral.ai",
4829
- reasoning: true,
4830
- input: ["text", "image"],
4831
- cost: {
4832
- input: 0.15,
4833
- output: 0.6,
4834
- cacheRead: 0,
4835
- cacheWrite: 0,
4836
- },
4837
- contextWindow: 256000,
4838
- maxTokens: 256000,
4839
- },
4840
5029
  "mistral-small-latest": {
4841
5030
  id: "mistral-small-latest",
4842
5031
  name: "Mistral Small (latest)",
4843
5032
  api: "mistral-conversations",
4844
5033
  provider: "mistral",
4845
5034
  baseUrl: "https://api.mistral.ai",
4846
- reasoning: true,
5035
+ reasoning: false,
4847
5036
  input: ["text", "image"],
4848
5037
  cost: {
4849
- input: 0.15,
4850
- output: 0.6,
5038
+ input: 0.1,
5039
+ output: 0.3,
4851
5040
  cacheRead: 0,
4852
5041
  cacheWrite: 0,
4853
5042
  },
4854
- contextWindow: 256000,
4855
- maxTokens: 256000,
5043
+ contextWindow: 128000,
5044
+ maxTokens: 16384,
4856
5045
  },
4857
5046
  "open-mistral-7b": {
4858
5047
  id: "open-mistral-7b",
@@ -4931,83 +5120,202 @@ export const MODELS = {
4931
5120
  reasoning: false,
4932
5121
  input: ["text", "image"],
4933
5122
  cost: {
4934
- input: 2,
4935
- output: 6,
4936
- cacheRead: 0,
5123
+ input: 2,
5124
+ output: 6,
5125
+ cacheRead: 0,
5126
+ cacheWrite: 0,
5127
+ },
5128
+ contextWindow: 128000,
5129
+ maxTokens: 128000,
5130
+ },
5131
+ },
5132
+ "openai": {
5133
+ "codex-mini-latest": {
5134
+ id: "codex-mini-latest",
5135
+ name: "Codex Mini",
5136
+ api: "openai-responses",
5137
+ provider: "openai",
5138
+ baseUrl: "https://api.openai.com/v1",
5139
+ reasoning: true,
5140
+ input: ["text"],
5141
+ cost: {
5142
+ input: 1.5,
5143
+ output: 6,
5144
+ cacheRead: 0.375,
5145
+ cacheWrite: 0,
5146
+ },
5147
+ contextWindow: 200000,
5148
+ maxTokens: 100000,
5149
+ },
5150
+ "gpt-4": {
5151
+ id: "gpt-4",
5152
+ name: "GPT-4",
5153
+ api: "openai-responses",
5154
+ provider: "openai",
5155
+ baseUrl: "https://api.openai.com/v1",
5156
+ reasoning: false,
5157
+ input: ["text"],
5158
+ cost: {
5159
+ input: 30,
5160
+ output: 60,
5161
+ cacheRead: 0,
5162
+ cacheWrite: 0,
5163
+ },
5164
+ contextWindow: 8192,
5165
+ maxTokens: 8192,
5166
+ },
5167
+ "gpt-4-turbo": {
5168
+ id: "gpt-4-turbo",
5169
+ name: "GPT-4 Turbo",
5170
+ api: "openai-responses",
5171
+ provider: "openai",
5172
+ baseUrl: "https://api.openai.com/v1",
5173
+ reasoning: false,
5174
+ input: ["text", "image"],
5175
+ cost: {
5176
+ input: 10,
5177
+ output: 30,
5178
+ cacheRead: 0,
5179
+ cacheWrite: 0,
5180
+ },
5181
+ contextWindow: 128000,
5182
+ maxTokens: 4096,
5183
+ },
5184
+ "gpt-4.1": {
5185
+ id: "gpt-4.1",
5186
+ name: "GPT-4.1",
5187
+ api: "openai-responses",
5188
+ provider: "openai",
5189
+ baseUrl: "https://api.openai.com/v1",
5190
+ reasoning: false,
5191
+ input: ["text", "image"],
5192
+ cost: {
5193
+ input: 2,
5194
+ output: 8,
5195
+ cacheRead: 0.5,
5196
+ cacheWrite: 0,
5197
+ },
5198
+ contextWindow: 1047576,
5199
+ maxTokens: 32768,
5200
+ },
5201
+ "gpt-4.1-mini": {
5202
+ id: "gpt-4.1-mini",
5203
+ name: "GPT-4.1 mini",
5204
+ api: "openai-responses",
5205
+ provider: "openai",
5206
+ baseUrl: "https://api.openai.com/v1",
5207
+ reasoning: false,
5208
+ input: ["text", "image"],
5209
+ cost: {
5210
+ input: 0.4,
5211
+ output: 1.6,
5212
+ cacheRead: 0.1,
5213
+ cacheWrite: 0,
5214
+ },
5215
+ contextWindow: 1047576,
5216
+ maxTokens: 32768,
5217
+ },
5218
+ "gpt-4.1-nano": {
5219
+ id: "gpt-4.1-nano",
5220
+ name: "GPT-4.1 nano",
5221
+ api: "openai-responses",
5222
+ provider: "openai",
5223
+ baseUrl: "https://api.openai.com/v1",
5224
+ reasoning: false,
5225
+ input: ["text", "image"],
5226
+ cost: {
5227
+ input: 0.1,
5228
+ output: 0.4,
5229
+ cacheRead: 0.03,
5230
+ cacheWrite: 0,
5231
+ },
5232
+ contextWindow: 1047576,
5233
+ maxTokens: 32768,
5234
+ },
5235
+ "gpt-4o": {
5236
+ id: "gpt-4o",
5237
+ name: "GPT-4o",
5238
+ api: "openai-responses",
5239
+ provider: "openai",
5240
+ baseUrl: "https://api.openai.com/v1",
5241
+ reasoning: false,
5242
+ input: ["text", "image"],
5243
+ cost: {
5244
+ input: 2.5,
5245
+ output: 10,
5246
+ cacheRead: 1.25,
4937
5247
  cacheWrite: 0,
4938
5248
  },
4939
5249
  contextWindow: 128000,
4940
- maxTokens: 128000,
5250
+ maxTokens: 16384,
4941
5251
  },
4942
- },
4943
- "openai": {
4944
- "codex-mini-latest": {
4945
- id: "codex-mini-latest",
4946
- name: "Codex Mini",
5252
+ "gpt-4o-2024-05-13": {
5253
+ id: "gpt-4o-2024-05-13",
5254
+ name: "GPT-4o (2024-05-13)",
4947
5255
  api: "openai-responses",
4948
5256
  provider: "openai",
4949
5257
  baseUrl: "https://api.openai.com/v1",
4950
- reasoning: true,
4951
- input: ["text"],
5258
+ reasoning: false,
5259
+ input: ["text", "image"],
4952
5260
  cost: {
4953
- input: 1.5,
4954
- output: 6,
4955
- cacheRead: 0.375,
5261
+ input: 5,
5262
+ output: 15,
5263
+ cacheRead: 0,
4956
5264
  cacheWrite: 0,
4957
5265
  },
4958
- contextWindow: 200000,
4959
- maxTokens: 100000,
5266
+ contextWindow: 128000,
5267
+ maxTokens: 4096,
4960
5268
  },
4961
- "gpt-4.1": {
4962
- id: "gpt-4.1",
4963
- name: "GPT-4.1",
5269
+ "gpt-4o-2024-08-06": {
5270
+ id: "gpt-4o-2024-08-06",
5271
+ name: "GPT-4o (2024-08-06)",
4964
5272
  api: "openai-responses",
4965
5273
  provider: "openai",
4966
5274
  baseUrl: "https://api.openai.com/v1",
4967
5275
  reasoning: false,
4968
5276
  input: ["text", "image"],
4969
5277
  cost: {
4970
- input: 2,
4971
- output: 8,
4972
- cacheRead: 0.5,
5278
+ input: 2.5,
5279
+ output: 10,
5280
+ cacheRead: 1.25,
4973
5281
  cacheWrite: 0,
4974
5282
  },
4975
- contextWindow: 1047576,
4976
- maxTokens: 32768,
5283
+ contextWindow: 128000,
5284
+ maxTokens: 16384,
4977
5285
  },
4978
- "gpt-4.1-mini": {
4979
- id: "gpt-4.1-mini",
4980
- name: "GPT-4.1 mini",
5286
+ "gpt-4o-2024-11-20": {
5287
+ id: "gpt-4o-2024-11-20",
5288
+ name: "GPT-4o (2024-11-20)",
4981
5289
  api: "openai-responses",
4982
5290
  provider: "openai",
4983
5291
  baseUrl: "https://api.openai.com/v1",
4984
5292
  reasoning: false,
4985
5293
  input: ["text", "image"],
4986
5294
  cost: {
4987
- input: 0.4,
4988
- output: 1.6,
4989
- cacheRead: 0.1,
5295
+ input: 2.5,
5296
+ output: 10,
5297
+ cacheRead: 1.25,
4990
5298
  cacheWrite: 0,
4991
5299
  },
4992
- contextWindow: 1047576,
4993
- maxTokens: 32768,
5300
+ contextWindow: 128000,
5301
+ maxTokens: 16384,
4994
5302
  },
4995
- "gpt-4.1-nano": {
4996
- id: "gpt-4.1-nano",
4997
- name: "GPT-4.1 nano",
5303
+ "gpt-4o-mini": {
5304
+ id: "gpt-4o-mini",
5305
+ name: "GPT-4o mini",
4998
5306
  api: "openai-responses",
4999
5307
  provider: "openai",
5000
5308
  baseUrl: "https://api.openai.com/v1",
5001
5309
  reasoning: false,
5002
5310
  input: ["text", "image"],
5003
5311
  cost: {
5004
- input: 0.1,
5005
- output: 0.4,
5006
- cacheRead: 0.03,
5312
+ input: 0.15,
5313
+ output: 0.6,
5314
+ cacheRead: 0.08,
5007
5315
  cacheWrite: 0,
5008
5316
  },
5009
- contextWindow: 1047576,
5010
- maxTokens: 32768,
5317
+ contextWindow: 128000,
5318
+ maxTokens: 16384,
5011
5319
  },
5012
5320
  "gpt-5": {
5013
5321
  id: "gpt-5",
@@ -5264,23 +5572,6 @@ export const MODELS = {
5264
5572
  contextWindow: 400000,
5265
5573
  maxTokens: 128000,
5266
5574
  },
5267
- "gpt-5.3-chat-latest": {
5268
- id: "gpt-5.3-chat-latest",
5269
- name: "GPT-5.3 Chat (latest)",
5270
- api: "openai-responses",
5271
- provider: "openai",
5272
- baseUrl: "https://api.openai.com/v1",
5273
- reasoning: false,
5274
- input: ["text", "image"],
5275
- cost: {
5276
- input: 1.75,
5277
- output: 14,
5278
- cacheRead: 0.175,
5279
- cacheWrite: 0,
5280
- },
5281
- contextWindow: 128000,
5282
- maxTokens: 16384,
5283
- },
5284
5575
  "gpt-5.3-codex": {
5285
5576
  id: "gpt-5.3-codex",
5286
5577
  name: "GPT-5.3 Codex",
@@ -6203,23 +6494,6 @@ export const MODELS = {
6203
6494
  contextWindow: 1000000,
6204
6495
  maxTokens: 128000,
6205
6496
  },
6206
- "qwen3.6-plus-free": {
6207
- id: "qwen3.6-plus-free",
6208
- name: "Qwen3.6 Plus Free",
6209
- api: "openai-completions",
6210
- provider: "opencode",
6211
- baseUrl: "https://opencode.ai/zen/v1",
6212
- reasoning: true,
6213
- input: ["text"],
6214
- cost: {
6215
- input: 0,
6216
- output: 0,
6217
- cacheRead: 0,
6218
- cacheWrite: 0,
6219
- },
6220
- contextWindow: 1048576,
6221
- maxTokens: 64000,
6222
- },
6223
6497
  },
6224
6498
  "opencode-go": {
6225
6499
  "glm-5": {
@@ -6428,6 +6702,23 @@ export const MODELS = {
6428
6702
  contextWindow: 300000,
6429
6703
  maxTokens: 5120,
6430
6704
  },
6705
+ "anthropic/claude-3-haiku": {
6706
+ id: "anthropic/claude-3-haiku",
6707
+ name: "Anthropic: Claude 3 Haiku",
6708
+ api: "openai-completions",
6709
+ provider: "openrouter",
6710
+ baseUrl: "https://openrouter.ai/api/v1",
6711
+ reasoning: false,
6712
+ input: ["text", "image"],
6713
+ cost: {
6714
+ input: 0.25,
6715
+ output: 1.25,
6716
+ cacheRead: 0.03,
6717
+ cacheWrite: 0.3,
6718
+ },
6719
+ contextWindow: 200000,
6720
+ maxTokens: 4096,
6721
+ },
6431
6722
  "anthropic/claude-3.5-haiku": {
6432
6723
  id: "anthropic/claude-3.5-haiku",
6433
6724
  name: "Anthropic: Claude 3.5 Haiku",
@@ -7159,6 +7450,23 @@ export const MODELS = {
7159
7450
  contextWindow: 1048576,
7160
7451
  maxTokens: 65536,
7161
7452
  },
7453
+ "google/gemini-3-pro-preview": {
7454
+ id: "google/gemini-3-pro-preview",
7455
+ name: "Google: Gemini 3 Pro Preview",
7456
+ api: "openai-completions",
7457
+ provider: "openrouter",
7458
+ baseUrl: "https://openrouter.ai/api/v1",
7459
+ reasoning: true,
7460
+ input: ["text", "image"],
7461
+ cost: {
7462
+ input: 2,
7463
+ output: 12,
7464
+ cacheRead: 0.19999999999999998,
7465
+ cacheWrite: 0.375,
7466
+ },
7467
+ contextWindow: 1048576,
7468
+ maxTokens: 65536,
7469
+ },
7162
7470
  "google/gemini-3.1-flash-lite-preview": {
7163
7471
  id: "google/gemini-3.1-flash-lite-preview",
7164
7472
  name: "Google: Gemini 3.1 Flash Lite Preview",
@@ -7261,22 +7569,22 @@ export const MODELS = {
7261
7569
  contextWindow: 128000,
7262
7570
  maxTokens: 32000,
7263
7571
  },
7264
- "kwaipilot/kat-coder-pro-v2": {
7265
- id: "kwaipilot/kat-coder-pro-v2",
7266
- name: "Kwaipilot: KAT-Coder-Pro V2",
7572
+ "kwaipilot/kat-coder-pro": {
7573
+ id: "kwaipilot/kat-coder-pro",
7574
+ name: "Kwaipilot: KAT-Coder-Pro V1",
7267
7575
  api: "openai-completions",
7268
7576
  provider: "openrouter",
7269
7577
  baseUrl: "https://openrouter.ai/api/v1",
7270
7578
  reasoning: false,
7271
7579
  input: ["text"],
7272
7580
  cost: {
7273
- input: 0.3,
7274
- output: 1.2,
7275
- cacheRead: 0.06,
7581
+ input: 0.207,
7582
+ output: 0.828,
7583
+ cacheRead: 0.0414,
7276
7584
  cacheWrite: 0,
7277
7585
  },
7278
7586
  contextWindow: 256000,
7279
- maxTokens: 80000,
7587
+ maxTokens: 128000,
7280
7588
  },
7281
7589
  "meituan/longcat-flash-chat": {
7282
7590
  id: "meituan/longcat-flash-chat",
@@ -7474,9 +7782,9 @@ export const MODELS = {
7474
7782
  reasoning: true,
7475
7783
  input: ["text"],
7476
7784
  cost: {
7477
- input: 0.16,
7478
- output: 1.1,
7479
- cacheRead: 0.08,
7785
+ input: 0.19999999999999998,
7786
+ output: 1.17,
7787
+ cacheRead: 0.09999999999999999,
7480
7788
  cacheWrite: 0,
7481
7789
  },
7482
7790
  contextWindow: 196608,
@@ -7805,6 +8113,23 @@ export const MODELS = {
7805
8113
  contextWindow: 262144,
7806
8114
  maxTokens: 4096,
7807
8115
  },
8116
+ "mistralai/mistral-small-3.1-24b-instruct:free": {
8117
+ id: "mistralai/mistral-small-3.1-24b-instruct:free",
8118
+ name: "Mistral: Mistral Small 3.1 24B (free)",
8119
+ api: "openai-completions",
8120
+ provider: "openrouter",
8121
+ baseUrl: "https://openrouter.ai/api/v1",
8122
+ reasoning: false,
8123
+ input: ["text", "image"],
8124
+ cost: {
8125
+ input: 0,
8126
+ output: 0,
8127
+ cacheRead: 0,
8128
+ cacheWrite: 0,
8129
+ },
8130
+ contextWindow: 128000,
8131
+ maxTokens: 4096,
8132
+ },
7808
8133
  "mistralai/mistral-small-3.2-24b-instruct": {
7809
8134
  id: "mistralai/mistral-small-3.2-24b-instruct",
7810
8135
  name: "Mistral: Mistral Small 3.2 24B",
@@ -7916,13 +8241,13 @@ export const MODELS = {
7916
8241
  reasoning: false,
7917
8242
  input: ["text"],
7918
8243
  cost: {
7919
- input: 0.5700000000000001,
7920
- output: 2.3,
8244
+ input: 0.55,
8245
+ output: 2.2,
7921
8246
  cacheRead: 0,
7922
8247
  cacheWrite: 0,
7923
8248
  },
7924
- contextWindow: 131072,
7925
- maxTokens: 131072,
8249
+ contextWindow: 131000,
8250
+ maxTokens: 4096,
7926
8251
  },
7927
8252
  "moonshotai/kimi-k2-0905": {
7928
8253
  id: "moonshotai/kimi-k2-0905",
@@ -7984,8 +8309,8 @@ export const MODELS = {
7984
8309
  reasoning: false,
7985
8310
  input: ["text"],
7986
8311
  cost: {
7987
- input: 0.135,
7988
- output: 0.5,
8312
+ input: 0.27,
8313
+ output: 1,
7989
8314
  cacheRead: 0,
7990
8315
  cacheWrite: 0,
7991
8316
  },
@@ -8071,7 +8396,7 @@ export const MODELS = {
8071
8396
  cost: {
8072
8397
  input: 0.09999999999999999,
8073
8398
  output: 0.5,
8074
- cacheRead: 0.09999999999999999,
8399
+ cacheRead: 0.04,
8075
8400
  cacheWrite: 0,
8076
8401
  },
8077
8402
  contextWindow: 262144,
@@ -8145,6 +8470,57 @@ export const MODELS = {
8145
8470
  contextWindow: 128000,
8146
8471
  maxTokens: 4096,
8147
8472
  },
8473
+ "openai/gpt-3.5-turbo": {
8474
+ id: "openai/gpt-3.5-turbo",
8475
+ name: "OpenAI: GPT-3.5 Turbo",
8476
+ api: "openai-completions",
8477
+ provider: "openrouter",
8478
+ baseUrl: "https://openrouter.ai/api/v1",
8479
+ reasoning: false,
8480
+ input: ["text"],
8481
+ cost: {
8482
+ input: 0.5,
8483
+ output: 1.5,
8484
+ cacheRead: 0,
8485
+ cacheWrite: 0,
8486
+ },
8487
+ contextWindow: 16385,
8488
+ maxTokens: 4096,
8489
+ },
8490
+ "openai/gpt-3.5-turbo-0613": {
8491
+ id: "openai/gpt-3.5-turbo-0613",
8492
+ name: "OpenAI: GPT-3.5 Turbo (older v0613)",
8493
+ api: "openai-completions",
8494
+ provider: "openrouter",
8495
+ baseUrl: "https://openrouter.ai/api/v1",
8496
+ reasoning: false,
8497
+ input: ["text"],
8498
+ cost: {
8499
+ input: 1,
8500
+ output: 2,
8501
+ cacheRead: 0,
8502
+ cacheWrite: 0,
8503
+ },
8504
+ contextWindow: 4095,
8505
+ maxTokens: 4096,
8506
+ },
8507
+ "openai/gpt-3.5-turbo-16k": {
8508
+ id: "openai/gpt-3.5-turbo-16k",
8509
+ name: "OpenAI: GPT-3.5 Turbo 16k",
8510
+ api: "openai-completions",
8511
+ provider: "openrouter",
8512
+ baseUrl: "https://openrouter.ai/api/v1",
8513
+ reasoning: false,
8514
+ input: ["text"],
8515
+ cost: {
8516
+ input: 3,
8517
+ output: 4,
8518
+ cacheRead: 0,
8519
+ cacheWrite: 0,
8520
+ },
8521
+ contextWindow: 16385,
8522
+ maxTokens: 4096,
8523
+ },
8148
8524
  "openai/gpt-4": {
8149
8525
  id: "openai/gpt-4",
8150
8526
  name: "OpenAI: GPT-4",
@@ -8196,6 +8572,40 @@ export const MODELS = {
8196
8572
  contextWindow: 128000,
8197
8573
  maxTokens: 4096,
8198
8574
  },
8575
+ "openai/gpt-4-turbo": {
8576
+ id: "openai/gpt-4-turbo",
8577
+ name: "OpenAI: GPT-4 Turbo",
8578
+ api: "openai-completions",
8579
+ provider: "openrouter",
8580
+ baseUrl: "https://openrouter.ai/api/v1",
8581
+ reasoning: false,
8582
+ input: ["text", "image"],
8583
+ cost: {
8584
+ input: 10,
8585
+ output: 30,
8586
+ cacheRead: 0,
8587
+ cacheWrite: 0,
8588
+ },
8589
+ contextWindow: 128000,
8590
+ maxTokens: 4096,
8591
+ },
8592
+ "openai/gpt-4-turbo-preview": {
8593
+ id: "openai/gpt-4-turbo-preview",
8594
+ name: "OpenAI: GPT-4 Turbo Preview",
8595
+ api: "openai-completions",
8596
+ provider: "openrouter",
8597
+ baseUrl: "https://openrouter.ai/api/v1",
8598
+ reasoning: false,
8599
+ input: ["text"],
8600
+ cost: {
8601
+ input: 10,
8602
+ output: 30,
8603
+ cacheRead: 0,
8604
+ cacheWrite: 0,
8605
+ },
8606
+ contextWindow: 128000,
8607
+ maxTokens: 4096,
8608
+ },
8199
8609
  "openai/gpt-4.1": {
8200
8610
  id: "openai/gpt-4.1",
8201
8611
  name: "OpenAI: GPT-4.1",
@@ -8741,55 +9151,21 @@ export const MODELS = {
8741
9151
  maxTokens: 128000,
8742
9152
  },
8743
9153
  "openai/gpt-5.4-pro": {
8744
- id: "openai/gpt-5.4-pro",
8745
- name: "OpenAI: GPT-5.4 Pro",
8746
- api: "openai-completions",
8747
- provider: "openrouter",
8748
- baseUrl: "https://openrouter.ai/api/v1",
8749
- reasoning: true,
8750
- input: ["text", "image"],
8751
- cost: {
8752
- input: 30,
8753
- output: 180,
8754
- cacheRead: 0,
8755
- cacheWrite: 0,
8756
- },
8757
- contextWindow: 1050000,
8758
- maxTokens: 128000,
8759
- },
8760
- "openai/gpt-audio": {
8761
- id: "openai/gpt-audio",
8762
- name: "OpenAI: GPT Audio",
8763
- api: "openai-completions",
8764
- provider: "openrouter",
8765
- baseUrl: "https://openrouter.ai/api/v1",
8766
- reasoning: false,
8767
- input: ["text"],
8768
- cost: {
8769
- input: 2.5,
8770
- output: 10,
8771
- cacheRead: 0,
8772
- cacheWrite: 0,
8773
- },
8774
- contextWindow: 128000,
8775
- maxTokens: 16384,
8776
- },
8777
- "openai/gpt-audio-mini": {
8778
- id: "openai/gpt-audio-mini",
8779
- name: "OpenAI: GPT Audio Mini",
9154
+ id: "openai/gpt-5.4-pro",
9155
+ name: "OpenAI: GPT-5.4 Pro",
8780
9156
  api: "openai-completions",
8781
9157
  provider: "openrouter",
8782
9158
  baseUrl: "https://openrouter.ai/api/v1",
8783
- reasoning: false,
8784
- input: ["text"],
9159
+ reasoning: true,
9160
+ input: ["text", "image"],
8785
9161
  cost: {
8786
- input: 0.6,
8787
- output: 2.4,
9162
+ input: 30,
9163
+ output: 180,
8788
9164
  cacheRead: 0,
8789
9165
  cacheWrite: 0,
8790
9166
  },
8791
- contextWindow: 128000,
8792
- maxTokens: 16384,
9167
+ contextWindow: 1050000,
9168
+ maxTokens: 128000,
8793
9169
  },
8794
9170
  "openai/gpt-oss-120b": {
8795
9171
  id: "openai/gpt-oss-120b",
@@ -9112,7 +9488,7 @@ export const MODELS = {
9112
9488
  cacheWrite: 0,
9113
9489
  },
9114
9490
  contextWindow: 32768,
9115
- maxTokens: 32768,
9491
+ maxTokens: 4096,
9116
9492
  },
9117
9493
  "qwen/qwen-max": {
9118
9494
  id: "qwen/qwen-max",
@@ -9352,6 +9728,23 @@ export const MODELS = {
9352
9728
  contextWindow: 40960,
9353
9729
  maxTokens: 40960,
9354
9730
  },
9731
+ "qwen/qwen3-4b:free": {
9732
+ id: "qwen/qwen3-4b:free",
9733
+ name: "Qwen: Qwen3 4B (free)",
9734
+ api: "openai-completions",
9735
+ provider: "openrouter",
9736
+ baseUrl: "https://openrouter.ai/api/v1",
9737
+ reasoning: true,
9738
+ input: ["text"],
9739
+ cost: {
9740
+ input: 0,
9741
+ output: 0,
9742
+ cacheRead: 0,
9743
+ cacheWrite: 0,
9744
+ },
9745
+ contextWindow: 40960,
9746
+ maxTokens: 4096,
9747
+ },
9355
9748
  "qwen/qwen3-8b": {
9356
9749
  id: "qwen/qwen3-8b",
9357
9750
  name: "Qwen: Qwen3 8B",
@@ -9758,7 +10151,7 @@ export const MODELS = {
9758
10151
  cacheWrite: 0,
9759
10152
  },
9760
10153
  contextWindow: 256000,
9761
- maxTokens: 65536,
10154
+ maxTokens: 4096,
9762
10155
  },
9763
10156
  "qwen/qwen3.5-flash-02-23": {
9764
10157
  id: "qwen/qwen3.5-flash-02-23",
@@ -9794,23 +10187,6 @@ export const MODELS = {
9794
10187
  contextWindow: 1000000,
9795
10188
  maxTokens: 65536,
9796
10189
  },
9797
- "qwen/qwen3.6-plus-preview:free": {
9798
- id: "qwen/qwen3.6-plus-preview:free",
9799
- name: "Qwen: Qwen3.6 Plus Preview (free)",
9800
- api: "openai-completions",
9801
- provider: "openrouter",
9802
- baseUrl: "https://openrouter.ai/api/v1",
9803
- reasoning: true,
9804
- input: ["text"],
9805
- cost: {
9806
- input: 0,
9807
- output: 0,
9808
- cacheRead: 0,
9809
- cacheWrite: 0,
9810
- },
9811
- contextWindow: 1000000,
9812
- maxTokens: 32000,
9813
- },
9814
10190
  "qwen/qwq-32b": {
9815
10191
  id: "qwen/qwq-32b",
9816
10192
  name: "Qwen: QwQ 32B",
@@ -9828,23 +10204,6 @@ export const MODELS = {
9828
10204
  contextWindow: 131072,
9829
10205
  maxTokens: 131072,
9830
10206
  },
9831
- "reka/reka-edge": {
9832
- id: "reka/reka-edge",
9833
- name: "Reka Edge",
9834
- api: "openai-completions",
9835
- provider: "openrouter",
9836
- baseUrl: "https://openrouter.ai/api/v1",
9837
- reasoning: false,
9838
- input: ["text", "image"],
9839
- cost: {
9840
- input: 0.09999999999999999,
9841
- output: 0.09999999999999999,
9842
- cacheRead: 0,
9843
- cacheWrite: 0,
9844
- },
9845
- contextWindow: 16384,
9846
- maxTokens: 16384,
9847
- },
9848
10207
  "relace/relace-search": {
9849
10208
  id: "relace/relace-search",
9850
10209
  name: "Relace: Relace Search",
@@ -9907,11 +10266,11 @@ export const MODELS = {
9907
10266
  cost: {
9908
10267
  input: 0.09999999999999999,
9909
10268
  output: 0.3,
9910
- cacheRead: 0,
10269
+ cacheRead: 0.02,
9911
10270
  cacheWrite: 0,
9912
10271
  },
9913
- contextWindow: 262144,
9914
- maxTokens: 65536,
10272
+ contextWindow: 256000,
10273
+ maxTokens: 256000,
9915
10274
  },
9916
10275
  "stepfun/step-3.5-flash:free": {
9917
10276
  id: "stepfun/step-3.5-flash:free",
@@ -10117,23 +10476,6 @@ export const MODELS = {
10117
10476
  contextWindow: 2000000,
10118
10477
  maxTokens: 30000,
10119
10478
  },
10120
- "x-ai/grok-4.20": {
10121
- id: "x-ai/grok-4.20",
10122
- name: "xAI: Grok 4.20",
10123
- api: "openai-completions",
10124
- provider: "openrouter",
10125
- baseUrl: "https://openrouter.ai/api/v1",
10126
- reasoning: true,
10127
- input: ["text", "image"],
10128
- cost: {
10129
- input: 2,
10130
- output: 6,
10131
- cacheRead: 0.19999999999999998,
10132
- cacheWrite: 0,
10133
- },
10134
- contextWindow: 2000000,
10135
- maxTokens: 4096,
10136
- },
10137
10479
  "x-ai/grok-4.20-beta": {
10138
10480
  id: "x-ai/grok-4.20-beta",
10139
10481
  name: "xAI: Grok 4.20 Beta",
@@ -10398,9 +10740,9 @@ export const MODELS = {
10398
10740
  reasoning: true,
10399
10741
  input: ["text"],
10400
10742
  cost: {
10401
- input: 1.2,
10402
- output: 4,
10403
- cacheRead: 0.24,
10743
+ input: 0.96,
10744
+ output: 3.1999999999999997,
10745
+ cacheRead: 0.192,
10404
10746
  cacheWrite: 0,
10405
10747
  },
10406
10748
  contextWindow: 202752,
@@ -10434,12 +10776,12 @@ export const MODELS = {
10434
10776
  reasoning: false,
10435
10777
  input: ["text"],
10436
10778
  cost: {
10437
- input: 0.22,
10438
- output: 0.88,
10439
- cacheRead: 0.11,
10779
+ input: 0.071,
10780
+ output: 0.463,
10781
+ cacheRead: 0,
10440
10782
  cacheWrite: 0,
10441
10783
  },
10442
- contextWindow: 32768,
10784
+ contextWindow: 40960,
10443
10785
  maxTokens: 16384,
10444
10786
  },
10445
10787
  "alibaba/qwen-3-30b": {
@@ -10468,13 +10810,13 @@ export const MODELS = {
10468
10810
  reasoning: true,
10469
10811
  input: ["text"],
10470
10812
  cost: {
10471
- input: 0.16,
10472
- output: 0.64,
10473
- cacheRead: 0,
10813
+ input: 0.29,
10814
+ output: 0.59,
10815
+ cacheRead: 0.145,
10474
10816
  cacheWrite: 0,
10475
10817
  },
10476
- contextWindow: 128000,
10477
- maxTokens: 8192,
10818
+ contextWindow: 131072,
10819
+ maxTokens: 40960,
10478
10820
  },
10479
10821
  "alibaba/qwen3-235b-a22b-thinking": {
10480
10822
  id: "alibaba/qwen3-235b-a22b-thinking",
@@ -10502,13 +10844,13 @@ export const MODELS = {
10502
10844
  reasoning: false,
10503
10845
  input: ["text"],
10504
10846
  cost: {
10505
- input: 1.5,
10506
- output: 7.5,
10507
- cacheRead: 0.3,
10847
+ input: 0.39999999999999997,
10848
+ output: 1.5999999999999999,
10849
+ cacheRead: 0.022,
10508
10850
  cacheWrite: 0,
10509
10851
  },
10510
10852
  contextWindow: 262144,
10511
- maxTokens: 65536,
10853
+ maxTokens: 66536,
10512
10854
  },
10513
10855
  "alibaba/qwen3-coder-30b-a3b": {
10514
10856
  id: "alibaba/qwen3-coder-30b-a3b",
@@ -10663,6 +11005,23 @@ export const MODELS = {
10663
11005
  contextWindow: 1000000,
10664
11006
  maxTokens: 64000,
10665
11007
  },
11008
+ "anthropic/claude-3-haiku": {
11009
+ id: "anthropic/claude-3-haiku",
11010
+ name: "Claude 3 Haiku",
11011
+ api: "anthropic-messages",
11012
+ provider: "vercel-ai-gateway",
11013
+ baseUrl: "https://ai-gateway.vercel.sh",
11014
+ reasoning: false,
11015
+ input: ["text", "image"],
11016
+ cost: {
11017
+ input: 0.25,
11018
+ output: 1.25,
11019
+ cacheRead: 0.03,
11020
+ cacheWrite: 0.3,
11021
+ },
11022
+ contextWindow: 200000,
11023
+ maxTokens: 4096,
11024
+ },
10666
11025
  "anthropic/claude-3.5-haiku": {
10667
11026
  id: "anthropic/claude-3.5-haiku",
10668
11027
  name: "Claude 3.5 Haiku",
@@ -10961,13 +11320,13 @@ export const MODELS = {
10961
11320
  reasoning: true,
10962
11321
  input: ["text"],
10963
11322
  cost: {
10964
- input: 0.56,
10965
- output: 1.68,
10966
- cacheRead: 0.28,
11323
+ input: 0.5,
11324
+ output: 1.5,
11325
+ cacheRead: 0,
10967
11326
  cacheWrite: 0,
10968
11327
  },
10969
11328
  contextWindow: 163840,
10970
- maxTokens: 8192,
11329
+ maxTokens: 16384,
10971
11330
  },
10972
11331
  "deepseek/deepseek-v3.1-terminus": {
10973
11332
  id: "deepseek/deepseek-v3.1-terminus",
@@ -11207,23 +11566,6 @@ export const MODELS = {
11207
11566
  contextWindow: 32000,
11208
11567
  maxTokens: 16384,
11209
11568
  },
11210
- "kwaipilot/kat-coder-pro-v2": {
11211
- id: "kwaipilot/kat-coder-pro-v2",
11212
- name: "Kat Coder Pro V2",
11213
- api: "anthropic-messages",
11214
- provider: "vercel-ai-gateway",
11215
- baseUrl: "https://ai-gateway.vercel.sh",
11216
- reasoning: true,
11217
- input: ["text"],
11218
- cost: {
11219
- input: 0.3,
11220
- output: 1.2,
11221
- cacheRead: 0.06,
11222
- cacheWrite: 0,
11223
- },
11224
- contextWindow: 256000,
11225
- maxTokens: 256000,
11226
- },
11227
11569
  "meituan/longcat-flash-chat": {
11228
11570
  id: "meituan/longcat-flash-chat",
11229
11571
  name: "LongCat Flash Chat",
@@ -11239,7 +11581,7 @@ export const MODELS = {
11239
11581
  cacheWrite: 0,
11240
11582
  },
11241
11583
  contextWindow: 128000,
11242
- maxTokens: 8192,
11584
+ maxTokens: 100000,
11243
11585
  },
11244
11586
  "meituan/longcat-flash-thinking": {
11245
11587
  id: "meituan/longcat-flash-thinking",
@@ -11284,13 +11626,13 @@ export const MODELS = {
11284
11626
  reasoning: false,
11285
11627
  input: ["text"],
11286
11628
  cost: {
11287
- input: 0.22,
11288
- output: 0.22,
11289
- cacheRead: 0,
11629
+ input: 0.09999999999999999,
11630
+ output: 0.09999999999999999,
11631
+ cacheRead: 0.09999999999999999,
11290
11632
  cacheWrite: 0,
11291
11633
  },
11292
11634
  contextWindow: 128000,
11293
- maxTokens: 8192,
11635
+ maxTokens: 16384,
11294
11636
  },
11295
11637
  "meta/llama-3.2-11b": {
11296
11638
  id: "meta/llama-3.2-11b",
@@ -11352,12 +11694,12 @@ export const MODELS = {
11352
11694
  reasoning: false,
11353
11695
  input: ["text", "image"],
11354
11696
  cost: {
11355
- input: 0.35,
11356
- output: 1.15,
11697
+ input: 0.24,
11698
+ output: 0.9700000000000001,
11357
11699
  cacheRead: 0,
11358
11700
  cacheWrite: 0,
11359
11701
  },
11360
- contextWindow: 524288,
11702
+ contextWindow: 128000,
11361
11703
  maxTokens: 8192,
11362
11704
  },
11363
11705
  "meta/llama-4-scout": {
@@ -11802,6 +12144,23 @@ export const MODELS = {
11802
12144
  contextWindow: 131072,
11803
12145
  maxTokens: 131072,
11804
12146
  },
12147
+ "openai/gpt-4-turbo": {
12148
+ id: "openai/gpt-4-turbo",
12149
+ name: "GPT-4 Turbo",
12150
+ api: "anthropic-messages",
12151
+ provider: "vercel-ai-gateway",
12152
+ baseUrl: "https://ai-gateway.vercel.sh",
12153
+ reasoning: false,
12154
+ input: ["text", "image"],
12155
+ cost: {
12156
+ input: 10,
12157
+ output: 30,
12158
+ cacheRead: 0,
12159
+ cacheWrite: 0,
12160
+ },
12161
+ contextWindow: 128000,
12162
+ maxTokens: 4096,
12163
+ },
11805
12164
  "openai/gpt-4.1": {
11806
12165
  id: "openai/gpt-4.1",
11807
12166
  name: "GPT-4.1",
@@ -12244,23 +12603,6 @@ export const MODELS = {
12244
12603
  contextWindow: 1050000,
12245
12604
  maxTokens: 128000,
12246
12605
  },
12247
- "openai/gpt-oss-120b": {
12248
- id: "openai/gpt-oss-120b",
12249
- name: "gpt-oss-120b",
12250
- api: "anthropic-messages",
12251
- provider: "vercel-ai-gateway",
12252
- baseUrl: "https://ai-gateway.vercel.sh",
12253
- reasoning: true,
12254
- input: ["text"],
12255
- cost: {
12256
- input: 0.15,
12257
- output: 0.6,
12258
- cacheRead: 0,
12259
- cacheWrite: 0,
12260
- },
12261
- contextWindow: 128000,
12262
- maxTokens: 8192,
12263
- },
12264
12606
  "openai/gpt-oss-20b": {
12265
12607
  id: "openai/gpt-oss-20b",
12266
12608
  name: "gpt-oss-20b",
@@ -12406,8 +12748,8 @@ export const MODELS = {
12406
12748
  reasoning: false,
12407
12749
  input: ["text", "image"],
12408
12750
  cost: {
12409
- input: 0,
12410
- output: 0,
12751
+ input: 1,
12752
+ output: 1,
12411
12753
  cacheRead: 0,
12412
12754
  cacheWrite: 0,
12413
12755
  },
@@ -12423,8 +12765,8 @@ export const MODELS = {
12423
12765
  reasoning: false,
12424
12766
  input: ["text", "image"],
12425
12767
  cost: {
12426
- input: 0,
12427
- output: 0,
12768
+ input: 3,
12769
+ output: 15,
12428
12770
  cacheRead: 0,
12429
12771
  cacheWrite: 0,
12430
12772
  },
@@ -12618,23 +12960,6 @@ export const MODELS = {
12618
12960
  contextWindow: 2000000,
12619
12961
  maxTokens: 30000,
12620
12962
  },
12621
- "xai/grok-4.20-multi-agent": {
12622
- id: "xai/grok-4.20-multi-agent",
12623
- name: "Grok 4.20 Multi-Agent",
12624
- api: "anthropic-messages",
12625
- provider: "vercel-ai-gateway",
12626
- baseUrl: "https://ai-gateway.vercel.sh",
12627
- reasoning: true,
12628
- input: ["text"],
12629
- cost: {
12630
- input: 2,
12631
- output: 6,
12632
- cacheRead: 0.19999999999999998,
12633
- cacheWrite: 0,
12634
- },
12635
- contextWindow: 2000000,
12636
- maxTokens: 2000000,
12637
- },
12638
12963
  "xai/grok-4.20-multi-agent-beta": {
12639
12964
  id: "xai/grok-4.20-multi-agent-beta",
12640
12965
  name: "Grok 4.20 Multi Agent Beta",
@@ -12652,23 +12977,6 @@ export const MODELS = {
12652
12977
  contextWindow: 2000000,
12653
12978
  maxTokens: 2000000,
12654
12979
  },
12655
- "xai/grok-4.20-non-reasoning": {
12656
- id: "xai/grok-4.20-non-reasoning",
12657
- name: "Grok 4.20 Non-Reasoning",
12658
- api: "anthropic-messages",
12659
- provider: "vercel-ai-gateway",
12660
- baseUrl: "https://ai-gateway.vercel.sh",
12661
- reasoning: false,
12662
- input: ["text", "image"],
12663
- cost: {
12664
- input: 2,
12665
- output: 6,
12666
- cacheRead: 0.19999999999999998,
12667
- cacheWrite: 0,
12668
- },
12669
- contextWindow: 2000000,
12670
- maxTokens: 2000000,
12671
- },
12672
12980
  "xai/grok-4.20-non-reasoning-beta": {
12673
12981
  id: "xai/grok-4.20-non-reasoning-beta",
12674
12982
  name: "Grok 4.20 Beta Non-Reasoning",
@@ -12686,23 +12994,6 @@ export const MODELS = {
12686
12994
  contextWindow: 2000000,
12687
12995
  maxTokens: 2000000,
12688
12996
  },
12689
- "xai/grok-4.20-reasoning": {
12690
- id: "xai/grok-4.20-reasoning",
12691
- name: "Grok 4.20 Reasoning",
12692
- api: "anthropic-messages",
12693
- provider: "vercel-ai-gateway",
12694
- baseUrl: "https://ai-gateway.vercel.sh",
12695
- reasoning: true,
12696
- input: ["text", "image"],
12697
- cost: {
12698
- input: 2,
12699
- output: 6,
12700
- cacheRead: 0.19999999999999998,
12701
- cacheWrite: 0,
12702
- },
12703
- contextWindow: 2000000,
12704
- maxTokens: 2000000,
12705
- },
12706
12997
  "xai/grok-4.20-reasoning-beta": {
12707
12998
  id: "xai/grok-4.20-reasoning-beta",
12708
12999
  name: "Grok 4.20 Beta Reasoning",
@@ -12746,9 +13037,9 @@ export const MODELS = {
12746
13037
  reasoning: true,
12747
13038
  input: ["text"],
12748
13039
  cost: {
12749
- input: 0.09,
12750
- output: 0.29,
12751
- cacheRead: 0.045,
13040
+ input: 0.09999999999999999,
13041
+ output: 0.3,
13042
+ cacheRead: 0.02,
12752
13043
  cacheWrite: 0,
12753
13044
  },
12754
13045
  contextWindow: 262144,
@@ -12884,7 +13175,7 @@ export const MODELS = {
12884
13175
  cost: {
12885
13176
  input: 0.6,
12886
13177
  output: 2.2,
12887
- cacheRead: 0.11,
13178
+ cacheRead: 0,
12888
13179
  cacheWrite: 0,
12889
13180
  },
12890
13181
  contextWindow: 200000,
@@ -13514,24 +13805,6 @@ export const MODELS = {
13514
13805
  contextWindow: 200000,
13515
13806
  maxTokens: 131072,
13516
13807
  },
13517
- "glm-4.7-flashx": {
13518
- id: "glm-4.7-flashx",
13519
- name: "GLM-4.7-FlashX",
13520
- api: "openai-completions",
13521
- provider: "zai",
13522
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
13523
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
13524
- reasoning: true,
13525
- input: ["text"],
13526
- cost: {
13527
- input: 0.07,
13528
- output: 0.4,
13529
- cacheRead: 0.01,
13530
- cacheWrite: 0,
13531
- },
13532
- contextWindow: 200000,
13533
- maxTokens: 131072,
13534
- },
13535
13808
  "glm-5": {
13536
13809
  id: "glm-5",
13537
13810
  name: "GLM-5",