gsd-pi 2.42.0-dev.97e9e30 → 2.42.0-dev.eedc83f

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (167) hide show
  1. package/README.md +23 -0
  2. package/dist/cli.js +15 -1
  3. package/dist/resource-loader.js +39 -6
  4. package/dist/resources/extensions/async-jobs/async-bash-tool.js +52 -4
  5. package/dist/resources/extensions/gsd/auto-prompts.js +1 -1
  6. package/dist/resources/extensions/gsd/bootstrap/register-hooks.js +11 -5
  7. package/dist/resources/extensions/gsd/detection.js +19 -0
  8. package/dist/resources/extensions/gsd/doctor-checks.js +31 -1
  9. package/dist/resources/extensions/gsd/doctor-providers.js +10 -0
  10. package/dist/resources/extensions/gsd/forensics.js +84 -0
  11. package/dist/resources/extensions/gsd/git-constants.js +1 -0
  12. package/dist/resources/extensions/gsd/git-service.js +68 -2
  13. package/dist/resources/extensions/gsd/native-git-bridge.js +1 -0
  14. package/dist/resources/extensions/gsd/preferences-types.js +1 -0
  15. package/dist/resources/extensions/gsd/preferences.js +59 -8
  16. package/dist/resources/extensions/gsd/prompts/forensics.md +12 -5
  17. package/dist/resources/extensions/gsd/repo-identity.js +46 -5
  18. package/dist/resources/extensions/gsd/service-tier.js +13 -4
  19. package/dist/resources/extensions/gsd/session-lock.js +2 -2
  20. package/dist/resources/extensions/gsd/worktree-resolver.js +2 -2
  21. package/dist/resources/extensions/mcp-client/index.js +2 -1
  22. package/dist/resources/extensions/search-the-web/tool-search.js +3 -3
  23. package/dist/web/standalone/.next/BUILD_ID +1 -1
  24. package/dist/web/standalone/.next/app-path-routes-manifest.json +12 -12
  25. package/dist/web/standalone/.next/build-manifest.json +2 -2
  26. package/dist/web/standalone/.next/prerender-manifest.json +3 -3
  27. package/dist/web/standalone/.next/server/app/_global-error.html +2 -2
  28. package/dist/web/standalone/.next/server/app/_global-error.rsc +1 -1
  29. package/dist/web/standalone/.next/server/app/_global-error.segments/_full.segment.rsc +1 -1
  30. package/dist/web/standalone/.next/server/app/_global-error.segments/_global-error/__PAGE__.segment.rsc +1 -1
  31. package/dist/web/standalone/.next/server/app/_global-error.segments/_global-error.segment.rsc +1 -1
  32. package/dist/web/standalone/.next/server/app/_global-error.segments/_head.segment.rsc +1 -1
  33. package/dist/web/standalone/.next/server/app/_global-error.segments/_index.segment.rsc +1 -1
  34. package/dist/web/standalone/.next/server/app/_global-error.segments/_tree.segment.rsc +1 -1
  35. package/dist/web/standalone/.next/server/app/_not-found.html +1 -1
  36. package/dist/web/standalone/.next/server/app/_not-found.rsc +1 -1
  37. package/dist/web/standalone/.next/server/app/_not-found.segments/_full.segment.rsc +1 -1
  38. package/dist/web/standalone/.next/server/app/_not-found.segments/_head.segment.rsc +1 -1
  39. package/dist/web/standalone/.next/server/app/_not-found.segments/_index.segment.rsc +1 -1
  40. package/dist/web/standalone/.next/server/app/_not-found.segments/_not-found/__PAGE__.segment.rsc +1 -1
  41. package/dist/web/standalone/.next/server/app/_not-found.segments/_not-found.segment.rsc +1 -1
  42. package/dist/web/standalone/.next/server/app/_not-found.segments/_tree.segment.rsc +1 -1
  43. package/dist/web/standalone/.next/server/app/api/git/route.js +1 -1
  44. package/dist/web/standalone/.next/server/app/index.html +1 -1
  45. package/dist/web/standalone/.next/server/app/index.rsc +1 -1
  46. package/dist/web/standalone/.next/server/app/index.segments/__PAGE__.segment.rsc +1 -1
  47. package/dist/web/standalone/.next/server/app/index.segments/_full.segment.rsc +1 -1
  48. package/dist/web/standalone/.next/server/app/index.segments/_head.segment.rsc +1 -1
  49. package/dist/web/standalone/.next/server/app/index.segments/_index.segment.rsc +1 -1
  50. package/dist/web/standalone/.next/server/app/index.segments/_tree.segment.rsc +1 -1
  51. package/dist/web/standalone/.next/server/app-paths-manifest.json +12 -12
  52. package/dist/web/standalone/.next/server/chunks/229.js +2 -2
  53. package/dist/web/standalone/.next/server/pages/404.html +1 -1
  54. package/dist/web/standalone/.next/server/pages/500.html +2 -2
  55. package/dist/web/standalone/.next/server/server-reference-manifest.json +1 -1
  56. package/dist/web-mode.d.ts +2 -0
  57. package/dist/web-mode.js +40 -4
  58. package/package.json +1 -1
  59. package/packages/pi-agent-core/dist/agent.d.ts.map +1 -1
  60. package/packages/pi-agent-core/dist/agent.js +2 -0
  61. package/packages/pi-agent-core/dist/agent.js.map +1 -1
  62. package/packages/pi-agent-core/dist/types.d.ts +6 -0
  63. package/packages/pi-agent-core/dist/types.d.ts.map +1 -1
  64. package/packages/pi-agent-core/dist/types.js.map +1 -1
  65. package/packages/pi-agent-core/src/agent.test.ts +53 -0
  66. package/packages/pi-agent-core/src/agent.ts +3 -0
  67. package/packages/pi-agent-core/src/types.ts +6 -0
  68. package/packages/pi-agent-core/tsconfig.json +1 -1
  69. package/packages/pi-ai/dist/models.d.ts +5 -3
  70. package/packages/pi-ai/dist/models.d.ts.map +1 -1
  71. package/packages/pi-ai/dist/models.generated.d.ts +801 -1468
  72. package/packages/pi-ai/dist/models.generated.d.ts.map +1 -1
  73. package/packages/pi-ai/dist/models.generated.js +1135 -1588
  74. package/packages/pi-ai/dist/models.generated.js.map +1 -1
  75. package/packages/pi-ai/dist/models.js.map +1 -1
  76. package/packages/pi-ai/dist/utils/oauth/github-copilot.d.ts.map +1 -1
  77. package/packages/pi-ai/dist/utils/oauth/github-copilot.js +60 -2
  78. package/packages/pi-ai/dist/utils/oauth/github-copilot.js.map +1 -1
  79. package/packages/pi-ai/scripts/generate-models.ts +1543 -0
  80. package/packages/pi-ai/src/models.generated.ts +1140 -1593
  81. package/packages/pi-ai/src/models.ts +7 -4
  82. package/packages/pi-ai/src/utils/oauth/github-copilot.ts +74 -2
  83. package/packages/pi-coding-agent/dist/core/agent-session.d.ts.map +1 -1
  84. package/packages/pi-coding-agent/dist/core/agent-session.js +8 -1
  85. package/packages/pi-coding-agent/dist/core/agent-session.js.map +1 -1
  86. package/packages/pi-coding-agent/dist/core/auth-storage.d.ts +7 -0
  87. package/packages/pi-coding-agent/dist/core/auth-storage.d.ts.map +1 -1
  88. package/packages/pi-coding-agent/dist/core/auth-storage.js +29 -2
  89. package/packages/pi-coding-agent/dist/core/auth-storage.js.map +1 -1
  90. package/packages/pi-coding-agent/dist/core/auth-storage.test.js +60 -0
  91. package/packages/pi-coding-agent/dist/core/auth-storage.test.js.map +1 -1
  92. package/packages/pi-coding-agent/dist/core/extensions/loader.d.ts.map +1 -1
  93. package/packages/pi-coding-agent/dist/core/extensions/loader.js +18 -0
  94. package/packages/pi-coding-agent/dist/core/extensions/loader.js.map +1 -1
  95. package/packages/pi-coding-agent/dist/core/lsp/client.d.ts.map +1 -1
  96. package/packages/pi-coding-agent/dist/core/lsp/client.js +23 -0
  97. package/packages/pi-coding-agent/dist/core/lsp/client.js.map +1 -1
  98. package/packages/pi-coding-agent/dist/core/model-registry.d.ts.map +1 -1
  99. package/packages/pi-coding-agent/dist/core/model-registry.js +2 -0
  100. package/packages/pi-coding-agent/dist/core/model-registry.js.map +1 -1
  101. package/packages/pi-coding-agent/dist/core/package-manager.d.ts +6 -0
  102. package/packages/pi-coding-agent/dist/core/package-manager.d.ts.map +1 -1
  103. package/packages/pi-coding-agent/dist/core/package-manager.js +63 -11
  104. package/packages/pi-coding-agent/dist/core/package-manager.js.map +1 -1
  105. package/packages/pi-coding-agent/dist/core/resource-loader.d.ts +9 -0
  106. package/packages/pi-coding-agent/dist/core/resource-loader.d.ts.map +1 -1
  107. package/packages/pi-coding-agent/dist/core/resource-loader.js +20 -6
  108. package/packages/pi-coding-agent/dist/core/resource-loader.js.map +1 -1
  109. package/packages/pi-coding-agent/dist/core/system-prompt.d.ts.map +1 -1
  110. package/packages/pi-coding-agent/dist/core/system-prompt.js +6 -5
  111. package/packages/pi-coding-agent/dist/core/system-prompt.js.map +1 -1
  112. package/packages/pi-coding-agent/dist/modes/interactive/components/extension-editor.d.ts.map +1 -1
  113. package/packages/pi-coding-agent/dist/modes/interactive/components/extension-editor.js +3 -0
  114. package/packages/pi-coding-agent/dist/modes/interactive/components/extension-editor.js.map +1 -1
  115. package/packages/pi-coding-agent/dist/modes/interactive/components/footer.d.ts.map +1 -1
  116. package/packages/pi-coding-agent/dist/modes/interactive/components/footer.js +9 -6
  117. package/packages/pi-coding-agent/dist/modes/interactive/components/footer.js.map +1 -1
  118. package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.d.ts.map +1 -1
  119. package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.js +30 -10
  120. package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.js.map +1 -1
  121. package/packages/pi-coding-agent/src/core/agent-session.ts +7 -1
  122. package/packages/pi-coding-agent/src/core/auth-storage.test.ts +68 -0
  123. package/packages/pi-coding-agent/src/core/auth-storage.ts +30 -2
  124. package/packages/pi-coding-agent/src/core/extensions/loader.ts +18 -0
  125. package/packages/pi-coding-agent/src/core/lsp/client.ts +29 -0
  126. package/packages/pi-coding-agent/src/core/model-registry.ts +3 -0
  127. package/packages/pi-coding-agent/src/core/package-manager.ts +99 -58
  128. package/packages/pi-coding-agent/src/core/resource-loader.ts +24 -6
  129. package/packages/pi-coding-agent/src/core/system-prompt.ts +6 -5
  130. package/packages/pi-coding-agent/src/modes/interactive/components/extension-editor.ts +3 -0
  131. package/packages/pi-coding-agent/src/modes/interactive/components/footer.ts +10 -6
  132. package/packages/pi-coding-agent/src/modes/interactive/interactive-mode.ts +31 -11
  133. package/src/resources/extensions/async-jobs/async-bash-timeout.test.ts +122 -0
  134. package/src/resources/extensions/async-jobs/async-bash-tool.ts +40 -4
  135. package/src/resources/extensions/gsd/auto-prompts.ts +1 -1
  136. package/src/resources/extensions/gsd/bootstrap/register-hooks.ts +13 -5
  137. package/src/resources/extensions/gsd/detection.ts +19 -0
  138. package/src/resources/extensions/gsd/doctor-checks.ts +32 -1
  139. package/src/resources/extensions/gsd/doctor-providers.ts +13 -0
  140. package/src/resources/extensions/gsd/doctor-types.ts +1 -0
  141. package/src/resources/extensions/gsd/forensics.ts +92 -0
  142. package/src/resources/extensions/gsd/git-constants.ts +1 -0
  143. package/src/resources/extensions/gsd/git-service.ts +71 -2
  144. package/src/resources/extensions/gsd/native-git-bridge.ts +1 -0
  145. package/src/resources/extensions/gsd/preferences-types.ts +3 -0
  146. package/src/resources/extensions/gsd/preferences.ts +62 -6
  147. package/src/resources/extensions/gsd/prompts/forensics.md +12 -5
  148. package/src/resources/extensions/gsd/repo-identity.ts +48 -5
  149. package/src/resources/extensions/gsd/service-tier.ts +17 -4
  150. package/src/resources/extensions/gsd/session-lock.ts +2 -2
  151. package/src/resources/extensions/gsd/tests/activity-log.test.ts +31 -69
  152. package/src/resources/extensions/gsd/tests/forensics-dedup.test.ts +48 -0
  153. package/src/resources/extensions/gsd/tests/forensics-issue-routing.test.ts +43 -0
  154. package/src/resources/extensions/gsd/tests/git-locale.test.ts +133 -0
  155. package/src/resources/extensions/gsd/tests/git-service.test.ts +49 -0
  156. package/src/resources/extensions/gsd/tests/journal.test.ts +82 -127
  157. package/src/resources/extensions/gsd/tests/manifest-status.test.ts +73 -82
  158. package/src/resources/extensions/gsd/tests/service-tier.test.ts +30 -1
  159. package/src/resources/extensions/gsd/tests/symlink-numbered-variants.test.ts +151 -0
  160. package/src/resources/extensions/gsd/tests/verification-gate.test.ts +156 -263
  161. package/src/resources/extensions/gsd/tests/worktree-health-dispatch.test.ts +35 -78
  162. package/src/resources/extensions/gsd/tests/worktree-manager.test.ts +81 -74
  163. package/src/resources/extensions/gsd/worktree-resolver.ts +2 -2
  164. package/src/resources/extensions/mcp-client/index.ts +5 -1
  165. package/src/resources/extensions/search-the-web/tool-search.ts +3 -3
  166. /package/dist/web/standalone/.next/static/{PXrI5DoWsm7rwAVnEU2rD → JUBX5FUR73jiViQU5a-Cx}/_buildManifest.js +0 -0
  167. /package/dist/web/standalone/.next/static/{PXrI5DoWsm7rwAVnEU2rD → JUBX5FUR73jiViQU5a-Cx}/_ssgManifest.js +0 -0
@@ -87,40 +87,6 @@ export const MODELS = {
87
87
  contextWindow: 300000,
88
88
  maxTokens: 8192,
89
89
  },
90
- "amazon.titan-text-express-v1": {
91
- id: "amazon.titan-text-express-v1",
92
- name: "Titan Text G1 - Express",
93
- api: "bedrock-converse-stream",
94
- provider: "amazon-bedrock",
95
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
96
- reasoning: false,
97
- input: ["text"],
98
- cost: {
99
- input: 0.2,
100
- output: 0.6,
101
- cacheRead: 0,
102
- cacheWrite: 0,
103
- },
104
- contextWindow: 128000,
105
- maxTokens: 4096,
106
- },
107
- "amazon.titan-text-express-v1:0:8k": {
108
- id: "amazon.titan-text-express-v1:0:8k",
109
- name: "Titan Text G1 - Express",
110
- api: "bedrock-converse-stream",
111
- provider: "amazon-bedrock",
112
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
113
- reasoning: false,
114
- input: ["text"],
115
- cost: {
116
- input: 0.2,
117
- output: 0.6,
118
- cacheRead: 0,
119
- cacheWrite: 0,
120
- },
121
- contextWindow: 128000,
122
- maxTokens: 4096,
123
- },
124
90
  "anthropic.claude-3-5-haiku-20241022-v1:0": {
125
91
  id: "anthropic.claude-3-5-haiku-20241022-v1:0",
126
92
  name: "Claude Haiku 3.5",
@@ -206,40 +172,6 @@ export const MODELS = {
206
172
  contextWindow: 200000,
207
173
  maxTokens: 4096,
208
174
  },
209
- "anthropic.claude-3-opus-20240229-v1:0": {
210
- id: "anthropic.claude-3-opus-20240229-v1:0",
211
- name: "Claude Opus 3",
212
- api: "bedrock-converse-stream",
213
- provider: "amazon-bedrock",
214
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
215
- reasoning: false,
216
- input: ["text", "image"],
217
- cost: {
218
- input: 15,
219
- output: 75,
220
- cacheRead: 0,
221
- cacheWrite: 0,
222
- },
223
- contextWindow: 200000,
224
- maxTokens: 4096,
225
- },
226
- "anthropic.claude-3-sonnet-20240229-v1:0": {
227
- id: "anthropic.claude-3-sonnet-20240229-v1:0",
228
- name: "Claude Sonnet 3",
229
- api: "bedrock-converse-stream",
230
- provider: "amazon-bedrock",
231
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
232
- reasoning: false,
233
- input: ["text", "image"],
234
- cost: {
235
- input: 3,
236
- output: 15,
237
- cacheRead: 0,
238
- cacheWrite: 0,
239
- },
240
- contextWindow: 200000,
241
- maxTokens: 4096,
242
- },
243
175
  "anthropic.claude-haiku-4-5-20251001-v1:0": {
244
176
  id: "anthropic.claude-haiku-4-5-20251001-v1:0",
245
177
  name: "Claude Haiku 4.5",
@@ -322,7 +254,7 @@ export const MODELS = {
322
254
  cacheRead: 0.5,
323
255
  cacheWrite: 6.25,
324
256
  },
325
- contextWindow: 200000,
257
+ contextWindow: 1000000,
326
258
  maxTokens: 128000,
327
259
  },
328
260
  "anthropic.claude-sonnet-4-20250514-v1:0": {
@@ -373,43 +305,9 @@ export const MODELS = {
373
305
  cacheRead: 0.3,
374
306
  cacheWrite: 3.75,
375
307
  },
376
- contextWindow: 200000,
308
+ contextWindow: 1000000,
377
309
  maxTokens: 64000,
378
310
  },
379
- "cohere.command-r-plus-v1:0": {
380
- id: "cohere.command-r-plus-v1:0",
381
- name: "Command R+",
382
- api: "bedrock-converse-stream",
383
- provider: "amazon-bedrock",
384
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
385
- reasoning: false,
386
- input: ["text"],
387
- cost: {
388
- input: 3,
389
- output: 15,
390
- cacheRead: 0,
391
- cacheWrite: 0,
392
- },
393
- contextWindow: 128000,
394
- maxTokens: 4096,
395
- },
396
- "cohere.command-r-v1:0": {
397
- id: "cohere.command-r-v1:0",
398
- name: "Command R",
399
- api: "bedrock-converse-stream",
400
- provider: "amazon-bedrock",
401
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
402
- reasoning: false,
403
- input: ["text"],
404
- cost: {
405
- input: 0.5,
406
- output: 1.5,
407
- cacheRead: 0,
408
- cacheWrite: 0,
409
- },
410
- contextWindow: 128000,
411
- maxTokens: 4096,
412
- },
413
311
  "deepseek.r1-v1:0": {
414
312
  id: "deepseek.r1-v1:0",
415
313
  name: "DeepSeek-R1",
@@ -444,8 +342,8 @@ export const MODELS = {
444
342
  contextWindow: 163840,
445
343
  maxTokens: 81920,
446
344
  },
447
- "deepseek.v3.2-v1:0": {
448
- id: "deepseek.v3.2-v1:0",
345
+ "deepseek.v3.2": {
346
+ id: "deepseek.v3.2",
449
347
  name: "DeepSeek-V3.2",
450
348
  api: "bedrock-converse-stream",
451
349
  provider: "amazon-bedrock",
@@ -509,7 +407,7 @@ export const MODELS = {
509
407
  cacheRead: 0.5,
510
408
  cacheWrite: 6.25,
511
409
  },
512
- contextWindow: 200000,
410
+ contextWindow: 1000000,
513
411
  maxTokens: 128000,
514
412
  },
515
413
  "eu.anthropic.claude-sonnet-4-20250514-v1:0": {
@@ -560,7 +458,7 @@ export const MODELS = {
560
458
  cacheRead: 0.3,
561
459
  cacheWrite: 3.75,
562
460
  },
563
- contextWindow: 200000,
461
+ contextWindow: 1000000,
564
462
  maxTokens: 64000,
565
463
  },
566
464
  "global.anthropic.claude-haiku-4-5-20251001-v1:0": {
@@ -611,7 +509,7 @@ export const MODELS = {
611
509
  cacheRead: 0.5,
612
510
  cacheWrite: 6.25,
613
511
  },
614
- contextWindow: 200000,
512
+ contextWindow: 1000000,
615
513
  maxTokens: 128000,
616
514
  },
617
515
  "global.anthropic.claude-sonnet-4-20250514-v1:0": {
@@ -662,7 +560,7 @@ export const MODELS = {
662
560
  cacheRead: 0.3,
663
561
  cacheWrite: 3.75,
664
562
  },
665
- contextWindow: 200000,
563
+ contextWindow: 1000000,
666
564
  maxTokens: 64000,
667
565
  },
668
566
  "google.gemma-3-27b-it": {
@@ -699,6 +597,23 @@ export const MODELS = {
699
597
  contextWindow: 128000,
700
598
  maxTokens: 4096,
701
599
  },
600
+ "meta.llama3-1-405b-instruct-v1:0": {
601
+ id: "meta.llama3-1-405b-instruct-v1:0",
602
+ name: "Llama 3.1 405B Instruct",
603
+ api: "bedrock-converse-stream",
604
+ provider: "amazon-bedrock",
605
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
606
+ reasoning: false,
607
+ input: ["text"],
608
+ cost: {
609
+ input: 2.4,
610
+ output: 2.4,
611
+ cacheRead: 0,
612
+ cacheWrite: 0,
613
+ },
614
+ contextWindow: 128000,
615
+ maxTokens: 4096,
616
+ },
702
617
  "meta.llama3-1-70b-instruct-v1:0": {
703
618
  id: "meta.llama3-1-70b-instruct-v1:0",
704
619
  name: "Llama 3.1 70B Instruct",
@@ -886,6 +801,40 @@ export const MODELS = {
886
801
  contextWindow: 204800,
887
802
  maxTokens: 131072,
888
803
  },
804
+ "mistral.devstral-2-123b": {
805
+ id: "mistral.devstral-2-123b",
806
+ name: "Devstral 2 123B",
807
+ api: "bedrock-converse-stream",
808
+ provider: "amazon-bedrock",
809
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
810
+ reasoning: false,
811
+ input: ["text"],
812
+ cost: {
813
+ input: 0.4,
814
+ output: 2,
815
+ cacheRead: 0,
816
+ cacheWrite: 0,
817
+ },
818
+ contextWindow: 256000,
819
+ maxTokens: 8192,
820
+ },
821
+ "mistral.magistral-small-2509": {
822
+ id: "mistral.magistral-small-2509",
823
+ name: "Magistral Small 1.2",
824
+ api: "bedrock-converse-stream",
825
+ provider: "amazon-bedrock",
826
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
827
+ reasoning: true,
828
+ input: ["text", "image"],
829
+ cost: {
830
+ input: 0.5,
831
+ output: 1.5,
832
+ cacheRead: 0,
833
+ cacheWrite: 0,
834
+ },
835
+ contextWindow: 128000,
836
+ maxTokens: 40000,
837
+ },
889
838
  "mistral.ministral-3-14b-instruct": {
890
839
  id: "mistral.ministral-3-14b-instruct",
891
840
  name: "Ministral 14B 3.0",
@@ -903,6 +852,23 @@ export const MODELS = {
903
852
  contextWindow: 128000,
904
853
  maxTokens: 4096,
905
854
  },
855
+ "mistral.ministral-3-3b-instruct": {
856
+ id: "mistral.ministral-3-3b-instruct",
857
+ name: "Ministral 3 3B",
858
+ api: "bedrock-converse-stream",
859
+ provider: "amazon-bedrock",
860
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
861
+ reasoning: false,
862
+ input: ["text", "image"],
863
+ cost: {
864
+ input: 0.1,
865
+ output: 0.1,
866
+ cacheRead: 0,
867
+ cacheWrite: 0,
868
+ },
869
+ contextWindow: 256000,
870
+ maxTokens: 8192,
871
+ },
906
872
  "mistral.ministral-3-8b-instruct": {
907
873
  id: "mistral.ministral-3-8b-instruct",
908
874
  name: "Ministral 3 8B",
@@ -920,22 +886,39 @@ export const MODELS = {
920
886
  contextWindow: 128000,
921
887
  maxTokens: 4096,
922
888
  },
923
- "mistral.mistral-large-2402-v1:0": {
924
- id: "mistral.mistral-large-2402-v1:0",
925
- name: "Mistral Large (24.02)",
889
+ "mistral.mistral-large-3-675b-instruct": {
890
+ id: "mistral.mistral-large-3-675b-instruct",
891
+ name: "Mistral Large 3",
926
892
  api: "bedrock-converse-stream",
927
893
  provider: "amazon-bedrock",
928
894
  baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
929
895
  reasoning: false,
930
- input: ["text"],
896
+ input: ["text", "image"],
931
897
  cost: {
932
898
  input: 0.5,
933
899
  output: 1.5,
934
900
  cacheRead: 0,
935
901
  cacheWrite: 0,
936
902
  },
903
+ contextWindow: 256000,
904
+ maxTokens: 8192,
905
+ },
906
+ "mistral.pixtral-large-2502-v1:0": {
907
+ id: "mistral.pixtral-large-2502-v1:0",
908
+ name: "Pixtral Large (25.02)",
909
+ api: "bedrock-converse-stream",
910
+ provider: "amazon-bedrock",
911
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
912
+ reasoning: false,
913
+ input: ["text", "image"],
914
+ cost: {
915
+ input: 2,
916
+ output: 6,
917
+ cacheRead: 0,
918
+ cacheWrite: 0,
919
+ },
937
920
  contextWindow: 128000,
938
- maxTokens: 4096,
921
+ maxTokens: 8192,
939
922
  },
940
923
  "mistral.voxtral-mini-3b-2507": {
941
924
  id: "mistral.voxtral-mini-3b-2507",
@@ -1022,6 +1005,23 @@ export const MODELS = {
1022
1005
  contextWindow: 128000,
1023
1006
  maxTokens: 4096,
1024
1007
  },
1008
+ "nvidia.nemotron-nano-3-30b": {
1009
+ id: "nvidia.nemotron-nano-3-30b",
1010
+ name: "NVIDIA Nemotron Nano 3 30B",
1011
+ api: "bedrock-converse-stream",
1012
+ provider: "amazon-bedrock",
1013
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
1014
+ reasoning: true,
1015
+ input: ["text"],
1016
+ cost: {
1017
+ input: 0.06,
1018
+ output: 0.24,
1019
+ cacheRead: 0,
1020
+ cacheWrite: 0,
1021
+ },
1022
+ contextWindow: 128000,
1023
+ maxTokens: 4096,
1024
+ },
1025
1025
  "nvidia.nemotron-nano-9b-v2": {
1026
1026
  id: "nvidia.nemotron-nano-9b-v2",
1027
1027
  name: "NVIDIA Nemotron Nano 9B v2",
@@ -1291,7 +1291,7 @@ export const MODELS = {
1291
1291
  cacheRead: 0.5,
1292
1292
  cacheWrite: 6.25,
1293
1293
  },
1294
- contextWindow: 200000,
1294
+ contextWindow: 1000000,
1295
1295
  maxTokens: 128000,
1296
1296
  },
1297
1297
  "us.anthropic.claude-sonnet-4-20250514-v1:0": {
@@ -1342,7 +1342,7 @@ export const MODELS = {
1342
1342
  cacheRead: 0.3,
1343
1343
  cacheWrite: 3.75,
1344
1344
  },
1345
- contextWindow: 200000,
1345
+ contextWindow: 1000000,
1346
1346
  maxTokens: 64000,
1347
1347
  },
1348
1348
  "writer.palmyra-x4-v1:0": {
@@ -1718,23 +1718,6 @@ export const MODELS = {
1718
1718
  cacheRead: 0.5,
1719
1719
  cacheWrite: 6.25,
1720
1720
  },
1721
- contextWindow: 200000,
1722
- maxTokens: 128000,
1723
- },
1724
- "claude-opus-4-6[1m]": {
1725
- id: "claude-opus-4-6[1m]",
1726
- name: "Claude Opus 4.6 (1M)",
1727
- api: "anthropic-messages",
1728
- provider: "anthropic",
1729
- baseUrl: "https://api.anthropic.com",
1730
- reasoning: true,
1731
- input: ["text", "image"],
1732
- cost: {
1733
- input: 5,
1734
- output: 25,
1735
- cacheRead: 0.5,
1736
- cacheWrite: 6.25,
1737
- },
1738
1721
  contextWindow: 1000000,
1739
1722
  maxTokens: 128000,
1740
1723
  },
@@ -1820,216 +1803,44 @@ export const MODELS = {
1820
1803
  cacheRead: 0.3,
1821
1804
  cacheWrite: 3.75,
1822
1805
  },
1823
- contextWindow: 200000,
1806
+ contextWindow: 1000000,
1824
1807
  maxTokens: 64000,
1825
1808
  },
1826
1809
  },
1827
- "anthropic-vertex": {
1828
- "claude-opus-4-6": {
1829
- id: "claude-opus-4-6",
1830
- name: "Claude Opus 4.6 (Vertex)",
1831
- api: "anthropic-vertex",
1832
- provider: "anthropic-vertex",
1833
- baseUrl: "https://us-central1-aiplatform.googleapis.com",
1810
+ "azure-openai-responses": {
1811
+ "codex-mini-latest": {
1812
+ id: "codex-mini-latest",
1813
+ name: "Codex Mini",
1814
+ api: "azure-openai-responses",
1815
+ provider: "azure-openai-responses",
1816
+ baseUrl: "",
1834
1817
  reasoning: true,
1835
- input: ["text", "image"],
1818
+ input: ["text"],
1836
1819
  cost: {
1837
- input: 15,
1838
- output: 75,
1839
- cacheRead: 1.5,
1840
- cacheWrite: 18.75,
1820
+ input: 1.5,
1821
+ output: 6,
1822
+ cacheRead: 0.375,
1823
+ cacheWrite: 0,
1841
1824
  },
1842
1825
  contextWindow: 200000,
1843
- maxTokens: 128000,
1826
+ maxTokens: 100000,
1844
1827
  },
1845
- "claude-opus-4-6[1m]": {
1846
- id: "claude-opus-4-6[1m]",
1847
- name: "Claude Opus 4.6 1M (Vertex)",
1848
- api: "anthropic-vertex",
1849
- provider: "anthropic-vertex",
1850
- baseUrl: "https://us-central1-aiplatform.googleapis.com",
1851
- reasoning: true,
1852
- input: ["text", "image"],
1828
+ "gpt-4": {
1829
+ id: "gpt-4",
1830
+ name: "GPT-4",
1831
+ api: "azure-openai-responses",
1832
+ provider: "azure-openai-responses",
1833
+ baseUrl: "",
1834
+ reasoning: false,
1835
+ input: ["text"],
1853
1836
  cost: {
1854
- input: 15,
1855
- output: 75,
1856
- cacheRead: 1.5,
1857
- cacheWrite: 18.75,
1837
+ input: 30,
1838
+ output: 60,
1839
+ cacheRead: 0,
1840
+ cacheWrite: 0,
1858
1841
  },
1859
- contextWindow: 1000000,
1860
- maxTokens: 128000,
1861
- },
1862
- "claude-sonnet-4-6": {
1863
- id: "claude-sonnet-4-6",
1864
- name: "Claude Sonnet 4.6 (Vertex)",
1865
- api: "anthropic-vertex",
1866
- provider: "anthropic-vertex",
1867
- baseUrl: "https://us-central1-aiplatform.googleapis.com",
1868
- reasoning: true,
1869
- input: ["text", "image"],
1870
- cost: {
1871
- input: 3,
1872
- output: 15,
1873
- cacheRead: 0.3,
1874
- cacheWrite: 3.75,
1875
- },
1876
- contextWindow: 200000,
1877
- maxTokens: 64000,
1878
- },
1879
- "claude-sonnet-4-6[1m]": {
1880
- id: "claude-sonnet-4-6[1m]",
1881
- name: "Claude Sonnet 4.6 1M (Vertex)",
1882
- api: "anthropic-vertex",
1883
- provider: "anthropic-vertex",
1884
- baseUrl: "https://us-central1-aiplatform.googleapis.com",
1885
- reasoning: true,
1886
- input: ["text", "image"],
1887
- cost: {
1888
- input: 3,
1889
- output: 15,
1890
- cacheRead: 0.3,
1891
- cacheWrite: 3.75,
1892
- },
1893
- contextWindow: 1000000,
1894
- maxTokens: 64000,
1895
- },
1896
- "claude-sonnet-4-5@20250929": {
1897
- id: "claude-sonnet-4-5@20250929",
1898
- name: "Claude Sonnet 4.5 (Vertex)",
1899
- api: "anthropic-vertex",
1900
- provider: "anthropic-vertex",
1901
- baseUrl: "https://us-central1-aiplatform.googleapis.com",
1902
- reasoning: true,
1903
- input: ["text", "image"],
1904
- cost: {
1905
- input: 3,
1906
- output: 15,
1907
- cacheRead: 0.3,
1908
- cacheWrite: 3.75,
1909
- },
1910
- contextWindow: 200000,
1911
- maxTokens: 64000,
1912
- },
1913
- "claude-sonnet-4@20250514": {
1914
- id: "claude-sonnet-4@20250514",
1915
- name: "Claude Sonnet 4 (Vertex)",
1916
- api: "anthropic-vertex",
1917
- provider: "anthropic-vertex",
1918
- baseUrl: "https://us-central1-aiplatform.googleapis.com",
1919
- reasoning: true,
1920
- input: ["text", "image"],
1921
- cost: {
1922
- input: 3,
1923
- output: 15,
1924
- cacheRead: 0.3,
1925
- cacheWrite: 3.75,
1926
- },
1927
- contextWindow: 200000,
1928
- maxTokens: 64000,
1929
- },
1930
- "claude-opus-4-5@20251101": {
1931
- id: "claude-opus-4-5@20251101",
1932
- name: "Claude Opus 4.5 (Vertex)",
1933
- api: "anthropic-vertex",
1934
- provider: "anthropic-vertex",
1935
- baseUrl: "https://us-central1-aiplatform.googleapis.com",
1936
- reasoning: true,
1937
- input: ["text", "image"],
1938
- cost: {
1939
- input: 15,
1940
- output: 75,
1941
- cacheRead: 1.5,
1942
- cacheWrite: 18.75,
1943
- },
1944
- contextWindow: 200000,
1945
- maxTokens: 32000,
1946
- },
1947
- "claude-opus-4-1@20250805": {
1948
- id: "claude-opus-4-1@20250805",
1949
- name: "Claude Opus 4.1 (Vertex)",
1950
- api: "anthropic-vertex",
1951
- provider: "anthropic-vertex",
1952
- baseUrl: "https://us-central1-aiplatform.googleapis.com",
1953
- reasoning: true,
1954
- input: ["text", "image"],
1955
- cost: {
1956
- input: 15,
1957
- output: 75,
1958
- cacheRead: 1.5,
1959
- cacheWrite: 18.75,
1960
- },
1961
- contextWindow: 200000,
1962
- maxTokens: 32000,
1963
- },
1964
- "claude-opus-4@20250514": {
1965
- id: "claude-opus-4@20250514",
1966
- name: "Claude Opus 4 (Vertex)",
1967
- api: "anthropic-vertex",
1968
- provider: "anthropic-vertex",
1969
- baseUrl: "https://us-central1-aiplatform.googleapis.com",
1970
- reasoning: true,
1971
- input: ["text", "image"],
1972
- cost: {
1973
- input: 15,
1974
- output: 75,
1975
- cacheRead: 1.5,
1976
- cacheWrite: 18.75,
1977
- },
1978
- contextWindow: 200000,
1979
- maxTokens: 32000,
1980
- },
1981
- "claude-haiku-4-5@20251001": {
1982
- id: "claude-haiku-4-5@20251001",
1983
- name: "Claude Haiku 4.5 (Vertex)",
1984
- api: "anthropic-vertex",
1985
- provider: "anthropic-vertex",
1986
- baseUrl: "https://us-central1-aiplatform.googleapis.com",
1987
- reasoning: false,
1988
- input: ["text", "image"],
1989
- cost: {
1990
- input: 0.8,
1991
- output: 4,
1992
- cacheRead: 0.08,
1993
- cacheWrite: 1,
1994
- },
1995
- contextWindow: 200000,
1996
- maxTokens: 8192,
1997
- },
1998
- },
1999
- "azure-openai-responses": {
2000
- "codex-mini-latest": {
2001
- id: "codex-mini-latest",
2002
- name: "Codex Mini",
2003
- api: "azure-openai-responses",
2004
- provider: "azure-openai-responses",
2005
- baseUrl: "",
2006
- reasoning: true,
2007
- input: ["text"],
2008
- cost: {
2009
- input: 1.5,
2010
- output: 6,
2011
- cacheRead: 0.375,
2012
- cacheWrite: 0,
2013
- },
2014
- contextWindow: 200000,
2015
- maxTokens: 100000,
2016
- },
2017
- "gpt-4": {
2018
- id: "gpt-4",
2019
- name: "GPT-4",
2020
- api: "azure-openai-responses",
2021
- provider: "azure-openai-responses",
2022
- baseUrl: "",
2023
- reasoning: false,
2024
- input: ["text"],
2025
- cost: {
2026
- input: 30,
2027
- output: 60,
2028
- cacheRead: 0,
2029
- cacheWrite: 0,
2030
- },
2031
- contextWindow: 8192,
2032
- maxTokens: 8192,
1842
+ contextWindow: 8192,
1843
+ maxTokens: 8192,
2033
1844
  },
2034
1845
  "gpt-4-turbo": {
2035
1846
  id: "gpt-4-turbo",
@@ -2490,6 +2301,40 @@ export const MODELS = {
2490
2301
  contextWindow: 272000,
2491
2302
  maxTokens: 128000,
2492
2303
  },
2304
+ "gpt-5.4-mini": {
2305
+ id: "gpt-5.4-mini",
2306
+ name: "GPT-5.4 mini",
2307
+ api: "azure-openai-responses",
2308
+ provider: "azure-openai-responses",
2309
+ baseUrl: "",
2310
+ reasoning: true,
2311
+ input: ["text", "image"],
2312
+ cost: {
2313
+ input: 0.75,
2314
+ output: 4.5,
2315
+ cacheRead: 0.075,
2316
+ cacheWrite: 0,
2317
+ },
2318
+ contextWindow: 400000,
2319
+ maxTokens: 128000,
2320
+ },
2321
+ "gpt-5.4-nano": {
2322
+ id: "gpt-5.4-nano",
2323
+ name: "GPT-5.4 nano",
2324
+ api: "azure-openai-responses",
2325
+ provider: "azure-openai-responses",
2326
+ baseUrl: "",
2327
+ reasoning: true,
2328
+ input: ["text", "image"],
2329
+ cost: {
2330
+ input: 0.2,
2331
+ output: 1.25,
2332
+ cacheRead: 0.02,
2333
+ cacheWrite: 0,
2334
+ },
2335
+ contextWindow: 400000,
2336
+ maxTokens: 128000,
2337
+ },
2493
2338
  "gpt-5.4-pro": {
2494
2339
  id: "gpt-5.4-pro",
2495
2340
  name: "GPT-5.4 Pro",
@@ -2730,7 +2575,7 @@ export const MODELS = {
2730
2575
  cacheRead: 0,
2731
2576
  cacheWrite: 0,
2732
2577
  },
2733
- contextWindow: 128000,
2578
+ contextWindow: 144000,
2734
2579
  maxTokens: 32000,
2735
2580
  },
2736
2581
  "claude-opus-4.5": {
@@ -2748,7 +2593,7 @@ export const MODELS = {
2748
2593
  cacheRead: 0,
2749
2594
  cacheWrite: 0,
2750
2595
  },
2751
- contextWindow: 128000,
2596
+ contextWindow: 160000,
2752
2597
  maxTokens: 32000,
2753
2598
  },
2754
2599
  "claude-opus-4.6": {
@@ -2766,7 +2611,7 @@ export const MODELS = {
2766
2611
  cacheRead: 0,
2767
2612
  cacheWrite: 0,
2768
2613
  },
2769
- contextWindow: 128000,
2614
+ contextWindow: 144000,
2770
2615
  maxTokens: 64000,
2771
2616
  },
2772
2617
  "claude-sonnet-4": {
@@ -2784,7 +2629,7 @@ export const MODELS = {
2784
2629
  cacheRead: 0,
2785
2630
  cacheWrite: 0,
2786
2631
  },
2787
- contextWindow: 128000,
2632
+ contextWindow: 216000,
2788
2633
  maxTokens: 16000,
2789
2634
  },
2790
2635
  "claude-sonnet-4.5": {
@@ -2802,7 +2647,7 @@ export const MODELS = {
2802
2647
  cacheRead: 0,
2803
2648
  cacheWrite: 0,
2804
2649
  },
2805
- contextWindow: 128000,
2650
+ contextWindow: 144000,
2806
2651
  maxTokens: 32000,
2807
2652
  },
2808
2653
  "claude-sonnet-4.6": {
@@ -2820,7 +2665,7 @@ export const MODELS = {
2820
2665
  cacheRead: 0,
2821
2666
  cacheWrite: 0,
2822
2667
  },
2823
- contextWindow: 128000,
2668
+ contextWindow: 200000,
2824
2669
  maxTokens: 32000,
2825
2670
  },
2826
2671
  "gemini-2.5-pro": {
@@ -2915,7 +2760,7 @@ export const MODELS = {
2915
2760
  cacheRead: 0,
2916
2761
  cacheWrite: 0,
2917
2762
  },
2918
- contextWindow: 64000,
2763
+ contextWindow: 128000,
2919
2764
  maxTokens: 16384,
2920
2765
  },
2921
2766
  "gpt-4o": {
@@ -2934,8 +2779,8 @@ export const MODELS = {
2934
2779
  cacheRead: 0,
2935
2780
  cacheWrite: 0,
2936
2781
  },
2937
- contextWindow: 64000,
2938
- maxTokens: 16384,
2782
+ contextWindow: 128000,
2783
+ maxTokens: 4096,
2939
2784
  },
2940
2785
  "gpt-5": {
2941
2786
  id: "gpt-5",
@@ -2970,7 +2815,7 @@ export const MODELS = {
2970
2815
  cacheRead: 0,
2971
2816
  cacheWrite: 0,
2972
2817
  },
2973
- contextWindow: 128000,
2818
+ contextWindow: 264000,
2974
2819
  maxTokens: 64000,
2975
2820
  },
2976
2821
  "gpt-5.1": {
@@ -2988,7 +2833,7 @@ export const MODELS = {
2988
2833
  cacheRead: 0,
2989
2834
  cacheWrite: 0,
2990
2835
  },
2991
- contextWindow: 128000,
2836
+ contextWindow: 264000,
2992
2837
  maxTokens: 64000,
2993
2838
  },
2994
2839
  "gpt-5.1-codex": {
@@ -3006,7 +2851,7 @@ export const MODELS = {
3006
2851
  cacheRead: 0,
3007
2852
  cacheWrite: 0,
3008
2853
  },
3009
- contextWindow: 128000,
2854
+ contextWindow: 400000,
3010
2855
  maxTokens: 128000,
3011
2856
  },
3012
2857
  "gpt-5.1-codex-max": {
@@ -3024,7 +2869,7 @@ export const MODELS = {
3024
2869
  cacheRead: 0,
3025
2870
  cacheWrite: 0,
3026
2871
  },
3027
- contextWindow: 128000,
2872
+ contextWindow: 400000,
3028
2873
  maxTokens: 128000,
3029
2874
  },
3030
2875
  "gpt-5.1-codex-mini": {
@@ -3042,7 +2887,7 @@ export const MODELS = {
3042
2887
  cacheRead: 0,
3043
2888
  cacheWrite: 0,
3044
2889
  },
3045
- contextWindow: 128000,
2890
+ contextWindow: 400000,
3046
2891
  maxTokens: 128000,
3047
2892
  },
3048
2893
  "gpt-5.2": {
@@ -3060,7 +2905,7 @@ export const MODELS = {
3060
2905
  cacheRead: 0,
3061
2906
  cacheWrite: 0,
3062
2907
  },
3063
- contextWindow: 128000,
2908
+ contextWindow: 264000,
3064
2909
  maxTokens: 64000,
3065
2910
  },
3066
2911
  "gpt-5.2-codex": {
@@ -3078,7 +2923,7 @@ export const MODELS = {
3078
2923
  cacheRead: 0,
3079
2924
  cacheWrite: 0,
3080
2925
  },
3081
- contextWindow: 272000,
2926
+ contextWindow: 400000,
3082
2927
  maxTokens: 128000,
3083
2928
  },
3084
2929
  "gpt-5.3-codex": {
@@ -3117,6 +2962,24 @@ export const MODELS = {
3117
2962
  contextWindow: 400000,
3118
2963
  maxTokens: 128000,
3119
2964
  },
2965
+ "gpt-5.4-mini": {
2966
+ id: "gpt-5.4-mini",
2967
+ name: "GPT-5.4 mini",
2968
+ api: "openai-responses",
2969
+ provider: "github-copilot",
2970
+ baseUrl: "https://api.individual.githubcopilot.com",
2971
+ headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
2972
+ reasoning: true,
2973
+ input: ["text", "image"],
2974
+ cost: {
2975
+ input: 0,
2976
+ output: 0,
2977
+ cacheRead: 0,
2978
+ cacheWrite: 0,
2979
+ },
2980
+ contextWindow: 400000,
2981
+ maxTokens: 128000,
2982
+ },
3120
2983
  "grok-code-fast-1": {
3121
2984
  id: "grok-code-fast-1",
3122
2985
  name: "Grok Code Fast 1",
@@ -3436,10 +3299,10 @@ export const MODELS = {
3436
3299
  reasoning: true,
3437
3300
  input: ["text", "image"],
3438
3301
  cost: {
3439
- input: 0,
3440
- output: 0,
3441
- cacheRead: 0,
3442
- cacheWrite: 0,
3302
+ input: 0.25,
3303
+ output: 1.5,
3304
+ cacheRead: 0.025,
3305
+ cacheWrite: 1,
3443
3306
  },
3444
3307
  contextWindow: 1048576,
3445
3308
  maxTokens: 65536,
@@ -4700,6 +4563,40 @@ export const MODELS = {
4700
4563
  contextWindow: 204800,
4701
4564
  maxTokens: 131072,
4702
4565
  },
4566
+ "MiniMax-M2.7": {
4567
+ id: "MiniMax-M2.7",
4568
+ name: "MiniMax-M2.7",
4569
+ api: "anthropic-messages",
4570
+ provider: "minimax",
4571
+ baseUrl: "https://api.minimax.io/anthropic",
4572
+ reasoning: true,
4573
+ input: ["text"],
4574
+ cost: {
4575
+ input: 0.3,
4576
+ output: 1.2,
4577
+ cacheRead: 0.06,
4578
+ cacheWrite: 0.375,
4579
+ },
4580
+ contextWindow: 204800,
4581
+ maxTokens: 131072,
4582
+ },
4583
+ "MiniMax-M2.7-highspeed": {
4584
+ id: "MiniMax-M2.7-highspeed",
4585
+ name: "MiniMax-M2.7-highspeed",
4586
+ api: "anthropic-messages",
4587
+ provider: "minimax",
4588
+ baseUrl: "https://api.minimax.io/anthropic",
4589
+ reasoning: true,
4590
+ input: ["text"],
4591
+ cost: {
4592
+ input: 0.6,
4593
+ output: 2.4,
4594
+ cacheRead: 0.06,
4595
+ cacheWrite: 0.375,
4596
+ },
4597
+ contextWindow: 204800,
4598
+ maxTokens: 131072,
4599
+ },
4703
4600
  },
4704
4601
  "minimax-cn": {
4705
4602
  "MiniMax-M2": {
@@ -4770,11 +4667,45 @@ export const MODELS = {
4770
4667
  contextWindow: 204800,
4771
4668
  maxTokens: 131072,
4772
4669
  },
4670
+ "MiniMax-M2.7": {
4671
+ id: "MiniMax-M2.7",
4672
+ name: "MiniMax-M2.7",
4673
+ api: "anthropic-messages",
4674
+ provider: "minimax-cn",
4675
+ baseUrl: "https://api.minimaxi.com/anthropic",
4676
+ reasoning: true,
4677
+ input: ["text"],
4678
+ cost: {
4679
+ input: 0.3,
4680
+ output: 1.2,
4681
+ cacheRead: 0.06,
4682
+ cacheWrite: 0.375,
4683
+ },
4684
+ contextWindow: 204800,
4685
+ maxTokens: 131072,
4686
+ },
4687
+ "MiniMax-M2.7-highspeed": {
4688
+ id: "MiniMax-M2.7-highspeed",
4689
+ name: "MiniMax-M2.7-highspeed",
4690
+ api: "anthropic-messages",
4691
+ provider: "minimax-cn",
4692
+ baseUrl: "https://api.minimaxi.com/anthropic",
4693
+ reasoning: true,
4694
+ input: ["text"],
4695
+ cost: {
4696
+ input: 0.6,
4697
+ output: 2.4,
4698
+ cacheRead: 0.06,
4699
+ cacheWrite: 0.375,
4700
+ },
4701
+ contextWindow: 204800,
4702
+ maxTokens: 131072,
4703
+ },
4773
4704
  },
4774
4705
  "mistral": {
4775
4706
  "codestral-latest": {
4776
4707
  id: "codestral-latest",
4777
- name: "Codestral",
4708
+ name: "Codestral (latest)",
4778
4709
  api: "mistral-conversations",
4779
4710
  provider: "mistral",
4780
4711
  baseUrl: "https://api.mistral.ai",
@@ -4825,7 +4756,7 @@ export const MODELS = {
4825
4756
  },
4826
4757
  "devstral-medium-latest": {
4827
4758
  id: "devstral-medium-latest",
4828
- name: "Devstral 2",
4759
+ name: "Devstral 2 (latest)",
4829
4760
  api: "mistral-conversations",
4830
4761
  provider: "mistral",
4831
4762
  baseUrl: "https://api.mistral.ai",
@@ -4893,7 +4824,7 @@ export const MODELS = {
4893
4824
  },
4894
4825
  "magistral-medium-latest": {
4895
4826
  id: "magistral-medium-latest",
4896
- name: "Magistral Medium",
4827
+ name: "Magistral Medium (latest)",
4897
4828
  api: "mistral-conversations",
4898
4829
  provider: "mistral",
4899
4830
  baseUrl: "https://api.mistral.ai",
@@ -4927,7 +4858,7 @@ export const MODELS = {
4927
4858
  },
4928
4859
  "ministral-3b-latest": {
4929
4860
  id: "ministral-3b-latest",
4930
- name: "Ministral 3B",
4861
+ name: "Ministral 3B (latest)",
4931
4862
  api: "mistral-conversations",
4932
4863
  provider: "mistral",
4933
4864
  baseUrl: "https://api.mistral.ai",
@@ -4944,7 +4875,7 @@ export const MODELS = {
4944
4875
  },
4945
4876
  "ministral-8b-latest": {
4946
4877
  id: "ministral-8b-latest",
4947
- name: "Ministral 8B",
4878
+ name: "Ministral 8B (latest)",
4948
4879
  api: "mistral-conversations",
4949
4880
  provider: "mistral",
4950
4881
  baseUrl: "https://api.mistral.ai",
@@ -4995,7 +4926,7 @@ export const MODELS = {
4995
4926
  },
4996
4927
  "mistral-large-latest": {
4997
4928
  id: "mistral-large-latest",
4998
- name: "Mistral Large",
4929
+ name: "Mistral Large (latest)",
4999
4930
  api: "mistral-conversations",
5000
4931
  provider: "mistral",
5001
4932
  baseUrl: "https://api.mistral.ai",
@@ -5046,7 +4977,7 @@ export const MODELS = {
5046
4977
  },
5047
4978
  "mistral-medium-latest": {
5048
4979
  id: "mistral-medium-latest",
5049
- name: "Mistral Medium",
4980
+ name: "Mistral Medium (latest)",
5050
4981
  api: "mistral-conversations",
5051
4982
  provider: "mistral",
5052
4983
  baseUrl: "https://api.mistral.ai",
@@ -5097,7 +5028,7 @@ export const MODELS = {
5097
5028
  },
5098
5029
  "mistral-small-latest": {
5099
5030
  id: "mistral-small-latest",
5100
- name: "Mistral Small",
5031
+ name: "Mistral Small (latest)",
5101
5032
  api: "mistral-conversations",
5102
5033
  provider: "mistral",
5103
5034
  baseUrl: "https://api.mistral.ai",
@@ -5182,7 +5113,7 @@ export const MODELS = {
5182
5113
  },
5183
5114
  "pixtral-large-latest": {
5184
5115
  id: "pixtral-large-latest",
5185
- name: "Pixtral Large",
5116
+ name: "Pixtral Large (latest)",
5186
5117
  api: "mistral-conversations",
5187
5118
  provider: "mistral",
5188
5119
  baseUrl: "https://api.mistral.ai",
@@ -5692,6 +5623,40 @@ export const MODELS = {
5692
5623
  contextWindow: 272000,
5693
5624
  maxTokens: 128000,
5694
5625
  },
5626
+ "gpt-5.4-mini": {
5627
+ id: "gpt-5.4-mini",
5628
+ name: "GPT-5.4 mini",
5629
+ api: "openai-responses",
5630
+ provider: "openai",
5631
+ baseUrl: "https://api.openai.com/v1",
5632
+ reasoning: true,
5633
+ input: ["text", "image"],
5634
+ cost: {
5635
+ input: 0.75,
5636
+ output: 4.5,
5637
+ cacheRead: 0.075,
5638
+ cacheWrite: 0,
5639
+ },
5640
+ contextWindow: 400000,
5641
+ maxTokens: 128000,
5642
+ },
5643
+ "gpt-5.4-nano": {
5644
+ id: "gpt-5.4-nano",
5645
+ name: "GPT-5.4 nano",
5646
+ api: "openai-responses",
5647
+ provider: "openai",
5648
+ baseUrl: "https://api.openai.com/v1",
5649
+ reasoning: true,
5650
+ input: ["text", "image"],
5651
+ cost: {
5652
+ input: 0.2,
5653
+ output: 1.25,
5654
+ cacheRead: 0.02,
5655
+ cacheWrite: 0,
5656
+ },
5657
+ contextWindow: 400000,
5658
+ maxTokens: 128000,
5659
+ },
5695
5660
  "gpt-5.4-pro": {
5696
5661
  id: "gpt-5.4-pro",
5697
5662
  name: "GPT-5.4 Pro",
@@ -6084,7 +6049,7 @@ export const MODELS = {
6084
6049
  cacheRead: 0.5,
6085
6050
  cacheWrite: 6.25,
6086
6051
  },
6087
- contextWindow: 200000,
6052
+ contextWindow: 1000000,
6088
6053
  maxTokens: 128000,
6089
6054
  },
6090
6055
  "claude-sonnet-4": {
@@ -6155,23 +6120,6 @@ export const MODELS = {
6155
6120
  contextWindow: 1048576,
6156
6121
  maxTokens: 65536,
6157
6122
  },
6158
- "gemini-3-pro": {
6159
- id: "gemini-3-pro",
6160
- name: "Gemini 3 Pro",
6161
- api: "google-generative-ai",
6162
- provider: "opencode",
6163
- baseUrl: "https://opencode.ai/zen/v1",
6164
- reasoning: true,
6165
- input: ["text", "image"],
6166
- cost: {
6167
- input: 2,
6168
- output: 12,
6169
- cacheRead: 0.2,
6170
- cacheWrite: 0,
6171
- },
6172
- contextWindow: 1048576,
6173
- maxTokens: 65536,
6174
- },
6175
6123
  "gemini-3.1-pro": {
6176
6124
  id: "gemini-3.1-pro",
6177
6125
  name: "Gemini 3.1 Pro Preview",
@@ -6189,40 +6137,6 @@ export const MODELS = {
6189
6137
  contextWindow: 1048576,
6190
6138
  maxTokens: 65536,
6191
6139
  },
6192
- "glm-4.6": {
6193
- id: "glm-4.6",
6194
- name: "GLM-4.6",
6195
- api: "openai-completions",
6196
- provider: "opencode",
6197
- baseUrl: "https://opencode.ai/zen/v1",
6198
- reasoning: true,
6199
- input: ["text"],
6200
- cost: {
6201
- input: 0.6,
6202
- output: 2.2,
6203
- cacheRead: 0.1,
6204
- cacheWrite: 0,
6205
- },
6206
- contextWindow: 204800,
6207
- maxTokens: 131072,
6208
- },
6209
- "glm-4.7": {
6210
- id: "glm-4.7",
6211
- name: "GLM-4.7",
6212
- api: "openai-completions",
6213
- provider: "opencode",
6214
- baseUrl: "https://opencode.ai/zen/v1",
6215
- reasoning: true,
6216
- input: ["text"],
6217
- cost: {
6218
- input: 0.6,
6219
- output: 2.2,
6220
- cacheRead: 0.1,
6221
- cacheWrite: 0,
6222
- },
6223
- contextWindow: 204800,
6224
- maxTokens: 131072,
6225
- },
6226
6140
  "glm-5": {
6227
6141
  id: "glm-5",
6228
6142
  name: "GLM-5",
@@ -6427,6 +6341,40 @@ export const MODELS = {
6427
6341
  contextWindow: 272000,
6428
6342
  maxTokens: 128000,
6429
6343
  },
6344
+ "gpt-5.4-mini": {
6345
+ id: "gpt-5.4-mini",
6346
+ name: "GPT-5.4 Mini",
6347
+ api: "openai-responses",
6348
+ provider: "opencode",
6349
+ baseUrl: "https://opencode.ai/zen/v1",
6350
+ reasoning: true,
6351
+ input: ["text", "image"],
6352
+ cost: {
6353
+ input: 0.75,
6354
+ output: 4.5,
6355
+ cacheRead: 0.075,
6356
+ cacheWrite: 0,
6357
+ },
6358
+ contextWindow: 400000,
6359
+ maxTokens: 128000,
6360
+ },
6361
+ "gpt-5.4-nano": {
6362
+ id: "gpt-5.4-nano",
6363
+ name: "GPT-5.4 Nano",
6364
+ api: "openai-responses",
6365
+ provider: "opencode",
6366
+ baseUrl: "https://opencode.ai/zen/v1",
6367
+ reasoning: true,
6368
+ input: ["text", "image"],
6369
+ cost: {
6370
+ input: 0.2,
6371
+ output: 1.25,
6372
+ cacheRead: 0.02,
6373
+ cacheWrite: 0,
6374
+ },
6375
+ contextWindow: 400000,
6376
+ maxTokens: 128000,
6377
+ },
6430
6378
  "gpt-5.4-pro": {
6431
6379
  id: "gpt-5.4-pro",
6432
6380
  name: "GPT-5.4 Pro",
@@ -6461,22 +6409,39 @@ export const MODELS = {
6461
6409
  contextWindow: 262144,
6462
6410
  maxTokens: 65536,
6463
6411
  },
6464
- "minimax-m2.1": {
6465
- id: "minimax-m2.1",
6466
- name: "MiniMax M2.1",
6412
+ "mimo-v2-omni-free": {
6413
+ id: "mimo-v2-omni-free",
6414
+ name: "MiMo V2 Omni Free",
6415
+ api: "openai-completions",
6416
+ provider: "opencode",
6417
+ baseUrl: "https://opencode.ai/zen/v1",
6418
+ reasoning: true,
6419
+ input: ["text", "image"],
6420
+ cost: {
6421
+ input: 0,
6422
+ output: 0,
6423
+ cacheRead: 0,
6424
+ cacheWrite: 0,
6425
+ },
6426
+ contextWindow: 262144,
6427
+ maxTokens: 64000,
6428
+ },
6429
+ "mimo-v2-pro-free": {
6430
+ id: "mimo-v2-pro-free",
6431
+ name: "MiMo V2 Pro Free",
6467
6432
  api: "openai-completions",
6468
6433
  provider: "opencode",
6469
6434
  baseUrl: "https://opencode.ai/zen/v1",
6470
6435
  reasoning: true,
6471
6436
  input: ["text"],
6472
6437
  cost: {
6473
- input: 0.3,
6474
- output: 1.2,
6475
- cacheRead: 0.1,
6438
+ input: 0,
6439
+ output: 0,
6440
+ cacheRead: 0,
6476
6441
  cacheWrite: 0,
6477
6442
  },
6478
- contextWindow: 204800,
6479
- maxTokens: 131072,
6443
+ contextWindow: 1048576,
6444
+ maxTokens: 64000,
6480
6445
  },
6481
6446
  "minimax-m2.5": {
6482
6447
  id: "minimax-m2.5",
@@ -6512,6 +6477,23 @@ export const MODELS = {
6512
6477
  contextWindow: 204800,
6513
6478
  maxTokens: 131072,
6514
6479
  },
6480
+ "nemotron-3-super-free": {
6481
+ id: "nemotron-3-super-free",
6482
+ name: "Nemotron 3 Super Free",
6483
+ api: "openai-completions",
6484
+ provider: "opencode",
6485
+ baseUrl: "https://opencode.ai/zen/v1",
6486
+ reasoning: true,
6487
+ input: ["text"],
6488
+ cost: {
6489
+ input: 0,
6490
+ output: 0,
6491
+ cacheRead: 0,
6492
+ cacheWrite: 0,
6493
+ },
6494
+ contextWindow: 1000000,
6495
+ maxTokens: 128000,
6496
+ },
6515
6497
  },
6516
6498
  "opencode-go": {
6517
6499
  "glm-5": {
@@ -6565,6 +6547,23 @@ export const MODELS = {
6565
6547
  contextWindow: 204800,
6566
6548
  maxTokens: 131072,
6567
6549
  },
6550
+ "minimax-m2.7": {
6551
+ id: "minimax-m2.7",
6552
+ name: "MiniMax M2.7",
6553
+ api: "anthropic-messages",
6554
+ provider: "opencode-go",
6555
+ baseUrl: "https://opencode.ai/zen/go",
6556
+ reasoning: true,
6557
+ input: ["text"],
6558
+ cost: {
6559
+ input: 0.3,
6560
+ output: 1.2,
6561
+ cacheRead: 0.06,
6562
+ cacheWrite: 0,
6563
+ },
6564
+ contextWindow: 204800,
6565
+ maxTokens: 131072,
6566
+ },
6568
6567
  },
6569
6568
  "openrouter": {
6570
6569
  "ai21/jamba-large-1.7": {
@@ -7077,6 +7076,23 @@ export const MODELS = {
7077
7076
  contextWindow: 262144,
7078
7077
  maxTokens: 32768,
7079
7078
  },
7079
+ "bytedance-seed/seed-2.0-lite": {
7080
+ id: "bytedance-seed/seed-2.0-lite",
7081
+ name: "ByteDance Seed: Seed-2.0-Lite",
7082
+ api: "openai-completions",
7083
+ provider: "openrouter",
7084
+ baseUrl: "https://openrouter.ai/api/v1",
7085
+ reasoning: true,
7086
+ input: ["text", "image"],
7087
+ cost: {
7088
+ input: 0.25,
7089
+ output: 2,
7090
+ cacheRead: 0,
7091
+ cacheWrite: 0,
7092
+ },
7093
+ contextWindow: 262144,
7094
+ maxTokens: 131072,
7095
+ },
7080
7096
  "bytedance-seed/seed-2.0-mini": {
7081
7097
  id: "bytedance-seed/seed-2.0-mini",
7082
7098
  name: "ByteDance Seed: Seed-2.0-Mini",
@@ -7156,11 +7172,11 @@ export const MODELS = {
7156
7172
  cost: {
7157
7173
  input: 0.19999999999999998,
7158
7174
  output: 0.77,
7159
- cacheRead: 0.13,
7175
+ cacheRead: 0.135,
7160
7176
  cacheWrite: 0,
7161
7177
  },
7162
7178
  contextWindow: 163840,
7163
- maxTokens: 163840,
7179
+ maxTokens: 4096,
7164
7180
  },
7165
7181
  "deepseek/deepseek-chat-v3.1": {
7166
7182
  id: "deepseek/deepseek-chat-v3.1",
@@ -7230,23 +7246,6 @@ export const MODELS = {
7230
7246
  contextWindow: 163840,
7231
7247
  maxTokens: 4096,
7232
7248
  },
7233
- "deepseek/deepseek-v3.1-terminus:exacto": {
7234
- id: "deepseek/deepseek-v3.1-terminus:exacto",
7235
- name: "DeepSeek: DeepSeek V3.1 Terminus (exacto)",
7236
- api: "openai-completions",
7237
- provider: "openrouter",
7238
- baseUrl: "https://openrouter.ai/api/v1",
7239
- reasoning: true,
7240
- input: ["text"],
7241
- cost: {
7242
- input: 0.21,
7243
- output: 0.7899999999999999,
7244
- cacheRead: 0.16799999999999998,
7245
- cacheWrite: 0,
7246
- },
7247
- contextWindow: 163840,
7248
- maxTokens: 4096,
7249
- },
7250
7249
  "deepseek/deepseek-v3.2": {
7251
7250
  id: "deepseek/deepseek-v3.2",
7252
7251
  name: "DeepSeek: DeepSeek V3.2",
@@ -7256,13 +7255,13 @@ export const MODELS = {
7256
7255
  reasoning: true,
7257
7256
  input: ["text"],
7258
7257
  cost: {
7259
- input: 0.25,
7260
- output: 0.39999999999999997,
7261
- cacheRead: 0,
7258
+ input: 0.26,
7259
+ output: 0.38,
7260
+ cacheRead: 0.13,
7262
7261
  cacheWrite: 0,
7263
7262
  },
7264
7263
  contextWindow: 163840,
7265
- maxTokens: 65536,
7264
+ maxTokens: 4096,
7266
7265
  },
7267
7266
  "deepseek/deepseek-v3.2-exp": {
7268
7267
  id: "deepseek/deepseek-v3.2-exp",
@@ -7519,40 +7518,6 @@ export const MODELS = {
7519
7518
  contextWindow: 1048576,
7520
7519
  maxTokens: 65536,
7521
7520
  },
7522
- "google/gemma-3-27b-it": {
7523
- id: "google/gemma-3-27b-it",
7524
- name: "Google: Gemma 3 27B",
7525
- api: "openai-completions",
7526
- provider: "openrouter",
7527
- baseUrl: "https://openrouter.ai/api/v1",
7528
- reasoning: false,
7529
- input: ["text", "image"],
7530
- cost: {
7531
- input: 0.04,
7532
- output: 0.15,
7533
- cacheRead: 0.02,
7534
- cacheWrite: 0,
7535
- },
7536
- contextWindow: 128000,
7537
- maxTokens: 65536,
7538
- },
7539
- "google/gemma-3-27b-it:free": {
7540
- id: "google/gemma-3-27b-it:free",
7541
- name: "Google: Gemma 3 27B (free)",
7542
- api: "openai-completions",
7543
- provider: "openrouter",
7544
- baseUrl: "https://openrouter.ai/api/v1",
7545
- reasoning: false,
7546
- input: ["text", "image"],
7547
- cost: {
7548
- input: 0,
7549
- output: 0,
7550
- cacheRead: 0,
7551
- cacheWrite: 0,
7552
- },
7553
- contextWindow: 131072,
7554
- maxTokens: 8192,
7555
- },
7556
7521
  "inception/mercury": {
7557
7522
  id: "inception/mercury",
7558
7523
  name: "Inception: Mercury",
@@ -7655,23 +7620,6 @@ export const MODELS = {
7655
7620
  contextWindow: 8192,
7656
7621
  maxTokens: 16384,
7657
7622
  },
7658
- "meta-llama/llama-3.1-405b-instruct": {
7659
- id: "meta-llama/llama-3.1-405b-instruct",
7660
- name: "Meta: Llama 3.1 405B Instruct",
7661
- api: "openai-completions",
7662
- provider: "openrouter",
7663
- baseUrl: "https://openrouter.ai/api/v1",
7664
- reasoning: false,
7665
- input: ["text"],
7666
- cost: {
7667
- input: 4,
7668
- output: 4,
7669
- cacheRead: 0,
7670
- cacheWrite: 0,
7671
- },
7672
- contextWindow: 131000,
7673
- maxTokens: 4096,
7674
- },
7675
7623
  "meta-llama/llama-3.1-70b-instruct": {
7676
7624
  id: "meta-llama/llama-3.1-70b-instruct",
7677
7625
  name: "Meta: Llama 3.1 70B Instruct",
@@ -7737,8 +7685,8 @@ export const MODELS = {
7737
7685
  cacheRead: 0,
7738
7686
  cacheWrite: 0,
7739
7687
  },
7740
- contextWindow: 128000,
7741
- maxTokens: 128000,
7688
+ contextWindow: 65536,
7689
+ maxTokens: 4096,
7742
7690
  },
7743
7691
  "meta-llama/llama-4-maverick": {
7744
7692
  id: "meta-llama/llama-4-maverick",
@@ -7834,28 +7782,62 @@ export const MODELS = {
7834
7782
  reasoning: true,
7835
7783
  input: ["text"],
7836
7784
  cost: {
7837
- input: 0.295,
7838
- output: 1.2,
7839
- cacheRead: 0.03,
7785
+ input: 0.19999999999999998,
7786
+ output: 1.17,
7787
+ cacheRead: 0.09999999999999999,
7840
7788
  cacheWrite: 0,
7841
7789
  },
7842
7790
  contextWindow: 196608,
7843
- maxTokens: 196608,
7791
+ maxTokens: 65536,
7844
7792
  },
7845
- "mistralai/codestral-2508": {
7846
- id: "mistralai/codestral-2508",
7847
- name: "Mistral: Codestral 2508",
7793
+ "minimax/minimax-m2.5:free": {
7794
+ id: "minimax/minimax-m2.5:free",
7795
+ name: "MiniMax: MiniMax M2.5 (free)",
7848
7796
  api: "openai-completions",
7849
7797
  provider: "openrouter",
7850
7798
  baseUrl: "https://openrouter.ai/api/v1",
7851
- reasoning: false,
7799
+ reasoning: true,
7852
7800
  input: ["text"],
7853
7801
  cost: {
7854
- input: 0.3,
7855
- output: 0.8999999999999999,
7802
+ input: 0,
7803
+ output: 0,
7856
7804
  cacheRead: 0,
7857
7805
  cacheWrite: 0,
7858
7806
  },
7807
+ contextWindow: 196608,
7808
+ maxTokens: 196608,
7809
+ },
7810
+ "minimax/minimax-m2.7": {
7811
+ id: "minimax/minimax-m2.7",
7812
+ name: "MiniMax: MiniMax M2.7",
7813
+ api: "openai-completions",
7814
+ provider: "openrouter",
7815
+ baseUrl: "https://openrouter.ai/api/v1",
7816
+ reasoning: true,
7817
+ input: ["text"],
7818
+ cost: {
7819
+ input: 0.3,
7820
+ output: 1.2,
7821
+ cacheRead: 0.06,
7822
+ cacheWrite: 0,
7823
+ },
7824
+ contextWindow: 204800,
7825
+ maxTokens: 131072,
7826
+ },
7827
+ "mistralai/codestral-2508": {
7828
+ id: "mistralai/codestral-2508",
7829
+ name: "Mistral: Codestral 2508",
7830
+ api: "openai-completions",
7831
+ provider: "openrouter",
7832
+ baseUrl: "https://openrouter.ai/api/v1",
7833
+ reasoning: false,
7834
+ input: ["text"],
7835
+ cost: {
7836
+ input: 0.3,
7837
+ output: 0.8999999999999999,
7838
+ cacheRead: 0.03,
7839
+ cacheWrite: 0,
7840
+ },
7859
7841
  contextWindow: 256000,
7860
7842
  maxTokens: 4096,
7861
7843
  },
@@ -7870,7 +7852,7 @@ export const MODELS = {
7870
7852
  cost: {
7871
7853
  input: 0.39999999999999997,
7872
7854
  output: 2,
7873
- cacheRead: 0,
7855
+ cacheRead: 0.04,
7874
7856
  cacheWrite: 0,
7875
7857
  },
7876
7858
  contextWindow: 262144,
@@ -7887,7 +7869,7 @@ export const MODELS = {
7887
7869
  cost: {
7888
7870
  input: 0.39999999999999997,
7889
7871
  output: 2,
7890
- cacheRead: 0,
7872
+ cacheRead: 0.04,
7891
7873
  cacheWrite: 0,
7892
7874
  },
7893
7875
  contextWindow: 131072,
@@ -7904,7 +7886,7 @@ export const MODELS = {
7904
7886
  cost: {
7905
7887
  input: 0.09999999999999999,
7906
7888
  output: 0.3,
7907
- cacheRead: 0,
7889
+ cacheRead: 0.01,
7908
7890
  cacheWrite: 0,
7909
7891
  },
7910
7892
  contextWindow: 131072,
@@ -7921,7 +7903,7 @@ export const MODELS = {
7921
7903
  cost: {
7922
7904
  input: 0.19999999999999998,
7923
7905
  output: 0.19999999999999998,
7924
- cacheRead: 0,
7906
+ cacheRead: 0.02,
7925
7907
  cacheWrite: 0,
7926
7908
  },
7927
7909
  contextWindow: 262144,
@@ -7938,7 +7920,7 @@ export const MODELS = {
7938
7920
  cost: {
7939
7921
  input: 0.09999999999999999,
7940
7922
  output: 0.09999999999999999,
7941
- cacheRead: 0,
7923
+ cacheRead: 0.01,
7942
7924
  cacheWrite: 0,
7943
7925
  },
7944
7926
  contextWindow: 131072,
@@ -7955,7 +7937,7 @@ export const MODELS = {
7955
7937
  cost: {
7956
7938
  input: 0.15,
7957
7939
  output: 0.15,
7958
- cacheRead: 0,
7940
+ cacheRead: 0.015,
7959
7941
  cacheWrite: 0,
7960
7942
  },
7961
7943
  contextWindow: 262144,
@@ -7972,7 +7954,7 @@ export const MODELS = {
7972
7954
  cost: {
7973
7955
  input: 2,
7974
7956
  output: 6,
7975
- cacheRead: 0,
7957
+ cacheRead: 0.19999999999999998,
7976
7958
  cacheWrite: 0,
7977
7959
  },
7978
7960
  contextWindow: 128000,
@@ -7989,7 +7971,7 @@ export const MODELS = {
7989
7971
  cost: {
7990
7972
  input: 2,
7991
7973
  output: 6,
7992
- cacheRead: 0,
7974
+ cacheRead: 0.19999999999999998,
7993
7975
  cacheWrite: 0,
7994
7976
  },
7995
7977
  contextWindow: 131072,
@@ -8006,7 +7988,7 @@ export const MODELS = {
8006
7988
  cost: {
8007
7989
  input: 2,
8008
7990
  output: 6,
8009
- cacheRead: 0,
7991
+ cacheRead: 0.19999999999999998,
8010
7992
  cacheWrite: 0,
8011
7993
  },
8012
7994
  contextWindow: 131072,
@@ -8023,7 +8005,7 @@ export const MODELS = {
8023
8005
  cost: {
8024
8006
  input: 0.5,
8025
8007
  output: 1.5,
8026
- cacheRead: 0,
8008
+ cacheRead: 0.049999999999999996,
8027
8009
  cacheWrite: 0,
8028
8010
  },
8029
8011
  contextWindow: 262144,
@@ -8040,7 +8022,7 @@ export const MODELS = {
8040
8022
  cost: {
8041
8023
  input: 0.39999999999999997,
8042
8024
  output: 2,
8043
- cacheRead: 0,
8025
+ cacheRead: 0.04,
8044
8026
  cacheWrite: 0,
8045
8027
  },
8046
8028
  contextWindow: 131072,
@@ -8057,7 +8039,7 @@ export const MODELS = {
8057
8039
  cost: {
8058
8040
  input: 0.39999999999999997,
8059
8041
  output: 2,
8060
- cacheRead: 0,
8042
+ cacheRead: 0.04,
8061
8043
  cacheWrite: 0,
8062
8044
  },
8063
8045
  contextWindow: 131072,
@@ -8091,7 +8073,7 @@ export const MODELS = {
8091
8073
  cost: {
8092
8074
  input: 0.19999999999999998,
8093
8075
  output: 0.6,
8094
- cacheRead: 0,
8076
+ cacheRead: 0.02,
8095
8077
  cacheWrite: 0,
8096
8078
  },
8097
8079
  contextWindow: 32768,
@@ -8114,6 +8096,23 @@ export const MODELS = {
8114
8096
  contextWindow: 32768,
8115
8097
  maxTokens: 16384,
8116
8098
  },
8099
+ "mistralai/mistral-small-2603": {
8100
+ id: "mistralai/mistral-small-2603",
8101
+ name: "Mistral: Mistral Small 4",
8102
+ api: "openai-completions",
8103
+ provider: "openrouter",
8104
+ baseUrl: "https://openrouter.ai/api/v1",
8105
+ reasoning: true,
8106
+ input: ["text", "image"],
8107
+ cost: {
8108
+ input: 0.15,
8109
+ output: 0.6,
8110
+ cacheRead: 0.015,
8111
+ cacheWrite: 0,
8112
+ },
8113
+ contextWindow: 262144,
8114
+ maxTokens: 4096,
8115
+ },
8117
8116
  "mistralai/mistral-small-3.1-24b-instruct:free": {
8118
8117
  id: "mistralai/mistral-small-3.1-24b-instruct:free",
8119
8118
  name: "Mistral: Mistral Small 3.1 24B (free)",
@@ -8140,13 +8139,13 @@ export const MODELS = {
8140
8139
  reasoning: false,
8141
8140
  input: ["text", "image"],
8142
8141
  cost: {
8143
- input: 0.06,
8144
- output: 0.18,
8145
- cacheRead: 0.03,
8142
+ input: 0.075,
8143
+ output: 0.19999999999999998,
8144
+ cacheRead: 0,
8146
8145
  cacheWrite: 0,
8147
8146
  },
8148
- contextWindow: 131072,
8149
- maxTokens: 131072,
8147
+ contextWindow: 128000,
8148
+ maxTokens: 4096,
8150
8149
  },
8151
8150
  "mistralai/mistral-small-creative": {
8152
8151
  id: "mistralai/mistral-small-creative",
@@ -8159,7 +8158,7 @@ export const MODELS = {
8159
8158
  cost: {
8160
8159
  input: 0.09999999999999999,
8161
8160
  output: 0.3,
8162
- cacheRead: 0,
8161
+ cacheRead: 0.01,
8163
8162
  cacheWrite: 0,
8164
8163
  },
8165
8164
  contextWindow: 32768,
@@ -8176,7 +8175,7 @@ export const MODELS = {
8176
8175
  cost: {
8177
8176
  input: 2,
8178
8177
  output: 6,
8179
- cacheRead: 0,
8178
+ cacheRead: 0.19999999999999998,
8180
8179
  cacheWrite: 0,
8181
8180
  },
8182
8181
  contextWindow: 65536,
@@ -8210,7 +8209,7 @@ export const MODELS = {
8210
8209
  cost: {
8211
8210
  input: 2,
8212
8211
  output: 6,
8213
- cacheRead: 0,
8212
+ cacheRead: 0.19999999999999998,
8214
8213
  cacheWrite: 0,
8215
8214
  },
8216
8215
  contextWindow: 131072,
@@ -8227,7 +8226,7 @@ export const MODELS = {
8227
8226
  cost: {
8228
8227
  input: 0.09999999999999999,
8229
8228
  output: 0.3,
8230
- cacheRead: 0,
8229
+ cacheRead: 0.01,
8231
8230
  cacheWrite: 0,
8232
8231
  },
8233
8232
  contextWindow: 32000,
@@ -8267,23 +8266,6 @@ export const MODELS = {
8267
8266
  contextWindow: 131072,
8268
8267
  maxTokens: 4096,
8269
8268
  },
8270
- "moonshotai/kimi-k2-0905:exacto": {
8271
- id: "moonshotai/kimi-k2-0905:exacto",
8272
- name: "MoonshotAI: Kimi K2 0905 (exacto)",
8273
- api: "openai-completions",
8274
- provider: "openrouter",
8275
- baseUrl: "https://openrouter.ai/api/v1",
8276
- reasoning: false,
8277
- input: ["text"],
8278
- cost: {
8279
- input: 0.6,
8280
- output: 2.5,
8281
- cacheRead: 0,
8282
- cacheWrite: 0,
8283
- },
8284
- contextWindow: 262144,
8285
- maxTokens: 4096,
8286
- },
8287
8269
  "moonshotai/kimi-k2-thinking": {
8288
8270
  id: "moonshotai/kimi-k2-thinking",
8289
8271
  name: "MoonshotAI: Kimi K2 Thinking",
@@ -8403,6 +8385,40 @@ export const MODELS = {
8403
8385
  contextWindow: 256000,
8404
8386
  maxTokens: 4096,
8405
8387
  },
8388
+ "nvidia/nemotron-3-super-120b-a12b": {
8389
+ id: "nvidia/nemotron-3-super-120b-a12b",
8390
+ name: "NVIDIA: Nemotron 3 Super",
8391
+ api: "openai-completions",
8392
+ provider: "openrouter",
8393
+ baseUrl: "https://openrouter.ai/api/v1",
8394
+ reasoning: true,
8395
+ input: ["text"],
8396
+ cost: {
8397
+ input: 0.09999999999999999,
8398
+ output: 0.5,
8399
+ cacheRead: 0.04,
8400
+ cacheWrite: 0,
8401
+ },
8402
+ contextWindow: 262144,
8403
+ maxTokens: 4096,
8404
+ },
8405
+ "nvidia/nemotron-3-super-120b-a12b:free": {
8406
+ id: "nvidia/nemotron-3-super-120b-a12b:free",
8407
+ name: "NVIDIA: Nemotron 3 Super (free)",
8408
+ api: "openai-completions",
8409
+ provider: "openrouter",
8410
+ baseUrl: "https://openrouter.ai/api/v1",
8411
+ reasoning: true,
8412
+ input: ["text"],
8413
+ cost: {
8414
+ input: 0,
8415
+ output: 0,
8416
+ cacheRead: 0,
8417
+ cacheWrite: 0,
8418
+ },
8419
+ contextWindow: 262144,
8420
+ maxTokens: 262144,
8421
+ },
8406
8422
  "nvidia/nemotron-nano-12b-v2-vl:free": {
8407
8423
  id: "nvidia/nemotron-nano-12b-v2-vl:free",
8408
8424
  name: "NVIDIA: Nemotron Nano 12B 2 VL (free)",
@@ -9100,6 +9116,40 @@ export const MODELS = {
9100
9116
  contextWindow: 1050000,
9101
9117
  maxTokens: 128000,
9102
9118
  },
9119
+ "openai/gpt-5.4-mini": {
9120
+ id: "openai/gpt-5.4-mini",
9121
+ name: "OpenAI: GPT-5.4 Mini",
9122
+ api: "openai-completions",
9123
+ provider: "openrouter",
9124
+ baseUrl: "https://openrouter.ai/api/v1",
9125
+ reasoning: true,
9126
+ input: ["text", "image"],
9127
+ cost: {
9128
+ input: 0.75,
9129
+ output: 4.5,
9130
+ cacheRead: 0.075,
9131
+ cacheWrite: 0,
9132
+ },
9133
+ contextWindow: 400000,
9134
+ maxTokens: 128000,
9135
+ },
9136
+ "openai/gpt-5.4-nano": {
9137
+ id: "openai/gpt-5.4-nano",
9138
+ name: "OpenAI: GPT-5.4 Nano",
9139
+ api: "openai-completions",
9140
+ provider: "openrouter",
9141
+ baseUrl: "https://openrouter.ai/api/v1",
9142
+ reasoning: true,
9143
+ input: ["text", "image"],
9144
+ cost: {
9145
+ input: 0.19999999999999998,
9146
+ output: 1.25,
9147
+ cacheRead: 0.02,
9148
+ cacheWrite: 0,
9149
+ },
9150
+ contextWindow: 400000,
9151
+ maxTokens: 128000,
9152
+ },
9103
9153
  "openai/gpt-5.4-pro": {
9104
9154
  id: "openai/gpt-5.4-pro",
9105
9155
  name: "OpenAI: GPT-5.4 Pro",
@@ -9134,23 +9184,6 @@ export const MODELS = {
9134
9184
  contextWindow: 131072,
9135
9185
  maxTokens: 4096,
9136
9186
  },
9137
- "openai/gpt-oss-120b:exacto": {
9138
- id: "openai/gpt-oss-120b:exacto",
9139
- name: "OpenAI: gpt-oss-120b (exacto)",
9140
- api: "openai-completions",
9141
- provider: "openrouter",
9142
- baseUrl: "https://openrouter.ai/api/v1",
9143
- reasoning: true,
9144
- input: ["text"],
9145
- cost: {
9146
- input: 0.039,
9147
- output: 0.19,
9148
- cacheRead: 0,
9149
- cacheWrite: 0,
9150
- },
9151
- contextWindow: 131072,
9152
- maxTokens: 4096,
9153
- },
9154
9187
  "openai/gpt-oss-120b:free": {
9155
9188
  id: "openai/gpt-oss-120b:free",
9156
9189
  name: "OpenAI: gpt-oss-120b (free)",
@@ -9178,12 +9211,12 @@ export const MODELS = {
9178
9211
  input: ["text"],
9179
9212
  cost: {
9180
9213
  input: 0.03,
9181
- output: 0.14,
9182
- cacheRead: 0,
9214
+ output: 0.11,
9215
+ cacheRead: 0.015,
9183
9216
  cacheWrite: 0,
9184
9217
  },
9185
9218
  contextWindow: 131072,
9186
- maxTokens: 4096,
9219
+ maxTokens: 131072,
9187
9220
  },
9188
9221
  "openai/gpt-oss-20b:free": {
9189
9222
  id: "openai/gpt-oss-20b:free",
@@ -9225,7 +9258,7 @@ export const MODELS = {
9225
9258
  api: "openai-completions",
9226
9259
  provider: "openrouter",
9227
9260
  baseUrl: "https://openrouter.ai/api/v1",
9228
- reasoning: false,
9261
+ reasoning: true,
9229
9262
  input: ["text", "image"],
9230
9263
  cost: {
9231
9264
  input: 15,
@@ -9276,7 +9309,7 @@ export const MODELS = {
9276
9309
  api: "openai-completions",
9277
9310
  provider: "openrouter",
9278
9311
  baseUrl: "https://openrouter.ai/api/v1",
9279
- reasoning: false,
9312
+ reasoning: true,
9280
9313
  input: ["text"],
9281
9314
  cost: {
9282
9315
  input: 1.1,
@@ -9293,7 +9326,7 @@ export const MODELS = {
9293
9326
  api: "openai-completions",
9294
9327
  provider: "openrouter",
9295
9328
  baseUrl: "https://openrouter.ai/api/v1",
9296
- reasoning: false,
9329
+ reasoning: true,
9297
9330
  input: ["text"],
9298
9331
  cost: {
9299
9332
  input: 1.1,
@@ -9483,9 +9516,9 @@ export const MODELS = {
9483
9516
  reasoning: false,
9484
9517
  input: ["text"],
9485
9518
  cost: {
9486
- input: 0.39999999999999997,
9487
- output: 1.2,
9488
- cacheRead: 0.08,
9519
+ input: 0.26,
9520
+ output: 0.78,
9521
+ cacheRead: 0.052000000000000005,
9489
9522
  cacheWrite: 0,
9490
9523
  },
9491
9524
  contextWindow: 1000000,
@@ -9551,8 +9584,8 @@ export const MODELS = {
9551
9584
  reasoning: false,
9552
9585
  input: ["text", "image"],
9553
9586
  cost: {
9554
- input: 0.7999999999999999,
9555
- output: 3.1999999999999997,
9587
+ input: 0.52,
9588
+ output: 2.08,
9556
9589
  cacheRead: 0,
9557
9590
  cacheWrite: 0,
9558
9591
  },
@@ -9619,13 +9652,13 @@ export const MODELS = {
9619
9652
  reasoning: true,
9620
9653
  input: ["text"],
9621
9654
  cost: {
9622
- input: 0.11,
9623
- output: 0.6,
9624
- cacheRead: 0.055,
9655
+ input: 0.14950000000000002,
9656
+ output: 1.495,
9657
+ cacheRead: 0,
9625
9658
  cacheWrite: 0,
9626
9659
  },
9627
- contextWindow: 262144,
9628
- maxTokens: 262144,
9660
+ contextWindow: 131072,
9661
+ maxTokens: 4096,
9629
9662
  },
9630
9663
  "qwen/qwen3-30b-a3b": {
9631
9664
  id: "qwen/qwen3-30b-a3b",
@@ -9670,13 +9703,13 @@ export const MODELS = {
9670
9703
  reasoning: true,
9671
9704
  input: ["text"],
9672
9705
  cost: {
9673
- input: 0.051,
9674
- output: 0.33999999999999997,
9675
- cacheRead: 0,
9706
+ input: 0.08,
9707
+ output: 0.39999999999999997,
9708
+ cacheRead: 0.08,
9676
9709
  cacheWrite: 0,
9677
9710
  },
9678
- contextWindow: 32768,
9679
- maxTokens: 4096,
9711
+ contextWindow: 131072,
9712
+ maxTokens: 131072,
9680
9713
  },
9681
9714
  "qwen/qwen3-32b": {
9682
9715
  id: "qwen/qwen3-32b",
@@ -9814,23 +9847,6 @@ export const MODELS = {
9814
9847
  contextWindow: 1000000,
9815
9848
  maxTokens: 65536,
9816
9849
  },
9817
- "qwen/qwen3-coder:exacto": {
9818
- id: "qwen/qwen3-coder:exacto",
9819
- name: "Qwen: Qwen3 Coder 480B A35B (exacto)",
9820
- api: "openai-completions",
9821
- provider: "openrouter",
9822
- baseUrl: "https://openrouter.ai/api/v1",
9823
- reasoning: false,
9824
- input: ["text"],
9825
- cost: {
9826
- input: 0.22,
9827
- output: 1.7999999999999998,
9828
- cacheRead: 0.022,
9829
- cacheWrite: 0,
9830
- },
9831
- contextWindow: 262144,
9832
- maxTokens: 65536,
9833
- },
9834
9850
  "qwen/qwen3-coder:free": {
9835
9851
  id: "qwen/qwen3-coder:free",
9836
9852
  name: "Qwen: Qwen3 Coder 480B A35B (free)",
@@ -9857,9 +9873,9 @@ export const MODELS = {
9857
9873
  reasoning: false,
9858
9874
  input: ["text"],
9859
9875
  cost: {
9860
- input: 1.2,
9861
- output: 6,
9862
- cacheRead: 0.24,
9876
+ input: 0.78,
9877
+ output: 3.9,
9878
+ cacheRead: 0.156,
9863
9879
  cacheWrite: 0,
9864
9880
  },
9865
9881
  contextWindow: 262144,
@@ -9925,13 +9941,13 @@ export const MODELS = {
9925
9941
  reasoning: true,
9926
9942
  input: ["text"],
9927
9943
  cost: {
9928
- input: 0.15,
9929
- output: 1.2,
9944
+ input: 0.0975,
9945
+ output: 0.78,
9930
9946
  cacheRead: 0,
9931
9947
  cacheWrite: 0,
9932
9948
  },
9933
- contextWindow: 128000,
9934
- maxTokens: 4096,
9949
+ contextWindow: 131072,
9950
+ maxTokens: 32768,
9935
9951
  },
9936
9952
  "qwen/qwen3-vl-235b-a22b-instruct": {
9937
9953
  id: "qwen/qwen3-vl-235b-a22b-instruct",
@@ -9959,8 +9975,8 @@ export const MODELS = {
9959
9975
  reasoning: true,
9960
9976
  input: ["text", "image"],
9961
9977
  cost: {
9962
- input: 0,
9963
- output: 0,
9978
+ input: 0.26,
9979
+ output: 2.6,
9964
9980
  cacheRead: 0,
9965
9981
  cacheWrite: 0,
9966
9982
  },
@@ -9993,8 +10009,8 @@ export const MODELS = {
9993
10009
  reasoning: true,
9994
10010
  input: ["text", "image"],
9995
10011
  cost: {
9996
- input: 0,
9997
- output: 0,
10012
+ input: 0.13,
10013
+ output: 1.56,
9998
10014
  cacheRead: 0,
9999
10015
  cacheWrite: 0,
10000
10016
  },
@@ -10120,6 +10136,23 @@ export const MODELS = {
10120
10136
  contextWindow: 262144,
10121
10137
  maxTokens: 65536,
10122
10138
  },
10139
+ "qwen/qwen3.5-9b": {
10140
+ id: "qwen/qwen3.5-9b",
10141
+ name: "Qwen: Qwen3.5-9B",
10142
+ api: "openai-completions",
10143
+ provider: "openrouter",
10144
+ baseUrl: "https://openrouter.ai/api/v1",
10145
+ reasoning: true,
10146
+ input: ["text", "image"],
10147
+ cost: {
10148
+ input: 0.049999999999999996,
10149
+ output: 0.15,
10150
+ cacheRead: 0,
10151
+ cacheWrite: 0,
10152
+ },
10153
+ contextWindow: 256000,
10154
+ maxTokens: 4096,
10155
+ },
10123
10156
  "qwen/qwen3.5-flash-02-23": {
10124
10157
  id: "qwen/qwen3.5-flash-02-23",
10125
10158
  name: "Qwen: Qwen3.5-Flash",
@@ -10129,8 +10162,8 @@ export const MODELS = {
10129
10162
  reasoning: true,
10130
10163
  input: ["text", "image"],
10131
10164
  cost: {
10132
- input: 0.09999999999999999,
10133
- output: 0.39999999999999997,
10165
+ input: 0.065,
10166
+ output: 0.26,
10134
10167
  cacheRead: 0,
10135
10168
  cacheWrite: 0,
10136
10169
  },
@@ -10164,12 +10197,12 @@ export const MODELS = {
10164
10197
  input: ["text"],
10165
10198
  cost: {
10166
10199
  input: 0.15,
10167
- output: 0.39999999999999997,
10200
+ output: 0.58,
10168
10201
  cacheRead: 0,
10169
10202
  cacheWrite: 0,
10170
10203
  },
10171
- contextWindow: 32768,
10172
- maxTokens: 32768,
10204
+ contextWindow: 131072,
10205
+ maxTokens: 131072,
10173
10206
  },
10174
10207
  "relace/relace-search": {
10175
10208
  id: "relace/relace-search",
@@ -10214,13 +10247,13 @@ export const MODELS = {
10214
10247
  reasoning: false,
10215
10248
  input: ["text"],
10216
10249
  cost: {
10217
- input: 0.65,
10218
- output: 0.75,
10250
+ input: 0.85,
10251
+ output: 0.85,
10219
10252
  cacheRead: 0,
10220
10253
  cacheWrite: 0,
10221
10254
  },
10222
- contextWindow: 32768,
10223
- maxTokens: 32768,
10255
+ contextWindow: 131072,
10256
+ maxTokens: 16384,
10224
10257
  },
10225
10258
  "stepfun/step-3.5-flash": {
10226
10259
  id: "stepfun/step-3.5-flash",
@@ -10299,9 +10332,9 @@ export const MODELS = {
10299
10332
  reasoning: true,
10300
10333
  input: ["text"],
10301
10334
  cost: {
10302
- input: 0.25,
10303
- output: 0.85,
10304
- cacheRead: 0.125,
10335
+ input: 0.3,
10336
+ output: 1.1,
10337
+ cacheRead: 0.15,
10305
10338
  cacheWrite: 0,
10306
10339
  },
10307
10340
  contextWindow: 163840,
@@ -10443,6 +10476,23 @@ export const MODELS = {
10443
10476
  contextWindow: 2000000,
10444
10477
  maxTokens: 30000,
10445
10478
  },
10479
+ "x-ai/grok-4.20-beta": {
10480
+ id: "x-ai/grok-4.20-beta",
10481
+ name: "xAI: Grok 4.20 Beta",
10482
+ api: "openai-completions",
10483
+ provider: "openrouter",
10484
+ baseUrl: "https://openrouter.ai/api/v1",
10485
+ reasoning: true,
10486
+ input: ["text", "image"],
10487
+ cost: {
10488
+ input: 2,
10489
+ output: 6,
10490
+ cacheRead: 0.19999999999999998,
10491
+ cacheWrite: 0,
10492
+ },
10493
+ contextWindow: 2000000,
10494
+ maxTokens: 4096,
10495
+ },
10446
10496
  "x-ai/grok-code-fast-1": {
10447
10497
  id: "x-ai/grok-code-fast-1",
10448
10498
  name: "xAI: Grok Code Fast 1",
@@ -10477,6 +10527,40 @@ export const MODELS = {
10477
10527
  contextWindow: 262144,
10478
10528
  maxTokens: 65536,
10479
10529
  },
10530
+ "xiaomi/mimo-v2-omni": {
10531
+ id: "xiaomi/mimo-v2-omni",
10532
+ name: "Xiaomi: MiMo-V2-Omni",
10533
+ api: "openai-completions",
10534
+ provider: "openrouter",
10535
+ baseUrl: "https://openrouter.ai/api/v1",
10536
+ reasoning: true,
10537
+ input: ["text", "image"],
10538
+ cost: {
10539
+ input: 0.39999999999999997,
10540
+ output: 2,
10541
+ cacheRead: 0.08,
10542
+ cacheWrite: 0,
10543
+ },
10544
+ contextWindow: 262144,
10545
+ maxTokens: 65536,
10546
+ },
10547
+ "xiaomi/mimo-v2-pro": {
10548
+ id: "xiaomi/mimo-v2-pro",
10549
+ name: "Xiaomi: MiMo-V2-Pro",
10550
+ api: "openai-completions",
10551
+ provider: "openrouter",
10552
+ baseUrl: "https://openrouter.ai/api/v1",
10553
+ reasoning: true,
10554
+ input: ["text"],
10555
+ cost: {
10556
+ input: 1,
10557
+ output: 3,
10558
+ cacheRead: 0.19999999999999998,
10559
+ cacheWrite: 0,
10560
+ },
10561
+ contextWindow: 1048576,
10562
+ maxTokens: 131072,
10563
+ },
10480
10564
  "z-ai/glm-4-32b": {
10481
10565
  id: "z-ai/glm-4-32b",
10482
10566
  name: "Z.ai: GLM 4 32B ",
@@ -10579,23 +10663,6 @@ export const MODELS = {
10579
10663
  contextWindow: 204800,
10580
10664
  maxTokens: 204800,
10581
10665
  },
10582
- "z-ai/glm-4.6:exacto": {
10583
- id: "z-ai/glm-4.6:exacto",
10584
- name: "Z.ai: GLM 4.6 (exacto)",
10585
- api: "openai-completions",
10586
- provider: "openrouter",
10587
- baseUrl: "https://openrouter.ai/api/v1",
10588
- reasoning: true,
10589
- input: ["text"],
10590
- cost: {
10591
- input: 0.44,
10592
- output: 1.76,
10593
- cacheRead: 0.11,
10594
- cacheWrite: 0,
10595
- },
10596
- contextWindow: 204800,
10597
- maxTokens: 131072,
10598
- },
10599
10666
  "z-ai/glm-4.6v": {
10600
10667
  id: "z-ai/glm-4.6v",
10601
10668
  name: "Z.ai: GLM 4.6V",
@@ -10622,13 +10689,13 @@ export const MODELS = {
10622
10689
  reasoning: true,
10623
10690
  input: ["text"],
10624
10691
  cost: {
10625
- input: 0.38,
10626
- output: 1.9800000000000002,
10627
- cacheRead: 0.19,
10692
+ input: 0.39,
10693
+ output: 1.75,
10694
+ cacheRead: 0.195,
10628
10695
  cacheWrite: 0,
10629
10696
  },
10630
10697
  contextWindow: 202752,
10631
- maxTokens: 4096,
10698
+ maxTokens: 65535,
10632
10699
  },
10633
10700
  "z-ai/glm-4.7-flash": {
10634
10701
  id: "z-ai/glm-4.7-flash",
@@ -10661,8 +10728,25 @@ export const MODELS = {
10661
10728
  cacheRead: 0.119,
10662
10729
  cacheWrite: 0,
10663
10730
  },
10731
+ contextWindow: 80000,
10732
+ maxTokens: 131072,
10733
+ },
10734
+ "z-ai/glm-5-turbo": {
10735
+ id: "z-ai/glm-5-turbo",
10736
+ name: "Z.ai: GLM 5 Turbo",
10737
+ api: "openai-completions",
10738
+ provider: "openrouter",
10739
+ baseUrl: "https://openrouter.ai/api/v1",
10740
+ reasoning: true,
10741
+ input: ["text"],
10742
+ cost: {
10743
+ input: 0.96,
10744
+ output: 3.1999999999999997,
10745
+ cacheRead: 0.192,
10746
+ cacheWrite: 0,
10747
+ },
10664
10748
  contextWindow: 202752,
10665
- maxTokens: 4096,
10749
+ maxTokens: 131072,
10666
10750
  },
10667
10751
  },
10668
10752
  "vercel-ai-gateway": {
@@ -10675,7 +10759,7 @@ export const MODELS = {
10675
10759
  reasoning: true,
10676
10760
  input: ["text"],
10677
10761
  cost: {
10678
- input: 0.06,
10762
+ input: 0.12,
10679
10763
  output: 0.24,
10680
10764
  cacheRead: 0,
10681
10765
  cacheWrite: 0,
@@ -10726,13 +10810,13 @@ export const MODELS = {
10726
10810
  reasoning: true,
10727
10811
  input: ["text"],
10728
10812
  cost: {
10729
- input: 0.09999999999999999,
10730
- output: 0.3,
10731
- cacheRead: 0,
10813
+ input: 0.29,
10814
+ output: 0.59,
10815
+ cacheRead: 0.145,
10732
10816
  cacheWrite: 0,
10733
10817
  },
10734
- contextWindow: 40960,
10735
- maxTokens: 16384,
10818
+ contextWindow: 131072,
10819
+ maxTokens: 40960,
10736
10820
  },
10737
10821
  "alibaba/qwen3-235b-a22b-thinking": {
10738
10822
  id: "alibaba/qwen3-235b-a22b-thinking",
@@ -10743,9 +10827,9 @@ export const MODELS = {
10743
10827
  reasoning: true,
10744
10828
  input: ["text", "image"],
10745
10829
  cost: {
10746
- input: 0.3,
10747
- output: 2.9000000000000004,
10748
- cacheRead: 0,
10830
+ input: 0.22999999999999998,
10831
+ output: 2.3,
10832
+ cacheRead: 0.19999999999999998,
10749
10833
  cacheWrite: 0,
10750
10834
  },
10751
10835
  contextWindow: 262114,
@@ -10762,7 +10846,7 @@ export const MODELS = {
10762
10846
  cost: {
10763
10847
  input: 0.39999999999999997,
10764
10848
  output: 1.5999999999999999,
10765
- cacheRead: 0,
10849
+ cacheRead: 0.022,
10766
10850
  cacheWrite: 0,
10767
10851
  },
10768
10852
  contextWindow: 262144,
@@ -10777,13 +10861,13 @@ export const MODELS = {
10777
10861
  reasoning: true,
10778
10862
  input: ["text"],
10779
10863
  cost: {
10780
- input: 0.07,
10781
- output: 0.27,
10864
+ input: 0.15,
10865
+ output: 0.6,
10782
10866
  cacheRead: 0,
10783
10867
  cacheWrite: 0,
10784
10868
  },
10785
- contextWindow: 160000,
10786
- maxTokens: 32768,
10869
+ contextWindow: 262144,
10870
+ maxTokens: 8192,
10787
10871
  },
10788
10872
  "alibaba/qwen3-coder-next": {
10789
10873
  id: "alibaba/qwen3-coder-next",
@@ -10791,7 +10875,7 @@ export const MODELS = {
10791
10875
  api: "anthropic-messages",
10792
10876
  provider: "vercel-ai-gateway",
10793
10877
  baseUrl: "https://ai-gateway.vercel.sh",
10794
- reasoning: true,
10878
+ reasoning: false,
10795
10879
  input: ["text"],
10796
10880
  cost: {
10797
10881
  input: 0.5,
@@ -10819,6 +10903,23 @@ export const MODELS = {
10819
10903
  contextWindow: 1000000,
10820
10904
  maxTokens: 65536,
10821
10905
  },
10906
+ "alibaba/qwen3-max": {
10907
+ id: "alibaba/qwen3-max",
10908
+ name: "Qwen3 Max",
10909
+ api: "anthropic-messages",
10910
+ provider: "vercel-ai-gateway",
10911
+ baseUrl: "https://ai-gateway.vercel.sh",
10912
+ reasoning: false,
10913
+ input: ["text"],
10914
+ cost: {
10915
+ input: 1.2,
10916
+ output: 6,
10917
+ cacheRead: 0.24,
10918
+ cacheWrite: 0,
10919
+ },
10920
+ contextWindow: 262144,
10921
+ maxTokens: 32768,
10922
+ },
10822
10923
  "alibaba/qwen3-max-preview": {
10823
10924
  id: "alibaba/qwen3-max-preview",
10824
10925
  name: "Qwen3 Max Preview",
@@ -10966,8 +11067,8 @@ export const MODELS = {
10966
11067
  cost: {
10967
11068
  input: 3,
10968
11069
  output: 15,
10969
- cacheRead: 0,
10970
- cacheWrite: 0,
11070
+ cacheRead: 0.3,
11071
+ cacheWrite: 3.75,
10971
11072
  },
10972
11073
  contextWindow: 200000,
10973
11074
  maxTokens: 8192,
@@ -11176,6 +11277,23 @@ export const MODELS = {
11176
11277
  contextWindow: 256000,
11177
11278
  maxTokens: 8000,
11178
11279
  },
11280
+ "deepseek/deepseek-r1": {
11281
+ id: "deepseek/deepseek-r1",
11282
+ name: "DeepSeek-R1",
11283
+ api: "anthropic-messages",
11284
+ provider: "vercel-ai-gateway",
11285
+ baseUrl: "https://ai-gateway.vercel.sh",
11286
+ reasoning: true,
11287
+ input: ["text"],
11288
+ cost: {
11289
+ input: 1.35,
11290
+ output: 5.4,
11291
+ cacheRead: 0,
11292
+ cacheWrite: 0,
11293
+ },
11294
+ contextWindow: 128000,
11295
+ maxTokens: 8192,
11296
+ },
11179
11297
  "deepseek/deepseek-v3": {
11180
11298
  id: "deepseek/deepseek-v3",
11181
11299
  name: "DeepSeek V3 0324",
@@ -11202,13 +11320,13 @@ export const MODELS = {
11202
11320
  reasoning: true,
11203
11321
  input: ["text"],
11204
11322
  cost: {
11205
- input: 0.21,
11206
- output: 0.7899999999999999,
11323
+ input: 0.5,
11324
+ output: 1.5,
11207
11325
  cacheRead: 0,
11208
11326
  cacheWrite: 0,
11209
11327
  },
11210
11328
  contextWindow: 163840,
11211
- maxTokens: 128000,
11329
+ maxTokens: 16384,
11212
11330
  },
11213
11331
  "deepseek/deepseek-v3.1-terminus": {
11214
11332
  id: "deepseek/deepseek-v3.1-terminus",
@@ -11221,7 +11339,7 @@ export const MODELS = {
11221
11339
  cost: {
11222
11340
  input: 0.27,
11223
11341
  output: 1,
11224
- cacheRead: 0,
11342
+ cacheRead: 0.135,
11225
11343
  cacheWrite: 0,
11226
11344
  },
11227
11345
  contextWindow: 131072,
@@ -11236,9 +11354,9 @@ export const MODELS = {
11236
11354
  reasoning: false,
11237
11355
  input: ["text"],
11238
11356
  cost: {
11239
- input: 0.26,
11240
- output: 0.38,
11241
- cacheRead: 0.13,
11357
+ input: 0.28,
11358
+ output: 0.42,
11359
+ cacheRead: 0.028,
11242
11360
  cacheWrite: 0,
11243
11361
  },
11244
11362
  contextWindow: 128000,
@@ -11261,72 +11379,72 @@ export const MODELS = {
11261
11379
  contextWindow: 128000,
11262
11380
  maxTokens: 64000,
11263
11381
  },
11264
- "google/gemini-2.5-flash": {
11265
- id: "google/gemini-2.5-flash",
11266
- name: "Gemini 2.5 Flash",
11382
+ "google/gemini-2.0-flash": {
11383
+ id: "google/gemini-2.0-flash",
11384
+ name: "Gemini 2.0 Flash",
11267
11385
  api: "anthropic-messages",
11268
11386
  provider: "vercel-ai-gateway",
11269
11387
  baseUrl: "https://ai-gateway.vercel.sh",
11270
- reasoning: true,
11271
- input: ["text"],
11388
+ reasoning: false,
11389
+ input: ["text", "image"],
11272
11390
  cost: {
11273
- input: 0.3,
11274
- output: 2.5,
11275
- cacheRead: 0,
11391
+ input: 0.15,
11392
+ output: 0.6,
11393
+ cacheRead: 0.024999999999999998,
11276
11394
  cacheWrite: 0,
11277
11395
  },
11278
- contextWindow: 1000000,
11279
- maxTokens: 65536,
11396
+ contextWindow: 1048576,
11397
+ maxTokens: 8192,
11280
11398
  },
11281
- "google/gemini-2.5-flash-lite": {
11282
- id: "google/gemini-2.5-flash-lite",
11283
- name: "Gemini 2.5 Flash Lite",
11399
+ "google/gemini-2.0-flash-lite": {
11400
+ id: "google/gemini-2.0-flash-lite",
11401
+ name: "Gemini 2.0 Flash Lite",
11284
11402
  api: "anthropic-messages",
11285
11403
  provider: "vercel-ai-gateway",
11286
11404
  baseUrl: "https://ai-gateway.vercel.sh",
11287
- reasoning: true,
11405
+ reasoning: false,
11288
11406
  input: ["text", "image"],
11289
11407
  cost: {
11290
- input: 0.09999999999999999,
11291
- output: 0.39999999999999997,
11292
- cacheRead: 0.01,
11408
+ input: 0.075,
11409
+ output: 0.3,
11410
+ cacheRead: 0.02,
11293
11411
  cacheWrite: 0,
11294
11412
  },
11295
11413
  contextWindow: 1048576,
11296
- maxTokens: 65536,
11414
+ maxTokens: 8192,
11297
11415
  },
11298
- "google/gemini-2.5-flash-lite-preview-09-2025": {
11299
- id: "google/gemini-2.5-flash-lite-preview-09-2025",
11300
- name: "Gemini 2.5 Flash Lite Preview 09-2025",
11416
+ "google/gemini-2.5-flash": {
11417
+ id: "google/gemini-2.5-flash",
11418
+ name: "Gemini 2.5 Flash",
11301
11419
  api: "anthropic-messages",
11302
11420
  provider: "vercel-ai-gateway",
11303
11421
  baseUrl: "https://ai-gateway.vercel.sh",
11304
11422
  reasoning: true,
11305
11423
  input: ["text", "image"],
11306
11424
  cost: {
11307
- input: 0.09999999999999999,
11308
- output: 0.39999999999999997,
11309
- cacheRead: 0.01,
11425
+ input: 0.3,
11426
+ output: 2.5,
11427
+ cacheRead: 0.03,
11310
11428
  cacheWrite: 0,
11311
11429
  },
11312
- contextWindow: 1048576,
11430
+ contextWindow: 1000000,
11313
11431
  maxTokens: 65536,
11314
11432
  },
11315
- "google/gemini-2.5-flash-preview-09-2025": {
11316
- id: "google/gemini-2.5-flash-preview-09-2025",
11317
- name: "Gemini 2.5 Flash Preview 09-2025",
11433
+ "google/gemini-2.5-flash-lite": {
11434
+ id: "google/gemini-2.5-flash-lite",
11435
+ name: "Gemini 2.5 Flash Lite",
11318
11436
  api: "anthropic-messages",
11319
11437
  provider: "vercel-ai-gateway",
11320
11438
  baseUrl: "https://ai-gateway.vercel.sh",
11321
11439
  reasoning: true,
11322
11440
  input: ["text", "image"],
11323
11441
  cost: {
11324
- input: 0.3,
11325
- output: 2.5,
11326
- cacheRead: 0.03,
11442
+ input: 0.09999999999999999,
11443
+ output: 0.39999999999999997,
11444
+ cacheRead: 0.01,
11327
11445
  cacheWrite: 0,
11328
11446
  },
11329
- contextWindow: 1000000,
11447
+ contextWindow: 1048576,
11330
11448
  maxTokens: 65536,
11331
11449
  },
11332
11450
  "google/gemini-2.5-pro": {
@@ -11336,11 +11454,11 @@ export const MODELS = {
11336
11454
  provider: "vercel-ai-gateway",
11337
11455
  baseUrl: "https://ai-gateway.vercel.sh",
11338
11456
  reasoning: true,
11339
- input: ["text"],
11457
+ input: ["text", "image"],
11340
11458
  cost: {
11341
11459
  input: 1.25,
11342
11460
  output: 10,
11343
- cacheRead: 0,
11461
+ cacheRead: 0.125,
11344
11462
  cacheWrite: 0,
11345
11463
  },
11346
11464
  contextWindow: 1048576,
@@ -11361,7 +11479,7 @@ export const MODELS = {
11361
11479
  cacheWrite: 0,
11362
11480
  },
11363
11481
  contextWindow: 1000000,
11364
- maxTokens: 64000,
11482
+ maxTokens: 65000,
11365
11483
  },
11366
11484
  "google/gemini-3-pro-preview": {
11367
11485
  id: "google/gemini-3-pro-preview",
@@ -11463,7 +11581,7 @@ export const MODELS = {
11463
11581
  cacheWrite: 0,
11464
11582
  },
11465
11583
  contextWindow: 128000,
11466
- maxTokens: 8192,
11584
+ maxTokens: 100000,
11467
11585
  },
11468
11586
  "meituan/longcat-flash-thinking": {
11469
11587
  id: "meituan/longcat-flash-thinking",
@@ -11491,13 +11609,13 @@ export const MODELS = {
11491
11609
  reasoning: false,
11492
11610
  input: ["text"],
11493
11611
  cost: {
11494
- input: 0.39999999999999997,
11495
- output: 0.39999999999999997,
11612
+ input: 0.72,
11613
+ output: 0.72,
11496
11614
  cacheRead: 0,
11497
11615
  cacheWrite: 0,
11498
11616
  },
11499
- contextWindow: 131072,
11500
- maxTokens: 16384,
11617
+ contextWindow: 128000,
11618
+ maxTokens: 8192,
11501
11619
  },
11502
11620
  "meta/llama-3.1-8b": {
11503
11621
  id: "meta/llama-3.1-8b",
@@ -11508,12 +11626,12 @@ export const MODELS = {
11508
11626
  reasoning: false,
11509
11627
  input: ["text"],
11510
11628
  cost: {
11511
- input: 0.03,
11512
- output: 0.049999999999999996,
11513
- cacheRead: 0,
11629
+ input: 0.09999999999999999,
11630
+ output: 0.09999999999999999,
11631
+ cacheRead: 0.09999999999999999,
11514
11632
  cacheWrite: 0,
11515
11633
  },
11516
- contextWindow: 131072,
11634
+ contextWindow: 128000,
11517
11635
  maxTokens: 16384,
11518
11636
  },
11519
11637
  "meta/llama-3.2-11b": {
@@ -11576,12 +11694,12 @@ export const MODELS = {
11576
11694
  reasoning: false,
11577
11695
  input: ["text", "image"],
11578
11696
  cost: {
11579
- input: 0.15,
11580
- output: 0.6,
11697
+ input: 0.24,
11698
+ output: 0.9700000000000001,
11581
11699
  cacheRead: 0,
11582
11700
  cacheWrite: 0,
11583
11701
  },
11584
- contextWindow: 131072,
11702
+ contextWindow: 128000,
11585
11703
  maxTokens: 8192,
11586
11704
  },
11587
11705
  "meta/llama-4-scout": {
@@ -11593,12 +11711,12 @@ export const MODELS = {
11593
11711
  reasoning: false,
11594
11712
  input: ["text", "image"],
11595
11713
  cost: {
11596
- input: 0.08,
11597
- output: 0.3,
11714
+ input: 0.16999999999999998,
11715
+ output: 0.66,
11598
11716
  cacheRead: 0,
11599
11717
  cacheWrite: 0,
11600
11718
  },
11601
- contextWindow: 131072,
11719
+ contextWindow: 128000,
11602
11720
  maxTokens: 8192,
11603
11721
  },
11604
11722
  "minimax/minimax-m2": {
@@ -11629,8 +11747,8 @@ export const MODELS = {
11629
11747
  cost: {
11630
11748
  input: 0.3,
11631
11749
  output: 1.2,
11632
- cacheRead: 0.15,
11633
- cacheWrite: 0,
11750
+ cacheRead: 0.03,
11751
+ cacheWrite: 0.375,
11634
11752
  },
11635
11753
  contextWindow: 204800,
11636
11754
  maxTokens: 131072,
@@ -11683,8 +11801,42 @@ export const MODELS = {
11683
11801
  cacheRead: 0.03,
11684
11802
  cacheWrite: 0.375,
11685
11803
  },
11686
- contextWindow: 4096,
11687
- maxTokens: 4096,
11804
+ contextWindow: 204800,
11805
+ maxTokens: 131000,
11806
+ },
11807
+ "minimax/minimax-m2.7": {
11808
+ id: "minimax/minimax-m2.7",
11809
+ name: "Minimax M2.7",
11810
+ api: "anthropic-messages",
11811
+ provider: "vercel-ai-gateway",
11812
+ baseUrl: "https://ai-gateway.vercel.sh",
11813
+ reasoning: true,
11814
+ input: ["text", "image"],
11815
+ cost: {
11816
+ input: 0.3,
11817
+ output: 1.2,
11818
+ cacheRead: 0.06,
11819
+ cacheWrite: 0.375,
11820
+ },
11821
+ contextWindow: 204800,
11822
+ maxTokens: 131000,
11823
+ },
11824
+ "minimax/minimax-m2.7-highspeed": {
11825
+ id: "minimax/minimax-m2.7-highspeed",
11826
+ name: "MiniMax M2.7 High Speed",
11827
+ api: "anthropic-messages",
11828
+ provider: "vercel-ai-gateway",
11829
+ baseUrl: "https://ai-gateway.vercel.sh",
11830
+ reasoning: true,
11831
+ input: ["text", "image"],
11832
+ cost: {
11833
+ input: 0.6,
11834
+ output: 2.4,
11835
+ cacheRead: 0.06,
11836
+ cacheWrite: 0.375,
11837
+ },
11838
+ contextWindow: 204800,
11839
+ maxTokens: 131100,
11688
11840
  },
11689
11841
  "mistral/codestral": {
11690
11842
  id: "mistral/codestral",
@@ -11712,8 +11864,8 @@ export const MODELS = {
11712
11864
  reasoning: false,
11713
11865
  input: ["text"],
11714
11866
  cost: {
11715
- input: 0,
11716
- output: 0,
11867
+ input: 0.39999999999999997,
11868
+ output: 2,
11717
11869
  cacheRead: 0,
11718
11870
  cacheWrite: 0,
11719
11871
  },
@@ -11746,8 +11898,8 @@ export const MODELS = {
11746
11898
  reasoning: false,
11747
11899
  input: ["text"],
11748
11900
  cost: {
11749
- input: 0,
11750
- output: 0,
11901
+ input: 0.09999999999999999,
11902
+ output: 0.3,
11751
11903
  cacheRead: 0,
11752
11904
  cacheWrite: 0,
11753
11905
  },
@@ -11763,8 +11915,8 @@ export const MODELS = {
11763
11915
  reasoning: false,
11764
11916
  input: ["text"],
11765
11917
  cost: {
11766
- input: 0.04,
11767
- output: 0.04,
11918
+ input: 0.09999999999999999,
11919
+ output: 0.09999999999999999,
11768
11920
  cacheRead: 0,
11769
11921
  cacheWrite: 0,
11770
11922
  },
@@ -11780,8 +11932,8 @@ export const MODELS = {
11780
11932
  reasoning: false,
11781
11933
  input: ["text"],
11782
11934
  cost: {
11783
- input: 0.09999999999999999,
11784
- output: 0.09999999999999999,
11935
+ input: 0.15,
11936
+ output: 0.15,
11785
11937
  cacheRead: 0,
11786
11938
  cacheWrite: 0,
11787
11939
  },
@@ -11865,14 +12017,31 @@ export const MODELS = {
11865
12017
  reasoning: false,
11866
12018
  input: ["text"],
11867
12019
  cost: {
11868
- input: 0.5,
11869
- output: 2,
11870
- cacheRead: 0,
12020
+ input: 0.6,
12021
+ output: 2.5,
12022
+ cacheRead: 0.15,
11871
12023
  cacheWrite: 0,
11872
12024
  },
11873
12025
  contextWindow: 131072,
11874
12026
  maxTokens: 16384,
11875
12027
  },
12028
+ "moonshotai/kimi-k2-0905": {
12029
+ id: "moonshotai/kimi-k2-0905",
12030
+ name: "Kimi K2 0905",
12031
+ api: "anthropic-messages",
12032
+ provider: "vercel-ai-gateway",
12033
+ baseUrl: "https://ai-gateway.vercel.sh",
12034
+ reasoning: false,
12035
+ input: ["text"],
12036
+ cost: {
12037
+ input: 0.6,
12038
+ output: 2.5,
12039
+ cacheRead: 0.15,
12040
+ cacheWrite: 0,
12041
+ },
12042
+ contextWindow: 256000,
12043
+ maxTokens: 16384,
12044
+ },
11876
12045
  "moonshotai/kimi-k2-thinking": {
11877
12046
  id: "moonshotai/kimi-k2-thinking",
11878
12047
  name: "Kimi K2 Thinking",
@@ -11882,13 +12051,13 @@ export const MODELS = {
11882
12051
  reasoning: true,
11883
12052
  input: ["text"],
11884
12053
  cost: {
11885
- input: 0.47,
11886
- output: 2,
11887
- cacheRead: 0.14100000000000001,
12054
+ input: 0.6,
12055
+ output: 2.5,
12056
+ cacheRead: 0.15,
11888
12057
  cacheWrite: 0,
11889
12058
  },
11890
- contextWindow: 216144,
11891
- maxTokens: 216144,
12059
+ contextWindow: 262114,
12060
+ maxTokens: 262114,
11892
12061
  },
11893
12062
  "moonshotai/kimi-k2-thinking-turbo": {
11894
12063
  id: "moonshotai/kimi-k2-thinking-turbo",
@@ -11916,9 +12085,9 @@ export const MODELS = {
11916
12085
  reasoning: false,
11917
12086
  input: ["text"],
11918
12087
  cost: {
11919
- input: 2.4,
11920
- output: 10,
11921
- cacheRead: 0,
12088
+ input: 1.15,
12089
+ output: 8,
12090
+ cacheRead: 0.15,
11922
12091
  cacheWrite: 0,
11923
12092
  },
11924
12093
  contextWindow: 256000,
@@ -11933,13 +12102,13 @@ export const MODELS = {
11933
12102
  reasoning: true,
11934
12103
  input: ["text", "image"],
11935
12104
  cost: {
11936
- input: 0.5,
11937
- output: 2.8,
11938
- cacheRead: 0,
12105
+ input: 0.6,
12106
+ output: 3,
12107
+ cacheRead: 0.09999999999999999,
11939
12108
  cacheWrite: 0,
11940
12109
  },
11941
- contextWindow: 256000,
11942
- maxTokens: 256000,
12110
+ contextWindow: 262114,
12111
+ maxTokens: 262114,
11943
12112
  },
11944
12113
  "nvidia/nemotron-nano-12b-v2-vl": {
11945
12114
  id: "nvidia/nemotron-nano-12b-v2-vl",
@@ -11967,31 +12136,14 @@ export const MODELS = {
11967
12136
  reasoning: true,
11968
12137
  input: ["text"],
11969
12138
  cost: {
11970
- input: 0.04,
11971
- output: 0.16,
12139
+ input: 0.06,
12140
+ output: 0.22999999999999998,
11972
12141
  cacheRead: 0,
11973
12142
  cacheWrite: 0,
11974
12143
  },
11975
12144
  contextWindow: 131072,
11976
12145
  maxTokens: 131072,
11977
12146
  },
11978
- "openai/codex-mini": {
11979
- id: "openai/codex-mini",
11980
- name: "Codex Mini",
11981
- api: "anthropic-messages",
11982
- provider: "vercel-ai-gateway",
11983
- baseUrl: "https://ai-gateway.vercel.sh",
11984
- reasoning: true,
11985
- input: ["text", "image"],
11986
- cost: {
11987
- input: 1.5,
11988
- output: 6,
11989
- cacheRead: 0.375,
11990
- cacheWrite: 0,
11991
- },
11992
- contextWindow: 200000,
11993
- maxTokens: 100000,
11994
- },
11995
12147
  "openai/gpt-4-turbo": {
11996
12148
  id: "openai/gpt-4-turbo",
11997
12149
  name: "GPT-4 Turbo",
@@ -12054,7 +12206,7 @@ export const MODELS = {
12054
12206
  cost: {
12055
12207
  input: 0.09999999999999999,
12056
12208
  output: 0.39999999999999997,
12057
- cacheRead: 0.03,
12209
+ cacheRead: 0.024999999999999998,
12058
12210
  cacheWrite: 0,
12059
12211
  },
12060
12212
  contextWindow: 1047576,
@@ -12105,7 +12257,7 @@ export const MODELS = {
12105
12257
  cost: {
12106
12258
  input: 1.25,
12107
12259
  output: 10,
12108
- cacheRead: 0.13,
12260
+ cacheRead: 0.125,
12109
12261
  cacheWrite: 0,
12110
12262
  },
12111
12263
  contextWindow: 400000,
@@ -12135,11 +12287,11 @@ export const MODELS = {
12135
12287
  provider: "vercel-ai-gateway",
12136
12288
  baseUrl: "https://ai-gateway.vercel.sh",
12137
12289
  reasoning: true,
12138
- input: ["text", "image"],
12290
+ input: ["text"],
12139
12291
  cost: {
12140
12292
  input: 1.25,
12141
12293
  output: 10,
12142
- cacheRead: 0.13,
12294
+ cacheRead: 0.125,
12143
12295
  cacheWrite: 0,
12144
12296
  },
12145
12297
  contextWindow: 400000,
@@ -12156,7 +12308,7 @@ export const MODELS = {
12156
12308
  cost: {
12157
12309
  input: 0.25,
12158
12310
  output: 2,
12159
- cacheRead: 0.03,
12311
+ cacheRead: 0.024999999999999998,
12160
12312
  cacheWrite: 0,
12161
12313
  },
12162
12314
  contextWindow: 400000,
@@ -12173,7 +12325,7 @@ export const MODELS = {
12173
12325
  cost: {
12174
12326
  input: 0.049999999999999996,
12175
12327
  output: 0.39999999999999997,
12176
- cacheRead: 0.01,
12328
+ cacheRead: 0.005,
12177
12329
  cacheWrite: 0,
12178
12330
  },
12179
12331
  contextWindow: 400000,
@@ -12207,7 +12359,7 @@ export const MODELS = {
12207
12359
  cost: {
12208
12360
  input: 1.25,
12209
12361
  output: 10,
12210
- cacheRead: 0.13,
12362
+ cacheRead: 0.125,
12211
12363
  cacheWrite: 0,
12212
12364
  },
12213
12365
  contextWindow: 400000,
@@ -12258,7 +12410,7 @@ export const MODELS = {
12258
12410
  cost: {
12259
12411
  input: 1.25,
12260
12412
  output: 10,
12261
- cacheRead: 0.13,
12413
+ cacheRead: 0.125,
12262
12414
  cacheWrite: 0,
12263
12415
  },
12264
12416
  contextWindow: 128000,
@@ -12275,7 +12427,7 @@ export const MODELS = {
12275
12427
  cost: {
12276
12428
  input: 1.25,
12277
12429
  output: 10,
12278
- cacheRead: 0.13,
12430
+ cacheRead: 0.125,
12279
12431
  cacheWrite: 0,
12280
12432
  },
12281
12433
  contextWindow: 400000,
@@ -12292,7 +12444,7 @@ export const MODELS = {
12292
12444
  cost: {
12293
12445
  input: 1.75,
12294
12446
  output: 14,
12295
- cacheRead: 0.18,
12447
+ cacheRead: 0.175,
12296
12448
  cacheWrite: 0,
12297
12449
  },
12298
12450
  contextWindow: 400000,
@@ -12397,42 +12549,59 @@ export const MODELS = {
12397
12549
  cacheRead: 0.25,
12398
12550
  cacheWrite: 0,
12399
12551
  },
12400
- contextWindow: 200000,
12552
+ contextWindow: 1050000,
12401
12553
  maxTokens: 128000,
12402
12554
  },
12403
- "openai/gpt-5.4-pro": {
12404
- id: "openai/gpt-5.4-pro",
12405
- name: "GPT 5.4 Pro",
12555
+ "openai/gpt-5.4-mini": {
12556
+ id: "openai/gpt-5.4-mini",
12557
+ name: "GPT 5.4 Mini",
12406
12558
  api: "anthropic-messages",
12407
12559
  provider: "vercel-ai-gateway",
12408
12560
  baseUrl: "https://ai-gateway.vercel.sh",
12409
12561
  reasoning: true,
12410
12562
  input: ["text", "image"],
12411
12563
  cost: {
12412
- input: 30,
12413
- output: 180,
12414
- cacheRead: 0,
12564
+ input: 0.75,
12565
+ output: 4.5,
12566
+ cacheRead: 0.075,
12415
12567
  cacheWrite: 0,
12416
12568
  },
12417
- contextWindow: 1050000,
12569
+ contextWindow: 400000,
12418
12570
  maxTokens: 128000,
12419
12571
  },
12420
- "openai/gpt-oss-120b": {
12421
- id: "openai/gpt-oss-120b",
12422
- name: "gpt-oss-120b",
12572
+ "openai/gpt-5.4-nano": {
12573
+ id: "openai/gpt-5.4-nano",
12574
+ name: "GPT 5.4 Nano",
12423
12575
  api: "anthropic-messages",
12424
12576
  provider: "vercel-ai-gateway",
12425
12577
  baseUrl: "https://ai-gateway.vercel.sh",
12426
12578
  reasoning: true,
12427
- input: ["text"],
12579
+ input: ["text", "image"],
12428
12580
  cost: {
12429
- input: 0.09999999999999999,
12430
- output: 0.5,
12581
+ input: 0.19999999999999998,
12582
+ output: 1.25,
12583
+ cacheRead: 0.02,
12584
+ cacheWrite: 0,
12585
+ },
12586
+ contextWindow: 400000,
12587
+ maxTokens: 128000,
12588
+ },
12589
+ "openai/gpt-5.4-pro": {
12590
+ id: "openai/gpt-5.4-pro",
12591
+ name: "GPT 5.4 Pro",
12592
+ api: "anthropic-messages",
12593
+ provider: "vercel-ai-gateway",
12594
+ baseUrl: "https://ai-gateway.vercel.sh",
12595
+ reasoning: true,
12596
+ input: ["text", "image"],
12597
+ cost: {
12598
+ input: 30,
12599
+ output: 180,
12431
12600
  cacheRead: 0,
12432
12601
  cacheWrite: 0,
12433
12602
  },
12434
- contextWindow: 131072,
12435
- maxTokens: 131072,
12603
+ contextWindow: 1050000,
12604
+ maxTokens: 128000,
12436
12605
  },
12437
12606
  "openai/gpt-oss-20b": {
12438
12607
  id: "openai/gpt-oss-20b",
@@ -12621,40 +12790,6 @@ export const MODELS = {
12621
12790
  contextWindow: 131072,
12622
12791
  maxTokens: 131072,
12623
12792
  },
12624
- "vercel/v0-1.0-md": {
12625
- id: "vercel/v0-1.0-md",
12626
- name: "v0-1.0-md",
12627
- api: "anthropic-messages",
12628
- provider: "vercel-ai-gateway",
12629
- baseUrl: "https://ai-gateway.vercel.sh",
12630
- reasoning: false,
12631
- input: ["text", "image"],
12632
- cost: {
12633
- input: 3,
12634
- output: 15,
12635
- cacheRead: 0,
12636
- cacheWrite: 0,
12637
- },
12638
- contextWindow: 128000,
12639
- maxTokens: 32000,
12640
- },
12641
- "vercel/v0-1.5-md": {
12642
- id: "vercel/v0-1.5-md",
12643
- name: "v0-1.5-md",
12644
- api: "anthropic-messages",
12645
- provider: "vercel-ai-gateway",
12646
- baseUrl: "https://ai-gateway.vercel.sh",
12647
- reasoning: false,
12648
- input: ["text", "image"],
12649
- cost: {
12650
- input: 3,
12651
- output: 15,
12652
- cacheRead: 0,
12653
- cacheWrite: 0,
12654
- },
12655
- contextWindow: 128000,
12656
- maxTokens: 32768,
12657
- },
12658
12793
  "xai/grok-2-vision": {
12659
12794
  id: "xai/grok-2-vision",
12660
12795
  name: "Grok 2 Vision",
@@ -12683,7 +12818,7 @@ export const MODELS = {
12683
12818
  cost: {
12684
12819
  input: 3,
12685
12820
  output: 15,
12686
- cacheRead: 0,
12821
+ cacheRead: 0.75,
12687
12822
  cacheWrite: 0,
12688
12823
  },
12689
12824
  contextWindow: 131072,
@@ -12700,7 +12835,7 @@ export const MODELS = {
12700
12835
  cost: {
12701
12836
  input: 5,
12702
12837
  output: 25,
12703
- cacheRead: 0,
12838
+ cacheRead: 1.25,
12704
12839
  cacheWrite: 0,
12705
12840
  },
12706
12841
  contextWindow: 131072,
@@ -12717,7 +12852,7 @@ export const MODELS = {
12717
12852
  cost: {
12718
12853
  input: 0.3,
12719
12854
  output: 0.5,
12720
- cacheRead: 0,
12855
+ cacheRead: 0.075,
12721
12856
  cacheWrite: 0,
12722
12857
  },
12723
12858
  contextWindow: 131072,
@@ -12751,7 +12886,7 @@ export const MODELS = {
12751
12886
  cost: {
12752
12887
  input: 3,
12753
12888
  output: 15,
12754
- cacheRead: 0,
12889
+ cacheRead: 0.75,
12755
12890
  cacheWrite: 0,
12756
12891
  },
12757
12892
  contextWindow: 256000,
@@ -12825,6 +12960,57 @@ export const MODELS = {
12825
12960
  contextWindow: 2000000,
12826
12961
  maxTokens: 30000,
12827
12962
  },
12963
+ "xai/grok-4.20-multi-agent-beta": {
12964
+ id: "xai/grok-4.20-multi-agent-beta",
12965
+ name: "Grok 4.20 Multi Agent Beta",
12966
+ api: "anthropic-messages",
12967
+ provider: "vercel-ai-gateway",
12968
+ baseUrl: "https://ai-gateway.vercel.sh",
12969
+ reasoning: true,
12970
+ input: ["text"],
12971
+ cost: {
12972
+ input: 2,
12973
+ output: 6,
12974
+ cacheRead: 0.19999999999999998,
12975
+ cacheWrite: 0,
12976
+ },
12977
+ contextWindow: 2000000,
12978
+ maxTokens: 2000000,
12979
+ },
12980
+ "xai/grok-4.20-non-reasoning-beta": {
12981
+ id: "xai/grok-4.20-non-reasoning-beta",
12982
+ name: "Grok 4.20 Beta Non-Reasoning",
12983
+ api: "anthropic-messages",
12984
+ provider: "vercel-ai-gateway",
12985
+ baseUrl: "https://ai-gateway.vercel.sh",
12986
+ reasoning: false,
12987
+ input: ["text", "image"],
12988
+ cost: {
12989
+ input: 2,
12990
+ output: 6,
12991
+ cacheRead: 0.19999999999999998,
12992
+ cacheWrite: 0,
12993
+ },
12994
+ contextWindow: 2000000,
12995
+ maxTokens: 2000000,
12996
+ },
12997
+ "xai/grok-4.20-reasoning-beta": {
12998
+ id: "xai/grok-4.20-reasoning-beta",
12999
+ name: "Grok 4.20 Beta Reasoning",
13000
+ api: "anthropic-messages",
13001
+ provider: "vercel-ai-gateway",
13002
+ baseUrl: "https://ai-gateway.vercel.sh",
13003
+ reasoning: true,
13004
+ input: ["text", "image"],
13005
+ cost: {
13006
+ input: 2,
13007
+ output: 6,
13008
+ cacheRead: 0.19999999999999998,
13009
+ cacheWrite: 0,
13010
+ },
13011
+ contextWindow: 2000000,
13012
+ maxTokens: 2000000,
13013
+ },
12828
13014
  "xai/grok-code-fast-1": {
12829
13015
  id: "xai/grok-code-fast-1",
12830
13016
  name: "Grok Code Fast 1",
@@ -12851,17 +13037,34 @@ export const MODELS = {
12851
13037
  reasoning: true,
12852
13038
  input: ["text"],
12853
13039
  cost: {
12854
- input: 0.09,
12855
- output: 0.29,
12856
- cacheRead: 0,
13040
+ input: 0.09999999999999999,
13041
+ output: 0.3,
13042
+ cacheRead: 0.02,
12857
13043
  cacheWrite: 0,
12858
13044
  },
12859
13045
  contextWindow: 262144,
12860
13046
  maxTokens: 32000,
12861
13047
  },
12862
- "zai/glm-4.5": {
12863
- id: "zai/glm-4.5",
12864
- name: "GLM-4.5",
13048
+ "xiaomi/mimo-v2-pro": {
13049
+ id: "xiaomi/mimo-v2-pro",
13050
+ name: "MiMo V2 Pro",
13051
+ api: "anthropic-messages",
13052
+ provider: "vercel-ai-gateway",
13053
+ baseUrl: "https://ai-gateway.vercel.sh",
13054
+ reasoning: true,
13055
+ input: ["text"],
13056
+ cost: {
13057
+ input: 1,
13058
+ output: 3,
13059
+ cacheRead: 0.19999999999999998,
13060
+ cacheWrite: 0,
13061
+ },
13062
+ contextWindow: 1000000,
13063
+ maxTokens: 128000,
13064
+ },
13065
+ "zai/glm-4.5": {
13066
+ id: "zai/glm-4.5",
13067
+ name: "GLM-4.5",
12865
13068
  api: "anthropic-messages",
12866
13069
  provider: "vercel-ai-gateway",
12867
13070
  baseUrl: "https://ai-gateway.vercel.sh",
@@ -12870,11 +13073,11 @@ export const MODELS = {
12870
13073
  cost: {
12871
13074
  input: 0.6,
12872
13075
  output: 2.2,
12873
- cacheRead: 0,
13076
+ cacheRead: 0.11,
12874
13077
  cacheWrite: 0,
12875
13078
  },
12876
- contextWindow: 131072,
12877
- maxTokens: 131072,
13079
+ contextWindow: 128000,
13080
+ maxTokens: 96000,
12878
13081
  },
12879
13082
  "zai/glm-4.5-air": {
12880
13083
  id: "zai/glm-4.5-air",
@@ -12899,16 +13102,16 @@ export const MODELS = {
12899
13102
  api: "anthropic-messages",
12900
13103
  provider: "vercel-ai-gateway",
12901
13104
  baseUrl: "https://ai-gateway.vercel.sh",
12902
- reasoning: true,
13105
+ reasoning: false,
12903
13106
  input: ["text", "image"],
12904
13107
  cost: {
12905
13108
  input: 0.6,
12906
13109
  output: 1.7999999999999998,
12907
- cacheRead: 0,
13110
+ cacheRead: 0.11,
12908
13111
  cacheWrite: 0,
12909
13112
  },
12910
- contextWindow: 65536,
12911
- maxTokens: 16384,
13113
+ contextWindow: 66000,
13114
+ maxTokens: 16000,
12912
13115
  },
12913
13116
  "zai/glm-4.6": {
12914
13117
  id: "zai/glm-4.6",
@@ -12919,8 +13122,8 @@ export const MODELS = {
12919
13122
  reasoning: true,
12920
13123
  input: ["text"],
12921
13124
  cost: {
12922
- input: 0.44999999999999996,
12923
- output: 1.7999999999999998,
13125
+ input: 0.6,
13126
+ output: 2.2,
12924
13127
  cacheRead: 0.11,
12925
13128
  cacheWrite: 0,
12926
13129
  },
@@ -12970,14 +13173,31 @@ export const MODELS = {
12970
13173
  reasoning: true,
12971
13174
  input: ["text"],
12972
13175
  cost: {
12973
- input: 0.43,
12974
- output: 1.75,
12975
- cacheRead: 0.08,
13176
+ input: 0.6,
13177
+ output: 2.2,
13178
+ cacheRead: 0,
12976
13179
  cacheWrite: 0,
12977
13180
  },
12978
- contextWindow: 202752,
13181
+ contextWindow: 200000,
12979
13182
  maxTokens: 120000,
12980
13183
  },
13184
+ "zai/glm-4.7-flash": {
13185
+ id: "zai/glm-4.7-flash",
13186
+ name: "GLM 4.7 Flash",
13187
+ api: "anthropic-messages",
13188
+ provider: "vercel-ai-gateway",
13189
+ baseUrl: "https://ai-gateway.vercel.sh",
13190
+ reasoning: true,
13191
+ input: ["text"],
13192
+ cost: {
13193
+ input: 0.07,
13194
+ output: 0.39999999999999997,
13195
+ cacheRead: 0,
13196
+ cacheWrite: 0,
13197
+ },
13198
+ contextWindow: 200000,
13199
+ maxTokens: 131000,
13200
+ },
12981
13201
  "zai/glm-4.7-flashx": {
12982
13202
  id: "zai/glm-4.7-flashx",
12983
13203
  name: "GLM 4.7 FlashX",
@@ -12997,7 +13217,7 @@ export const MODELS = {
12997
13217
  },
12998
13218
  "zai/glm-5": {
12999
13219
  id: "zai/glm-5",
13000
- name: "GLM-5",
13220
+ name: "GLM 5",
13001
13221
  api: "anthropic-messages",
13002
13222
  provider: "vercel-ai-gateway",
13003
13223
  baseUrl: "https://ai-gateway.vercel.sh",
@@ -13010,7 +13230,24 @@ export const MODELS = {
13010
13230
  cacheWrite: 0,
13011
13231
  },
13012
13232
  contextWindow: 202800,
13013
- maxTokens: 131072,
13233
+ maxTokens: 131100,
13234
+ },
13235
+ "zai/glm-5-turbo": {
13236
+ id: "zai/glm-5-turbo",
13237
+ name: "GLM 5 Turbo",
13238
+ api: "anthropic-messages",
13239
+ provider: "vercel-ai-gateway",
13240
+ baseUrl: "https://ai-gateway.vercel.sh",
13241
+ reasoning: true,
13242
+ input: ["text"],
13243
+ cost: {
13244
+ input: 1.2,
13245
+ output: 4,
13246
+ cacheRead: 0.24,
13247
+ cacheWrite: 0,
13248
+ },
13249
+ contextWindow: 202800,
13250
+ maxTokens: 131100,
13014
13251
  },
13015
13252
  },
13016
13253
  "xai": {
@@ -13337,6 +13574,40 @@ export const MODELS = {
13337
13574
  contextWindow: 2000000,
13338
13575
  maxTokens: 30000,
13339
13576
  },
13577
+ "grok-4.20-0309-non-reasoning": {
13578
+ id: "grok-4.20-0309-non-reasoning",
13579
+ name: "Grok 4.20 (Non-Reasoning)",
13580
+ api: "openai-completions",
13581
+ provider: "xai",
13582
+ baseUrl: "https://api.x.ai/v1",
13583
+ reasoning: false,
13584
+ input: ["text", "image"],
13585
+ cost: {
13586
+ input: 2,
13587
+ output: 6,
13588
+ cacheRead: 0.2,
13589
+ cacheWrite: 0,
13590
+ },
13591
+ contextWindow: 2000000,
13592
+ maxTokens: 30000,
13593
+ },
13594
+ "grok-4.20-0309-reasoning": {
13595
+ id: "grok-4.20-0309-reasoning",
13596
+ name: "Grok 4.20 (Reasoning)",
13597
+ api: "openai-completions",
13598
+ provider: "xai",
13599
+ baseUrl: "https://api.x.ai/v1",
13600
+ reasoning: true,
13601
+ input: ["text", "image"],
13602
+ cost: {
13603
+ input: 2,
13604
+ output: 6,
13605
+ cacheRead: 0.2,
13606
+ cacheWrite: 0,
13607
+ },
13608
+ contextWindow: 2000000,
13609
+ maxTokens: 30000,
13610
+ },
13340
13611
  "grok-beta": {
13341
13612
  id: "grok-beta",
13342
13613
  name: "Grok Beta",
@@ -13552,748 +13823,24 @@ export const MODELS = {
13552
13823
  contextWindow: 204800,
13553
13824
  maxTokens: 131072,
13554
13825
  },
13555
- },
13556
- "alibaba-coding-plan": {
13557
- "qwen3.5-plus": {
13558
- id: "qwen3.5-plus",
13559
- name: "Qwen3.5 Plus",
13560
- api: "openai-completions",
13561
- provider: "alibaba-coding-plan",
13562
- baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
13563
- reasoning: true,
13564
- input: ["text"],
13565
- cost: {
13566
- input: 0,
13567
- output: 0,
13568
- cacheRead: 0,
13569
- cacheWrite: 0,
13570
- },
13571
- contextWindow: 983616,
13572
- maxTokens: 65536,
13573
- compat: { thinkingFormat: "qwen", supportsDeveloperRole: false },
13574
- },
13575
- "qwen3-max-2026-01-23": {
13576
- id: "qwen3-max-2026-01-23",
13577
- name: "Qwen3 Max 2026-01-23",
13578
- api: "openai-completions",
13579
- provider: "alibaba-coding-plan",
13580
- baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
13581
- reasoning: true,
13582
- input: ["text"],
13583
- cost: {
13584
- input: 0,
13585
- output: 0,
13586
- cacheRead: 0,
13587
- cacheWrite: 0,
13588
- },
13589
- contextWindow: 258048,
13590
- maxTokens: 32768,
13591
- compat: { thinkingFormat: "qwen", supportsDeveloperRole: false },
13592
- },
13593
- "qwen3-coder-next": {
13594
- id: "qwen3-coder-next",
13595
- name: "Qwen3 Coder Next",
13596
- api: "openai-completions",
13597
- provider: "alibaba-coding-plan",
13598
- baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
13599
- reasoning: false,
13600
- input: ["text"],
13601
- cost: {
13602
- input: 0,
13603
- output: 0,
13604
- cacheRead: 0,
13605
- cacheWrite: 0,
13606
- },
13607
- contextWindow: 204800,
13608
- maxTokens: 65536,
13609
- compat: { supportsDeveloperRole: false },
13610
- },
13611
- "qwen3-coder-plus": {
13612
- id: "qwen3-coder-plus",
13613
- name: "Qwen3 Coder Plus",
13614
- api: "openai-completions",
13615
- provider: "alibaba-coding-plan",
13616
- baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
13617
- reasoning: false,
13618
- input: ["text"],
13619
- cost: {
13620
- input: 0,
13621
- output: 0,
13622
- cacheRead: 0,
13623
- cacheWrite: 0,
13624
- },
13625
- contextWindow: 997952,
13626
- maxTokens: 65536,
13627
- compat: { supportsDeveloperRole: false },
13628
- },
13629
- "MiniMax-M2.5": {
13630
- id: "MiniMax-M2.5",
13631
- name: "MiniMax M2.5",
13632
- api: "openai-completions",
13633
- provider: "alibaba-coding-plan",
13634
- baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
13635
- reasoning: true,
13636
- input: ["text"],
13637
- cost: {
13638
- input: 0,
13639
- output: 0,
13640
- cacheRead: 0,
13641
- cacheWrite: 0,
13642
- },
13643
- contextWindow: 196608,
13644
- maxTokens: 65536,
13645
- compat: { supportsStore: false, supportsDeveloperRole: false, supportsReasoningEffort: true, maxTokensField: "max_tokens" },
13646
- },
13647
- "glm-5": {
13648
- id: "glm-5",
13649
- name: "GLM-5",
13650
- api: "openai-completions",
13651
- provider: "alibaba-coding-plan",
13652
- baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
13653
- reasoning: true,
13654
- input: ["text"],
13655
- cost: {
13656
- input: 0,
13657
- output: 0,
13658
- cacheRead: 0,
13659
- cacheWrite: 0,
13660
- },
13661
- contextWindow: 202752,
13662
- maxTokens: 16384,
13663
- compat: { thinkingFormat: "qwen", supportsDeveloperRole: false },
13664
- },
13665
- "glm-4.7": {
13666
- id: "glm-4.7",
13667
- name: "GLM-4.7",
13668
- api: "openai-completions",
13669
- provider: "alibaba-coding-plan",
13670
- baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
13671
- reasoning: true,
13672
- input: ["text"],
13673
- cost: {
13674
- input: 0,
13675
- output: 0,
13676
- cacheRead: 0,
13677
- cacheWrite: 0,
13678
- },
13679
- contextWindow: 169984,
13680
- maxTokens: 16384,
13681
- compat: { thinkingFormat: "qwen", supportsDeveloperRole: false },
13682
- },
13683
- "kimi-k2.5": {
13684
- id: "kimi-k2.5",
13685
- name: "Kimi K2.5",
13686
- api: "openai-completions",
13687
- provider: "alibaba-coding-plan",
13688
- baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
13689
- reasoning: true,
13690
- input: ["text"],
13691
- cost: {
13692
- input: 0,
13693
- output: 0,
13694
- cacheRead: 0,
13695
- cacheWrite: 0,
13696
- },
13697
- contextWindow: 258048,
13698
- maxTokens: 32768,
13699
- compat: { thinkingFormat: "zai", supportsDeveloperRole: false },
13700
- },
13701
- },
13702
- "ollama-cloud": {
13703
- "cogito-2.1:671b": {
13704
- id: "cogito-2.1:671b",
13705
- name: "Cogito 2.1 671B",
13706
- api: "openai-completions",
13707
- provider: "ollama-cloud",
13708
- baseUrl: "https://ollama.com/v1",
13709
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13710
- reasoning: true,
13711
- input: ["text"],
13712
- cost: {
13713
- input: 0,
13714
- output: 0,
13715
- cacheRead: 0,
13716
- cacheWrite: 0,
13717
- },
13718
- contextWindow: 163840,
13719
- maxTokens: 32000,
13720
- },
13721
- "deepseek-v3.1:671b": {
13722
- id: "deepseek-v3.1:671b",
13723
- name: "DeepSeek V3.1 671B",
13724
- api: "openai-completions",
13725
- provider: "ollama-cloud",
13726
- baseUrl: "https://ollama.com/v1",
13727
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13728
- reasoning: true,
13729
- input: ["text"],
13730
- cost: {
13731
- input: 0,
13732
- output: 0,
13733
- cacheRead: 0,
13734
- cacheWrite: 0,
13735
- },
13736
- contextWindow: 163840,
13737
- maxTokens: 163840,
13738
- },
13739
- "deepseek-v3.2": {
13740
- id: "deepseek-v3.2",
13741
- name: "DeepSeek V3.2",
13826
+ "glm-5-turbo": {
13827
+ id: "glm-5-turbo",
13828
+ name: "GLM-5-Turbo",
13742
13829
  api: "openai-completions",
13743
- provider: "ollama-cloud",
13744
- baseUrl: "https://ollama.com/v1",
13745
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13830
+ provider: "zai",
13831
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
13832
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
13746
13833
  reasoning: true,
13747
13834
  input: ["text"],
13748
13835
  cost: {
13749
- input: 0,
13750
- output: 0,
13751
- cacheRead: 0,
13752
- cacheWrite: 0,
13753
- },
13754
- contextWindow: 163840,
13755
- maxTokens: 65536,
13756
- },
13757
- "devstral-2:123b": {
13758
- id: "devstral-2:123b",
13759
- name: "Devstral 2 123B",
13760
- api: "openai-completions",
13761
- provider: "ollama-cloud",
13762
- baseUrl: "https://ollama.com/v1",
13763
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13764
- reasoning: false,
13765
- input: ["text"],
13766
- cost: {
13767
- input: 0,
13768
- output: 0,
13769
- cacheRead: 0,
13836
+ input: 1.2,
13837
+ output: 4,
13838
+ cacheRead: 0.24,
13770
13839
  cacheWrite: 0,
13771
13840
  },
13772
- contextWindow: 262144,
13773
- maxTokens: 262144,
13774
- },
13775
- "devstral-small-2:24b": {
13776
- id: "devstral-small-2:24b",
13777
- name: "Devstral Small 2 24B",
13778
- api: "openai-completions",
13779
- provider: "ollama-cloud",
13780
- baseUrl: "https://ollama.com/v1",
13781
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13782
- reasoning: false,
13783
- input: ["text", "image"],
13784
- cost: {
13785
- input: 0,
13786
- output: 0,
13787
- cacheRead: 0,
13788
- cacheWrite: 0,
13789
- },
13790
- contextWindow: 262144,
13791
- maxTokens: 262144,
13792
- },
13793
- "gemini-3-flash-preview": {
13794
- id: "gemini-3-flash-preview",
13795
- name: "Gemini 3 Flash Preview",
13796
- api: "openai-completions",
13797
- provider: "ollama-cloud",
13798
- baseUrl: "https://ollama.com/v1",
13799
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13800
- reasoning: true,
13801
- input: ["text"],
13802
- cost: {
13803
- input: 0,
13804
- output: 0,
13805
- cacheRead: 0,
13806
- cacheWrite: 0,
13807
- },
13808
- contextWindow: 1048576,
13809
- maxTokens: 65536,
13810
- },
13811
- "gemma3:12b": {
13812
- id: "gemma3:12b",
13813
- name: "Gemma 3 12B",
13814
- api: "openai-completions",
13815
- provider: "ollama-cloud",
13816
- baseUrl: "https://ollama.com/v1",
13817
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13818
- reasoning: false,
13819
- input: ["text", "image"],
13820
- cost: {
13821
- input: 0,
13822
- output: 0,
13823
- cacheRead: 0,
13824
- cacheWrite: 0,
13825
- },
13826
- contextWindow: 131072,
13827
- maxTokens: 131072,
13828
- },
13829
- "gemma3:27b": {
13830
- id: "gemma3:27b",
13831
- name: "Gemma 3 27B",
13832
- api: "openai-completions",
13833
- provider: "ollama-cloud",
13834
- baseUrl: "https://ollama.com/v1",
13835
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13836
- reasoning: false,
13837
- input: ["text", "image"],
13838
- cost: {
13839
- input: 0,
13840
- output: 0,
13841
- cacheRead: 0,
13842
- cacheWrite: 0,
13843
- },
13844
- contextWindow: 131072,
13845
- maxTokens: 131072,
13846
- },
13847
- "gemma3:4b": {
13848
- id: "gemma3:4b",
13849
- name: "Gemma 3 4B",
13850
- api: "openai-completions",
13851
- provider: "ollama-cloud",
13852
- baseUrl: "https://ollama.com/v1",
13853
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13854
- reasoning: false,
13855
- input: ["text", "image"],
13856
- cost: {
13857
- input: 0,
13858
- output: 0,
13859
- cacheRead: 0,
13860
- cacheWrite: 0,
13861
- },
13862
- contextWindow: 131072,
13863
- maxTokens: 131072,
13864
- },
13865
- "glm-4.6": {
13866
- id: "glm-4.6",
13867
- name: "GLM 4.6",
13868
- api: "openai-completions",
13869
- provider: "ollama-cloud",
13870
- baseUrl: "https://ollama.com/v1",
13871
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13872
- reasoning: true,
13873
- input: ["text"],
13874
- cost: {
13875
- input: 0,
13876
- output: 0,
13877
- cacheRead: 0,
13878
- cacheWrite: 0,
13879
- },
13880
- contextWindow: 202752,
13881
- maxTokens: 131072,
13882
- },
13883
- "glm-4.7": {
13884
- id: "glm-4.7",
13885
- name: "GLM 4.7",
13886
- api: "openai-completions",
13887
- provider: "ollama-cloud",
13888
- baseUrl: "https://ollama.com/v1",
13889
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13890
- reasoning: true,
13891
- input: ["text"],
13892
- cost: {
13893
- input: 0,
13894
- output: 0,
13895
- cacheRead: 0,
13896
- cacheWrite: 0,
13897
- },
13898
- contextWindow: 202752,
13899
- maxTokens: 131072,
13900
- },
13901
- "glm-5": {
13902
- id: "glm-5",
13903
- name: "GLM 5",
13904
- api: "openai-completions",
13905
- provider: "ollama-cloud",
13906
- baseUrl: "https://ollama.com/v1",
13907
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13908
- reasoning: true,
13909
- input: ["text"],
13910
- cost: {
13911
- input: 0,
13912
- output: 0,
13913
- cacheRead: 0,
13914
- cacheWrite: 0,
13915
- },
13916
- contextWindow: 202752,
13917
- maxTokens: 131072,
13918
- },
13919
- "gpt-oss:120b": {
13920
- id: "gpt-oss:120b",
13921
- name: "GPT-OSS 120B",
13922
- api: "openai-completions",
13923
- provider: "ollama-cloud",
13924
- baseUrl: "https://ollama.com/v1",
13925
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13926
- reasoning: true,
13927
- input: ["text"],
13928
- cost: {
13929
- input: 0,
13930
- output: 0,
13931
- cacheRead: 0,
13932
- cacheWrite: 0,
13933
- },
13934
- contextWindow: 131072,
13935
- maxTokens: 32768,
13936
- },
13937
- "gpt-oss:20b": {
13938
- id: "gpt-oss:20b",
13939
- name: "GPT-OSS 20B",
13940
- api: "openai-completions",
13941
- provider: "ollama-cloud",
13942
- baseUrl: "https://ollama.com/v1",
13943
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13944
- reasoning: true,
13945
- input: ["text"],
13946
- cost: {
13947
- input: 0,
13948
- output: 0,
13949
- cacheRead: 0,
13950
- cacheWrite: 0,
13951
- },
13952
- contextWindow: 131072,
13953
- maxTokens: 32768,
13954
- },
13955
- "kimi-k2:1t": {
13956
- id: "kimi-k2:1t",
13957
- name: "Kimi K2 1T",
13958
- api: "openai-completions",
13959
- provider: "ollama-cloud",
13960
- baseUrl: "https://ollama.com/v1",
13961
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13962
- reasoning: false,
13963
- input: ["text"],
13964
- cost: {
13965
- input: 0,
13966
- output: 0,
13967
- cacheRead: 0,
13968
- cacheWrite: 0,
13969
- },
13970
- contextWindow: 262144,
13971
- maxTokens: 262144,
13972
- },
13973
- "kimi-k2.5": {
13974
- id: "kimi-k2.5",
13975
- name: "Kimi K2.5",
13976
- api: "openai-completions",
13977
- provider: "ollama-cloud",
13978
- baseUrl: "https://ollama.com/v1",
13979
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13980
- reasoning: true,
13981
- input: ["text", "image"],
13982
- cost: {
13983
- input: 0,
13984
- output: 0,
13985
- cacheRead: 0,
13986
- cacheWrite: 0,
13987
- },
13988
- contextWindow: 262144,
13989
- maxTokens: 262144,
13990
- },
13991
- "kimi-k2-thinking": {
13992
- id: "kimi-k2-thinking",
13993
- name: "Kimi K2 Thinking",
13994
- api: "openai-completions",
13995
- provider: "ollama-cloud",
13996
- baseUrl: "https://ollama.com/v1",
13997
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
13998
- reasoning: true,
13999
- input: ["text"],
14000
- cost: {
14001
- input: 0,
14002
- output: 0,
14003
- cacheRead: 0,
14004
- cacheWrite: 0,
14005
- },
14006
- contextWindow: 262144,
14007
- maxTokens: 262144,
14008
- },
14009
- "minimax-m2.1": {
14010
- id: "minimax-m2.1",
14011
- name: "Minimax M2.1",
14012
- api: "openai-completions",
14013
- provider: "ollama-cloud",
14014
- baseUrl: "https://ollama.com/v1",
14015
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14016
- reasoning: true,
14017
- input: ["text"],
14018
- cost: {
14019
- input: 0,
14020
- output: 0,
14021
- cacheRead: 0,
14022
- cacheWrite: 0,
14023
- },
14024
- contextWindow: 204800,
14025
- maxTokens: 131072,
14026
- },
14027
- "minimax-m2.5": {
14028
- id: "minimax-m2.5",
14029
- name: "Minimax M2.5",
14030
- api: "openai-completions",
14031
- provider: "ollama-cloud",
14032
- baseUrl: "https://ollama.com/v1",
14033
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14034
- reasoning: true,
14035
- input: ["text"],
14036
- cost: {
14037
- input: 0,
14038
- output: 0,
14039
- cacheRead: 0,
14040
- cacheWrite: 0,
14041
- },
14042
- contextWindow: 204800,
14043
- maxTokens: 131072,
14044
- },
14045
- "minimax-m2": {
14046
- id: "minimax-m2",
14047
- name: "Minimax M2",
14048
- api: "openai-completions",
14049
- provider: "ollama-cloud",
14050
- baseUrl: "https://ollama.com/v1",
14051
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14052
- reasoning: false,
14053
- input: ["text"],
14054
- cost: {
14055
- input: 0,
14056
- output: 0,
14057
- cacheRead: 0,
14058
- cacheWrite: 0,
14059
- },
14060
- contextWindow: 204800,
14061
- maxTokens: 128000,
14062
- },
14063
- "ministral-3:14b": {
14064
- id: "ministral-3:14b",
14065
- name: "Ministral 3 14B",
14066
- api: "openai-completions",
14067
- provider: "ollama-cloud",
14068
- baseUrl: "https://ollama.com/v1",
14069
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14070
- reasoning: false,
14071
- input: ["text", "image"],
14072
- cost: {
14073
- input: 0,
14074
- output: 0,
14075
- cacheRead: 0,
14076
- cacheWrite: 0,
14077
- },
14078
- contextWindow: 262144,
14079
- maxTokens: 128000,
14080
- },
14081
- "ministral-3:3b": {
14082
- id: "ministral-3:3b",
14083
- name: "Ministral 3 3B",
14084
- api: "openai-completions",
14085
- provider: "ollama-cloud",
14086
- baseUrl: "https://ollama.com/v1",
14087
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14088
- reasoning: false,
14089
- input: ["text", "image"],
14090
- cost: {
14091
- input: 0,
14092
- output: 0,
14093
- cacheRead: 0,
14094
- cacheWrite: 0,
14095
- },
14096
- contextWindow: 262144,
14097
- maxTokens: 128000,
14098
- },
14099
- "ministral-3:8b": {
14100
- id: "ministral-3:8b",
14101
- name: "Ministral 3 8B",
14102
- api: "openai-completions",
14103
- provider: "ollama-cloud",
14104
- baseUrl: "https://ollama.com/v1",
14105
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14106
- reasoning: false,
14107
- input: ["text", "image"],
14108
- cost: {
14109
- input: 0,
14110
- output: 0,
14111
- cacheRead: 0,
14112
- cacheWrite: 0,
14113
- },
14114
- contextWindow: 262144,
14115
- maxTokens: 128000,
14116
- },
14117
- "mistral-large-3:675b": {
14118
- id: "mistral-large-3:675b",
14119
- name: "Mistral Large 3 675B",
14120
- api: "openai-completions",
14121
- provider: "ollama-cloud",
14122
- baseUrl: "https://ollama.com/v1",
14123
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14124
- reasoning: false,
14125
- input: ["text", "image"],
14126
- cost: {
14127
- input: 0,
14128
- output: 0,
14129
- cacheRead: 0,
14130
- cacheWrite: 0,
14131
- },
14132
- contextWindow: 262144,
14133
- maxTokens: 262144,
14134
- },
14135
- "nemotron-3-nano:30b": {
14136
- id: "nemotron-3-nano:30b",
14137
- name: "Nemotron 3 Nano 30B",
14138
- api: "openai-completions",
14139
- provider: "ollama-cloud",
14140
- baseUrl: "https://ollama.com/v1",
14141
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14142
- reasoning: true,
14143
- input: ["text"],
14144
- cost: {
14145
- input: 0,
14146
- output: 0,
14147
- cacheRead: 0,
14148
- cacheWrite: 0,
14149
- },
14150
- contextWindow: 1048576,
14151
- maxTokens: 131072,
14152
- },
14153
- "nemotron-3-super": {
14154
- id: "nemotron-3-super",
14155
- name: "Nemotron 3 Super",
14156
- api: "openai-completions",
14157
- provider: "ollama-cloud",
14158
- baseUrl: "https://ollama.com/v1",
14159
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14160
- reasoning: true,
14161
- input: ["text"],
14162
- cost: {
14163
- input: 0,
14164
- output: 0,
14165
- cacheRead: 0,
14166
- cacheWrite: 0,
14167
- },
14168
- contextWindow: 262144,
14169
- maxTokens: 65536,
14170
- },
14171
- "qwen3.5:397b": {
14172
- id: "qwen3.5:397b",
14173
- name: "Qwen 3.5 397B",
14174
- api: "openai-completions",
14175
- provider: "ollama-cloud",
14176
- baseUrl: "https://ollama.com/v1",
14177
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14178
- reasoning: true,
14179
- input: ["text", "image"],
14180
- cost: {
14181
- input: 0,
14182
- output: 0,
14183
- cacheRead: 0,
14184
- cacheWrite: 0,
14185
- },
14186
- contextWindow: 262144,
14187
- maxTokens: 81920,
14188
- },
14189
- "qwen3-coder:480b": {
14190
- id: "qwen3-coder:480b",
14191
- name: "Qwen 3 Coder 480B",
14192
- api: "openai-completions",
14193
- provider: "ollama-cloud",
14194
- baseUrl: "https://ollama.com/v1",
14195
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14196
- reasoning: false,
14197
- input: ["text"],
14198
- cost: {
14199
- input: 0,
14200
- output: 0,
14201
- cacheRead: 0,
14202
- cacheWrite: 0,
14203
- },
14204
- contextWindow: 262144,
14205
- maxTokens: 65536,
14206
- },
14207
- "qwen3-coder-next": {
14208
- id: "qwen3-coder-next",
14209
- name: "Qwen 3 Coder Next",
14210
- api: "openai-completions",
14211
- provider: "ollama-cloud",
14212
- baseUrl: "https://ollama.com/v1",
14213
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14214
- reasoning: false,
14215
- input: ["text"],
14216
- cost: {
14217
- input: 0,
14218
- output: 0,
14219
- cacheRead: 0,
14220
- cacheWrite: 0,
14221
- },
14222
- contextWindow: 262144,
14223
- maxTokens: 65536,
14224
- },
14225
- "qwen3-next:80b": {
14226
- id: "qwen3-next:80b",
14227
- name: "Qwen 3 Next 80B",
14228
- api: "openai-completions",
14229
- provider: "ollama-cloud",
14230
- baseUrl: "https://ollama.com/v1",
14231
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14232
- reasoning: true,
14233
- input: ["text"],
14234
- cost: {
14235
- input: 0,
14236
- output: 0,
14237
- cacheRead: 0,
14238
- cacheWrite: 0,
14239
- },
14240
- contextWindow: 262144,
14241
- maxTokens: 32768,
14242
- },
14243
- "qwen3-vl:235b-instruct": {
14244
- id: "qwen3-vl:235b-instruct",
14245
- name: "Qwen 3 VL 235B Instruct",
14246
- api: "openai-completions",
14247
- provider: "ollama-cloud",
14248
- baseUrl: "https://ollama.com/v1",
14249
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14250
- reasoning: false,
14251
- input: ["text", "image"],
14252
- cost: {
14253
- input: 0,
14254
- output: 0,
14255
- cacheRead: 0,
14256
- cacheWrite: 0,
14257
- },
14258
- contextWindow: 262144,
13841
+ contextWindow: 200000,
14259
13842
  maxTokens: 131072,
14260
13843
  },
14261
- "qwen3-vl:235b": {
14262
- id: "qwen3-vl:235b",
14263
- name: "Qwen 3 VL 235B",
14264
- api: "openai-completions",
14265
- provider: "ollama-cloud",
14266
- baseUrl: "https://ollama.com/v1",
14267
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14268
- reasoning: true,
14269
- input: ["text", "image"],
14270
- cost: {
14271
- input: 0,
14272
- output: 0,
14273
- cacheRead: 0,
14274
- cacheWrite: 0,
14275
- },
14276
- contextWindow: 262144,
14277
- maxTokens: 32768,
14278
- },
14279
- "rnj-1:8b": {
14280
- id: "rnj-1:8b",
14281
- name: "RNJ 1 8B",
14282
- api: "openai-completions",
14283
- provider: "ollama-cloud",
14284
- baseUrl: "https://ollama.com/v1",
14285
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
14286
- reasoning: false,
14287
- input: ["text"],
14288
- cost: {
14289
- input: 0,
14290
- output: 0,
14291
- cacheRead: 0,
14292
- cacheWrite: 0,
14293
- },
14294
- contextWindow: 32768,
14295
- maxTokens: 4096,
14296
- },
14297
13844
  },
14298
13845
  };
14299
13846
  //# sourceMappingURL=models.generated.js.map