@mariozechner/pi-ai 0.58.3 → 0.59.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (41) hide show
  1. package/README.md +5 -0
  2. package/dist/index.d.ts +9 -8
  3. package/dist/index.d.ts.map +1 -1
  4. package/dist/index.js +0 -8
  5. package/dist/index.js.map +1 -1
  6. package/dist/models.generated.d.ts +92 -139
  7. package/dist/models.generated.d.ts.map +1 -1
  8. package/dist/models.generated.js +135 -185
  9. package/dist/models.generated.js.map +1 -1
  10. package/dist/providers/anthropic.d.ts +7 -0
  11. package/dist/providers/anthropic.d.ts.map +1 -1
  12. package/dist/providers/anthropic.js +22 -11
  13. package/dist/providers/anthropic.js.map +1 -1
  14. package/dist/providers/google-gemini-cli.d.ts.map +1 -1
  15. package/dist/providers/google-gemini-cli.js +3 -0
  16. package/dist/providers/google-gemini-cli.js.map +1 -1
  17. package/dist/providers/google-vertex.d.ts.map +1 -1
  18. package/dist/providers/google-vertex.js +3 -0
  19. package/dist/providers/google-vertex.js.map +1 -1
  20. package/dist/providers/google.d.ts.map +1 -1
  21. package/dist/providers/google.js +3 -0
  22. package/dist/providers/google.js.map +1 -1
  23. package/dist/providers/mistral.d.ts.map +1 -1
  24. package/dist/providers/mistral.js +3 -0
  25. package/dist/providers/mistral.js.map +1 -1
  26. package/dist/providers/openai-completions.d.ts.map +1 -1
  27. package/dist/providers/openai-completions.js +3 -0
  28. package/dist/providers/openai-completions.js.map +1 -1
  29. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  30. package/dist/providers/openai-responses-shared.js +7 -1
  31. package/dist/providers/openai-responses-shared.js.map +1 -1
  32. package/dist/providers/register-builtins.d.ts +28 -1
  33. package/dist/providers/register-builtins.d.ts.map +1 -1
  34. package/dist/providers/register-builtins.js +178 -45
  35. package/dist/providers/register-builtins.js.map +1 -1
  36. package/dist/types.d.ts +1 -0
  37. package/dist/types.d.ts.map +1 -1
  38. package/dist/types.js.map +1 -1
  39. package/package.json +39 -5
  40. package/bedrock-provider.d.ts +0 -1
  41. package/bedrock-provider.js +0 -1
@@ -2577,7 +2577,7 @@ export const MODELS = {
2577
2577
  cacheRead: 0,
2578
2578
  cacheWrite: 0,
2579
2579
  },
2580
- contextWindow: 128000,
2580
+ contextWindow: 1000000,
2581
2581
  maxTokens: 64000,
2582
2582
  },
2583
2583
  "claude-sonnet-4": {
@@ -2631,7 +2631,7 @@ export const MODELS = {
2631
2631
  cacheRead: 0,
2632
2632
  cacheWrite: 0,
2633
2633
  },
2634
- contextWindow: 128000,
2634
+ contextWindow: 1000000,
2635
2635
  maxTokens: 32000,
2636
2636
  },
2637
2637
  "gemini-2.5-pro": {
@@ -5966,23 +5966,6 @@ export const MODELS = {
5966
5966
  contextWindow: 1048576,
5967
5967
  maxTokens: 65536,
5968
5968
  },
5969
- "gemini-3-pro": {
5970
- id: "gemini-3-pro",
5971
- name: "Gemini 3 Pro",
5972
- api: "google-generative-ai",
5973
- provider: "opencode",
5974
- baseUrl: "https://opencode.ai/zen/v1",
5975
- reasoning: true,
5976
- input: ["text", "image"],
5977
- cost: {
5978
- input: 2,
5979
- output: 12,
5980
- cacheRead: 0.2,
5981
- cacheWrite: 0,
5982
- },
5983
- contextWindow: 1048576,
5984
- maxTokens: 65536,
5985
- },
5986
5969
  "gemini-3.1-pro": {
5987
5970
  id: "gemini-3.1-pro",
5988
5971
  name: "Gemini 3.1 Pro Preview",
@@ -6000,40 +5983,6 @@ export const MODELS = {
6000
5983
  contextWindow: 1048576,
6001
5984
  maxTokens: 65536,
6002
5985
  },
6003
- "glm-4.6": {
6004
- id: "glm-4.6",
6005
- name: "GLM-4.6",
6006
- api: "openai-completions",
6007
- provider: "opencode",
6008
- baseUrl: "https://opencode.ai/zen/v1",
6009
- reasoning: true,
6010
- input: ["text"],
6011
- cost: {
6012
- input: 0.6,
6013
- output: 2.2,
6014
- cacheRead: 0.1,
6015
- cacheWrite: 0,
6016
- },
6017
- contextWindow: 204800,
6018
- maxTokens: 131072,
6019
- },
6020
- "glm-4.7": {
6021
- id: "glm-4.7",
6022
- name: "GLM-4.7",
6023
- api: "openai-completions",
6024
- provider: "opencode",
6025
- baseUrl: "https://opencode.ai/zen/v1",
6026
- reasoning: true,
6027
- input: ["text"],
6028
- cost: {
6029
- input: 0.6,
6030
- output: 2.2,
6031
- cacheRead: 0.1,
6032
- cacheWrite: 0,
6033
- },
6034
- contextWindow: 204800,
6035
- maxTokens: 131072,
6036
- },
6037
5986
  "glm-5": {
6038
5987
  id: "glm-5",
6039
5988
  name: "GLM-5",
@@ -6289,23 +6238,6 @@ export const MODELS = {
6289
6238
  contextWindow: 262144,
6290
6239
  maxTokens: 65536,
6291
6240
  },
6292
- "minimax-m2.1": {
6293
- id: "minimax-m2.1",
6294
- name: "MiniMax M2.1",
6295
- api: "openai-completions",
6296
- provider: "opencode",
6297
- baseUrl: "https://opencode.ai/zen/v1",
6298
- reasoning: true,
6299
- input: ["text"],
6300
- cost: {
6301
- input: 0.3,
6302
- output: 1.2,
6303
- cacheRead: 0.1,
6304
- cacheWrite: 0,
6305
- },
6306
- contextWindow: 204800,
6307
- maxTokens: 131072,
6308
- },
6309
6241
  "minimax-m2.5": {
6310
6242
  id: "minimax-m2.5",
6311
6243
  name: "MiniMax M2.5",
@@ -7662,7 +7594,7 @@ export const MODELS = {
7662
7594
  reasoning: true,
7663
7595
  input: ["text"],
7664
7596
  cost: {
7665
- input: 0.25,
7597
+ input: 0.19999999999999998,
7666
7598
  output: 1.2,
7667
7599
  cacheRead: 0,
7668
7600
  cacheWrite: 0,
@@ -7959,6 +7891,23 @@ export const MODELS = {
7959
7891
  contextWindow: 32768,
7960
7892
  maxTokens: 16384,
7961
7893
  },
7894
+ "mistralai/mistral-small-2603": {
7895
+ id: "mistralai/mistral-small-2603",
7896
+ name: "Mistral: Mistral Small 4",
7897
+ api: "openai-completions",
7898
+ provider: "openrouter",
7899
+ baseUrl: "https://openrouter.ai/api/v1",
7900
+ reasoning: true,
7901
+ input: ["text", "image"],
7902
+ cost: {
7903
+ input: 0.15,
7904
+ output: 0.6,
7905
+ cacheRead: 0,
7906
+ cacheWrite: 0,
7907
+ },
7908
+ contextWindow: 262144,
7909
+ maxTokens: 4096,
7910
+ },
7962
7911
  "mistralai/mistral-small-3.1-24b-instruct:free": {
7963
7912
  id: "mistralai/mistral-small-3.1-24b-instruct:free",
7964
7913
  name: "Mistral: Mistral Small 3.1 24B (free)",
@@ -8945,6 +8894,40 @@ export const MODELS = {
8945
8894
  contextWindow: 1050000,
8946
8895
  maxTokens: 128000,
8947
8896
  },
8897
+ "openai/gpt-5.4-mini": {
8898
+ id: "openai/gpt-5.4-mini",
8899
+ name: "OpenAI: GPT-5.4 Mini",
8900
+ api: "openai-completions",
8901
+ provider: "openrouter",
8902
+ baseUrl: "https://openrouter.ai/api/v1",
8903
+ reasoning: true,
8904
+ input: ["text", "image"],
8905
+ cost: {
8906
+ input: 0.75,
8907
+ output: 4.5,
8908
+ cacheRead: 0.075,
8909
+ cacheWrite: 0,
8910
+ },
8911
+ contextWindow: 400000,
8912
+ maxTokens: 128000,
8913
+ },
8914
+ "openai/gpt-5.4-nano": {
8915
+ id: "openai/gpt-5.4-nano",
8916
+ name: "OpenAI: GPT-5.4 Nano",
8917
+ api: "openai-completions",
8918
+ provider: "openrouter",
8919
+ baseUrl: "https://openrouter.ai/api/v1",
8920
+ reasoning: true,
8921
+ input: ["text", "image"],
8922
+ cost: {
8923
+ input: 0.19999999999999998,
8924
+ output: 1.25,
8925
+ cacheRead: 0.02,
8926
+ cacheWrite: 0,
8927
+ },
8928
+ contextWindow: 400000,
8929
+ maxTokens: 128000,
8930
+ },
8948
8931
  "openai/gpt-5.4-pro": {
8949
8932
  id: "openai/gpt-5.4-pro",
8950
8933
  name: "OpenAI: GPT-5.4 Pro",
@@ -9053,7 +9036,7 @@ export const MODELS = {
9053
9036
  api: "openai-completions",
9054
9037
  provider: "openrouter",
9055
9038
  baseUrl: "https://openrouter.ai/api/v1",
9056
- reasoning: false,
9039
+ reasoning: true,
9057
9040
  input: ["text", "image"],
9058
9041
  cost: {
9059
9042
  input: 15,
@@ -9104,7 +9087,7 @@ export const MODELS = {
9104
9087
  api: "openai-completions",
9105
9088
  provider: "openrouter",
9106
9089
  baseUrl: "https://openrouter.ai/api/v1",
9107
- reasoning: false,
9090
+ reasoning: true,
9108
9091
  input: ["text"],
9109
9092
  cost: {
9110
9093
  input: 1.1,
@@ -9121,7 +9104,7 @@ export const MODELS = {
9121
9104
  api: "openai-completions",
9122
9105
  provider: "openrouter",
9123
9106
  baseUrl: "https://openrouter.ai/api/v1",
9124
- reasoning: false,
9107
+ reasoning: true,
9125
9108
  input: ["text"],
9126
9109
  cost: {
9127
9110
  input: 1.1,
@@ -9345,9 +9328,9 @@ export const MODELS = {
9345
9328
  reasoning: false,
9346
9329
  input: ["text"],
9347
9330
  cost: {
9348
- input: 0.39999999999999997,
9349
- output: 1.2,
9350
- cacheRead: 0.08,
9331
+ input: 0.26,
9332
+ output: 0.78,
9333
+ cacheRead: 0.052000000000000005,
9351
9334
  cacheWrite: 0,
9352
9335
  },
9353
9336
  contextWindow: 1000000,
@@ -9413,8 +9396,8 @@ export const MODELS = {
9413
9396
  reasoning: false,
9414
9397
  input: ["text", "image"],
9415
9398
  cost: {
9416
- input: 0.7999999999999999,
9417
- output: 3.1999999999999997,
9399
+ input: 0.52,
9400
+ output: 2.08,
9418
9401
  cacheRead: 0,
9419
9402
  cacheWrite: 0,
9420
9403
  },
@@ -9702,9 +9685,9 @@ export const MODELS = {
9702
9685
  reasoning: false,
9703
9686
  input: ["text"],
9704
9687
  cost: {
9705
- input: 1.2,
9706
- output: 6,
9707
- cacheRead: 0.24,
9688
+ input: 0.78,
9689
+ output: 3.9,
9690
+ cacheRead: 0.156,
9708
9691
  cacheWrite: 0,
9709
9692
  },
9710
9693
  contextWindow: 262144,
@@ -9991,8 +9974,8 @@ export const MODELS = {
9991
9974
  reasoning: true,
9992
9975
  input: ["text", "image"],
9993
9976
  cost: {
9994
- input: 0.09999999999999999,
9995
- output: 0.39999999999999997,
9977
+ input: 0.065,
9978
+ output: 0.26,
9996
9979
  cacheRead: 0,
9997
9980
  cacheWrite: 0,
9998
9981
  },
@@ -10554,7 +10537,7 @@ export const MODELS = {
10554
10537
  reasoning: true,
10555
10538
  input: ["text"],
10556
10539
  cost: {
10557
- input: 0.06,
10540
+ input: 0.12,
10558
10541
  output: 0.24,
10559
10542
  cacheRead: 0,
10560
10543
  cacheWrite: 0,
@@ -10607,7 +10590,7 @@ export const MODELS = {
10607
10590
  cost: {
10608
10591
  input: 0.29,
10609
10592
  output: 0.59,
10610
- cacheRead: 0,
10593
+ cacheRead: 0.145,
10611
10594
  cacheWrite: 0,
10612
10595
  },
10613
10596
  contextWindow: 131072,
@@ -10622,9 +10605,9 @@ export const MODELS = {
10622
10605
  reasoning: true,
10623
10606
  input: ["text", "image"],
10624
10607
  cost: {
10625
- input: 0.3,
10626
- output: 2.9000000000000004,
10627
- cacheRead: 0,
10608
+ input: 0.22999999999999998,
10609
+ output: 2.3,
10610
+ cacheRead: 0.19999999999999998,
10628
10611
  cacheWrite: 0,
10629
10612
  },
10630
10613
  contextWindow: 262114,
@@ -10641,7 +10624,7 @@ export const MODELS = {
10641
10624
  cost: {
10642
10625
  input: 0.39999999999999997,
10643
10626
  output: 1.5999999999999999,
10644
- cacheRead: 0,
10627
+ cacheRead: 0.022,
10645
10628
  cacheWrite: 0,
10646
10629
  },
10647
10630
  contextWindow: 262144,
@@ -11134,7 +11117,7 @@ export const MODELS = {
11134
11117
  cost: {
11135
11118
  input: 0.27,
11136
11119
  output: 1,
11137
- cacheRead: 0,
11120
+ cacheRead: 0.135,
11138
11121
  cacheWrite: 0,
11139
11122
  },
11140
11123
  contextWindow: 131072,
@@ -11149,9 +11132,9 @@ export const MODELS = {
11149
11132
  reasoning: false,
11150
11133
  input: ["text"],
11151
11134
  cost: {
11152
- input: 0.26,
11153
- output: 0.38,
11154
- cacheRead: 0.13,
11135
+ input: 0.28,
11136
+ output: 0.42,
11137
+ cacheRead: 0.028,
11155
11138
  cacheWrite: 0,
11156
11139
  },
11157
11140
  contextWindow: 128000,
@@ -11185,7 +11168,7 @@ export const MODELS = {
11185
11168
  cost: {
11186
11169
  input: 0.15,
11187
11170
  output: 0.6,
11188
- cacheRead: 0,
11171
+ cacheRead: 0.024999999999999998,
11189
11172
  cacheWrite: 0,
11190
11173
  },
11191
11174
  contextWindow: 1048576,
@@ -11202,7 +11185,7 @@ export const MODELS = {
11202
11185
  cost: {
11203
11186
  input: 0.075,
11204
11187
  output: 0.3,
11205
- cacheRead: 0,
11188
+ cacheRead: 0.02,
11206
11189
  cacheWrite: 0,
11207
11190
  },
11208
11191
  contextWindow: 1048576,
@@ -11242,40 +11225,6 @@ export const MODELS = {
11242
11225
  contextWindow: 1048576,
11243
11226
  maxTokens: 65536,
11244
11227
  },
11245
- "google/gemini-2.5-flash-lite-preview-09-2025": {
11246
- id: "google/gemini-2.5-flash-lite-preview-09-2025",
11247
- name: "Gemini 2.5 Flash Lite Preview 09-2025",
11248
- api: "anthropic-messages",
11249
- provider: "vercel-ai-gateway",
11250
- baseUrl: "https://ai-gateway.vercel.sh",
11251
- reasoning: true,
11252
- input: ["text", "image"],
11253
- cost: {
11254
- input: 0.09999999999999999,
11255
- output: 0.39999999999999997,
11256
- cacheRead: 0.01,
11257
- cacheWrite: 0,
11258
- },
11259
- contextWindow: 1048576,
11260
- maxTokens: 65536,
11261
- },
11262
- "google/gemini-2.5-flash-preview-09-2025": {
11263
- id: "google/gemini-2.5-flash-preview-09-2025",
11264
- name: "Gemini 2.5 Flash Preview 09-2025",
11265
- api: "anthropic-messages",
11266
- provider: "vercel-ai-gateway",
11267
- baseUrl: "https://ai-gateway.vercel.sh",
11268
- reasoning: true,
11269
- input: ["text", "image"],
11270
- cost: {
11271
- input: 0.3,
11272
- output: 2.5,
11273
- cacheRead: 0.03,
11274
- cacheWrite: 0,
11275
- },
11276
- contextWindow: 1000000,
11277
- maxTokens: 65536,
11278
- },
11279
11228
  "google/gemini-2.5-pro": {
11280
11229
  id: "google/gemini-2.5-pro",
11281
11230
  name: "Gemini 2.5 Pro",
@@ -11457,7 +11406,7 @@ export const MODELS = {
11457
11406
  cost: {
11458
11407
  input: 0.09999999999999999,
11459
11408
  output: 0.09999999999999999,
11460
- cacheRead: 0,
11409
+ cacheRead: 0.09999999999999999,
11461
11410
  cacheWrite: 0,
11462
11411
  },
11463
11412
  contextWindow: 128000,
@@ -11630,8 +11579,8 @@ export const MODELS = {
11630
11579
  cacheRead: 0.03,
11631
11580
  cacheWrite: 0.375,
11632
11581
  },
11633
- contextWindow: 4096,
11634
- maxTokens: 4096,
11582
+ contextWindow: 204800,
11583
+ maxTokens: 131000,
11635
11584
  },
11636
11585
  "mistral/codestral": {
11637
11586
  id: "mistral/codestral",
@@ -11659,8 +11608,8 @@ export const MODELS = {
11659
11608
  reasoning: false,
11660
11609
  input: ["text"],
11661
11610
  cost: {
11662
- input: 0,
11663
- output: 0,
11611
+ input: 0.39999999999999997,
11612
+ output: 2,
11664
11613
  cacheRead: 0,
11665
11614
  cacheWrite: 0,
11666
11615
  },
@@ -11693,8 +11642,8 @@ export const MODELS = {
11693
11642
  reasoning: false,
11694
11643
  input: ["text"],
11695
11644
  cost: {
11696
- input: 0,
11697
- output: 0,
11645
+ input: 0.09999999999999999,
11646
+ output: 0.3,
11698
11647
  cacheRead: 0,
11699
11648
  cacheWrite: 0,
11700
11649
  },
@@ -11710,8 +11659,8 @@ export const MODELS = {
11710
11659
  reasoning: false,
11711
11660
  input: ["text"],
11712
11661
  cost: {
11713
- input: 0.04,
11714
- output: 0.04,
11662
+ input: 0.09999999999999999,
11663
+ output: 0.09999999999999999,
11715
11664
  cacheRead: 0,
11716
11665
  cacheWrite: 0,
11717
11666
  },
@@ -11727,8 +11676,8 @@ export const MODELS = {
11727
11676
  reasoning: false,
11728
11677
  input: ["text"],
11729
11678
  cost: {
11730
- input: 0.09999999999999999,
11731
- output: 0.09999999999999999,
11679
+ input: 0.15,
11680
+ output: 0.15,
11732
11681
  cacheRead: 0,
11733
11682
  cacheWrite: 0,
11734
11683
  },
@@ -11814,7 +11763,7 @@ export const MODELS = {
11814
11763
  cost: {
11815
11764
  input: 0.6,
11816
11765
  output: 2.5,
11817
- cacheRead: 0,
11766
+ cacheRead: 0.15,
11818
11767
  cacheWrite: 0,
11819
11768
  },
11820
11769
  contextWindow: 131072,
@@ -11831,7 +11780,7 @@ export const MODELS = {
11831
11780
  cost: {
11832
11781
  input: 0.6,
11833
11782
  output: 2.5,
11834
- cacheRead: 0,
11783
+ cacheRead: 0.15,
11835
11784
  cacheWrite: 0,
11836
11785
  },
11837
11786
  contextWindow: 256000,
@@ -11880,9 +11829,9 @@ export const MODELS = {
11880
11829
  reasoning: false,
11881
11830
  input: ["text"],
11882
11831
  cost: {
11883
- input: 2.4,
11884
- output: 10,
11885
- cacheRead: 0,
11832
+ input: 1.15,
11833
+ output: 8,
11834
+ cacheRead: 0.15,
11886
11835
  cacheWrite: 0,
11887
11836
  },
11888
11837
  contextWindow: 256000,
@@ -12551,40 +12500,6 @@ export const MODELS = {
12551
12500
  contextWindow: 131072,
12552
12501
  maxTokens: 131072,
12553
12502
  },
12554
- "vercel/v0-1.0-md": {
12555
- id: "vercel/v0-1.0-md",
12556
- name: "v0-1.0-md",
12557
- api: "anthropic-messages",
12558
- provider: "vercel-ai-gateway",
12559
- baseUrl: "https://ai-gateway.vercel.sh",
12560
- reasoning: false,
12561
- input: ["text", "image"],
12562
- cost: {
12563
- input: 3,
12564
- output: 15,
12565
- cacheRead: 0,
12566
- cacheWrite: 0,
12567
- },
12568
- contextWindow: 128000,
12569
- maxTokens: 32000,
12570
- },
12571
- "vercel/v0-1.5-md": {
12572
- id: "vercel/v0-1.5-md",
12573
- name: "v0-1.5-md",
12574
- api: "anthropic-messages",
12575
- provider: "vercel-ai-gateway",
12576
- baseUrl: "https://ai-gateway.vercel.sh",
12577
- reasoning: false,
12578
- input: ["text", "image"],
12579
- cost: {
12580
- input: 3,
12581
- output: 15,
12582
- cacheRead: 0,
12583
- cacheWrite: 0,
12584
- },
12585
- contextWindow: 128000,
12586
- maxTokens: 32768,
12587
- },
12588
12503
  "xai/grok-2-vision": {
12589
12504
  id: "xai/grok-2-vision",
12590
12505
  name: "Grok 2 Vision",
@@ -12613,7 +12528,7 @@ export const MODELS = {
12613
12528
  cost: {
12614
12529
  input: 3,
12615
12530
  output: 15,
12616
- cacheRead: 0,
12531
+ cacheRead: 0.75,
12617
12532
  cacheWrite: 0,
12618
12533
  },
12619
12534
  contextWindow: 131072,
@@ -12630,7 +12545,7 @@ export const MODELS = {
12630
12545
  cost: {
12631
12546
  input: 5,
12632
12547
  output: 25,
12633
- cacheRead: 0,
12548
+ cacheRead: 1.25,
12634
12549
  cacheWrite: 0,
12635
12550
  },
12636
12551
  contextWindow: 131072,
@@ -12647,7 +12562,7 @@ export const MODELS = {
12647
12562
  cost: {
12648
12563
  input: 0.3,
12649
12564
  output: 0.5,
12650
- cacheRead: 0,
12565
+ cacheRead: 0.075,
12651
12566
  cacheWrite: 0,
12652
12567
  },
12653
12568
  contextWindow: 131072,
@@ -12900,8 +12815,8 @@ export const MODELS = {
12900
12815
  reasoning: true,
12901
12816
  input: ["text"],
12902
12817
  cost: {
12903
- input: 0.44999999999999996,
12904
- output: 1.7999999999999998,
12818
+ input: 0.6,
12819
+ output: 2.2,
12905
12820
  cacheRead: 0.11,
12906
12821
  cacheWrite: 0,
12907
12822
  },
@@ -13010,6 +12925,23 @@ export const MODELS = {
13010
12925
  contextWindow: 202800,
13011
12926
  maxTokens: 131100,
13012
12927
  },
12928
+ "zai/glm-5-turbo": {
12929
+ id: "zai/glm-5-turbo",
12930
+ name: "GLM 5 Turbo",
12931
+ api: "anthropic-messages",
12932
+ provider: "vercel-ai-gateway",
12933
+ baseUrl: "https://ai-gateway.vercel.sh",
12934
+ reasoning: true,
12935
+ input: ["text"],
12936
+ cost: {
12937
+ input: 1.2,
12938
+ output: 4,
12939
+ cacheRead: 0.24,
12940
+ cacheWrite: 0,
12941
+ },
12942
+ contextWindow: 202800,
12943
+ maxTokens: 131100,
12944
+ },
13013
12945
  },
13014
12946
  "xai": {
13015
12947
  "grok-2": {
@@ -13601,6 +13533,24 @@ export const MODELS = {
13601
13533
  contextWindow: 204800,
13602
13534
  maxTokens: 131072,
13603
13535
  },
13536
+ "glm-5-turbo": {
13537
+ id: "glm-5-turbo",
13538
+ name: "GLM-5-Turbo",
13539
+ api: "openai-completions",
13540
+ provider: "zai",
13541
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
13542
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
13543
+ reasoning: true,
13544
+ input: ["text"],
13545
+ cost: {
13546
+ input: 1.2,
13547
+ output: 4,
13548
+ cacheRead: 0.24,
13549
+ cacheWrite: 0,
13550
+ },
13551
+ contextWindow: 200000,
13552
+ maxTokens: 131072,
13553
+ },
13604
13554
  },
13605
13555
  };
13606
13556
  //# sourceMappingURL=models.generated.js.map