@mariozechner/pi-ai 0.65.0 → 0.65.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3477,6 +3477,40 @@ export const MODELS = {
3477
3477
  contextWindow: 131072,
3478
3478
  maxTokens: 8192,
3479
3479
  },
3480
+ "gemma-4-26b": {
3481
+ id: "gemma-4-26b",
3482
+ name: "Gemma 4 26B",
3483
+ api: "google-generative-ai",
3484
+ provider: "google",
3485
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
3486
+ reasoning: true,
3487
+ input: ["text", "image"],
3488
+ cost: {
3489
+ input: 0,
3490
+ output: 0,
3491
+ cacheRead: 0,
3492
+ cacheWrite: 0,
3493
+ },
3494
+ contextWindow: 256000,
3495
+ maxTokens: 8192,
3496
+ },
3497
+ "gemma-4-31b": {
3498
+ id: "gemma-4-31b",
3499
+ name: "Gemma 4 31B",
3500
+ api: "google-generative-ai",
3501
+ provider: "google",
3502
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
3503
+ reasoning: true,
3504
+ input: ["text", "image"],
3505
+ cost: {
3506
+ input: 0,
3507
+ output: 0,
3508
+ cacheRead: 0,
3509
+ cacheWrite: 0,
3510
+ },
3511
+ contextWindow: 256000,
3512
+ maxTokens: 8192,
3513
+ },
3480
3514
  },
3481
3515
  "google-antigravity": {
3482
3516
  "claude-opus-4-5-thinking": {
@@ -6804,23 +6838,6 @@ export const MODELS = {
6804
6838
  contextWindow: 200000,
6805
6839
  maxTokens: 8192,
6806
6840
  },
6807
- "anthropic/claude-3.5-sonnet": {
6808
- id: "anthropic/claude-3.5-sonnet",
6809
- name: "Anthropic: Claude 3.5 Sonnet",
6810
- api: "openai-completions",
6811
- provider: "openrouter",
6812
- baseUrl: "https://openrouter.ai/api/v1",
6813
- reasoning: false,
6814
- input: ["text", "image"],
6815
- cost: {
6816
- input: 6,
6817
- output: 30,
6818
- cacheRead: 0.6,
6819
- cacheWrite: 7.5,
6820
- },
6821
- contextWindow: 200000,
6822
- maxTokens: 8192,
6823
- },
6824
6841
  "anthropic/claude-3.7-sonnet": {
6825
6842
  id: "anthropic/claude-3.7-sonnet",
6826
6843
  name: "Anthropic: Claude 3.7 Sonnet",
@@ -10905,7 +10922,7 @@ export const MODELS = {
10905
10922
  },
10906
10923
  "alibaba/qwen-3-235b": {
10907
10924
  id: "alibaba/qwen-3-235b",
10908
- name: "Qwen3-235B-A22B",
10925
+ name: "Qwen3 235B A22b Instruct 2507",
10909
10926
  api: "anthropic-messages",
10910
10927
  provider: "vercel-ai-gateway",
10911
10928
  baseUrl: "https://ai-gateway.vercel.sh",
@@ -11099,13 +11116,13 @@ export const MODELS = {
11099
11116
  reasoning: true,
11100
11117
  input: ["text", "image"],
11101
11118
  cost: {
11102
- input: 0.22,
11103
- output: 0.88,
11119
+ input: 0.39999999999999997,
11120
+ output: 4,
11104
11121
  cacheRead: 0,
11105
11122
  cacheWrite: 0,
11106
11123
  },
11107
- contextWindow: 256000,
11108
- maxTokens: 256000,
11124
+ contextWindow: 131072,
11125
+ maxTokens: 32768,
11109
11126
  },
11110
11127
  "alibaba/qwen3.5-flash": {
11111
11128
  id: "alibaba/qwen3.5-flash",
@@ -11192,40 +11209,6 @@ export const MODELS = {
11192
11209
  contextWindow: 200000,
11193
11210
  maxTokens: 8192,
11194
11211
  },
11195
- "anthropic/claude-3.5-sonnet": {
11196
- id: "anthropic/claude-3.5-sonnet",
11197
- name: "Claude 3.5 Sonnet",
11198
- api: "anthropic-messages",
11199
- provider: "vercel-ai-gateway",
11200
- baseUrl: "https://ai-gateway.vercel.sh",
11201
- reasoning: false,
11202
- input: ["text", "image"],
11203
- cost: {
11204
- input: 3,
11205
- output: 15,
11206
- cacheRead: 0.3,
11207
- cacheWrite: 3.75,
11208
- },
11209
- contextWindow: 200000,
11210
- maxTokens: 8192,
11211
- },
11212
- "anthropic/claude-3.5-sonnet-20240620": {
11213
- id: "anthropic/claude-3.5-sonnet-20240620",
11214
- name: "Claude 3.5 Sonnet (2024-06-20)",
11215
- api: "anthropic-messages",
11216
- provider: "vercel-ai-gateway",
11217
- baseUrl: "https://ai-gateway.vercel.sh",
11218
- reasoning: false,
11219
- input: ["text", "image"],
11220
- cost: {
11221
- input: 3,
11222
- output: 15,
11223
- cacheRead: 0.3,
11224
- cacheWrite: 3.75,
11225
- },
11226
- contextWindow: 200000,
11227
- maxTokens: 8192,
11228
- },
11229
11212
  "anthropic/claude-3.7-sonnet": {
11230
11213
  id: "anthropic/claude-3.7-sonnet",
11231
11214
  name: "Claude 3.7 Sonnet",
@@ -11708,7 +11691,7 @@ export const MODELS = {
11708
11691
  api: "anthropic-messages",
11709
11692
  provider: "vercel-ai-gateway",
11710
11693
  baseUrl: "https://ai-gateway.vercel.sh",
11711
- reasoning: true,
11694
+ reasoning: false,
11712
11695
  input: ["text", "image"],
11713
11696
  cost: {
11714
11697
  input: 0.13,
@@ -11725,7 +11708,7 @@ export const MODELS = {
11725
11708
  api: "anthropic-messages",
11726
11709
  provider: "vercel-ai-gateway",
11727
11710
  baseUrl: "https://ai-gateway.vercel.sh",
11728
- reasoning: true,
11711
+ reasoning: false,
11729
11712
  input: ["text", "image"],
11730
11713
  cost: {
11731
11714
  input: 0.14,
@@ -11804,23 +11787,6 @@ export const MODELS = {
11804
11787
  contextWindow: 128000,
11805
11788
  maxTokens: 8192,
11806
11789
  },
11807
- "meituan/longcat-flash-thinking": {
11808
- id: "meituan/longcat-flash-thinking",
11809
- name: "LongCat Flash Thinking",
11810
- api: "anthropic-messages",
11811
- provider: "vercel-ai-gateway",
11812
- baseUrl: "https://ai-gateway.vercel.sh",
11813
- reasoning: true,
11814
- input: ["text"],
11815
- cost: {
11816
- input: 0.15,
11817
- output: 1.5,
11818
- cacheRead: 0,
11819
- cacheWrite: 0,
11820
- },
11821
- contextWindow: 128000,
11822
- maxTokens: 8192,
11823
- },
11824
11790
  "meta/llama-3.1-70b": {
11825
11791
  id: "meta/llama-3.1-70b",
11826
11792
  name: "Llama 3.1 70B Instruct",
@@ -11915,12 +11881,12 @@ export const MODELS = {
11915
11881
  reasoning: false,
11916
11882
  input: ["text", "image"],
11917
11883
  cost: {
11918
- input: 0.35,
11919
- output: 1.15,
11884
+ input: 0.24,
11885
+ output: 0.9700000000000001,
11920
11886
  cacheRead: 0,
11921
11887
  cacheWrite: 0,
11922
11888
  },
11923
- contextWindow: 524288,
11889
+ contextWindow: 128000,
11924
11890
  maxTokens: 8192,
11925
11891
  },
11926
11892
  "meta/llama-4-scout": {
@@ -12231,20 +12197,20 @@ export const MODELS = {
12231
12197
  },
12232
12198
  "moonshotai/kimi-k2": {
12233
12199
  id: "moonshotai/kimi-k2",
12234
- name: "Kimi K2",
12200
+ name: "Kimi K2 Instruct",
12235
12201
  api: "anthropic-messages",
12236
12202
  provider: "vercel-ai-gateway",
12237
12203
  baseUrl: "https://ai-gateway.vercel.sh",
12238
12204
  reasoning: false,
12239
12205
  input: ["text"],
12240
12206
  cost: {
12241
- input: 0.6,
12242
- output: 2.5,
12207
+ input: 0.5700000000000001,
12208
+ output: 2.3,
12243
12209
  cacheRead: 0,
12244
12210
  cacheWrite: 0,
12245
12211
  },
12246
12212
  contextWindow: 131072,
12247
- maxTokens: 16384,
12213
+ maxTokens: 131072,
12248
12214
  },
12249
12215
  "moonshotai/kimi-k2-0905": {
12250
12216
  id: "moonshotai/kimi-k2-0905",
@@ -13977,9 +13943,9 @@ export const MODELS = {
13977
13943
  reasoning: true,
13978
13944
  input: ["text"],
13979
13945
  cost: {
13980
- input: 0.6,
13981
- output: 2.2,
13982
- cacheRead: 0.11,
13946
+ input: 0,
13947
+ output: 0,
13948
+ cacheRead: 0,
13983
13949
  cacheWrite: 0,
13984
13950
  },
13985
13951
  contextWindow: 131072,
@@ -13995,9 +13961,9 @@ export const MODELS = {
13995
13961
  reasoning: true,
13996
13962
  input: ["text"],
13997
13963
  cost: {
13998
- input: 0.2,
13999
- output: 1.1,
14000
- cacheRead: 0.03,
13964
+ input: 0,
13965
+ output: 0,
13966
+ cacheRead: 0,
14001
13967
  cacheWrite: 0,
14002
13968
  },
14003
13969
  contextWindow: 131072,
@@ -14031,8 +13997,8 @@ export const MODELS = {
14031
13997
  reasoning: true,
14032
13998
  input: ["text", "image"],
14033
13999
  cost: {
14034
- input: 0.6,
14035
- output: 1.8,
14000
+ input: 0,
14001
+ output: 0,
14036
14002
  cacheRead: 0,
14037
14003
  cacheWrite: 0,
14038
14004
  },
@@ -14049,9 +14015,9 @@ export const MODELS = {
14049
14015
  reasoning: true,
14050
14016
  input: ["text"],
14051
14017
  cost: {
14052
- input: 0.6,
14053
- output: 2.2,
14054
- cacheRead: 0.11,
14018
+ input: 0,
14019
+ output: 0,
14020
+ cacheRead: 0,
14055
14021
  cacheWrite: 0,
14056
14022
  },
14057
14023
  contextWindow: 204800,
@@ -14067,8 +14033,8 @@ export const MODELS = {
14067
14033
  reasoning: true,
14068
14034
  input: ["text", "image"],
14069
14035
  cost: {
14070
- input: 0.3,
14071
- output: 0.9,
14036
+ input: 0,
14037
+ output: 0,
14072
14038
  cacheRead: 0,
14073
14039
  cacheWrite: 0,
14074
14040
  },
@@ -14085,9 +14051,9 @@ export const MODELS = {
14085
14051
  reasoning: true,
14086
14052
  input: ["text"],
14087
14053
  cost: {
14088
- input: 0.6,
14089
- output: 2.2,
14090
- cacheRead: 0.11,
14054
+ input: 0,
14055
+ output: 0,
14056
+ cacheRead: 0,
14091
14057
  cacheWrite: 0,
14092
14058
  },
14093
14059
  contextWindow: 204800,
@@ -14139,9 +14105,9 @@ export const MODELS = {
14139
14105
  reasoning: true,
14140
14106
  input: ["text"],
14141
14107
  cost: {
14142
- input: 1,
14143
- output: 3.2,
14144
- cacheRead: 0.2,
14108
+ input: 0,
14109
+ output: 0,
14110
+ cacheRead: 0,
14145
14111
  cacheWrite: 0,
14146
14112
  },
14147
14113
  contextWindow: 204800,
@@ -14157,9 +14123,27 @@ export const MODELS = {
14157
14123
  reasoning: true,
14158
14124
  input: ["text"],
14159
14125
  cost: {
14160
- input: 1.2,
14161
- output: 4,
14162
- cacheRead: 0.24,
14126
+ input: 0,
14127
+ output: 0,
14128
+ cacheRead: 0,
14129
+ cacheWrite: 0,
14130
+ },
14131
+ contextWindow: 200000,
14132
+ maxTokens: 131072,
14133
+ },
14134
+ "glm-5.1": {
14135
+ id: "glm-5.1",
14136
+ name: "GLM-5.1",
14137
+ api: "openai-completions",
14138
+ provider: "zai",
14139
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
14140
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
14141
+ reasoning: true,
14142
+ input: ["text"],
14143
+ cost: {
14144
+ input: 0,
14145
+ output: 0,
14146
+ cacheRead: 0,
14163
14147
  cacheWrite: 0,
14164
14148
  },
14165
14149
  contextWindow: 200000,
@@ -14175,9 +14159,9 @@ export const MODELS = {
14175
14159
  reasoning: true,
14176
14160
  input: ["text", "image"],
14177
14161
  cost: {
14178
- input: 1.2,
14179
- output: 4,
14180
- cacheRead: 0.24,
14162
+ input: 0,
14163
+ output: 0,
14164
+ cacheRead: 0,
14181
14165
  cacheWrite: 0,
14182
14166
  },
14183
14167
  contextWindow: 200000,