@mariozechner/pi-ai 0.65.0 → 0.65.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +58 -70
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +91 -107
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +13 -6
- package/dist/providers/openai-completions.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -3477,6 +3477,40 @@ export const MODELS = {
|
|
|
3477
3477
|
contextWindow: 131072,
|
|
3478
3478
|
maxTokens: 8192,
|
|
3479
3479
|
},
|
|
3480
|
+
"gemma-4-26b": {
|
|
3481
|
+
id: "gemma-4-26b",
|
|
3482
|
+
name: "Gemma 4 26B",
|
|
3483
|
+
api: "google-generative-ai",
|
|
3484
|
+
provider: "google",
|
|
3485
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
3486
|
+
reasoning: true,
|
|
3487
|
+
input: ["text", "image"],
|
|
3488
|
+
cost: {
|
|
3489
|
+
input: 0,
|
|
3490
|
+
output: 0,
|
|
3491
|
+
cacheRead: 0,
|
|
3492
|
+
cacheWrite: 0,
|
|
3493
|
+
},
|
|
3494
|
+
contextWindow: 256000,
|
|
3495
|
+
maxTokens: 8192,
|
|
3496
|
+
},
|
|
3497
|
+
"gemma-4-31b": {
|
|
3498
|
+
id: "gemma-4-31b",
|
|
3499
|
+
name: "Gemma 4 31B",
|
|
3500
|
+
api: "google-generative-ai",
|
|
3501
|
+
provider: "google",
|
|
3502
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
3503
|
+
reasoning: true,
|
|
3504
|
+
input: ["text", "image"],
|
|
3505
|
+
cost: {
|
|
3506
|
+
input: 0,
|
|
3507
|
+
output: 0,
|
|
3508
|
+
cacheRead: 0,
|
|
3509
|
+
cacheWrite: 0,
|
|
3510
|
+
},
|
|
3511
|
+
contextWindow: 256000,
|
|
3512
|
+
maxTokens: 8192,
|
|
3513
|
+
},
|
|
3480
3514
|
},
|
|
3481
3515
|
"google-antigravity": {
|
|
3482
3516
|
"claude-opus-4-5-thinking": {
|
|
@@ -6804,23 +6838,6 @@ export const MODELS = {
|
|
|
6804
6838
|
contextWindow: 200000,
|
|
6805
6839
|
maxTokens: 8192,
|
|
6806
6840
|
},
|
|
6807
|
-
"anthropic/claude-3.5-sonnet": {
|
|
6808
|
-
id: "anthropic/claude-3.5-sonnet",
|
|
6809
|
-
name: "Anthropic: Claude 3.5 Sonnet",
|
|
6810
|
-
api: "openai-completions",
|
|
6811
|
-
provider: "openrouter",
|
|
6812
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
6813
|
-
reasoning: false,
|
|
6814
|
-
input: ["text", "image"],
|
|
6815
|
-
cost: {
|
|
6816
|
-
input: 6,
|
|
6817
|
-
output: 30,
|
|
6818
|
-
cacheRead: 0.6,
|
|
6819
|
-
cacheWrite: 7.5,
|
|
6820
|
-
},
|
|
6821
|
-
contextWindow: 200000,
|
|
6822
|
-
maxTokens: 8192,
|
|
6823
|
-
},
|
|
6824
6841
|
"anthropic/claude-3.7-sonnet": {
|
|
6825
6842
|
id: "anthropic/claude-3.7-sonnet",
|
|
6826
6843
|
name: "Anthropic: Claude 3.7 Sonnet",
|
|
@@ -10905,7 +10922,7 @@ export const MODELS = {
|
|
|
10905
10922
|
},
|
|
10906
10923
|
"alibaba/qwen-3-235b": {
|
|
10907
10924
|
id: "alibaba/qwen-3-235b",
|
|
10908
|
-
name: "Qwen3
|
|
10925
|
+
name: "Qwen3 235B A22b Instruct 2507",
|
|
10909
10926
|
api: "anthropic-messages",
|
|
10910
10927
|
provider: "vercel-ai-gateway",
|
|
10911
10928
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
@@ -11099,13 +11116,13 @@ export const MODELS = {
|
|
|
11099
11116
|
reasoning: true,
|
|
11100
11117
|
input: ["text", "image"],
|
|
11101
11118
|
cost: {
|
|
11102
|
-
input: 0.
|
|
11103
|
-
output:
|
|
11119
|
+
input: 0.39999999999999997,
|
|
11120
|
+
output: 4,
|
|
11104
11121
|
cacheRead: 0,
|
|
11105
11122
|
cacheWrite: 0,
|
|
11106
11123
|
},
|
|
11107
|
-
contextWindow:
|
|
11108
|
-
maxTokens:
|
|
11124
|
+
contextWindow: 131072,
|
|
11125
|
+
maxTokens: 32768,
|
|
11109
11126
|
},
|
|
11110
11127
|
"alibaba/qwen3.5-flash": {
|
|
11111
11128
|
id: "alibaba/qwen3.5-flash",
|
|
@@ -11192,40 +11209,6 @@ export const MODELS = {
|
|
|
11192
11209
|
contextWindow: 200000,
|
|
11193
11210
|
maxTokens: 8192,
|
|
11194
11211
|
},
|
|
11195
|
-
"anthropic/claude-3.5-sonnet": {
|
|
11196
|
-
id: "anthropic/claude-3.5-sonnet",
|
|
11197
|
-
name: "Claude 3.5 Sonnet",
|
|
11198
|
-
api: "anthropic-messages",
|
|
11199
|
-
provider: "vercel-ai-gateway",
|
|
11200
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11201
|
-
reasoning: false,
|
|
11202
|
-
input: ["text", "image"],
|
|
11203
|
-
cost: {
|
|
11204
|
-
input: 3,
|
|
11205
|
-
output: 15,
|
|
11206
|
-
cacheRead: 0.3,
|
|
11207
|
-
cacheWrite: 3.75,
|
|
11208
|
-
},
|
|
11209
|
-
contextWindow: 200000,
|
|
11210
|
-
maxTokens: 8192,
|
|
11211
|
-
},
|
|
11212
|
-
"anthropic/claude-3.5-sonnet-20240620": {
|
|
11213
|
-
id: "anthropic/claude-3.5-sonnet-20240620",
|
|
11214
|
-
name: "Claude 3.5 Sonnet (2024-06-20)",
|
|
11215
|
-
api: "anthropic-messages",
|
|
11216
|
-
provider: "vercel-ai-gateway",
|
|
11217
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11218
|
-
reasoning: false,
|
|
11219
|
-
input: ["text", "image"],
|
|
11220
|
-
cost: {
|
|
11221
|
-
input: 3,
|
|
11222
|
-
output: 15,
|
|
11223
|
-
cacheRead: 0.3,
|
|
11224
|
-
cacheWrite: 3.75,
|
|
11225
|
-
},
|
|
11226
|
-
contextWindow: 200000,
|
|
11227
|
-
maxTokens: 8192,
|
|
11228
|
-
},
|
|
11229
11212
|
"anthropic/claude-3.7-sonnet": {
|
|
11230
11213
|
id: "anthropic/claude-3.7-sonnet",
|
|
11231
11214
|
name: "Claude 3.7 Sonnet",
|
|
@@ -11708,7 +11691,7 @@ export const MODELS = {
|
|
|
11708
11691
|
api: "anthropic-messages",
|
|
11709
11692
|
provider: "vercel-ai-gateway",
|
|
11710
11693
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11711
|
-
reasoning:
|
|
11694
|
+
reasoning: false,
|
|
11712
11695
|
input: ["text", "image"],
|
|
11713
11696
|
cost: {
|
|
11714
11697
|
input: 0.13,
|
|
@@ -11725,7 +11708,7 @@ export const MODELS = {
|
|
|
11725
11708
|
api: "anthropic-messages",
|
|
11726
11709
|
provider: "vercel-ai-gateway",
|
|
11727
11710
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11728
|
-
reasoning:
|
|
11711
|
+
reasoning: false,
|
|
11729
11712
|
input: ["text", "image"],
|
|
11730
11713
|
cost: {
|
|
11731
11714
|
input: 0.14,
|
|
@@ -11804,23 +11787,6 @@ export const MODELS = {
|
|
|
11804
11787
|
contextWindow: 128000,
|
|
11805
11788
|
maxTokens: 8192,
|
|
11806
11789
|
},
|
|
11807
|
-
"meituan/longcat-flash-thinking": {
|
|
11808
|
-
id: "meituan/longcat-flash-thinking",
|
|
11809
|
-
name: "LongCat Flash Thinking",
|
|
11810
|
-
api: "anthropic-messages",
|
|
11811
|
-
provider: "vercel-ai-gateway",
|
|
11812
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11813
|
-
reasoning: true,
|
|
11814
|
-
input: ["text"],
|
|
11815
|
-
cost: {
|
|
11816
|
-
input: 0.15,
|
|
11817
|
-
output: 1.5,
|
|
11818
|
-
cacheRead: 0,
|
|
11819
|
-
cacheWrite: 0,
|
|
11820
|
-
},
|
|
11821
|
-
contextWindow: 128000,
|
|
11822
|
-
maxTokens: 8192,
|
|
11823
|
-
},
|
|
11824
11790
|
"meta/llama-3.1-70b": {
|
|
11825
11791
|
id: "meta/llama-3.1-70b",
|
|
11826
11792
|
name: "Llama 3.1 70B Instruct",
|
|
@@ -11915,12 +11881,12 @@ export const MODELS = {
|
|
|
11915
11881
|
reasoning: false,
|
|
11916
11882
|
input: ["text", "image"],
|
|
11917
11883
|
cost: {
|
|
11918
|
-
input: 0.
|
|
11919
|
-
output:
|
|
11884
|
+
input: 0.24,
|
|
11885
|
+
output: 0.9700000000000001,
|
|
11920
11886
|
cacheRead: 0,
|
|
11921
11887
|
cacheWrite: 0,
|
|
11922
11888
|
},
|
|
11923
|
-
contextWindow:
|
|
11889
|
+
contextWindow: 128000,
|
|
11924
11890
|
maxTokens: 8192,
|
|
11925
11891
|
},
|
|
11926
11892
|
"meta/llama-4-scout": {
|
|
@@ -12231,20 +12197,20 @@ export const MODELS = {
|
|
|
12231
12197
|
},
|
|
12232
12198
|
"moonshotai/kimi-k2": {
|
|
12233
12199
|
id: "moonshotai/kimi-k2",
|
|
12234
|
-
name: "Kimi K2",
|
|
12200
|
+
name: "Kimi K2 Instruct",
|
|
12235
12201
|
api: "anthropic-messages",
|
|
12236
12202
|
provider: "vercel-ai-gateway",
|
|
12237
12203
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12238
12204
|
reasoning: false,
|
|
12239
12205
|
input: ["text"],
|
|
12240
12206
|
cost: {
|
|
12241
|
-
input: 0.
|
|
12242
|
-
output: 2.
|
|
12207
|
+
input: 0.5700000000000001,
|
|
12208
|
+
output: 2.3,
|
|
12243
12209
|
cacheRead: 0,
|
|
12244
12210
|
cacheWrite: 0,
|
|
12245
12211
|
},
|
|
12246
12212
|
contextWindow: 131072,
|
|
12247
|
-
maxTokens:
|
|
12213
|
+
maxTokens: 131072,
|
|
12248
12214
|
},
|
|
12249
12215
|
"moonshotai/kimi-k2-0905": {
|
|
12250
12216
|
id: "moonshotai/kimi-k2-0905",
|
|
@@ -13977,9 +13943,9 @@ export const MODELS = {
|
|
|
13977
13943
|
reasoning: true,
|
|
13978
13944
|
input: ["text"],
|
|
13979
13945
|
cost: {
|
|
13980
|
-
input: 0
|
|
13981
|
-
output:
|
|
13982
|
-
cacheRead: 0
|
|
13946
|
+
input: 0,
|
|
13947
|
+
output: 0,
|
|
13948
|
+
cacheRead: 0,
|
|
13983
13949
|
cacheWrite: 0,
|
|
13984
13950
|
},
|
|
13985
13951
|
contextWindow: 131072,
|
|
@@ -13995,9 +13961,9 @@ export const MODELS = {
|
|
|
13995
13961
|
reasoning: true,
|
|
13996
13962
|
input: ["text"],
|
|
13997
13963
|
cost: {
|
|
13998
|
-
input: 0
|
|
13999
|
-
output:
|
|
14000
|
-
cacheRead: 0
|
|
13964
|
+
input: 0,
|
|
13965
|
+
output: 0,
|
|
13966
|
+
cacheRead: 0,
|
|
14001
13967
|
cacheWrite: 0,
|
|
14002
13968
|
},
|
|
14003
13969
|
contextWindow: 131072,
|
|
@@ -14031,8 +13997,8 @@ export const MODELS = {
|
|
|
14031
13997
|
reasoning: true,
|
|
14032
13998
|
input: ["text", "image"],
|
|
14033
13999
|
cost: {
|
|
14034
|
-
input: 0
|
|
14035
|
-
output:
|
|
14000
|
+
input: 0,
|
|
14001
|
+
output: 0,
|
|
14036
14002
|
cacheRead: 0,
|
|
14037
14003
|
cacheWrite: 0,
|
|
14038
14004
|
},
|
|
@@ -14049,9 +14015,9 @@ export const MODELS = {
|
|
|
14049
14015
|
reasoning: true,
|
|
14050
14016
|
input: ["text"],
|
|
14051
14017
|
cost: {
|
|
14052
|
-
input: 0
|
|
14053
|
-
output:
|
|
14054
|
-
cacheRead: 0
|
|
14018
|
+
input: 0,
|
|
14019
|
+
output: 0,
|
|
14020
|
+
cacheRead: 0,
|
|
14055
14021
|
cacheWrite: 0,
|
|
14056
14022
|
},
|
|
14057
14023
|
contextWindow: 204800,
|
|
@@ -14067,8 +14033,8 @@ export const MODELS = {
|
|
|
14067
14033
|
reasoning: true,
|
|
14068
14034
|
input: ["text", "image"],
|
|
14069
14035
|
cost: {
|
|
14070
|
-
input: 0
|
|
14071
|
-
output: 0
|
|
14036
|
+
input: 0,
|
|
14037
|
+
output: 0,
|
|
14072
14038
|
cacheRead: 0,
|
|
14073
14039
|
cacheWrite: 0,
|
|
14074
14040
|
},
|
|
@@ -14085,9 +14051,9 @@ export const MODELS = {
|
|
|
14085
14051
|
reasoning: true,
|
|
14086
14052
|
input: ["text"],
|
|
14087
14053
|
cost: {
|
|
14088
|
-
input: 0
|
|
14089
|
-
output:
|
|
14090
|
-
cacheRead: 0
|
|
14054
|
+
input: 0,
|
|
14055
|
+
output: 0,
|
|
14056
|
+
cacheRead: 0,
|
|
14091
14057
|
cacheWrite: 0,
|
|
14092
14058
|
},
|
|
14093
14059
|
contextWindow: 204800,
|
|
@@ -14139,9 +14105,9 @@ export const MODELS = {
|
|
|
14139
14105
|
reasoning: true,
|
|
14140
14106
|
input: ["text"],
|
|
14141
14107
|
cost: {
|
|
14142
|
-
input:
|
|
14143
|
-
output:
|
|
14144
|
-
cacheRead: 0
|
|
14108
|
+
input: 0,
|
|
14109
|
+
output: 0,
|
|
14110
|
+
cacheRead: 0,
|
|
14145
14111
|
cacheWrite: 0,
|
|
14146
14112
|
},
|
|
14147
14113
|
contextWindow: 204800,
|
|
@@ -14157,9 +14123,27 @@ export const MODELS = {
|
|
|
14157
14123
|
reasoning: true,
|
|
14158
14124
|
input: ["text"],
|
|
14159
14125
|
cost: {
|
|
14160
|
-
input:
|
|
14161
|
-
output:
|
|
14162
|
-
cacheRead: 0
|
|
14126
|
+
input: 0,
|
|
14127
|
+
output: 0,
|
|
14128
|
+
cacheRead: 0,
|
|
14129
|
+
cacheWrite: 0,
|
|
14130
|
+
},
|
|
14131
|
+
contextWindow: 200000,
|
|
14132
|
+
maxTokens: 131072,
|
|
14133
|
+
},
|
|
14134
|
+
"glm-5.1": {
|
|
14135
|
+
id: "glm-5.1",
|
|
14136
|
+
name: "GLM-5.1",
|
|
14137
|
+
api: "openai-completions",
|
|
14138
|
+
provider: "zai",
|
|
14139
|
+
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
14140
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
14141
|
+
reasoning: true,
|
|
14142
|
+
input: ["text"],
|
|
14143
|
+
cost: {
|
|
14144
|
+
input: 0,
|
|
14145
|
+
output: 0,
|
|
14146
|
+
cacheRead: 0,
|
|
14163
14147
|
cacheWrite: 0,
|
|
14164
14148
|
},
|
|
14165
14149
|
contextWindow: 200000,
|
|
@@ -14175,9 +14159,9 @@ export const MODELS = {
|
|
|
14175
14159
|
reasoning: true,
|
|
14176
14160
|
input: ["text", "image"],
|
|
14177
14161
|
cost: {
|
|
14178
|
-
input:
|
|
14179
|
-
output:
|
|
14180
|
-
cacheRead: 0
|
|
14162
|
+
input: 0,
|
|
14163
|
+
output: 0,
|
|
14164
|
+
cacheRead: 0,
|
|
14181
14165
|
cacheWrite: 0,
|
|
14182
14166
|
},
|
|
14183
14167
|
contextWindow: 200000,
|