@mariozechner/pi-ai 0.67.4 → 0.67.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.d.ts +2 -2
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js.map +1 -1
- package/dist/models.d.ts +1 -1
- package/dist/models.d.ts.map +1 -1
- package/dist/models.generated.d.ts +1 -1
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +17 -17
- package/dist/models.generated.js.map +1 -1
- package/dist/models.js +5 -2
- package/dist/models.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts +12 -0
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +22 -3
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts +16 -2
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +29 -9
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/azure-openai-responses.d.ts.map +1 -1
- package/dist/providers/azure-openai-responses.js +5 -1
- package/dist/providers/azure-openai-responses.js.map +1 -1
- package/dist/providers/faux.d.ts.map +1 -1
- package/dist/providers/faux.js +1 -0
- package/dist/providers/faux.js.map +1 -1
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js +3 -0
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +2 -7
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +5 -1
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +6 -2
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/providers/simple-options.d.ts.map +1 -1
- package/dist/providers/simple-options.js +1 -0
- package/dist/providers/simple-options.js.map +1 -1
- package/dist/types.d.ts +9 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/headers.d.ts +2 -0
- package/dist/utils/headers.d.ts.map +1 -0
- package/dist/utils/headers.js +8 -0
- package/dist/utils/headers.js.map +1 -0
- package/package.json +3 -3
package/dist/models.generated.js
CHANGED
|
@@ -9895,7 +9895,7 @@ export const MODELS = {
|
|
|
9895
9895
|
input: 0.26,
|
|
9896
9896
|
output: 0.78,
|
|
9897
9897
|
cacheRead: 0.052000000000000005,
|
|
9898
|
-
cacheWrite: 0,
|
|
9898
|
+
cacheWrite: 0.325,
|
|
9899
9899
|
},
|
|
9900
9900
|
contextWindow: 1000000,
|
|
9901
9901
|
maxTokens: 32768,
|
|
@@ -9912,7 +9912,7 @@ export const MODELS = {
|
|
|
9912
9912
|
input: 0.26,
|
|
9913
9913
|
output: 0.78,
|
|
9914
9914
|
cacheRead: 0,
|
|
9915
|
-
cacheWrite: 0,
|
|
9915
|
+
cacheWrite: 0.325,
|
|
9916
9916
|
},
|
|
9917
9917
|
contextWindow: 1000000,
|
|
9918
9918
|
maxTokens: 32768,
|
|
@@ -9929,7 +9929,7 @@ export const MODELS = {
|
|
|
9929
9929
|
input: 0.26,
|
|
9930
9930
|
output: 0.78,
|
|
9931
9931
|
cacheRead: 0,
|
|
9932
|
-
cacheWrite: 0,
|
|
9932
|
+
cacheWrite: 0.325,
|
|
9933
9933
|
},
|
|
9934
9934
|
contextWindow: 1000000,
|
|
9935
9935
|
maxTokens: 32768,
|
|
@@ -10028,13 +10028,13 @@ export const MODELS = {
|
|
|
10028
10028
|
reasoning: true,
|
|
10029
10029
|
input: ["text"],
|
|
10030
10030
|
cost: {
|
|
10031
|
-
input: 0.
|
|
10032
|
-
output:
|
|
10031
|
+
input: 0.13,
|
|
10032
|
+
output: 0.6,
|
|
10033
10033
|
cacheRead: 0,
|
|
10034
10034
|
cacheWrite: 0,
|
|
10035
10035
|
},
|
|
10036
|
-
contextWindow:
|
|
10037
|
-
maxTokens:
|
|
10036
|
+
contextWindow: 262144,
|
|
10037
|
+
maxTokens: 262144,
|
|
10038
10038
|
},
|
|
10039
10039
|
"qwen/qwen3-30b-a3b": {
|
|
10040
10040
|
id: "qwen/qwen3-30b-a3b",
|
|
@@ -10167,7 +10167,7 @@ export const MODELS = {
|
|
|
10167
10167
|
input: 0.195,
|
|
10168
10168
|
output: 0.975,
|
|
10169
10169
|
cacheRead: 0.039,
|
|
10170
|
-
cacheWrite: 0,
|
|
10170
|
+
cacheWrite: 0.24375,
|
|
10171
10171
|
},
|
|
10172
10172
|
contextWindow: 1000000,
|
|
10173
10173
|
maxTokens: 65536,
|
|
@@ -10201,7 +10201,7 @@ export const MODELS = {
|
|
|
10201
10201
|
input: 0.65,
|
|
10202
10202
|
output: 3.25,
|
|
10203
10203
|
cacheRead: 0.13,
|
|
10204
|
-
cacheWrite: 0,
|
|
10204
|
+
cacheWrite: 0.8125,
|
|
10205
10205
|
},
|
|
10206
10206
|
contextWindow: 1000000,
|
|
10207
10207
|
maxTokens: 65536,
|
|
@@ -10235,7 +10235,7 @@ export const MODELS = {
|
|
|
10235
10235
|
input: 0.78,
|
|
10236
10236
|
output: 3.9,
|
|
10237
10237
|
cacheRead: 0.156,
|
|
10238
|
-
cacheWrite: 0,
|
|
10238
|
+
cacheWrite: 0.975,
|
|
10239
10239
|
},
|
|
10240
10240
|
contextWindow: 262144,
|
|
10241
10241
|
maxTokens: 32768,
|
|
@@ -10489,7 +10489,7 @@ export const MODELS = {
|
|
|
10489
10489
|
cost: {
|
|
10490
10490
|
input: 0.39,
|
|
10491
10491
|
output: 2.34,
|
|
10492
|
-
cacheRead: 0,
|
|
10492
|
+
cacheRead: 0.195,
|
|
10493
10493
|
cacheWrite: 0,
|
|
10494
10494
|
},
|
|
10495
10495
|
contextWindow: 262144,
|
|
@@ -10524,7 +10524,7 @@ export const MODELS = {
|
|
|
10524
10524
|
input: 0.065,
|
|
10525
10525
|
output: 0.26,
|
|
10526
10526
|
cacheRead: 0,
|
|
10527
|
-
cacheWrite: 0,
|
|
10527
|
+
cacheWrite: 0.08125,
|
|
10528
10528
|
},
|
|
10529
10529
|
contextWindow: 1000000,
|
|
10530
10530
|
maxTokens: 65536,
|
|
@@ -10541,7 +10541,7 @@ export const MODELS = {
|
|
|
10541
10541
|
input: 0.26,
|
|
10542
10542
|
output: 1.56,
|
|
10543
10543
|
cacheRead: 0,
|
|
10544
|
-
cacheWrite: 0,
|
|
10544
|
+
cacheWrite: 0.325,
|
|
10545
10545
|
},
|
|
10546
10546
|
contextWindow: 1000000,
|
|
10547
10547
|
maxTokens: 65536,
|
|
@@ -10558,7 +10558,7 @@ export const MODELS = {
|
|
|
10558
10558
|
input: 0.325,
|
|
10559
10559
|
output: 1.95,
|
|
10560
10560
|
cacheRead: 0,
|
|
10561
|
-
cacheWrite: 0,
|
|
10561
|
+
cacheWrite: 0.40625,
|
|
10562
10562
|
},
|
|
10563
10563
|
contextWindow: 1000000,
|
|
10564
10564
|
maxTokens: 65536,
|
|
@@ -13751,15 +13751,15 @@ export const MODELS = {
|
|
|
13751
13751
|
provider: "vercel-ai-gateway",
|
|
13752
13752
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13753
13753
|
reasoning: true,
|
|
13754
|
-
input: ["text"],
|
|
13754
|
+
input: ["text", "image"],
|
|
13755
13755
|
cost: {
|
|
13756
13756
|
input: 1.4,
|
|
13757
13757
|
output: 4.4,
|
|
13758
13758
|
cacheRead: 0.26,
|
|
13759
13759
|
cacheWrite: 0,
|
|
13760
13760
|
},
|
|
13761
|
-
contextWindow:
|
|
13762
|
-
maxTokens:
|
|
13761
|
+
contextWindow: 202752,
|
|
13762
|
+
maxTokens: 202752,
|
|
13763
13763
|
},
|
|
13764
13764
|
"zai/glm-5v-turbo": {
|
|
13765
13765
|
id: "zai/glm-5v-turbo",
|