@mariozechner/pi-ai 0.67.4 → 0.67.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -9895,7 +9895,7 @@ export const MODELS = {
9895
9895
  input: 0.26,
9896
9896
  output: 0.78,
9897
9897
  cacheRead: 0.052000000000000005,
9898
- cacheWrite: 0,
9898
+ cacheWrite: 0.325,
9899
9899
  },
9900
9900
  contextWindow: 1000000,
9901
9901
  maxTokens: 32768,
@@ -9912,7 +9912,7 @@ export const MODELS = {
9912
9912
  input: 0.26,
9913
9913
  output: 0.78,
9914
9914
  cacheRead: 0,
9915
- cacheWrite: 0,
9915
+ cacheWrite: 0.325,
9916
9916
  },
9917
9917
  contextWindow: 1000000,
9918
9918
  maxTokens: 32768,
@@ -9929,7 +9929,7 @@ export const MODELS = {
9929
9929
  input: 0.26,
9930
9930
  output: 0.78,
9931
9931
  cacheRead: 0,
9932
- cacheWrite: 0,
9932
+ cacheWrite: 0.325,
9933
9933
  },
9934
9934
  contextWindow: 1000000,
9935
9935
  maxTokens: 32768,
@@ -10167,7 +10167,7 @@ export const MODELS = {
10167
10167
  input: 0.195,
10168
10168
  output: 0.975,
10169
10169
  cacheRead: 0.039,
10170
- cacheWrite: 0,
10170
+ cacheWrite: 0.24375,
10171
10171
  },
10172
10172
  contextWindow: 1000000,
10173
10173
  maxTokens: 65536,
@@ -10201,7 +10201,7 @@ export const MODELS = {
10201
10201
  input: 0.65,
10202
10202
  output: 3.25,
10203
10203
  cacheRead: 0.13,
10204
- cacheWrite: 0,
10204
+ cacheWrite: 0.8125,
10205
10205
  },
10206
10206
  contextWindow: 1000000,
10207
10207
  maxTokens: 65536,
@@ -10235,7 +10235,7 @@ export const MODELS = {
10235
10235
  input: 0.78,
10236
10236
  output: 3.9,
10237
10237
  cacheRead: 0.156,
10238
- cacheWrite: 0,
10238
+ cacheWrite: 0.975,
10239
10239
  },
10240
10240
  contextWindow: 262144,
10241
10241
  maxTokens: 32768,
@@ -10524,7 +10524,7 @@ export const MODELS = {
10524
10524
  input: 0.065,
10525
10525
  output: 0.26,
10526
10526
  cacheRead: 0,
10527
- cacheWrite: 0,
10527
+ cacheWrite: 0.08125,
10528
10528
  },
10529
10529
  contextWindow: 1000000,
10530
10530
  maxTokens: 65536,
@@ -10541,7 +10541,7 @@ export const MODELS = {
10541
10541
  input: 0.26,
10542
10542
  output: 1.56,
10543
10543
  cacheRead: 0,
10544
- cacheWrite: 0,
10544
+ cacheWrite: 0.325,
10545
10545
  },
10546
10546
  contextWindow: 1000000,
10547
10547
  maxTokens: 65536,
@@ -10558,7 +10558,7 @@ export const MODELS = {
10558
10558
  input: 0.325,
10559
10559
  output: 1.95,
10560
10560
  cacheRead: 0,
10561
- cacheWrite: 0,
10561
+ cacheWrite: 0.40625,
10562
10562
  },
10563
10563
  contextWindow: 1000000,
10564
10564
  maxTokens: 65536,
@@ -13751,15 +13751,15 @@ export const MODELS = {
13751
13751
  provider: "vercel-ai-gateway",
13752
13752
  baseUrl: "https://ai-gateway.vercel.sh",
13753
13753
  reasoning: true,
13754
- input: ["text"],
13754
+ input: ["text", "image"],
13755
13755
  cost: {
13756
13756
  input: 1.4,
13757
13757
  output: 4.4,
13758
13758
  cacheRead: 0.26,
13759
13759
  cacheWrite: 0,
13760
13760
  },
13761
- contextWindow: 202800,
13762
- maxTokens: 64000,
13761
+ contextWindow: 202752,
13762
+ maxTokens: 202752,
13763
13763
  },
13764
13764
  "zai/glm-5v-turbo": {
13765
13765
  id: "zai/glm-5v-turbo",