@mariozechner/pi-ai 0.66.1 → 0.67.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +37 -34
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +74 -73
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js +1 -1
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/google.d.ts.map +1 -1
- package/dist/providers/google.js +28 -3
- package/dist/providers/google.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts +2 -0
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +33 -4
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/types.d.ts +61 -4
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -4356,6 +4356,24 @@ export const MODELS = {
|
|
|
4356
4356
|
contextWindow: 204800,
|
|
4357
4357
|
maxTokens: 131072,
|
|
4358
4358
|
},
|
|
4359
|
+
"MiniMaxAI/MiniMax-M2.7": {
|
|
4360
|
+
id: "MiniMaxAI/MiniMax-M2.7",
|
|
4361
|
+
name: "MiniMax-M2.7",
|
|
4362
|
+
api: "openai-completions",
|
|
4363
|
+
provider: "huggingface",
|
|
4364
|
+
baseUrl: "https://router.huggingface.co/v1",
|
|
4365
|
+
compat: { "supportsDeveloperRole": false },
|
|
4366
|
+
reasoning: true,
|
|
4367
|
+
input: ["text"],
|
|
4368
|
+
cost: {
|
|
4369
|
+
input: 0.3,
|
|
4370
|
+
output: 1.2,
|
|
4371
|
+
cacheRead: 0.06,
|
|
4372
|
+
cacheWrite: 0,
|
|
4373
|
+
},
|
|
4374
|
+
contextWindow: 204800,
|
|
4375
|
+
maxTokens: 131072,
|
|
4376
|
+
},
|
|
4359
4377
|
"Qwen/Qwen3-235B-A22B-Thinking-2507": {
|
|
4360
4378
|
id: "Qwen/Qwen3-235B-A22B-Thinking-2507",
|
|
4361
4379
|
name: "Qwen3-235B-A22B-Thinking-2507",
|
|
@@ -6905,7 +6923,7 @@ export const MODELS = {
|
|
|
6905
6923
|
cacheWrite: 3.75,
|
|
6906
6924
|
},
|
|
6907
6925
|
contextWindow: 200000,
|
|
6908
|
-
maxTokens:
|
|
6926
|
+
maxTokens: 128000,
|
|
6909
6927
|
},
|
|
6910
6928
|
"anthropic/claude-3.7-sonnet:thinking": {
|
|
6911
6929
|
id: "anthropic/claude-3.7-sonnet:thinking",
|
|
@@ -7040,7 +7058,7 @@ export const MODELS = {
|
|
|
7040
7058
|
cacheRead: 0.3,
|
|
7041
7059
|
cacheWrite: 3.75,
|
|
7042
7060
|
},
|
|
7043
|
-
contextWindow:
|
|
7061
|
+
contextWindow: 1000000,
|
|
7044
7062
|
maxTokens: 64000,
|
|
7045
7063
|
},
|
|
7046
7064
|
"anthropic/claude-sonnet-4.5": {
|
|
@@ -7105,7 +7123,7 @@ export const MODELS = {
|
|
|
7105
7123
|
cost: {
|
|
7106
7124
|
input: 0.22,
|
|
7107
7125
|
output: 0.85,
|
|
7108
|
-
cacheRead: 0,
|
|
7126
|
+
cacheRead: 0.06,
|
|
7109
7127
|
cacheWrite: 0,
|
|
7110
7128
|
},
|
|
7111
7129
|
contextWindow: 262144,
|
|
@@ -7128,23 +7146,6 @@ export const MODELS = {
|
|
|
7128
7146
|
contextWindow: 131072,
|
|
7129
7147
|
maxTokens: 131072,
|
|
7130
7148
|
},
|
|
7131
|
-
"arcee-ai/trinity-mini:free": {
|
|
7132
|
-
id: "arcee-ai/trinity-mini:free",
|
|
7133
|
-
name: "Arcee AI: Trinity Mini (free)",
|
|
7134
|
-
api: "openai-completions",
|
|
7135
|
-
provider: "openrouter",
|
|
7136
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7137
|
-
reasoning: true,
|
|
7138
|
-
input: ["text"],
|
|
7139
|
-
cost: {
|
|
7140
|
-
input: 0,
|
|
7141
|
-
output: 0,
|
|
7142
|
-
cacheRead: 0,
|
|
7143
|
-
cacheWrite: 0,
|
|
7144
|
-
},
|
|
7145
|
-
contextWindow: 131072,
|
|
7146
|
-
maxTokens: 4096,
|
|
7147
|
-
},
|
|
7148
7149
|
"arcee-ai/virtuoso-large": {
|
|
7149
7150
|
id: "arcee-ai/virtuoso-large",
|
|
7150
7151
|
name: "Arcee AI: Virtuoso Large",
|
|
@@ -7392,13 +7393,13 @@ export const MODELS = {
|
|
|
7392
7393
|
reasoning: true,
|
|
7393
7394
|
input: ["text"],
|
|
7394
7395
|
cost: {
|
|
7395
|
-
input: 0.
|
|
7396
|
+
input: 0.5,
|
|
7396
7397
|
output: 2.1500000000000004,
|
|
7397
|
-
cacheRead: 0.
|
|
7398
|
+
cacheRead: 0.35,
|
|
7398
7399
|
cacheWrite: 0,
|
|
7399
7400
|
},
|
|
7400
7401
|
contextWindow: 163840,
|
|
7401
|
-
maxTokens:
|
|
7402
|
+
maxTokens: 4096,
|
|
7402
7403
|
},
|
|
7403
7404
|
"deepseek/deepseek-v3.1-terminus": {
|
|
7404
7405
|
id: "deepseek/deepseek-v3.1-terminus",
|
|
@@ -7681,13 +7682,13 @@ export const MODELS = {
|
|
|
7681
7682
|
reasoning: true,
|
|
7682
7683
|
input: ["text", "image"],
|
|
7683
7684
|
cost: {
|
|
7684
|
-
input: 0.
|
|
7685
|
-
output: 0.
|
|
7686
|
-
cacheRead: 0,
|
|
7685
|
+
input: 0.08,
|
|
7686
|
+
output: 0.35,
|
|
7687
|
+
cacheRead: 0.01,
|
|
7687
7688
|
cacheWrite: 0,
|
|
7688
7689
|
},
|
|
7689
7690
|
contextWindow: 262144,
|
|
7690
|
-
maxTokens:
|
|
7691
|
+
maxTokens: 4096,
|
|
7691
7692
|
},
|
|
7692
7693
|
"google/gemma-4-26b-a4b-it:free": {
|
|
7693
7694
|
id: "google/gemma-4-26b-a4b-it:free",
|
|
@@ -7715,13 +7716,13 @@ export const MODELS = {
|
|
|
7715
7716
|
reasoning: true,
|
|
7716
7717
|
input: ["text", "image"],
|
|
7717
7718
|
cost: {
|
|
7718
|
-
input: 0.
|
|
7719
|
-
output: 0.
|
|
7720
|
-
cacheRead: 0,
|
|
7719
|
+
input: 0.13,
|
|
7720
|
+
output: 0.38,
|
|
7721
|
+
cacheRead: 0.019999999499999997,
|
|
7721
7722
|
cacheWrite: 0,
|
|
7722
7723
|
},
|
|
7723
7724
|
contextWindow: 262144,
|
|
7724
|
-
maxTokens:
|
|
7725
|
+
maxTokens: 4096,
|
|
7725
7726
|
},
|
|
7726
7727
|
"google/gemma-4-31b-it:free": {
|
|
7727
7728
|
id: "google/gemma-4-31b-it:free",
|
|
@@ -8027,7 +8028,7 @@ export const MODELS = {
|
|
|
8027
8028
|
cacheWrite: 0,
|
|
8028
8029
|
},
|
|
8029
8030
|
contextWindow: 196608,
|
|
8030
|
-
maxTokens:
|
|
8031
|
+
maxTokens: 8192,
|
|
8031
8032
|
},
|
|
8032
8033
|
"minimax/minimax-m2.7": {
|
|
8033
8034
|
id: "minimax/minimax-m2.7",
|
|
@@ -8040,11 +8041,11 @@ export const MODELS = {
|
|
|
8040
8041
|
cost: {
|
|
8041
8042
|
input: 0.3,
|
|
8042
8043
|
output: 1.2,
|
|
8043
|
-
cacheRead: 0.
|
|
8044
|
+
cacheRead: 0.059,
|
|
8044
8045
|
cacheWrite: 0,
|
|
8045
8046
|
},
|
|
8046
|
-
contextWindow:
|
|
8047
|
-
maxTokens:
|
|
8047
|
+
contextWindow: 196608,
|
|
8048
|
+
maxTokens: 4096,
|
|
8048
8049
|
},
|
|
8049
8050
|
"mistralai/codestral-2508": {
|
|
8050
8051
|
id: "mistralai/codestral-2508",
|
|
@@ -8809,7 +8810,7 @@ export const MODELS = {
|
|
|
8809
8810
|
cacheWrite: 0,
|
|
8810
8811
|
},
|
|
8811
8812
|
contextWindow: 1047576,
|
|
8812
|
-
maxTokens:
|
|
8813
|
+
maxTokens: 4096,
|
|
8813
8814
|
},
|
|
8814
8815
|
"openai/gpt-4.1-mini": {
|
|
8815
8816
|
id: "openai/gpt-4.1-mini",
|
|
@@ -8856,7 +8857,7 @@ export const MODELS = {
|
|
|
8856
8857
|
cost: {
|
|
8857
8858
|
input: 2.5,
|
|
8858
8859
|
output: 10,
|
|
8859
|
-
cacheRead:
|
|
8860
|
+
cacheRead: 0,
|
|
8860
8861
|
cacheWrite: 0,
|
|
8861
8862
|
},
|
|
8862
8863
|
contextWindow: 128000,
|
|
@@ -9077,11 +9078,11 @@ export const MODELS = {
|
|
|
9077
9078
|
cost: {
|
|
9078
9079
|
input: 0.049999999999999996,
|
|
9079
9080
|
output: 0.39999999999999997,
|
|
9080
|
-
cacheRead: 0.
|
|
9081
|
+
cacheRead: 0.01,
|
|
9081
9082
|
cacheWrite: 0,
|
|
9082
9083
|
},
|
|
9083
9084
|
contextWindow: 400000,
|
|
9084
|
-
maxTokens:
|
|
9085
|
+
maxTokens: 4096,
|
|
9085
9086
|
},
|
|
9086
9087
|
"openai/gpt-5-pro": {
|
|
9087
9088
|
id: "openai/gpt-5-pro",
|
|
@@ -9111,7 +9112,7 @@ export const MODELS = {
|
|
|
9111
9112
|
cost: {
|
|
9112
9113
|
input: 1.25,
|
|
9113
9114
|
output: 10,
|
|
9114
|
-
cacheRead: 0.
|
|
9115
|
+
cacheRead: 0.13,
|
|
9115
9116
|
cacheWrite: 0,
|
|
9116
9117
|
},
|
|
9117
9118
|
contextWindow: 400000,
|
|
@@ -9179,11 +9180,11 @@ export const MODELS = {
|
|
|
9179
9180
|
cost: {
|
|
9180
9181
|
input: 0.25,
|
|
9181
9182
|
output: 2,
|
|
9182
|
-
cacheRead: 0.
|
|
9183
|
+
cacheRead: 0.03,
|
|
9183
9184
|
cacheWrite: 0,
|
|
9184
9185
|
},
|
|
9185
9186
|
contextWindow: 400000,
|
|
9186
|
-
maxTokens:
|
|
9187
|
+
maxTokens: 128000,
|
|
9187
9188
|
},
|
|
9188
9189
|
"openai/gpt-5.2": {
|
|
9189
9190
|
id: "openai/gpt-5.2",
|
|
@@ -9217,7 +9218,7 @@ export const MODELS = {
|
|
|
9217
9218
|
cacheWrite: 0,
|
|
9218
9219
|
},
|
|
9219
9220
|
contextWindow: 128000,
|
|
9220
|
-
maxTokens:
|
|
9221
|
+
maxTokens: 32000,
|
|
9221
9222
|
},
|
|
9222
9223
|
"openai/gpt-5.2-codex": {
|
|
9223
9224
|
id: "openai/gpt-5.2-codex",
|
|
@@ -9433,12 +9434,12 @@ export const MODELS = {
|
|
|
9433
9434
|
input: ["text"],
|
|
9434
9435
|
cost: {
|
|
9435
9436
|
input: 0.03,
|
|
9436
|
-
output: 0.
|
|
9437
|
-
cacheRead: 0
|
|
9437
|
+
output: 0.14,
|
|
9438
|
+
cacheRead: 0,
|
|
9438
9439
|
cacheWrite: 0,
|
|
9439
9440
|
},
|
|
9440
9441
|
contextWindow: 131072,
|
|
9441
|
-
maxTokens:
|
|
9442
|
+
maxTokens: 4096,
|
|
9442
9443
|
},
|
|
9443
9444
|
"openai/gpt-oss-20b:free": {
|
|
9444
9445
|
id: "openai/gpt-oss-20b:free",
|
|
@@ -9455,7 +9456,7 @@ export const MODELS = {
|
|
|
9455
9456
|
cacheWrite: 0,
|
|
9456
9457
|
},
|
|
9457
9458
|
contextWindow: 131072,
|
|
9458
|
-
maxTokens:
|
|
9459
|
+
maxTokens: 8192,
|
|
9459
9460
|
},
|
|
9460
9461
|
"openai/gpt-oss-safeguard-20b": {
|
|
9461
9462
|
id: "openai/gpt-oss-safeguard-20b",
|
|
@@ -9644,6 +9645,23 @@ export const MODELS = {
|
|
|
9644
9645
|
contextWindow: 2000000,
|
|
9645
9646
|
maxTokens: 4096,
|
|
9646
9647
|
},
|
|
9648
|
+
"openrouter/elephant-alpha": {
|
|
9649
|
+
id: "openrouter/elephant-alpha",
|
|
9650
|
+
name: "Elephant",
|
|
9651
|
+
api: "openai-completions",
|
|
9652
|
+
provider: "openrouter",
|
|
9653
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9654
|
+
reasoning: false,
|
|
9655
|
+
input: ["text"],
|
|
9656
|
+
cost: {
|
|
9657
|
+
input: 0,
|
|
9658
|
+
output: 0,
|
|
9659
|
+
cacheRead: 0,
|
|
9660
|
+
cacheWrite: 0,
|
|
9661
|
+
},
|
|
9662
|
+
contextWindow: 262144,
|
|
9663
|
+
maxTokens: 32768,
|
|
9664
|
+
},
|
|
9647
9665
|
"openrouter/free": {
|
|
9648
9666
|
id: "openrouter/free",
|
|
9649
9667
|
name: "Free Models Router",
|
|
@@ -10027,13 +10045,13 @@ export const MODELS = {
|
|
|
10027
10045
|
reasoning: false,
|
|
10028
10046
|
input: ["text"],
|
|
10029
10047
|
cost: {
|
|
10030
|
-
input: 0.
|
|
10031
|
-
output: 0.
|
|
10032
|
-
cacheRead: 0.
|
|
10048
|
+
input: 0.15,
|
|
10049
|
+
output: 0.7999999999999999,
|
|
10050
|
+
cacheRead: 0.12,
|
|
10033
10051
|
cacheWrite: 0,
|
|
10034
10052
|
},
|
|
10035
10053
|
contextWindow: 262144,
|
|
10036
|
-
maxTokens:
|
|
10054
|
+
maxTokens: 262144,
|
|
10037
10055
|
},
|
|
10038
10056
|
"qwen/qwen3-coder-plus": {
|
|
10039
10057
|
id: "qwen/qwen3-coder-plus",
|
|
@@ -10511,23 +10529,6 @@ export const MODELS = {
|
|
|
10511
10529
|
contextWindow: 262144,
|
|
10512
10530
|
maxTokens: 65536,
|
|
10513
10531
|
},
|
|
10514
|
-
"stepfun/step-3.5-flash:free": {
|
|
10515
|
-
id: "stepfun/step-3.5-flash:free",
|
|
10516
|
-
name: "StepFun: Step 3.5 Flash (free)",
|
|
10517
|
-
api: "openai-completions",
|
|
10518
|
-
provider: "openrouter",
|
|
10519
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
10520
|
-
reasoning: true,
|
|
10521
|
-
input: ["text"],
|
|
10522
|
-
cost: {
|
|
10523
|
-
input: 0,
|
|
10524
|
-
output: 0,
|
|
10525
|
-
cacheRead: 0,
|
|
10526
|
-
cacheWrite: 0,
|
|
10527
|
-
},
|
|
10528
|
-
contextWindow: 256000,
|
|
10529
|
-
maxTokens: 256000,
|
|
10530
|
-
},
|
|
10531
10532
|
"thedrummer/rocinante-12b": {
|
|
10532
10533
|
id: "thedrummer/rocinante-12b",
|
|
10533
10534
|
name: "TheDrummer: Rocinante 12B",
|
|
@@ -10996,13 +10997,13 @@ export const MODELS = {
|
|
|
10996
10997
|
reasoning: true,
|
|
10997
10998
|
input: ["text"],
|
|
10998
10999
|
cost: {
|
|
10999
|
-
input:
|
|
11000
|
-
output:
|
|
11001
|
-
cacheRead: 0.
|
|
11000
|
+
input: 0.95,
|
|
11001
|
+
output: 3.15,
|
|
11002
|
+
cacheRead: 0.475,
|
|
11002
11003
|
cacheWrite: 0,
|
|
11003
11004
|
},
|
|
11004
11005
|
contextWindow: 202752,
|
|
11005
|
-
maxTokens:
|
|
11006
|
+
maxTokens: 65535,
|
|
11006
11007
|
},
|
|
11007
11008
|
"z-ai/glm-5v-turbo": {
|
|
11008
11009
|
id: "z-ai/glm-5v-turbo",
|
|
@@ -11782,7 +11783,7 @@ export const MODELS = {
|
|
|
11782
11783
|
cost: {
|
|
11783
11784
|
input: 0.25,
|
|
11784
11785
|
output: 1.5,
|
|
11785
|
-
cacheRead: 0,
|
|
11786
|
+
cacheRead: 0.03,
|
|
11786
11787
|
cacheWrite: 0,
|
|
11787
11788
|
},
|
|
11788
11789
|
contextWindow: 1000000,
|