@mariozechner/pi-ai 0.66.0 → 0.67.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -4356,6 +4356,24 @@ export const MODELS = {
4356
4356
  contextWindow: 204800,
4357
4357
  maxTokens: 131072,
4358
4358
  },
4359
+ "MiniMaxAI/MiniMax-M2.7": {
4360
+ id: "MiniMaxAI/MiniMax-M2.7",
4361
+ name: "MiniMax-M2.7",
4362
+ api: "openai-completions",
4363
+ provider: "huggingface",
4364
+ baseUrl: "https://router.huggingface.co/v1",
4365
+ compat: { "supportsDeveloperRole": false },
4366
+ reasoning: true,
4367
+ input: ["text"],
4368
+ cost: {
4369
+ input: 0.3,
4370
+ output: 1.2,
4371
+ cacheRead: 0.06,
4372
+ cacheWrite: 0,
4373
+ },
4374
+ contextWindow: 204800,
4375
+ maxTokens: 131072,
4376
+ },
4359
4377
  "Qwen/Qwen3-235B-A22B-Thinking-2507": {
4360
4378
  id: "Qwen/Qwen3-235B-A22B-Thinking-2507",
4361
4379
  name: "Qwen3-235B-A22B-Thinking-2507",
@@ -6905,7 +6923,7 @@ export const MODELS = {
6905
6923
  cacheWrite: 3.75,
6906
6924
  },
6907
6925
  contextWindow: 200000,
6908
- maxTokens: 64000,
6926
+ maxTokens: 128000,
6909
6927
  },
6910
6928
  "anthropic/claude-3.7-sonnet:thinking": {
6911
6929
  id: "anthropic/claude-3.7-sonnet:thinking",
@@ -7040,7 +7058,7 @@ export const MODELS = {
7040
7058
  cacheRead: 0.3,
7041
7059
  cacheWrite: 3.75,
7042
7060
  },
7043
- contextWindow: 200000,
7061
+ contextWindow: 1000000,
7044
7062
  maxTokens: 64000,
7045
7063
  },
7046
7064
  "anthropic/claude-sonnet-4.5": {
@@ -7105,7 +7123,7 @@ export const MODELS = {
7105
7123
  cost: {
7106
7124
  input: 0.22,
7107
7125
  output: 0.85,
7108
- cacheRead: 0,
7126
+ cacheRead: 0.06,
7109
7127
  cacheWrite: 0,
7110
7128
  },
7111
7129
  contextWindow: 262144,
@@ -7128,23 +7146,6 @@ export const MODELS = {
7128
7146
  contextWindow: 131072,
7129
7147
  maxTokens: 131072,
7130
7148
  },
7131
- "arcee-ai/trinity-mini:free": {
7132
- id: "arcee-ai/trinity-mini:free",
7133
- name: "Arcee AI: Trinity Mini (free)",
7134
- api: "openai-completions",
7135
- provider: "openrouter",
7136
- baseUrl: "https://openrouter.ai/api/v1",
7137
- reasoning: true,
7138
- input: ["text"],
7139
- cost: {
7140
- input: 0,
7141
- output: 0,
7142
- cacheRead: 0,
7143
- cacheWrite: 0,
7144
- },
7145
- contextWindow: 131072,
7146
- maxTokens: 4096,
7147
- },
7148
7149
  "arcee-ai/virtuoso-large": {
7149
7150
  id: "arcee-ai/virtuoso-large",
7150
7151
  name: "Arcee AI: Virtuoso Large",
@@ -7392,13 +7393,13 @@ export const MODELS = {
7392
7393
  reasoning: true,
7393
7394
  input: ["text"],
7394
7395
  cost: {
7395
- input: 0.44999999999999996,
7396
+ input: 0.5,
7396
7397
  output: 2.1500000000000004,
7397
- cacheRead: 0.22499999999999998,
7398
+ cacheRead: 0.35,
7398
7399
  cacheWrite: 0,
7399
7400
  },
7400
7401
  contextWindow: 163840,
7401
- maxTokens: 65536,
7402
+ maxTokens: 4096,
7402
7403
  },
7403
7404
  "deepseek/deepseek-v3.1-terminus": {
7404
7405
  id: "deepseek/deepseek-v3.1-terminus",
@@ -7681,13 +7682,13 @@ export const MODELS = {
7681
7682
  reasoning: true,
7682
7683
  input: ["text", "image"],
7683
7684
  cost: {
7684
- input: 0.13,
7685
- output: 0.39999999999999997,
7686
- cacheRead: 0,
7685
+ input: 0.08,
7686
+ output: 0.35,
7687
+ cacheRead: 0.01,
7687
7688
  cacheWrite: 0,
7688
7689
  },
7689
7690
  contextWindow: 262144,
7690
- maxTokens: 262144,
7691
+ maxTokens: 4096,
7691
7692
  },
7692
7693
  "google/gemma-4-26b-a4b-it:free": {
7693
7694
  id: "google/gemma-4-26b-a4b-it:free",
@@ -7715,13 +7716,13 @@ export const MODELS = {
7715
7716
  reasoning: true,
7716
7717
  input: ["text", "image"],
7717
7718
  cost: {
7718
- input: 0.14,
7719
- output: 0.39999999999999997,
7720
- cacheRead: 0,
7719
+ input: 0.13,
7720
+ output: 0.38,
7721
+ cacheRead: 0.019999999499999997,
7721
7722
  cacheWrite: 0,
7722
7723
  },
7723
7724
  contextWindow: 262144,
7724
- maxTokens: 131072,
7725
+ maxTokens: 4096,
7725
7726
  },
7726
7727
  "google/gemma-4-31b-it:free": {
7727
7728
  id: "google/gemma-4-31b-it:free",
@@ -7987,13 +7988,13 @@ export const MODELS = {
7987
7988
  reasoning: true,
7988
7989
  input: ["text"],
7989
7990
  cost: {
7990
- input: 0.27,
7991
+ input: 0.29,
7991
7992
  output: 0.95,
7992
- cacheRead: 0.0290000007,
7993
+ cacheRead: 0.03,
7993
7994
  cacheWrite: 0,
7994
7995
  },
7995
7996
  contextWindow: 196608,
7996
- maxTokens: 4096,
7997
+ maxTokens: 196608,
7997
7998
  },
7998
7999
  "minimax/minimax-m2.5": {
7999
8000
  id: "minimax/minimax-m2.5",
@@ -8027,7 +8028,7 @@ export const MODELS = {
8027
8028
  cacheWrite: 0,
8028
8029
  },
8029
8030
  contextWindow: 196608,
8030
- maxTokens: 196608,
8031
+ maxTokens: 8192,
8031
8032
  },
8032
8033
  "minimax/minimax-m2.7": {
8033
8034
  id: "minimax/minimax-m2.7",
@@ -8040,11 +8041,11 @@ export const MODELS = {
8040
8041
  cost: {
8041
8042
  input: 0.3,
8042
8043
  output: 1.2,
8043
- cacheRead: 0.06,
8044
+ cacheRead: 0.059,
8044
8045
  cacheWrite: 0,
8045
8046
  },
8046
- contextWindow: 204800,
8047
- maxTokens: 131072,
8047
+ contextWindow: 196608,
8048
+ maxTokens: 4096,
8048
8049
  },
8049
8050
  "mistralai/codestral-2508": {
8050
8051
  id: "mistralai/codestral-2508",
@@ -8809,7 +8810,7 @@ export const MODELS = {
8809
8810
  cacheWrite: 0,
8810
8811
  },
8811
8812
  contextWindow: 1047576,
8812
- maxTokens: 32768,
8813
+ maxTokens: 4096,
8813
8814
  },
8814
8815
  "openai/gpt-4.1-mini": {
8815
8816
  id: "openai/gpt-4.1-mini",
@@ -8856,7 +8857,7 @@ export const MODELS = {
8856
8857
  cost: {
8857
8858
  input: 2.5,
8858
8859
  output: 10,
8859
- cacheRead: 1.25,
8860
+ cacheRead: 0,
8860
8861
  cacheWrite: 0,
8861
8862
  },
8862
8863
  contextWindow: 128000,
@@ -9077,11 +9078,11 @@ export const MODELS = {
9077
9078
  cost: {
9078
9079
  input: 0.049999999999999996,
9079
9080
  output: 0.39999999999999997,
9080
- cacheRead: 0.005,
9081
+ cacheRead: 0.01,
9081
9082
  cacheWrite: 0,
9082
9083
  },
9083
9084
  contextWindow: 400000,
9084
- maxTokens: 128000,
9085
+ maxTokens: 4096,
9085
9086
  },
9086
9087
  "openai/gpt-5-pro": {
9087
9088
  id: "openai/gpt-5-pro",
@@ -9111,7 +9112,7 @@ export const MODELS = {
9111
9112
  cost: {
9112
9113
  input: 1.25,
9113
9114
  output: 10,
9114
- cacheRead: 0.125,
9115
+ cacheRead: 0.13,
9115
9116
  cacheWrite: 0,
9116
9117
  },
9117
9118
  contextWindow: 400000,
@@ -9179,11 +9180,11 @@ export const MODELS = {
9179
9180
  cost: {
9180
9181
  input: 0.25,
9181
9182
  output: 2,
9182
- cacheRead: 0.024999999999999998,
9183
+ cacheRead: 0.03,
9183
9184
  cacheWrite: 0,
9184
9185
  },
9185
9186
  contextWindow: 400000,
9186
- maxTokens: 100000,
9187
+ maxTokens: 128000,
9187
9188
  },
9188
9189
  "openai/gpt-5.2": {
9189
9190
  id: "openai/gpt-5.2",
@@ -9217,7 +9218,7 @@ export const MODELS = {
9217
9218
  cacheWrite: 0,
9218
9219
  },
9219
9220
  contextWindow: 128000,
9220
- maxTokens: 16384,
9221
+ maxTokens: 32000,
9221
9222
  },
9222
9223
  "openai/gpt-5.2-codex": {
9223
9224
  id: "openai/gpt-5.2-codex",
@@ -9433,12 +9434,12 @@ export const MODELS = {
9433
9434
  input: ["text"],
9434
9435
  cost: {
9435
9436
  input: 0.03,
9436
- output: 0.11,
9437
- cacheRead: 0.015,
9437
+ output: 0.14,
9438
+ cacheRead: 0,
9438
9439
  cacheWrite: 0,
9439
9440
  },
9440
9441
  contextWindow: 131072,
9441
- maxTokens: 131072,
9442
+ maxTokens: 4096,
9442
9443
  },
9443
9444
  "openai/gpt-oss-20b:free": {
9444
9445
  id: "openai/gpt-oss-20b:free",
@@ -9455,7 +9456,7 @@ export const MODELS = {
9455
9456
  cacheWrite: 0,
9456
9457
  },
9457
9458
  contextWindow: 131072,
9458
- maxTokens: 131072,
9459
+ maxTokens: 8192,
9459
9460
  },
9460
9461
  "openai/gpt-oss-safeguard-20b": {
9461
9462
  id: "openai/gpt-oss-safeguard-20b",
@@ -9644,6 +9645,23 @@ export const MODELS = {
9644
9645
  contextWindow: 2000000,
9645
9646
  maxTokens: 4096,
9646
9647
  },
9648
+ "openrouter/elephant-alpha": {
9649
+ id: "openrouter/elephant-alpha",
9650
+ name: "Elephant",
9651
+ api: "openai-completions",
9652
+ provider: "openrouter",
9653
+ baseUrl: "https://openrouter.ai/api/v1",
9654
+ reasoning: false,
9655
+ input: ["text"],
9656
+ cost: {
9657
+ input: 0,
9658
+ output: 0,
9659
+ cacheRead: 0,
9660
+ cacheWrite: 0,
9661
+ },
9662
+ contextWindow: 262144,
9663
+ maxTokens: 32768,
9664
+ },
9647
9665
  "openrouter/free": {
9648
9666
  id: "openrouter/free",
9649
9667
  name: "Free Models Router",
@@ -10027,13 +10045,13 @@ export const MODELS = {
10027
10045
  reasoning: false,
10028
10046
  input: ["text"],
10029
10047
  cost: {
10030
- input: 0.12,
10031
- output: 0.75,
10032
- cacheRead: 0.06,
10048
+ input: 0.15,
10049
+ output: 0.7999999999999999,
10050
+ cacheRead: 0.12,
10033
10051
  cacheWrite: 0,
10034
10052
  },
10035
10053
  contextWindow: 262144,
10036
- maxTokens: 65536,
10054
+ maxTokens: 262144,
10037
10055
  },
10038
10056
  "qwen/qwen3-coder-plus": {
10039
10057
  id: "qwen/qwen3-coder-plus",
@@ -10511,23 +10529,6 @@ export const MODELS = {
10511
10529
  contextWindow: 262144,
10512
10530
  maxTokens: 65536,
10513
10531
  },
10514
- "stepfun/step-3.5-flash:free": {
10515
- id: "stepfun/step-3.5-flash:free",
10516
- name: "StepFun: Step 3.5 Flash (free)",
10517
- api: "openai-completions",
10518
- provider: "openrouter",
10519
- baseUrl: "https://openrouter.ai/api/v1",
10520
- reasoning: true,
10521
- input: ["text"],
10522
- cost: {
10523
- input: 0,
10524
- output: 0,
10525
- cacheRead: 0,
10526
- cacheWrite: 0,
10527
- },
10528
- contextWindow: 256000,
10529
- maxTokens: 256000,
10530
- },
10531
10532
  "thedrummer/rocinante-12b": {
10532
10533
  id: "thedrummer/rocinante-12b",
10533
10534
  name: "TheDrummer: Rocinante 12B",
@@ -10996,13 +10997,13 @@ export const MODELS = {
10996
10997
  reasoning: true,
10997
10998
  input: ["text"],
10998
10999
  cost: {
10999
- input: 1.26,
11000
- output: 3.9600000000000004,
11001
- cacheRead: 0,
11000
+ input: 0.95,
11001
+ output: 3.15,
11002
+ cacheRead: 0.475,
11002
11003
  cacheWrite: 0,
11003
11004
  },
11004
11005
  contextWindow: 202752,
11005
- maxTokens: 4096,
11006
+ maxTokens: 65535,
11006
11007
  },
11007
11008
  "z-ai/glm-5v-turbo": {
11008
11009
  id: "z-ai/glm-5v-turbo",
@@ -11782,7 +11783,7 @@ export const MODELS = {
11782
11783
  cost: {
11783
11784
  input: 0.25,
11784
11785
  output: 1.5,
11785
- cacheRead: 0,
11786
+ cacheRead: 0.03,
11786
11787
  cacheWrite: 0,
11787
11788
  },
11788
11789
  contextWindow: 1000000,