@mariozechner/pi-ai 0.7.16 → 0.7.17

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -361,6 +361,23 @@ export const MODELS = {
361
361
  contextWindow: 1048576,
362
362
  maxTokens: 65536,
363
363
  },
364
+ "gemini-3-pro-preview": {
365
+ id: "gemini-3-pro-preview",
366
+ name: "Gemini 3 Pro Preview",
367
+ api: "google-generative-ai",
368
+ provider: "google",
369
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
370
+ reasoning: true,
371
+ input: ["text", "image"],
372
+ cost: {
373
+ input: 2,
374
+ output: 12,
375
+ cacheRead: 0.2,
376
+ cacheWrite: 0,
377
+ },
378
+ contextWindow: 200000,
379
+ maxTokens: 65536,
380
+ },
364
381
  "gemini-2.5-flash": {
365
382
  id: "gemini-2.5-flash",
366
383
  name: "Gemini 2.5 Flash",
@@ -1903,6 +1920,23 @@ export const MODELS = {
1903
1920
  },
1904
1921
  },
1905
1922
  openrouter: {
1923
+ "google/gemini-3-pro-preview": {
1924
+ id: "google/gemini-3-pro-preview",
1925
+ name: "Google: Gemini 3 Pro Preview",
1926
+ api: "openai-completions",
1927
+ provider: "openrouter",
1928
+ baseUrl: "https://openrouter.ai/api/v1",
1929
+ reasoning: true,
1930
+ input: ["text", "image"],
1931
+ cost: {
1932
+ input: 2,
1933
+ output: 12,
1934
+ cacheRead: 0.19999999999999998,
1935
+ cacheWrite: 2.375,
1936
+ },
1937
+ contextWindow: 1048576,
1938
+ maxTokens: 65536,
1939
+ },
1906
1940
  "openrouter/sherlock-dash-alpha": {
1907
1941
  id: "openrouter/sherlock-dash-alpha",
1908
1942
  name: "Sherlock Dash Alpha",
@@ -1954,6 +1988,23 @@ export const MODELS = {
1954
1988
  contextWindow: 400000,
1955
1989
  maxTokens: 128000,
1956
1990
  },
1991
+ "openai/gpt-5.1-chat": {
1992
+ id: "openai/gpt-5.1-chat",
1993
+ name: "OpenAI: GPT-5.1 Chat",
1994
+ api: "openai-completions",
1995
+ provider: "openrouter",
1996
+ baseUrl: "https://openrouter.ai/api/v1",
1997
+ reasoning: false,
1998
+ input: ["text", "image"],
1999
+ cost: {
2000
+ input: 1.25,
2001
+ output: 10,
2002
+ cacheRead: 0.125,
2003
+ cacheWrite: 0,
2004
+ },
2005
+ contextWindow: 128000,
2006
+ maxTokens: 16384,
2007
+ },
1957
2008
  "openai/gpt-5.1-codex": {
1958
2009
  id: "openai/gpt-5.1-codex",
1959
2010
  name: "OpenAI: GPT-5.1-Codex",
@@ -2014,13 +2065,13 @@ export const MODELS = {
2014
2065
  reasoning: true,
2015
2066
  input: ["text"],
2016
2067
  cost: {
2017
- input: 0.55,
2018
- output: 2.25,
2068
+ input: 0.5,
2069
+ output: 2.5,
2019
2070
  cacheRead: 0,
2020
2071
  cacheWrite: 0,
2021
2072
  },
2022
2073
  contextWindow: 262144,
2023
- maxTokens: 16384,
2074
+ maxTokens: 262144,
2024
2075
  },
2025
2076
  "amazon/nova-premier-v1": {
2026
2077
  id: "amazon/nova-premier-v1",
@@ -2209,40 +2260,6 @@ export const MODELS = {
2209
2260
  contextWindow: 400000,
2210
2261
  maxTokens: 128000,
2211
2262
  },
2212
- "inclusionai/ring-1t": {
2213
- id: "inclusionai/ring-1t",
2214
- name: "inclusionAI: Ring 1T",
2215
- api: "openai-completions",
2216
- provider: "openrouter",
2217
- baseUrl: "https://openrouter.ai/api/v1",
2218
- reasoning: true,
2219
- input: ["text"],
2220
- cost: {
2221
- input: 0.5700000000000001,
2222
- output: 2.2800000000000002,
2223
- cacheRead: 0,
2224
- cacheWrite: 0,
2225
- },
2226
- contextWindow: 131072,
2227
- maxTokens: 131072,
2228
- },
2229
- "inclusionai/ling-1t": {
2230
- id: "inclusionai/ling-1t",
2231
- name: "inclusionAI: Ling-1T",
2232
- api: "openai-completions",
2233
- provider: "openrouter",
2234
- baseUrl: "https://openrouter.ai/api/v1",
2235
- reasoning: false,
2236
- input: ["text"],
2237
- cost: {
2238
- input: 0.5700000000000001,
2239
- output: 2.2800000000000002,
2240
- cacheRead: 0,
2241
- cacheWrite: 0,
2242
- },
2243
- contextWindow: 131072,
2244
- maxTokens: 131072,
2245
- },
2246
2263
  "openai/o3-deep-research": {
2247
2264
  id: "openai/o3-deep-research",
2248
2265
  name: "OpenAI: o3 Deep Research",
@@ -2473,13 +2490,13 @@ export const MODELS = {
2473
2490
  reasoning: false,
2474
2491
  input: ["text", "image"],
2475
2492
  cost: {
2476
- input: 0.22,
2477
- output: 0.88,
2493
+ input: 0.21,
2494
+ output: 1.9,
2478
2495
  cacheRead: 0,
2479
2496
  cacheWrite: 0,
2480
2497
  },
2481
- contextWindow: 262144,
2482
- maxTokens: 4096,
2498
+ contextWindow: 131072,
2499
+ maxTokens: 32768,
2483
2500
  },
2484
2501
  "qwen/qwen3-max": {
2485
2502
  id: "qwen/qwen3-max",
@@ -2847,13 +2864,13 @@ export const MODELS = {
2847
2864
  reasoning: true,
2848
2865
  input: ["text"],
2849
2866
  cost: {
2850
- input: 0.09,
2851
- output: 0.3,
2867
+ input: 0.051,
2868
+ output: 0.33999999999999997,
2852
2869
  cacheRead: 0,
2853
2870
  cacheWrite: 0,
2854
2871
  },
2855
- contextWindow: 262144,
2856
- maxTokens: 131072,
2872
+ contextWindow: 32768,
2873
+ maxTokens: 4096,
2857
2874
  },
2858
2875
  "x-ai/grok-code-fast-1": {
2859
2876
  id: "x-ai/grok-code-fast-1",
@@ -3731,8 +3748,8 @@ export const MODELS = {
3731
3748
  reasoning: true,
3732
3749
  input: ["text"],
3733
3750
  cost: {
3734
- input: 0.39999999999999997,
3735
- output: 1.75,
3751
+ input: 0.19999999999999998,
3752
+ output: 4.5,
3736
3753
  cacheRead: 0,
3737
3754
  cacheWrite: 0,
3738
3755
  },
@@ -4258,13 +4275,13 @@ export const MODELS = {
4258
4275
  reasoning: false,
4259
4276
  input: ["text", "image"],
4260
4277
  cost: {
4261
- input: 0.09,
4262
- output: 0.16,
4278
+ input: 0.07,
4279
+ output: 0.5,
4263
4280
  cacheRead: 0,
4264
4281
  cacheWrite: 0,
4265
4282
  },
4266
4283
  contextWindow: 131072,
4267
- maxTokens: 16384,
4284
+ maxTokens: 131072,
4268
4285
  },
4269
4286
  "qwen/qwq-32b": {
4270
4287
  id: "qwen/qwq-32b",
@@ -4810,34 +4827,34 @@ export const MODELS = {
4810
4827
  contextWindow: 200000,
4811
4828
  maxTokens: 8192,
4812
4829
  },
4813
- "mistralai/ministral-3b": {
4814
- id: "mistralai/ministral-3b",
4815
- name: "Mistral: Ministral 3B",
4830
+ "mistralai/ministral-8b": {
4831
+ id: "mistralai/ministral-8b",
4832
+ name: "Mistral: Ministral 8B",
4816
4833
  api: "openai-completions",
4817
4834
  provider: "openrouter",
4818
4835
  baseUrl: "https://openrouter.ai/api/v1",
4819
4836
  reasoning: false,
4820
4837
  input: ["text"],
4821
4838
  cost: {
4822
- input: 0.04,
4823
- output: 0.04,
4839
+ input: 0.09999999999999999,
4840
+ output: 0.09999999999999999,
4824
4841
  cacheRead: 0,
4825
4842
  cacheWrite: 0,
4826
4843
  },
4827
4844
  contextWindow: 131072,
4828
4845
  maxTokens: 4096,
4829
4846
  },
4830
- "mistralai/ministral-8b": {
4831
- id: "mistralai/ministral-8b",
4832
- name: "Mistral: Ministral 8B",
4847
+ "mistralai/ministral-3b": {
4848
+ id: "mistralai/ministral-3b",
4849
+ name: "Mistral: Ministral 3B",
4833
4850
  api: "openai-completions",
4834
4851
  provider: "openrouter",
4835
4852
  baseUrl: "https://openrouter.ai/api/v1",
4836
4853
  reasoning: false,
4837
4854
  input: ["text"],
4838
4855
  cost: {
4839
- input: 0.09999999999999999,
4840
- output: 0.09999999999999999,
4856
+ input: 0.04,
4857
+ output: 0.04,
4841
4858
  cacheRead: 0,
4842
4859
  cacheWrite: 0,
4843
4860
  },
@@ -4946,34 +4963,34 @@ export const MODELS = {
4946
4963
  contextWindow: 32768,
4947
4964
  maxTokens: 4096,
4948
4965
  },
4949
- "cohere/command-r-plus-08-2024": {
4950
- id: "cohere/command-r-plus-08-2024",
4951
- name: "Cohere: Command R+ (08-2024)",
4966
+ "cohere/command-r-08-2024": {
4967
+ id: "cohere/command-r-08-2024",
4968
+ name: "Cohere: Command R (08-2024)",
4952
4969
  api: "openai-completions",
4953
4970
  provider: "openrouter",
4954
4971
  baseUrl: "https://openrouter.ai/api/v1",
4955
4972
  reasoning: false,
4956
4973
  input: ["text"],
4957
4974
  cost: {
4958
- input: 2.5,
4959
- output: 10,
4975
+ input: 0.15,
4976
+ output: 0.6,
4960
4977
  cacheRead: 0,
4961
4978
  cacheWrite: 0,
4962
4979
  },
4963
4980
  contextWindow: 128000,
4964
4981
  maxTokens: 4000,
4965
4982
  },
4966
- "cohere/command-r-08-2024": {
4967
- id: "cohere/command-r-08-2024",
4968
- name: "Cohere: Command R (08-2024)",
4983
+ "cohere/command-r-plus-08-2024": {
4984
+ id: "cohere/command-r-plus-08-2024",
4985
+ name: "Cohere: Command R+ (08-2024)",
4969
4986
  api: "openai-completions",
4970
4987
  provider: "openrouter",
4971
4988
  baseUrl: "https://openrouter.ai/api/v1",
4972
4989
  reasoning: false,
4973
4990
  input: ["text"],
4974
4991
  cost: {
4975
- input: 0.15,
4976
- output: 0.6,
4992
+ input: 2.5,
4993
+ output: 10,
4977
4994
  cacheRead: 0,
4978
4995
  cacheWrite: 0,
4979
4996
  },
@@ -5558,38 +5575,38 @@ export const MODELS = {
5558
5575
  contextWindow: 8191,
5559
5576
  maxTokens: 4096,
5560
5577
  },
5561
- "openai/gpt-4": {
5562
- id: "openai/gpt-4",
5563
- name: "OpenAI: GPT-4",
5578
+ "openai/gpt-3.5-turbo": {
5579
+ id: "openai/gpt-3.5-turbo",
5580
+ name: "OpenAI: GPT-3.5 Turbo",
5564
5581
  api: "openai-completions",
5565
5582
  provider: "openrouter",
5566
5583
  baseUrl: "https://openrouter.ai/api/v1",
5567
5584
  reasoning: false,
5568
5585
  input: ["text"],
5569
5586
  cost: {
5570
- input: 30,
5571
- output: 60,
5587
+ input: 0.5,
5588
+ output: 1.5,
5572
5589
  cacheRead: 0,
5573
5590
  cacheWrite: 0,
5574
5591
  },
5575
- contextWindow: 8191,
5592
+ contextWindow: 16385,
5576
5593
  maxTokens: 4096,
5577
5594
  },
5578
- "openai/gpt-3.5-turbo": {
5579
- id: "openai/gpt-3.5-turbo",
5580
- name: "OpenAI: GPT-3.5 Turbo",
5595
+ "openai/gpt-4": {
5596
+ id: "openai/gpt-4",
5597
+ name: "OpenAI: GPT-4",
5581
5598
  api: "openai-completions",
5582
5599
  provider: "openrouter",
5583
5600
  baseUrl: "https://openrouter.ai/api/v1",
5584
5601
  reasoning: false,
5585
5602
  input: ["text"],
5586
5603
  cost: {
5587
- input: 0.5,
5588
- output: 1.5,
5604
+ input: 30,
5605
+ output: 60,
5589
5606
  cacheRead: 0,
5590
5607
  cacheWrite: 0,
5591
5608
  },
5592
- contextWindow: 16385,
5609
+ contextWindow: 8191,
5593
5610
  maxTokens: 4096,
5594
5611
  },
5595
5612
  "openrouter/auto": {