@mariozechner/pi-ai 0.5.39 → 0.5.41

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1410,6 +1410,40 @@ export const MODELS = {
1410
1410
  },
1411
1411
  },
1412
1412
  openrouter: {
1413
+ "qwen/qwen3-coder-flash": {
1414
+ id: "qwen/qwen3-coder-flash",
1415
+ name: "Qwen: Qwen3 Coder Flash",
1416
+ api: "openai-completions",
1417
+ provider: "openrouter",
1418
+ baseUrl: "https://openrouter.ai/api/v1",
1419
+ reasoning: false,
1420
+ input: ["text"],
1421
+ cost: {
1422
+ input: 0.3,
1423
+ output: 1.5,
1424
+ cacheRead: 0.08,
1425
+ cacheWrite: 0,
1426
+ },
1427
+ contextWindow: 128000,
1428
+ maxTokens: 65536,
1429
+ },
1430
+ "qwen/qwen3-coder-plus": {
1431
+ id: "qwen/qwen3-coder-plus",
1432
+ name: "Qwen: Qwen3 Coder Plus",
1433
+ api: "openai-completions",
1434
+ provider: "openrouter",
1435
+ baseUrl: "https://openrouter.ai/api/v1",
1436
+ reasoning: false,
1437
+ input: ["text"],
1438
+ cost: {
1439
+ input: 1,
1440
+ output: 5,
1441
+ cacheRead: 0.09999999999999999,
1442
+ cacheWrite: 0,
1443
+ },
1444
+ contextWindow: 128000,
1445
+ maxTokens: 65536,
1446
+ },
1413
1447
  "qwen/qwen3-next-80b-a3b-thinking": {
1414
1448
  id: "qwen/qwen3-next-80b-a3b-thinking",
1415
1449
  name: "Qwen: Qwen3 Next 80B A3B Thinking",
@@ -1419,8 +1453,8 @@ export const MODELS = {
1419
1453
  reasoning: true,
1420
1454
  input: ["text"],
1421
1455
  cost: {
1422
- input: 0.09782604,
1423
- output: 0.391304304,
1456
+ input: 0.09999999999999999,
1457
+ output: 0.7999999999999999,
1424
1458
  cacheRead: 0,
1425
1459
  cacheWrite: 0,
1426
1460
  },
@@ -1436,8 +1470,8 @@ export const MODELS = {
1436
1470
  reasoning: false,
1437
1471
  input: ["text"],
1438
1472
  cost: {
1439
- input: 0.09782604,
1440
- output: 0.391304304,
1473
+ input: 0.09999999999999999,
1474
+ output: 0.7999999999999999,
1441
1475
  cacheRead: 0,
1442
1476
  cacheWrite: 0,
1443
1477
  },
@@ -1453,13 +1487,13 @@ export const MODELS = {
1453
1487
  reasoning: false,
1454
1488
  input: ["text"],
1455
1489
  cost: {
1456
- input: 0.15,
1457
- output: 0.75,
1490
+ input: 0.12,
1491
+ output: 0.6,
1458
1492
  cacheRead: 0,
1459
1493
  cacheWrite: 0,
1460
1494
  },
1461
1495
  contextWindow: 131072,
1462
- maxTokens: 131072,
1496
+ maxTokens: 4096,
1463
1497
  },
1464
1498
  "qwen/qwen-plus-2025-07-28": {
1465
1499
  id: "qwen/qwen-plus-2025-07-28",
@@ -1589,8 +1623,8 @@ export const MODELS = {
1589
1623
  reasoning: false,
1590
1624
  input: ["text"],
1591
1625
  cost: {
1592
- input: 0.38043459999999996,
1593
- output: 1.52173896,
1626
+ input: 0.38,
1627
+ output: 1.52,
1594
1628
  cacheRead: 0,
1595
1629
  cacheWrite: 0,
1596
1630
  },
@@ -1640,8 +1674,8 @@ export const MODELS = {
1640
1674
  reasoning: true,
1641
1675
  input: ["text"],
1642
1676
  cost: {
1643
- input: 0.08967387,
1644
- output: 0.358695612,
1677
+ input: 0.08,
1678
+ output: 0.29,
1645
1679
  cacheRead: 0,
1646
1680
  cacheWrite: 0,
1647
1681
  },
@@ -1657,8 +1691,8 @@ export const MODELS = {
1657
1691
  reasoning: true,
1658
1692
  input: ["text"],
1659
1693
  cost: {
1660
- input: 0.127173852,
1661
- output: 0.5086955952000001,
1694
+ input: 0.11,
1695
+ output: 0.38,
1662
1696
  cacheRead: 0,
1663
1697
  cacheWrite: 0,
1664
1698
  },
@@ -1696,7 +1730,7 @@ export const MODELS = {
1696
1730
  cacheRead: 0,
1697
1731
  cacheWrite: 0,
1698
1732
  },
1699
- contextWindow: 32768,
1733
+ contextWindow: 163840,
1700
1734
  maxTokens: 4096,
1701
1735
  },
1702
1736
  "deepseek/deepseek-chat-v3.1": {
@@ -1816,7 +1850,7 @@ export const MODELS = {
1816
1850
  cacheWrite: 0,
1817
1851
  },
1818
1852
  contextWindow: 262144,
1819
- maxTokens: 262144,
1853
+ maxTokens: 4096,
1820
1854
  },
1821
1855
  "qwen/qwen3-30b-a3b-instruct-2507": {
1822
1856
  id: "qwen/qwen3-30b-a3b-instruct-2507",
@@ -1827,8 +1861,8 @@ export const MODELS = {
1827
1861
  reasoning: false,
1828
1862
  input: ["text"],
1829
1863
  cost: {
1830
- input: 0.07065213999999999,
1831
- output: 0.282608664,
1864
+ input: 0.07,
1865
+ output: 0.28,
1832
1866
  cacheRead: 0,
1833
1867
  cacheWrite: 0,
1834
1868
  },
@@ -1844,8 +1878,8 @@ export const MODELS = {
1844
1878
  reasoning: true,
1845
1879
  input: ["text"],
1846
1880
  cost: {
1847
- input: 0.41249980199999997,
1848
- output: 1.6499998152000002,
1881
+ input: 0.41,
1882
+ output: 1.6500000000000001,
1849
1883
  cacheRead: 0,
1850
1884
  cacheWrite: 0,
1851
1885
  },
@@ -1895,8 +1929,8 @@ export const MODELS = {
1895
1929
  reasoning: true,
1896
1930
  input: ["text"],
1897
1931
  cost: {
1898
- input: 0.0974999532,
1899
- output: 0.38999995632,
1932
+ input: 0.09999999999999999,
1933
+ output: 0.39,
1900
1934
  cacheRead: 0,
1901
1935
  cacheWrite: 0,
1902
1936
  },
@@ -1946,8 +1980,8 @@ export const MODELS = {
1946
1980
  reasoning: false,
1947
1981
  input: ["text"],
1948
1982
  cost: {
1949
- input: 0.24999987999999998,
1950
- output: 0.999999888,
1983
+ input: 0.22,
1984
+ output: 0.95,
1951
1985
  cacheRead: 0,
1952
1986
  cacheWrite: 0,
1953
1987
  },
@@ -1963,13 +1997,13 @@ export const MODELS = {
1963
1997
  reasoning: false,
1964
1998
  input: ["text"],
1965
1999
  cost: {
1966
- input: 0.0974999532,
1967
- output: 0.38999995632,
2000
+ input: 0.09999999999999999,
2001
+ output: 0.09999999999999999,
1968
2002
  cacheRead: 0,
1969
2003
  cacheWrite: 0,
1970
2004
  },
1971
2005
  contextWindow: 262144,
1972
- maxTokens: 4096,
2006
+ maxTokens: 262144,
1973
2007
  },
1974
2008
  "moonshotai/kimi-k2:free": {
1975
2009
  id: "moonshotai/kimi-k2:free",
@@ -2167,8 +2201,8 @@ export const MODELS = {
2167
2201
  reasoning: true,
2168
2202
  input: ["text"],
2169
2203
  cost: {
2170
- input: 0.24999987999999998,
2171
- output: 0.999999888,
2204
+ input: 0.39999999999999997,
2205
+ output: 1.75,
2172
2206
  cacheRead: 0,
2173
2207
  cacheWrite: 0,
2174
2208
  },
@@ -2201,8 +2235,8 @@ export const MODELS = {
2201
2235
  reasoning: false,
2202
2236
  input: ["text"],
2203
2237
  cost: {
2204
- input: 0.035869548,
2205
- output: 0.14347824480000002,
2238
+ input: 0.04,
2239
+ output: 0.14,
2206
2240
  cacheRead: 0,
2207
2241
  cacheWrite: 0,
2208
2242
  },
@@ -2303,8 +2337,8 @@ export const MODELS = {
2303
2337
  reasoning: true,
2304
2338
  input: ["text"],
2305
2339
  cost: {
2306
- input: 0.035869548,
2307
- output: 0.14347824480000002,
2340
+ input: 0.06,
2341
+ output: 0.22,
2308
2342
  cacheRead: 0,
2309
2343
  cacheWrite: 0,
2310
2344
  },
@@ -2337,8 +2371,8 @@ export const MODELS = {
2337
2371
  reasoning: true,
2338
2372
  input: ["text"],
2339
2373
  cost: {
2340
- input: 0.0322825932,
2341
- output: 0.12913042032,
2374
+ input: 0.03,
2375
+ output: 0.13,
2342
2376
  cacheRead: 0,
2343
2377
  cacheWrite: 0,
2344
2378
  },
@@ -2507,14 +2541,31 @@ export const MODELS = {
2507
2541
  reasoning: false,
2508
2542
  input: ["text", "image"],
2509
2543
  cost: {
2510
- input: 0.03804346,
2511
- output: 0.152173896,
2544
+ input: 0.04,
2545
+ output: 0.15,
2512
2546
  cacheRead: 0,
2513
2547
  cacheWrite: 0,
2514
2548
  },
2515
2549
  contextWindow: 131072,
2516
2550
  maxTokens: 96000,
2517
2551
  },
2552
+ "microsoft/phi-4-multimodal-instruct": {
2553
+ id: "microsoft/phi-4-multimodal-instruct",
2554
+ name: "Microsoft: Phi 4 Multimodal Instruct",
2555
+ api: "openai-completions",
2556
+ provider: "openrouter",
2557
+ baseUrl: "https://openrouter.ai/api/v1",
2558
+ reasoning: false,
2559
+ input: ["text", "image"],
2560
+ cost: {
2561
+ input: 0.049999999999999996,
2562
+ output: 0.09999999999999999,
2563
+ cacheRead: 0,
2564
+ cacheWrite: 0,
2565
+ },
2566
+ contextWindow: 131072,
2567
+ maxTokens: 4096,
2568
+ },
2518
2569
  "qwen/qwq-32b": {
2519
2570
  id: "qwen/qwq-32b",
2520
2571
  name: "Qwen: QwQ 32B",
@@ -2609,8 +2660,8 @@ export const MODELS = {
2609
2660
  reasoning: false,
2610
2661
  input: ["text"],
2611
2662
  cost: {
2612
- input: 0.03804346,
2613
- output: 0.152173896,
2663
+ input: 0.04,
2664
+ output: 0.15,
2614
2665
  cacheRead: 0,
2615
2666
  cacheWrite: 0,
2616
2667
  },
@@ -2626,8 +2677,8 @@ export const MODELS = {
2626
2677
  reasoning: true,
2627
2678
  input: ["text"],
2628
2679
  cost: {
2629
- input: 0.03260868,
2630
- output: 0.130434768,
2680
+ input: 0.03,
2681
+ output: 0.13,
2631
2682
  cacheRead: 0,
2632
2683
  cacheWrite: 0,
2633
2684
  },
@@ -2711,13 +2762,13 @@ export const MODELS = {
2711
2762
  reasoning: false,
2712
2763
  input: ["text"],
2713
2764
  cost: {
2714
- input: 0.038000000000000006,
2715
- output: 0.12,
2765
+ input: 0.012,
2766
+ output: 0.036,
2716
2767
  cacheRead: 0,
2717
2768
  cacheWrite: 0,
2718
2769
  },
2719
2770
  contextWindow: 131072,
2720
- maxTokens: 16384,
2771
+ maxTokens: 8192,
2721
2772
  },
2722
2773
  "amazon/nova-lite-v1": {
2723
2774
  id: "amazon/nova-lite-v1",
@@ -2915,8 +2966,8 @@ export const MODELS = {
2915
2966
  reasoning: false,
2916
2967
  input: ["text"],
2917
2968
  cost: {
2918
- input: 0.06521736,
2919
- output: 0.260869536,
2969
+ input: 0.07,
2970
+ output: 0.26,
2920
2971
  cacheRead: 0,
2921
2972
  cacheWrite: 0,
2922
2973
  },
@@ -2940,34 +2991,34 @@ export const MODELS = {
2940
2991
  contextWindow: 32768,
2941
2992
  maxTokens: 4096,
2942
2993
  },
2943
- "cohere/command-r-plus-08-2024": {
2944
- id: "cohere/command-r-plus-08-2024",
2945
- name: "Cohere: Command R+ (08-2024)",
2994
+ "cohere/command-r-08-2024": {
2995
+ id: "cohere/command-r-08-2024",
2996
+ name: "Cohere: Command R (08-2024)",
2946
2997
  api: "openai-completions",
2947
2998
  provider: "openrouter",
2948
2999
  baseUrl: "https://openrouter.ai/api/v1",
2949
3000
  reasoning: false,
2950
3001
  input: ["text"],
2951
3002
  cost: {
2952
- input: 2.5,
2953
- output: 10,
3003
+ input: 0.15,
3004
+ output: 0.6,
2954
3005
  cacheRead: 0,
2955
3006
  cacheWrite: 0,
2956
3007
  },
2957
3008
  contextWindow: 128000,
2958
3009
  maxTokens: 4000,
2959
3010
  },
2960
- "cohere/command-r-08-2024": {
2961
- id: "cohere/command-r-08-2024",
2962
- name: "Cohere: Command R (08-2024)",
3011
+ "cohere/command-r-plus-08-2024": {
3012
+ id: "cohere/command-r-plus-08-2024",
3013
+ name: "Cohere: Command R+ (08-2024)",
2963
3014
  api: "openai-completions",
2964
3015
  provider: "openrouter",
2965
3016
  baseUrl: "https://openrouter.ai/api/v1",
2966
3017
  reasoning: false,
2967
3018
  input: ["text"],
2968
3019
  cost: {
2969
- input: 0.15,
2970
- output: 0.6,
3020
+ input: 2.5,
3021
+ output: 10,
2971
3022
  cacheRead: 0,
2972
3023
  cacheWrite: 0,
2973
3024
  },
@@ -3068,13 +3119,13 @@ export const MODELS = {
3068
3119
  reasoning: false,
3069
3120
  input: ["text"],
3070
3121
  cost: {
3071
- input: 0.017934774,
3072
- output: 0.07173912240000001,
3122
+ input: 0.02,
3123
+ output: 0.04,
3073
3124
  cacheRead: 0,
3074
3125
  cacheWrite: 0,
3075
3126
  },
3076
3127
  contextWindow: 131072,
3077
- maxTokens: 128000,
3128
+ maxTokens: 16384,
3078
3129
  },
3079
3130
  "mistralai/mistral-7b-instruct-v0.3": {
3080
3131
  id: "mistralai/mistral-7b-instruct-v0.3",
@@ -3161,34 +3212,34 @@ export const MODELS = {
3161
3212
  contextWindow: 128000,
3162
3213
  maxTokens: 4096,
3163
3214
  },
3164
- "meta-llama/llama-3-70b-instruct": {
3165
- id: "meta-llama/llama-3-70b-instruct",
3166
- name: "Meta: Llama 3 70B Instruct",
3215
+ "meta-llama/llama-3-8b-instruct": {
3216
+ id: "meta-llama/llama-3-8b-instruct",
3217
+ name: "Meta: Llama 3 8B Instruct",
3167
3218
  api: "openai-completions",
3168
3219
  provider: "openrouter",
3169
3220
  baseUrl: "https://openrouter.ai/api/v1",
3170
3221
  reasoning: false,
3171
3222
  input: ["text"],
3172
3223
  cost: {
3173
- input: 0.3,
3174
- output: 0.39999999999999997,
3224
+ input: 0.03,
3225
+ output: 0.06,
3175
3226
  cacheRead: 0,
3176
3227
  cacheWrite: 0,
3177
3228
  },
3178
3229
  contextWindow: 8192,
3179
3230
  maxTokens: 16384,
3180
3231
  },
3181
- "meta-llama/llama-3-8b-instruct": {
3182
- id: "meta-llama/llama-3-8b-instruct",
3183
- name: "Meta: Llama 3 8B Instruct",
3232
+ "meta-llama/llama-3-70b-instruct": {
3233
+ id: "meta-llama/llama-3-70b-instruct",
3234
+ name: "Meta: Llama 3 70B Instruct",
3184
3235
  api: "openai-completions",
3185
3236
  provider: "openrouter",
3186
3237
  baseUrl: "https://openrouter.ai/api/v1",
3187
3238
  reasoning: false,
3188
3239
  input: ["text"],
3189
3240
  cost: {
3190
- input: 0.03,
3191
- output: 0.06,
3241
+ input: 0.3,
3242
+ output: 0.39999999999999997,
3192
3243
  cacheRead: 0,
3193
3244
  cacheWrite: 0,
3194
3245
  },