@mariozechner/pi-ai 0.5.31 → 0.5.34

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (74) hide show
  1. package/README.md +275 -10
  2. package/dist/agent/agent.d.ts +5 -0
  3. package/dist/agent/agent.d.ts.map +1 -0
  4. package/dist/agent/agent.js +161 -0
  5. package/dist/agent/agent.js.map +1 -0
  6. package/dist/agent/index.d.ts +4 -0
  7. package/dist/agent/index.d.ts.map +1 -0
  8. package/dist/agent/index.js +3 -0
  9. package/dist/agent/index.js.map +1 -0
  10. package/dist/agent/tools/calculate.d.ts +17 -0
  11. package/dist/agent/tools/calculate.d.ts.map +1 -0
  12. package/dist/agent/tools/calculate.js +23 -0
  13. package/dist/agent/tools/calculate.js.map +1 -0
  14. package/dist/agent/tools/get-current-time.d.ts +20 -0
  15. package/dist/agent/tools/get-current-time.d.ts.map +1 -0
  16. package/dist/agent/tools/get-current-time.js +36 -0
  17. package/dist/agent/tools/get-current-time.js.map +1 -0
  18. package/dist/agent/tools/index.d.ts +3 -0
  19. package/dist/agent/tools/index.d.ts.map +1 -0
  20. package/dist/agent/tools/index.js +3 -0
  21. package/dist/agent/tools/index.js.map +1 -0
  22. package/dist/agent/types.d.ts +53 -0
  23. package/dist/agent/types.d.ts.map +1 -0
  24. package/dist/agent/types.js +2 -0
  25. package/dist/agent/types.js.map +1 -0
  26. package/dist/event-stream.d.ts +19 -0
  27. package/dist/event-stream.d.ts.map +1 -0
  28. package/dist/event-stream.js +77 -0
  29. package/dist/event-stream.js.map +1 -0
  30. package/dist/index.d.ts +3 -1
  31. package/dist/index.d.ts.map +1 -1
  32. package/dist/index.js +3 -1
  33. package/dist/index.js.map +1 -1
  34. package/dist/models.generated.d.ts +129 -10
  35. package/dist/models.generated.d.ts.map +1 -1
  36. package/dist/models.generated.js +234 -115
  37. package/dist/models.generated.js.map +1 -1
  38. package/dist/providers/anthropic.d.ts +3 -3
  39. package/dist/providers/anthropic.d.ts.map +1 -1
  40. package/dist/providers/anthropic.js +108 -56
  41. package/dist/providers/anthropic.js.map +1 -1
  42. package/dist/providers/google.d.ts +3 -3
  43. package/dist/providers/google.d.ts.map +1 -1
  44. package/dist/providers/google.js +56 -13
  45. package/dist/providers/google.js.map +1 -1
  46. package/dist/providers/openai-completions.d.ts +3 -3
  47. package/dist/providers/openai-completions.d.ts.map +1 -1
  48. package/dist/providers/openai-completions.js +92 -150
  49. package/dist/providers/openai-completions.js.map +1 -1
  50. package/dist/providers/openai-responses.d.ts +3 -3
  51. package/dist/providers/openai-responses.d.ts.map +1 -1
  52. package/dist/providers/openai-responses.js +59 -9
  53. package/dist/providers/openai-responses.js.map +1 -1
  54. package/dist/providers/{utils.d.ts → transorm-messages.d.ts} +1 -1
  55. package/dist/providers/transorm-messages.d.ts.map +1 -0
  56. package/dist/providers/{utils.js → transorm-messages.js} +1 -1
  57. package/dist/providers/transorm-messages.js.map +1 -0
  58. package/dist/stream.d.ts +10 -0
  59. package/dist/stream.d.ts.map +1 -0
  60. package/dist/{generate.js → stream.js} +3 -65
  61. package/dist/stream.js.map +1 -0
  62. package/dist/types.d.ts +28 -11
  63. package/dist/types.d.ts.map +1 -1
  64. package/dist/types.js.map +1 -1
  65. package/dist/validation.d.ts +10 -0
  66. package/dist/validation.d.ts.map +1 -0
  67. package/dist/validation.js +30 -0
  68. package/dist/validation.js.map +1 -0
  69. package/package.json +7 -5
  70. package/dist/generate.d.ts +0 -22
  71. package/dist/generate.d.ts.map +0 -1
  72. package/dist/generate.js.map +0 -1
  73. package/dist/providers/utils.d.ts.map +0 -1
  74. package/dist/providers/utils.js.map +0 -1
@@ -1044,7 +1044,7 @@ export const MODELS = {
1044
1044
  input: 0.6,
1045
1045
  output: 4,
1046
1046
  cacheRead: 0.15,
1047
- cacheWrite: 4,
1047
+ cacheWrite: 0,
1048
1048
  },
1049
1049
  contextWindow: 131072,
1050
1050
  maxTokens: 8192,
@@ -1061,7 +1061,7 @@ export const MODELS = {
1061
1061
  input: 0.3,
1062
1062
  output: 0.5,
1063
1063
  cacheRead: 0.075,
1064
- cacheWrite: 0.5,
1064
+ cacheWrite: 0,
1065
1065
  },
1066
1066
  contextWindow: 131072,
1067
1067
  maxTokens: 8192,
@@ -1078,7 +1078,7 @@ export const MODELS = {
1078
1078
  input: 5,
1079
1079
  output: 15,
1080
1080
  cacheRead: 5,
1081
- cacheWrite: 15,
1081
+ cacheWrite: 0,
1082
1082
  },
1083
1083
  contextWindow: 131072,
1084
1084
  maxTokens: 4096,
@@ -1095,7 +1095,7 @@ export const MODELS = {
1095
1095
  input: 5,
1096
1096
  output: 25,
1097
1097
  cacheRead: 1.25,
1098
- cacheWrite: 25,
1098
+ cacheWrite: 0,
1099
1099
  },
1100
1100
  contextWindow: 131072,
1101
1101
  maxTokens: 8192,
@@ -1112,7 +1112,7 @@ export const MODELS = {
1112
1112
  input: 3,
1113
1113
  output: 15,
1114
1114
  cacheRead: 0.75,
1115
- cacheWrite: 15,
1115
+ cacheWrite: 0,
1116
1116
  },
1117
1117
  contextWindow: 131072,
1118
1118
  maxTokens: 8192,
@@ -1129,7 +1129,7 @@ export const MODELS = {
1129
1129
  input: 0.3,
1130
1130
  output: 0.5,
1131
1131
  cacheRead: 0.075,
1132
- cacheWrite: 0.5,
1132
+ cacheWrite: 0,
1133
1133
  },
1134
1134
  contextWindow: 131072,
1135
1135
  maxTokens: 8192,
@@ -1146,7 +1146,7 @@ export const MODELS = {
1146
1146
  input: 2,
1147
1147
  output: 10,
1148
1148
  cacheRead: 2,
1149
- cacheWrite: 10,
1149
+ cacheWrite: 0,
1150
1150
  },
1151
1151
  contextWindow: 8192,
1152
1152
  maxTokens: 4096,
@@ -1163,7 +1163,7 @@ export const MODELS = {
1163
1163
  input: 2,
1164
1164
  output: 10,
1165
1165
  cacheRead: 2,
1166
- cacheWrite: 10,
1166
+ cacheWrite: 0,
1167
1167
  },
1168
1168
  contextWindow: 131072,
1169
1169
  maxTokens: 8192,
@@ -1180,7 +1180,7 @@ export const MODELS = {
1180
1180
  input: 2,
1181
1181
  output: 10,
1182
1182
  cacheRead: 2,
1183
- cacheWrite: 10,
1183
+ cacheWrite: 0,
1184
1184
  },
1185
1185
  contextWindow: 8192,
1186
1186
  maxTokens: 4096,
@@ -1197,7 +1197,7 @@ export const MODELS = {
1197
1197
  input: 3,
1198
1198
  output: 15,
1199
1199
  cacheRead: 0.75,
1200
- cacheWrite: 15,
1200
+ cacheWrite: 0,
1201
1201
  },
1202
1202
  contextWindow: 131072,
1203
1203
  maxTokens: 8192,
@@ -1214,7 +1214,7 @@ export const MODELS = {
1214
1214
  input: 2,
1215
1215
  output: 10,
1216
1216
  cacheRead: 2,
1217
- cacheWrite: 10,
1217
+ cacheWrite: 0,
1218
1218
  },
1219
1219
  contextWindow: 8192,
1220
1220
  maxTokens: 4096,
@@ -1231,7 +1231,7 @@ export const MODELS = {
1231
1231
  input: 2,
1232
1232
  output: 10,
1233
1233
  cacheRead: 2,
1234
- cacheWrite: 10,
1234
+ cacheWrite: 0,
1235
1235
  },
1236
1236
  contextWindow: 131072,
1237
1237
  maxTokens: 8192,
@@ -1248,7 +1248,7 @@ export const MODELS = {
1248
1248
  input: 5,
1249
1249
  output: 25,
1250
1250
  cacheRead: 1.25,
1251
- cacheWrite: 25,
1251
+ cacheWrite: 0,
1252
1252
  },
1253
1253
  contextWindow: 131072,
1254
1254
  maxTokens: 8192,
@@ -1265,7 +1265,7 @@ export const MODELS = {
1265
1265
  input: 2,
1266
1266
  output: 10,
1267
1267
  cacheRead: 2,
1268
- cacheWrite: 10,
1268
+ cacheWrite: 0,
1269
1269
  },
1270
1270
  contextWindow: 131072,
1271
1271
  maxTokens: 8192,
@@ -1282,7 +1282,7 @@ export const MODELS = {
1282
1282
  input: 3,
1283
1283
  output: 15,
1284
1284
  cacheRead: 0.75,
1285
- cacheWrite: 15,
1285
+ cacheWrite: 0,
1286
1286
  },
1287
1287
  contextWindow: 256000,
1288
1288
  maxTokens: 64000,
@@ -1299,7 +1299,7 @@ export const MODELS = {
1299
1299
  input: 5,
1300
1300
  output: 15,
1301
1301
  cacheRead: 5,
1302
- cacheWrite: 15,
1302
+ cacheWrite: 0,
1303
1303
  },
1304
1304
  contextWindow: 8192,
1305
1305
  maxTokens: 4096,
@@ -1316,7 +1316,7 @@ export const MODELS = {
1316
1316
  input: 0.6,
1317
1317
  output: 4,
1318
1318
  cacheRead: 0.15,
1319
- cacheWrite: 4,
1319
+ cacheWrite: 0,
1320
1320
  },
1321
1321
  contextWindow: 131072,
1322
1322
  maxTokens: 8192,
@@ -1410,6 +1410,125 @@ export const MODELS = {
1410
1410
  },
1411
1411
  },
1412
1412
  openrouter: {
1413
+ "qwen/qwen3-next-80b-a3b-thinking": {
1414
+ id: "qwen/qwen3-next-80b-a3b-thinking",
1415
+ name: "Qwen: Qwen3 Next 80B A3B Thinking",
1416
+ api: "openai-completions",
1417
+ provider: "openrouter",
1418
+ baseUrl: "https://openrouter.ai/api/v1",
1419
+ reasoning: true,
1420
+ input: ["text"],
1421
+ cost: {
1422
+ input: 0.09782604,
1423
+ output: 0.391304304,
1424
+ cacheRead: 0,
1425
+ cacheWrite: 0,
1426
+ },
1427
+ contextWindow: 262144,
1428
+ maxTokens: 4096,
1429
+ },
1430
+ "qwen/qwen3-next-80b-a3b-instruct": {
1431
+ id: "qwen/qwen3-next-80b-a3b-instruct",
1432
+ name: "Qwen: Qwen3 Next 80B A3B Instruct",
1433
+ api: "openai-completions",
1434
+ provider: "openrouter",
1435
+ baseUrl: "https://openrouter.ai/api/v1",
1436
+ reasoning: false,
1437
+ input: ["text"],
1438
+ cost: {
1439
+ input: 0.09782604,
1440
+ output: 0.391304304,
1441
+ cacheRead: 0,
1442
+ cacheWrite: 0,
1443
+ },
1444
+ contextWindow: 262144,
1445
+ maxTokens: 4096,
1446
+ },
1447
+ "meituan/longcat-flash-chat": {
1448
+ id: "meituan/longcat-flash-chat",
1449
+ name: "Meituan: LongCat Flash Chat",
1450
+ api: "openai-completions",
1451
+ provider: "openrouter",
1452
+ baseUrl: "https://openrouter.ai/api/v1",
1453
+ reasoning: false,
1454
+ input: ["text"],
1455
+ cost: {
1456
+ input: 0.15,
1457
+ output: 0.75,
1458
+ cacheRead: 0,
1459
+ cacheWrite: 0,
1460
+ },
1461
+ contextWindow: 131072,
1462
+ maxTokens: 131072,
1463
+ },
1464
+ "qwen/qwen-plus-2025-07-28": {
1465
+ id: "qwen/qwen-plus-2025-07-28",
1466
+ name: "Qwen: Qwen Plus 0728",
1467
+ api: "openai-completions",
1468
+ provider: "openrouter",
1469
+ baseUrl: "https://openrouter.ai/api/v1",
1470
+ reasoning: false,
1471
+ input: ["text"],
1472
+ cost: {
1473
+ input: 0.39999999999999997,
1474
+ output: 1.2,
1475
+ cacheRead: 0,
1476
+ cacheWrite: 0,
1477
+ },
1478
+ contextWindow: 1000000,
1479
+ maxTokens: 32768,
1480
+ },
1481
+ "qwen/qwen-plus-2025-07-28:thinking": {
1482
+ id: "qwen/qwen-plus-2025-07-28:thinking",
1483
+ name: "Qwen: Qwen Plus 0728 (thinking)",
1484
+ api: "openai-completions",
1485
+ provider: "openrouter",
1486
+ baseUrl: "https://openrouter.ai/api/v1",
1487
+ reasoning: true,
1488
+ input: ["text"],
1489
+ cost: {
1490
+ input: 0.39999999999999997,
1491
+ output: 4,
1492
+ cacheRead: 0,
1493
+ cacheWrite: 0,
1494
+ },
1495
+ contextWindow: 1000000,
1496
+ maxTokens: 32768,
1497
+ },
1498
+ "nvidia/nemotron-nano-9b-v2:free": {
1499
+ id: "nvidia/nemotron-nano-9b-v2:free",
1500
+ name: "NVIDIA: Nemotron Nano 9B V2 (free)",
1501
+ api: "openai-completions",
1502
+ provider: "openrouter",
1503
+ baseUrl: "https://openrouter.ai/api/v1",
1504
+ reasoning: true,
1505
+ input: ["text"],
1506
+ cost: {
1507
+ input: 0,
1508
+ output: 0,
1509
+ cacheRead: 0,
1510
+ cacheWrite: 0,
1511
+ },
1512
+ contextWindow: 128000,
1513
+ maxTokens: 4096,
1514
+ },
1515
+ "nvidia/nemotron-nano-9b-v2": {
1516
+ id: "nvidia/nemotron-nano-9b-v2",
1517
+ name: "NVIDIA: Nemotron Nano 9B V2",
1518
+ api: "openai-completions",
1519
+ provider: "openrouter",
1520
+ baseUrl: "https://openrouter.ai/api/v1",
1521
+ reasoning: true,
1522
+ input: ["text"],
1523
+ cost: {
1524
+ input: 0.04,
1525
+ output: 0.16,
1526
+ cacheRead: 0,
1527
+ cacheWrite: 0,
1528
+ },
1529
+ contextWindow: 131072,
1530
+ maxTokens: 4096,
1531
+ },
1413
1532
  "openrouter/sonoma-dusk-alpha": {
1414
1533
  id: "openrouter/sonoma-dusk-alpha",
1415
1534
  name: "Sonoma Dusk Alpha",
@@ -1470,8 +1589,8 @@ export const MODELS = {
1470
1589
  reasoning: false,
1471
1590
  input: ["text"],
1472
1591
  cost: {
1473
- input: 0.2962,
1474
- output: 1.1852999999999998,
1592
+ input: 0.38043459999999996,
1593
+ output: 1.52173896,
1475
1594
  cacheRead: 0,
1476
1595
  cacheWrite: 0,
1477
1596
  },
@@ -1521,8 +1640,8 @@ export const MODELS = {
1521
1640
  reasoning: true,
1522
1641
  input: ["text"],
1523
1642
  cost: {
1524
- input: 0.0713,
1525
- output: 0.2852,
1643
+ input: 0.08967387,
1644
+ output: 0.358695612,
1526
1645
  cacheRead: 0,
1527
1646
  cacheWrite: 0,
1528
1647
  },
@@ -1538,8 +1657,8 @@ export const MODELS = {
1538
1657
  reasoning: true,
1539
1658
  input: ["text"],
1540
1659
  cost: {
1541
- input: 0.09329544,
1542
- output: 0.3733632,
1660
+ input: 0.127173852,
1661
+ output: 0.5086955952000001,
1543
1662
  cacheRead: 0,
1544
1663
  cacheWrite: 0,
1545
1664
  },
@@ -1555,8 +1674,8 @@ export const MODELS = {
1555
1674
  reasoning: true,
1556
1675
  input: ["text"],
1557
1676
  cost: {
1558
- input: 0.1999188,
1559
- output: 0.800064,
1677
+ input: 0.24999987999999998,
1678
+ output: 0.999999888,
1560
1679
  cacheRead: 0,
1561
1680
  cacheWrite: 0,
1562
1681
  },
@@ -1589,8 +1708,8 @@ export const MODELS = {
1589
1708
  reasoning: true,
1590
1709
  input: ["text"],
1591
1710
  cost: {
1592
- input: 0.19999999999999998,
1593
- output: 0.7999999999999999,
1711
+ input: 0.24999987999999998,
1712
+ output: 0.999999888,
1594
1713
  cacheRead: 0,
1595
1714
  cacheWrite: 0,
1596
1715
  },
@@ -1691,8 +1810,8 @@ export const MODELS = {
1691
1810
  reasoning: false,
1692
1811
  input: ["text"],
1693
1812
  cost: {
1694
- input: 0.051830799999999996,
1695
- output: 0.207424,
1813
+ input: 0.07065213999999999,
1814
+ output: 0.282608664,
1696
1815
  cacheRead: 0,
1697
1816
  cacheWrite: 0,
1698
1817
  },
@@ -1708,8 +1827,8 @@ export const MODELS = {
1708
1827
  reasoning: false,
1709
1828
  input: ["text"],
1710
1829
  cost: {
1711
- input: 0.051830799999999996,
1712
- output: 0.207424,
1830
+ input: 0.07065213999999999,
1831
+ output: 0.282608664,
1713
1832
  cacheRead: 0,
1714
1833
  cacheWrite: 0,
1715
1834
  },
@@ -1725,8 +1844,8 @@ export const MODELS = {
1725
1844
  reasoning: true,
1726
1845
  input: ["text"],
1727
1846
  cost: {
1728
- input: 0.32986602,
1729
- output: 1.3201056,
1847
+ input: 0.41249980199999997,
1848
+ output: 1.6499998152000002,
1730
1849
  cacheRead: 0,
1731
1850
  cacheWrite: 0,
1732
1851
  },
@@ -1776,8 +1895,8 @@ export const MODELS = {
1776
1895
  reasoning: true,
1777
1896
  input: ["text"],
1778
1897
  cost: {
1779
- input: 0.077968332,
1780
- output: 0.31202496,
1898
+ input: 0.0974999532,
1899
+ output: 0.38999995632,
1781
1900
  cacheRead: 0,
1782
1901
  cacheWrite: 0,
1783
1902
  },
@@ -1827,8 +1946,8 @@ export const MODELS = {
1827
1946
  reasoning: false,
1828
1947
  input: ["text"],
1829
1948
  cost: {
1830
- input: 0.19999999999999998,
1831
- output: 0.7999999999999999,
1949
+ input: 0.24999987999999998,
1950
+ output: 0.999999888,
1832
1951
  cacheRead: 0,
1833
1952
  cacheWrite: 0,
1834
1953
  },
@@ -1844,8 +1963,8 @@ export const MODELS = {
1844
1963
  reasoning: false,
1845
1964
  input: ["text"],
1846
1965
  cost: {
1847
- input: 0.077968332,
1848
- output: 0.31202496,
1966
+ input: 0.0974999532,
1967
+ output: 0.38999995632,
1849
1968
  cacheRead: 0,
1850
1969
  cacheWrite: 0,
1851
1970
  },
@@ -2048,8 +2167,8 @@ export const MODELS = {
2048
2167
  reasoning: true,
2049
2168
  input: ["text"],
2050
2169
  cost: {
2051
- input: 0.1999188,
2052
- output: 0.800064,
2170
+ input: 0.24999987999999998,
2171
+ output: 0.999999888,
2053
2172
  cacheRead: 0,
2054
2173
  cacheWrite: 0,
2055
2174
  },
@@ -2082,8 +2201,8 @@ export const MODELS = {
2082
2201
  reasoning: false,
2083
2202
  input: ["text"],
2084
2203
  cost: {
2085
- input: 0.01999188,
2086
- output: 0.0800064,
2204
+ input: 0.035869548,
2205
+ output: 0.14347824480000002,
2087
2206
  cacheRead: 0,
2088
2207
  cacheWrite: 0,
2089
2208
  },
@@ -2184,8 +2303,8 @@ export const MODELS = {
2184
2303
  reasoning: true,
2185
2304
  input: ["text"],
2186
2305
  cost: {
2187
- input: 0.01999188,
2188
- output: 0.0800064,
2306
+ input: 0.035869548,
2307
+ output: 0.14347824480000002,
2189
2308
  cacheRead: 0,
2190
2309
  cacheWrite: 0,
2191
2310
  },
@@ -2218,8 +2337,8 @@ export const MODELS = {
2218
2337
  reasoning: true,
2219
2338
  input: ["text"],
2220
2339
  cost: {
2221
- input: 0.017992691999999998,
2222
- output: 0.07200576,
2340
+ input: 0.0322825932,
2341
+ output: 0.12913042032,
2223
2342
  cacheRead: 0,
2224
2343
  cacheWrite: 0,
2225
2344
  },
@@ -2354,8 +2473,8 @@ export const MODELS = {
2354
2473
  reasoning: false,
2355
2474
  input: ["text"],
2356
2475
  cost: {
2357
- input: 0.1999188,
2358
- output: 0.800064,
2476
+ input: 0.24999987999999998,
2477
+ output: 0.999999888,
2359
2478
  cacheRead: 0,
2360
2479
  cacheWrite: 0,
2361
2480
  },
@@ -2388,8 +2507,8 @@ export const MODELS = {
2388
2507
  reasoning: false,
2389
2508
  input: ["text", "image"],
2390
2509
  cost: {
2391
- input: 0.01999188,
2392
- output: 0.0800064,
2510
+ input: 0.03804346,
2511
+ output: 0.152173896,
2393
2512
  cacheRead: 0,
2394
2513
  cacheWrite: 0,
2395
2514
  },
@@ -2490,8 +2609,8 @@ export const MODELS = {
2490
2609
  reasoning: false,
2491
2610
  input: ["text"],
2492
2611
  cost: {
2493
- input: 0.01999188,
2494
- output: 0.0800064,
2612
+ input: 0.03804346,
2613
+ output: 0.152173896,
2495
2614
  cacheRead: 0,
2496
2615
  cacheWrite: 0,
2497
2616
  },
@@ -2507,8 +2626,8 @@ export const MODELS = {
2507
2626
  reasoning: true,
2508
2627
  input: ["text"],
2509
2628
  cost: {
2510
- input: 0.025915399999999998,
2511
- output: 0.103712,
2629
+ input: 0.03260868,
2630
+ output: 0.130434768,
2512
2631
  cacheRead: 0,
2513
2632
  cacheWrite: 0,
2514
2633
  },
@@ -2558,8 +2677,8 @@ export const MODELS = {
2558
2677
  reasoning: false,
2559
2678
  input: ["text"],
2560
2679
  cost: {
2561
- input: 0.1999188,
2562
- output: 0.800064,
2680
+ input: 0.24999987999999998,
2681
+ output: 0.999999888,
2563
2682
  cacheRead: 0,
2564
2683
  cacheWrite: 0,
2565
2684
  },
@@ -2745,8 +2864,8 @@ export const MODELS = {
2745
2864
  reasoning: false,
2746
2865
  input: ["text"],
2747
2866
  cost: {
2748
- input: 0.12,
2749
- output: 0.3,
2867
+ input: 0.6,
2868
+ output: 0.6,
2750
2869
  cacheRead: 0,
2751
2870
  cacheWrite: 0,
2752
2871
  },
@@ -2796,8 +2915,8 @@ export const MODELS = {
2796
2915
  reasoning: false,
2797
2916
  input: ["text"],
2798
2917
  cost: {
2799
- input: 0.051830799999999996,
2800
- output: 0.207424,
2918
+ input: 0.06521736,
2919
+ output: 0.260869536,
2801
2920
  cacheRead: 0,
2802
2921
  cacheWrite: 0,
2803
2922
  },
@@ -2821,34 +2940,34 @@ export const MODELS = {
2821
2940
  contextWindow: 32768,
2822
2941
  maxTokens: 4096,
2823
2942
  },
2824
- "cohere/command-r-08-2024": {
2825
- id: "cohere/command-r-08-2024",
2826
- name: "Cohere: Command R (08-2024)",
2943
+ "cohere/command-r-plus-08-2024": {
2944
+ id: "cohere/command-r-plus-08-2024",
2945
+ name: "Cohere: Command R+ (08-2024)",
2827
2946
  api: "openai-completions",
2828
2947
  provider: "openrouter",
2829
2948
  baseUrl: "https://openrouter.ai/api/v1",
2830
2949
  reasoning: false,
2831
2950
  input: ["text"],
2832
2951
  cost: {
2833
- input: 0.15,
2834
- output: 0.6,
2952
+ input: 2.5,
2953
+ output: 10,
2835
2954
  cacheRead: 0,
2836
2955
  cacheWrite: 0,
2837
2956
  },
2838
2957
  contextWindow: 128000,
2839
2958
  maxTokens: 4000,
2840
2959
  },
2841
- "cohere/command-r-plus-08-2024": {
2842
- id: "cohere/command-r-plus-08-2024",
2843
- name: "Cohere: Command R+ (08-2024)",
2960
+ "cohere/command-r-08-2024": {
2961
+ id: "cohere/command-r-08-2024",
2962
+ name: "Cohere: Command R (08-2024)",
2844
2963
  api: "openai-completions",
2845
2964
  provider: "openrouter",
2846
2965
  baseUrl: "https://openrouter.ai/api/v1",
2847
2966
  reasoning: false,
2848
2967
  input: ["text"],
2849
2968
  cost: {
2850
- input: 2.5,
2851
- output: 10,
2969
+ input: 0.15,
2970
+ output: 0.6,
2852
2971
  cacheRead: 0,
2853
2972
  cacheWrite: 0,
2854
2973
  },
@@ -2881,14 +3000,31 @@ export const MODELS = {
2881
3000
  reasoning: false,
2882
3001
  input: ["text"],
2883
3002
  cost: {
2884
- input: 0.09999999999999999,
2885
- output: 0.28,
3003
+ input: 0.12,
3004
+ output: 0.3,
2886
3005
  cacheRead: 0,
2887
3006
  cacheWrite: 0,
2888
3007
  },
2889
3008
  contextWindow: 131072,
2890
3009
  maxTokens: 4096,
2891
3010
  },
3011
+ "meta-llama/llama-3.1-8b-instruct": {
3012
+ id: "meta-llama/llama-3.1-8b-instruct",
3013
+ name: "Meta: Llama 3.1 8B Instruct",
3014
+ api: "openai-completions",
3015
+ provider: "openrouter",
3016
+ baseUrl: "https://openrouter.ai/api/v1",
3017
+ reasoning: false,
3018
+ input: ["text"],
3019
+ cost: {
3020
+ input: 0.015,
3021
+ output: 0.02,
3022
+ cacheRead: 0,
3023
+ cacheWrite: 0,
3024
+ },
3025
+ contextWindow: 131072,
3026
+ maxTokens: 16384,
3027
+ },
2892
3028
  "meta-llama/llama-3.1-405b-instruct": {
2893
3029
  id: "meta-llama/llama-3.1-405b-instruct",
2894
3030
  name: "Meta: Llama 3.1 405B Instruct",
@@ -2923,39 +3059,39 @@ export const MODELS = {
2923
3059
  contextWindow: 131072,
2924
3060
  maxTokens: 16384,
2925
3061
  },
2926
- "meta-llama/llama-3.1-8b-instruct": {
2927
- id: "meta-llama/llama-3.1-8b-instruct",
2928
- name: "Meta: Llama 3.1 8B Instruct",
3062
+ "mistralai/mistral-nemo": {
3063
+ id: "mistralai/mistral-nemo",
3064
+ name: "Mistral: Mistral Nemo",
2929
3065
  api: "openai-completions",
2930
3066
  provider: "openrouter",
2931
3067
  baseUrl: "https://openrouter.ai/api/v1",
2932
3068
  reasoning: false,
2933
3069
  input: ["text"],
2934
3070
  cost: {
2935
- input: 0.015,
2936
- output: 0.02,
3071
+ input: 0.017934774,
3072
+ output: 0.07173912240000001,
2937
3073
  cacheRead: 0,
2938
3074
  cacheWrite: 0,
2939
3075
  },
2940
3076
  contextWindow: 131072,
2941
- maxTokens: 16384,
3077
+ maxTokens: 128000,
2942
3078
  },
2943
- "mistralai/mistral-nemo": {
2944
- id: "mistralai/mistral-nemo",
2945
- name: "Mistral: Mistral Nemo",
3079
+ "mistralai/mistral-7b-instruct-v0.3": {
3080
+ id: "mistralai/mistral-7b-instruct-v0.3",
3081
+ name: "Mistral: Mistral 7B Instruct v0.3",
2946
3082
  api: "openai-completions",
2947
3083
  provider: "openrouter",
2948
3084
  baseUrl: "https://openrouter.ai/api/v1",
2949
3085
  reasoning: false,
2950
3086
  input: ["text"],
2951
3087
  cost: {
2952
- input: 0.01,
2953
- output: 0.0400032,
3088
+ input: 0.028,
3089
+ output: 0.054,
2954
3090
  cacheRead: 0,
2955
3091
  cacheWrite: 0,
2956
3092
  },
2957
- contextWindow: 131072,
2958
- maxTokens: 128000,
3093
+ contextWindow: 32768,
3094
+ maxTokens: 16384,
2959
3095
  },
2960
3096
  "mistralai/mistral-7b-instruct:free": {
2961
3097
  id: "mistralai/mistral-7b-instruct:free",
@@ -2991,23 +3127,6 @@ export const MODELS = {
2991
3127
  contextWindow: 32768,
2992
3128
  maxTokens: 16384,
2993
3129
  },
2994
- "mistralai/mistral-7b-instruct-v0.3": {
2995
- id: "mistralai/mistral-7b-instruct-v0.3",
2996
- name: "Mistral: Mistral 7B Instruct v0.3",
2997
- api: "openai-completions",
2998
- provider: "openrouter",
2999
- baseUrl: "https://openrouter.ai/api/v1",
3000
- reasoning: false,
3001
- input: ["text"],
3002
- cost: {
3003
- input: 0.028,
3004
- output: 0.054,
3005
- cacheRead: 0,
3006
- cacheWrite: 0,
3007
- },
3008
- contextWindow: 32768,
3009
- maxTokens: 16384,
3010
- },
3011
3130
  "microsoft/phi-3-mini-128k-instruct": {
3012
3131
  id: "microsoft/phi-3-mini-128k-instruct",
3013
3132
  name: "Microsoft: Phi-3 Mini 128K Instruct",
@@ -3178,34 +3297,34 @@ export const MODELS = {
3178
3297
  contextWindow: 128000,
3179
3298
  maxTokens: 4096,
3180
3299
  },
3181
- "mistralai/mistral-small": {
3182
- id: "mistralai/mistral-small",
3183
- name: "Mistral Small",
3300
+ "mistralai/mistral-tiny": {
3301
+ id: "mistralai/mistral-tiny",
3302
+ name: "Mistral Tiny",
3184
3303
  api: "openai-completions",
3185
3304
  provider: "openrouter",
3186
3305
  baseUrl: "https://openrouter.ai/api/v1",
3187
3306
  reasoning: false,
3188
3307
  input: ["text"],
3189
3308
  cost: {
3190
- input: 0.19999999999999998,
3191
- output: 0.6,
3309
+ input: 0.25,
3310
+ output: 0.25,
3192
3311
  cacheRead: 0,
3193
3312
  cacheWrite: 0,
3194
3313
  },
3195
3314
  contextWindow: 32768,
3196
3315
  maxTokens: 4096,
3197
3316
  },
3198
- "mistralai/mistral-tiny": {
3199
- id: "mistralai/mistral-tiny",
3200
- name: "Mistral Tiny",
3317
+ "mistralai/mistral-small": {
3318
+ id: "mistralai/mistral-small",
3319
+ name: "Mistral Small",
3201
3320
  api: "openai-completions",
3202
3321
  provider: "openrouter",
3203
3322
  baseUrl: "https://openrouter.ai/api/v1",
3204
3323
  reasoning: false,
3205
3324
  input: ["text"],
3206
3325
  cost: {
3207
- input: 0.25,
3208
- output: 0.25,
3326
+ input: 0.19999999999999998,
3327
+ output: 0.6,
3209
3328
  cacheRead: 0,
3210
3329
  cacheWrite: 0,
3211
3330
  },
@@ -3221,8 +3340,8 @@ export const MODELS = {
3221
3340
  reasoning: false,
3222
3341
  input: ["text"],
3223
3342
  cost: {
3224
- input: 0.08,
3225
- output: 0.24,
3343
+ input: 0.39999999999999997,
3344
+ output: 0.39999999999999997,
3226
3345
  cacheRead: 0,
3227
3346
  cacheWrite: 0,
3228
3347
  },