@mariozechner/pi-ai 0.49.3 → 0.50.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (108) hide show
  1. package/README.md +32 -22
  2. package/dist/api-registry.d.ts +20 -0
  3. package/dist/api-registry.d.ts.map +1 -0
  4. package/dist/api-registry.js +44 -0
  5. package/dist/api-registry.js.map +1 -0
  6. package/dist/cli.d.ts.map +1 -1
  7. package/dist/cli.js +22 -67
  8. package/dist/cli.js.map +1 -1
  9. package/dist/env-api-keys.d.ts +9 -0
  10. package/dist/env-api-keys.d.ts.map +1 -0
  11. package/dist/env-api-keys.js +91 -0
  12. package/dist/env-api-keys.js.map +1 -0
  13. package/dist/index.d.ts +4 -0
  14. package/dist/index.d.ts.map +1 -1
  15. package/dist/index.js +4 -0
  16. package/dist/index.js.map +1 -1
  17. package/dist/models.generated.d.ts +583 -105
  18. package/dist/models.generated.d.ts.map +1 -1
  19. package/dist/models.generated.js +605 -127
  20. package/dist/models.generated.js.map +1 -1
  21. package/dist/providers/amazon-bedrock.d.ts +3 -2
  22. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  23. package/dist/providers/amazon-bedrock.js +52 -5
  24. package/dist/providers/amazon-bedrock.js.map +1 -1
  25. package/dist/providers/anthropic.d.ts +3 -2
  26. package/dist/providers/anthropic.d.ts.map +1 -1
  27. package/dist/providers/anthropic.js +20 -2
  28. package/dist/providers/anthropic.js.map +1 -1
  29. package/dist/providers/azure-openai-responses.d.ts +15 -0
  30. package/dist/providers/azure-openai-responses.d.ts.map +1 -0
  31. package/dist/providers/azure-openai-responses.js +184 -0
  32. package/dist/providers/azure-openai-responses.js.map +1 -0
  33. package/dist/providers/google-gemini-cli.d.ts +3 -2
  34. package/dist/providers/google-gemini-cli.d.ts.map +1 -1
  35. package/dist/providers/google-gemini-cli.js +68 -1
  36. package/dist/providers/google-gemini-cli.js.map +1 -1
  37. package/dist/providers/google-vertex.d.ts +3 -2
  38. package/dist/providers/google-vertex.d.ts.map +1 -1
  39. package/dist/providers/google-vertex.js +81 -1
  40. package/dist/providers/google-vertex.js.map +1 -1
  41. package/dist/providers/google.d.ts +3 -2
  42. package/dist/providers/google.d.ts.map +1 -1
  43. package/dist/providers/google.js +84 -3
  44. package/dist/providers/google.js.map +1 -1
  45. package/dist/providers/openai-codex-responses.d.ts +3 -2
  46. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  47. package/dist/providers/openai-codex-responses.js +57 -307
  48. package/dist/providers/openai-codex-responses.js.map +1 -1
  49. package/dist/providers/openai-completions.d.ts +5 -2
  50. package/dist/providers/openai-completions.d.ts.map +1 -1
  51. package/dist/providers/openai-completions.js +78 -41
  52. package/dist/providers/openai-completions.js.map +1 -1
  53. package/dist/providers/openai-responses-shared.d.ts +17 -0
  54. package/dist/providers/openai-responses-shared.d.ts.map +1 -0
  55. package/dist/providers/openai-responses-shared.js +424 -0
  56. package/dist/providers/openai-responses-shared.js.map +1 -0
  57. package/dist/providers/openai-responses.d.ts +3 -2
  58. package/dist/providers/openai-responses.d.ts.map +1 -1
  59. package/dist/providers/openai-responses.js +25 -415
  60. package/dist/providers/openai-responses.js.map +1 -1
  61. package/dist/providers/register-builtins.d.ts +3 -0
  62. package/dist/providers/register-builtins.d.ts.map +1 -0
  63. package/dist/providers/register-builtins.js +63 -0
  64. package/dist/providers/register-builtins.js.map +1 -0
  65. package/dist/providers/simple-options.d.ts +8 -0
  66. package/dist/providers/simple-options.d.ts.map +1 -0
  67. package/dist/providers/simple-options.js +32 -0
  68. package/dist/providers/simple-options.js.map +1 -0
  69. package/dist/stream.d.ts +5 -10
  70. package/dist/stream.d.ts.map +1 -1
  71. package/dist/stream.js +17 -420
  72. package/dist/stream.js.map +1 -1
  73. package/dist/types.d.ts +18 -22
  74. package/dist/types.d.ts.map +1 -1
  75. package/dist/types.js +0 -1
  76. package/dist/types.js.map +1 -1
  77. package/dist/utils/event-stream.d.ts +2 -0
  78. package/dist/utils/event-stream.d.ts.map +1 -1
  79. package/dist/utils/event-stream.js +4 -0
  80. package/dist/utils/event-stream.js.map +1 -1
  81. package/dist/utils/oauth/anthropic.d.ts +2 -1
  82. package/dist/utils/oauth/anthropic.d.ts.map +1 -1
  83. package/dist/utils/oauth/anthropic.js +13 -0
  84. package/dist/utils/oauth/anthropic.js.map +1 -1
  85. package/dist/utils/oauth/github-copilot.d.ts +2 -1
  86. package/dist/utils/oauth/github-copilot.d.ts.map +1 -1
  87. package/dist/utils/oauth/github-copilot.js +25 -0
  88. package/dist/utils/oauth/github-copilot.js.map +1 -1
  89. package/dist/utils/oauth/google-antigravity.d.ts +2 -1
  90. package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
  91. package/dist/utils/oauth/google-antigravity.js +19 -0
  92. package/dist/utils/oauth/google-antigravity.js.map +1 -1
  93. package/dist/utils/oauth/google-gemini-cli.d.ts +2 -1
  94. package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
  95. package/dist/utils/oauth/google-gemini-cli.js +19 -0
  96. package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
  97. package/dist/utils/oauth/index.d.ts +26 -16
  98. package/dist/utils/oauth/index.d.ts.map +1 -1
  99. package/dist/utils/oauth/index.js +65 -84
  100. package/dist/utils/oauth/index.js.map +1 -1
  101. package/dist/utils/oauth/openai-codex.d.ts +2 -1
  102. package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
  103. package/dist/utils/oauth/openai-codex.js +20 -1
  104. package/dist/utils/oauth/openai-codex.js.map +1 -1
  105. package/dist/utils/oauth/types.d.ts +28 -6
  106. package/dist/utils/oauth/types.d.ts.map +1 -1
  107. package/dist/utils/oauth/types.js.map +1 -1
  108. package/package.json +3 -1
@@ -1297,6 +1297,586 @@ export const MODELS = {
1297
1297
  maxTokens: 64000,
1298
1298
  },
1299
1299
  },
1300
+ "azure-openai-responses": {
1301
+ "codex-mini-latest": {
1302
+ id: "codex-mini-latest",
1303
+ name: "Codex Mini",
1304
+ api: "azure-openai-responses",
1305
+ provider: "azure-openai-responses",
1306
+ baseUrl: "",
1307
+ reasoning: true,
1308
+ input: ["text"],
1309
+ cost: {
1310
+ input: 1.5,
1311
+ output: 6,
1312
+ cacheRead: 0.375,
1313
+ cacheWrite: 0,
1314
+ },
1315
+ contextWindow: 200000,
1316
+ maxTokens: 100000,
1317
+ },
1318
+ "gpt-4": {
1319
+ id: "gpt-4",
1320
+ name: "GPT-4",
1321
+ api: "azure-openai-responses",
1322
+ provider: "azure-openai-responses",
1323
+ baseUrl: "",
1324
+ reasoning: false,
1325
+ input: ["text"],
1326
+ cost: {
1327
+ input: 30,
1328
+ output: 60,
1329
+ cacheRead: 0,
1330
+ cacheWrite: 0,
1331
+ },
1332
+ contextWindow: 8192,
1333
+ maxTokens: 8192,
1334
+ },
1335
+ "gpt-4-turbo": {
1336
+ id: "gpt-4-turbo",
1337
+ name: "GPT-4 Turbo",
1338
+ api: "azure-openai-responses",
1339
+ provider: "azure-openai-responses",
1340
+ baseUrl: "",
1341
+ reasoning: false,
1342
+ input: ["text", "image"],
1343
+ cost: {
1344
+ input: 10,
1345
+ output: 30,
1346
+ cacheRead: 0,
1347
+ cacheWrite: 0,
1348
+ },
1349
+ contextWindow: 128000,
1350
+ maxTokens: 4096,
1351
+ },
1352
+ "gpt-4.1": {
1353
+ id: "gpt-4.1",
1354
+ name: "GPT-4.1",
1355
+ api: "azure-openai-responses",
1356
+ provider: "azure-openai-responses",
1357
+ baseUrl: "",
1358
+ reasoning: false,
1359
+ input: ["text", "image"],
1360
+ cost: {
1361
+ input: 2,
1362
+ output: 8,
1363
+ cacheRead: 0.5,
1364
+ cacheWrite: 0,
1365
+ },
1366
+ contextWindow: 1047576,
1367
+ maxTokens: 32768,
1368
+ },
1369
+ "gpt-4.1-mini": {
1370
+ id: "gpt-4.1-mini",
1371
+ name: "GPT-4.1 mini",
1372
+ api: "azure-openai-responses",
1373
+ provider: "azure-openai-responses",
1374
+ baseUrl: "",
1375
+ reasoning: false,
1376
+ input: ["text", "image"],
1377
+ cost: {
1378
+ input: 0.4,
1379
+ output: 1.6,
1380
+ cacheRead: 0.1,
1381
+ cacheWrite: 0,
1382
+ },
1383
+ contextWindow: 1047576,
1384
+ maxTokens: 32768,
1385
+ },
1386
+ "gpt-4.1-nano": {
1387
+ id: "gpt-4.1-nano",
1388
+ name: "GPT-4.1 nano",
1389
+ api: "azure-openai-responses",
1390
+ provider: "azure-openai-responses",
1391
+ baseUrl: "",
1392
+ reasoning: false,
1393
+ input: ["text", "image"],
1394
+ cost: {
1395
+ input: 0.1,
1396
+ output: 0.4,
1397
+ cacheRead: 0.03,
1398
+ cacheWrite: 0,
1399
+ },
1400
+ contextWindow: 1047576,
1401
+ maxTokens: 32768,
1402
+ },
1403
+ "gpt-4o": {
1404
+ id: "gpt-4o",
1405
+ name: "GPT-4o",
1406
+ api: "azure-openai-responses",
1407
+ provider: "azure-openai-responses",
1408
+ baseUrl: "",
1409
+ reasoning: false,
1410
+ input: ["text", "image"],
1411
+ cost: {
1412
+ input: 2.5,
1413
+ output: 10,
1414
+ cacheRead: 1.25,
1415
+ cacheWrite: 0,
1416
+ },
1417
+ contextWindow: 128000,
1418
+ maxTokens: 16384,
1419
+ },
1420
+ "gpt-4o-2024-05-13": {
1421
+ id: "gpt-4o-2024-05-13",
1422
+ name: "GPT-4o (2024-05-13)",
1423
+ api: "azure-openai-responses",
1424
+ provider: "azure-openai-responses",
1425
+ baseUrl: "",
1426
+ reasoning: false,
1427
+ input: ["text", "image"],
1428
+ cost: {
1429
+ input: 5,
1430
+ output: 15,
1431
+ cacheRead: 0,
1432
+ cacheWrite: 0,
1433
+ },
1434
+ contextWindow: 128000,
1435
+ maxTokens: 4096,
1436
+ },
1437
+ "gpt-4o-2024-08-06": {
1438
+ id: "gpt-4o-2024-08-06",
1439
+ name: "GPT-4o (2024-08-06)",
1440
+ api: "azure-openai-responses",
1441
+ provider: "azure-openai-responses",
1442
+ baseUrl: "",
1443
+ reasoning: false,
1444
+ input: ["text", "image"],
1445
+ cost: {
1446
+ input: 2.5,
1447
+ output: 10,
1448
+ cacheRead: 1.25,
1449
+ cacheWrite: 0,
1450
+ },
1451
+ contextWindow: 128000,
1452
+ maxTokens: 16384,
1453
+ },
1454
+ "gpt-4o-2024-11-20": {
1455
+ id: "gpt-4o-2024-11-20",
1456
+ name: "GPT-4o (2024-11-20)",
1457
+ api: "azure-openai-responses",
1458
+ provider: "azure-openai-responses",
1459
+ baseUrl: "",
1460
+ reasoning: false,
1461
+ input: ["text", "image"],
1462
+ cost: {
1463
+ input: 2.5,
1464
+ output: 10,
1465
+ cacheRead: 1.25,
1466
+ cacheWrite: 0,
1467
+ },
1468
+ contextWindow: 128000,
1469
+ maxTokens: 16384,
1470
+ },
1471
+ "gpt-4o-mini": {
1472
+ id: "gpt-4o-mini",
1473
+ name: "GPT-4o mini",
1474
+ api: "azure-openai-responses",
1475
+ provider: "azure-openai-responses",
1476
+ baseUrl: "",
1477
+ reasoning: false,
1478
+ input: ["text", "image"],
1479
+ cost: {
1480
+ input: 0.15,
1481
+ output: 0.6,
1482
+ cacheRead: 0.08,
1483
+ cacheWrite: 0,
1484
+ },
1485
+ contextWindow: 128000,
1486
+ maxTokens: 16384,
1487
+ },
1488
+ "gpt-5": {
1489
+ id: "gpt-5",
1490
+ name: "GPT-5",
1491
+ api: "azure-openai-responses",
1492
+ provider: "azure-openai-responses",
1493
+ baseUrl: "",
1494
+ reasoning: true,
1495
+ input: ["text", "image"],
1496
+ cost: {
1497
+ input: 1.25,
1498
+ output: 10,
1499
+ cacheRead: 0.125,
1500
+ cacheWrite: 0,
1501
+ },
1502
+ contextWindow: 400000,
1503
+ maxTokens: 128000,
1504
+ },
1505
+ "gpt-5-chat-latest": {
1506
+ id: "gpt-5-chat-latest",
1507
+ name: "GPT-5 Chat Latest",
1508
+ api: "azure-openai-responses",
1509
+ provider: "azure-openai-responses",
1510
+ baseUrl: "",
1511
+ reasoning: false,
1512
+ input: ["text", "image"],
1513
+ cost: {
1514
+ input: 1.25,
1515
+ output: 10,
1516
+ cacheRead: 0.125,
1517
+ cacheWrite: 0,
1518
+ },
1519
+ contextWindow: 128000,
1520
+ maxTokens: 16384,
1521
+ },
1522
+ "gpt-5-codex": {
1523
+ id: "gpt-5-codex",
1524
+ name: "GPT-5-Codex",
1525
+ api: "azure-openai-responses",
1526
+ provider: "azure-openai-responses",
1527
+ baseUrl: "",
1528
+ reasoning: true,
1529
+ input: ["text", "image"],
1530
+ cost: {
1531
+ input: 1.25,
1532
+ output: 10,
1533
+ cacheRead: 0.125,
1534
+ cacheWrite: 0,
1535
+ },
1536
+ contextWindow: 400000,
1537
+ maxTokens: 128000,
1538
+ },
1539
+ "gpt-5-mini": {
1540
+ id: "gpt-5-mini",
1541
+ name: "GPT-5 Mini",
1542
+ api: "azure-openai-responses",
1543
+ provider: "azure-openai-responses",
1544
+ baseUrl: "",
1545
+ reasoning: true,
1546
+ input: ["text", "image"],
1547
+ cost: {
1548
+ input: 0.25,
1549
+ output: 2,
1550
+ cacheRead: 0.025,
1551
+ cacheWrite: 0,
1552
+ },
1553
+ contextWindow: 400000,
1554
+ maxTokens: 128000,
1555
+ },
1556
+ "gpt-5-nano": {
1557
+ id: "gpt-5-nano",
1558
+ name: "GPT-5 Nano",
1559
+ api: "azure-openai-responses",
1560
+ provider: "azure-openai-responses",
1561
+ baseUrl: "",
1562
+ reasoning: true,
1563
+ input: ["text", "image"],
1564
+ cost: {
1565
+ input: 0.05,
1566
+ output: 0.4,
1567
+ cacheRead: 0.005,
1568
+ cacheWrite: 0,
1569
+ },
1570
+ contextWindow: 400000,
1571
+ maxTokens: 128000,
1572
+ },
1573
+ "gpt-5-pro": {
1574
+ id: "gpt-5-pro",
1575
+ name: "GPT-5 Pro",
1576
+ api: "azure-openai-responses",
1577
+ provider: "azure-openai-responses",
1578
+ baseUrl: "",
1579
+ reasoning: true,
1580
+ input: ["text", "image"],
1581
+ cost: {
1582
+ input: 15,
1583
+ output: 120,
1584
+ cacheRead: 0,
1585
+ cacheWrite: 0,
1586
+ },
1587
+ contextWindow: 400000,
1588
+ maxTokens: 272000,
1589
+ },
1590
+ "gpt-5.1": {
1591
+ id: "gpt-5.1",
1592
+ name: "GPT-5.1",
1593
+ api: "azure-openai-responses",
1594
+ provider: "azure-openai-responses",
1595
+ baseUrl: "",
1596
+ reasoning: true,
1597
+ input: ["text", "image"],
1598
+ cost: {
1599
+ input: 1.25,
1600
+ output: 10,
1601
+ cacheRead: 0.13,
1602
+ cacheWrite: 0,
1603
+ },
1604
+ contextWindow: 400000,
1605
+ maxTokens: 128000,
1606
+ },
1607
+ "gpt-5.1-chat-latest": {
1608
+ id: "gpt-5.1-chat-latest",
1609
+ name: "GPT-5.1 Chat",
1610
+ api: "azure-openai-responses",
1611
+ provider: "azure-openai-responses",
1612
+ baseUrl: "",
1613
+ reasoning: true,
1614
+ input: ["text", "image"],
1615
+ cost: {
1616
+ input: 1.25,
1617
+ output: 10,
1618
+ cacheRead: 0.125,
1619
+ cacheWrite: 0,
1620
+ },
1621
+ contextWindow: 128000,
1622
+ maxTokens: 16384,
1623
+ },
1624
+ "gpt-5.1-codex": {
1625
+ id: "gpt-5.1-codex",
1626
+ name: "GPT-5.1 Codex",
1627
+ api: "azure-openai-responses",
1628
+ provider: "azure-openai-responses",
1629
+ baseUrl: "",
1630
+ reasoning: true,
1631
+ input: ["text", "image"],
1632
+ cost: {
1633
+ input: 1.25,
1634
+ output: 10,
1635
+ cacheRead: 0.125,
1636
+ cacheWrite: 0,
1637
+ },
1638
+ contextWindow: 400000,
1639
+ maxTokens: 128000,
1640
+ },
1641
+ "gpt-5.1-codex-max": {
1642
+ id: "gpt-5.1-codex-max",
1643
+ name: "GPT-5.1 Codex Max",
1644
+ api: "azure-openai-responses",
1645
+ provider: "azure-openai-responses",
1646
+ baseUrl: "",
1647
+ reasoning: true,
1648
+ input: ["text", "image"],
1649
+ cost: {
1650
+ input: 1.25,
1651
+ output: 10,
1652
+ cacheRead: 0.125,
1653
+ cacheWrite: 0,
1654
+ },
1655
+ contextWindow: 400000,
1656
+ maxTokens: 128000,
1657
+ },
1658
+ "gpt-5.1-codex-mini": {
1659
+ id: "gpt-5.1-codex-mini",
1660
+ name: "GPT-5.1 Codex mini",
1661
+ api: "azure-openai-responses",
1662
+ provider: "azure-openai-responses",
1663
+ baseUrl: "",
1664
+ reasoning: true,
1665
+ input: ["text", "image"],
1666
+ cost: {
1667
+ input: 0.25,
1668
+ output: 2,
1669
+ cacheRead: 0.025,
1670
+ cacheWrite: 0,
1671
+ },
1672
+ contextWindow: 400000,
1673
+ maxTokens: 128000,
1674
+ },
1675
+ "gpt-5.2": {
1676
+ id: "gpt-5.2",
1677
+ name: "GPT-5.2",
1678
+ api: "azure-openai-responses",
1679
+ provider: "azure-openai-responses",
1680
+ baseUrl: "",
1681
+ reasoning: true,
1682
+ input: ["text", "image"],
1683
+ cost: {
1684
+ input: 1.75,
1685
+ output: 14,
1686
+ cacheRead: 0.175,
1687
+ cacheWrite: 0,
1688
+ },
1689
+ contextWindow: 400000,
1690
+ maxTokens: 128000,
1691
+ },
1692
+ "gpt-5.2-chat-latest": {
1693
+ id: "gpt-5.2-chat-latest",
1694
+ name: "GPT-5.2 Chat",
1695
+ api: "azure-openai-responses",
1696
+ provider: "azure-openai-responses",
1697
+ baseUrl: "",
1698
+ reasoning: true,
1699
+ input: ["text", "image"],
1700
+ cost: {
1701
+ input: 1.75,
1702
+ output: 14,
1703
+ cacheRead: 0.175,
1704
+ cacheWrite: 0,
1705
+ },
1706
+ contextWindow: 128000,
1707
+ maxTokens: 16384,
1708
+ },
1709
+ "gpt-5.2-codex": {
1710
+ id: "gpt-5.2-codex",
1711
+ name: "GPT-5.2 Codex",
1712
+ api: "azure-openai-responses",
1713
+ provider: "azure-openai-responses",
1714
+ baseUrl: "",
1715
+ reasoning: true,
1716
+ input: ["text", "image"],
1717
+ cost: {
1718
+ input: 1.75,
1719
+ output: 14,
1720
+ cacheRead: 0.175,
1721
+ cacheWrite: 0,
1722
+ },
1723
+ contextWindow: 400000,
1724
+ maxTokens: 128000,
1725
+ },
1726
+ "gpt-5.2-pro": {
1727
+ id: "gpt-5.2-pro",
1728
+ name: "GPT-5.2 Pro",
1729
+ api: "azure-openai-responses",
1730
+ provider: "azure-openai-responses",
1731
+ baseUrl: "",
1732
+ reasoning: true,
1733
+ input: ["text", "image"],
1734
+ cost: {
1735
+ input: 21,
1736
+ output: 168,
1737
+ cacheRead: 0,
1738
+ cacheWrite: 0,
1739
+ },
1740
+ contextWindow: 400000,
1741
+ maxTokens: 128000,
1742
+ },
1743
+ "o1": {
1744
+ id: "o1",
1745
+ name: "o1",
1746
+ api: "azure-openai-responses",
1747
+ provider: "azure-openai-responses",
1748
+ baseUrl: "",
1749
+ reasoning: true,
1750
+ input: ["text", "image"],
1751
+ cost: {
1752
+ input: 15,
1753
+ output: 60,
1754
+ cacheRead: 7.5,
1755
+ cacheWrite: 0,
1756
+ },
1757
+ contextWindow: 200000,
1758
+ maxTokens: 100000,
1759
+ },
1760
+ "o1-pro": {
1761
+ id: "o1-pro",
1762
+ name: "o1-pro",
1763
+ api: "azure-openai-responses",
1764
+ provider: "azure-openai-responses",
1765
+ baseUrl: "",
1766
+ reasoning: true,
1767
+ input: ["text", "image"],
1768
+ cost: {
1769
+ input: 150,
1770
+ output: 600,
1771
+ cacheRead: 0,
1772
+ cacheWrite: 0,
1773
+ },
1774
+ contextWindow: 200000,
1775
+ maxTokens: 100000,
1776
+ },
1777
+ "o3": {
1778
+ id: "o3",
1779
+ name: "o3",
1780
+ api: "azure-openai-responses",
1781
+ provider: "azure-openai-responses",
1782
+ baseUrl: "",
1783
+ reasoning: true,
1784
+ input: ["text", "image"],
1785
+ cost: {
1786
+ input: 2,
1787
+ output: 8,
1788
+ cacheRead: 0.5,
1789
+ cacheWrite: 0,
1790
+ },
1791
+ contextWindow: 200000,
1792
+ maxTokens: 100000,
1793
+ },
1794
+ "o3-deep-research": {
1795
+ id: "o3-deep-research",
1796
+ name: "o3-deep-research",
1797
+ api: "azure-openai-responses",
1798
+ provider: "azure-openai-responses",
1799
+ baseUrl: "",
1800
+ reasoning: true,
1801
+ input: ["text", "image"],
1802
+ cost: {
1803
+ input: 10,
1804
+ output: 40,
1805
+ cacheRead: 2.5,
1806
+ cacheWrite: 0,
1807
+ },
1808
+ contextWindow: 200000,
1809
+ maxTokens: 100000,
1810
+ },
1811
+ "o3-mini": {
1812
+ id: "o3-mini",
1813
+ name: "o3-mini",
1814
+ api: "azure-openai-responses",
1815
+ provider: "azure-openai-responses",
1816
+ baseUrl: "",
1817
+ reasoning: true,
1818
+ input: ["text"],
1819
+ cost: {
1820
+ input: 1.1,
1821
+ output: 4.4,
1822
+ cacheRead: 0.55,
1823
+ cacheWrite: 0,
1824
+ },
1825
+ contextWindow: 200000,
1826
+ maxTokens: 100000,
1827
+ },
1828
+ "o3-pro": {
1829
+ id: "o3-pro",
1830
+ name: "o3-pro",
1831
+ api: "azure-openai-responses",
1832
+ provider: "azure-openai-responses",
1833
+ baseUrl: "",
1834
+ reasoning: true,
1835
+ input: ["text", "image"],
1836
+ cost: {
1837
+ input: 20,
1838
+ output: 80,
1839
+ cacheRead: 0,
1840
+ cacheWrite: 0,
1841
+ },
1842
+ contextWindow: 200000,
1843
+ maxTokens: 100000,
1844
+ },
1845
+ "o4-mini": {
1846
+ id: "o4-mini",
1847
+ name: "o4-mini",
1848
+ api: "azure-openai-responses",
1849
+ provider: "azure-openai-responses",
1850
+ baseUrl: "",
1851
+ reasoning: true,
1852
+ input: ["text", "image"],
1853
+ cost: {
1854
+ input: 1.1,
1855
+ output: 4.4,
1856
+ cacheRead: 0.28,
1857
+ cacheWrite: 0,
1858
+ },
1859
+ contextWindow: 200000,
1860
+ maxTokens: 100000,
1861
+ },
1862
+ "o4-mini-deep-research": {
1863
+ id: "o4-mini-deep-research",
1864
+ name: "o4-mini-deep-research",
1865
+ api: "azure-openai-responses",
1866
+ provider: "azure-openai-responses",
1867
+ baseUrl: "",
1868
+ reasoning: true,
1869
+ input: ["text", "image"],
1870
+ cost: {
1871
+ input: 2,
1872
+ output: 8,
1873
+ cacheRead: 0.5,
1874
+ cacheWrite: 0,
1875
+ },
1876
+ contextWindow: 200000,
1877
+ maxTokens: 100000,
1878
+ },
1879
+ },
1300
1880
  "cerebras": {
1301
1881
  "gpt-oss-120b": {
1302
1882
  id: "gpt-oss-120b",
@@ -1500,7 +2080,7 @@ export const MODELS = {
1500
2080
  cacheRead: 0,
1501
2081
  cacheWrite: 0,
1502
2082
  },
1503
- contextWindow: 128000,
2083
+ contextWindow: 64000,
1504
2084
  maxTokens: 16384,
1505
2085
  },
1506
2086
  "gpt-4o": {
@@ -3866,40 +4446,6 @@ export const MODELS = {
3866
4446
  },
3867
4447
  },
3868
4448
  "opencode": {
3869
- "alpha-gd4": {
3870
- id: "alpha-gd4",
3871
- name: "Alpha GD4",
3872
- api: "anthropic-messages",
3873
- provider: "opencode",
3874
- baseUrl: "https://opencode.ai/zen",
3875
- reasoning: true,
3876
- input: ["text"],
3877
- cost: {
3878
- input: 0.5,
3879
- output: 2,
3880
- cacheRead: 0.15,
3881
- cacheWrite: 0,
3882
- },
3883
- contextWindow: 262144,
3884
- maxTokens: 32768,
3885
- },
3886
- "alpha-glm-4.7": {
3887
- id: "alpha-glm-4.7",
3888
- name: "Alpha GLM-4.7",
3889
- api: "openai-completions",
3890
- provider: "opencode",
3891
- baseUrl: "https://opencode.ai/zen/v1",
3892
- reasoning: true,
3893
- input: ["text"],
3894
- cost: {
3895
- input: 0.6,
3896
- output: 2.2,
3897
- cacheRead: 0.6,
3898
- cacheWrite: 0,
3899
- },
3900
- contextWindow: 204800,
3901
- maxTokens: 131072,
3902
- },
3903
4449
  "big-pickle": {
3904
4450
  id: "big-pickle",
3905
4451
  name: "Big Pickle",
@@ -4070,8 +4616,8 @@ export const MODELS = {
4070
4616
  contextWindow: 204800,
4071
4617
  maxTokens: 131072,
4072
4618
  },
4073
- "glm-4.7-free": {
4074
- id: "glm-4.7-free",
4619
+ "glm-4.7": {
4620
+ id: "glm-4.7",
4075
4621
  name: "GLM-4.7",
4076
4622
  api: "openai-completions",
4077
4623
  provider: "opencode",
@@ -4079,9 +4625,9 @@ export const MODELS = {
4079
4625
  reasoning: true,
4080
4626
  input: ["text"],
4081
4627
  cost: {
4082
- input: 0,
4083
- output: 0,
4084
- cacheRead: 0,
4628
+ input: 0.6,
4629
+ output: 2.2,
4630
+ cacheRead: 0.1,
4085
4631
  cacheWrite: 0,
4086
4632
  },
4087
4633
  contextWindow: 204800,
@@ -4240,23 +4786,6 @@ export const MODELS = {
4240
4786
  contextWindow: 400000,
4241
4787
  maxTokens: 128000,
4242
4788
  },
4243
- "grok-code": {
4244
- id: "grok-code",
4245
- name: "Grok Code Fast 1",
4246
- api: "openai-completions",
4247
- provider: "opencode",
4248
- baseUrl: "https://opencode.ai/zen/v1",
4249
- reasoning: true,
4250
- input: ["text"],
4251
- cost: {
4252
- input: 0,
4253
- output: 0,
4254
- cacheRead: 0,
4255
- cacheWrite: 0,
4256
- },
4257
- contextWindow: 256000,
4258
- maxTokens: 256000,
4259
- },
4260
4789
  "kimi-k2": {
4261
4790
  id: "kimi-k2",
4262
4791
  name: "Kimi K2",
@@ -4291,23 +4820,6 @@ export const MODELS = {
4291
4820
  contextWindow: 262144,
4292
4821
  maxTokens: 262144,
4293
4822
  },
4294
- "minimax-m2.1-free": {
4295
- id: "minimax-m2.1-free",
4296
- name: "MiniMax M2.1",
4297
- api: "anthropic-messages",
4298
- provider: "opencode",
4299
- baseUrl: "https://opencode.ai/zen",
4300
- reasoning: true,
4301
- input: ["text"],
4302
- cost: {
4303
- input: 0,
4304
- output: 0,
4305
- cacheRead: 0,
4306
- cacheWrite: 0,
4307
- },
4308
- contextWindow: 204800,
4309
- maxTokens: 131072,
4310
- },
4311
4823
  "qwen3-coder": {
4312
4824
  id: "qwen3-coder",
4313
4825
  name: "Qwen3 Coder",
@@ -5053,7 +5565,7 @@ export const MODELS = {
5053
5565
  input: 0.09999999999999999,
5054
5566
  output: 0.39999999999999997,
5055
5567
  cacheRead: 0.024999999999999998,
5056
- cacheWrite: 0.0833,
5568
+ cacheWrite: 0.08333333333333334,
5057
5569
  },
5058
5570
  contextWindow: 1048576,
5059
5571
  maxTokens: 8192,
@@ -5121,7 +5633,7 @@ export const MODELS = {
5121
5633
  input: 0.09999999999999999,
5122
5634
  output: 0.39999999999999997,
5123
5635
  cacheRead: 0.01,
5124
- cacheWrite: 0.0833,
5636
+ cacheWrite: 0.08333333333333334,
5125
5637
  },
5126
5638
  contextWindow: 1048576,
5127
5639
  maxTokens: 65535,
@@ -5138,7 +5650,7 @@ export const MODELS = {
5138
5650
  input: 0.09999999999999999,
5139
5651
  output: 0.39999999999999997,
5140
5652
  cacheRead: 0.01,
5141
- cacheWrite: 0.0833,
5653
+ cacheWrite: 0.08333333333333334,
5142
5654
  },
5143
5655
  contextWindow: 1048576,
5144
5656
  maxTokens: 65535,
@@ -5155,7 +5667,7 @@ export const MODELS = {
5155
5667
  input: 0.3,
5156
5668
  output: 2.5,
5157
5669
  cacheRead: 0.03,
5158
- cacheWrite: 0.0833,
5670
+ cacheWrite: 0.08333333333333334,
5159
5671
  },
5160
5672
  contextWindow: 1048576,
5161
5673
  maxTokens: 65535,
@@ -5188,7 +5700,7 @@ export const MODELS = {
5188
5700
  cost: {
5189
5701
  input: 1.25,
5190
5702
  output: 10,
5191
- cacheRead: 0.31,
5703
+ cacheRead: 0.125,
5192
5704
  cacheWrite: 0.375,
5193
5705
  },
5194
5706
  contextWindow: 1048576,
@@ -5205,7 +5717,7 @@ export const MODELS = {
5205
5717
  cost: {
5206
5718
  input: 1.25,
5207
5719
  output: 10,
5208
- cacheRead: 0.31,
5720
+ cacheRead: 0.125,
5209
5721
  cacheWrite: 0.375,
5210
5722
  },
5211
5723
  contextWindow: 1048576,
@@ -5223,7 +5735,7 @@ export const MODELS = {
5223
5735
  input: 0.5,
5224
5736
  output: 3,
5225
5737
  cacheRead: 0.049999999999999996,
5226
- cacheWrite: 0,
5738
+ cacheWrite: 0.08333333333333334,
5227
5739
  },
5228
5740
  contextWindow: 1048576,
5229
5741
  maxTokens: 65535,
@@ -5510,12 +6022,12 @@ export const MODELS = {
5510
6022
  input: ["text"],
5511
6023
  cost: {
5512
6024
  input: 0.27,
5513
- output: 1.12,
6025
+ output: 1.1,
5514
6026
  cacheRead: 0,
5515
6027
  cacheWrite: 0,
5516
6028
  },
5517
6029
  contextWindow: 196608,
5518
- maxTokens: 65536,
6030
+ maxTokens: 196608,
5519
6031
  },
5520
6032
  "mistralai/codestral-2508": {
5521
6033
  id: "mistralai/codestral-2508",
@@ -7268,23 +7780,6 @@ export const MODELS = {
7268
7780
  contextWindow: 131072,
7269
7781
  maxTokens: 8192,
7270
7782
  },
7271
- "qwen/qwen2.5-vl-72b-instruct": {
7272
- id: "qwen/qwen2.5-vl-72b-instruct",
7273
- name: "Qwen: Qwen2.5 VL 72B Instruct",
7274
- api: "openai-completions",
7275
- provider: "openrouter",
7276
- baseUrl: "https://openrouter.ai/api/v1",
7277
- reasoning: false,
7278
- input: ["text", "image"],
7279
- cost: {
7280
- input: 0.15,
7281
- output: 0.6,
7282
- cacheRead: 0,
7283
- cacheWrite: 0,
7284
- },
7285
- contextWindow: 32768,
7286
- maxTokens: 32768,
7287
- },
7288
7783
  "qwen/qwen3-14b": {
7289
7784
  id: "qwen/qwen3-14b",
7290
7785
  name: "Qwen: Qwen3 14B",
@@ -7534,7 +8029,7 @@ export const MODELS = {
7534
8029
  cost: {
7535
8030
  input: 0.22,
7536
8031
  output: 1.7999999999999998,
7537
- cacheRead: 0,
8032
+ cacheRead: 0.022,
7538
8033
  cacheWrite: 0,
7539
8034
  },
7540
8035
  contextWindow: 262144,
@@ -7670,7 +8165,7 @@ export const MODELS = {
7670
8165
  cost: {
7671
8166
  input: 0.15,
7672
8167
  output: 0.6,
7673
- cacheRead: 0,
8168
+ cacheRead: 0.075,
7674
8169
  cacheWrite: 0,
7675
8170
  },
7676
8171
  contextWindow: 262144,
@@ -8050,23 +8545,6 @@ export const MODELS = {
8050
8545
  contextWindow: 262144,
8051
8546
  maxTokens: 4096,
8052
8547
  },
8053
- "xiaomi/mimo-v2-flash:free": {
8054
- id: "xiaomi/mimo-v2-flash:free",
8055
- name: "Xiaomi: MiMo-V2-Flash (free)",
8056
- api: "openai-completions",
8057
- provider: "openrouter",
8058
- baseUrl: "https://openrouter.ai/api/v1",
8059
- reasoning: true,
8060
- input: ["text"],
8061
- cost: {
8062
- input: 0,
8063
- output: 0,
8064
- cacheRead: 0,
8065
- cacheWrite: 0,
8066
- },
8067
- contextWindow: 262144,
8068
- maxTokens: 65536,
8069
- },
8070
8548
  "z-ai/glm-4-32b": {
8071
8549
  id: "z-ai/glm-4-32b",
8072
8550
  name: "Z.AI: GLM 4 32B ",
@@ -8180,7 +8658,7 @@ export const MODELS = {
8180
8658
  cost: {
8181
8659
  input: 0.44,
8182
8660
  output: 1.76,
8183
- cacheRead: 0,
8661
+ cacheRead: 0.11,
8184
8662
  cacheWrite: 0,
8185
8663
  },
8186
8664
  contextWindow: 204800,
@@ -8369,7 +8847,7 @@ export const MODELS = {
8369
8847
  cost: {
8370
8848
  input: 1,
8371
8849
  output: 5,
8372
- cacheRead: 0,
8850
+ cacheRead: 0.19999999999999998,
8373
8851
  cacheWrite: 0,
8374
8852
  },
8375
8853
  contextWindow: 1000000,
@@ -8722,15 +9200,15 @@ export const MODELS = {
8722
9200
  provider: "vercel-ai-gateway",
8723
9201
  baseUrl: "https://ai-gateway.vercel.sh",
8724
9202
  reasoning: true,
8725
- input: ["text", "image"],
9203
+ input: ["text"],
8726
9204
  cost: {
8727
9205
  input: 0.3,
8728
9206
  output: 2.5,
8729
- cacheRead: 0.03,
9207
+ cacheRead: 0,
8730
9208
  cacheWrite: 0,
8731
9209
  },
8732
9210
  contextWindow: 1000000,
8733
- maxTokens: 64000,
9211
+ maxTokens: 65536,
8734
9212
  },
8735
9213
  "google/gemini-2.5-flash-lite": {
8736
9214
  id: "google/gemini-2.5-flash-lite",
@@ -8790,11 +9268,11 @@ export const MODELS = {
8790
9268
  provider: "vercel-ai-gateway",
8791
9269
  baseUrl: "https://ai-gateway.vercel.sh",
8792
9270
  reasoning: true,
8793
- input: ["text", "image"],
9271
+ input: ["text"],
8794
9272
  cost: {
8795
9273
  input: 1.25,
8796
9274
  output: 10,
8797
- cacheRead: 0.125,
9275
+ cacheRead: 0,
8798
9276
  cacheWrite: 0,
8799
9277
  },
8800
9278
  contextWindow: 1048576,
@@ -10205,7 +10683,7 @@ export const MODELS = {
10205
10683
  cost: {
10206
10684
  input: 0.19999999999999998,
10207
10685
  output: 1.1,
10208
- cacheRead: 0,
10686
+ cacheRead: 0.03,
10209
10687
  cacheWrite: 0,
10210
10688
  },
10211
10689
  contextWindow: 128000,