@mariozechner/pi-ai 0.49.3 → 0.50.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (108) hide show
  1. package/README.md +32 -22
  2. package/dist/api-registry.d.ts +20 -0
  3. package/dist/api-registry.d.ts.map +1 -0
  4. package/dist/api-registry.js +44 -0
  5. package/dist/api-registry.js.map +1 -0
  6. package/dist/cli.d.ts.map +1 -1
  7. package/dist/cli.js +22 -67
  8. package/dist/cli.js.map +1 -1
  9. package/dist/env-api-keys.d.ts +9 -0
  10. package/dist/env-api-keys.d.ts.map +1 -0
  11. package/dist/env-api-keys.js +91 -0
  12. package/dist/env-api-keys.js.map +1 -0
  13. package/dist/index.d.ts +4 -0
  14. package/dist/index.d.ts.map +1 -1
  15. package/dist/index.js +4 -0
  16. package/dist/index.js.map +1 -1
  17. package/dist/models.generated.d.ts +599 -70
  18. package/dist/models.generated.d.ts.map +1 -1
  19. package/dist/models.generated.js +619 -90
  20. package/dist/models.generated.js.map +1 -1
  21. package/dist/providers/amazon-bedrock.d.ts +3 -2
  22. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  23. package/dist/providers/amazon-bedrock.js +52 -5
  24. package/dist/providers/amazon-bedrock.js.map +1 -1
  25. package/dist/providers/anthropic.d.ts +3 -2
  26. package/dist/providers/anthropic.d.ts.map +1 -1
  27. package/dist/providers/anthropic.js +20 -2
  28. package/dist/providers/anthropic.js.map +1 -1
  29. package/dist/providers/azure-openai-responses.d.ts +15 -0
  30. package/dist/providers/azure-openai-responses.d.ts.map +1 -0
  31. package/dist/providers/azure-openai-responses.js +184 -0
  32. package/dist/providers/azure-openai-responses.js.map +1 -0
  33. package/dist/providers/google-gemini-cli.d.ts +3 -2
  34. package/dist/providers/google-gemini-cli.d.ts.map +1 -1
  35. package/dist/providers/google-gemini-cli.js +68 -1
  36. package/dist/providers/google-gemini-cli.js.map +1 -1
  37. package/dist/providers/google-vertex.d.ts +3 -2
  38. package/dist/providers/google-vertex.d.ts.map +1 -1
  39. package/dist/providers/google-vertex.js +81 -1
  40. package/dist/providers/google-vertex.js.map +1 -1
  41. package/dist/providers/google.d.ts +3 -2
  42. package/dist/providers/google.d.ts.map +1 -1
  43. package/dist/providers/google.js +84 -3
  44. package/dist/providers/google.js.map +1 -1
  45. package/dist/providers/openai-codex-responses.d.ts +3 -2
  46. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  47. package/dist/providers/openai-codex-responses.js +57 -307
  48. package/dist/providers/openai-codex-responses.js.map +1 -1
  49. package/dist/providers/openai-completions.d.ts +5 -2
  50. package/dist/providers/openai-completions.d.ts.map +1 -1
  51. package/dist/providers/openai-completions.js +78 -41
  52. package/dist/providers/openai-completions.js.map +1 -1
  53. package/dist/providers/openai-responses-shared.d.ts +17 -0
  54. package/dist/providers/openai-responses-shared.d.ts.map +1 -0
  55. package/dist/providers/openai-responses-shared.js +424 -0
  56. package/dist/providers/openai-responses-shared.js.map +1 -0
  57. package/dist/providers/openai-responses.d.ts +3 -2
  58. package/dist/providers/openai-responses.d.ts.map +1 -1
  59. package/dist/providers/openai-responses.js +25 -415
  60. package/dist/providers/openai-responses.js.map +1 -1
  61. package/dist/providers/register-builtins.d.ts +3 -0
  62. package/dist/providers/register-builtins.d.ts.map +1 -0
  63. package/dist/providers/register-builtins.js +63 -0
  64. package/dist/providers/register-builtins.js.map +1 -0
  65. package/dist/providers/simple-options.d.ts +8 -0
  66. package/dist/providers/simple-options.d.ts.map +1 -0
  67. package/dist/providers/simple-options.js +32 -0
  68. package/dist/providers/simple-options.js.map +1 -0
  69. package/dist/stream.d.ts +5 -10
  70. package/dist/stream.d.ts.map +1 -1
  71. package/dist/stream.js +17 -420
  72. package/dist/stream.js.map +1 -1
  73. package/dist/types.d.ts +18 -22
  74. package/dist/types.d.ts.map +1 -1
  75. package/dist/types.js +0 -1
  76. package/dist/types.js.map +1 -1
  77. package/dist/utils/event-stream.d.ts +2 -0
  78. package/dist/utils/event-stream.d.ts.map +1 -1
  79. package/dist/utils/event-stream.js +4 -0
  80. package/dist/utils/event-stream.js.map +1 -1
  81. package/dist/utils/oauth/anthropic.d.ts +2 -1
  82. package/dist/utils/oauth/anthropic.d.ts.map +1 -1
  83. package/dist/utils/oauth/anthropic.js +13 -0
  84. package/dist/utils/oauth/anthropic.js.map +1 -1
  85. package/dist/utils/oauth/github-copilot.d.ts +2 -1
  86. package/dist/utils/oauth/github-copilot.d.ts.map +1 -1
  87. package/dist/utils/oauth/github-copilot.js +25 -0
  88. package/dist/utils/oauth/github-copilot.js.map +1 -1
  89. package/dist/utils/oauth/google-antigravity.d.ts +2 -1
  90. package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
  91. package/dist/utils/oauth/google-antigravity.js +19 -0
  92. package/dist/utils/oauth/google-antigravity.js.map +1 -1
  93. package/dist/utils/oauth/google-gemini-cli.d.ts +2 -1
  94. package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
  95. package/dist/utils/oauth/google-gemini-cli.js +19 -0
  96. package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
  97. package/dist/utils/oauth/index.d.ts +26 -16
  98. package/dist/utils/oauth/index.d.ts.map +1 -1
  99. package/dist/utils/oauth/index.js +65 -84
  100. package/dist/utils/oauth/index.js.map +1 -1
  101. package/dist/utils/oauth/openai-codex.d.ts +2 -1
  102. package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
  103. package/dist/utils/oauth/openai-codex.js +20 -1
  104. package/dist/utils/oauth/openai-codex.js.map +1 -1
  105. package/dist/utils/oauth/types.d.ts +28 -6
  106. package/dist/utils/oauth/types.d.ts.map +1 -1
  107. package/dist/utils/oauth/types.js.map +1 -1
  108. package/package.json +3 -1
@@ -1297,6 +1297,586 @@ export const MODELS = {
1297
1297
  maxTokens: 64000,
1298
1298
  },
1299
1299
  },
1300
+ "azure-openai-responses": {
1301
+ "codex-mini-latest": {
1302
+ id: "codex-mini-latest",
1303
+ name: "Codex Mini",
1304
+ api: "azure-openai-responses",
1305
+ provider: "azure-openai-responses",
1306
+ baseUrl: "",
1307
+ reasoning: true,
1308
+ input: ["text"],
1309
+ cost: {
1310
+ input: 1.5,
1311
+ output: 6,
1312
+ cacheRead: 0.375,
1313
+ cacheWrite: 0,
1314
+ },
1315
+ contextWindow: 200000,
1316
+ maxTokens: 100000,
1317
+ },
1318
+ "gpt-4": {
1319
+ id: "gpt-4",
1320
+ name: "GPT-4",
1321
+ api: "azure-openai-responses",
1322
+ provider: "azure-openai-responses",
1323
+ baseUrl: "",
1324
+ reasoning: false,
1325
+ input: ["text"],
1326
+ cost: {
1327
+ input: 30,
1328
+ output: 60,
1329
+ cacheRead: 0,
1330
+ cacheWrite: 0,
1331
+ },
1332
+ contextWindow: 8192,
1333
+ maxTokens: 8192,
1334
+ },
1335
+ "gpt-4-turbo": {
1336
+ id: "gpt-4-turbo",
1337
+ name: "GPT-4 Turbo",
1338
+ api: "azure-openai-responses",
1339
+ provider: "azure-openai-responses",
1340
+ baseUrl: "",
1341
+ reasoning: false,
1342
+ input: ["text", "image"],
1343
+ cost: {
1344
+ input: 10,
1345
+ output: 30,
1346
+ cacheRead: 0,
1347
+ cacheWrite: 0,
1348
+ },
1349
+ contextWindow: 128000,
1350
+ maxTokens: 4096,
1351
+ },
1352
+ "gpt-4.1": {
1353
+ id: "gpt-4.1",
1354
+ name: "GPT-4.1",
1355
+ api: "azure-openai-responses",
1356
+ provider: "azure-openai-responses",
1357
+ baseUrl: "",
1358
+ reasoning: false,
1359
+ input: ["text", "image"],
1360
+ cost: {
1361
+ input: 2,
1362
+ output: 8,
1363
+ cacheRead: 0.5,
1364
+ cacheWrite: 0,
1365
+ },
1366
+ contextWindow: 1047576,
1367
+ maxTokens: 32768,
1368
+ },
1369
+ "gpt-4.1-mini": {
1370
+ id: "gpt-4.1-mini",
1371
+ name: "GPT-4.1 mini",
1372
+ api: "azure-openai-responses",
1373
+ provider: "azure-openai-responses",
1374
+ baseUrl: "",
1375
+ reasoning: false,
1376
+ input: ["text", "image"],
1377
+ cost: {
1378
+ input: 0.4,
1379
+ output: 1.6,
1380
+ cacheRead: 0.1,
1381
+ cacheWrite: 0,
1382
+ },
1383
+ contextWindow: 1047576,
1384
+ maxTokens: 32768,
1385
+ },
1386
+ "gpt-4.1-nano": {
1387
+ id: "gpt-4.1-nano",
1388
+ name: "GPT-4.1 nano",
1389
+ api: "azure-openai-responses",
1390
+ provider: "azure-openai-responses",
1391
+ baseUrl: "",
1392
+ reasoning: false,
1393
+ input: ["text", "image"],
1394
+ cost: {
1395
+ input: 0.1,
1396
+ output: 0.4,
1397
+ cacheRead: 0.03,
1398
+ cacheWrite: 0,
1399
+ },
1400
+ contextWindow: 1047576,
1401
+ maxTokens: 32768,
1402
+ },
1403
+ "gpt-4o": {
1404
+ id: "gpt-4o",
1405
+ name: "GPT-4o",
1406
+ api: "azure-openai-responses",
1407
+ provider: "azure-openai-responses",
1408
+ baseUrl: "",
1409
+ reasoning: false,
1410
+ input: ["text", "image"],
1411
+ cost: {
1412
+ input: 2.5,
1413
+ output: 10,
1414
+ cacheRead: 1.25,
1415
+ cacheWrite: 0,
1416
+ },
1417
+ contextWindow: 128000,
1418
+ maxTokens: 16384,
1419
+ },
1420
+ "gpt-4o-2024-05-13": {
1421
+ id: "gpt-4o-2024-05-13",
1422
+ name: "GPT-4o (2024-05-13)",
1423
+ api: "azure-openai-responses",
1424
+ provider: "azure-openai-responses",
1425
+ baseUrl: "",
1426
+ reasoning: false,
1427
+ input: ["text", "image"],
1428
+ cost: {
1429
+ input: 5,
1430
+ output: 15,
1431
+ cacheRead: 0,
1432
+ cacheWrite: 0,
1433
+ },
1434
+ contextWindow: 128000,
1435
+ maxTokens: 4096,
1436
+ },
1437
+ "gpt-4o-2024-08-06": {
1438
+ id: "gpt-4o-2024-08-06",
1439
+ name: "GPT-4o (2024-08-06)",
1440
+ api: "azure-openai-responses",
1441
+ provider: "azure-openai-responses",
1442
+ baseUrl: "",
1443
+ reasoning: false,
1444
+ input: ["text", "image"],
1445
+ cost: {
1446
+ input: 2.5,
1447
+ output: 10,
1448
+ cacheRead: 1.25,
1449
+ cacheWrite: 0,
1450
+ },
1451
+ contextWindow: 128000,
1452
+ maxTokens: 16384,
1453
+ },
1454
+ "gpt-4o-2024-11-20": {
1455
+ id: "gpt-4o-2024-11-20",
1456
+ name: "GPT-4o (2024-11-20)",
1457
+ api: "azure-openai-responses",
1458
+ provider: "azure-openai-responses",
1459
+ baseUrl: "",
1460
+ reasoning: false,
1461
+ input: ["text", "image"],
1462
+ cost: {
1463
+ input: 2.5,
1464
+ output: 10,
1465
+ cacheRead: 1.25,
1466
+ cacheWrite: 0,
1467
+ },
1468
+ contextWindow: 128000,
1469
+ maxTokens: 16384,
1470
+ },
1471
+ "gpt-4o-mini": {
1472
+ id: "gpt-4o-mini",
1473
+ name: "GPT-4o mini",
1474
+ api: "azure-openai-responses",
1475
+ provider: "azure-openai-responses",
1476
+ baseUrl: "",
1477
+ reasoning: false,
1478
+ input: ["text", "image"],
1479
+ cost: {
1480
+ input: 0.15,
1481
+ output: 0.6,
1482
+ cacheRead: 0.08,
1483
+ cacheWrite: 0,
1484
+ },
1485
+ contextWindow: 128000,
1486
+ maxTokens: 16384,
1487
+ },
1488
+ "gpt-5": {
1489
+ id: "gpt-5",
1490
+ name: "GPT-5",
1491
+ api: "azure-openai-responses",
1492
+ provider: "azure-openai-responses",
1493
+ baseUrl: "",
1494
+ reasoning: true,
1495
+ input: ["text", "image"],
1496
+ cost: {
1497
+ input: 1.25,
1498
+ output: 10,
1499
+ cacheRead: 0.125,
1500
+ cacheWrite: 0,
1501
+ },
1502
+ contextWindow: 400000,
1503
+ maxTokens: 128000,
1504
+ },
1505
+ "gpt-5-chat-latest": {
1506
+ id: "gpt-5-chat-latest",
1507
+ name: "GPT-5 Chat Latest",
1508
+ api: "azure-openai-responses",
1509
+ provider: "azure-openai-responses",
1510
+ baseUrl: "",
1511
+ reasoning: false,
1512
+ input: ["text", "image"],
1513
+ cost: {
1514
+ input: 1.25,
1515
+ output: 10,
1516
+ cacheRead: 0.125,
1517
+ cacheWrite: 0,
1518
+ },
1519
+ contextWindow: 128000,
1520
+ maxTokens: 16384,
1521
+ },
1522
+ "gpt-5-codex": {
1523
+ id: "gpt-5-codex",
1524
+ name: "GPT-5-Codex",
1525
+ api: "azure-openai-responses",
1526
+ provider: "azure-openai-responses",
1527
+ baseUrl: "",
1528
+ reasoning: true,
1529
+ input: ["text", "image"],
1530
+ cost: {
1531
+ input: 1.25,
1532
+ output: 10,
1533
+ cacheRead: 0.125,
1534
+ cacheWrite: 0,
1535
+ },
1536
+ contextWindow: 400000,
1537
+ maxTokens: 128000,
1538
+ },
1539
+ "gpt-5-mini": {
1540
+ id: "gpt-5-mini",
1541
+ name: "GPT-5 Mini",
1542
+ api: "azure-openai-responses",
1543
+ provider: "azure-openai-responses",
1544
+ baseUrl: "",
1545
+ reasoning: true,
1546
+ input: ["text", "image"],
1547
+ cost: {
1548
+ input: 0.25,
1549
+ output: 2,
1550
+ cacheRead: 0.025,
1551
+ cacheWrite: 0,
1552
+ },
1553
+ contextWindow: 400000,
1554
+ maxTokens: 128000,
1555
+ },
1556
+ "gpt-5-nano": {
1557
+ id: "gpt-5-nano",
1558
+ name: "GPT-5 Nano",
1559
+ api: "azure-openai-responses",
1560
+ provider: "azure-openai-responses",
1561
+ baseUrl: "",
1562
+ reasoning: true,
1563
+ input: ["text", "image"],
1564
+ cost: {
1565
+ input: 0.05,
1566
+ output: 0.4,
1567
+ cacheRead: 0.005,
1568
+ cacheWrite: 0,
1569
+ },
1570
+ contextWindow: 400000,
1571
+ maxTokens: 128000,
1572
+ },
1573
+ "gpt-5-pro": {
1574
+ id: "gpt-5-pro",
1575
+ name: "GPT-5 Pro",
1576
+ api: "azure-openai-responses",
1577
+ provider: "azure-openai-responses",
1578
+ baseUrl: "",
1579
+ reasoning: true,
1580
+ input: ["text", "image"],
1581
+ cost: {
1582
+ input: 15,
1583
+ output: 120,
1584
+ cacheRead: 0,
1585
+ cacheWrite: 0,
1586
+ },
1587
+ contextWindow: 400000,
1588
+ maxTokens: 272000,
1589
+ },
1590
+ "gpt-5.1": {
1591
+ id: "gpt-5.1",
1592
+ name: "GPT-5.1",
1593
+ api: "azure-openai-responses",
1594
+ provider: "azure-openai-responses",
1595
+ baseUrl: "",
1596
+ reasoning: true,
1597
+ input: ["text", "image"],
1598
+ cost: {
1599
+ input: 1.25,
1600
+ output: 10,
1601
+ cacheRead: 0.13,
1602
+ cacheWrite: 0,
1603
+ },
1604
+ contextWindow: 400000,
1605
+ maxTokens: 128000,
1606
+ },
1607
+ "gpt-5.1-chat-latest": {
1608
+ id: "gpt-5.1-chat-latest",
1609
+ name: "GPT-5.1 Chat",
1610
+ api: "azure-openai-responses",
1611
+ provider: "azure-openai-responses",
1612
+ baseUrl: "",
1613
+ reasoning: true,
1614
+ input: ["text", "image"],
1615
+ cost: {
1616
+ input: 1.25,
1617
+ output: 10,
1618
+ cacheRead: 0.125,
1619
+ cacheWrite: 0,
1620
+ },
1621
+ contextWindow: 128000,
1622
+ maxTokens: 16384,
1623
+ },
1624
+ "gpt-5.1-codex": {
1625
+ id: "gpt-5.1-codex",
1626
+ name: "GPT-5.1 Codex",
1627
+ api: "azure-openai-responses",
1628
+ provider: "azure-openai-responses",
1629
+ baseUrl: "",
1630
+ reasoning: true,
1631
+ input: ["text", "image"],
1632
+ cost: {
1633
+ input: 1.25,
1634
+ output: 10,
1635
+ cacheRead: 0.125,
1636
+ cacheWrite: 0,
1637
+ },
1638
+ contextWindow: 400000,
1639
+ maxTokens: 128000,
1640
+ },
1641
+ "gpt-5.1-codex-max": {
1642
+ id: "gpt-5.1-codex-max",
1643
+ name: "GPT-5.1 Codex Max",
1644
+ api: "azure-openai-responses",
1645
+ provider: "azure-openai-responses",
1646
+ baseUrl: "",
1647
+ reasoning: true,
1648
+ input: ["text", "image"],
1649
+ cost: {
1650
+ input: 1.25,
1651
+ output: 10,
1652
+ cacheRead: 0.125,
1653
+ cacheWrite: 0,
1654
+ },
1655
+ contextWindow: 400000,
1656
+ maxTokens: 128000,
1657
+ },
1658
+ "gpt-5.1-codex-mini": {
1659
+ id: "gpt-5.1-codex-mini",
1660
+ name: "GPT-5.1 Codex mini",
1661
+ api: "azure-openai-responses",
1662
+ provider: "azure-openai-responses",
1663
+ baseUrl: "",
1664
+ reasoning: true,
1665
+ input: ["text", "image"],
1666
+ cost: {
1667
+ input: 0.25,
1668
+ output: 2,
1669
+ cacheRead: 0.025,
1670
+ cacheWrite: 0,
1671
+ },
1672
+ contextWindow: 400000,
1673
+ maxTokens: 128000,
1674
+ },
1675
+ "gpt-5.2": {
1676
+ id: "gpt-5.2",
1677
+ name: "GPT-5.2",
1678
+ api: "azure-openai-responses",
1679
+ provider: "azure-openai-responses",
1680
+ baseUrl: "",
1681
+ reasoning: true,
1682
+ input: ["text", "image"],
1683
+ cost: {
1684
+ input: 1.75,
1685
+ output: 14,
1686
+ cacheRead: 0.175,
1687
+ cacheWrite: 0,
1688
+ },
1689
+ contextWindow: 400000,
1690
+ maxTokens: 128000,
1691
+ },
1692
+ "gpt-5.2-chat-latest": {
1693
+ id: "gpt-5.2-chat-latest",
1694
+ name: "GPT-5.2 Chat",
1695
+ api: "azure-openai-responses",
1696
+ provider: "azure-openai-responses",
1697
+ baseUrl: "",
1698
+ reasoning: true,
1699
+ input: ["text", "image"],
1700
+ cost: {
1701
+ input: 1.75,
1702
+ output: 14,
1703
+ cacheRead: 0.175,
1704
+ cacheWrite: 0,
1705
+ },
1706
+ contextWindow: 128000,
1707
+ maxTokens: 16384,
1708
+ },
1709
+ "gpt-5.2-codex": {
1710
+ id: "gpt-5.2-codex",
1711
+ name: "GPT-5.2 Codex",
1712
+ api: "azure-openai-responses",
1713
+ provider: "azure-openai-responses",
1714
+ baseUrl: "",
1715
+ reasoning: true,
1716
+ input: ["text", "image"],
1717
+ cost: {
1718
+ input: 1.75,
1719
+ output: 14,
1720
+ cacheRead: 0.175,
1721
+ cacheWrite: 0,
1722
+ },
1723
+ contextWindow: 400000,
1724
+ maxTokens: 128000,
1725
+ },
1726
+ "gpt-5.2-pro": {
1727
+ id: "gpt-5.2-pro",
1728
+ name: "GPT-5.2 Pro",
1729
+ api: "azure-openai-responses",
1730
+ provider: "azure-openai-responses",
1731
+ baseUrl: "",
1732
+ reasoning: true,
1733
+ input: ["text", "image"],
1734
+ cost: {
1735
+ input: 21,
1736
+ output: 168,
1737
+ cacheRead: 0,
1738
+ cacheWrite: 0,
1739
+ },
1740
+ contextWindow: 400000,
1741
+ maxTokens: 128000,
1742
+ },
1743
+ "o1": {
1744
+ id: "o1",
1745
+ name: "o1",
1746
+ api: "azure-openai-responses",
1747
+ provider: "azure-openai-responses",
1748
+ baseUrl: "",
1749
+ reasoning: true,
1750
+ input: ["text", "image"],
1751
+ cost: {
1752
+ input: 15,
1753
+ output: 60,
1754
+ cacheRead: 7.5,
1755
+ cacheWrite: 0,
1756
+ },
1757
+ contextWindow: 200000,
1758
+ maxTokens: 100000,
1759
+ },
1760
+ "o1-pro": {
1761
+ id: "o1-pro",
1762
+ name: "o1-pro",
1763
+ api: "azure-openai-responses",
1764
+ provider: "azure-openai-responses",
1765
+ baseUrl: "",
1766
+ reasoning: true,
1767
+ input: ["text", "image"],
1768
+ cost: {
1769
+ input: 150,
1770
+ output: 600,
1771
+ cacheRead: 0,
1772
+ cacheWrite: 0,
1773
+ },
1774
+ contextWindow: 200000,
1775
+ maxTokens: 100000,
1776
+ },
1777
+ "o3": {
1778
+ id: "o3",
1779
+ name: "o3",
1780
+ api: "azure-openai-responses",
1781
+ provider: "azure-openai-responses",
1782
+ baseUrl: "",
1783
+ reasoning: true,
1784
+ input: ["text", "image"],
1785
+ cost: {
1786
+ input: 2,
1787
+ output: 8,
1788
+ cacheRead: 0.5,
1789
+ cacheWrite: 0,
1790
+ },
1791
+ contextWindow: 200000,
1792
+ maxTokens: 100000,
1793
+ },
1794
+ "o3-deep-research": {
1795
+ id: "o3-deep-research",
1796
+ name: "o3-deep-research",
1797
+ api: "azure-openai-responses",
1798
+ provider: "azure-openai-responses",
1799
+ baseUrl: "",
1800
+ reasoning: true,
1801
+ input: ["text", "image"],
1802
+ cost: {
1803
+ input: 10,
1804
+ output: 40,
1805
+ cacheRead: 2.5,
1806
+ cacheWrite: 0,
1807
+ },
1808
+ contextWindow: 200000,
1809
+ maxTokens: 100000,
1810
+ },
1811
+ "o3-mini": {
1812
+ id: "o3-mini",
1813
+ name: "o3-mini",
1814
+ api: "azure-openai-responses",
1815
+ provider: "azure-openai-responses",
1816
+ baseUrl: "",
1817
+ reasoning: true,
1818
+ input: ["text"],
1819
+ cost: {
1820
+ input: 1.1,
1821
+ output: 4.4,
1822
+ cacheRead: 0.55,
1823
+ cacheWrite: 0,
1824
+ },
1825
+ contextWindow: 200000,
1826
+ maxTokens: 100000,
1827
+ },
1828
+ "o3-pro": {
1829
+ id: "o3-pro",
1830
+ name: "o3-pro",
1831
+ api: "azure-openai-responses",
1832
+ provider: "azure-openai-responses",
1833
+ baseUrl: "",
1834
+ reasoning: true,
1835
+ input: ["text", "image"],
1836
+ cost: {
1837
+ input: 20,
1838
+ output: 80,
1839
+ cacheRead: 0,
1840
+ cacheWrite: 0,
1841
+ },
1842
+ contextWindow: 200000,
1843
+ maxTokens: 100000,
1844
+ },
1845
+ "o4-mini": {
1846
+ id: "o4-mini",
1847
+ name: "o4-mini",
1848
+ api: "azure-openai-responses",
1849
+ provider: "azure-openai-responses",
1850
+ baseUrl: "",
1851
+ reasoning: true,
1852
+ input: ["text", "image"],
1853
+ cost: {
1854
+ input: 1.1,
1855
+ output: 4.4,
1856
+ cacheRead: 0.28,
1857
+ cacheWrite: 0,
1858
+ },
1859
+ contextWindow: 200000,
1860
+ maxTokens: 100000,
1861
+ },
1862
+ "o4-mini-deep-research": {
1863
+ id: "o4-mini-deep-research",
1864
+ name: "o4-mini-deep-research",
1865
+ api: "azure-openai-responses",
1866
+ provider: "azure-openai-responses",
1867
+ baseUrl: "",
1868
+ reasoning: true,
1869
+ input: ["text", "image"],
1870
+ cost: {
1871
+ input: 2,
1872
+ output: 8,
1873
+ cacheRead: 0.5,
1874
+ cacheWrite: 0,
1875
+ },
1876
+ contextWindow: 200000,
1877
+ maxTokens: 100000,
1878
+ },
1879
+ },
1300
1880
  "cerebras": {
1301
1881
  "gpt-oss-120b": {
1302
1882
  id: "gpt-oss-120b",
@@ -1500,7 +2080,7 @@ export const MODELS = {
1500
2080
  cacheRead: 0,
1501
2081
  cacheWrite: 0,
1502
2082
  },
1503
- contextWindow: 128000,
2083
+ contextWindow: 64000,
1504
2084
  maxTokens: 16384,
1505
2085
  },
1506
2086
  "gpt-4o": {
@@ -3866,40 +4446,6 @@ export const MODELS = {
3866
4446
  },
3867
4447
  },
3868
4448
  "opencode": {
3869
- "alpha-gd4": {
3870
- id: "alpha-gd4",
3871
- name: "Alpha GD4",
3872
- api: "anthropic-messages",
3873
- provider: "opencode",
3874
- baseUrl: "https://opencode.ai/zen",
3875
- reasoning: true,
3876
- input: ["text"],
3877
- cost: {
3878
- input: 0.5,
3879
- output: 2,
3880
- cacheRead: 0.15,
3881
- cacheWrite: 0,
3882
- },
3883
- contextWindow: 262144,
3884
- maxTokens: 32768,
3885
- },
3886
- "alpha-glm-4.7": {
3887
- id: "alpha-glm-4.7",
3888
- name: "Alpha GLM-4.7",
3889
- api: "openai-completions",
3890
- provider: "opencode",
3891
- baseUrl: "https://opencode.ai/zen/v1",
3892
- reasoning: true,
3893
- input: ["text"],
3894
- cost: {
3895
- input: 0.6,
3896
- output: 2.2,
3897
- cacheRead: 0.6,
3898
- cacheWrite: 0,
3899
- },
3900
- contextWindow: 204800,
3901
- maxTokens: 131072,
3902
- },
3903
4449
  "big-pickle": {
3904
4450
  id: "big-pickle",
3905
4451
  name: "Big Pickle",
@@ -4070,6 +4616,23 @@ export const MODELS = {
4070
4616
  contextWindow: 204800,
4071
4617
  maxTokens: 131072,
4072
4618
  },
4619
+ "glm-4.7": {
4620
+ id: "glm-4.7",
4621
+ name: "GLM-4.7",
4622
+ api: "openai-completions",
4623
+ provider: "opencode",
4624
+ baseUrl: "https://opencode.ai/zen/v1",
4625
+ reasoning: true,
4626
+ input: ["text"],
4627
+ cost: {
4628
+ input: 0.6,
4629
+ output: 2.2,
4630
+ cacheRead: 0.1,
4631
+ cacheWrite: 0,
4632
+ },
4633
+ contextWindow: 204800,
4634
+ maxTokens: 131072,
4635
+ },
4073
4636
  "glm-4.7-free": {
4074
4637
  id: "glm-4.7-free",
4075
4638
  name: "GLM-4.7",
@@ -4508,8 +5071,8 @@ export const MODELS = {
4508
5071
  cost: {
4509
5072
  input: 0.7999999999999999,
4510
5073
  output: 4,
4511
- cacheRead: 0.08,
4512
- cacheWrite: 1,
5074
+ cacheRead: 0,
5075
+ cacheWrite: 0,
4513
5076
  },
4514
5077
  contextWindow: 200000,
4515
5078
  maxTokens: 8192,
@@ -5053,7 +5616,7 @@ export const MODELS = {
5053
5616
  input: 0.09999999999999999,
5054
5617
  output: 0.39999999999999997,
5055
5618
  cacheRead: 0.024999999999999998,
5056
- cacheWrite: 0.0833,
5619
+ cacheWrite: 0.08333333333333334,
5057
5620
  },
5058
5621
  contextWindow: 1048576,
5059
5622
  maxTokens: 8192,
@@ -5121,7 +5684,7 @@ export const MODELS = {
5121
5684
  input: 0.09999999999999999,
5122
5685
  output: 0.39999999999999997,
5123
5686
  cacheRead: 0.01,
5124
- cacheWrite: 0.0833,
5687
+ cacheWrite: 0.08333333333333334,
5125
5688
  },
5126
5689
  contextWindow: 1048576,
5127
5690
  maxTokens: 65535,
@@ -5138,7 +5701,7 @@ export const MODELS = {
5138
5701
  input: 0.09999999999999999,
5139
5702
  output: 0.39999999999999997,
5140
5703
  cacheRead: 0.01,
5141
- cacheWrite: 0.0833,
5704
+ cacheWrite: 0.08333333333333334,
5142
5705
  },
5143
5706
  contextWindow: 1048576,
5144
5707
  maxTokens: 65535,
@@ -5155,7 +5718,7 @@ export const MODELS = {
5155
5718
  input: 0.3,
5156
5719
  output: 2.5,
5157
5720
  cacheRead: 0.03,
5158
- cacheWrite: 0.0833,
5721
+ cacheWrite: 0.08333333333333334,
5159
5722
  },
5160
5723
  contextWindow: 1048576,
5161
5724
  maxTokens: 65535,
@@ -5188,7 +5751,7 @@ export const MODELS = {
5188
5751
  cost: {
5189
5752
  input: 1.25,
5190
5753
  output: 10,
5191
- cacheRead: 0.31,
5754
+ cacheRead: 0.125,
5192
5755
  cacheWrite: 0.375,
5193
5756
  },
5194
5757
  contextWindow: 1048576,
@@ -5205,7 +5768,7 @@ export const MODELS = {
5205
5768
  cost: {
5206
5769
  input: 1.25,
5207
5770
  output: 10,
5208
- cacheRead: 0.31,
5771
+ cacheRead: 0.125,
5209
5772
  cacheWrite: 0.375,
5210
5773
  },
5211
5774
  contextWindow: 1048576,
@@ -5223,7 +5786,7 @@ export const MODELS = {
5223
5786
  input: 0.5,
5224
5787
  output: 3,
5225
5788
  cacheRead: 0.049999999999999996,
5226
- cacheWrite: 0,
5789
+ cacheWrite: 0.08333333333333334,
5227
5790
  },
5228
5791
  contextWindow: 1048576,
5229
5792
  maxTokens: 65535,
@@ -5510,12 +6073,12 @@ export const MODELS = {
5510
6073
  input: ["text"],
5511
6074
  cost: {
5512
6075
  input: 0.27,
5513
- output: 1.12,
6076
+ output: 1.1,
5514
6077
  cacheRead: 0,
5515
6078
  cacheWrite: 0,
5516
6079
  },
5517
6080
  contextWindow: 196608,
5518
- maxTokens: 65536,
6081
+ maxTokens: 196608,
5519
6082
  },
5520
6083
  "mistralai/codestral-2508": {
5521
6084
  id: "mistralai/codestral-2508",
@@ -7268,23 +7831,6 @@ export const MODELS = {
7268
7831
  contextWindow: 131072,
7269
7832
  maxTokens: 8192,
7270
7833
  },
7271
- "qwen/qwen2.5-vl-72b-instruct": {
7272
- id: "qwen/qwen2.5-vl-72b-instruct",
7273
- name: "Qwen: Qwen2.5 VL 72B Instruct",
7274
- api: "openai-completions",
7275
- provider: "openrouter",
7276
- baseUrl: "https://openrouter.ai/api/v1",
7277
- reasoning: false,
7278
- input: ["text", "image"],
7279
- cost: {
7280
- input: 0.15,
7281
- output: 0.6,
7282
- cacheRead: 0,
7283
- cacheWrite: 0,
7284
- },
7285
- contextWindow: 32768,
7286
- maxTokens: 32768,
7287
- },
7288
7834
  "qwen/qwen3-14b": {
7289
7835
  id: "qwen/qwen3-14b",
7290
7836
  name: "Qwen: Qwen3 14B",
@@ -7534,7 +8080,7 @@ export const MODELS = {
7534
8080
  cost: {
7535
8081
  input: 0.22,
7536
8082
  output: 1.7999999999999998,
7537
- cacheRead: 0,
8083
+ cacheRead: 0.022,
7538
8084
  cacheWrite: 0,
7539
8085
  },
7540
8086
  contextWindow: 262144,
@@ -7670,7 +8216,7 @@ export const MODELS = {
7670
8216
  cost: {
7671
8217
  input: 0.15,
7672
8218
  output: 0.6,
7673
- cacheRead: 0,
8219
+ cacheRead: 0.075,
7674
8220
  cacheWrite: 0,
7675
8221
  },
7676
8222
  contextWindow: 262144,
@@ -8050,23 +8596,6 @@ export const MODELS = {
8050
8596
  contextWindow: 262144,
8051
8597
  maxTokens: 4096,
8052
8598
  },
8053
- "xiaomi/mimo-v2-flash:free": {
8054
- id: "xiaomi/mimo-v2-flash:free",
8055
- name: "Xiaomi: MiMo-V2-Flash (free)",
8056
- api: "openai-completions",
8057
- provider: "openrouter",
8058
- baseUrl: "https://openrouter.ai/api/v1",
8059
- reasoning: true,
8060
- input: ["text"],
8061
- cost: {
8062
- input: 0,
8063
- output: 0,
8064
- cacheRead: 0,
8065
- cacheWrite: 0,
8066
- },
8067
- contextWindow: 262144,
8068
- maxTokens: 65536,
8069
- },
8070
8599
  "z-ai/glm-4-32b": {
8071
8600
  id: "z-ai/glm-4-32b",
8072
8601
  name: "Z.AI: GLM 4 32B ",
@@ -8180,7 +8709,7 @@ export const MODELS = {
8180
8709
  cost: {
8181
8710
  input: 0.44,
8182
8711
  output: 1.76,
8183
- cacheRead: 0,
8712
+ cacheRead: 0.11,
8184
8713
  cacheWrite: 0,
8185
8714
  },
8186
8715
  contextWindow: 204800,
@@ -8369,7 +8898,7 @@ export const MODELS = {
8369
8898
  cost: {
8370
8899
  input: 1,
8371
8900
  output: 5,
8372
- cacheRead: 0,
8901
+ cacheRead: 0.19999999999999998,
8373
8902
  cacheWrite: 0,
8374
8903
  },
8375
8904
  contextWindow: 1000000,
@@ -8722,15 +9251,15 @@ export const MODELS = {
8722
9251
  provider: "vercel-ai-gateway",
8723
9252
  baseUrl: "https://ai-gateway.vercel.sh",
8724
9253
  reasoning: true,
8725
- input: ["text", "image"],
9254
+ input: ["text"],
8726
9255
  cost: {
8727
9256
  input: 0.3,
8728
9257
  output: 2.5,
8729
- cacheRead: 0.03,
9258
+ cacheRead: 0,
8730
9259
  cacheWrite: 0,
8731
9260
  },
8732
9261
  contextWindow: 1000000,
8733
- maxTokens: 64000,
9262
+ maxTokens: 65536,
8734
9263
  },
8735
9264
  "google/gemini-2.5-flash-lite": {
8736
9265
  id: "google/gemini-2.5-flash-lite",
@@ -8790,11 +9319,11 @@ export const MODELS = {
8790
9319
  provider: "vercel-ai-gateway",
8791
9320
  baseUrl: "https://ai-gateway.vercel.sh",
8792
9321
  reasoning: true,
8793
- input: ["text", "image"],
9322
+ input: ["text"],
8794
9323
  cost: {
8795
9324
  input: 1.25,
8796
9325
  output: 10,
8797
- cacheRead: 0.125,
9326
+ cacheRead: 0,
8798
9327
  cacheWrite: 0,
8799
9328
  },
8800
9329
  contextWindow: 1048576,
@@ -10205,7 +10734,7 @@ export const MODELS = {
10205
10734
  cost: {
10206
10735
  input: 0.19999999999999998,
10207
10736
  output: 1.1,
10208
- cacheRead: 0,
10737
+ cacheRead: 0.03,
10209
10738
  cacheWrite: 0,
10210
10739
  },
10211
10740
  contextWindow: 128000,