@ax-llm/ax 11.0.56 → 11.0.57

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/index.d.cts CHANGED
@@ -166,6 +166,7 @@ type AxModelInfo = {
166
166
  aliases?: string[];
167
167
  hasThinkingBudget?: boolean;
168
168
  hasShowThoughts?: boolean;
169
+ maxTokens?: number;
169
170
  };
170
171
  type AxTokenUsage = {
171
172
  promptTokens: number;
@@ -386,6 +387,8 @@ interface AxAIServiceImpl<TModel, TEmbedModel, TChatRequest, TEmbedRequest, TCha
386
387
  }
387
388
 
388
389
  declare enum AxAIAnthropicModel {
390
+ Claude4Opus = "claude-opus-4-20250514",
391
+ Claude4Sonnet = "claude-sonnet-4-20250514",
389
392
  Claude37Sonnet = "claude-3-7-sonnet-latest",
390
393
  Claude35Sonnet = "claude-3-5-sonnet-latest",
391
394
  Claude35Haiku = "claude-3-5-haiku-latest",
package/index.d.ts CHANGED
@@ -166,6 +166,7 @@ type AxModelInfo = {
166
166
  aliases?: string[];
167
167
  hasThinkingBudget?: boolean;
168
168
  hasShowThoughts?: boolean;
169
+ maxTokens?: number;
169
170
  };
170
171
  type AxTokenUsage = {
171
172
  promptTokens: number;
@@ -386,6 +387,8 @@ interface AxAIServiceImpl<TModel, TEmbedModel, TChatRequest, TEmbedRequest, TCha
386
387
  }
387
388
 
388
389
  declare enum AxAIAnthropicModel {
390
+ Claude4Opus = "claude-opus-4-20250514",
391
+ Claude4Sonnet = "claude-sonnet-4-20250514",
389
392
  Claude37Sonnet = "claude-3-7-sonnet-latest",
390
393
  Claude35Sonnet = "claude-3-5-sonnet-latest",
391
394
  Claude35Haiku = "claude-3-5-haiku-latest",
package/index.js CHANGED
@@ -1425,6 +1425,8 @@ var GoogleVertexAuth = class {
1425
1425
 
1426
1426
  // ai/anthropic/types.ts
1427
1427
  var AxAIAnthropicModel = /* @__PURE__ */ ((AxAIAnthropicModel2) => {
1428
+ AxAIAnthropicModel2["Claude4Opus"] = "claude-opus-4-20250514";
1429
+ AxAIAnthropicModel2["Claude4Sonnet"] = "claude-sonnet-4-20250514";
1428
1430
  AxAIAnthropicModel2["Claude37Sonnet"] = "claude-3-7-sonnet-latest";
1429
1431
  AxAIAnthropicModel2["Claude35Sonnet"] = "claude-3-5-sonnet-latest";
1430
1432
  AxAIAnthropicModel2["Claude35Haiku"] = "claude-3-5-haiku-latest";
@@ -1447,50 +1449,80 @@ var AxAIAnthropicVertexModel = /* @__PURE__ */ ((AxAIAnthropicVertexModel2) => {
1447
1449
 
1448
1450
  // ai/anthropic/info.ts
1449
1451
  var axModelInfoAnthropic = [
1450
- // 35
1452
+ // 4
1453
+ {
1454
+ name: "claude-opus-4-20250514" /* Claude4Opus */,
1455
+ currency: "usd",
1456
+ promptTokenCostPer1M: 15,
1457
+ completionTokenCostPer1M: 75,
1458
+ maxTokens: 32e3
1459
+ },
1460
+ {
1461
+ name: "claude-sonnet-4-20250514" /* Claude4Sonnet */,
1462
+ currency: "usd",
1463
+ promptTokenCostPer1M: 3,
1464
+ completionTokenCostPer1M: 15,
1465
+ maxTokens: 64e3
1466
+ },
1467
+ // 3.7
1468
+ {
1469
+ name: "claude-3-7-sonnet-latest" /* Claude37Sonnet */,
1470
+ currency: "usd",
1471
+ promptTokenCostPer1M: 3,
1472
+ completionTokenCostPer1M: 15,
1473
+ maxTokens: 64e3
1474
+ },
1475
+ // 3.5
1451
1476
  {
1452
1477
  name: "claude-3-5-sonnet-latest" /* Claude35Sonnet */,
1453
1478
  currency: "usd",
1454
1479
  promptTokenCostPer1M: 3,
1455
- completionTokenCostPer1M: 15
1480
+ completionTokenCostPer1M: 15,
1481
+ maxTokens: 8192
1456
1482
  },
1457
1483
  {
1458
1484
  name: "claude-3-5-haiku-latest" /* Claude35Haiku */,
1459
1485
  currency: "usd",
1460
1486
  promptTokenCostPer1M: 0.8,
1461
- completionTokenCostPer1M: 4
1487
+ completionTokenCostPer1M: 4,
1488
+ maxTokens: 8192
1462
1489
  },
1463
1490
  // 3
1464
1491
  {
1465
1492
  name: "claude-3-opus-latest" /* Claude3Opus */,
1466
1493
  currency: "usd",
1467
1494
  promptTokenCostPer1M: 15,
1468
- completionTokenCostPer1M: 75
1495
+ completionTokenCostPer1M: 75,
1496
+ maxTokens: 4096
1469
1497
  },
1470
1498
  {
1471
1499
  name: "claude-3-sonnet-20240229" /* Claude3Sonnet */,
1472
1500
  currency: "usd",
1473
1501
  promptTokenCostPer1M: 3,
1474
- completionTokenCostPer1M: 15
1502
+ completionTokenCostPer1M: 15,
1503
+ maxTokens: 4096
1475
1504
  },
1476
1505
  {
1477
1506
  name: "claude-3-haiku-20240307" /* Claude3Haiku */,
1478
1507
  currency: "usd",
1479
1508
  promptTokenCostPer1M: 0.25,
1480
- completionTokenCostPer1M: 1.25
1509
+ completionTokenCostPer1M: 1.25,
1510
+ maxTokens: 4096
1481
1511
  },
1482
- // 21
1512
+ // 2.1
1483
1513
  {
1484
1514
  name: "claude-2.1" /* Claude21 */,
1485
1515
  currency: "usd",
1486
1516
  promptTokenCostPer1M: 8,
1487
- completionTokenCostPer1M: 25
1517
+ completionTokenCostPer1M: 25,
1518
+ maxTokens: 4096
1488
1519
  },
1489
1520
  {
1490
1521
  name: "claude-instant-1.2" /* ClaudeInstant12 */,
1491
1522
  currency: "usd",
1492
1523
  promptTokenCostPer1M: 0.8,
1493
- completionTokenCostPer1M: 2.24
1524
+ completionTokenCostPer1M: 2.24,
1525
+ maxTokens: 4096
1494
1526
  }
1495
1527
  ];
1496
1528
 
@@ -1515,7 +1547,7 @@ var AxAIAnthropicImpl = class {
1515
1547
  getModelConfig() {
1516
1548
  const { config } = this;
1517
1549
  return {
1518
- maxTokens: config.maxTokens,
1550
+ maxTokens: config.maxTokens ?? 4096,
1519
1551
  temperature: config.temperature,
1520
1552
  topP: config.topP,
1521
1553
  topK: config.topK,
@@ -1578,13 +1610,18 @@ var AxAIAnthropicImpl = class {
1578
1610
  input_schema: v.parameters
1579
1611
  })
1580
1612
  );
1613
+ const maxTokens = req.modelConfig?.maxTokens ?? this.config.maxTokens;
1614
+ const stopSequences = req.modelConfig?.stopSequences ?? this.config.stopSequences;
1615
+ const temperature = req.modelConfig?.temperature ?? this.config.temperature;
1616
+ const topP = req.modelConfig?.topP ?? this.config.topP;
1617
+ const topK = req.modelConfig?.topK ?? this.config.topK;
1581
1618
  const reqValue = {
1582
1619
  ...this.isVertex ? { anthropic_version: "vertex-2023-10-16" } : { model },
1583
- max_tokens: req.modelConfig?.maxTokens ?? this.config.maxTokens,
1584
- stop_sequences: req.modelConfig?.stopSequences ?? this.config.stopSequences,
1585
- temperature: req.modelConfig?.temperature ?? this.config.temperature,
1586
- top_p: req.modelConfig?.topP ?? this.config.topP,
1587
- top_k: req.modelConfig?.topK ?? this.config.topK,
1620
+ ...maxTokens ? { max_tokens: maxTokens } : {},
1621
+ ...stopSequences && stopSequences.length > 0 ? { stop_sequences: stopSequences } : {},
1622
+ ...temperature ? { temperature } : {},
1623
+ ...topP ? { top_p: topP } : {},
1624
+ ...topK ? { top_k: topK } : {},
1588
1625
  ...toolsChoice,
1589
1626
  ...tools && tools.length > 0 ? { tools } : {},
1590
1627
  ...stream ? { stream: true } : {},