@ai-sdk/anthropic 4.0.0-beta.32 → 4.0.0-beta.34

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -1,5 +1,18 @@
1
1
  # @ai-sdk/anthropic
2
2
 
3
+ ## 4.0.0-beta.34
4
+
5
+ ### Patch Changes
6
+
7
+ - Updated dependencies [350ea38]
8
+ - @ai-sdk/provider-utils@5.0.0-beta.23
9
+
10
+ ## 4.0.0-beta.33
11
+
12
+ ### Major Changes
13
+
14
+ - 832f86f: fix(anthropic): remove cacheCreationInputTokens from providerMetadata
15
+
3
16
  ## 4.0.0-beta.32
4
17
 
5
18
  ### Patch Changes
package/dist/index.d.ts CHANGED
@@ -21,7 +21,6 @@ interface AnthropicUsageIteration {
21
21
  }
22
22
  interface AnthropicMessageMetadata {
23
23
  usage: JSONObject;
24
- cacheCreationInputTokens: number | null;
25
24
  stopSequence: string | null;
26
25
  /**
27
26
  * Usage breakdown by iteration when compaction is triggered.
package/dist/index.js CHANGED
@@ -3945,11 +3945,10 @@ var AnthropicMessagesLanguageModel = class _AnthropicMessagesLanguageModel {
3945
3945
  },
3946
3946
  warnings,
3947
3947
  providerMetadata: (() => {
3948
- var _a2, _b2, _c2, _d2, _e2;
3948
+ var _a2, _b2, _c2, _d2;
3949
3949
  const anthropicMetadata = {
3950
3950
  usage: response.usage,
3951
- cacheCreationInputTokens: (_a2 = response.usage.cache_creation_input_tokens) != null ? _a2 : null,
3952
- stopSequence: (_b2 = response.stop_sequence) != null ? _b2 : null,
3951
+ stopSequence: (_a2 = response.stop_sequence) != null ? _a2 : null,
3953
3952
  iterations: response.usage.iterations ? response.usage.iterations.map((iter) => ({
3954
3953
  type: iter.type,
3955
3954
  inputTokens: iter.input_tokens,
@@ -3958,15 +3957,15 @@ var AnthropicMessagesLanguageModel = class _AnthropicMessagesLanguageModel {
3958
3957
  container: response.container ? {
3959
3958
  expiresAt: response.container.expires_at,
3960
3959
  id: response.container.id,
3961
- skills: (_d2 = (_c2 = response.container.skills) == null ? void 0 : _c2.map((skill) => ({
3960
+ skills: (_c2 = (_b2 = response.container.skills) == null ? void 0 : _b2.map((skill) => ({
3962
3961
  type: skill.type,
3963
3962
  skillId: skill.skill_id,
3964
3963
  version: skill.version
3965
- }))) != null ? _d2 : null
3964
+ }))) != null ? _c2 : null
3966
3965
  } : null,
3967
- contextManagement: (_e2 = mapAnthropicResponseContextManagement(
3966
+ contextManagement: (_d2 = mapAnthropicResponseContextManagement(
3968
3967
  response.context_management
3969
- )) != null ? _e2 : null
3968
+ )) != null ? _d2 : null
3970
3969
  };
3971
3970
  const providerMetadata = {
3972
3971
  anthropic: anthropicMetadata
@@ -4028,7 +4027,6 @@ var AnthropicMessagesLanguageModel = class _AnthropicMessagesLanguageModel {
4028
4027
  const serverToolCalls = {};
4029
4028
  let contextManagement = null;
4030
4029
  let rawUsage = void 0;
4031
- let cacheCreationInputTokens = null;
4032
4030
  let stopSequence = null;
4033
4031
  let container = null;
4034
4032
  let isJsonResponseFromTool = false;
@@ -4040,7 +4038,7 @@ var AnthropicMessagesLanguageModel = class _AnthropicMessagesLanguageModel {
4040
4038
  controller.enqueue({ type: "stream-start", warnings });
4041
4039
  },
4042
4040
  transform(chunk, controller) {
4043
- var _a2, _b2, _c, _d, _e, _f, _g, _h, _i, _j, _k, _l, _m, _n;
4041
+ var _a2, _b2, _c, _d, _e, _f, _g, _h, _i, _j, _k, _l, _m;
4044
4042
  if (options.includeRawChunks) {
4045
4043
  controller.enqueue({ type: "raw", rawValue: chunk.rawValue });
4046
4044
  }
@@ -4575,7 +4573,6 @@ var AnthropicMessagesLanguageModel = class _AnthropicMessagesLanguageModel {
4575
4573
  rawUsage = {
4576
4574
  ...value.message.usage
4577
4575
  };
4578
- cacheCreationInputTokens = (_g = value.message.usage.cache_creation_input_tokens) != null ? _g : null;
4579
4576
  if (value.message.container != null) {
4580
4577
  container = {
4581
4578
  expiresAt: value.message.container.expires_at,
@@ -4594,8 +4591,8 @@ var AnthropicMessagesLanguageModel = class _AnthropicMessagesLanguageModel {
4594
4591
  }
4595
4592
  controller.enqueue({
4596
4593
  type: "response-metadata",
4597
- id: (_h = value.message.id) != null ? _h : void 0,
4598
- modelId: (_i = value.message.model) != null ? _i : void 0
4594
+ id: (_g = value.message.id) != null ? _g : void 0,
4595
+ modelId: (_h = value.message.model) != null ? _h : void 0
4599
4596
  });
4600
4597
  if (value.message.content != null) {
4601
4598
  for (let contentIndex = 0; contentIndex < value.message.content.length; contentIndex++) {
@@ -4611,7 +4608,7 @@ var AnthropicMessagesLanguageModel = class _AnthropicMessagesLanguageModel {
4611
4608
  id: part.id,
4612
4609
  toolName: part.name
4613
4610
  });
4614
- const inputStr = JSON.stringify((_j = part.input) != null ? _j : {});
4611
+ const inputStr = JSON.stringify((_i = part.input) != null ? _i : {});
4615
4612
  controller.enqueue({
4616
4613
  type: "tool-input-delta",
4617
4614
  id: part.id,
@@ -4649,7 +4646,6 @@ var AnthropicMessagesLanguageModel = class _AnthropicMessagesLanguageModel {
4649
4646
  }
4650
4647
  if (value.usage.cache_creation_input_tokens != null) {
4651
4648
  usage.cache_creation_input_tokens = value.usage.cache_creation_input_tokens;
4652
- cacheCreationInputTokens = value.usage.cache_creation_input_tokens;
4653
4649
  }
4654
4650
  if (value.usage.iterations != null) {
4655
4651
  usage.iterations = value.usage.iterations;
@@ -4659,17 +4655,17 @@ var AnthropicMessagesLanguageModel = class _AnthropicMessagesLanguageModel {
4659
4655
  finishReason: value.delta.stop_reason,
4660
4656
  isJsonResponseFromTool
4661
4657
  }),
4662
- raw: (_k = value.delta.stop_reason) != null ? _k : void 0
4658
+ raw: (_j = value.delta.stop_reason) != null ? _j : void 0
4663
4659
  };
4664
- stopSequence = (_l = value.delta.stop_sequence) != null ? _l : null;
4660
+ stopSequence = (_k = value.delta.stop_sequence) != null ? _k : null;
4665
4661
  container = value.delta.container != null ? {
4666
4662
  expiresAt: value.delta.container.expires_at,
4667
4663
  id: value.delta.container.id,
4668
- skills: (_n = (_m = value.delta.container.skills) == null ? void 0 : _m.map((skill) => ({
4664
+ skills: (_m = (_l = value.delta.container.skills) == null ? void 0 : _l.map((skill) => ({
4669
4665
  type: skill.type,
4670
4666
  skillId: skill.skill_id,
4671
4667
  version: skill.version
4672
- }))) != null ? _n : null
4668
+ }))) != null ? _m : null
4673
4669
  } : null;
4674
4670
  if (value.context_management) {
4675
4671
  contextManagement = mapAnthropicResponseContextManagement(
@@ -4685,7 +4681,6 @@ var AnthropicMessagesLanguageModel = class _AnthropicMessagesLanguageModel {
4685
4681
  case "message_stop": {
4686
4682
  const anthropicMetadata = {
4687
4683
  usage: rawUsage != null ? rawUsage : null,
4688
- cacheCreationInputTokens,
4689
4684
  stopSequence,
4690
4685
  iterations: usage.iterations ? usage.iterations.map((iter) => ({
4691
4686
  type: iter.type,
@@ -5574,7 +5569,7 @@ var AnthropicSkills = class {
5574
5569
  };
5575
5570
 
5576
5571
  // src/version.ts
5577
- var VERSION = true ? "4.0.0-beta.32" : "0.0.0-test";
5572
+ var VERSION = true ? "4.0.0-beta.34" : "0.0.0-test";
5578
5573
 
5579
5574
  // src/anthropic-provider.ts
5580
5575
  function createAnthropic(options = {}) {