ai 4.0.38 → 4.0.40

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.mjs CHANGED
@@ -4,19 +4,17 @@ var __export = (target, all) => {
4
4
  __defProp(target, name14, { get: all[name14], enumerable: true });
5
5
  };
6
6
 
7
- // streams/index.ts
7
+ // core/index.ts
8
+ import { createIdGenerator as createIdGenerator5, generateId as generateId2 } from "@ai-sdk/provider-utils";
8
9
  import {
9
- formatAssistantStreamPart as formatAssistantStreamPart2,
10
- formatDataStreamPart as formatDataStreamPart6,
10
+ formatAssistantStreamPart,
11
+ formatDataStreamPart as formatDataStreamPart3,
12
+ jsonSchema,
11
13
  parseAssistantStreamPart,
12
14
  parseDataStreamPart,
13
15
  processDataStream,
14
16
  processTextStream
15
17
  } from "@ai-sdk/ui-utils";
16
- import { generateId as generateId3 } from "@ai-sdk/provider-utils";
17
-
18
- // core/index.ts
19
- import { jsonSchema } from "@ai-sdk/ui-utils";
20
18
 
21
19
  // core/data-stream/create-data-stream.ts
22
20
  import { formatDataStreamPart } from "@ai-sdk/ui-utils";
@@ -1907,7 +1905,7 @@ var arrayOutputStrategy = (schema) => {
1907
1905
  // be able to generate an array directly:
1908
1906
  // possible future optimization: use arrays directly when model supports grammar-guided generation
1909
1907
  jsonSchema: {
1910
- $schema: "https://json-schema.org/draft/2019-09/schema#",
1908
+ $schema: "http://json-schema.org/draft-07/schema#",
1911
1909
  type: "object",
1912
1910
  properties: {
1913
1911
  elements: { type: "array", items: itemSchema }
@@ -2019,7 +2017,7 @@ var enumOutputStrategy = (enumValues) => {
2019
2017
  // be able to generate an enum value directly:
2020
2018
  // possible future optimization: use enums directly when model supports top-level enums
2021
2019
  jsonSchema: {
2022
- $schema: "https://json-schema.org/draft/2019-09/schema#",
2020
+ $schema: "http://json-schema.org/draft-07/schema#",
2023
2021
  type: "object",
2024
2022
  properties: {
2025
2023
  result: { type: "string", enum: enumValues }
@@ -2225,7 +2223,7 @@ async function generateObject({
2225
2223
  experimental_telemetry: telemetry,
2226
2224
  experimental_providerMetadata: providerMetadata,
2227
2225
  _internal: {
2228
- generateId: generateId4 = originalGenerateId,
2226
+ generateId: generateId3 = originalGenerateId,
2229
2227
  currentDate = () => /* @__PURE__ */ new Date()
2230
2228
  } = {},
2231
2229
  ...settings
@@ -2355,7 +2353,7 @@ async function generateObject({
2355
2353
  headers
2356
2354
  });
2357
2355
  const responseData = {
2358
- id: (_b2 = (_a15 = result2.response) == null ? void 0 : _a15.id) != null ? _b2 : generateId4(),
2356
+ id: (_b2 = (_a15 = result2.response) == null ? void 0 : _a15.id) != null ? _b2 : generateId3(),
2359
2357
  timestamp: (_d = (_c = result2.response) == null ? void 0 : _c.timestamp) != null ? _d : currentDate(),
2360
2358
  modelId: (_f = (_e = result2.response) == null ? void 0 : _e.modelId) != null ? _f : model.modelId
2361
2359
  };
@@ -2463,7 +2461,7 @@ async function generateObject({
2463
2461
  });
2464
2462
  const objectText = (_b2 = (_a15 = result2.toolCalls) == null ? void 0 : _a15[0]) == null ? void 0 : _b2.args;
2465
2463
  const responseData = {
2466
- id: (_d = (_c = result2.response) == null ? void 0 : _c.id) != null ? _d : generateId4(),
2464
+ id: (_d = (_c = result2.response) == null ? void 0 : _c.id) != null ? _d : generateId3(),
2467
2465
  timestamp: (_f = (_e = result2.response) == null ? void 0 : _e.timestamp) != null ? _f : currentDate(),
2468
2466
  modelId: (_h = (_g = result2.response) == null ? void 0 : _g.modelId) != null ? _h : model.modelId
2469
2467
  };
@@ -2764,7 +2762,7 @@ function streamObject({
2764
2762
  experimental_providerMetadata: providerMetadata,
2765
2763
  onFinish,
2766
2764
  _internal: {
2767
- generateId: generateId4 = originalGenerateId2,
2765
+ generateId: generateId3 = originalGenerateId2,
2768
2766
  currentDate = () => /* @__PURE__ */ new Date(),
2769
2767
  now: now2 = now
2770
2768
  } = {},
@@ -2797,7 +2795,7 @@ function streamObject({
2797
2795
  inputProviderMetadata: providerMetadata,
2798
2796
  mode,
2799
2797
  onFinish,
2800
- generateId: generateId4,
2798
+ generateId: generateId3,
2801
2799
  currentDate,
2802
2800
  now: now2
2803
2801
  });
@@ -2819,7 +2817,7 @@ var DefaultStreamObjectResult = class {
2819
2817
  inputProviderMetadata,
2820
2818
  mode,
2821
2819
  onFinish,
2822
- generateId: generateId4,
2820
+ generateId: generateId3,
2823
2821
  currentDate,
2824
2822
  now: now2
2825
2823
  }) {
@@ -3020,7 +3018,7 @@ var DefaultStreamObjectResult = class {
3020
3018
  let accumulatedText = "";
3021
3019
  let textDelta = "";
3022
3020
  let response = {
3023
- id: generateId4(),
3021
+ id: generateId3(),
3024
3022
  timestamp: currentDate(),
3025
3023
  modelId: model.modelId
3026
3024
  };
@@ -3561,16 +3559,20 @@ function toResponseMessages({
3561
3559
  text: text2 = "",
3562
3560
  tools,
3563
3561
  toolCalls,
3564
- toolResults
3562
+ toolResults,
3563
+ messageId,
3564
+ generateMessageId
3565
3565
  }) {
3566
3566
  const responseMessages = [];
3567
3567
  responseMessages.push({
3568
3568
  role: "assistant",
3569
- content: [{ type: "text", text: text2 }, ...toolCalls]
3569
+ content: [{ type: "text", text: text2 }, ...toolCalls],
3570
+ id: messageId
3570
3571
  });
3571
3572
  if (toolResults.length > 0) {
3572
3573
  responseMessages.push({
3573
3574
  role: "tool",
3575
+ id: generateMessageId(),
3574
3576
  content: toolResults.map((toolResult) => {
3575
3577
  const tool2 = tools[toolResult.toolName];
3576
3578
  return (tool2 == null ? void 0 : tool2.experimental_toToolResultContent) != null ? {
@@ -3594,7 +3596,14 @@ function toResponseMessages({
3594
3596
  }
3595
3597
 
3596
3598
  // core/generate-text/generate-text.ts
3597
- var originalGenerateId3 = createIdGenerator3({ prefix: "aitxt", size: 24 });
3599
+ var originalGenerateId3 = createIdGenerator3({
3600
+ prefix: "aitxt",
3601
+ size: 24
3602
+ });
3603
+ var originalGenerateMessageId = createIdGenerator3({
3604
+ prefix: "msg",
3605
+ size: 24
3606
+ });
3598
3607
  async function generateText({
3599
3608
  model,
3600
3609
  tools,
@@ -3606,6 +3615,7 @@ async function generateText({
3606
3615
  abortSignal,
3607
3616
  headers,
3608
3617
  maxSteps = 1,
3618
+ experimental_generateMessageId: generateMessageId = originalGenerateMessageId,
3609
3619
  experimental_output: output,
3610
3620
  experimental_continueSteps: continueSteps = false,
3611
3621
  experimental_telemetry: telemetry,
@@ -3613,7 +3623,7 @@ async function generateText({
3613
3623
  experimental_activeTools: activeTools,
3614
3624
  experimental_repairToolCall: repairToolCall,
3615
3625
  _internal: {
3616
- generateId: generateId4 = originalGenerateId3,
3626
+ generateId: generateId3 = originalGenerateId3,
3617
3627
  currentDate = () => /* @__PURE__ */ new Date()
3618
3628
  } = {},
3619
3629
  onStepFinish,
@@ -3747,7 +3757,7 @@ async function generateText({
3747
3757
  headers
3748
3758
  });
3749
3759
  const responseData = {
3750
- id: (_b2 = (_a16 = result.response) == null ? void 0 : _a16.id) != null ? _b2 : generateId4(),
3760
+ id: (_b2 = (_a16 = result.response) == null ? void 0 : _a16.id) != null ? _b2 : generateId3(),
3751
3761
  timestamp: (_d2 = (_c2 = result.response) == null ? void 0 : _c2.timestamp) != null ? _d2 : currentDate(),
3752
3762
  modelId: (_f2 = (_e2 = result.response) == null ? void 0 : _e2.modelId) != null ? _f2 : model.modelId
3753
3763
  };
@@ -3837,7 +3847,9 @@ async function generateText({
3837
3847
  text: text2,
3838
3848
  tools: tools != null ? tools : {},
3839
3849
  toolCalls: currentToolCalls,
3840
- toolResults: currentToolResults
3850
+ toolResults: currentToolResults,
3851
+ messageId: generateMessageId(),
3852
+ generateMessageId
3841
3853
  })
3842
3854
  );
3843
3855
  }
@@ -3855,7 +3867,7 @@ async function generateText({
3855
3867
  ...currentModelResponse.response,
3856
3868
  headers: (_d = currentModelResponse.rawResponse) == null ? void 0 : _d.headers,
3857
3869
  // deep clone msgs to avoid mutating past messages in multi-step:
3858
- messages: JSON.parse(JSON.stringify(responseMessages))
3870
+ messages: structuredClone(responseMessages)
3859
3871
  },
3860
3872
  experimental_providerMetadata: currentModelResponse.providerMetadata,
3861
3873
  isContinued: nextStepType === "continue"
@@ -4159,6 +4171,11 @@ function smoothStream({
4159
4171
  import { createIdGenerator as createIdGenerator4 } from "@ai-sdk/provider-utils";
4160
4172
  import { formatDataStreamPart as formatDataStreamPart2 } from "@ai-sdk/ui-utils";
4161
4173
 
4174
+ // util/as-array.ts
4175
+ function asArray(value) {
4176
+ return value === void 0 ? [] : Array.isArray(value) ? value : [value];
4177
+ }
4178
+
4162
4179
  // core/util/merge-streams.ts
4163
4180
  function mergeStreams(stream1, stream2) {
4164
4181
  const reader1 = stream1.getReader();
@@ -4438,7 +4455,14 @@ function runToolsTransformation({
4438
4455
  }
4439
4456
 
4440
4457
  // core/generate-text/stream-text.ts
4441
- var originalGenerateId4 = createIdGenerator4({ prefix: "aitxt", size: 24 });
4458
+ var originalGenerateId4 = createIdGenerator4({
4459
+ prefix: "aitxt",
4460
+ size: 24
4461
+ });
4462
+ var originalGenerateMessageId2 = createIdGenerator4({
4463
+ prefix: "msg",
4464
+ size: 24
4465
+ });
4442
4466
  function streamText({
4443
4467
  model,
4444
4468
  tools,
@@ -4450,6 +4474,7 @@ function streamText({
4450
4474
  abortSignal,
4451
4475
  headers,
4452
4476
  maxSteps = 1,
4477
+ experimental_generateMessageId: generateMessageId = originalGenerateMessageId2,
4453
4478
  experimental_output: output,
4454
4479
  experimental_continueSteps: continueSteps = false,
4455
4480
  experimental_telemetry: telemetry,
@@ -4463,7 +4488,7 @@ function streamText({
4463
4488
  onStepFinish,
4464
4489
  _internal: {
4465
4490
  now: now2 = now,
4466
- generateId: generateId4 = originalGenerateId4,
4491
+ generateId: generateId3 = originalGenerateId4,
4467
4492
  currentDate = () => /* @__PURE__ */ new Date()
4468
4493
  } = {},
4469
4494
  ...settings
@@ -4481,7 +4506,7 @@ function streamText({
4481
4506
  tools,
4482
4507
  toolChoice,
4483
4508
  toolCallStreaming,
4484
- transforms: transform == null ? [] : Array.isArray(transform) ? transform : [transform],
4509
+ transforms: asArray(transform),
4485
4510
  activeTools,
4486
4511
  repairToolCall,
4487
4512
  maxSteps,
@@ -4493,7 +4518,8 @@ function streamText({
4493
4518
  onStepFinish,
4494
4519
  now: now2,
4495
4520
  currentDate,
4496
- generateId: generateId4
4521
+ generateId: generateId3,
4522
+ generateMessageId
4497
4523
  });
4498
4524
  }
4499
4525
  function createOutputTransformStream(output) {
@@ -4573,7 +4599,8 @@ var DefaultStreamTextResult = class {
4573
4599
  onStepFinish,
4574
4600
  now: now2,
4575
4601
  currentDate,
4576
- generateId: generateId4
4602
+ generateId: generateId3,
4603
+ generateMessageId
4577
4604
  }) {
4578
4605
  this.warningsPromise = new DelayedPromise();
4579
4606
  this.usagePromise = new DelayedPromise();
@@ -4598,7 +4625,7 @@ var DefaultStreamTextResult = class {
4598
4625
  let recordedContinuationText = "";
4599
4626
  let recordedFullText = "";
4600
4627
  const recordedResponse = {
4601
- id: generateId4(),
4628
+ id: generateId3(),
4602
4629
  timestamp: currentDate(),
4603
4630
  modelId: model.modelId,
4604
4631
  messages: []
@@ -4633,7 +4660,9 @@ var DefaultStreamTextResult = class {
4633
4660
  text: recordedContinuationText,
4634
4661
  tools: tools != null ? tools : {},
4635
4662
  toolCalls: recordedToolCalls,
4636
- toolResults: recordedToolResults
4663
+ toolResults: recordedToolResults,
4664
+ messageId: part.messageId,
4665
+ generateMessageId
4637
4666
  });
4638
4667
  const currentStep = recordedSteps.length;
4639
4668
  let nextStepType = "done";
@@ -4808,7 +4837,8 @@ var DefaultStreamTextResult = class {
4808
4837
  usage,
4809
4838
  stepType: stepType2,
4810
4839
  previousStepText,
4811
- hasLeadingWhitespace
4840
+ hasLeadingWhitespace,
4841
+ messageId
4812
4842
  }) {
4813
4843
  const promptFormat = responseMessages.length === 0 ? initialPrompt.type : "messages";
4814
4844
  const stepInputMessages = [
@@ -4916,7 +4946,7 @@ var DefaultStreamTextResult = class {
4916
4946
  let fullStepText = stepType2 === "continue" ? previousStepText : "";
4917
4947
  let stepLogProbs;
4918
4948
  let stepResponse = {
4919
- id: generateId4(),
4949
+ id: generateId3(),
4920
4950
  timestamp: currentDate(),
4921
4951
  modelId: model.modelId
4922
4952
  };
@@ -4948,6 +4978,12 @@ var DefaultStreamTextResult = class {
4948
4978
  doStreamSpan.setAttributes({
4949
4979
  "ai.response.msToFirstChunk": msToFirstChunk
4950
4980
  });
4981
+ controller.enqueue({
4982
+ type: "step-start",
4983
+ messageId,
4984
+ request: stepRequest,
4985
+ warnings: warnings != null ? warnings : []
4986
+ });
4951
4987
  }
4952
4988
  if (chunk.type === "text-delta" && chunk.textDelta.length === 0) {
4953
4989
  return;
@@ -5092,7 +5128,8 @@ var DefaultStreamTextResult = class {
5092
5128
  headers: rawResponse == null ? void 0 : rawResponse.headers
5093
5129
  },
5094
5130
  warnings,
5095
- isContinued: nextStepType === "continue"
5131
+ isContinued: nextStepType === "continue",
5132
+ messageId
5096
5133
  });
5097
5134
  const combinedUsage = addLanguageModelUsage(usage, stepUsage);
5098
5135
  if (nextStepType === "done") {
@@ -5125,7 +5162,9 @@ var DefaultStreamTextResult = class {
5125
5162
  text: stepText,
5126
5163
  tools: tools != null ? tools : {},
5127
5164
  toolCalls: stepToolCalls,
5128
- toolResults: stepToolResults
5165
+ toolResults: stepToolResults,
5166
+ messageId,
5167
+ generateMessageId
5129
5168
  })
5130
5169
  );
5131
5170
  }
@@ -5135,7 +5174,11 @@ var DefaultStreamTextResult = class {
5135
5174
  usage: combinedUsage,
5136
5175
  stepType: nextStepType,
5137
5176
  previousStepText: fullStepText,
5138
- hasLeadingWhitespace: hasWhitespaceSuffix
5177
+ hasLeadingWhitespace: hasWhitespaceSuffix,
5178
+ messageId: (
5179
+ // keep the same id when continuing a step:
5180
+ nextStepType === "continue" ? messageId : generateMessageId()
5181
+ )
5139
5182
  });
5140
5183
  }
5141
5184
  }
@@ -5153,7 +5196,8 @@ var DefaultStreamTextResult = class {
5153
5196
  },
5154
5197
  previousStepText: "",
5155
5198
  stepType: "initial",
5156
- hasLeadingWhitespace: false
5199
+ hasLeadingWhitespace: false,
5200
+ messageId: generateMessageId()
5157
5201
  });
5158
5202
  }
5159
5203
  }).catch((error) => {
@@ -5318,6 +5362,14 @@ var DefaultStreamTextResult = class {
5318
5362
  );
5319
5363
  break;
5320
5364
  }
5365
+ case "step-start": {
5366
+ controller.enqueue(
5367
+ formatDataStreamPart2("start_step", {
5368
+ messageId: chunk.messageId
5369
+ })
5370
+ );
5371
+ break;
5372
+ }
5321
5373
  case "step-finish": {
5322
5374
  controller.enqueue(
5323
5375
  formatDataStreamPart2("finish_step", {
@@ -5463,7 +5515,6 @@ var experimental_wrapLanguageModel = ({
5463
5515
  };
5464
5516
 
5465
5517
  // core/prompt/append-response-messages.ts
5466
- import { generateId as generateId2 } from "@ai-sdk/provider-utils";
5467
5518
  function appendResponseMessages({
5468
5519
  messages,
5469
5520
  responseMessages
@@ -5473,17 +5524,10 @@ function appendResponseMessages({
5473
5524
  for (const message of responseMessages) {
5474
5525
  const role = message.role;
5475
5526
  switch (role) {
5476
- case "system":
5477
- case "user": {
5478
- throw new Error(
5479
- "AI response must not contain system or user messages: " + role
5480
- );
5481
- }
5482
5527
  case "assistant": {
5483
5528
  clonedMessages.push({
5484
5529
  role: "assistant",
5485
- id: generateId2(),
5486
- // generate an id for the message, will be overridden by the client
5530
+ id: message.id,
5487
5531
  createdAt: /* @__PURE__ */ new Date(),
5488
5532
  // generate a createdAt date for the message, will be overridden by the client
5489
5533
  // only include text in the content:
@@ -5517,6 +5561,10 @@ function appendResponseMessages({
5517
5561
  }
5518
5562
  break;
5519
5563
  }
5564
+ default: {
5565
+ const _exhaustiveCheck = role;
5566
+ throw new Error(`Unsupported message role: ${_exhaustiveCheck}`);
5567
+ }
5520
5568
  }
5521
5569
  }
5522
5570
  return clonedMessages;
@@ -5708,7 +5756,7 @@ function simulateReadableStream({
5708
5756
 
5709
5757
  // streams/assistant-response.ts
5710
5758
  import {
5711
- formatAssistantStreamPart
5759
+ formatAssistantStreamPart as formatAssistantStreamPart2
5712
5760
  } from "@ai-sdk/ui-utils";
5713
5761
  function AssistantResponse({ threadId, messageId }, process2) {
5714
5762
  const stream = new ReadableStream({
@@ -5718,20 +5766,20 @@ function AssistantResponse({ threadId, messageId }, process2) {
5718
5766
  const sendMessage = (message) => {
5719
5767
  controller.enqueue(
5720
5768
  textEncoder.encode(
5721
- formatAssistantStreamPart("assistant_message", message)
5769
+ formatAssistantStreamPart2("assistant_message", message)
5722
5770
  )
5723
5771
  );
5724
5772
  };
5725
5773
  const sendDataMessage = (message) => {
5726
5774
  controller.enqueue(
5727
5775
  textEncoder.encode(
5728
- formatAssistantStreamPart("data_message", message)
5776
+ formatAssistantStreamPart2("data_message", message)
5729
5777
  )
5730
5778
  );
5731
5779
  };
5732
5780
  const sendError = (errorMessage) => {
5733
5781
  controller.enqueue(
5734
- textEncoder.encode(formatAssistantStreamPart("error", errorMessage))
5782
+ textEncoder.encode(formatAssistantStreamPart2("error", errorMessage))
5735
5783
  );
5736
5784
  };
5737
5785
  const forwardStream = async (stream2) => {
@@ -5742,7 +5790,7 @@ function AssistantResponse({ threadId, messageId }, process2) {
5742
5790
  case "thread.message.created": {
5743
5791
  controller.enqueue(
5744
5792
  textEncoder.encode(
5745
- formatAssistantStreamPart("assistant_message", {
5793
+ formatAssistantStreamPart2("assistant_message", {
5746
5794
  id: value.data.id,
5747
5795
  role: "assistant",
5748
5796
  content: [{ type: "text", text: { value: "" } }]
@@ -5756,7 +5804,7 @@ function AssistantResponse({ threadId, messageId }, process2) {
5756
5804
  if ((content == null ? void 0 : content.type) === "text" && ((_b = content.text) == null ? void 0 : _b.value) != null) {
5757
5805
  controller.enqueue(
5758
5806
  textEncoder.encode(
5759
- formatAssistantStreamPart("text", content.text.value)
5807
+ formatAssistantStreamPart2("text", content.text.value)
5760
5808
  )
5761
5809
  );
5762
5810
  }
@@ -5773,7 +5821,7 @@ function AssistantResponse({ threadId, messageId }, process2) {
5773
5821
  };
5774
5822
  controller.enqueue(
5775
5823
  textEncoder.encode(
5776
- formatAssistantStreamPart("assistant_control_data", {
5824
+ formatAssistantStreamPart2("assistant_control_data", {
5777
5825
  threadId,
5778
5826
  messageId
5779
5827
  })
@@ -5811,7 +5859,7 @@ __export(langchain_adapter_exports, {
5811
5859
  toDataStream: () => toDataStream,
5812
5860
  toDataStreamResponse: () => toDataStreamResponse
5813
5861
  });
5814
- import { formatDataStreamPart as formatDataStreamPart3 } from "@ai-sdk/ui-utils";
5862
+ import { formatDataStreamPart as formatDataStreamPart4 } from "@ai-sdk/ui-utils";
5815
5863
 
5816
5864
  // streams/stream-callbacks.ts
5817
5865
  function createCallbacksTransformer(callbacks = {}) {
@@ -5867,7 +5915,7 @@ function toDataStreamInternal(stream, callbacks) {
5867
5915
  ).pipeThrough(createCallbacksTransformer(callbacks)).pipeThrough(new TextDecoderStream()).pipeThrough(
5868
5916
  new TransformStream({
5869
5917
  transform: async (chunk, controller) => {
5870
- controller.enqueue(formatDataStreamPart3("text", chunk));
5918
+ controller.enqueue(formatDataStreamPart4("text", chunk));
5871
5919
  }
5872
5920
  })
5873
5921
  );
@@ -5919,7 +5967,7 @@ __export(llamaindex_adapter_exports, {
5919
5967
  toDataStreamResponse: () => toDataStreamResponse2
5920
5968
  });
5921
5969
  import { convertAsyncIteratorToReadableStream } from "@ai-sdk/provider-utils";
5922
- import { formatDataStreamPart as formatDataStreamPart4 } from "@ai-sdk/ui-utils";
5970
+ import { formatDataStreamPart as formatDataStreamPart5 } from "@ai-sdk/ui-utils";
5923
5971
  function toDataStreamInternal2(stream, callbacks) {
5924
5972
  const trimStart = trimStartOfStream();
5925
5973
  return convertAsyncIteratorToReadableStream(stream[Symbol.asyncIterator]()).pipeThrough(
@@ -5931,7 +5979,7 @@ function toDataStreamInternal2(stream, callbacks) {
5931
5979
  ).pipeThrough(createCallbacksTransformer(callbacks)).pipeThrough(new TextDecoderStream()).pipeThrough(
5932
5980
  new TransformStream({
5933
5981
  transform: async (chunk, controller) => {
5934
- controller.enqueue(formatDataStreamPart4("text", chunk));
5982
+ controller.enqueue(formatDataStreamPart5("text", chunk));
5935
5983
  }
5936
5984
  })
5937
5985
  );
@@ -5973,7 +6021,7 @@ function trimStartOfStream() {
5973
6021
  }
5974
6022
 
5975
6023
  // streams/stream-data.ts
5976
- import { formatDataStreamPart as formatDataStreamPart5 } from "@ai-sdk/ui-utils";
6024
+ import { formatDataStreamPart as formatDataStreamPart6 } from "@ai-sdk/ui-utils";
5977
6025
 
5978
6026
  // util/constants.ts
5979
6027
  var HANGING_STREAM_WARNING_TIME_MS = 15 * 1e3;
@@ -6025,7 +6073,7 @@ var StreamData = class {
6025
6073
  throw new Error("Stream controller is not initialized.");
6026
6074
  }
6027
6075
  this.controller.enqueue(
6028
- this.encoder.encode(formatDataStreamPart5("data", [value]))
6076
+ this.encoder.encode(formatDataStreamPart6("data", [value]))
6029
6077
  );
6030
6078
  }
6031
6079
  appendMessageAnnotation(value) {
@@ -6036,7 +6084,7 @@ var StreamData = class {
6036
6084
  throw new Error("Stream controller is not initialized.");
6037
6085
  }
6038
6086
  this.controller.enqueue(
6039
- this.encoder.encode(formatDataStreamPart5("message_annotations", [value]))
6087
+ this.encoder.encode(formatDataStreamPart6("message_annotations", [value]))
6040
6088
  );
6041
6089
  }
6042
6090
  };
@@ -6075,15 +6123,16 @@ export {
6075
6123
  cosineSimilarity,
6076
6124
  createDataStream,
6077
6125
  createDataStreamResponse,
6126
+ createIdGenerator5 as createIdGenerator,
6078
6127
  embed,
6079
6128
  embedMany,
6080
6129
  experimental_createProviderRegistry,
6081
6130
  experimental_customProvider,
6082
6131
  generateImage as experimental_generateImage,
6083
6132
  experimental_wrapLanguageModel,
6084
- formatAssistantStreamPart2 as formatAssistantStreamPart,
6085
- formatDataStreamPart6 as formatDataStreamPart,
6086
- generateId3 as generateId,
6133
+ formatAssistantStreamPart,
6134
+ formatDataStreamPart3 as formatDataStreamPart,
6135
+ generateId2 as generateId,
6087
6136
  generateObject,
6088
6137
  generateText,
6089
6138
  jsonSchema,