ai 5.0.0-canary.13 → 5.0.0-canary.15

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -1,9 +1,7 @@
1
1
  "use strict";
2
- var __create = Object.create;
3
2
  var __defProp = Object.defineProperty;
4
3
  var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
5
4
  var __getOwnPropNames = Object.getOwnPropertyNames;
6
- var __getProtoOf = Object.getPrototypeOf;
7
5
  var __hasOwnProp = Object.prototype.hasOwnProperty;
8
6
  var __export = (target, all) => {
9
7
  for (var name17 in all)
@@ -17,14 +15,6 @@ var __copyProps = (to, from, except, desc) => {
17
15
  }
18
16
  return to;
19
17
  };
20
- var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__getProtoOf(mod)) : {}, __copyProps(
21
- // If the importer is in node compatibility mode or this is not an ESM
22
- // file that has been converted to a CommonJS file using a Babel-
23
- // compatible transform (i.e. "__esModule" has not been set), then set
24
- // "default" to the CommonJS "module.exports" for node compatibility.
25
- isNodeMode || !mod || !mod.__esModule ? __defProp(target, "default", { value: mod, enumerable: true }) : target,
26
- mod
27
- ));
28
18
  var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
29
19
 
30
20
  // index.ts
@@ -61,7 +51,7 @@ __export(ai_exports, {
61
51
  UnsupportedFunctionalityError: () => import_provider21.UnsupportedFunctionalityError,
62
52
  appendClientMessage: () => appendClientMessage,
63
53
  appendResponseMessages: () => appendResponseMessages,
64
- asSchema: () => asSchema,
54
+ asSchema: () => import_provider_utils4.asSchema,
65
55
  callChatApi: () => callChatApi,
66
56
  callCompletionApi: () => callCompletionApi,
67
57
  convertToCoreMessages: () => convertToCoreMessages,
@@ -73,7 +63,7 @@ __export(ai_exports, {
73
63
  cosineSimilarity: () => cosineSimilarity,
74
64
  createDataStream: () => createDataStream,
75
65
  createDataStreamResponse: () => createDataStreamResponse,
76
- createIdGenerator: () => import_provider_utils22.createIdGenerator,
66
+ createIdGenerator: () => import_provider_utils21.createIdGenerator,
77
67
  createProviderRegistry: () => createProviderRegistry,
78
68
  customProvider: () => customProvider,
79
69
  defaultSettingsMiddleware: () => defaultSettingsMiddleware,
@@ -89,14 +79,14 @@ __export(ai_exports, {
89
79
  extractReasoningMiddleware: () => extractReasoningMiddleware,
90
80
  fillMessageParts: () => fillMessageParts,
91
81
  formatDataStreamPart: () => formatDataStreamPart,
92
- generateId: () => import_provider_utils22.generateId,
82
+ generateId: () => import_provider_utils21.generateId,
93
83
  generateObject: () => generateObject,
94
84
  generateText: () => generateText,
95
85
  getMessageParts: () => getMessageParts,
96
86
  getTextFromDataUrl: () => getTextFromDataUrl,
97
87
  isAssistantMessageWithCompletedToolCalls: () => isAssistantMessageWithCompletedToolCalls,
98
88
  isDeepEqualData: () => isDeepEqualData,
99
- jsonSchema: () => jsonSchema,
89
+ jsonSchema: () => import_provider_utils4.jsonSchema,
100
90
  parseDataStreamPart: () => parseDataStreamPart,
101
91
  parsePartialJson: () => parsePartialJson,
102
92
  pipeDataStreamToResponse: () => pipeDataStreamToResponse,
@@ -111,16 +101,15 @@ __export(ai_exports, {
111
101
  streamText: () => streamText,
112
102
  tool: () => tool,
113
103
  updateToolCallResult: () => updateToolCallResult,
114
- wrapLanguageModel: () => wrapLanguageModel,
115
- zodSchema: () => zodSchema
104
+ wrapLanguageModel: () => wrapLanguageModel
116
105
  });
117
106
  module.exports = __toCommonJS(ai_exports);
118
107
 
119
108
  // core/index.ts
120
- var import_provider_utils22 = require("@ai-sdk/provider-utils");
109
+ var import_provider_utils21 = require("@ai-sdk/provider-utils");
121
110
 
122
111
  // core/util/index.ts
123
- var import_provider_utils5 = require("@ai-sdk/provider-utils");
112
+ var import_provider_utils4 = require("@ai-sdk/provider-utils");
124
113
 
125
114
  // core/util/process-chat-response.ts
126
115
  var import_provider_utils2 = require("@ai-sdk/provider-utils");
@@ -935,7 +924,6 @@ async function processChatResponse({
935
924
  execUpdate();
936
925
  },
937
926
  onReasoningPart(value) {
938
- var _a18;
939
927
  if (currentReasoningPart == null) {
940
928
  currentReasoningPart = {
941
929
  type: "reasoning",
@@ -947,7 +935,6 @@ async function processChatResponse({
947
935
  currentReasoningPart.reasoning += value.text;
948
936
  currentReasoningPart.providerMetadata = value.providerMetadata;
949
937
  }
950
- message.reasoning = ((_a18 = message.reasoning) != null ? _a18 : "") + value.text;
951
938
  execUpdate();
952
939
  },
953
940
  onReasoningPartFinish(value) {
@@ -1370,12 +1357,6 @@ function getMessageParts(message) {
1370
1357
  type: "tool-invocation",
1371
1358
  toolInvocation
1372
1359
  })) : [],
1373
- ...message.reasoning ? [
1374
- {
1375
- type: "reasoning",
1376
- reasoning: message.reasoning
1377
- }
1378
- ] : [],
1379
1360
  ...message.content ? [{ type: "text", text: message.content }] : []
1380
1361
  ];
1381
1362
  }
@@ -1455,51 +1436,29 @@ async function prepareAttachmentsForRequest(attachmentsFromOptions) {
1455
1436
  throw new Error("Invalid attachments type");
1456
1437
  }
1457
1438
 
1458
- // core/util/schema.ts
1459
- var import_provider_utils4 = require("@ai-sdk/provider-utils");
1460
-
1461
- // core/util/zod-schema.ts
1462
- var import_zod_to_json_schema = __toESM(require("zod-to-json-schema"));
1463
- function zodSchema(zodSchema2, options) {
1439
+ // core/util/update-tool-call-result.ts
1440
+ function updateToolCallResult({
1441
+ messages,
1442
+ toolCallId,
1443
+ toolResult: result
1444
+ }) {
1464
1445
  var _a17;
1465
- const useReferences = (_a17 = options == null ? void 0 : options.useReferences) != null ? _a17 : false;
1466
- return jsonSchema(
1467
- (0, import_zod_to_json_schema.default)(zodSchema2, {
1468
- $refStrategy: useReferences ? "root" : "none",
1469
- target: "jsonSchema7"
1470
- // note: openai mode breaks various gemini conversions
1471
- }),
1472
- {
1473
- validate: (value) => {
1474
- const result = zodSchema2.safeParse(value);
1475
- return result.success ? { success: true, value: result.data } : { success: false, error: result.error };
1476
- }
1477
- }
1446
+ const lastMessage = messages[messages.length - 1];
1447
+ const invocationPart = lastMessage.parts.find(
1448
+ (part) => part.type === "tool-invocation" && part.toolInvocation.toolCallId === toolCallId
1478
1449
  );
1479
- }
1480
-
1481
- // core/util/schema.ts
1482
- var schemaSymbol = Symbol.for("vercel.ai.schema");
1483
- function jsonSchema(jsonSchema2, {
1484
- validate
1485
- } = {}) {
1486
- return {
1487
- [schemaSymbol]: true,
1488
- _type: void 0,
1489
- // should never be used directly
1490
- [import_provider_utils4.validatorSymbol]: true,
1491
- jsonSchema: jsonSchema2,
1492
- validate
1450
+ if (invocationPart == null) {
1451
+ return;
1452
+ }
1453
+ const toolResult = {
1454
+ ...invocationPart.toolInvocation,
1455
+ state: "result",
1456
+ result
1493
1457
  };
1494
- }
1495
- function isSchema(value) {
1496
- return typeof value === "object" && value !== null && schemaSymbol in value && value[schemaSymbol] === true && "jsonSchema" in value && "validate" in value;
1497
- }
1498
- function asSchema(schema) {
1499
- return schema == null ? jsonSchema({
1500
- properties: {},
1501
- additionalProperties: false
1502
- }) : isSchema(schema) ? schema : zodSchema(schema);
1458
+ invocationPart.toolInvocation = toolResult;
1459
+ lastMessage.toolInvocations = (_a17 = lastMessage.toolInvocations) == null ? void 0 : _a17.map(
1460
+ (toolInvocation) => toolInvocation.toolCallId === toolCallId ? toolResult : toolInvocation
1461
+ );
1503
1462
  }
1504
1463
 
1505
1464
  // core/util/should-resubmit-messages.ts
@@ -1531,31 +1490,6 @@ function isAssistantMessageWithCompletedToolCalls(message) {
1531
1490
  return lastStepToolInvocations.length > 0 && lastStepToolInvocations.every((part) => "result" in part.toolInvocation);
1532
1491
  }
1533
1492
 
1534
- // core/util/update-tool-call-result.ts
1535
- function updateToolCallResult({
1536
- messages,
1537
- toolCallId,
1538
- toolResult: result
1539
- }) {
1540
- var _a17;
1541
- const lastMessage = messages[messages.length - 1];
1542
- const invocationPart = lastMessage.parts.find(
1543
- (part) => part.type === "tool-invocation" && part.toolInvocation.toolCallId === toolCallId
1544
- );
1545
- if (invocationPart == null) {
1546
- return;
1547
- }
1548
- const toolResult = {
1549
- ...invocationPart.toolInvocation,
1550
- state: "result",
1551
- result
1552
- };
1553
- invocationPart.toolInvocation = toolResult;
1554
- lastMessage.toolInvocations = (_a17 = lastMessage.toolInvocations) == null ? void 0 : _a17.map(
1555
- (toolInvocation) => toolInvocation.toolCallId === toolCallId ? toolResult : toolInvocation
1556
- );
1557
- }
1558
-
1559
1493
  // core/data-stream/create-data-stream.ts
1560
1494
  function createDataStream({
1561
1495
  execute,
@@ -1764,7 +1698,7 @@ _a = symbol;
1764
1698
 
1765
1699
  // util/retry-with-exponential-backoff.ts
1766
1700
  var import_provider3 = require("@ai-sdk/provider");
1767
- var import_provider_utils6 = require("@ai-sdk/provider-utils");
1701
+ var import_provider_utils5 = require("@ai-sdk/provider-utils");
1768
1702
 
1769
1703
  // util/retry-error.ts
1770
1704
  var import_provider2 = require("@ai-sdk/provider");
@@ -1808,13 +1742,13 @@ async function _retryWithExponentialBackoff(f, {
1808
1742
  try {
1809
1743
  return await f();
1810
1744
  } catch (error) {
1811
- if ((0, import_provider_utils6.isAbortError)(error)) {
1745
+ if ((0, import_provider_utils5.isAbortError)(error)) {
1812
1746
  throw error;
1813
1747
  }
1814
1748
  if (maxRetries === 0) {
1815
1749
  throw error;
1816
1750
  }
1817
- const errorMessage = (0, import_provider_utils6.getErrorMessage)(error);
1751
+ const errorMessage = (0, import_provider_utils5.getErrorMessage)(error);
1818
1752
  const newErrors = [...errors, error];
1819
1753
  const tryNumber = newErrors.length;
1820
1754
  if (tryNumber > maxRetries) {
@@ -1825,7 +1759,7 @@ async function _retryWithExponentialBackoff(f, {
1825
1759
  });
1826
1760
  }
1827
1761
  if (error instanceof Error && import_provider3.APICallError.isInstance(error) && error.isRetryable === true && tryNumber <= maxRetries) {
1828
- await (0, import_provider_utils6.delay)(delayInMs);
1762
+ await (0, import_provider_utils5.delay)(delayInMs);
1829
1763
  return _retryWithExponentialBackoff(
1830
1764
  f,
1831
1765
  { maxRetries, delayInMs: backoffFactor * delayInMs, backoffFactor },
@@ -2215,8 +2149,8 @@ async function embedMany({
2215
2149
  }),
2216
2150
  tracer,
2217
2151
  fn: async (span) => {
2218
- const maxEmbeddingsPerCall = model.maxEmbeddingsPerCall;
2219
- if (maxEmbeddingsPerCall == null) {
2152
+ const maxEmbeddingsPerCall = await model.maxEmbeddingsPerCall;
2153
+ if (maxEmbeddingsPerCall == null || maxEmbeddingsPerCall === Infinity) {
2220
2154
  const { embeddings: embeddings2, usage, response } = await retry(() => {
2221
2155
  return recordSpan({
2222
2156
  name: "ai.embedMany.doEmbed",
@@ -2394,7 +2328,7 @@ var NoImageGeneratedError = class extends import_provider4.AISDKError {
2394
2328
  _a3 = symbol3;
2395
2329
 
2396
2330
  // core/generate-text/generated-file.ts
2397
- var import_provider_utils7 = require("@ai-sdk/provider-utils");
2331
+ var import_provider_utils6 = require("@ai-sdk/provider-utils");
2398
2332
  var DefaultGeneratedFile = class {
2399
2333
  constructor({
2400
2334
  data,
@@ -2408,14 +2342,14 @@ var DefaultGeneratedFile = class {
2408
2342
  // lazy conversion with caching to avoid unnecessary conversion overhead:
2409
2343
  get base64() {
2410
2344
  if (this.base64Data == null) {
2411
- this.base64Data = (0, import_provider_utils7.convertUint8ArrayToBase64)(this.uint8ArrayData);
2345
+ this.base64Data = (0, import_provider_utils6.convertUint8ArrayToBase64)(this.uint8ArrayData);
2412
2346
  }
2413
2347
  return this.base64Data;
2414
2348
  }
2415
2349
  // lazy conversion with caching to avoid unnecessary conversion overhead:
2416
2350
  get uint8Array() {
2417
2351
  if (this.uint8ArrayData == null) {
2418
- this.uint8ArrayData = (0, import_provider_utils7.convertBase64ToUint8Array)(this.base64Data);
2352
+ this.uint8ArrayData = (0, import_provider_utils6.convertBase64ToUint8Array)(this.base64Data);
2419
2353
  }
2420
2354
  return this.uint8ArrayData;
2421
2355
  }
@@ -2428,7 +2362,7 @@ var DefaultGeneratedFileWithType = class extends DefaultGeneratedFile {
2428
2362
  };
2429
2363
 
2430
2364
  // core/util/detect-media-type.ts
2431
- var import_provider_utils8 = require("@ai-sdk/provider-utils");
2365
+ var import_provider_utils7 = require("@ai-sdk/provider-utils");
2432
2366
  var imageMediaTypeSignatures = [
2433
2367
  {
2434
2368
  mediaType: "image/gif",
@@ -2535,7 +2469,7 @@ var audioMediaTypeSignatures = [
2535
2469
  }
2536
2470
  ];
2537
2471
  var stripID3 = (data) => {
2538
- const bytes = typeof data === "string" ? (0, import_provider_utils8.convertBase64ToUint8Array)(data) : data;
2472
+ const bytes = typeof data === "string" ? (0, import_provider_utils7.convertBase64ToUint8Array)(data) : data;
2539
2473
  const id3Size = (bytes[6] & 127) << 21 | (bytes[7] & 127) << 14 | (bytes[8] & 127) << 7 | bytes[9] & 127;
2540
2474
  return bytes.slice(id3Size + 10);
2541
2475
  };
@@ -2639,7 +2573,7 @@ var DefaultGenerateImageResult = class {
2639
2573
 
2640
2574
  // core/generate-object/generate-object.ts
2641
2575
  var import_provider13 = require("@ai-sdk/provider");
2642
- var import_provider_utils13 = require("@ai-sdk/provider-utils");
2576
+ var import_provider_utils12 = require("@ai-sdk/provider-utils");
2643
2577
 
2644
2578
  // errors/no-object-generated-error.ts
2645
2579
  var import_provider5 = require("@ai-sdk/provider");
@@ -2733,7 +2667,7 @@ async function download({ url }) {
2733
2667
 
2734
2668
  // core/prompt/data-content.ts
2735
2669
  var import_provider8 = require("@ai-sdk/provider");
2736
- var import_provider_utils9 = require("@ai-sdk/provider-utils");
2670
+ var import_provider_utils8 = require("@ai-sdk/provider-utils");
2737
2671
  var import_zod = require("zod");
2738
2672
 
2739
2673
  // core/prompt/invalid-data-content-error.ts
@@ -2820,9 +2754,9 @@ function convertDataContentToBase64String(content) {
2820
2754
  return content;
2821
2755
  }
2822
2756
  if (content instanceof ArrayBuffer) {
2823
- return (0, import_provider_utils9.convertUint8ArrayToBase64)(new Uint8Array(content));
2757
+ return (0, import_provider_utils8.convertUint8ArrayToBase64)(new Uint8Array(content));
2824
2758
  }
2825
- return (0, import_provider_utils9.convertUint8ArrayToBase64)(content);
2759
+ return (0, import_provider_utils8.convertUint8ArrayToBase64)(content);
2826
2760
  }
2827
2761
  function convertDataContentToUint8Array(content) {
2828
2762
  if (content instanceof Uint8Array) {
@@ -2830,7 +2764,7 @@ function convertDataContentToUint8Array(content) {
2830
2764
  }
2831
2765
  if (typeof content === "string") {
2832
2766
  try {
2833
- return (0, import_provider_utils9.convertBase64ToUint8Array)(content);
2767
+ return (0, import_provider_utils8.convertBase64ToUint8Array)(content);
2834
2768
  } catch (error) {
2835
2769
  throw new InvalidDataContentError({
2836
2770
  message: "Invalid data content. Content string is not a base64-encoded media.",
@@ -2874,7 +2808,7 @@ var InvalidMessageRoleError = class extends import_provider9.AISDKError {
2874
2808
  _a7 = symbol7;
2875
2809
 
2876
2810
  // core/prompt/convert-to-language-model-prompt.ts
2877
- var import_provider_utils10 = require("@ai-sdk/provider-utils");
2811
+ var import_provider_utils9 = require("@ai-sdk/provider-utils");
2878
2812
  async function convertToLanguageModelPrompt({
2879
2813
  prompt,
2880
2814
  supportedUrls,
@@ -3011,7 +2945,7 @@ async function downloadAssets(messages, downloadImplementation, supportedUrls) {
3011
2945
  }
3012
2946
  return { mediaType, data };
3013
2947
  }).filter(
3014
- (part) => part.data instanceof URL && part.mediaType != null && !(0, import_provider_utils10.isUrlSupported)({
2948
+ (part) => part.data instanceof URL && part.mediaType != null && !(0, import_provider_utils9.isUrlSupported)({
3015
2949
  url: part.data.toString(),
3016
2950
  mediaType: part.mediaType,
3017
2951
  supportedUrls
@@ -3182,7 +3116,7 @@ function prepareCallSettings({
3182
3116
 
3183
3117
  // core/prompt/standardize-prompt.ts
3184
3118
  var import_provider11 = require("@ai-sdk/provider");
3185
- var import_provider_utils11 = require("@ai-sdk/provider-utils");
3119
+ var import_provider_utils10 = require("@ai-sdk/provider-utils");
3186
3120
  var import_zod7 = require("zod");
3187
3121
 
3188
3122
  // core/prompt/attachments-to-parts.ts
@@ -3705,7 +3639,6 @@ async function standardizePrompt({
3705
3639
  });
3706
3640
  }
3707
3641
  return {
3708
- type: "prompt",
3709
3642
  system: prompt.system,
3710
3643
  messages: [
3711
3644
  {
@@ -3732,7 +3665,7 @@ async function standardizePrompt({
3732
3665
  message: "messages must not be empty"
3733
3666
  });
3734
3667
  }
3735
- const validationResult = await (0, import_provider_utils11.safeValidateTypes)({
3668
+ const validationResult = await (0, import_provider_utils10.safeValidateTypes)({
3736
3669
  value: messages,
3737
3670
  schema: import_zod7.z.array(coreMessageSchema)
3738
3671
  });
@@ -3744,7 +3677,6 @@ async function standardizePrompt({
3744
3677
  });
3745
3678
  }
3746
3679
  return {
3747
- type: "messages",
3748
3680
  messages,
3749
3681
  system: prompt.system
3750
3682
  };
@@ -3773,7 +3705,7 @@ function addLanguageModelUsage(usage1, usage2) {
3773
3705
 
3774
3706
  // core/generate-object/output-strategy.ts
3775
3707
  var import_provider12 = require("@ai-sdk/provider");
3776
- var import_provider_utils12 = require("@ai-sdk/provider-utils");
3708
+ var import_provider_utils11 = require("@ai-sdk/provider-utils");
3777
3709
 
3778
3710
  // core/util/async-iterable-stream.ts
3779
3711
  function createAsyncIterableStream(source) {
@@ -3829,7 +3761,7 @@ var objectOutputStrategy = (schema) => ({
3829
3761
  };
3830
3762
  },
3831
3763
  async validateFinalResult(value) {
3832
- return (0, import_provider_utils12.safeValidateTypes)({ value, schema });
3764
+ return (0, import_provider_utils11.safeValidateTypes)({ value, schema });
3833
3765
  },
3834
3766
  createElementStream() {
3835
3767
  throw new import_provider12.UnsupportedFunctionalityError({
@@ -3873,7 +3805,7 @@ var arrayOutputStrategy = (schema) => {
3873
3805
  const resultArray = [];
3874
3806
  for (let i = 0; i < inputArray.length; i++) {
3875
3807
  const element = inputArray[i];
3876
- const result = await (0, import_provider_utils12.safeValidateTypes)({ value: element, schema });
3808
+ const result = await (0, import_provider_utils11.safeValidateTypes)({ value: element, schema });
3877
3809
  if (i === inputArray.length - 1 && !isFinalDelta) {
3878
3810
  continue;
3879
3811
  }
@@ -3914,7 +3846,7 @@ var arrayOutputStrategy = (schema) => {
3914
3846
  }
3915
3847
  const inputArray = value.elements;
3916
3848
  for (const element of inputArray) {
3917
- const result = await (0, import_provider_utils12.safeValidateTypes)({ value: element, schema });
3849
+ const result = await (0, import_provider_utils11.safeValidateTypes)({ value: element, schema });
3918
3850
  if (!result.success) {
3919
3851
  return result;
3920
3852
  }
@@ -4006,9 +3938,9 @@ function getOutputStrategy({
4006
3938
  }) {
4007
3939
  switch (output) {
4008
3940
  case "object":
4009
- return objectOutputStrategy(asSchema(schema));
3941
+ return objectOutputStrategy((0, import_provider_utils4.asSchema)(schema));
4010
3942
  case "array":
4011
- return arrayOutputStrategy(asSchema(schema));
3943
+ return arrayOutputStrategy((0, import_provider_utils4.asSchema)(schema));
4012
3944
  case "enum":
4013
3945
  return enumOutputStrategy(enumValues);
4014
3946
  case "no-schema":
@@ -4139,30 +4071,32 @@ function validateObjectGenerationInput({
4139
4071
  }
4140
4072
 
4141
4073
  // core/generate-object/generate-object.ts
4142
- var originalGenerateId = (0, import_provider_utils13.createIdGenerator)({ prefix: "aiobj", size: 24 });
4143
- async function generateObject({
4144
- model,
4145
- enum: enumValues,
4146
- // rename bc enum is reserved by typescript
4147
- schema: inputSchema,
4148
- schemaName,
4149
- schemaDescription,
4150
- output = "object",
4151
- system,
4152
- prompt,
4153
- messages,
4154
- maxRetries: maxRetriesArg,
4155
- abortSignal,
4156
- headers,
4157
- experimental_repairText: repairText,
4158
- experimental_telemetry: telemetry,
4159
- providerOptions,
4160
- _internal: {
4161
- generateId: generateId3 = originalGenerateId,
4162
- currentDate = () => /* @__PURE__ */ new Date()
4163
- } = {},
4164
- ...settings
4165
- }) {
4074
+ var originalGenerateId = (0, import_provider_utils12.createIdGenerator)({ prefix: "aiobj", size: 24 });
4075
+ async function generateObject(options) {
4076
+ const {
4077
+ model,
4078
+ output = "object",
4079
+ system,
4080
+ prompt,
4081
+ messages,
4082
+ maxRetries: maxRetriesArg,
4083
+ abortSignal,
4084
+ headers,
4085
+ experimental_repairText: repairText,
4086
+ experimental_telemetry: telemetry,
4087
+ providerOptions,
4088
+ _internal: {
4089
+ generateId: generateId3 = originalGenerateId,
4090
+ currentDate = () => /* @__PURE__ */ new Date()
4091
+ } = {},
4092
+ ...settings
4093
+ } = options;
4094
+ const enumValues = "enum" in options ? options.enum : void 0;
4095
+ const {
4096
+ schema: inputSchema,
4097
+ schemaDescription,
4098
+ schemaName
4099
+ } = "schema" in options ? options : {};
4166
4100
  validateObjectGenerationInput({
4167
4101
  output,
4168
4102
  schema: inputSchema,
@@ -4220,7 +4154,7 @@ async function generateObject({
4220
4154
  });
4221
4155
  const promptMessages = await convertToLanguageModelPrompt({
4222
4156
  prompt: standardizedPrompt,
4223
- supportedUrls: await model.getSupportedUrls()
4157
+ supportedUrls: await model.supportedUrls
4224
4158
  });
4225
4159
  const generateResult = await retry(
4226
4160
  () => recordSpan({
@@ -4233,9 +4167,6 @@ async function generateObject({
4233
4167
  telemetry
4234
4168
  }),
4235
4169
  ...baseTelemetryAttributes,
4236
- "ai.prompt.format": {
4237
- input: () => standardizedPrompt.type
4238
- },
4239
4170
  "ai.prompt.messages": {
4240
4171
  input: () => JSON.stringify(promptMessages)
4241
4172
  },
@@ -4261,7 +4192,6 @@ async function generateObject({
4261
4192
  description: schemaDescription
4262
4193
  },
4263
4194
  ...prepareCallSettings(settings),
4264
- inputFormat: standardizedPrompt.type,
4265
4195
  prompt: promptMessages,
4266
4196
  providerOptions,
4267
4197
  abortSignal,
@@ -4316,7 +4246,7 @@ async function generateObject({
4316
4246
  request = (_a17 = generateResult.request) != null ? _a17 : {};
4317
4247
  response = generateResult.responseData;
4318
4248
  async function processResult(result2) {
4319
- const parseResult = await (0, import_provider_utils13.safeParseJSON)({ text: result2 });
4249
+ const parseResult = await (0, import_provider_utils12.safeParseJSON)({ text: result2 });
4320
4250
  if (!parseResult.success) {
4321
4251
  throw new NoObjectGeneratedError({
4322
4252
  message: "No object generated: could not parse the response.",
@@ -4412,7 +4342,7 @@ var DefaultGenerateObjectResult = class {
4412
4342
  };
4413
4343
 
4414
4344
  // core/generate-object/stream-object.ts
4415
- var import_provider_utils14 = require("@ai-sdk/provider-utils");
4345
+ var import_provider_utils13 = require("@ai-sdk/provider-utils");
4416
4346
 
4417
4347
  // util/delayed-promise.ts
4418
4348
  var DelayedPromise = class {
@@ -4556,30 +4486,33 @@ function now() {
4556
4486
  }
4557
4487
 
4558
4488
  // core/generate-object/stream-object.ts
4559
- var originalGenerateId2 = (0, import_provider_utils14.createIdGenerator)({ prefix: "aiobj", size: 24 });
4560
- function streamObject({
4561
- model,
4562
- schema: inputSchema,
4563
- schemaName,
4564
- schemaDescription,
4565
- output = "object",
4566
- system,
4567
- prompt,
4568
- messages,
4569
- maxRetries,
4570
- abortSignal,
4571
- headers,
4572
- experimental_telemetry: telemetry,
4573
- providerOptions,
4574
- onError,
4575
- onFinish,
4576
- _internal: {
4577
- generateId: generateId3 = originalGenerateId2,
4578
- currentDate = () => /* @__PURE__ */ new Date(),
4579
- now: now2 = now
4580
- } = {},
4581
- ...settings
4582
- }) {
4489
+ var originalGenerateId2 = (0, import_provider_utils13.createIdGenerator)({ prefix: "aiobj", size: 24 });
4490
+ function streamObject(options) {
4491
+ const {
4492
+ model,
4493
+ output = "object",
4494
+ system,
4495
+ prompt,
4496
+ messages,
4497
+ maxRetries,
4498
+ abortSignal,
4499
+ headers,
4500
+ experimental_telemetry: telemetry,
4501
+ providerOptions,
4502
+ onError,
4503
+ onFinish,
4504
+ _internal: {
4505
+ generateId: generateId3 = originalGenerateId2,
4506
+ currentDate = () => /* @__PURE__ */ new Date(),
4507
+ now: now2 = now
4508
+ } = {},
4509
+ ...settings
4510
+ } = options;
4511
+ const {
4512
+ schema: inputSchema,
4513
+ schemaDescription,
4514
+ schemaName
4515
+ } = "schema" in options ? options : {};
4583
4516
  validateObjectGenerationInput({
4584
4517
  output,
4585
4518
  schema: inputSchema,
@@ -4692,10 +4625,9 @@ var DefaultStreamObjectResult = class {
4692
4625
  description: schemaDescription
4693
4626
  },
4694
4627
  ...prepareCallSettings(settings),
4695
- inputFormat: standardizedPrompt.type,
4696
4628
  prompt: await convertToLanguageModelPrompt({
4697
4629
  prompt: standardizedPrompt,
4698
- supportedUrls: await model.getSupportedUrls()
4630
+ supportedUrls: await model.supportedUrls
4699
4631
  }),
4700
4632
  providerOptions,
4701
4633
  abortSignal,
@@ -4730,9 +4662,6 @@ var DefaultStreamObjectResult = class {
4730
4662
  telemetry
4731
4663
  }),
4732
4664
  ...baseTelemetryAttributes,
4733
- "ai.prompt.format": {
4734
- input: () => callOptions.inputFormat
4735
- },
4736
4665
  "ai.prompt.messages": {
4737
4666
  input: () => JSON.stringify(callOptions.prompt)
4738
4667
  },
@@ -5052,7 +4981,7 @@ var DefaultStreamObjectResult = class {
5052
4981
  };
5053
4982
 
5054
4983
  // core/generate-text/generate-text.ts
5055
- var import_provider_utils16 = require("@ai-sdk/provider-utils");
4984
+ var import_provider_utils15 = require("@ai-sdk/provider-utils");
5056
4985
 
5057
4986
  // errors/no-output-specified-error.ts
5058
4987
  var import_provider14 = require("@ai-sdk/provider");
@@ -5128,7 +5057,7 @@ function prepareToolsAndToolChoice({
5128
5057
  type: "function",
5129
5058
  name: name17,
5130
5059
  description: tool2.description,
5131
- parameters: asSchema(tool2.parameters).jsonSchema
5060
+ parameters: (0, import_provider_utils4.asSchema)(tool2.parameters).jsonSchema
5132
5061
  };
5133
5062
  case "provider-defined":
5134
5063
  return {
@@ -5161,7 +5090,7 @@ function removeTextAfterLastWhitespace(text2) {
5161
5090
  }
5162
5091
 
5163
5092
  // core/generate-text/parse-tool-call.ts
5164
- var import_provider_utils15 = require("@ai-sdk/provider-utils");
5093
+ var import_provider_utils14 = require("@ai-sdk/provider-utils");
5165
5094
 
5166
5095
  // errors/invalid-tool-arguments-error.ts
5167
5096
  var import_provider16 = require("@ai-sdk/provider");
@@ -5258,7 +5187,7 @@ async function parseToolCall({
5258
5187
  tools,
5259
5188
  parameterSchema: ({ toolName }) => {
5260
5189
  const { parameters } = tools[toolName];
5261
- return asSchema(parameters).jsonSchema;
5190
+ return (0, import_provider_utils4.asSchema)(parameters).jsonSchema;
5262
5191
  },
5263
5192
  system,
5264
5193
  messages,
@@ -5288,8 +5217,8 @@ async function doParseToolCall({
5288
5217
  availableTools: Object.keys(tools)
5289
5218
  });
5290
5219
  }
5291
- const schema = asSchema(tool2.parameters);
5292
- const parseResult = toolCall.args.trim() === "" ? await (0, import_provider_utils15.safeValidateTypes)({ value: {}, schema }) : await (0, import_provider_utils15.safeParseJSON)({ text: toolCall.args, schema });
5220
+ const schema = (0, import_provider_utils4.asSchema)(tool2.parameters);
5221
+ const parseResult = toolCall.args.trim() === "" ? await (0, import_provider_utils14.safeValidateTypes)({ value: {}, schema }) : await (0, import_provider_utils14.safeParseJSON)({ text: toolCall.args, schema });
5293
5222
  if (parseResult.success === false) {
5294
5223
  throw new InvalidToolArgumentsError({
5295
5224
  toolName,
@@ -5383,11 +5312,11 @@ function toResponseMessages({
5383
5312
  }
5384
5313
 
5385
5314
  // core/generate-text/generate-text.ts
5386
- var originalGenerateId3 = (0, import_provider_utils16.createIdGenerator)({
5315
+ var originalGenerateId3 = (0, import_provider_utils15.createIdGenerator)({
5387
5316
  prefix: "aitxt",
5388
5317
  size: 24
5389
5318
  });
5390
- var originalGenerateMessageId = (0, import_provider_utils16.createIdGenerator)({
5319
+ var originalGenerateMessageId = (0, import_provider_utils15.createIdGenerator)({
5391
5320
  prefix: "msg",
5392
5321
  size: 24
5393
5322
  });
@@ -5408,6 +5337,7 @@ async function generateText({
5408
5337
  experimental_telemetry: telemetry,
5409
5338
  providerOptions,
5410
5339
  experimental_activeTools: activeTools,
5340
+ experimental_prepareStep: prepareStep,
5411
5341
  experimental_repairToolCall: repairToolCall,
5412
5342
  _internal: {
5413
5343
  generateId: generateId3 = originalGenerateId3,
@@ -5446,6 +5376,9 @@ async function generateText({
5446
5376
  telemetry
5447
5377
  }),
5448
5378
  ...baseTelemetryAttributes,
5379
+ // model:
5380
+ "ai.model.provider": model.provider,
5381
+ "ai.model.id": model.modelId,
5449
5382
  // specific settings that only make sense on the outer level:
5450
5383
  "ai.prompt": {
5451
5384
  input: () => JSON.stringify({ system, prompt, messages })
@@ -5455,10 +5388,8 @@ async function generateText({
5455
5388
  }),
5456
5389
  tracer,
5457
5390
  fn: async (span) => {
5458
- var _a17, _b, _c;
5459
- const toolsAndToolChoice = {
5460
- ...prepareToolsAndToolChoice({ tools, toolChoice, activeTools })
5461
- };
5391
+ var _a17, _b, _c, _d, _e, _f;
5392
+ const callSettings2 = prepareCallSettings(settings);
5462
5393
  let currentModelResponse;
5463
5394
  let currentToolCalls = [];
5464
5395
  let currentToolResults = [];
@@ -5475,18 +5406,28 @@ async function generateText({
5475
5406
  };
5476
5407
  let stepType = "initial";
5477
5408
  do {
5478
- const promptFormat = stepCount === 0 ? initialPrompt.type : "messages";
5479
5409
  const stepInputMessages = [
5480
5410
  ...initialPrompt.messages,
5481
5411
  ...responseMessages
5482
5412
  ];
5413
+ const prepareStepResult = await (prepareStep == null ? void 0 : prepareStep({
5414
+ model,
5415
+ steps,
5416
+ maxSteps,
5417
+ stepNumber: stepCount
5418
+ }));
5483
5419
  const promptMessages = await convertToLanguageModelPrompt({
5484
5420
  prompt: {
5485
- type: promptFormat,
5486
5421
  system: initialPrompt.system,
5487
5422
  messages: stepInputMessages
5488
5423
  },
5489
- supportedUrls: await model.getSupportedUrls()
5424
+ supportedUrls: await model.supportedUrls
5425
+ });
5426
+ const stepModel = (_a17 = prepareStepResult == null ? void 0 : prepareStepResult.model) != null ? _a17 : model;
5427
+ const { toolChoice: stepToolChoice, tools: stepTools } = prepareToolsAndToolChoice({
5428
+ tools,
5429
+ toolChoice: (_b = prepareStepResult == null ? void 0 : prepareStepResult.toolChoice) != null ? _b : toolChoice,
5430
+ activeTools: (_c = prepareStepResult == null ? void 0 : prepareStepResult.experimental_activeTools) != null ? _c : activeTools
5490
5431
  });
5491
5432
  currentModelResponse = await retry(
5492
5433
  () => {
@@ -5501,23 +5442,23 @@ async function generateText({
5501
5442
  telemetry
5502
5443
  }),
5503
5444
  ...baseTelemetryAttributes,
5504
- "ai.prompt.format": { input: () => promptFormat },
5445
+ // model:
5446
+ "ai.model.provider": stepModel.provider,
5447
+ "ai.model.id": stepModel.modelId,
5448
+ // prompt:
5505
5449
  "ai.prompt.messages": {
5506
5450
  input: () => JSON.stringify(promptMessages)
5507
5451
  },
5508
5452
  "ai.prompt.tools": {
5509
5453
  // convert the language model level tools:
5510
- input: () => {
5511
- var _a19;
5512
- return (_a19 = toolsAndToolChoice.tools) == null ? void 0 : _a19.map((tool2) => JSON.stringify(tool2));
5513
- }
5454
+ input: () => stepTools == null ? void 0 : stepTools.map((tool2) => JSON.stringify(tool2))
5514
5455
  },
5515
5456
  "ai.prompt.toolChoice": {
5516
- input: () => toolsAndToolChoice.toolChoice != null ? JSON.stringify(toolsAndToolChoice.toolChoice) : void 0
5457
+ input: () => stepToolChoice != null ? JSON.stringify(stepToolChoice) : void 0
5517
5458
  },
5518
5459
  // standardized gen-ai llm span attributes:
5519
- "gen_ai.system": model.provider,
5520
- "gen_ai.request.model": model.modelId,
5460
+ "gen_ai.system": stepModel.provider,
5461
+ "gen_ai.request.model": stepModel.modelId,
5521
5462
  "gen_ai.request.frequency_penalty": settings.frequencyPenalty,
5522
5463
  "gen_ai.request.max_tokens": settings.maxOutputTokens,
5523
5464
  "gen_ai.request.presence_penalty": settings.presencePenalty,
@@ -5529,11 +5470,11 @@ async function generateText({
5529
5470
  }),
5530
5471
  tracer,
5531
5472
  fn: async (span2) => {
5532
- var _a19, _b2, _c2, _d, _e, _f, _g, _h;
5533
- const result = await model.doGenerate({
5534
- ...callSettings,
5535
- ...toolsAndToolChoice,
5536
- inputFormat: promptFormat,
5473
+ var _a19, _b2, _c2, _d2, _e2, _f2, _g, _h;
5474
+ const result = await stepModel.doGenerate({
5475
+ ...callSettings2,
5476
+ tools: stepTools,
5477
+ toolChoice: stepToolChoice,
5537
5478
  responseFormat: output == null ? void 0 : output.responseFormat,
5538
5479
  prompt: promptMessages,
5539
5480
  providerOptions,
@@ -5542,8 +5483,8 @@ async function generateText({
5542
5483
  });
5543
5484
  const responseData = {
5544
5485
  id: (_b2 = (_a19 = result.response) == null ? void 0 : _a19.id) != null ? _b2 : generateId3(),
5545
- timestamp: (_d = (_c2 = result.response) == null ? void 0 : _c2.timestamp) != null ? _d : currentDate(),
5546
- modelId: (_f = (_e = result.response) == null ? void 0 : _e.modelId) != null ? _f : model.modelId,
5486
+ timestamp: (_d2 = (_c2 = result.response) == null ? void 0 : _c2.timestamp) != null ? _d2 : currentDate(),
5487
+ modelId: (_f2 = (_e2 = result.response) == null ? void 0 : _e2.modelId) != null ? _f2 : stepModel.modelId,
5547
5488
  headers: (_g = result.response) == null ? void 0 : _g.headers,
5548
5489
  body: (_h = result.response) == null ? void 0 : _h.body
5549
5490
  };
@@ -5619,7 +5560,7 @@ async function generateText({
5619
5560
  nextStepType = "tool-result";
5620
5561
  }
5621
5562
  }
5622
- const originalText = (_a17 = extractContentText(currentModelResponse.content)) != null ? _a17 : "";
5563
+ const originalText = (_d = extractContentText(currentModelResponse.content)) != null ? _d : "";
5623
5564
  const stepTextLeadingWhitespaceTrimmed = stepType === "continue" && // only for continue steps
5624
5565
  text2.trimEnd() !== text2 ? originalText.trimStart() : originalText;
5625
5566
  const stepText = nextStepType === "continue" ? removeTextAfterLastWhitespace(stepTextLeadingWhitespaceTrimmed) : stepTextLeadingWhitespaceTrimmed;
@@ -5672,7 +5613,7 @@ async function generateText({
5672
5613
  finishReason: currentModelResponse.finishReason,
5673
5614
  usage: currentUsage,
5674
5615
  warnings: currentModelResponse.warnings,
5675
- request: (_b = currentModelResponse.request) != null ? _b : {},
5616
+ request: (_e = currentModelResponse.request) != null ? _e : {},
5676
5617
  response: {
5677
5618
  ...currentModelResponse.response,
5678
5619
  // deep clone msgs to avoid mutating past messages in multi-step:
@@ -5725,7 +5666,7 @@ async function generateText({
5725
5666
  finishReason: currentModelResponse.finishReason,
5726
5667
  usage,
5727
5668
  warnings: currentModelResponse.warnings,
5728
- request: (_c = currentModelResponse.request) != null ? _c : {},
5669
+ request: (_f = currentModelResponse.request) != null ? _f : {},
5729
5670
  response: {
5730
5671
  ...currentModelResponse.response,
5731
5672
  messages: responseMessages
@@ -5860,7 +5801,7 @@ __export(output_exports, {
5860
5801
  object: () => object,
5861
5802
  text: () => text
5862
5803
  });
5863
- var import_provider_utils17 = require("@ai-sdk/provider-utils");
5804
+ var import_provider_utils16 = require("@ai-sdk/provider-utils");
5864
5805
 
5865
5806
  // errors/index.ts
5866
5807
  var import_provider21 = require("@ai-sdk/provider");
@@ -5921,7 +5862,7 @@ var text = () => ({
5921
5862
  var object = ({
5922
5863
  schema: inputSchema
5923
5864
  }) => {
5924
- const schema = asSchema(inputSchema);
5865
+ const schema = (0, import_provider_utils4.asSchema)(inputSchema);
5925
5866
  return {
5926
5867
  type: "object",
5927
5868
  responseFormat: {
@@ -5947,7 +5888,7 @@ var object = ({
5947
5888
  }
5948
5889
  },
5949
5890
  async parseOutput({ text: text2 }, context) {
5950
- const parseResult = await (0, import_provider_utils17.safeParseJSON)({ text: text2 });
5891
+ const parseResult = await (0, import_provider_utils16.safeParseJSON)({ text: text2 });
5951
5892
  if (!parseResult.success) {
5952
5893
  throw new NoObjectGeneratedError({
5953
5894
  message: "No object generated: could not parse the response.",
@@ -5958,7 +5899,7 @@ var object = ({
5958
5899
  finishReason: context.finishReason
5959
5900
  });
5960
5901
  }
5961
- const validationResult = await (0, import_provider_utils17.safeValidateTypes)({
5902
+ const validationResult = await (0, import_provider_utils16.safeValidateTypes)({
5962
5903
  value: parseResult.value,
5963
5904
  schema
5964
5905
  });
@@ -5978,7 +5919,7 @@ var object = ({
5978
5919
  };
5979
5920
 
5980
5921
  // core/generate-text/smooth-stream.ts
5981
- var import_provider_utils18 = require("@ai-sdk/provider-utils");
5922
+ var import_provider_utils17 = require("@ai-sdk/provider-utils");
5982
5923
  var import_provider22 = require("@ai-sdk/provider");
5983
5924
  var CHUNKING_REGEXPS = {
5984
5925
  word: /\S+\s+/m,
@@ -5987,7 +5928,7 @@ var CHUNKING_REGEXPS = {
5987
5928
  function smoothStream({
5988
5929
  delayInMs = 10,
5989
5930
  chunking = "word",
5990
- _internal: { delay: delay2 = import_provider_utils18.delay } = {}
5931
+ _internal: { delay: delay2 = import_provider_utils17.delay } = {}
5991
5932
  } = {}) {
5992
5933
  let detectChunk;
5993
5934
  if (typeof chunking === "function") {
@@ -6047,7 +5988,7 @@ function smoothStream({
6047
5988
  }
6048
5989
 
6049
5990
  // core/generate-text/stream-text.ts
6050
- var import_provider_utils19 = require("@ai-sdk/provider-utils");
5991
+ var import_provider_utils18 = require("@ai-sdk/provider-utils");
6051
5992
 
6052
5993
  // util/as-array.ts
6053
5994
  function asArray(value) {
@@ -6246,7 +6187,7 @@ function runToolsTransformation({
6246
6187
  controller.enqueue(toolCall);
6247
6188
  const tool2 = tools[toolCall.toolName];
6248
6189
  if (tool2.execute != null) {
6249
- const toolExecutionId = (0, import_provider_utils5.generateId)();
6190
+ const toolExecutionId = (0, import_provider_utils4.generateId)();
6250
6191
  outstandingToolResults.add(toolExecutionId);
6251
6192
  recordSpan({
6252
6193
  name: "ai.toolCall",
@@ -6364,11 +6305,11 @@ function runToolsTransformation({
6364
6305
  }
6365
6306
 
6366
6307
  // core/generate-text/stream-text.ts
6367
- var originalGenerateId4 = (0, import_provider_utils19.createIdGenerator)({
6308
+ var originalGenerateId4 = (0, import_provider_utils18.createIdGenerator)({
6368
6309
  prefix: "aitxt",
6369
6310
  size: 24
6370
6311
  });
6371
- var originalGenerateMessageId2 = (0, import_provider_utils19.createIdGenerator)({
6312
+ var originalGenerateMessageId2 = (0, import_provider_utils18.createIdGenerator)({
6372
6313
  prefix: "msg",
6373
6314
  size: 24
6374
6315
  });
@@ -6795,18 +6736,16 @@ var DefaultStreamTextResult = class {
6795
6736
  prompt: { system, prompt, messages },
6796
6737
  tools
6797
6738
  });
6798
- const promptFormat = responseMessages.length === 0 ? initialPrompt.type : "messages";
6799
6739
  const stepInputMessages = [
6800
6740
  ...initialPrompt.messages,
6801
6741
  ...responseMessages
6802
6742
  ];
6803
6743
  const promptMessages = await convertToLanguageModelPrompt({
6804
6744
  prompt: {
6805
- type: promptFormat,
6806
6745
  system: initialPrompt.system,
6807
6746
  messages: stepInputMessages
6808
6747
  },
6809
- supportedUrls: await model.getSupportedUrls()
6748
+ supportedUrls: await model.supportedUrls
6810
6749
  });
6811
6750
  const toolsAndToolChoice = {
6812
6751
  ...prepareToolsAndToolChoice({ tools, toolChoice, activeTools })
@@ -6826,9 +6765,6 @@ var DefaultStreamTextResult = class {
6826
6765
  telemetry
6827
6766
  }),
6828
6767
  ...baseTelemetryAttributes,
6829
- "ai.prompt.format": {
6830
- input: () => promptFormat
6831
- },
6832
6768
  "ai.prompt.messages": {
6833
6769
  input: () => JSON.stringify(promptMessages)
6834
6770
  },
@@ -6866,7 +6802,6 @@ var DefaultStreamTextResult = class {
6866
6802
  result: await model.doStream({
6867
6803
  ...callSettings,
6868
6804
  ...toolsAndToolChoice,
6869
- inputFormat: promptFormat,
6870
6805
  responseFormat: output == null ? void 0 : output.responseFormat,
6871
6806
  prompt: promptMessages,
6872
6807
  providerOptions,
@@ -7959,8 +7894,8 @@ var doWrap = ({
7959
7894
  provider: providerId != null ? providerId : model.provider,
7960
7895
  modelId: modelId != null ? modelId : model.modelId,
7961
7896
  // TODO middleware should be able to modify the supported urls
7962
- async getSupportedUrls() {
7963
- return model.getSupportedUrls();
7897
+ get supportedUrls() {
7898
+ return model.supportedUrls;
7964
7899
  },
7965
7900
  async doGenerate(params) {
7966
7901
  const transformedParams = await doTransform({ params, type: "generate" });
@@ -8077,7 +8012,6 @@ function appendResponseMessages({
8077
8012
  );
8078
8013
  (_b = lastMessage.parts) != null ? _b : lastMessage.parts = [];
8079
8014
  lastMessage.content = textContent;
8080
- lastMessage.reasoning = reasoningTextContent;
8081
8015
  lastMessage.parts.push(...parts);
8082
8016
  lastMessage.toolInvocations = [
8083
8017
  ...(_c = lastMessage.toolInvocations) != null ? _c : [],
@@ -8096,7 +8030,6 @@ function appendResponseMessages({
8096
8030
  createdAt: currentDate(),
8097
8031
  // generate a createdAt date for the message, will be overridden by the client
8098
8032
  content: textContent,
8099
- reasoning: reasoningTextContent,
8100
8033
  toolInvocations: getToolInvocations2(0),
8101
8034
  parts: [
8102
8035
  ...parts,
@@ -8303,7 +8236,7 @@ function tool(tool2) {
8303
8236
  }
8304
8237
 
8305
8238
  // core/tool/mcp/mcp-sse-transport.ts
8306
- var import_provider_utils20 = require("@ai-sdk/provider-utils");
8239
+ var import_provider_utils19 = require("@ai-sdk/provider-utils");
8307
8240
 
8308
8241
  // core/tool/mcp/json-rpc-message.ts
8309
8242
  var import_zod9 = require("zod");
@@ -8474,7 +8407,7 @@ var SseMCPTransport = class {
8474
8407
  (_b = this.onerror) == null ? void 0 : _b.call(this, error);
8475
8408
  return reject(error);
8476
8409
  }
8477
- const stream = response.body.pipeThrough(new TextDecoderStream()).pipeThrough((0, import_provider_utils20.createEventSourceParserStream)());
8410
+ const stream = response.body.pipeThrough(new TextDecoderStream()).pipeThrough((0, import_provider_utils19.createEventSourceParserStream)());
8478
8411
  const reader = stream.getReader();
8479
8412
  const processEvents = async () => {
8480
8413
  var _a18, _b2, _c2;
@@ -8789,7 +8722,7 @@ var MCPClient = class {
8789
8722
  if (schemas !== "automatic" && !(name17 in schemas)) {
8790
8723
  continue;
8791
8724
  }
8792
- const parameters = schemas === "automatic" ? jsonSchema({
8725
+ const parameters = schemas === "automatic" ? (0, import_provider_utils4.jsonSchema)({
8793
8726
  ...inputSchema,
8794
8727
  properties: (_a17 = inputSchema.properties) != null ? _a17 : {},
8795
8728
  additionalProperties: false
@@ -8879,7 +8812,7 @@ function cosineSimilarity(vector1, vector2) {
8879
8812
  }
8880
8813
 
8881
8814
  // core/util/simulate-readable-stream.ts
8882
- var import_provider_utils21 = require("@ai-sdk/provider-utils");
8815
+ var import_provider_utils20 = require("@ai-sdk/provider-utils");
8883
8816
  function simulateReadableStream({
8884
8817
  chunks,
8885
8818
  initialDelayInMs = 0,
@@ -8887,7 +8820,7 @@ function simulateReadableStream({
8887
8820
  _internal
8888
8821
  }) {
8889
8822
  var _a17;
8890
- const delay2 = (_a17 = _internal == null ? void 0 : _internal.delay) != null ? _a17 : import_provider_utils21.delay;
8823
+ const delay2 = (_a17 = _internal == null ? void 0 : _internal.delay) != null ? _a17 : import_provider_utils20.delay;
8891
8824
  let index = 0;
8892
8825
  return new ReadableStream({
8893
8826
  async pull(controller) {
@@ -9049,7 +8982,6 @@ var StreamData = class {
9049
8982
  streamText,
9050
8983
  tool,
9051
8984
  updateToolCallResult,
9052
- wrapLanguageModel,
9053
- zodSchema
8985
+ wrapLanguageModel
9054
8986
  });
9055
8987
  //# sourceMappingURL=index.js.map