ai 6.0.0-beta.70 → 6.0.0-beta.72

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -4,8 +4,8 @@ var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
4
4
  var __getOwnPropNames = Object.getOwnPropertyNames;
5
5
  var __hasOwnProp = Object.prototype.hasOwnProperty;
6
6
  var __export = (target, all) => {
7
- for (var name17 in all)
8
- __defProp(target, name17, { get: all[name17], enumerable: true });
7
+ for (var name16 in all)
8
+ __defProp(target, name16, { get: all[name16], enumerable: true });
9
9
  };
10
10
  var __copyProps = (to, from, except, desc) => {
11
11
  if (from && typeof from === "object" || typeof from === "function") {
@@ -20,48 +20,47 @@ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: tru
20
20
  // src/index.ts
21
21
  var src_exports = {};
22
22
  __export(src_exports, {
23
- AISDKError: () => import_provider18.AISDKError,
24
- APICallError: () => import_provider18.APICallError,
23
+ AISDKError: () => import_provider17.AISDKError,
24
+ APICallError: () => import_provider17.APICallError,
25
25
  AbstractChat: () => AbstractChat,
26
26
  DefaultChatTransport: () => DefaultChatTransport,
27
27
  DownloadError: () => DownloadError,
28
- EmptyResponseBodyError: () => import_provider18.EmptyResponseBodyError,
28
+ EmptyResponseBodyError: () => import_provider17.EmptyResponseBodyError,
29
29
  Experimental_Agent: () => ToolLoopAgent,
30
30
  HttpChatTransport: () => HttpChatTransport,
31
31
  InvalidArgumentError: () => InvalidArgumentError,
32
32
  InvalidDataContentError: () => InvalidDataContentError,
33
33
  InvalidMessageRoleError: () => InvalidMessageRoleError,
34
- InvalidPromptError: () => import_provider18.InvalidPromptError,
35
- InvalidResponseDataError: () => import_provider18.InvalidResponseDataError,
34
+ InvalidPromptError: () => import_provider17.InvalidPromptError,
35
+ InvalidResponseDataError: () => import_provider17.InvalidResponseDataError,
36
36
  InvalidStreamPartError: () => InvalidStreamPartError,
37
37
  InvalidToolInputError: () => InvalidToolInputError,
38
- JSONParseError: () => import_provider18.JSONParseError,
38
+ JSONParseError: () => import_provider17.JSONParseError,
39
39
  JsonToSseTransformStream: () => JsonToSseTransformStream,
40
- LoadAPIKeyError: () => import_provider18.LoadAPIKeyError,
41
- LoadSettingError: () => import_provider18.LoadSettingError,
42
- MCPClientError: () => MCPClientError,
40
+ LoadAPIKeyError: () => import_provider17.LoadAPIKeyError,
41
+ LoadSettingError: () => import_provider17.LoadSettingError,
43
42
  MessageConversionError: () => MessageConversionError,
44
- NoContentGeneratedError: () => import_provider18.NoContentGeneratedError,
43
+ NoContentGeneratedError: () => import_provider17.NoContentGeneratedError,
45
44
  NoImageGeneratedError: () => NoImageGeneratedError,
46
45
  NoObjectGeneratedError: () => NoObjectGeneratedError,
47
46
  NoOutputGeneratedError: () => NoOutputGeneratedError,
48
47
  NoOutputSpecifiedError: () => NoOutputSpecifiedError,
49
48
  NoSpeechGeneratedError: () => NoSpeechGeneratedError,
50
- NoSuchModelError: () => import_provider18.NoSuchModelError,
49
+ NoSuchModelError: () => import_provider17.NoSuchModelError,
51
50
  NoSuchProviderError: () => NoSuchProviderError,
52
51
  NoSuchToolError: () => NoSuchToolError,
53
52
  Output: () => output_exports,
54
53
  RetryError: () => RetryError,
55
54
  SerialJobExecutor: () => SerialJobExecutor,
56
55
  TextStreamChatTransport: () => TextStreamChatTransport,
57
- TooManyEmbeddingValuesForCallError: () => import_provider18.TooManyEmbeddingValuesForCallError,
56
+ TooManyEmbeddingValuesForCallError: () => import_provider17.TooManyEmbeddingValuesForCallError,
58
57
  ToolCallRepairError: () => ToolCallRepairError,
59
58
  ToolLoopAgent: () => ToolLoopAgent,
60
- TypeValidationError: () => import_provider18.TypeValidationError,
59
+ TypeValidationError: () => import_provider17.TypeValidationError,
61
60
  UI_MESSAGE_STREAM_HEADERS: () => UI_MESSAGE_STREAM_HEADERS,
62
- UnsupportedFunctionalityError: () => import_provider18.UnsupportedFunctionalityError,
61
+ UnsupportedFunctionalityError: () => import_provider17.UnsupportedFunctionalityError,
63
62
  UnsupportedModelVersionError: () => UnsupportedModelVersionError,
64
- asSchema: () => import_provider_utils38.asSchema,
63
+ asSchema: () => import_provider_utils36.asSchema,
65
64
  assistantModelMessageSchema: () => assistantModelMessageSchema,
66
65
  callCompletionApi: () => callCompletionApi,
67
66
  consumeStream: () => consumeStream,
@@ -77,17 +76,16 @@ __export(src_exports, {
77
76
  createAgentUIStream: () => createAgentUIStream,
78
77
  createAgentUIStreamResponse: () => createAgentUIStreamResponse,
79
78
  createGateway: () => import_gateway3.createGateway,
80
- createIdGenerator: () => import_provider_utils38.createIdGenerator,
79
+ createIdGenerator: () => import_provider_utils36.createIdGenerator,
81
80
  createProviderRegistry: () => createProviderRegistry,
82
81
  createTextStreamResponse: () => createTextStreamResponse,
83
82
  createUIMessageStream: () => createUIMessageStream,
84
83
  createUIMessageStreamResponse: () => createUIMessageStreamResponse,
85
84
  customProvider: () => customProvider,
86
85
  defaultSettingsMiddleware: () => defaultSettingsMiddleware,
87
- dynamicTool: () => import_provider_utils38.dynamicTool,
86
+ dynamicTool: () => import_provider_utils36.dynamicTool,
88
87
  embed: () => embed,
89
88
  embedMany: () => embedMany,
90
- experimental_createMCPClient: () => createMCPClient,
91
89
  experimental_createProviderRegistry: () => experimental_createProviderRegistry,
92
90
  experimental_customProvider: () => experimental_customProvider,
93
91
  experimental_generateImage: () => generateImage,
@@ -95,7 +93,7 @@ __export(src_exports, {
95
93
  experimental_transcribe: () => transcribe,
96
94
  extractReasoningMiddleware: () => extractReasoningMiddleware,
97
95
  gateway: () => import_gateway3.gateway,
98
- generateId: () => import_provider_utils38.generateId,
96
+ generateId: () => import_provider_utils36.generateId,
99
97
  generateObject: () => generateObject,
100
98
  generateText: () => generateText,
101
99
  getTextFromDataUrl: () => getTextFromDataUrl,
@@ -109,11 +107,11 @@ __export(src_exports, {
109
107
  isTextUIPart: () => isTextUIPart,
110
108
  isToolOrDynamicToolUIPart: () => isToolOrDynamicToolUIPart,
111
109
  isToolUIPart: () => isToolUIPart,
112
- jsonSchema: () => import_provider_utils38.jsonSchema,
110
+ jsonSchema: () => import_provider_utils36.jsonSchema,
113
111
  lastAssistantMessageIsCompleteWithApprovalResponses: () => lastAssistantMessageIsCompleteWithApprovalResponses,
114
112
  lastAssistantMessageIsCompleteWithToolCalls: () => lastAssistantMessageIsCompleteWithToolCalls,
115
113
  modelMessageSchema: () => modelMessageSchema,
116
- parseJsonEventStream: () => import_provider_utils38.parseJsonEventStream,
114
+ parseJsonEventStream: () => import_provider_utils36.parseJsonEventStream,
117
115
  parsePartialJson: () => parsePartialJson,
118
116
  pipeAgentUIStreamToResponse: () => pipeAgentUIStreamToResponse,
119
117
  pipeTextStreamToResponse: () => pipeTextStreamToResponse,
@@ -128,18 +126,18 @@ __export(src_exports, {
128
126
  streamObject: () => streamObject,
129
127
  streamText: () => streamText,
130
128
  systemModelMessageSchema: () => systemModelMessageSchema,
131
- tool: () => import_provider_utils38.tool,
129
+ tool: () => import_provider_utils36.tool,
132
130
  toolModelMessageSchema: () => toolModelMessageSchema,
133
131
  uiMessageChunkSchema: () => uiMessageChunkSchema,
134
132
  userModelMessageSchema: () => userModelMessageSchema,
135
133
  validateUIMessages: () => validateUIMessages,
136
134
  wrapLanguageModel: () => wrapLanguageModel,
137
135
  wrapProvider: () => wrapProvider,
138
- zodSchema: () => import_provider_utils38.zodSchema
136
+ zodSchema: () => import_provider_utils36.zodSchema
139
137
  });
140
138
  module.exports = __toCommonJS(src_exports);
141
139
  var import_gateway3 = require("@ai-sdk/gateway");
142
- var import_provider_utils38 = require("@ai-sdk/provider-utils");
140
+ var import_provider_utils36 = require("@ai-sdk/provider-utils");
143
141
 
144
142
  // src/generate-text/generate-text.ts
145
143
  var import_provider_utils11 = require("@ai-sdk/provider-utils");
@@ -216,7 +214,7 @@ var logWarnings = (warnings) => {
216
214
  var import_gateway = require("@ai-sdk/gateway");
217
215
 
218
216
  // src/error/index.ts
219
- var import_provider18 = require("@ai-sdk/provider");
217
+ var import_provider17 = require("@ai-sdk/provider");
220
218
 
221
219
  // src/error/invalid-argument-error.ts
222
220
  var import_provider2 = require("@ai-sdk/provider");
@@ -289,24 +287,21 @@ var InvalidToolInputError = class extends import_provider4.AISDKError {
289
287
  };
290
288
  _a4 = symbol4;
291
289
 
292
- // src/error/mcp-client-error.ts
290
+ // src/error/no-image-generated-error.ts
293
291
  var import_provider5 = require("@ai-sdk/provider");
294
- var name5 = "AI_MCPClientError";
292
+ var name5 = "AI_NoImageGeneratedError";
295
293
  var marker5 = `vercel.ai.error.${name5}`;
296
294
  var symbol5 = Symbol.for(marker5);
297
295
  var _a5;
298
- var MCPClientError = class extends import_provider5.AISDKError {
296
+ var NoImageGeneratedError = class extends import_provider5.AISDKError {
299
297
  constructor({
300
- name: name17 = "MCPClientError",
301
- message,
298
+ message = "No image generated.",
302
299
  cause,
303
- data,
304
- code
300
+ responses
305
301
  }) {
306
- super({ name: name17, message, cause });
302
+ super({ name: name5, message, cause });
307
303
  this[_a5] = true;
308
- this.data = data;
309
- this.code = code;
304
+ this.responses = responses;
310
305
  }
311
306
  static isInstance(error) {
312
307
  return import_provider5.AISDKError.hasMarker(error, marker5);
@@ -314,35 +309,13 @@ var MCPClientError = class extends import_provider5.AISDKError {
314
309
  };
315
310
  _a5 = symbol5;
316
311
 
317
- // src/error/no-image-generated-error.ts
312
+ // src/error/no-object-generated-error.ts
318
313
  var import_provider6 = require("@ai-sdk/provider");
319
- var name6 = "AI_NoImageGeneratedError";
314
+ var name6 = "AI_NoObjectGeneratedError";
320
315
  var marker6 = `vercel.ai.error.${name6}`;
321
316
  var symbol6 = Symbol.for(marker6);
322
317
  var _a6;
323
- var NoImageGeneratedError = class extends import_provider6.AISDKError {
324
- constructor({
325
- message = "No image generated.",
326
- cause,
327
- responses
328
- }) {
329
- super({ name: name6, message, cause });
330
- this[_a6] = true;
331
- this.responses = responses;
332
- }
333
- static isInstance(error) {
334
- return import_provider6.AISDKError.hasMarker(error, marker6);
335
- }
336
- };
337
- _a6 = symbol6;
338
-
339
- // src/error/no-object-generated-error.ts
340
- var import_provider7 = require("@ai-sdk/provider");
341
- var name7 = "AI_NoObjectGeneratedError";
342
- var marker7 = `vercel.ai.error.${name7}`;
343
- var symbol7 = Symbol.for(marker7);
344
- var _a7;
345
- var NoObjectGeneratedError = class extends import_provider7.AISDKError {
318
+ var NoObjectGeneratedError = class extends import_provider6.AISDKError {
346
319
  constructor({
347
320
  message = "No object generated.",
348
321
  cause,
@@ -351,43 +324,43 @@ var NoObjectGeneratedError = class extends import_provider7.AISDKError {
351
324
  usage,
352
325
  finishReason
353
326
  }) {
354
- super({ name: name7, message, cause });
355
- this[_a7] = true;
327
+ super({ name: name6, message, cause });
328
+ this[_a6] = true;
356
329
  this.text = text2;
357
330
  this.response = response;
358
331
  this.usage = usage;
359
332
  this.finishReason = finishReason;
360
333
  }
361
334
  static isInstance(error) {
362
- return import_provider7.AISDKError.hasMarker(error, marker7);
335
+ return import_provider6.AISDKError.hasMarker(error, marker6);
363
336
  }
364
337
  };
365
- _a7 = symbol7;
338
+ _a6 = symbol6;
366
339
 
367
340
  // src/error/no-output-generated-error.ts
368
- var import_provider8 = require("@ai-sdk/provider");
369
- var name8 = "AI_NoOutputGeneratedError";
370
- var marker8 = `vercel.ai.error.${name8}`;
371
- var symbol8 = Symbol.for(marker8);
372
- var _a8;
373
- var NoOutputGeneratedError = class extends import_provider8.AISDKError {
341
+ var import_provider7 = require("@ai-sdk/provider");
342
+ var name7 = "AI_NoOutputGeneratedError";
343
+ var marker7 = `vercel.ai.error.${name7}`;
344
+ var symbol7 = Symbol.for(marker7);
345
+ var _a7;
346
+ var NoOutputGeneratedError = class extends import_provider7.AISDKError {
374
347
  // used in isInstance
375
348
  constructor({
376
349
  message = "No output generated.",
377
350
  cause
378
351
  } = {}) {
379
- super({ name: name8, message, cause });
380
- this[_a8] = true;
352
+ super({ name: name7, message, cause });
353
+ this[_a7] = true;
381
354
  }
382
355
  static isInstance(error) {
383
- return import_provider8.AISDKError.hasMarker(error, marker8);
356
+ return import_provider7.AISDKError.hasMarker(error, marker7);
384
357
  }
385
358
  };
386
- _a8 = symbol8;
359
+ _a7 = symbol7;
387
360
 
388
361
  // src/error/no-speech-generated-error.ts
389
- var import_provider9 = require("@ai-sdk/provider");
390
- var NoSpeechGeneratedError = class extends import_provider9.AISDKError {
362
+ var import_provider8 = require("@ai-sdk/provider");
363
+ var NoSpeechGeneratedError = class extends import_provider8.AISDKError {
391
364
  constructor(options) {
392
365
  super({
393
366
  name: "AI_NoSpeechGeneratedError",
@@ -398,53 +371,53 @@ var NoSpeechGeneratedError = class extends import_provider9.AISDKError {
398
371
  };
399
372
 
400
373
  // src/error/no-such-tool-error.ts
401
- var import_provider10 = require("@ai-sdk/provider");
402
- var name9 = "AI_NoSuchToolError";
403
- var marker9 = `vercel.ai.error.${name9}`;
404
- var symbol9 = Symbol.for(marker9);
405
- var _a9;
406
- var NoSuchToolError = class extends import_provider10.AISDKError {
374
+ var import_provider9 = require("@ai-sdk/provider");
375
+ var name8 = "AI_NoSuchToolError";
376
+ var marker8 = `vercel.ai.error.${name8}`;
377
+ var symbol8 = Symbol.for(marker8);
378
+ var _a8;
379
+ var NoSuchToolError = class extends import_provider9.AISDKError {
407
380
  constructor({
408
381
  toolName,
409
382
  availableTools = void 0,
410
383
  message = `Model tried to call unavailable tool '${toolName}'. ${availableTools === void 0 ? "No tools are available." : `Available tools: ${availableTools.join(", ")}.`}`
411
384
  }) {
412
- super({ name: name9, message });
413
- this[_a9] = true;
385
+ super({ name: name8, message });
386
+ this[_a8] = true;
414
387
  this.toolName = toolName;
415
388
  this.availableTools = availableTools;
416
389
  }
417
390
  static isInstance(error) {
418
- return import_provider10.AISDKError.hasMarker(error, marker9);
391
+ return import_provider9.AISDKError.hasMarker(error, marker8);
419
392
  }
420
393
  };
421
- _a9 = symbol9;
394
+ _a8 = symbol8;
422
395
 
423
396
  // src/error/tool-call-repair-error.ts
424
- var import_provider11 = require("@ai-sdk/provider");
425
- var name10 = "AI_ToolCallRepairError";
426
- var marker10 = `vercel.ai.error.${name10}`;
427
- var symbol10 = Symbol.for(marker10);
428
- var _a10;
429
- var ToolCallRepairError = class extends import_provider11.AISDKError {
397
+ var import_provider10 = require("@ai-sdk/provider");
398
+ var name9 = "AI_ToolCallRepairError";
399
+ var marker9 = `vercel.ai.error.${name9}`;
400
+ var symbol9 = Symbol.for(marker9);
401
+ var _a9;
402
+ var ToolCallRepairError = class extends import_provider10.AISDKError {
430
403
  constructor({
431
404
  cause,
432
405
  originalError,
433
- message = `Error repairing tool call: ${(0, import_provider11.getErrorMessage)(cause)}`
406
+ message = `Error repairing tool call: ${(0, import_provider10.getErrorMessage)(cause)}`
434
407
  }) {
435
- super({ name: name10, message, cause });
436
- this[_a10] = true;
408
+ super({ name: name9, message, cause });
409
+ this[_a9] = true;
437
410
  this.originalError = originalError;
438
411
  }
439
412
  static isInstance(error) {
440
- return import_provider11.AISDKError.hasMarker(error, marker10);
413
+ return import_provider10.AISDKError.hasMarker(error, marker9);
441
414
  }
442
415
  };
443
- _a10 = symbol10;
416
+ _a9 = symbol9;
444
417
 
445
418
  // src/error/unsupported-model-version-error.ts
446
- var import_provider12 = require("@ai-sdk/provider");
447
- var UnsupportedModelVersionError = class extends import_provider12.AISDKError {
419
+ var import_provider11 = require("@ai-sdk/provider");
420
+ var UnsupportedModelVersionError = class extends import_provider11.AISDKError {
448
421
  constructor(options) {
449
422
  super({
450
423
  name: "AI_UnsupportedModelVersionError",
@@ -457,76 +430,76 @@ var UnsupportedModelVersionError = class extends import_provider12.AISDKError {
457
430
  };
458
431
 
459
432
  // src/prompt/invalid-data-content-error.ts
460
- var import_provider13 = require("@ai-sdk/provider");
461
- var name11 = "AI_InvalidDataContentError";
462
- var marker11 = `vercel.ai.error.${name11}`;
463
- var symbol11 = Symbol.for(marker11);
464
- var _a11;
465
- var InvalidDataContentError = class extends import_provider13.AISDKError {
433
+ var import_provider12 = require("@ai-sdk/provider");
434
+ var name10 = "AI_InvalidDataContentError";
435
+ var marker10 = `vercel.ai.error.${name10}`;
436
+ var symbol10 = Symbol.for(marker10);
437
+ var _a10;
438
+ var InvalidDataContentError = class extends import_provider12.AISDKError {
466
439
  constructor({
467
440
  content,
468
441
  cause,
469
442
  message = `Invalid data content. Expected a base64 string, Uint8Array, ArrayBuffer, or Buffer, but got ${typeof content}.`
470
443
  }) {
471
- super({ name: name11, message, cause });
472
- this[_a11] = true;
444
+ super({ name: name10, message, cause });
445
+ this[_a10] = true;
473
446
  this.content = content;
474
447
  }
475
448
  static isInstance(error) {
476
- return import_provider13.AISDKError.hasMarker(error, marker11);
449
+ return import_provider12.AISDKError.hasMarker(error, marker10);
477
450
  }
478
451
  };
479
- _a11 = symbol11;
452
+ _a10 = symbol10;
480
453
 
481
454
  // src/prompt/invalid-message-role-error.ts
482
- var import_provider14 = require("@ai-sdk/provider");
483
- var name12 = "AI_InvalidMessageRoleError";
484
- var marker12 = `vercel.ai.error.${name12}`;
485
- var symbol12 = Symbol.for(marker12);
486
- var _a12;
487
- var InvalidMessageRoleError = class extends import_provider14.AISDKError {
455
+ var import_provider13 = require("@ai-sdk/provider");
456
+ var name11 = "AI_InvalidMessageRoleError";
457
+ var marker11 = `vercel.ai.error.${name11}`;
458
+ var symbol11 = Symbol.for(marker11);
459
+ var _a11;
460
+ var InvalidMessageRoleError = class extends import_provider13.AISDKError {
488
461
  constructor({
489
462
  role,
490
463
  message = `Invalid message role: '${role}'. Must be one of: "system", "user", "assistant", "tool".`
491
464
  }) {
492
- super({ name: name12, message });
493
- this[_a12] = true;
465
+ super({ name: name11, message });
466
+ this[_a11] = true;
494
467
  this.role = role;
495
468
  }
496
469
  static isInstance(error) {
497
- return import_provider14.AISDKError.hasMarker(error, marker12);
470
+ return import_provider13.AISDKError.hasMarker(error, marker11);
498
471
  }
499
472
  };
500
- _a12 = symbol12;
473
+ _a11 = symbol11;
501
474
 
502
475
  // src/prompt/message-conversion-error.ts
503
- var import_provider15 = require("@ai-sdk/provider");
504
- var name13 = "AI_MessageConversionError";
505
- var marker13 = `vercel.ai.error.${name13}`;
506
- var symbol13 = Symbol.for(marker13);
507
- var _a13;
508
- var MessageConversionError = class extends import_provider15.AISDKError {
476
+ var import_provider14 = require("@ai-sdk/provider");
477
+ var name12 = "AI_MessageConversionError";
478
+ var marker12 = `vercel.ai.error.${name12}`;
479
+ var symbol12 = Symbol.for(marker12);
480
+ var _a12;
481
+ var MessageConversionError = class extends import_provider14.AISDKError {
509
482
  constructor({
510
483
  originalMessage,
511
484
  message
512
485
  }) {
513
- super({ name: name13, message });
514
- this[_a13] = true;
486
+ super({ name: name12, message });
487
+ this[_a12] = true;
515
488
  this.originalMessage = originalMessage;
516
489
  }
517
490
  static isInstance(error) {
518
- return import_provider15.AISDKError.hasMarker(error, marker13);
491
+ return import_provider14.AISDKError.hasMarker(error, marker12);
519
492
  }
520
493
  };
521
- _a13 = symbol13;
494
+ _a12 = symbol12;
522
495
 
523
496
  // src/util/download/download-error.ts
524
- var import_provider16 = require("@ai-sdk/provider");
525
- var name14 = "AI_DownloadError";
526
- var marker14 = `vercel.ai.error.${name14}`;
527
- var symbol14 = Symbol.for(marker14);
528
- var _a14;
529
- var DownloadError = class extends import_provider16.AISDKError {
497
+ var import_provider15 = require("@ai-sdk/provider");
498
+ var name13 = "AI_DownloadError";
499
+ var marker13 = `vercel.ai.error.${name13}`;
500
+ var symbol13 = Symbol.for(marker13);
501
+ var _a13;
502
+ var DownloadError = class extends import_provider15.AISDKError {
530
503
  constructor({
531
504
  url,
532
505
  statusCode,
@@ -534,41 +507,41 @@ var DownloadError = class extends import_provider16.AISDKError {
534
507
  cause,
535
508
  message = cause == null ? `Failed to download ${url}: ${statusCode} ${statusText}` : `Failed to download ${url}: ${cause}`
536
509
  }) {
537
- super({ name: name14, message, cause });
538
- this[_a14] = true;
510
+ super({ name: name13, message, cause });
511
+ this[_a13] = true;
539
512
  this.url = url;
540
513
  this.statusCode = statusCode;
541
514
  this.statusText = statusText;
542
515
  }
543
516
  static isInstance(error) {
544
- return import_provider16.AISDKError.hasMarker(error, marker14);
517
+ return import_provider15.AISDKError.hasMarker(error, marker13);
545
518
  }
546
519
  };
547
- _a14 = symbol14;
520
+ _a13 = symbol13;
548
521
 
549
522
  // src/util/retry-error.ts
550
- var import_provider17 = require("@ai-sdk/provider");
551
- var name15 = "AI_RetryError";
552
- var marker15 = `vercel.ai.error.${name15}`;
553
- var symbol15 = Symbol.for(marker15);
554
- var _a15;
555
- var RetryError = class extends import_provider17.AISDKError {
523
+ var import_provider16 = require("@ai-sdk/provider");
524
+ var name14 = "AI_RetryError";
525
+ var marker14 = `vercel.ai.error.${name14}`;
526
+ var symbol14 = Symbol.for(marker14);
527
+ var _a14;
528
+ var RetryError = class extends import_provider16.AISDKError {
556
529
  constructor({
557
530
  message,
558
531
  reason,
559
532
  errors
560
533
  }) {
561
- super({ name: name15, message });
562
- this[_a15] = true;
534
+ super({ name: name14, message });
535
+ this[_a14] = true;
563
536
  this.reason = reason;
564
537
  this.errors = errors;
565
538
  this.lastError = errors[errors.length - 1];
566
539
  }
567
540
  static isInstance(error) {
568
- return import_provider17.AISDKError.hasMarker(error, marker15);
541
+ return import_provider16.AISDKError.hasMarker(error, marker14);
569
542
  }
570
543
  };
571
- _a15 = symbol15;
544
+ _a14 = symbol14;
572
545
 
573
546
  // src/model/resolve-model.ts
574
547
  function transformToV3LanguageModel(model) {
@@ -644,7 +617,7 @@ function resolveEmbeddingModel(model) {
644
617
  );
645
618
  }
646
619
  function resolveTranscriptionModel(model) {
647
- var _a17, _b;
620
+ var _a16, _b;
648
621
  if (typeof model !== "string") {
649
622
  if (model.specificationVersion !== "v3" && model.specificationVersion !== "v2") {
650
623
  const unsupportedModel = model;
@@ -659,10 +632,10 @@ function resolveTranscriptionModel(model) {
659
632
  }
660
633
  return model;
661
634
  }
662
- return (_b = (_a17 = getGlobalProvider()).transcriptionModel) == null ? void 0 : _b.call(_a17, model);
635
+ return (_b = (_a16 = getGlobalProvider()).transcriptionModel) == null ? void 0 : _b.call(_a16, model);
663
636
  }
664
637
  function resolveSpeechModel(model) {
665
- var _a17, _b;
638
+ var _a16, _b;
666
639
  if (typeof model !== "string") {
667
640
  if (model.specificationVersion !== "v3" && model.specificationVersion !== "v2") {
668
641
  const unsupportedModel = model;
@@ -677,11 +650,11 @@ function resolveSpeechModel(model) {
677
650
  }
678
651
  return model;
679
652
  }
680
- return (_b = (_a17 = getGlobalProvider()).speechModel) == null ? void 0 : _b.call(_a17, model);
653
+ return (_b = (_a16 = getGlobalProvider()).speechModel) == null ? void 0 : _b.call(_a16, model);
681
654
  }
682
655
  function getGlobalProvider() {
683
- var _a17;
684
- return (_a17 = globalThis.AI_SDK_DEFAULT_PROVIDER) != null ? _a17 : import_gateway.gateway;
656
+ var _a16;
657
+ return (_a16 = globalThis.AI_SDK_DEFAULT_PROVIDER) != null ? _a16 : import_gateway.gateway;
685
658
  }
686
659
 
687
660
  // src/prompt/convert-to-language-model-prompt.ts
@@ -876,11 +849,11 @@ function detectMediaType({
876
849
  var import_provider_utils2 = require("@ai-sdk/provider-utils");
877
850
 
878
851
  // src/version.ts
879
- var VERSION = true ? "6.0.0-beta.70" : "0.0.0-test";
852
+ var VERSION = true ? "6.0.0-beta.72" : "0.0.0-test";
880
853
 
881
854
  // src/util/download/download.ts
882
855
  var download = async ({ url }) => {
883
- var _a17;
856
+ var _a16;
884
857
  const urlText = url.toString();
885
858
  try {
886
859
  const response = await fetch(urlText, {
@@ -899,7 +872,7 @@ var download = async ({ url }) => {
899
872
  }
900
873
  return {
901
874
  data: new Uint8Array(await response.arrayBuffer()),
902
- mediaType: (_a17 = response.headers.get("content-type")) != null ? _a17 : void 0
875
+ mediaType: (_a16 = response.headers.get("content-type")) != null ? _a16 : void 0
903
876
  };
904
877
  } catch (error) {
905
878
  if (DownloadError.isInstance(error)) {
@@ -917,7 +890,7 @@ var createDefaultDownloadFunction = (download2 = download) => (requestedDownload
917
890
  );
918
891
 
919
892
  // src/prompt/data-content.ts
920
- var import_provider19 = require("@ai-sdk/provider");
893
+ var import_provider18 = require("@ai-sdk/provider");
921
894
  var import_provider_utils3 = require("@ai-sdk/provider-utils");
922
895
  var import_v4 = require("zod/v4");
923
896
 
@@ -945,8 +918,8 @@ var dataContentSchema = import_v4.z.union([
945
918
  import_v4.z.custom(
946
919
  // Buffer might not be available in some environments such as CloudFlare:
947
920
  (value) => {
948
- var _a17, _b;
949
- return (_b = (_a17 = globalThis.Buffer) == null ? void 0 : _a17.isBuffer(value)) != null ? _b : false;
921
+ var _a16, _b;
922
+ return (_b = (_a16 = globalThis.Buffer) == null ? void 0 : _a16.isBuffer(value)) != null ? _b : false;
950
923
  },
951
924
  { message: "Must be a Buffer" }
952
925
  )
@@ -969,7 +942,7 @@ function convertToLanguageModelV3DataContent(content) {
969
942
  content.toString()
970
943
  );
971
944
  if (dataUrlMediaType == null || base64Content == null) {
972
- throw new import_provider19.AISDKError({
945
+ throw new import_provider18.AISDKError({
973
946
  name: "InvalidDataContentError",
974
947
  message: `Invalid data URL format in content ${content.toString()}`
975
948
  });
@@ -1160,8 +1133,8 @@ async function downloadAssets(messages, download2, supportedUrls) {
1160
1133
  ).flat().filter(
1161
1134
  (part) => part.type === "image" || part.type === "file"
1162
1135
  ).map((part) => {
1163
- var _a17;
1164
- const mediaType = (_a17 = part.mediaType) != null ? _a17 : part.type === "image" ? "image/*" : void 0;
1136
+ var _a16;
1137
+ const mediaType = (_a16 = part.mediaType) != null ? _a16 : part.type === "image" ? "image/*" : void 0;
1165
1138
  let data = part.type === "image" ? part.image : part.data;
1166
1139
  if (typeof data === "string") {
1167
1140
  try {
@@ -1191,7 +1164,7 @@ async function downloadAssets(messages, download2, supportedUrls) {
1191
1164
  );
1192
1165
  }
1193
1166
  function convertPartToLanguageModelPart(part, downloadedAssets) {
1194
- var _a17;
1167
+ var _a16;
1195
1168
  if (part.type === "text") {
1196
1169
  return {
1197
1170
  type: "text",
@@ -1224,7 +1197,7 @@ function convertPartToLanguageModelPart(part, downloadedAssets) {
1224
1197
  switch (type) {
1225
1198
  case "image": {
1226
1199
  if (data instanceof Uint8Array || typeof data === "string") {
1227
- mediaType = (_a17 = detectMediaType({ data, signatures: imageMediaTypeSignatures })) != null ? _a17 : mediaType;
1200
+ mediaType = (_a16 = detectMediaType({ data, signatures: imageMediaTypeSignatures })) != null ? _a16 : mediaType;
1228
1201
  }
1229
1202
  return {
1230
1203
  type: "file",
@@ -1276,19 +1249,19 @@ function mapToolResultOutput(output) {
1276
1249
  }
1277
1250
 
1278
1251
  // src/prompt/create-tool-model-output.ts
1279
- var import_provider20 = require("@ai-sdk/provider");
1252
+ var import_provider19 = require("@ai-sdk/provider");
1280
1253
  function createToolModelOutput({
1281
1254
  output,
1282
- tool: tool3,
1255
+ tool: tool2,
1283
1256
  errorMode
1284
1257
  }) {
1285
1258
  if (errorMode === "text") {
1286
- return { type: "error-text", value: (0, import_provider20.getErrorMessage)(output) };
1259
+ return { type: "error-text", value: (0, import_provider19.getErrorMessage)(output) };
1287
1260
  } else if (errorMode === "json") {
1288
1261
  return { type: "error-json", value: toJSONValue(output) };
1289
1262
  }
1290
- if (tool3 == null ? void 0 : tool3.toModelOutput) {
1291
- return tool3.toModelOutput(output);
1263
+ if (tool2 == null ? void 0 : tool2.toModelOutput) {
1264
+ return tool2.toModelOutput(output);
1292
1265
  }
1293
1266
  return typeof output === "string" ? { type: "text", value: output } : { type: "json", value: toJSONValue(output) };
1294
1267
  }
@@ -1410,29 +1383,29 @@ async function prepareToolsAndToolChoice({
1410
1383
  };
1411
1384
  }
1412
1385
  const filteredTools = activeTools != null ? Object.entries(tools).filter(
1413
- ([name17]) => activeTools.includes(name17)
1386
+ ([name16]) => activeTools.includes(name16)
1414
1387
  ) : Object.entries(tools);
1415
1388
  const languageModelTools = [];
1416
- for (const [name17, tool3] of filteredTools) {
1417
- const toolType = tool3.type;
1389
+ for (const [name16, tool2] of filteredTools) {
1390
+ const toolType = tool2.type;
1418
1391
  switch (toolType) {
1419
1392
  case void 0:
1420
1393
  case "dynamic":
1421
1394
  case "function":
1422
1395
  languageModelTools.push({
1423
1396
  type: "function",
1424
- name: name17,
1425
- description: tool3.description,
1426
- inputSchema: await (0, import_provider_utils5.asSchema)(tool3.inputSchema).jsonSchema,
1427
- providerOptions: tool3.providerOptions
1397
+ name: name16,
1398
+ description: tool2.description,
1399
+ inputSchema: await (0, import_provider_utils5.asSchema)(tool2.inputSchema).jsonSchema,
1400
+ providerOptions: tool2.providerOptions
1428
1401
  });
1429
1402
  break;
1430
1403
  case "provider-defined":
1431
1404
  languageModelTools.push({
1432
1405
  type: "provider-defined",
1433
- name: name17,
1434
- id: tool3.id,
1435
- args: tool3.args
1406
+ name: name16,
1407
+ id: tool2.id,
1408
+ args: tool2.args
1436
1409
  });
1437
1410
  break;
1438
1411
  default: {
@@ -1448,7 +1421,7 @@ async function prepareToolsAndToolChoice({
1448
1421
  }
1449
1422
 
1450
1423
  // src/prompt/standardize-prompt.ts
1451
- var import_provider21 = require("@ai-sdk/provider");
1424
+ var import_provider20 = require("@ai-sdk/provider");
1452
1425
  var import_provider_utils6 = require("@ai-sdk/provider-utils");
1453
1426
  var import_v46 = require("zod/v4");
1454
1427
 
@@ -1666,19 +1639,19 @@ var coreMessageSchema = modelMessageSchema;
1666
1639
  // src/prompt/standardize-prompt.ts
1667
1640
  async function standardizePrompt(prompt) {
1668
1641
  if (prompt.prompt == null && prompt.messages == null) {
1669
- throw new import_provider21.InvalidPromptError({
1642
+ throw new import_provider20.InvalidPromptError({
1670
1643
  prompt,
1671
1644
  message: "prompt or messages must be defined"
1672
1645
  });
1673
1646
  }
1674
1647
  if (prompt.prompt != null && prompt.messages != null) {
1675
- throw new import_provider21.InvalidPromptError({
1648
+ throw new import_provider20.InvalidPromptError({
1676
1649
  prompt,
1677
1650
  message: "prompt and messages cannot be defined at the same time"
1678
1651
  });
1679
1652
  }
1680
1653
  if (prompt.system != null && typeof prompt.system !== "string") {
1681
- throw new import_provider21.InvalidPromptError({
1654
+ throw new import_provider20.InvalidPromptError({
1682
1655
  prompt,
1683
1656
  message: "system must be a string"
1684
1657
  });
@@ -1691,13 +1664,13 @@ async function standardizePrompt(prompt) {
1691
1664
  } else if (prompt.messages != null) {
1692
1665
  messages = prompt.messages;
1693
1666
  } else {
1694
- throw new import_provider21.InvalidPromptError({
1667
+ throw new import_provider20.InvalidPromptError({
1695
1668
  prompt,
1696
1669
  message: "prompt or messages must be defined"
1697
1670
  });
1698
1671
  }
1699
1672
  if (messages.length === 0) {
1700
- throw new import_provider21.InvalidPromptError({
1673
+ throw new import_provider20.InvalidPromptError({
1701
1674
  prompt,
1702
1675
  message: "messages must not be empty"
1703
1676
  });
@@ -1707,7 +1680,7 @@ async function standardizePrompt(prompt) {
1707
1680
  schema: import_v46.z.array(modelMessageSchema)
1708
1681
  });
1709
1682
  if (!validationResult.success) {
1710
- throw new import_provider21.InvalidPromptError({
1683
+ throw new import_provider20.InvalidPromptError({
1711
1684
  prompt,
1712
1685
  message: "The messages must be a ModelMessage[]. If you have passed a UIMessage[], you can use convertToModelMessages to convert them.",
1713
1686
  cause: validationResult.error
@@ -1721,10 +1694,10 @@ async function standardizePrompt(prompt) {
1721
1694
 
1722
1695
  // src/prompt/wrap-gateway-error.ts
1723
1696
  var import_gateway2 = require("@ai-sdk/gateway");
1724
- var import_provider22 = require("@ai-sdk/provider");
1697
+ var import_provider21 = require("@ai-sdk/provider");
1725
1698
  function wrapGatewayError(error) {
1726
1699
  if (import_gateway2.GatewayAuthenticationError.isInstance(error) || import_gateway2.GatewayModelNotFoundError.isInstance(error)) {
1727
- return new import_provider22.AISDKError({
1700
+ return new import_provider21.AISDKError({
1728
1701
  name: "GatewayError",
1729
1702
  message: "Vercel AI Gateway access failed. If you want to use AI SDK providers directly, use the providers, e.g. @ai-sdk/openai, or register a different global default provider.",
1730
1703
  cause: error
@@ -1755,7 +1728,7 @@ function getBaseTelemetryAttributes({
1755
1728
  telemetry,
1756
1729
  headers
1757
1730
  }) {
1758
- var _a17;
1731
+ var _a16;
1759
1732
  return {
1760
1733
  "ai.model.provider": model.provider,
1761
1734
  "ai.model.id": model.modelId,
@@ -1765,7 +1738,7 @@ function getBaseTelemetryAttributes({
1765
1738
  return attributes;
1766
1739
  }, {}),
1767
1740
  // add metadata as attributes:
1768
- ...Object.entries((_a17 = telemetry == null ? void 0 : telemetry.metadata) != null ? _a17 : {}).reduce(
1741
+ ...Object.entries((_a16 = telemetry == null ? void 0 : telemetry.metadata) != null ? _a16 : {}).reduce(
1769
1742
  (attributes, [key, value]) => {
1770
1743
  attributes[`ai.telemetry.metadata.${key}`] = value;
1771
1744
  return attributes;
@@ -1790,7 +1763,7 @@ var noopTracer = {
1790
1763
  startSpan() {
1791
1764
  return noopSpan;
1792
1765
  },
1793
- startActiveSpan(name17, arg1, arg2, arg3) {
1766
+ startActiveSpan(name16, arg1, arg2, arg3) {
1794
1767
  if (typeof arg1 === "function") {
1795
1768
  return arg1(noopSpan);
1796
1769
  }
@@ -1860,14 +1833,14 @@ function getTracer({
1860
1833
  // src/telemetry/record-span.ts
1861
1834
  var import_api2 = require("@opentelemetry/api");
1862
1835
  async function recordSpan({
1863
- name: name17,
1836
+ name: name16,
1864
1837
  tracer,
1865
1838
  attributes,
1866
1839
  fn,
1867
1840
  endWhenDone = true
1868
1841
  }) {
1869
1842
  return tracer.startActiveSpan(
1870
- name17,
1843
+ name16,
1871
1844
  { attributes: await attributes },
1872
1845
  async (span) => {
1873
1846
  try {
@@ -1982,7 +1955,7 @@ function asArray(value) {
1982
1955
  }
1983
1956
 
1984
1957
  // src/util/retry-with-exponential-backoff.ts
1985
- var import_provider23 = require("@ai-sdk/provider");
1958
+ var import_provider22 = require("@ai-sdk/provider");
1986
1959
  var import_provider_utils7 = require("@ai-sdk/provider-utils");
1987
1960
  function getRetryDelayInMs({
1988
1961
  error,
@@ -2049,7 +2022,7 @@ async function _retryWithExponentialBackoff(f, {
2049
2022
  errors: newErrors
2050
2023
  });
2051
2024
  }
2052
- if (error instanceof Error && import_provider23.APICallError.isInstance(error) && error.isRetryable === true && tryNumber <= maxRetries) {
2025
+ if (error instanceof Error && import_provider22.APICallError.isInstance(error) && error.isRetryable === true && tryNumber <= maxRetries) {
2053
2026
  await (0, import_provider_utils7.delay)(
2054
2027
  getRetryDelayInMs({
2055
2028
  error,
@@ -2186,8 +2159,8 @@ async function executeToolCall({
2186
2159
  onPreliminaryToolResult
2187
2160
  }) {
2188
2161
  const { toolName, toolCallId, input } = toolCall;
2189
- const tool3 = tools == null ? void 0 : tools[toolName];
2190
- if ((tool3 == null ? void 0 : tool3.execute) == null) {
2162
+ const tool2 = tools == null ? void 0 : tools[toolName];
2163
+ if ((tool2 == null ? void 0 : tool2.execute) == null) {
2191
2164
  return void 0;
2192
2165
  }
2193
2166
  return recordSpan({
@@ -2211,7 +2184,7 @@ async function executeToolCall({
2211
2184
  let output;
2212
2185
  try {
2213
2186
  const stream = (0, import_provider_utils8.executeTool)({
2214
- execute: tool3.execute.bind(tool3),
2187
+ execute: tool2.execute.bind(tool2),
2215
2188
  input,
2216
2189
  options: {
2217
2190
  toolCallId,
@@ -2240,7 +2213,7 @@ async function executeToolCall({
2240
2213
  toolName,
2241
2214
  input,
2242
2215
  error,
2243
- dynamic: tool3.type === "dynamic"
2216
+ dynamic: tool2.type === "dynamic"
2244
2217
  };
2245
2218
  }
2246
2219
  try {
@@ -2262,7 +2235,7 @@ async function executeToolCall({
2262
2235
  toolName,
2263
2236
  input,
2264
2237
  output,
2265
- dynamic: tool3.type === "dynamic"
2238
+ dynamic: tool2.type === "dynamic"
2266
2239
  };
2267
2240
  }
2268
2241
  });
@@ -2315,18 +2288,18 @@ var DefaultGeneratedFileWithType = class extends DefaultGeneratedFile {
2315
2288
 
2316
2289
  // src/generate-text/is-approval-needed.ts
2317
2290
  async function isApprovalNeeded({
2318
- tool: tool3,
2291
+ tool: tool2,
2319
2292
  toolCall,
2320
2293
  messages,
2321
2294
  experimental_context
2322
2295
  }) {
2323
- if (tool3.needsApproval == null) {
2296
+ if (tool2.needsApproval == null) {
2324
2297
  return false;
2325
2298
  }
2326
- if (typeof tool3.needsApproval === "boolean") {
2327
- return tool3.needsApproval;
2299
+ if (typeof tool2.needsApproval === "boolean") {
2300
+ return tool2.needsApproval;
2328
2301
  }
2329
- return await tool3.needsApproval(toolCall.input, {
2302
+ return await tool2.needsApproval(toolCall.input, {
2330
2303
  toolCallId: toolCall.toolCallId,
2331
2304
  messages,
2332
2305
  experimental_context
@@ -2417,8 +2390,8 @@ async function doParseToolCall({
2417
2390
  tools
2418
2391
  }) {
2419
2392
  const toolName = toolCall.toolName;
2420
- const tool3 = tools[toolName];
2421
- if (tool3 == null) {
2393
+ const tool2 = tools[toolName];
2394
+ if (tool2 == null) {
2422
2395
  if (toolCall.providerExecuted && toolCall.dynamic) {
2423
2396
  return await parseProviderExecutedDynamicToolCall(toolCall);
2424
2397
  }
@@ -2427,7 +2400,7 @@ async function doParseToolCall({
2427
2400
  availableTools: Object.keys(tools)
2428
2401
  });
2429
2402
  }
2430
- const schema = (0, import_provider_utils10.asSchema)(tool3.inputSchema);
2403
+ const schema = (0, import_provider_utils10.asSchema)(tool2.inputSchema);
2431
2404
  const parseResult = toolCall.input.trim() === "" ? await (0, import_provider_utils10.safeValidateTypes)({ value: {}, schema }) : await (0, import_provider_utils10.safeParseJSON)({ text: toolCall.input, schema });
2432
2405
  if (parseResult.success === false) {
2433
2406
  throw new InvalidToolInputError({
@@ -2436,7 +2409,7 @@ async function doParseToolCall({
2436
2409
  cause: parseResult.error
2437
2410
  });
2438
2411
  }
2439
- return tool3.type === "dynamic" ? {
2412
+ return tool2.type === "dynamic" ? {
2440
2413
  type: "tool-call",
2441
2414
  toolCallId: toolCall.toolCallId,
2442
2415
  toolName: toolCall.toolName,
@@ -2522,8 +2495,8 @@ function stepCountIs(stepCount) {
2522
2495
  }
2523
2496
  function hasToolCall(toolName) {
2524
2497
  return ({ steps }) => {
2525
- var _a17, _b, _c;
2526
- return (_c = (_b = (_a17 = steps[steps.length - 1]) == null ? void 0 : _a17.toolCalls) == null ? void 0 : _b.some(
2498
+ var _a16, _b, _c;
2499
+ return (_c = (_b = (_a16 = steps[steps.length - 1]) == null ? void 0 : _a16.toolCalls) == null ? void 0 : _b.some(
2527
2500
  (toolCall) => toolCall.toolName === toolName
2528
2501
  )) != null ? _c : false;
2529
2502
  };
@@ -2647,7 +2620,8 @@ async function generateText({
2647
2620
  abortSignal,
2648
2621
  headers,
2649
2622
  stopWhen = stepCountIs(1),
2650
- experimental_output: output,
2623
+ experimental_output,
2624
+ output = experimental_output,
2651
2625
  experimental_telemetry: telemetry,
2652
2626
  providerOptions,
2653
2627
  experimental_activeTools,
@@ -2710,7 +2684,7 @@ async function generateText({
2710
2684
  }),
2711
2685
  tracer,
2712
2686
  fn: async (span) => {
2713
- var _a17, _b, _c, _d, _e, _f, _g;
2687
+ var _a16, _b, _c, _d, _e, _f, _g;
2714
2688
  const initialMessages = initialPrompt.messages;
2715
2689
  const responseMessages = [];
2716
2690
  const { approvedToolApprovals, deniedToolApprovals } = collectToolApprovals({ messages: initialMessages });
@@ -2767,7 +2741,7 @@ async function generateText({
2767
2741
  messages: stepInputMessages
2768
2742
  }));
2769
2743
  const stepModel = resolveLanguageModel(
2770
- (_a17 = prepareStepResult == null ? void 0 : prepareStepResult.model) != null ? _a17 : model
2744
+ (_a16 = prepareStepResult == null ? void 0 : prepareStepResult.model) != null ? _a16 : model
2771
2745
  );
2772
2746
  const promptMessages = await convertToLanguageModelPrompt({
2773
2747
  prompt: {
@@ -2784,7 +2758,7 @@ async function generateText({
2784
2758
  });
2785
2759
  currentModelResponse = await retry(
2786
2760
  () => {
2787
- var _a18;
2761
+ var _a17;
2788
2762
  return recordSpan({
2789
2763
  name: "ai.generateText.doGenerate",
2790
2764
  attributes: selectTelemetryAttributes({
@@ -2804,7 +2778,7 @@ async function generateText({
2804
2778
  },
2805
2779
  "ai.prompt.tools": {
2806
2780
  // convert the language model level tools:
2807
- input: () => stepTools == null ? void 0 : stepTools.map((tool3) => JSON.stringify(tool3))
2781
+ input: () => stepTools == null ? void 0 : stepTools.map((tool2) => JSON.stringify(tool2))
2808
2782
  },
2809
2783
  "ai.prompt.toolChoice": {
2810
2784
  input: () => stepToolChoice != null ? JSON.stringify(stepToolChoice) : void 0
@@ -2816,14 +2790,14 @@ async function generateText({
2816
2790
  "gen_ai.request.max_tokens": settings.maxOutputTokens,
2817
2791
  "gen_ai.request.presence_penalty": settings.presencePenalty,
2818
2792
  "gen_ai.request.stop_sequences": settings.stopSequences,
2819
- "gen_ai.request.temperature": (_a18 = settings.temperature) != null ? _a18 : void 0,
2793
+ "gen_ai.request.temperature": (_a17 = settings.temperature) != null ? _a17 : void 0,
2820
2794
  "gen_ai.request.top_k": settings.topK,
2821
2795
  "gen_ai.request.top_p": settings.topP
2822
2796
  }
2823
2797
  }),
2824
2798
  tracer,
2825
2799
  fn: async (span2) => {
2826
- var _a19, _b2, _c2, _d2, _e2, _f2, _g2, _h;
2800
+ var _a18, _b2, _c2, _d2, _e2, _f2, _g2, _h;
2827
2801
  const result = await stepModel.doGenerate({
2828
2802
  ...callSettings2,
2829
2803
  tools: stepTools,
@@ -2835,7 +2809,7 @@ async function generateText({
2835
2809
  headers: headersWithUserAgent
2836
2810
  });
2837
2811
  const responseData = {
2838
- id: (_b2 = (_a19 = result.response) == null ? void 0 : _a19.id) != null ? _b2 : generateId2(),
2812
+ id: (_b2 = (_a18 = result.response) == null ? void 0 : _a18.id) != null ? _b2 : generateId2(),
2839
2813
  timestamp: (_d2 = (_c2 = result.response) == null ? void 0 : _c2.timestamp) != null ? _d2 : currentDate(),
2840
2814
  modelId: (_f2 = (_e2 = result.response) == null ? void 0 : _e2.modelId) != null ? _f2 : stepModel.modelId,
2841
2815
  headers: (_g2 = result.response) == null ? void 0 : _g2.headers,
@@ -2896,12 +2870,12 @@ async function generateText({
2896
2870
  if (toolCall.invalid) {
2897
2871
  continue;
2898
2872
  }
2899
- const tool3 = tools == null ? void 0 : tools[toolCall.toolName];
2900
- if (tool3 == null) {
2873
+ const tool2 = tools == null ? void 0 : tools[toolCall.toolName];
2874
+ if (tool2 == null) {
2901
2875
  continue;
2902
2876
  }
2903
- if ((tool3 == null ? void 0 : tool3.onInputAvailable) != null) {
2904
- await tool3.onInputAvailable({
2877
+ if ((tool2 == null ? void 0 : tool2.onInputAvailable) != null) {
2878
+ await tool2.onInputAvailable({
2905
2879
  input: toolCall.input,
2906
2880
  toolCallId: toolCall.toolCallId,
2907
2881
  messages: stepInputMessages,
@@ -2910,7 +2884,7 @@ async function generateText({
2910
2884
  });
2911
2885
  }
2912
2886
  if (await isApprovalNeeded({
2913
- tool: tool3,
2887
+ tool: tool2,
2914
2888
  toolCall,
2915
2889
  messages: stepInputMessages,
2916
2890
  experimental_context
@@ -3158,6 +3132,9 @@ var DefaultGenerateTextResult = class {
3158
3132
  return this.finalStep.usage;
3159
3133
  }
3160
3134
  get experimental_output() {
3135
+ return this.output;
3136
+ }
3137
+ get output() {
3161
3138
  if (this.resolvedOutput == null) {
3162
3139
  throw new NoOutputSpecifiedError();
3163
3140
  }
@@ -3237,7 +3214,7 @@ function asContent({
3237
3214
  }
3238
3215
 
3239
3216
  // src/generate-text/stream-text.ts
3240
- var import_provider24 = require("@ai-sdk/provider");
3217
+ var import_provider23 = require("@ai-sdk/provider");
3241
3218
  var import_provider_utils16 = require("@ai-sdk/provider-utils");
3242
3219
 
3243
3220
  // src/util/prepare-headers.ts
@@ -3967,7 +3944,7 @@ function processUIMessageStream({
3967
3944
  new TransformStream({
3968
3945
  async transform(chunk, controller) {
3969
3946
  await runUpdateMessageJob(async ({ state, write }) => {
3970
- var _a17, _b, _c, _d;
3947
+ var _a16, _b, _c, _d;
3971
3948
  function getToolInvocation(toolCallId) {
3972
3949
  const toolInvocations = state.message.parts.filter(
3973
3950
  isToolOrDynamicToolUIPart
@@ -3983,7 +3960,7 @@ function processUIMessageStream({
3983
3960
  return toolInvocation;
3984
3961
  }
3985
3962
  function updateToolPart(options) {
3986
- var _a18;
3963
+ var _a17;
3987
3964
  const part = state.message.parts.find(
3988
3965
  (part2) => isToolUIPart(part2) && part2.toolCallId === options.toolCallId
3989
3966
  );
@@ -3996,7 +3973,7 @@ function processUIMessageStream({
3996
3973
  anyPart.errorText = anyOptions.errorText;
3997
3974
  anyPart.rawInput = anyOptions.rawInput;
3998
3975
  anyPart.preliminary = anyOptions.preliminary;
3999
- anyPart.providerExecuted = (_a18 = anyOptions.providerExecuted) != null ? _a18 : part.providerExecuted;
3976
+ anyPart.providerExecuted = (_a17 = anyOptions.providerExecuted) != null ? _a17 : part.providerExecuted;
4000
3977
  if (anyOptions.providerMetadata != null && part.state === "input-available") {
4001
3978
  part.callProviderMetadata = anyOptions.providerMetadata;
4002
3979
  }
@@ -4016,7 +3993,7 @@ function processUIMessageStream({
4016
3993
  }
4017
3994
  }
4018
3995
  function updateDynamicToolPart(options) {
4019
- var _a18, _b2;
3996
+ var _a17, _b2;
4020
3997
  const part = state.message.parts.find(
4021
3998
  (part2) => part2.type === "dynamic-tool" && part2.toolCallId === options.toolCallId
4022
3999
  );
@@ -4028,7 +4005,7 @@ function processUIMessageStream({
4028
4005
  anyPart.input = anyOptions.input;
4029
4006
  anyPart.output = anyOptions.output;
4030
4007
  anyPart.errorText = anyOptions.errorText;
4031
- anyPart.rawInput = (_a18 = anyOptions.rawInput) != null ? _a18 : anyPart.rawInput;
4008
+ anyPart.rawInput = (_a17 = anyOptions.rawInput) != null ? _a17 : anyPart.rawInput;
4032
4009
  anyPart.preliminary = anyOptions.preliminary;
4033
4010
  anyPart.providerExecuted = (_b2 = anyOptions.providerExecuted) != null ? _b2 : part.providerExecuted;
4034
4011
  if (anyOptions.providerMetadata != null && part.state === "input-available") {
@@ -4077,7 +4054,7 @@ function processUIMessageStream({
4077
4054
  case "text-delta": {
4078
4055
  const textPart = state.activeTextParts[chunk.id];
4079
4056
  textPart.text += chunk.delta;
4080
- textPart.providerMetadata = (_a17 = chunk.providerMetadata) != null ? _a17 : textPart.providerMetadata;
4057
+ textPart.providerMetadata = (_a16 = chunk.providerMetadata) != null ? _a16 : textPart.providerMetadata;
4081
4058
  write();
4082
4059
  break;
4083
4060
  }
@@ -4501,11 +4478,11 @@ function createAsyncIterableStream(source) {
4501
4478
  const reader = this.getReader();
4502
4479
  let finished = false;
4503
4480
  async function cleanup(cancelStream) {
4504
- var _a17;
4481
+ var _a16;
4505
4482
  finished = true;
4506
4483
  try {
4507
4484
  if (cancelStream) {
4508
- await ((_a17 = reader.cancel) == null ? void 0 : _a17.call(reader));
4485
+ await ((_a16 = reader.cancel) == null ? void 0 : _a16.call(reader));
4509
4486
  }
4510
4487
  } finally {
4511
4488
  try {
@@ -4692,25 +4669,25 @@ var DelayedPromise = class {
4692
4669
  return this._promise;
4693
4670
  }
4694
4671
  resolve(value) {
4695
- var _a17;
4672
+ var _a16;
4696
4673
  this.status = { type: "resolved", value };
4697
4674
  if (this._promise) {
4698
- (_a17 = this._resolve) == null ? void 0 : _a17.call(this, value);
4675
+ (_a16 = this._resolve) == null ? void 0 : _a16.call(this, value);
4699
4676
  }
4700
4677
  }
4701
4678
  reject(error) {
4702
- var _a17;
4679
+ var _a16;
4703
4680
  this.status = { type: "rejected", error };
4704
4681
  if (this._promise) {
4705
- (_a17 = this._reject) == null ? void 0 : _a17.call(this, error);
4682
+ (_a16 = this._reject) == null ? void 0 : _a16.call(this, error);
4706
4683
  }
4707
4684
  }
4708
4685
  };
4709
4686
 
4710
4687
  // src/util/now.ts
4711
4688
  function now() {
4712
- var _a17, _b;
4713
- return (_b = (_a17 = globalThis == null ? void 0 : globalThis.performance) == null ? void 0 : _a17.now()) != null ? _b : Date.now();
4689
+ var _a16, _b;
4690
+ return (_b = (_a16 = globalThis == null ? void 0 : globalThis.performance) == null ? void 0 : _a16.now()) != null ? _b : Date.now();
4714
4691
  }
4715
4692
 
4716
4693
  // src/generate-text/run-tools-transformation.ts
@@ -4806,12 +4783,12 @@ function runToolsTransformation({
4806
4783
  });
4807
4784
  break;
4808
4785
  }
4809
- const tool3 = tools == null ? void 0 : tools[toolCall.toolName];
4810
- if (tool3 == null) {
4786
+ const tool2 = tools == null ? void 0 : tools[toolCall.toolName];
4787
+ if (tool2 == null) {
4811
4788
  break;
4812
4789
  }
4813
- if (tool3.onInputAvailable != null) {
4814
- await tool3.onInputAvailable({
4790
+ if (tool2.onInputAvailable != null) {
4791
+ await tool2.onInputAvailable({
4815
4792
  input: toolCall.input,
4816
4793
  toolCallId: toolCall.toolCallId,
4817
4794
  messages,
@@ -4820,7 +4797,7 @@ function runToolsTransformation({
4820
4797
  });
4821
4798
  }
4822
4799
  if (await isApprovalNeeded({
4823
- tool: tool3,
4800
+ tool: tool2,
4824
4801
  toolCall,
4825
4802
  messages,
4826
4803
  experimental_context
@@ -4833,7 +4810,7 @@ function runToolsTransformation({
4833
4810
  break;
4834
4811
  }
4835
4812
  toolInputs.set(toolCall.toolCallId, toolCall.input);
4836
- if (tool3.execute != null && toolCall.providerExecuted !== true) {
4813
+ if (tool2.execute != null && toolCall.providerExecuted !== true) {
4837
4814
  const toolExecutionId = generateId2();
4838
4815
  outstandingToolResults.add(toolExecutionId);
4839
4816
  executeToolCall({
@@ -4937,7 +4914,8 @@ function streamText({
4937
4914
  abortSignal,
4938
4915
  headers,
4939
4916
  stopWhen = stepCountIs(1),
4940
- experimental_output: output,
4917
+ experimental_output,
4918
+ output = experimental_output,
4941
4919
  experimental_telemetry: telemetry,
4942
4920
  prepareStep,
4943
4921
  providerOptions,
@@ -5110,7 +5088,7 @@ var DefaultStreamTextResult = class {
5110
5088
  let activeReasoningContent = {};
5111
5089
  const eventProcessor = new TransformStream({
5112
5090
  async transform(chunk, controller) {
5113
- var _a17, _b, _c, _d;
5091
+ var _a16, _b, _c, _d;
5114
5092
  controller.enqueue(chunk);
5115
5093
  const { part } = chunk;
5116
5094
  if (part.type === "text-delta" || part.type === "reasoning-delta" || part.type === "source" || part.type === "tool-call" || part.type === "tool-result" || part.type === "tool-input-start" || part.type === "tool-input-delta" || part.type === "raw") {
@@ -5140,7 +5118,7 @@ var DefaultStreamTextResult = class {
5140
5118
  return;
5141
5119
  }
5142
5120
  activeText.text += part.text;
5143
- activeText.providerMetadata = (_a17 = part.providerMetadata) != null ? _a17 : activeText.providerMetadata;
5121
+ activeText.providerMetadata = (_a16 = part.providerMetadata) != null ? _a16 : activeText.providerMetadata;
5144
5122
  }
5145
5123
  if (part.type === "text-end") {
5146
5124
  const activeText = activeTextContent[part.id];
@@ -5299,8 +5277,8 @@ var DefaultStreamTextResult = class {
5299
5277
  "ai.response.text": { output: () => finalStep.text },
5300
5278
  "ai.response.toolCalls": {
5301
5279
  output: () => {
5302
- var _a17;
5303
- return ((_a17 = finalStep.toolCalls) == null ? void 0 : _a17.length) ? JSON.stringify(finalStep.toolCalls) : void 0;
5280
+ var _a16;
5281
+ return ((_a16 = finalStep.toolCalls) == null ? void 0 : _a16.length) ? JSON.stringify(finalStep.toolCalls) : void 0;
5304
5282
  }
5305
5283
  },
5306
5284
  "ai.response.providerMetadata": JSON.stringify(
@@ -5480,7 +5458,7 @@ var DefaultStreamTextResult = class {
5480
5458
  responseMessages,
5481
5459
  usage
5482
5460
  }) {
5483
- var _a17, _b, _c, _d, _e;
5461
+ var _a16, _b, _c, _d, _e;
5484
5462
  const includeRawChunks2 = self.includeRawChunks;
5485
5463
  stepFinish = new DelayedPromise();
5486
5464
  const stepInputMessages = [...initialMessages, ...responseMessages];
@@ -5491,7 +5469,7 @@ var DefaultStreamTextResult = class {
5491
5469
  messages: stepInputMessages
5492
5470
  }));
5493
5471
  const stepModel = resolveLanguageModel(
5494
- (_a17 = prepareStepResult == null ? void 0 : prepareStepResult.model) != null ? _a17 : model
5472
+ (_a16 = prepareStepResult == null ? void 0 : prepareStepResult.model) != null ? _a16 : model
5495
5473
  );
5496
5474
  const promptMessages = await convertToLanguageModelPrompt({
5497
5475
  prompt: {
@@ -5530,7 +5508,7 @@ var DefaultStreamTextResult = class {
5530
5508
  },
5531
5509
  "ai.prompt.tools": {
5532
5510
  // convert the language model level tools:
5533
- input: () => stepTools == null ? void 0 : stepTools.map((tool3) => JSON.stringify(tool3))
5511
+ input: () => stepTools == null ? void 0 : stepTools.map((tool2) => JSON.stringify(tool2))
5534
5512
  },
5535
5513
  "ai.prompt.toolChoice": {
5536
5514
  input: () => stepToolChoice != null ? JSON.stringify(stepToolChoice) : void 0
@@ -5602,7 +5580,7 @@ var DefaultStreamTextResult = class {
5602
5580
  streamWithToolResults.pipeThrough(
5603
5581
  new TransformStream({
5604
5582
  async transform(chunk, controller) {
5605
- var _a18, _b2, _c2, _d2, _e2;
5583
+ var _a17, _b2, _c2, _d2, _e2;
5606
5584
  if (chunk.type === "stream-start") {
5607
5585
  warnings = chunk.warnings;
5608
5586
  return;
@@ -5675,7 +5653,7 @@ var DefaultStreamTextResult = class {
5675
5653
  }
5676
5654
  case "response-metadata": {
5677
5655
  stepResponse = {
5678
- id: (_a18 = chunk.id) != null ? _a18 : stepResponse.id,
5656
+ id: (_a17 = chunk.id) != null ? _a17 : stepResponse.id,
5679
5657
  timestamp: (_b2 = chunk.timestamp) != null ? _b2 : stepResponse.timestamp,
5680
5658
  modelId: (_c2 = chunk.modelId) != null ? _c2 : stepResponse.modelId
5681
5659
  };
@@ -5703,9 +5681,9 @@ var DefaultStreamTextResult = class {
5703
5681
  }
5704
5682
  case "tool-input-start": {
5705
5683
  activeToolCallToolNames[chunk.id] = chunk.toolName;
5706
- const tool3 = tools == null ? void 0 : tools[chunk.toolName];
5707
- if ((tool3 == null ? void 0 : tool3.onInputStart) != null) {
5708
- await tool3.onInputStart({
5684
+ const tool2 = tools == null ? void 0 : tools[chunk.toolName];
5685
+ if ((tool2 == null ? void 0 : tool2.onInputStart) != null) {
5686
+ await tool2.onInputStart({
5709
5687
  toolCallId: chunk.id,
5710
5688
  messages: stepInputMessages,
5711
5689
  abortSignal,
@@ -5714,7 +5692,7 @@ var DefaultStreamTextResult = class {
5714
5692
  }
5715
5693
  controller.enqueue({
5716
5694
  ...chunk,
5717
- dynamic: (_e2 = chunk.dynamic) != null ? _e2 : (tool3 == null ? void 0 : tool3.type) === "dynamic"
5695
+ dynamic: (_e2 = chunk.dynamic) != null ? _e2 : (tool2 == null ? void 0 : tool2.type) === "dynamic"
5718
5696
  });
5719
5697
  break;
5720
5698
  }
@@ -5725,9 +5703,9 @@ var DefaultStreamTextResult = class {
5725
5703
  }
5726
5704
  case "tool-input-delta": {
5727
5705
  const toolName = activeToolCallToolNames[chunk.id];
5728
- const tool3 = tools == null ? void 0 : tools[toolName];
5729
- if ((tool3 == null ? void 0 : tool3.onInputDelta) != null) {
5730
- await tool3.onInputDelta({
5706
+ const tool2 = tools == null ? void 0 : tools[toolName];
5707
+ if ((tool2 == null ? void 0 : tool2.onInputDelta) != null) {
5708
+ await tool2.onInputDelta({
5731
5709
  inputTextDelta: chunk.delta,
5732
5710
  toolCallId: chunk.id,
5733
5711
  messages: stepInputMessages,
@@ -5977,17 +5955,20 @@ var DefaultStreamTextResult = class {
5977
5955
  );
5978
5956
  }
5979
5957
  async consumeStream(options) {
5980
- var _a17;
5958
+ var _a16;
5981
5959
  try {
5982
5960
  await consumeStream({
5983
5961
  stream: this.fullStream,
5984
5962
  onError: options == null ? void 0 : options.onError
5985
5963
  });
5986
5964
  } catch (error) {
5987
- (_a17 = options == null ? void 0 : options.onError) == null ? void 0 : _a17.call(options, error);
5965
+ (_a16 = options == null ? void 0 : options.onError) == null ? void 0 : _a16.call(options, error);
5988
5966
  }
5989
5967
  }
5990
5968
  get experimental_partialOutputStream() {
5969
+ return this.partialOutputStream;
5970
+ }
5971
+ get partialOutputStream() {
5991
5972
  if (this.output == null) {
5992
5973
  throw new NoOutputSpecifiedError();
5993
5974
  }
@@ -6012,19 +5993,19 @@ var DefaultStreamTextResult = class {
6012
5993
  sendSources = false,
6013
5994
  sendStart = true,
6014
5995
  sendFinish = true,
6015
- onError = import_provider24.getErrorMessage
5996
+ onError = import_provider23.getErrorMessage
6016
5997
  } = {}) {
6017
5998
  const responseMessageId = generateMessageId != null ? getResponseUIMessageId({
6018
5999
  originalMessages,
6019
6000
  responseMessageId: generateMessageId
6020
6001
  }) : void 0;
6021
6002
  const isDynamic = (part) => {
6022
- var _a17;
6023
- const tool3 = (_a17 = this.tools) == null ? void 0 : _a17[part.toolName];
6024
- if (tool3 == null) {
6003
+ var _a16;
6004
+ const tool2 = (_a16 = this.tools) == null ? void 0 : _a16[part.toolName];
6005
+ if (tool2 == null) {
6025
6006
  return part.dynamic;
6026
6007
  }
6027
- return (tool3 == null ? void 0 : tool3.type) === "dynamic" ? true : void 0;
6008
+ return (tool2 == null ? void 0 : tool2.type) === "dynamic" ? true : void 0;
6028
6009
  };
6029
6010
  const baseStream = this.fullStream.pipeThrough(
6030
6011
  new TransformStream({
@@ -6354,10 +6335,10 @@ var ToolLoopAgent = class {
6354
6335
  return this.settings.tools;
6355
6336
  }
6356
6337
  async prepareCall(options) {
6357
- var _a17, _b, _c, _d;
6338
+ var _a16, _b, _c, _d;
6358
6339
  const baseCallArgs = {
6359
6340
  ...this.settings,
6360
- stopWhen: (_a17 = this.settings.stopWhen) != null ? _a17 : stepCountIs(20),
6341
+ stopWhen: (_a16 = this.settings.stopWhen) != null ? _a16 : stepCountIs(20),
6361
6342
  ...options
6362
6343
  };
6363
6344
  const preparedCallArgs = (_d = await ((_c = (_b = this.settings).prepareCall) == null ? void 0 : _c.call(_b, baseCallArgs))) != null ? _d : baseCallArgs;
@@ -6475,7 +6456,7 @@ function readUIMessageStream({
6475
6456
  onError,
6476
6457
  terminateOnError = false
6477
6458
  }) {
6478
- var _a17;
6459
+ var _a16;
6479
6460
  let controller;
6480
6461
  let hasErrored = false;
6481
6462
  const outputStream = new ReadableStream({
@@ -6484,7 +6465,7 @@ function readUIMessageStream({
6484
6465
  }
6485
6466
  });
6486
6467
  const state = createStreamingUIMessageState({
6487
- messageId: (_a17 = message == null ? void 0 : message.id) != null ? _a17 : "",
6468
+ messageId: (_a16 = message == null ? void 0 : message.id) != null ? _a16 : "",
6488
6469
  lastMessage: message
6489
6470
  });
6490
6471
  const handleError = (error) => {
@@ -6551,7 +6532,7 @@ function convertToModelMessages(messages, options) {
6551
6532
  modelMessages.push({
6552
6533
  role: "user",
6553
6534
  content: message.parts.map((part) => {
6554
- var _a17;
6535
+ var _a16;
6555
6536
  if (isTextUIPart(part)) {
6556
6537
  return {
6557
6538
  type: "text",
@@ -6569,7 +6550,7 @@ function convertToModelMessages(messages, options) {
6569
6550
  };
6570
6551
  }
6571
6552
  if (isDataUIPart(part)) {
6572
- return (_a17 = options == null ? void 0 : options.convertDataPart) == null ? void 0 : _a17.call(
6553
+ return (_a16 = options == null ? void 0 : options.convertDataPart) == null ? void 0 : _a16.call(
6573
6554
  options,
6574
6555
  part
6575
6556
  );
@@ -6581,7 +6562,7 @@ function convertToModelMessages(messages, options) {
6581
6562
  case "assistant": {
6582
6563
  if (message.parts != null) {
6583
6564
  let processBlock2 = function() {
6584
- var _a17, _b, _c;
6565
+ var _a16, _b, _c;
6585
6566
  if (block.length === 0) {
6586
6567
  return;
6587
6568
  }
@@ -6613,7 +6594,7 @@ function convertToModelMessages(messages, options) {
6613
6594
  type: "tool-call",
6614
6595
  toolCallId: part.toolCallId,
6615
6596
  toolName,
6616
- input: part.state === "output-error" ? (_a17 = part.input) != null ? _a17 : "rawInput" in part ? part.rawInput : void 0 : part.input,
6597
+ input: part.state === "output-error" ? (_a16 = part.input) != null ? _a16 : "rawInput" in part ? part.rawInput : void 0 : part.input,
6617
6598
  providerExecuted: part.providerExecuted,
6618
6599
  ...part.callProviderMetadata != null ? { providerOptions: part.callProviderMetadata } : {}
6619
6600
  });
@@ -6662,9 +6643,9 @@ function convertToModelMessages(messages, options) {
6662
6643
  role: "tool",
6663
6644
  content: toolParts.flatMap(
6664
6645
  (toolPart) => {
6665
- var _a18, _b2, _c2;
6646
+ var _a17, _b2, _c2;
6666
6647
  const outputs = [];
6667
- if (((_a18 = toolPart.approval) == null ? void 0 : _a18.approved) != null) {
6648
+ if (((_a17 = toolPart.approval) == null ? void 0 : _a17.approved) != null) {
6668
6649
  outputs.push({
6669
6650
  type: "tool-approval-response",
6670
6651
  approvalId: toolPart.approval.id,
@@ -6736,7 +6717,7 @@ function convertToModelMessages(messages, options) {
6736
6717
  var convertToCoreMessages = convertToModelMessages;
6737
6718
 
6738
6719
  // src/ui/validate-ui-messages.ts
6739
- var import_provider25 = require("@ai-sdk/provider");
6720
+ var import_provider24 = require("@ai-sdk/provider");
6740
6721
  var import_provider_utils19 = require("@ai-sdk/provider-utils");
6741
6722
  var import_v48 = require("zod/v4");
6742
6723
  var uiMessagesSchema = (0, import_provider_utils19.lazySchema)(
@@ -7037,7 +7018,7 @@ async function safeValidateUIMessages({
7037
7018
  if (!dataSchema) {
7038
7019
  return {
7039
7020
  success: false,
7040
- error: new import_provider25.TypeValidationError({
7021
+ error: new import_provider24.TypeValidationError({
7041
7022
  value: dataPart.data,
7042
7023
  cause: `No data schema found for data part ${dataName}`
7043
7024
  })
@@ -7057,11 +7038,11 @@ async function safeValidateUIMessages({
7057
7038
  );
7058
7039
  for (const toolPart of toolParts) {
7059
7040
  const toolName = toolPart.type.slice(5);
7060
- const tool3 = tools[toolName];
7061
- if (!tool3) {
7041
+ const tool2 = tools[toolName];
7042
+ if (!tool2) {
7062
7043
  return {
7063
7044
  success: false,
7064
- error: new import_provider25.TypeValidationError({
7045
+ error: new import_provider24.TypeValidationError({
7065
7046
  value: toolPart.input,
7066
7047
  cause: `No tool schema found for tool part ${toolName}`
7067
7048
  })
@@ -7070,13 +7051,13 @@ async function safeValidateUIMessages({
7070
7051
  if (toolPart.state === "input-available" || toolPart.state === "output-available" || toolPart.state === "output-error") {
7071
7052
  await (0, import_provider_utils19.validateTypes)({
7072
7053
  value: toolPart.input,
7073
- schema: tool3.inputSchema
7054
+ schema: tool2.inputSchema
7074
7055
  });
7075
7056
  }
7076
- if (toolPart.state === "output-available" && tool3.outputSchema) {
7057
+ if (toolPart.state === "output-available" && tool2.outputSchema) {
7077
7058
  await (0, import_provider_utils19.validateTypes)({
7078
7059
  value: toolPart.output,
7079
- schema: tool3.outputSchema
7060
+ schema: tool2.outputSchema
7080
7061
  });
7081
7062
  }
7082
7063
  }
@@ -7226,7 +7207,7 @@ async function embed({
7226
7207
  }),
7227
7208
  tracer,
7228
7209
  fn: async (doEmbedSpan) => {
7229
- var _a17;
7210
+ var _a16;
7230
7211
  const modelResponse = await model.doEmbed({
7231
7212
  values: [value],
7232
7213
  abortSignal,
@@ -7234,7 +7215,7 @@ async function embed({
7234
7215
  providerOptions
7235
7216
  });
7236
7217
  const embedding2 = modelResponse.embeddings[0];
7237
- const usage2 = (_a17 = modelResponse.usage) != null ? _a17 : { tokens: NaN };
7218
+ const usage2 = (_a16 = modelResponse.usage) != null ? _a16 : { tokens: NaN };
7238
7219
  doEmbedSpan.setAttributes(
7239
7220
  await selectTelemetryAttributes({
7240
7221
  telemetry,
@@ -7344,7 +7325,7 @@ async function embedMany({
7344
7325
  }),
7345
7326
  tracer,
7346
7327
  fn: async (span) => {
7347
- var _a17;
7328
+ var _a16;
7348
7329
  const [maxEmbeddingsPerCall, supportsParallelCalls] = await Promise.all([
7349
7330
  model.maxEmbeddingsPerCall,
7350
7331
  model.supportsParallelCalls
@@ -7370,7 +7351,7 @@ async function embedMany({
7370
7351
  }),
7371
7352
  tracer,
7372
7353
  fn: async (doEmbedSpan) => {
7373
- var _a18;
7354
+ var _a17;
7374
7355
  const modelResponse = await model.doEmbed({
7375
7356
  values,
7376
7357
  abortSignal,
@@ -7378,7 +7359,7 @@ async function embedMany({
7378
7359
  providerOptions
7379
7360
  });
7380
7361
  const embeddings3 = modelResponse.embeddings;
7381
- const usage2 = (_a18 = modelResponse.usage) != null ? _a18 : { tokens: NaN };
7362
+ const usage2 = (_a17 = modelResponse.usage) != null ? _a17 : { tokens: NaN };
7382
7363
  doEmbedSpan.setAttributes(
7383
7364
  await selectTelemetryAttributes({
7384
7365
  telemetry,
@@ -7452,7 +7433,7 @@ async function embedMany({
7452
7433
  }),
7453
7434
  tracer,
7454
7435
  fn: async (doEmbedSpan) => {
7455
- var _a18;
7436
+ var _a17;
7456
7437
  const modelResponse = await model.doEmbed({
7457
7438
  values: chunk,
7458
7439
  abortSignal,
@@ -7460,7 +7441,7 @@ async function embedMany({
7460
7441
  providerOptions
7461
7442
  });
7462
7443
  const embeddings2 = modelResponse.embeddings;
7463
- const usage = (_a18 = modelResponse.usage) != null ? _a18 : { tokens: NaN };
7444
+ const usage = (_a17 = modelResponse.usage) != null ? _a17 : { tokens: NaN };
7464
7445
  doEmbedSpan.setAttributes(
7465
7446
  await selectTelemetryAttributes({
7466
7447
  telemetry,
@@ -7497,7 +7478,7 @@ async function embedMany({
7497
7478
  result.providerMetadata
7498
7479
  )) {
7499
7480
  providerMetadata[providerName] = {
7500
- ...(_a17 = providerMetadata[providerName]) != null ? _a17 : {},
7481
+ ...(_a16 = providerMetadata[providerName]) != null ? _a16 : {},
7501
7482
  ...metadata
7502
7483
  };
7503
7484
  }
@@ -7551,7 +7532,7 @@ async function generateImage({
7551
7532
  abortSignal,
7552
7533
  headers
7553
7534
  }) {
7554
- var _a17, _b;
7535
+ var _a16, _b;
7555
7536
  if (model.specificationVersion !== "v3") {
7556
7537
  throw new UnsupportedModelVersionError({
7557
7538
  version: model.specificationVersion,
@@ -7567,7 +7548,7 @@ async function generateImage({
7567
7548
  maxRetries: maxRetriesArg,
7568
7549
  abortSignal
7569
7550
  });
7570
- const maxImagesPerCallWithDefault = (_a17 = maxImagesPerCall != null ? maxImagesPerCall : await invokeModelMaxImagesPerCall(model)) != null ? _a17 : 1;
7551
+ const maxImagesPerCallWithDefault = (_a16 = maxImagesPerCall != null ? maxImagesPerCall : await invokeModelMaxImagesPerCall(model)) != null ? _a16 : 1;
7571
7552
  const callCount = Math.ceil(n / maxImagesPerCallWithDefault);
7572
7553
  const callImageCounts = Array.from({ length: callCount }, (_, i) => {
7573
7554
  if (i < callCount - 1) {
@@ -7600,13 +7581,13 @@ async function generateImage({
7600
7581
  images.push(
7601
7582
  ...result.images.map(
7602
7583
  (image) => {
7603
- var _a18;
7584
+ var _a17;
7604
7585
  return new DefaultGeneratedFile({
7605
7586
  data: image,
7606
- mediaType: (_a18 = detectMediaType({
7587
+ mediaType: (_a17 = detectMediaType({
7607
7588
  data: image,
7608
7589
  signatures: imageMediaTypeSignatures
7609
- })) != null ? _a18 : "image/png"
7590
+ })) != null ? _a17 : "image/png"
7610
7591
  });
7611
7592
  }
7612
7593
  )
@@ -7666,7 +7647,7 @@ function extractReasoningContent(content) {
7666
7647
  }
7667
7648
 
7668
7649
  // src/generate-object/output-strategy.ts
7669
- var import_provider26 = require("@ai-sdk/provider");
7650
+ var import_provider25 = require("@ai-sdk/provider");
7670
7651
  var import_provider_utils23 = require("@ai-sdk/provider-utils");
7671
7652
  var noSchemaOutputStrategy = {
7672
7653
  type: "no-schema",
@@ -7687,7 +7668,7 @@ var noSchemaOutputStrategy = {
7687
7668
  } : { success: true, value };
7688
7669
  },
7689
7670
  createElementStream() {
7690
- throw new import_provider26.UnsupportedFunctionalityError({
7671
+ throw new import_provider25.UnsupportedFunctionalityError({
7691
7672
  functionality: "element streams in no-schema mode"
7692
7673
  });
7693
7674
  }
@@ -7709,7 +7690,7 @@ var objectOutputStrategy = (schema) => ({
7709
7690
  return (0, import_provider_utils23.safeValidateTypes)({ value, schema });
7710
7691
  },
7711
7692
  createElementStream() {
7712
- throw new import_provider26.UnsupportedFunctionalityError({
7693
+ throw new import_provider25.UnsupportedFunctionalityError({
7713
7694
  functionality: "element streams in object mode"
7714
7695
  });
7715
7696
  }
@@ -7738,11 +7719,11 @@ var arrayOutputStrategy = (schema) => {
7738
7719
  isFirstDelta,
7739
7720
  isFinalDelta
7740
7721
  }) {
7741
- var _a17;
7742
- if (!(0, import_provider26.isJSONObject)(value) || !(0, import_provider26.isJSONArray)(value.elements)) {
7722
+ var _a16;
7723
+ if (!(0, import_provider25.isJSONObject)(value) || !(0, import_provider25.isJSONArray)(value.elements)) {
7743
7724
  return {
7744
7725
  success: false,
7745
- error: new import_provider26.TypeValidationError({
7726
+ error: new import_provider25.TypeValidationError({
7746
7727
  value,
7747
7728
  cause: "value must be an object that contains an array of elements"
7748
7729
  })
@@ -7761,7 +7742,7 @@ var arrayOutputStrategy = (schema) => {
7761
7742
  }
7762
7743
  resultArray.push(result.value);
7763
7744
  }
7764
- const publishedElementCount = (_a17 = latestObject == null ? void 0 : latestObject.length) != null ? _a17 : 0;
7745
+ const publishedElementCount = (_a16 = latestObject == null ? void 0 : latestObject.length) != null ? _a16 : 0;
7765
7746
  let textDelta = "";
7766
7747
  if (isFirstDelta) {
7767
7748
  textDelta += "[";
@@ -7782,10 +7763,10 @@ var arrayOutputStrategy = (schema) => {
7782
7763
  };
7783
7764
  },
7784
7765
  async validateFinalResult(value) {
7785
- if (!(0, import_provider26.isJSONObject)(value) || !(0, import_provider26.isJSONArray)(value.elements)) {
7766
+ if (!(0, import_provider25.isJSONObject)(value) || !(0, import_provider25.isJSONArray)(value.elements)) {
7786
7767
  return {
7787
7768
  success: false,
7788
- error: new import_provider26.TypeValidationError({
7769
+ error: new import_provider25.TypeValidationError({
7789
7770
  value,
7790
7771
  cause: "value must be an object that contains an array of elements"
7791
7772
  })
@@ -7848,10 +7829,10 @@ var enumOutputStrategy = (enumValues) => {
7848
7829
  additionalProperties: false
7849
7830
  }),
7850
7831
  async validateFinalResult(value) {
7851
- if (!(0, import_provider26.isJSONObject)(value) || typeof value.result !== "string") {
7832
+ if (!(0, import_provider25.isJSONObject)(value) || typeof value.result !== "string") {
7852
7833
  return {
7853
7834
  success: false,
7854
- error: new import_provider26.TypeValidationError({
7835
+ error: new import_provider25.TypeValidationError({
7855
7836
  value,
7856
7837
  cause: 'value must be an object that contains a string in the "result" property.'
7857
7838
  })
@@ -7860,17 +7841,17 @@ var enumOutputStrategy = (enumValues) => {
7860
7841
  const result = value.result;
7861
7842
  return enumValues.includes(result) ? { success: true, value: result } : {
7862
7843
  success: false,
7863
- error: new import_provider26.TypeValidationError({
7844
+ error: new import_provider25.TypeValidationError({
7864
7845
  value,
7865
7846
  cause: "value must be a string in the enum"
7866
7847
  })
7867
7848
  };
7868
7849
  },
7869
7850
  async validatePartialResult({ value, textDelta }) {
7870
- if (!(0, import_provider26.isJSONObject)(value) || typeof value.result !== "string") {
7851
+ if (!(0, import_provider25.isJSONObject)(value) || typeof value.result !== "string") {
7871
7852
  return {
7872
7853
  success: false,
7873
- error: new import_provider26.TypeValidationError({
7854
+ error: new import_provider25.TypeValidationError({
7874
7855
  value,
7875
7856
  cause: 'value must be an object that contains a string in the "result" property.'
7876
7857
  })
@@ -7883,7 +7864,7 @@ var enumOutputStrategy = (enumValues) => {
7883
7864
  if (value.result.length === 0 || possibleEnumValues.length === 0) {
7884
7865
  return {
7885
7866
  success: false,
7886
- error: new import_provider26.TypeValidationError({
7867
+ error: new import_provider25.TypeValidationError({
7887
7868
  value,
7888
7869
  cause: "value must be a string in the enum"
7889
7870
  })
@@ -7898,7 +7879,7 @@ var enumOutputStrategy = (enumValues) => {
7898
7879
  };
7899
7880
  },
7900
7881
  createElementStream() {
7901
- throw new import_provider26.UnsupportedFunctionalityError({
7882
+ throw new import_provider25.UnsupportedFunctionalityError({
7902
7883
  functionality: "element streams in enum mode"
7903
7884
  });
7904
7885
  }
@@ -7926,7 +7907,7 @@ function getOutputStrategy({
7926
7907
  }
7927
7908
 
7928
7909
  // src/generate-object/parse-and-validate-object-result.ts
7929
- var import_provider27 = require("@ai-sdk/provider");
7910
+ var import_provider26 = require("@ai-sdk/provider");
7930
7911
  var import_provider_utils24 = require("@ai-sdk/provider-utils");
7931
7912
  async function parseAndValidateObjectResult(result, outputStrategy, context) {
7932
7913
  const parseResult = await (0, import_provider_utils24.safeParseJSON)({ text: result });
@@ -7964,7 +7945,7 @@ async function parseAndValidateObjectResultWithRepair(result, outputStrategy, re
7964
7945
  try {
7965
7946
  return await parseAndValidateObjectResult(result, outputStrategy, context);
7966
7947
  } catch (error) {
7967
- if (repairText != null && NoObjectGeneratedError.isInstance(error) && (import_provider27.JSONParseError.isInstance(error.cause) || import_provider27.TypeValidationError.isInstance(error.cause))) {
7948
+ if (repairText != null && NoObjectGeneratedError.isInstance(error) && (import_provider26.JSONParseError.isInstance(error.cause) || import_provider26.TypeValidationError.isInstance(error.cause))) {
7968
7949
  const repairedText = await repairText({
7969
7950
  text: result,
7970
7951
  error: error.cause
@@ -8157,7 +8138,7 @@ async function generateObject(options) {
8157
8138
  settings: { ...callSettings, maxRetries }
8158
8139
  });
8159
8140
  const tracer = getTracer(telemetry);
8160
- const jsonSchema3 = await outputStrategy.jsonSchema();
8141
+ const jsonSchema2 = await outputStrategy.jsonSchema();
8161
8142
  try {
8162
8143
  return await recordSpan({
8163
8144
  name: "ai.generateObject",
@@ -8173,7 +8154,7 @@ async function generateObject(options) {
8173
8154
  "ai.prompt": {
8174
8155
  input: () => JSON.stringify({ system, prompt, messages })
8175
8156
  },
8176
- "ai.schema": jsonSchema3 != null ? { input: () => JSON.stringify(jsonSchema3) } : void 0,
8157
+ "ai.schema": jsonSchema2 != null ? { input: () => JSON.stringify(jsonSchema2) } : void 0,
8177
8158
  "ai.schema.name": schemaName,
8178
8159
  "ai.schema.description": schemaDescription,
8179
8160
  "ai.settings.output": outputStrategy.type
@@ -8181,7 +8162,7 @@ async function generateObject(options) {
8181
8162
  }),
8182
8163
  tracer,
8183
8164
  fn: async (span) => {
8184
- var _a17;
8165
+ var _a16;
8185
8166
  let result;
8186
8167
  let finishReason;
8187
8168
  let usage;
@@ -8227,11 +8208,11 @@ async function generateObject(options) {
8227
8208
  }),
8228
8209
  tracer,
8229
8210
  fn: async (span2) => {
8230
- var _a18, _b, _c, _d, _e, _f, _g, _h;
8211
+ var _a17, _b, _c, _d, _e, _f, _g, _h;
8231
8212
  const result2 = await model.doGenerate({
8232
8213
  responseFormat: {
8233
8214
  type: "json",
8234
- schema: jsonSchema3,
8215
+ schema: jsonSchema2,
8235
8216
  name: schemaName,
8236
8217
  description: schemaDescription
8237
8218
  },
@@ -8242,7 +8223,7 @@ async function generateObject(options) {
8242
8223
  headers: headersWithUserAgent
8243
8224
  });
8244
8225
  const responseData = {
8245
- id: (_b = (_a18 = result2.response) == null ? void 0 : _a18.id) != null ? _b : generateId2(),
8226
+ id: (_b = (_a17 = result2.response) == null ? void 0 : _a17.id) != null ? _b : generateId2(),
8246
8227
  timestamp: (_d = (_c = result2.response) == null ? void 0 : _c.timestamp) != null ? _d : currentDate(),
8247
8228
  modelId: (_f = (_e = result2.response) == null ? void 0 : _e.modelId) != null ? _f : model.modelId,
8248
8229
  headers: (_g = result2.response) == null ? void 0 : _g.headers,
@@ -8296,7 +8277,7 @@ async function generateObject(options) {
8296
8277
  usage = generateResult.usage;
8297
8278
  warnings = generateResult.warnings;
8298
8279
  resultProviderMetadata = generateResult.providerMetadata;
8299
- request = (_a17 = generateResult.request) != null ? _a17 : {};
8280
+ request = (_a16 = generateResult.request) != null ? _a16 : {};
8300
8281
  response = generateResult.responseData;
8301
8282
  reasoning = generateResult.reasoning;
8302
8283
  logWarnings(warnings);
@@ -8355,9 +8336,9 @@ var DefaultGenerateObjectResult = class {
8355
8336
  this.reasoning = options.reasoning;
8356
8337
  }
8357
8338
  toJsonResponse(init) {
8358
- var _a17;
8339
+ var _a16;
8359
8340
  return new Response(JSON.stringify(this.object), {
8360
- status: (_a17 = init == null ? void 0 : init.status) != null ? _a17 : 200,
8341
+ status: (_a16 = init == null ? void 0 : init.status) != null ? _a16 : 200,
8361
8342
  headers: prepareHeaders(init == null ? void 0 : init.headers, {
8362
8343
  "content-type": "application/json; charset=utf-8"
8363
8344
  })
@@ -8483,8 +8464,8 @@ function simulateReadableStream({
8483
8464
  chunkDelayInMs = 0,
8484
8465
  _internal
8485
8466
  }) {
8486
- var _a17;
8487
- const delay2 = (_a17 = _internal == null ? void 0 : _internal.delay) != null ? _a17 : import_provider_utils26.delay;
8467
+ var _a16;
8468
+ const delay2 = (_a16 = _internal == null ? void 0 : _internal.delay) != null ? _a16 : import_provider_utils26.delay;
8488
8469
  let index = 0;
8489
8470
  return new ReadableStream({
8490
8471
  async pull(controller) {
@@ -8746,7 +8727,7 @@ var DefaultStreamObjectResult = class {
8746
8727
  const transformedStream = stream.pipeThrough(new TransformStream(transformer)).pipeThrough(
8747
8728
  new TransformStream({
8748
8729
  async transform(chunk, controller) {
8749
- var _a17, _b, _c;
8730
+ var _a16, _b, _c;
8750
8731
  if (typeof chunk === "object" && chunk.type === "stream-start") {
8751
8732
  warnings = chunk.warnings;
8752
8733
  return;
@@ -8796,7 +8777,7 @@ var DefaultStreamObjectResult = class {
8796
8777
  switch (chunk.type) {
8797
8778
  case "response-metadata": {
8798
8779
  fullResponse = {
8799
- id: (_a17 = chunk.id) != null ? _a17 : fullResponse.id,
8780
+ id: (_a16 = chunk.id) != null ? _a16 : fullResponse.id,
8800
8781
  timestamp: (_b = chunk.timestamp) != null ? _b : fullResponse.timestamp,
8801
8782
  modelId: (_c = chunk.modelId) != null ? _c : fullResponse.modelId
8802
8783
  };
@@ -9063,7 +9044,7 @@ async function generateSpeech({
9063
9044
  abortSignal,
9064
9045
  headers
9065
9046
  }) {
9066
- var _a17;
9047
+ var _a16;
9067
9048
  const resolvedModel = resolveSpeechModel(model);
9068
9049
  if (!resolvedModel) {
9069
9050
  throw new Error("Model could not be resolved");
@@ -9096,10 +9077,10 @@ async function generateSpeech({
9096
9077
  return new DefaultSpeechResult({
9097
9078
  audio: new DefaultGeneratedAudioFile({
9098
9079
  data: result.audio,
9099
- mediaType: (_a17 = detectMediaType({
9080
+ mediaType: (_a16 = detectMediaType({
9100
9081
  data: result.audio,
9101
9082
  signatures: audioMediaTypeSignatures
9102
- })) != null ? _a17 : "audio/mp3"
9083
+ })) != null ? _a16 : "audio/mp3"
9103
9084
  }),
9104
9085
  warnings: result.warnings,
9105
9086
  responses: [result.response],
@@ -9108,11 +9089,11 @@ async function generateSpeech({
9108
9089
  }
9109
9090
  var DefaultSpeechResult = class {
9110
9091
  constructor(options) {
9111
- var _a17;
9092
+ var _a16;
9112
9093
  this.audio = options.audio;
9113
9094
  this.warnings = options.warnings;
9114
9095
  this.responses = options.responses;
9115
- this.providerMetadata = (_a17 = options.providerMetadata) != null ? _a17 : {};
9096
+ this.providerMetadata = (_a16 = options.providerMetadata) != null ? _a16 : {};
9116
9097
  }
9117
9098
  };
9118
9099
 
@@ -9124,7 +9105,7 @@ __export(output_exports, {
9124
9105
  object: () => object,
9125
9106
  text: () => text
9126
9107
  });
9127
- var import_provider28 = require("@ai-sdk/provider");
9108
+ var import_provider27 = require("@ai-sdk/provider");
9128
9109
  var import_provider_utils29 = require("@ai-sdk/provider-utils");
9129
9110
  var text = () => ({
9130
9111
  type: "text",
@@ -9142,9 +9123,9 @@ var object = ({
9142
9123
  const schema = (0, import_provider_utils29.asSchema)(inputSchema);
9143
9124
  return {
9144
9125
  type: "object",
9145
- responseFormat: (0, import_provider_utils29.resolve)(schema.jsonSchema).then((jsonSchema3) => ({
9126
+ responseFormat: (0, import_provider_utils29.resolve)(schema.jsonSchema).then((jsonSchema2) => ({
9146
9127
  type: "json",
9147
- schema: jsonSchema3
9128
+ schema: jsonSchema2
9148
9129
  })),
9149
9130
  async parseOutput({ text: text2 }, context) {
9150
9131
  const parseResult = await (0, import_provider_utils29.safeParseJSON)({ text: text2 });
@@ -9203,8 +9184,8 @@ var array = ({
9203
9184
  return {
9204
9185
  type: "object",
9205
9186
  // JSON schema that describes an array of elements:
9206
- responseFormat: (0, import_provider_utils29.resolve)(elementSchema.jsonSchema).then((jsonSchema3) => {
9207
- const { $schema, ...itemSchema } = jsonSchema3;
9187
+ responseFormat: (0, import_provider_utils29.resolve)(elementSchema.jsonSchema).then((jsonSchema2) => {
9188
+ const { $schema, ...itemSchema } = jsonSchema2;
9208
9189
  return {
9209
9190
  type: "json",
9210
9191
  schema: {
@@ -9234,7 +9215,7 @@ var array = ({
9234
9215
  if (outerValue == null || typeof outerValue !== "object" || !("elements" in outerValue) || !Array.isArray(outerValue.elements)) {
9235
9216
  throw new NoObjectGeneratedError({
9236
9217
  message: "No object generated: response did not match schema.",
9237
- cause: new import_provider28.TypeValidationError({
9218
+ cause: new import_provider27.TypeValidationError({
9238
9219
  value: outerValue,
9239
9220
  cause: "response must be an object with an elements array"
9240
9221
  }),
@@ -9330,7 +9311,7 @@ var choice = ({
9330
9311
  if (outerValue == null || typeof outerValue !== "object" || !("result" in outerValue) || typeof outerValue.result !== "string" || !choiceOptions.includes(outerValue.result)) {
9331
9312
  throw new NoObjectGeneratedError({
9332
9313
  message: "No object generated: response did not match schema.",
9333
- cause: new import_provider28.TypeValidationError({
9314
+ cause: new import_provider27.TypeValidationError({
9334
9315
  value: outerValue,
9335
9316
  cause: "response must be an object that contains a choice value."
9336
9317
  }),
@@ -9454,7 +9435,7 @@ function pruneMessages({
9454
9435
 
9455
9436
  // src/generate-text/smooth-stream.ts
9456
9437
  var import_provider_utils30 = require("@ai-sdk/provider-utils");
9457
- var import_provider29 = require("@ai-sdk/provider");
9438
+ var import_provider28 = require("@ai-sdk/provider");
9458
9439
  var CHUNKING_REGEXPS = {
9459
9440
  word: /\S+\s+/m,
9460
9441
  line: /\n+/m
@@ -9484,7 +9465,7 @@ function smoothStream({
9484
9465
  } else {
9485
9466
  const chunkingRegex = typeof chunking === "string" ? CHUNKING_REGEXPS[chunking] : chunking;
9486
9467
  if (chunkingRegex == null) {
9487
- throw new import_provider29.InvalidArgumentError({
9468
+ throw new import_provider28.InvalidArgumentError({
9488
9469
  argument: "chunking",
9489
9470
  message: `Chunking must be "word" or "line" or a RegExp. Received: ${chunking}`
9490
9471
  });
@@ -9801,7 +9782,7 @@ var doWrap = ({
9801
9782
  modelId,
9802
9783
  providerId
9803
9784
  }) => {
9804
- var _a17, _b, _c;
9785
+ var _a16, _b, _c;
9805
9786
  async function doTransform({
9806
9787
  params,
9807
9788
  type
@@ -9810,7 +9791,7 @@ var doWrap = ({
9810
9791
  }
9811
9792
  return {
9812
9793
  specificationVersion: "v3",
9813
- provider: (_a17 = providerId != null ? providerId : overrideProvider == null ? void 0 : overrideProvider({ model })) != null ? _a17 : model.provider,
9794
+ provider: (_a16 = providerId != null ? providerId : overrideProvider == null ? void 0 : overrideProvider({ model })) != null ? _a16 : model.provider,
9814
9795
  modelId: (_b = modelId != null ? modelId : overrideModelId == null ? void 0 : overrideModelId({ model })) != null ? _b : model.modelId,
9815
9796
  supportedUrls: (_c = overrideSupportedUrls == null ? void 0 : overrideSupportedUrls({ model })) != null ? _c : model.supportedUrls,
9816
9797
  async doGenerate(params) {
@@ -9856,7 +9837,7 @@ function wrapProvider({
9856
9837
  }
9857
9838
 
9858
9839
  // src/registry/custom-provider.ts
9859
- var import_provider30 = require("@ai-sdk/provider");
9840
+ var import_provider29 = require("@ai-sdk/provider");
9860
9841
  function customProvider({
9861
9842
  languageModels,
9862
9843
  textEmbeddingModels,
@@ -9873,7 +9854,7 @@ function customProvider({
9873
9854
  if (fallbackProvider) {
9874
9855
  return fallbackProvider.languageModel(modelId);
9875
9856
  }
9876
- throw new import_provider30.NoSuchModelError({ modelId, modelType: "languageModel" });
9857
+ throw new import_provider29.NoSuchModelError({ modelId, modelType: "languageModel" });
9877
9858
  },
9878
9859
  textEmbeddingModel(modelId) {
9879
9860
  if (textEmbeddingModels != null && modelId in textEmbeddingModels) {
@@ -9882,7 +9863,7 @@ function customProvider({
9882
9863
  if (fallbackProvider) {
9883
9864
  return fallbackProvider.textEmbeddingModel(modelId);
9884
9865
  }
9885
- throw new import_provider30.NoSuchModelError({ modelId, modelType: "textEmbeddingModel" });
9866
+ throw new import_provider29.NoSuchModelError({ modelId, modelType: "textEmbeddingModel" });
9886
9867
  },
9887
9868
  imageModel(modelId) {
9888
9869
  if (imageModels != null && modelId in imageModels) {
@@ -9891,7 +9872,7 @@ function customProvider({
9891
9872
  if (fallbackProvider == null ? void 0 : fallbackProvider.imageModel) {
9892
9873
  return fallbackProvider.imageModel(modelId);
9893
9874
  }
9894
- throw new import_provider30.NoSuchModelError({ modelId, modelType: "imageModel" });
9875
+ throw new import_provider29.NoSuchModelError({ modelId, modelType: "imageModel" });
9895
9876
  },
9896
9877
  transcriptionModel(modelId) {
9897
9878
  if (transcriptionModels != null && modelId in transcriptionModels) {
@@ -9900,7 +9881,7 @@ function customProvider({
9900
9881
  if (fallbackProvider == null ? void 0 : fallbackProvider.transcriptionModel) {
9901
9882
  return fallbackProvider.transcriptionModel(modelId);
9902
9883
  }
9903
- throw new import_provider30.NoSuchModelError({ modelId, modelType: "transcriptionModel" });
9884
+ throw new import_provider29.NoSuchModelError({ modelId, modelType: "transcriptionModel" });
9904
9885
  },
9905
9886
  speechModel(modelId) {
9906
9887
  if (speechModels != null && modelId in speechModels) {
@@ -9909,19 +9890,19 @@ function customProvider({
9909
9890
  if (fallbackProvider == null ? void 0 : fallbackProvider.speechModel) {
9910
9891
  return fallbackProvider.speechModel(modelId);
9911
9892
  }
9912
- throw new import_provider30.NoSuchModelError({ modelId, modelType: "speechModel" });
9893
+ throw new import_provider29.NoSuchModelError({ modelId, modelType: "speechModel" });
9913
9894
  }
9914
9895
  };
9915
9896
  }
9916
9897
  var experimental_customProvider = customProvider;
9917
9898
 
9918
9899
  // src/registry/no-such-provider-error.ts
9919
- var import_provider31 = require("@ai-sdk/provider");
9920
- var name16 = "AI_NoSuchProviderError";
9921
- var marker16 = `vercel.ai.error.${name16}`;
9922
- var symbol16 = Symbol.for(marker16);
9923
- var _a16;
9924
- var NoSuchProviderError = class extends import_provider31.NoSuchModelError {
9900
+ var import_provider30 = require("@ai-sdk/provider");
9901
+ var name15 = "AI_NoSuchProviderError";
9902
+ var marker15 = `vercel.ai.error.${name15}`;
9903
+ var symbol15 = Symbol.for(marker15);
9904
+ var _a15;
9905
+ var NoSuchProviderError = class extends import_provider30.NoSuchModelError {
9925
9906
  constructor({
9926
9907
  modelId,
9927
9908
  modelType,
@@ -9929,19 +9910,19 @@ var NoSuchProviderError = class extends import_provider31.NoSuchModelError {
9929
9910
  availableProviders,
9930
9911
  message = `No such provider: ${providerId} (available providers: ${availableProviders.join()})`
9931
9912
  }) {
9932
- super({ errorName: name16, modelId, modelType, message });
9933
- this[_a16] = true;
9913
+ super({ errorName: name15, modelId, modelType, message });
9914
+ this[_a15] = true;
9934
9915
  this.providerId = providerId;
9935
9916
  this.availableProviders = availableProviders;
9936
9917
  }
9937
9918
  static isInstance(error) {
9938
- return import_provider31.AISDKError.hasMarker(error, marker16);
9919
+ return import_provider30.AISDKError.hasMarker(error, marker15);
9939
9920
  }
9940
9921
  };
9941
- _a16 = symbol16;
9922
+ _a15 = symbol15;
9942
9923
 
9943
9924
  // src/registry/provider-registry.ts
9944
- var import_provider32 = require("@ai-sdk/provider");
9925
+ var import_provider31 = require("@ai-sdk/provider");
9945
9926
  function createProviderRegistry(providers, {
9946
9927
  separator = ":",
9947
9928
  languageModelMiddleware
@@ -9986,7 +9967,7 @@ var DefaultProviderRegistry = class {
9986
9967
  splitId(id, modelType) {
9987
9968
  const index = id.indexOf(this.separator);
9988
9969
  if (index === -1) {
9989
- throw new import_provider32.NoSuchModelError({
9970
+ throw new import_provider31.NoSuchModelError({
9990
9971
  modelId: id,
9991
9972
  modelType,
9992
9973
  message: `Invalid ${modelType} id for registry: ${id} (must be in the format "providerId${this.separator}modelId")`
@@ -9995,14 +9976,14 @@ var DefaultProviderRegistry = class {
9995
9976
  return [id.slice(0, index), id.slice(index + this.separator.length)];
9996
9977
  }
9997
9978
  languageModel(id) {
9998
- var _a17, _b;
9979
+ var _a16, _b;
9999
9980
  const [providerId, modelId] = this.splitId(id, "languageModel");
10000
- let model = (_b = (_a17 = this.getProvider(providerId, "languageModel")).languageModel) == null ? void 0 : _b.call(
10001
- _a17,
9981
+ let model = (_b = (_a16 = this.getProvider(providerId, "languageModel")).languageModel) == null ? void 0 : _b.call(
9982
+ _a16,
10002
9983
  modelId
10003
9984
  );
10004
9985
  if (model == null) {
10005
- throw new import_provider32.NoSuchModelError({ modelId: id, modelType: "languageModel" });
9986
+ throw new import_provider31.NoSuchModelError({ modelId: id, modelType: "languageModel" });
10006
9987
  }
10007
9988
  if (this.languageModelMiddleware != null) {
10008
9989
  model = wrapLanguageModel({
@@ -10013,12 +9994,12 @@ var DefaultProviderRegistry = class {
10013
9994
  return model;
10014
9995
  }
10015
9996
  textEmbeddingModel(id) {
10016
- var _a17;
9997
+ var _a16;
10017
9998
  const [providerId, modelId] = this.splitId(id, "textEmbeddingModel");
10018
9999
  const provider = this.getProvider(providerId, "textEmbeddingModel");
10019
- const model = (_a17 = provider.textEmbeddingModel) == null ? void 0 : _a17.call(provider, modelId);
10000
+ const model = (_a16 = provider.textEmbeddingModel) == null ? void 0 : _a16.call(provider, modelId);
10020
10001
  if (model == null) {
10021
- throw new import_provider32.NoSuchModelError({
10002
+ throw new import_provider31.NoSuchModelError({
10022
10003
  modelId: id,
10023
10004
  modelType: "textEmbeddingModel"
10024
10005
  });
@@ -10026,22 +10007,22 @@ var DefaultProviderRegistry = class {
10026
10007
  return model;
10027
10008
  }
10028
10009
  imageModel(id) {
10029
- var _a17;
10010
+ var _a16;
10030
10011
  const [providerId, modelId] = this.splitId(id, "imageModel");
10031
10012
  const provider = this.getProvider(providerId, "imageModel");
10032
- const model = (_a17 = provider.imageModel) == null ? void 0 : _a17.call(provider, modelId);
10013
+ const model = (_a16 = provider.imageModel) == null ? void 0 : _a16.call(provider, modelId);
10033
10014
  if (model == null) {
10034
- throw new import_provider32.NoSuchModelError({ modelId: id, modelType: "imageModel" });
10015
+ throw new import_provider31.NoSuchModelError({ modelId: id, modelType: "imageModel" });
10035
10016
  }
10036
10017
  return model;
10037
10018
  }
10038
10019
  transcriptionModel(id) {
10039
- var _a17;
10020
+ var _a16;
10040
10021
  const [providerId, modelId] = this.splitId(id, "transcriptionModel");
10041
10022
  const provider = this.getProvider(providerId, "transcriptionModel");
10042
- const model = (_a17 = provider.transcriptionModel) == null ? void 0 : _a17.call(provider, modelId);
10023
+ const model = (_a16 = provider.transcriptionModel) == null ? void 0 : _a16.call(provider, modelId);
10043
10024
  if (model == null) {
10044
- throw new import_provider32.NoSuchModelError({
10025
+ throw new import_provider31.NoSuchModelError({
10045
10026
  modelId: id,
10046
10027
  modelType: "transcriptionModel"
10047
10028
  });
@@ -10049,600 +10030,23 @@ var DefaultProviderRegistry = class {
10049
10030
  return model;
10050
10031
  }
10051
10032
  speechModel(id) {
10052
- var _a17;
10033
+ var _a16;
10053
10034
  const [providerId, modelId] = this.splitId(id, "speechModel");
10054
10035
  const provider = this.getProvider(providerId, "speechModel");
10055
- const model = (_a17 = provider.speechModel) == null ? void 0 : _a17.call(provider, modelId);
10036
+ const model = (_a16 = provider.speechModel) == null ? void 0 : _a16.call(provider, modelId);
10056
10037
  if (model == null) {
10057
- throw new import_provider32.NoSuchModelError({ modelId: id, modelType: "speechModel" });
10038
+ throw new import_provider31.NoSuchModelError({ modelId: id, modelType: "speechModel" });
10058
10039
  }
10059
10040
  return model;
10060
10041
  }
10061
10042
  };
10062
10043
 
10063
- // src/tool/mcp/mcp-client.ts
10064
- var import_provider_utils32 = require("@ai-sdk/provider-utils");
10065
-
10066
- // src/tool/mcp/mcp-sse-transport.ts
10067
- var import_provider_utils31 = require("@ai-sdk/provider-utils");
10068
-
10069
- // src/tool/mcp/json-rpc-message.ts
10070
- var import_v410 = require("zod/v4");
10071
-
10072
- // src/tool/mcp/types.ts
10073
- var import_v49 = require("zod/v4");
10074
- var LATEST_PROTOCOL_VERSION = "2025-06-18";
10075
- var SUPPORTED_PROTOCOL_VERSIONS = [
10076
- LATEST_PROTOCOL_VERSION,
10077
- "2025-03-26",
10078
- "2024-11-05"
10079
- ];
10080
- var ClientOrServerImplementationSchema = import_v49.z.looseObject({
10081
- name: import_v49.z.string(),
10082
- version: import_v49.z.string()
10083
- });
10084
- var BaseParamsSchema = import_v49.z.looseObject({
10085
- _meta: import_v49.z.optional(import_v49.z.object({}).loose())
10086
- });
10087
- var ResultSchema = BaseParamsSchema;
10088
- var RequestSchema = import_v49.z.object({
10089
- method: import_v49.z.string(),
10090
- params: import_v49.z.optional(BaseParamsSchema)
10091
- });
10092
- var ServerCapabilitiesSchema = import_v49.z.looseObject({
10093
- experimental: import_v49.z.optional(import_v49.z.object({}).loose()),
10094
- logging: import_v49.z.optional(import_v49.z.object({}).loose()),
10095
- prompts: import_v49.z.optional(
10096
- import_v49.z.looseObject({
10097
- listChanged: import_v49.z.optional(import_v49.z.boolean())
10098
- })
10099
- ),
10100
- resources: import_v49.z.optional(
10101
- import_v49.z.looseObject({
10102
- subscribe: import_v49.z.optional(import_v49.z.boolean()),
10103
- listChanged: import_v49.z.optional(import_v49.z.boolean())
10104
- })
10105
- ),
10106
- tools: import_v49.z.optional(
10107
- import_v49.z.looseObject({
10108
- listChanged: import_v49.z.optional(import_v49.z.boolean())
10109
- })
10110
- )
10111
- });
10112
- var InitializeResultSchema = ResultSchema.extend({
10113
- protocolVersion: import_v49.z.string(),
10114
- capabilities: ServerCapabilitiesSchema,
10115
- serverInfo: ClientOrServerImplementationSchema,
10116
- instructions: import_v49.z.optional(import_v49.z.string())
10117
- });
10118
- var PaginatedResultSchema = ResultSchema.extend({
10119
- nextCursor: import_v49.z.optional(import_v49.z.string())
10120
- });
10121
- var ToolSchema = import_v49.z.object({
10122
- name: import_v49.z.string(),
10123
- description: import_v49.z.optional(import_v49.z.string()),
10124
- inputSchema: import_v49.z.object({
10125
- type: import_v49.z.literal("object"),
10126
- properties: import_v49.z.optional(import_v49.z.object({}).loose())
10127
- }).loose()
10128
- }).loose();
10129
- var ListToolsResultSchema = PaginatedResultSchema.extend({
10130
- tools: import_v49.z.array(ToolSchema)
10131
- });
10132
- var TextContentSchema = import_v49.z.object({
10133
- type: import_v49.z.literal("text"),
10134
- text: import_v49.z.string()
10135
- }).loose();
10136
- var ImageContentSchema = import_v49.z.object({
10137
- type: import_v49.z.literal("image"),
10138
- data: import_v49.z.base64(),
10139
- mimeType: import_v49.z.string()
10140
- }).loose();
10141
- var ResourceContentsSchema = import_v49.z.object({
10142
- /**
10143
- * The URI of this resource.
10144
- */
10145
- uri: import_v49.z.string(),
10146
- /**
10147
- * The MIME type of this resource, if known.
10148
- */
10149
- mimeType: import_v49.z.optional(import_v49.z.string())
10150
- }).loose();
10151
- var TextResourceContentsSchema = ResourceContentsSchema.extend({
10152
- text: import_v49.z.string()
10153
- });
10154
- var BlobResourceContentsSchema = ResourceContentsSchema.extend({
10155
- blob: import_v49.z.base64()
10156
- });
10157
- var EmbeddedResourceSchema = import_v49.z.object({
10158
- type: import_v49.z.literal("resource"),
10159
- resource: import_v49.z.union([TextResourceContentsSchema, BlobResourceContentsSchema])
10160
- }).loose();
10161
- var CallToolResultSchema = ResultSchema.extend({
10162
- content: import_v49.z.array(
10163
- import_v49.z.union([TextContentSchema, ImageContentSchema, EmbeddedResourceSchema])
10164
- ),
10165
- isError: import_v49.z.boolean().default(false).optional()
10166
- }).or(
10167
- ResultSchema.extend({
10168
- toolResult: import_v49.z.unknown()
10169
- })
10170
- );
10171
-
10172
- // src/tool/mcp/json-rpc-message.ts
10173
- var JSONRPC_VERSION = "2.0";
10174
- var JSONRPCRequestSchema = import_v410.z.object({
10175
- jsonrpc: import_v410.z.literal(JSONRPC_VERSION),
10176
- id: import_v410.z.union([import_v410.z.string(), import_v410.z.number().int()])
10177
- }).merge(RequestSchema).strict();
10178
- var JSONRPCResponseSchema = import_v410.z.object({
10179
- jsonrpc: import_v410.z.literal(JSONRPC_VERSION),
10180
- id: import_v410.z.union([import_v410.z.string(), import_v410.z.number().int()]),
10181
- result: ResultSchema
10182
- }).strict();
10183
- var JSONRPCErrorSchema = import_v410.z.object({
10184
- jsonrpc: import_v410.z.literal(JSONRPC_VERSION),
10185
- id: import_v410.z.union([import_v410.z.string(), import_v410.z.number().int()]),
10186
- error: import_v410.z.object({
10187
- code: import_v410.z.number().int(),
10188
- message: import_v410.z.string(),
10189
- data: import_v410.z.optional(import_v410.z.unknown())
10190
- })
10191
- }).strict();
10192
- var JSONRPCNotificationSchema = import_v410.z.object({
10193
- jsonrpc: import_v410.z.literal(JSONRPC_VERSION)
10194
- }).merge(
10195
- import_v410.z.object({
10196
- method: import_v410.z.string(),
10197
- params: import_v410.z.optional(BaseParamsSchema)
10198
- })
10199
- ).strict();
10200
- var JSONRPCMessageSchema = import_v410.z.union([
10201
- JSONRPCRequestSchema,
10202
- JSONRPCNotificationSchema,
10203
- JSONRPCResponseSchema,
10204
- JSONRPCErrorSchema
10205
- ]);
10206
-
10207
- // src/tool/mcp/mcp-sse-transport.ts
10208
- var SseMCPTransport = class {
10209
- constructor({
10210
- url,
10211
- headers
10212
- }) {
10213
- this.connected = false;
10214
- this.url = new URL(url);
10215
- this.headers = headers;
10216
- }
10217
- async start() {
10218
- return new Promise((resolve3, reject) => {
10219
- if (this.connected) {
10220
- return resolve3();
10221
- }
10222
- this.abortController = new AbortController();
10223
- const establishConnection = async () => {
10224
- var _a17, _b, _c;
10225
- try {
10226
- const headers = (0, import_provider_utils31.withUserAgentSuffix)(
10227
- {
10228
- ...this.headers,
10229
- Accept: "text/event-stream"
10230
- },
10231
- `ai-sdk/${VERSION}`,
10232
- (0, import_provider_utils31.getRuntimeEnvironmentUserAgent)()
10233
- );
10234
- const response = await fetch(this.url.href, {
10235
- headers,
10236
- signal: (_a17 = this.abortController) == null ? void 0 : _a17.signal
10237
- });
10238
- if (!response.ok || !response.body) {
10239
- const error = new MCPClientError({
10240
- message: `MCP SSE Transport Error: ${response.status} ${response.statusText}`
10241
- });
10242
- (_b = this.onerror) == null ? void 0 : _b.call(this, error);
10243
- return reject(error);
10244
- }
10245
- const stream = response.body.pipeThrough(new TextDecoderStream()).pipeThrough(new import_provider_utils31.EventSourceParserStream());
10246
- const reader = stream.getReader();
10247
- const processEvents = async () => {
10248
- var _a18, _b2, _c2;
10249
- try {
10250
- while (true) {
10251
- const { done, value } = await reader.read();
10252
- if (done) {
10253
- if (this.connected) {
10254
- this.connected = false;
10255
- throw new MCPClientError({
10256
- message: "MCP SSE Transport Error: Connection closed unexpectedly"
10257
- });
10258
- }
10259
- return;
10260
- }
10261
- const { event, data } = value;
10262
- if (event === "endpoint") {
10263
- this.endpoint = new URL(data, this.url);
10264
- if (this.endpoint.origin !== this.url.origin) {
10265
- throw new MCPClientError({
10266
- message: `MCP SSE Transport Error: Endpoint origin does not match connection origin: ${this.endpoint.origin}`
10267
- });
10268
- }
10269
- this.connected = true;
10270
- resolve3();
10271
- } else if (event === "message") {
10272
- try {
10273
- const message = JSONRPCMessageSchema.parse(
10274
- JSON.parse(data)
10275
- );
10276
- (_a18 = this.onmessage) == null ? void 0 : _a18.call(this, message);
10277
- } catch (error) {
10278
- const e = new MCPClientError({
10279
- message: "MCP SSE Transport Error: Failed to parse message",
10280
- cause: error
10281
- });
10282
- (_b2 = this.onerror) == null ? void 0 : _b2.call(this, e);
10283
- }
10284
- }
10285
- }
10286
- } catch (error) {
10287
- if (error instanceof Error && error.name === "AbortError") {
10288
- return;
10289
- }
10290
- (_c2 = this.onerror) == null ? void 0 : _c2.call(this, error);
10291
- reject(error);
10292
- }
10293
- };
10294
- this.sseConnection = {
10295
- close: () => reader.cancel()
10296
- };
10297
- processEvents();
10298
- } catch (error) {
10299
- if (error instanceof Error && error.name === "AbortError") {
10300
- return;
10301
- }
10302
- (_c = this.onerror) == null ? void 0 : _c.call(this, error);
10303
- reject(error);
10304
- }
10305
- };
10306
- establishConnection();
10307
- });
10308
- }
10309
- async close() {
10310
- var _a17, _b, _c;
10311
- this.connected = false;
10312
- (_a17 = this.sseConnection) == null ? void 0 : _a17.close();
10313
- (_b = this.abortController) == null ? void 0 : _b.abort();
10314
- (_c = this.onclose) == null ? void 0 : _c.call(this);
10315
- }
10316
- async send(message) {
10317
- var _a17, _b, _c;
10318
- if (!this.endpoint || !this.connected) {
10319
- throw new MCPClientError({
10320
- message: "MCP SSE Transport Error: Not connected"
10321
- });
10322
- }
10323
- try {
10324
- const headers = (0, import_provider_utils31.withUserAgentSuffix)(
10325
- {
10326
- ...this.headers,
10327
- "Content-Type": "application/json"
10328
- },
10329
- `ai-sdk/${VERSION}`,
10330
- (0, import_provider_utils31.getRuntimeEnvironmentUserAgent)()
10331
- );
10332
- const init = {
10333
- method: "POST",
10334
- headers,
10335
- body: JSON.stringify(message),
10336
- signal: (_a17 = this.abortController) == null ? void 0 : _a17.signal
10337
- };
10338
- const response = await fetch(this.endpoint, init);
10339
- if (!response.ok) {
10340
- const text2 = await response.text().catch(() => null);
10341
- const error = new MCPClientError({
10342
- message: `MCP SSE Transport Error: POSTing to endpoint (HTTP ${response.status}): ${text2}`
10343
- });
10344
- (_b = this.onerror) == null ? void 0 : _b.call(this, error);
10345
- return;
10346
- }
10347
- } catch (error) {
10348
- (_c = this.onerror) == null ? void 0 : _c.call(this, error);
10349
- return;
10350
- }
10351
- }
10352
- };
10353
-
10354
- // src/tool/mcp/mcp-transport.ts
10355
- function createMcpTransport(config) {
10356
- if (config.type !== "sse") {
10357
- throw new MCPClientError({
10358
- message: "Unsupported or invalid transport configuration. If you are using a custom transport, make sure it implements the MCPTransport interface."
10359
- });
10360
- }
10361
- return new SseMCPTransport(config);
10362
- }
10363
- function isCustomMcpTransport(transport) {
10364
- return "start" in transport && typeof transport.start === "function" && "send" in transport && typeof transport.send === "function" && "close" in transport && typeof transport.close === "function";
10365
- }
10366
-
10367
- // src/tool/mcp/mcp-client.ts
10368
- var CLIENT_VERSION = "1.0.0";
10369
- async function createMCPClient(config) {
10370
- const client = new DefaultMCPClient(config);
10371
- await client.init();
10372
- return client;
10373
- }
10374
- var DefaultMCPClient = class {
10375
- constructor({
10376
- transport: transportConfig,
10377
- name: name17 = "ai-sdk-mcp-client",
10378
- onUncaughtError
10379
- }) {
10380
- this.requestMessageId = 0;
10381
- this.responseHandlers = /* @__PURE__ */ new Map();
10382
- this.serverCapabilities = {};
10383
- this.isClosed = true;
10384
- this.onUncaughtError = onUncaughtError;
10385
- if (isCustomMcpTransport(transportConfig)) {
10386
- this.transport = transportConfig;
10387
- } else {
10388
- this.transport = createMcpTransport(transportConfig);
10389
- }
10390
- this.transport.onclose = () => this.onClose();
10391
- this.transport.onerror = (error) => this.onError(error);
10392
- this.transport.onmessage = (message) => {
10393
- if ("method" in message) {
10394
- this.onError(
10395
- new MCPClientError({
10396
- message: "Unsupported message type"
10397
- })
10398
- );
10399
- return;
10400
- }
10401
- this.onResponse(message);
10402
- };
10403
- this.clientInfo = {
10404
- name: name17,
10405
- version: CLIENT_VERSION
10406
- };
10407
- }
10408
- async init() {
10409
- try {
10410
- await this.transport.start();
10411
- this.isClosed = false;
10412
- const result = await this.request({
10413
- request: {
10414
- method: "initialize",
10415
- params: {
10416
- protocolVersion: LATEST_PROTOCOL_VERSION,
10417
- capabilities: {},
10418
- clientInfo: this.clientInfo
10419
- }
10420
- },
10421
- resultSchema: InitializeResultSchema
10422
- });
10423
- if (result === void 0) {
10424
- throw new MCPClientError({
10425
- message: "Server sent invalid initialize result"
10426
- });
10427
- }
10428
- if (!SUPPORTED_PROTOCOL_VERSIONS.includes(result.protocolVersion)) {
10429
- throw new MCPClientError({
10430
- message: `Server's protocol version is not supported: ${result.protocolVersion}`
10431
- });
10432
- }
10433
- this.serverCapabilities = result.capabilities;
10434
- await this.notification({
10435
- method: "notifications/initialized"
10436
- });
10437
- return this;
10438
- } catch (error) {
10439
- await this.close();
10440
- throw error;
10441
- }
10442
- }
10443
- async close() {
10444
- var _a17;
10445
- if (this.isClosed)
10446
- return;
10447
- await ((_a17 = this.transport) == null ? void 0 : _a17.close());
10448
- this.onClose();
10449
- }
10450
- assertCapability(method) {
10451
- switch (method) {
10452
- case "initialize":
10453
- break;
10454
- case "tools/list":
10455
- case "tools/call":
10456
- if (!this.serverCapabilities.tools) {
10457
- throw new MCPClientError({
10458
- message: `Server does not support tools`
10459
- });
10460
- }
10461
- break;
10462
- default:
10463
- throw new MCPClientError({
10464
- message: `Unsupported method: ${method}`
10465
- });
10466
- }
10467
- }
10468
- async request({
10469
- request,
10470
- resultSchema,
10471
- options
10472
- }) {
10473
- return new Promise((resolve3, reject) => {
10474
- if (this.isClosed) {
10475
- return reject(
10476
- new MCPClientError({
10477
- message: "Attempted to send a request from a closed client"
10478
- })
10479
- );
10480
- }
10481
- this.assertCapability(request.method);
10482
- const signal = options == null ? void 0 : options.signal;
10483
- signal == null ? void 0 : signal.throwIfAborted();
10484
- const messageId = this.requestMessageId++;
10485
- const jsonrpcRequest = {
10486
- ...request,
10487
- jsonrpc: "2.0",
10488
- id: messageId
10489
- };
10490
- const cleanup = () => {
10491
- this.responseHandlers.delete(messageId);
10492
- };
10493
- this.responseHandlers.set(messageId, (response) => {
10494
- if (signal == null ? void 0 : signal.aborted) {
10495
- return reject(
10496
- new MCPClientError({
10497
- message: "Request was aborted",
10498
- cause: signal.reason
10499
- })
10500
- );
10501
- }
10502
- if (response instanceof Error) {
10503
- return reject(response);
10504
- }
10505
- try {
10506
- const result = resultSchema.parse(response.result);
10507
- resolve3(result);
10508
- } catch (error) {
10509
- const parseError = new MCPClientError({
10510
- message: "Failed to parse server response",
10511
- cause: error
10512
- });
10513
- reject(parseError);
10514
- }
10515
- });
10516
- this.transport.send(jsonrpcRequest).catch((error) => {
10517
- cleanup();
10518
- reject(error);
10519
- });
10520
- });
10521
- }
10522
- async listTools({
10523
- params,
10524
- options
10525
- } = {}) {
10526
- try {
10527
- return this.request({
10528
- request: { method: "tools/list", params },
10529
- resultSchema: ListToolsResultSchema,
10530
- options
10531
- });
10532
- } catch (error) {
10533
- throw error;
10534
- }
10535
- }
10536
- async callTool({
10537
- name: name17,
10538
- args,
10539
- options
10540
- }) {
10541
- try {
10542
- return this.request({
10543
- request: { method: "tools/call", params: { name: name17, arguments: args } },
10544
- resultSchema: CallToolResultSchema,
10545
- options: {
10546
- signal: options == null ? void 0 : options.abortSignal
10547
- }
10548
- });
10549
- } catch (error) {
10550
- throw error;
10551
- }
10552
- }
10553
- async notification(notification) {
10554
- const jsonrpcNotification = {
10555
- ...notification,
10556
- jsonrpc: "2.0"
10557
- };
10558
- await this.transport.send(jsonrpcNotification);
10559
- }
10560
- /**
10561
- * Returns a set of AI SDK tools from the MCP server
10562
- * @returns A record of tool names to their implementations
10563
- */
10564
- async tools({
10565
- schemas = "automatic"
10566
- } = {}) {
10567
- var _a17;
10568
- const tools = {};
10569
- try {
10570
- const listToolsResult = await this.listTools();
10571
- for (const { name: name17, description, inputSchema } of listToolsResult.tools) {
10572
- if (schemas !== "automatic" && !(name17 in schemas)) {
10573
- continue;
10574
- }
10575
- const self = this;
10576
- const execute = async (args, options) => {
10577
- var _a18;
10578
- (_a18 = options == null ? void 0 : options.abortSignal) == null ? void 0 : _a18.throwIfAborted();
10579
- return self.callTool({ name: name17, args, options });
10580
- };
10581
- const toolWithExecute = schemas === "automatic" ? (0, import_provider_utils32.dynamicTool)({
10582
- description,
10583
- inputSchema: (0, import_provider_utils32.jsonSchema)({
10584
- ...inputSchema,
10585
- properties: (_a17 = inputSchema.properties) != null ? _a17 : {},
10586
- additionalProperties: false
10587
- }),
10588
- execute
10589
- }) : (0, import_provider_utils32.tool)({
10590
- description,
10591
- inputSchema: schemas[name17].inputSchema,
10592
- execute
10593
- });
10594
- tools[name17] = toolWithExecute;
10595
- }
10596
- return tools;
10597
- } catch (error) {
10598
- throw error;
10599
- }
10600
- }
10601
- onClose() {
10602
- if (this.isClosed)
10603
- return;
10604
- this.isClosed = true;
10605
- const error = new MCPClientError({
10606
- message: "Connection closed"
10607
- });
10608
- for (const handler of this.responseHandlers.values()) {
10609
- handler(error);
10610
- }
10611
- this.responseHandlers.clear();
10612
- }
10613
- onError(error) {
10614
- if (this.onUncaughtError) {
10615
- this.onUncaughtError(error);
10616
- }
10617
- }
10618
- onResponse(response) {
10619
- const messageId = Number(response.id);
10620
- const handler = this.responseHandlers.get(messageId);
10621
- if (handler === void 0) {
10622
- throw new MCPClientError({
10623
- message: `Protocol error: Received a response for an unknown message ID: ${JSON.stringify(
10624
- response
10625
- )}`
10626
- });
10627
- }
10628
- this.responseHandlers.delete(messageId);
10629
- handler(
10630
- "result" in response ? response : new MCPClientError({
10631
- message: response.error.message,
10632
- code: response.error.code,
10633
- data: response.error.data,
10634
- cause: response.error
10635
- })
10636
- );
10637
- }
10638
- };
10639
-
10640
10044
  // src/transcribe/transcribe.ts
10641
- var import_provider_utils33 = require("@ai-sdk/provider-utils");
10045
+ var import_provider_utils31 = require("@ai-sdk/provider-utils");
10642
10046
 
10643
10047
  // src/error/no-transcript-generated-error.ts
10644
- var import_provider33 = require("@ai-sdk/provider");
10645
- var NoTranscriptGeneratedError = class extends import_provider33.AISDKError {
10048
+ var import_provider32 = require("@ai-sdk/provider");
10049
+ var NoTranscriptGeneratedError = class extends import_provider32.AISDKError {
10646
10050
  constructor(options) {
10647
10051
  super({
10648
10052
  name: "AI_NoTranscriptGeneratedError",
@@ -10669,23 +10073,23 @@ async function transcribe({
10669
10073
  maxRetries: maxRetriesArg,
10670
10074
  abortSignal
10671
10075
  });
10672
- const headersWithUserAgent = (0, import_provider_utils33.withUserAgentSuffix)(
10076
+ const headersWithUserAgent = (0, import_provider_utils31.withUserAgentSuffix)(
10673
10077
  headers != null ? headers : {},
10674
10078
  `ai/${VERSION}`
10675
10079
  );
10676
10080
  const audioData = audio instanceof URL ? (await download({ url: audio })).data : convertDataContentToUint8Array(audio);
10677
10081
  const result = await retry(
10678
10082
  () => {
10679
- var _a17;
10083
+ var _a16;
10680
10084
  return resolvedModel.doGenerate({
10681
10085
  audio: audioData,
10682
10086
  abortSignal,
10683
10087
  headers: headersWithUserAgent,
10684
10088
  providerOptions,
10685
- mediaType: (_a17 = detectMediaType({
10089
+ mediaType: (_a16 = detectMediaType({
10686
10090
  data: audioData,
10687
10091
  signatures: audioMediaTypeSignatures
10688
- })) != null ? _a17 : "audio/wav"
10092
+ })) != null ? _a16 : "audio/wav"
10689
10093
  });
10690
10094
  }
10691
10095
  );
@@ -10705,19 +10109,19 @@ async function transcribe({
10705
10109
  }
10706
10110
  var DefaultTranscriptionResult = class {
10707
10111
  constructor(options) {
10708
- var _a17;
10112
+ var _a16;
10709
10113
  this.text = options.text;
10710
10114
  this.segments = options.segments;
10711
10115
  this.language = options.language;
10712
10116
  this.durationInSeconds = options.durationInSeconds;
10713
10117
  this.warnings = options.warnings;
10714
10118
  this.responses = options.responses;
10715
- this.providerMetadata = (_a17 = options.providerMetadata) != null ? _a17 : {};
10119
+ this.providerMetadata = (_a16 = options.providerMetadata) != null ? _a16 : {};
10716
10120
  }
10717
10121
  };
10718
10122
 
10719
10123
  // src/ui/call-completion-api.ts
10720
- var import_provider_utils34 = require("@ai-sdk/provider-utils");
10124
+ var import_provider_utils32 = require("@ai-sdk/provider-utils");
10721
10125
 
10722
10126
  // src/ui/process-text-stream.ts
10723
10127
  async function processTextStream({
@@ -10751,7 +10155,7 @@ async function callCompletionApi({
10751
10155
  onError,
10752
10156
  fetch: fetch2 = getOriginalFetch()
10753
10157
  }) {
10754
- var _a17;
10158
+ var _a16;
10755
10159
  try {
10756
10160
  setLoading(true);
10757
10161
  setError(void 0);
@@ -10765,13 +10169,13 @@ async function callCompletionApi({
10765
10169
  ...body
10766
10170
  }),
10767
10171
  credentials,
10768
- headers: (0, import_provider_utils34.withUserAgentSuffix)(
10172
+ headers: (0, import_provider_utils32.withUserAgentSuffix)(
10769
10173
  {
10770
10174
  "Content-Type": "application/json",
10771
10175
  ...headers
10772
10176
  },
10773
10177
  `ai-sdk/${VERSION}`,
10774
- (0, import_provider_utils34.getRuntimeEnvironmentUserAgent)()
10178
+ (0, import_provider_utils32.getRuntimeEnvironmentUserAgent)()
10775
10179
  ),
10776
10180
  signal: abortController.signal
10777
10181
  }).catch((err) => {
@@ -10779,7 +10183,7 @@ async function callCompletionApi({
10779
10183
  });
10780
10184
  if (!response.ok) {
10781
10185
  throw new Error(
10782
- (_a17 = await response.text()) != null ? _a17 : "Failed to fetch the chat response."
10186
+ (_a16 = await response.text()) != null ? _a16 : "Failed to fetch the chat response."
10783
10187
  );
10784
10188
  }
10785
10189
  if (!response.body) {
@@ -10799,7 +10203,7 @@ async function callCompletionApi({
10799
10203
  }
10800
10204
  case "data": {
10801
10205
  await consumeStream({
10802
- stream: (0, import_provider_utils34.parseJsonEventStream)({
10206
+ stream: (0, import_provider_utils32.parseJsonEventStream)({
10803
10207
  stream: response.body,
10804
10208
  schema: uiMessageChunkSchema
10805
10209
  }).pipeThrough(
@@ -10851,7 +10255,7 @@ async function callCompletionApi({
10851
10255
  }
10852
10256
 
10853
10257
  // src/ui/chat.ts
10854
- var import_provider_utils37 = require("@ai-sdk/provider-utils");
10258
+ var import_provider_utils35 = require("@ai-sdk/provider-utils");
10855
10259
 
10856
10260
  // src/ui/convert-file-list-to-file-ui-parts.ts
10857
10261
  async function convertFileListToFileUIParts(files) {
@@ -10863,12 +10267,12 @@ async function convertFileListToFileUIParts(files) {
10863
10267
  }
10864
10268
  return Promise.all(
10865
10269
  Array.from(files).map(async (file) => {
10866
- const { name: name17, type } = file;
10270
+ const { name: name16, type } = file;
10867
10271
  const dataUrl = await new Promise((resolve3, reject) => {
10868
10272
  const reader = new FileReader();
10869
10273
  reader.onload = (readerEvent) => {
10870
- var _a17;
10871
- resolve3((_a17 = readerEvent.target) == null ? void 0 : _a17.result);
10274
+ var _a16;
10275
+ resolve3((_a16 = readerEvent.target) == null ? void 0 : _a16.result);
10872
10276
  };
10873
10277
  reader.onerror = (error) => reject(error);
10874
10278
  reader.readAsDataURL(file);
@@ -10876,7 +10280,7 @@ async function convertFileListToFileUIParts(files) {
10876
10280
  return {
10877
10281
  type: "file",
10878
10282
  mediaType: type,
10879
- filename: name17,
10283
+ filename: name16,
10880
10284
  url: dataUrl
10881
10285
  };
10882
10286
  })
@@ -10884,10 +10288,10 @@ async function convertFileListToFileUIParts(files) {
10884
10288
  }
10885
10289
 
10886
10290
  // src/ui/default-chat-transport.ts
10887
- var import_provider_utils36 = require("@ai-sdk/provider-utils");
10291
+ var import_provider_utils34 = require("@ai-sdk/provider-utils");
10888
10292
 
10889
10293
  // src/ui/http-chat-transport.ts
10890
- var import_provider_utils35 = require("@ai-sdk/provider-utils");
10294
+ var import_provider_utils33 = require("@ai-sdk/provider-utils");
10891
10295
  var HttpChatTransport = class {
10892
10296
  constructor({
10893
10297
  api = "/api/chat",
@@ -10910,11 +10314,11 @@ var HttpChatTransport = class {
10910
10314
  abortSignal,
10911
10315
  ...options
10912
10316
  }) {
10913
- var _a17, _b, _c, _d, _e;
10914
- const resolvedBody = await (0, import_provider_utils35.resolve)(this.body);
10915
- const resolvedHeaders = await (0, import_provider_utils35.resolve)(this.headers);
10916
- const resolvedCredentials = await (0, import_provider_utils35.resolve)(this.credentials);
10917
- const preparedRequest = await ((_a17 = this.prepareSendMessagesRequest) == null ? void 0 : _a17.call(this, {
10317
+ var _a16, _b, _c, _d, _e;
10318
+ const resolvedBody = await (0, import_provider_utils33.resolve)(this.body);
10319
+ const resolvedHeaders = await (0, import_provider_utils33.resolve)(this.headers);
10320
+ const resolvedCredentials = await (0, import_provider_utils33.resolve)(this.credentials);
10321
+ const preparedRequest = await ((_a16 = this.prepareSendMessagesRequest) == null ? void 0 : _a16.call(this, {
10918
10322
  api: this.api,
10919
10323
  id: options.chatId,
10920
10324
  messages: options.messages,
@@ -10939,13 +10343,13 @@ var HttpChatTransport = class {
10939
10343
  const fetch2 = (_d = this.fetch) != null ? _d : globalThis.fetch;
10940
10344
  const response = await fetch2(api, {
10941
10345
  method: "POST",
10942
- headers: (0, import_provider_utils35.withUserAgentSuffix)(
10346
+ headers: (0, import_provider_utils33.withUserAgentSuffix)(
10943
10347
  {
10944
10348
  "Content-Type": "application/json",
10945
10349
  ...headers
10946
10350
  },
10947
10351
  `ai-sdk/${VERSION}`,
10948
- (0, import_provider_utils35.getRuntimeEnvironmentUserAgent)()
10352
+ (0, import_provider_utils33.getRuntimeEnvironmentUserAgent)()
10949
10353
  ),
10950
10354
  body: JSON.stringify(body),
10951
10355
  credentials,
@@ -10962,11 +10366,11 @@ var HttpChatTransport = class {
10962
10366
  return this.processResponseStream(response.body);
10963
10367
  }
10964
10368
  async reconnectToStream(options) {
10965
- var _a17, _b, _c, _d, _e;
10966
- const resolvedBody = await (0, import_provider_utils35.resolve)(this.body);
10967
- const resolvedHeaders = await (0, import_provider_utils35.resolve)(this.headers);
10968
- const resolvedCredentials = await (0, import_provider_utils35.resolve)(this.credentials);
10969
- const preparedRequest = await ((_a17 = this.prepareReconnectToStreamRequest) == null ? void 0 : _a17.call(this, {
10369
+ var _a16, _b, _c, _d, _e;
10370
+ const resolvedBody = await (0, import_provider_utils33.resolve)(this.body);
10371
+ const resolvedHeaders = await (0, import_provider_utils33.resolve)(this.headers);
10372
+ const resolvedCredentials = await (0, import_provider_utils33.resolve)(this.credentials);
10373
+ const preparedRequest = await ((_a16 = this.prepareReconnectToStreamRequest) == null ? void 0 : _a16.call(this, {
10970
10374
  api: this.api,
10971
10375
  id: options.chatId,
10972
10376
  body: { ...resolvedBody, ...options.body },
@@ -10980,10 +10384,10 @@ var HttpChatTransport = class {
10980
10384
  const fetch2 = (_d = this.fetch) != null ? _d : globalThis.fetch;
10981
10385
  const response = await fetch2(api, {
10982
10386
  method: "GET",
10983
- headers: (0, import_provider_utils35.withUserAgentSuffix)(
10387
+ headers: (0, import_provider_utils33.withUserAgentSuffix)(
10984
10388
  headers,
10985
10389
  `ai-sdk/${VERSION}`,
10986
- (0, import_provider_utils35.getRuntimeEnvironmentUserAgent)()
10390
+ (0, import_provider_utils33.getRuntimeEnvironmentUserAgent)()
10987
10391
  ),
10988
10392
  credentials
10989
10393
  });
@@ -11008,7 +10412,7 @@ var DefaultChatTransport = class extends HttpChatTransport {
11008
10412
  super(options);
11009
10413
  }
11010
10414
  processResponseStream(stream) {
11011
- return (0, import_provider_utils36.parseJsonEventStream)({
10415
+ return (0, import_provider_utils34.parseJsonEventStream)({
11012
10416
  stream,
11013
10417
  schema: uiMessageChunkSchema
11014
10418
  }).pipeThrough(
@@ -11027,7 +10431,7 @@ var DefaultChatTransport = class extends HttpChatTransport {
11027
10431
  // src/ui/chat.ts
11028
10432
  var AbstractChat = class {
11029
10433
  constructor({
11030
- generateId: generateId2 = import_provider_utils37.generateId,
10434
+ generateId: generateId2 = import_provider_utils35.generateId,
11031
10435
  id = generateId2(),
11032
10436
  transport = new DefaultChatTransport(),
11033
10437
  messageMetadataSchema,
@@ -11048,11 +10452,11 @@ var AbstractChat = class {
11048
10452
  * If a messageId is provided, the message will be replaced.
11049
10453
  */
11050
10454
  this.sendMessage = async (message, options) => {
11051
- var _a17, _b, _c, _d;
10455
+ var _a16, _b, _c, _d;
11052
10456
  if (message == null) {
11053
10457
  await this.makeRequest({
11054
10458
  trigger: "submit-message",
11055
- messageId: (_a17 = this.lastMessage) == null ? void 0 : _a17.id,
10459
+ messageId: (_a16 = this.lastMessage) == null ? void 0 : _a16.id,
11056
10460
  ...options
11057
10461
  });
11058
10462
  return;
@@ -11145,7 +10549,7 @@ var AbstractChat = class {
11145
10549
  approved,
11146
10550
  reason
11147
10551
  }) => this.jobExecutor.run(async () => {
11148
- var _a17, _b;
10552
+ var _a16, _b;
11149
10553
  const messages = this.state.messages;
11150
10554
  const lastMessage = messages[messages.length - 1];
11151
10555
  const updatePart = (part) => isToolOrDynamicToolUIPart(part) && part.state === "approval-requested" && part.approval.id === id ? {
@@ -11160,7 +10564,7 @@ var AbstractChat = class {
11160
10564
  if (this.activeResponse) {
11161
10565
  this.activeResponse.state.message.parts = this.activeResponse.state.message.parts.map(updatePart);
11162
10566
  }
11163
- if (this.status !== "streaming" && this.status !== "submitted" && ((_a17 = this.sendAutomaticallyWhen) == null ? void 0 : _a17.call(this, { messages: this.state.messages }))) {
10567
+ if (this.status !== "streaming" && this.status !== "submitted" && ((_a16 = this.sendAutomaticallyWhen) == null ? void 0 : _a16.call(this, { messages: this.state.messages }))) {
11164
10568
  this.makeRequest({
11165
10569
  trigger: "submit-message",
11166
10570
  messageId: (_b = this.lastMessage) == null ? void 0 : _b.id
@@ -11169,12 +10573,12 @@ var AbstractChat = class {
11169
10573
  });
11170
10574
  this.addToolResult = async ({
11171
10575
  state = "output-available",
11172
- tool: tool3,
10576
+ tool: tool2,
11173
10577
  toolCallId,
11174
10578
  output,
11175
10579
  errorText
11176
10580
  }) => this.jobExecutor.run(async () => {
11177
- var _a17, _b;
10581
+ var _a16, _b;
11178
10582
  const messages = this.state.messages;
11179
10583
  const lastMessage = messages[messages.length - 1];
11180
10584
  const updatePart = (part) => isToolOrDynamicToolUIPart(part) && part.toolCallId === toolCallId ? { ...part, state, output, errorText } : part;
@@ -11185,7 +10589,7 @@ var AbstractChat = class {
11185
10589
  if (this.activeResponse) {
11186
10590
  this.activeResponse.state.message.parts = this.activeResponse.state.message.parts.map(updatePart);
11187
10591
  }
11188
- if (this.status !== "streaming" && this.status !== "submitted" && ((_a17 = this.sendAutomaticallyWhen) == null ? void 0 : _a17.call(this, { messages: this.state.messages }))) {
10592
+ if (this.status !== "streaming" && this.status !== "submitted" && ((_a16 = this.sendAutomaticallyWhen) == null ? void 0 : _a16.call(this, { messages: this.state.messages }))) {
11189
10593
  this.makeRequest({
11190
10594
  trigger: "submit-message",
11191
10595
  messageId: (_b = this.lastMessage) == null ? void 0 : _b.id
@@ -11196,10 +10600,10 @@ var AbstractChat = class {
11196
10600
  * Abort the current request immediately, keep the generated tokens if any.
11197
10601
  */
11198
10602
  this.stop = async () => {
11199
- var _a17;
10603
+ var _a16;
11200
10604
  if (this.status !== "streaming" && this.status !== "submitted")
11201
10605
  return;
11202
- if ((_a17 = this.activeResponse) == null ? void 0 : _a17.abortController) {
10606
+ if ((_a16 = this.activeResponse) == null ? void 0 : _a16.abortController) {
11203
10607
  this.activeResponse.abortController.abort();
11204
10608
  }
11205
10609
  };
@@ -11254,7 +10658,7 @@ var AbstractChat = class {
11254
10658
  body,
11255
10659
  messageId
11256
10660
  }) {
11257
- var _a17, _b, _c;
10661
+ var _a16, _b, _c;
11258
10662
  this.setStatus({ status: "submitted", error: void 0 });
11259
10663
  const lastMessage = this.lastMessage;
11260
10664
  let isAbort = false;
@@ -11303,9 +10707,9 @@ var AbstractChat = class {
11303
10707
  () => job({
11304
10708
  state: activeResponse.state,
11305
10709
  write: () => {
11306
- var _a18;
10710
+ var _a17;
11307
10711
  this.setStatus({ status: "streaming" });
11308
- const replaceLastMessage = activeResponse.state.message.id === ((_a18 = this.lastMessage) == null ? void 0 : _a18.id);
10712
+ const replaceLastMessage = activeResponse.state.message.id === ((_a17 = this.lastMessage) == null ? void 0 : _a17.id);
11309
10713
  if (replaceLastMessage) {
11310
10714
  this.state.replaceMessage(
11311
10715
  this.state.messages.length - 1,
@@ -11351,7 +10755,7 @@ var AbstractChat = class {
11351
10755
  this.setStatus({ status: "error", error: err });
11352
10756
  } finally {
11353
10757
  try {
11354
- (_a17 = this.onFinish) == null ? void 0 : _a17.call(this, {
10758
+ (_a16 = this.onFinish) == null ? void 0 : _a16.call(this, {
11355
10759
  message: this.activeResponse.state.message,
11356
10760
  messages: this.state.messages,
11357
10761
  isAbort,
@@ -11474,7 +10878,6 @@ var TextStreamChatTransport = class extends HttpChatTransport {
11474
10878
  JsonToSseTransformStream,
11475
10879
  LoadAPIKeyError,
11476
10880
  LoadSettingError,
11477
- MCPClientError,
11478
10881
  MessageConversionError,
11479
10882
  NoContentGeneratedError,
11480
10883
  NoImageGeneratedError,
@@ -11522,7 +10925,6 @@ var TextStreamChatTransport = class extends HttpChatTransport {
11522
10925
  dynamicTool,
11523
10926
  embed,
11524
10927
  embedMany,
11525
- experimental_createMCPClient,
11526
10928
  experimental_createProviderRegistry,
11527
10929
  experimental_customProvider,
11528
10930
  experimental_generateImage,