ai 3.0.33 → 3.0.35

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.d.mts CHANGED
@@ -1,5 +1,6 @@
1
- import { LanguageModelV1, LanguageModelV1FinishReason, LanguageModelV1CallWarning, LanguageModelV1LogProbs } from '@ai-sdk/provider';
2
1
  import { z } from 'zod';
2
+ import { LanguageModelV1, LanguageModelV1FinishReason, LanguageModelV1LogProbs, LanguageModelV1CallWarning } from '@ai-sdk/provider';
3
+ export { APICallError, EmptyResponseBodyError, InvalidArgumentError, InvalidDataContentError, InvalidPromptError, InvalidResponseDataError, InvalidToolArgumentsError, JSONParseError, LoadAPIKeyError, NoObjectGeneratedError, NoSuchToolError, RetryError, ToolCallParseError, TypeValidationError, UnsupportedFunctionalityError, UnsupportedJSONSchemaError } from '@ai-sdk/provider';
3
4
  import { ServerResponse } from 'node:http';
4
5
  import { AssistantStream } from 'openai/lib/AssistantStream';
5
6
  import { Run } from 'openai/resources/beta/threads/runs/runs';
@@ -214,6 +215,32 @@ type Prompt = {
214
215
  messages?: Array<ExperimentalMessage>;
215
216
  };
216
217
 
218
+ /**
219
+ Language model that is used by the AI SDK Core functions.
220
+ */
221
+ type LanguageModel = LanguageModelV1;
222
+ /**
223
+ Reason why a language model finished generating a response.
224
+
225
+ Can be one of the following:
226
+ - `stop`: model generated stop sequence
227
+ - `length`: model generated maximum number of tokens
228
+ - `content-filter`: content filter violation stopped the model
229
+ - `tool-calls`: model triggered tool calls
230
+ - `error`: model stopped because of an error
231
+ - `other`: model stopped for other reasons
232
+ */
233
+ type FinishReason = LanguageModelV1FinishReason;
234
+ /**
235
+ Log probabilities for each token and its top log probabilities.
236
+ */
237
+ type LogProbs = LanguageModelV1LogProbs;
238
+ /**
239
+ Warning from the model provider for this call. The call will proceed, but e.g.
240
+ some settings might not be supported, which can lead to suboptimal results.
241
+ */
242
+ type CallWarning = LanguageModelV1CallWarning;
243
+
217
244
  /**
218
245
  Generate a structured, typed object for a given prompt and schema using a language model.
219
246
 
@@ -254,7 +281,7 @@ declare function experimental_generateObject<T>({ model, schema, mode, system, p
254
281
  /**
255
282
  The language model to use.
256
283
  */
257
- model: LanguageModelV1;
284
+ model: LanguageModel;
258
285
  /**
259
286
  The schema of the object that the model should generate.
260
287
  */
@@ -277,7 +304,7 @@ declare class GenerateObjectResult<T> {
277
304
  /**
278
305
  The reason why the generation finished.
279
306
  */
280
- readonly finishReason: LanguageModelV1FinishReason;
307
+ readonly finishReason: FinishReason;
281
308
  /**
282
309
  The token usage of the generated text.
283
310
  */
@@ -285,7 +312,7 @@ declare class GenerateObjectResult<T> {
285
312
  /**
286
313
  Warnings from the model provider (e.g. unsupported settings)
287
314
  */
288
- readonly warnings: LanguageModelV1CallWarning[] | undefined;
315
+ readonly warnings: CallWarning[] | undefined;
289
316
  /**
290
317
  Optional raw response data.
291
318
  */
@@ -299,16 +326,16 @@ declare class GenerateObjectResult<T> {
299
326
  Logprobs for the completion.
300
327
  `undefined` if the mode does not support logprobs or if was not enabled
301
328
  */
302
- readonly logprobs: LanguageModelV1LogProbs | undefined;
329
+ readonly logprobs: LogProbs | undefined;
303
330
  constructor(options: {
304
331
  object: T;
305
- finishReason: LanguageModelV1FinishReason;
332
+ finishReason: FinishReason;
306
333
  usage: TokenUsage;
307
- warnings: LanguageModelV1CallWarning[] | undefined;
334
+ warnings: CallWarning[] | undefined;
308
335
  rawResponse?: {
309
336
  headers?: Record<string, string>;
310
337
  };
311
- logprobs: LanguageModelV1LogProbs | undefined;
338
+ logprobs: LogProbs | undefined;
312
339
  });
313
340
  }
314
341
 
@@ -370,7 +397,7 @@ declare function experimental_streamObject<T>({ model, schema, mode, system, pro
370
397
  /**
371
398
  The language model to use.
372
399
  */
373
- model: LanguageModelV1;
400
+ model: LanguageModel;
374
401
  /**
375
402
  The schema of the object that the model should generate.
376
403
  */
@@ -387,8 +414,8 @@ type ObjectStreamPartInput = {
387
414
  error: unknown;
388
415
  } | {
389
416
  type: 'finish';
390
- finishReason: LanguageModelV1FinishReason;
391
- logprobs?: LanguageModelV1LogProbs;
417
+ finishReason: FinishReason;
418
+ logprobs?: LogProbs;
392
419
  usage: {
393
420
  promptTokens: number;
394
421
  completionTokens: number;
@@ -407,7 +434,7 @@ declare class StreamObjectResult<T> {
407
434
  /**
408
435
  Warnings from the model provider (e.g. unsupported settings)
409
436
  */
410
- readonly warnings: LanguageModelV1CallWarning[] | undefined;
437
+ readonly warnings: CallWarning[] | undefined;
411
438
  /**
412
439
  Optional raw response data.
413
440
  */
@@ -419,7 +446,7 @@ declare class StreamObjectResult<T> {
419
446
  };
420
447
  constructor({ stream, warnings, rawResponse, }: {
421
448
  stream: ReadableStream<string | ObjectStreamPartInput>;
422
- warnings: LanguageModelV1CallWarning[] | undefined;
449
+ warnings: CallWarning[] | undefined;
423
450
  rawResponse?: {
424
451
  headers?: Record<string, string>;
425
452
  };
@@ -574,7 +601,7 @@ declare function experimental_generateText<TOOLS extends Record<string, Experime
574
601
  /**
575
602
  The language model to use.
576
603
  */
577
- model: LanguageModelV1;
604
+ model: LanguageModel;
578
605
  /**
579
606
  The tools that the model can call. The model needs to support calling tools.
580
607
  */
@@ -600,7 +627,7 @@ declare class GenerateTextResult<TOOLS extends Record<string, ExperimentalTool>>
600
627
  /**
601
628
  The reason why the generation finished.
602
629
  */
603
- readonly finishReason: LanguageModelV1FinishReason;
630
+ readonly finishReason: FinishReason;
604
631
  /**
605
632
  The token usage of the generated text.
606
633
  */
@@ -608,7 +635,7 @@ declare class GenerateTextResult<TOOLS extends Record<string, ExperimentalTool>>
608
635
  /**
609
636
  Warnings from the model provider (e.g. unsupported settings)
610
637
  */
611
- readonly warnings: LanguageModelV1CallWarning[] | undefined;
638
+ readonly warnings: CallWarning[] | undefined;
612
639
  /**
613
640
  Optional raw response data.
614
641
  */
@@ -622,18 +649,18 @@ declare class GenerateTextResult<TOOLS extends Record<string, ExperimentalTool>>
622
649
  Logprobs for the completion.
623
650
  `undefined` if the mode does not support logprobs or if was not enabled
624
651
  */
625
- readonly logprobs: LanguageModelV1LogProbs | undefined;
652
+ readonly logprobs: LogProbs | undefined;
626
653
  constructor(options: {
627
654
  text: string;
628
655
  toolCalls: ToToolCallArray<TOOLS>;
629
656
  toolResults: ToToolResultArray<TOOLS>;
630
- finishReason: LanguageModelV1FinishReason;
657
+ finishReason: FinishReason;
631
658
  usage: TokenUsage;
632
- warnings: LanguageModelV1CallWarning[] | undefined;
659
+ warnings: CallWarning[] | undefined;
633
660
  rawResponse?: {
634
661
  headers?: Record<string, string>;
635
662
  };
636
- logprobs: LanguageModelV1LogProbs | undefined;
663
+ logprobs: LogProbs | undefined;
637
664
  });
638
665
  }
639
666
 
@@ -675,7 +702,7 @@ declare function experimental_streamText<TOOLS extends Record<string, Experiment
675
702
  /**
676
703
  The language model to use.
677
704
  */
678
- model: LanguageModelV1;
705
+ model: LanguageModel;
679
706
  /**
680
707
  The tools that the model can call. The model needs to support calling tools.
681
708
  */
@@ -693,8 +720,8 @@ type TextStreamPart<TOOLS extends Record<string, ExperimentalTool>> = {
693
720
  type: 'tool-result';
694
721
  } & ToToolResult<TOOLS>) | {
695
722
  type: 'finish';
696
- finishReason: LanguageModelV1FinishReason;
697
- logprobs?: LanguageModelV1LogProbs;
723
+ finishReason: FinishReason;
724
+ logprobs?: LogProbs;
698
725
  usage: {
699
726
  promptTokens: number;
700
727
  completionTokens: number;
@@ -709,7 +736,7 @@ declare class StreamTextResult<TOOLS extends Record<string, ExperimentalTool>> {
709
736
  /**
710
737
  Warnings from the model provider (e.g. unsupported settings)
711
738
  */
712
- readonly warnings: LanguageModelV1CallWarning[] | undefined;
739
+ readonly warnings: CallWarning[] | undefined;
713
740
  /**
714
741
  Optional raw response data.
715
742
  */
@@ -721,7 +748,7 @@ declare class StreamTextResult<TOOLS extends Record<string, ExperimentalTool>> {
721
748
  };
722
749
  constructor({ stream, warnings, rawResponse, }: {
723
750
  stream: ReadableStream<TextStreamPart<TOOLS>>;
724
- warnings: LanguageModelV1CallWarning[] | undefined;
751
+ warnings: CallWarning[] | undefined;
725
752
  rawResponse?: {
726
753
  headers?: Record<string, string>;
727
754
  };
@@ -1893,4 +1920,4 @@ declare function streamToResponse(res: ReadableStream, response: ServerResponse,
1893
1920
  status?: number;
1894
1921
  }): void;
1895
1922
 
1896
- export { AIStream, AIStreamCallbacksAndOptions, AIStreamParser, AIStreamParserOptions, AWSBedrockAnthropicMessagesStream, AWSBedrockAnthropicStream, AWSBedrockCohereStream, AWSBedrockLlama2Stream, AWSBedrockStream, AnthropicStream, AssistantContent, AssistantMessage, AssistantResponse, ChatRequest, ChatRequestOptions, CohereStream, CompletionUsage, CreateMessage, DataContent, DataMessage, DeepPartial, ExperimentalAssistantMessage, ExperimentalMessage, ExperimentalTool, ExperimentalToolMessage, ExperimentalUserMessage, Function, FunctionCall$1 as FunctionCall, FunctionCallHandler, FunctionCallPayload, GenerateObjectResult, GenerateTextResult, GoogleGenerativeAIStream, HuggingFaceStream, IdGenerator, ImagePart, InkeepAIStreamCallbacksAndOptions, InkeepChatResultCallbacks, InkeepOnFinalMetadata, InkeepStream, JSONValue, LangChainStream, Message$1 as Message, MistralStream, ObjectStreamPart, ObjectStreamPartInput, OpenAIStream, OpenAIStreamCallbacks, ReactResponseRow, ReplicateStream, RequestOptions, StreamData, StreamObjectResult, StreamPart, StreamString, StreamTextResult, StreamingTextResponse, TextPart$1 as TextPart, TextStreamPart, Tool, ToolCall, ToolCallHandler, ToolCallPart, ToolCallPayload, ToolChoice, ToolContent, ToolResultPart, UseChatOptions, UseCompletionOptions, UserContent, convertDataContentToBase64String, convertDataContentToUint8Array, createCallbacksTransformer, createChunkDecoder, createEventStreamTransformer, createStreamDataTransformer, experimental_AssistantResponse, experimental_StreamData, experimental_StreamingReactResponse, experimental_generateObject, experimental_generateText, experimental_streamObject, experimental_streamText, formatStreamPart, generateId, isStreamStringEqualToType, generateId as nanoid, parseStreamPart, readDataStream, readableFromAsyncIterable, streamToResponse, tool, trimStartOfStreamHelper };
1923
+ export { AIStream, AIStreamCallbacksAndOptions, AIStreamParser, AIStreamParserOptions, AWSBedrockAnthropicMessagesStream, AWSBedrockAnthropicStream, AWSBedrockCohereStream, AWSBedrockLlama2Stream, AWSBedrockStream, AnthropicStream, AssistantContent, AssistantMessage, AssistantResponse, CallWarning, ChatRequest, ChatRequestOptions, CohereStream, CompletionUsage, CreateMessage, DataContent, DataMessage, DeepPartial, ExperimentalAssistantMessage, ExperimentalMessage, ExperimentalTool, ExperimentalToolMessage, ExperimentalUserMessage, FinishReason, Function, FunctionCall$1 as FunctionCall, FunctionCallHandler, FunctionCallPayload, GenerateObjectResult, GenerateTextResult, GoogleGenerativeAIStream, HuggingFaceStream, IdGenerator, ImagePart, InkeepAIStreamCallbacksAndOptions, InkeepChatResultCallbacks, InkeepOnFinalMetadata, InkeepStream, JSONValue, LangChainStream, LanguageModel, LogProbs, Message$1 as Message, MistralStream, ObjectStreamPart, ObjectStreamPartInput, OpenAIStream, OpenAIStreamCallbacks, ReactResponseRow, ReplicateStream, RequestOptions, StreamData, StreamObjectResult, StreamPart, StreamString, StreamTextResult, StreamingTextResponse, TextPart$1 as TextPart, TextStreamPart, Tool, ToolCall, ToolCallHandler, ToolCallPart, ToolCallPayload, ToolChoice, ToolContent, ToolResultPart, UseChatOptions, UseCompletionOptions, UserContent, convertDataContentToBase64String, convertDataContentToUint8Array, createCallbacksTransformer, createChunkDecoder, createEventStreamTransformer, createStreamDataTransformer, experimental_AssistantResponse, experimental_StreamData, experimental_StreamingReactResponse, experimental_generateObject, experimental_generateText, experimental_streamObject, experimental_streamText, formatStreamPart, generateId, isStreamStringEqualToType, generateId as nanoid, parseStreamPart, readDataStream, readableFromAsyncIterable, streamToResponse, tool, trimStartOfStreamHelper };
package/dist/index.d.ts CHANGED
@@ -1,5 +1,6 @@
1
- import { LanguageModelV1, LanguageModelV1FinishReason, LanguageModelV1CallWarning, LanguageModelV1LogProbs } from '@ai-sdk/provider';
2
1
  import { z } from 'zod';
2
+ import { LanguageModelV1, LanguageModelV1FinishReason, LanguageModelV1LogProbs, LanguageModelV1CallWarning } from '@ai-sdk/provider';
3
+ export { APICallError, EmptyResponseBodyError, InvalidArgumentError, InvalidDataContentError, InvalidPromptError, InvalidResponseDataError, InvalidToolArgumentsError, JSONParseError, LoadAPIKeyError, NoObjectGeneratedError, NoSuchToolError, RetryError, ToolCallParseError, TypeValidationError, UnsupportedFunctionalityError, UnsupportedJSONSchemaError } from '@ai-sdk/provider';
3
4
  import { ServerResponse } from 'node:http';
4
5
  import { AssistantStream } from 'openai/lib/AssistantStream';
5
6
  import { Run } from 'openai/resources/beta/threads/runs/runs';
@@ -214,6 +215,32 @@ type Prompt = {
214
215
  messages?: Array<ExperimentalMessage>;
215
216
  };
216
217
 
218
+ /**
219
+ Language model that is used by the AI SDK Core functions.
220
+ */
221
+ type LanguageModel = LanguageModelV1;
222
+ /**
223
+ Reason why a language model finished generating a response.
224
+
225
+ Can be one of the following:
226
+ - `stop`: model generated stop sequence
227
+ - `length`: model generated maximum number of tokens
228
+ - `content-filter`: content filter violation stopped the model
229
+ - `tool-calls`: model triggered tool calls
230
+ - `error`: model stopped because of an error
231
+ - `other`: model stopped for other reasons
232
+ */
233
+ type FinishReason = LanguageModelV1FinishReason;
234
+ /**
235
+ Log probabilities for each token and its top log probabilities.
236
+ */
237
+ type LogProbs = LanguageModelV1LogProbs;
238
+ /**
239
+ Warning from the model provider for this call. The call will proceed, but e.g.
240
+ some settings might not be supported, which can lead to suboptimal results.
241
+ */
242
+ type CallWarning = LanguageModelV1CallWarning;
243
+
217
244
  /**
218
245
  Generate a structured, typed object for a given prompt and schema using a language model.
219
246
 
@@ -254,7 +281,7 @@ declare function experimental_generateObject<T>({ model, schema, mode, system, p
254
281
  /**
255
282
  The language model to use.
256
283
  */
257
- model: LanguageModelV1;
284
+ model: LanguageModel;
258
285
  /**
259
286
  The schema of the object that the model should generate.
260
287
  */
@@ -277,7 +304,7 @@ declare class GenerateObjectResult<T> {
277
304
  /**
278
305
  The reason why the generation finished.
279
306
  */
280
- readonly finishReason: LanguageModelV1FinishReason;
307
+ readonly finishReason: FinishReason;
281
308
  /**
282
309
  The token usage of the generated text.
283
310
  */
@@ -285,7 +312,7 @@ declare class GenerateObjectResult<T> {
285
312
  /**
286
313
  Warnings from the model provider (e.g. unsupported settings)
287
314
  */
288
- readonly warnings: LanguageModelV1CallWarning[] | undefined;
315
+ readonly warnings: CallWarning[] | undefined;
289
316
  /**
290
317
  Optional raw response data.
291
318
  */
@@ -299,16 +326,16 @@ declare class GenerateObjectResult<T> {
299
326
  Logprobs for the completion.
300
327
  `undefined` if the mode does not support logprobs or if was not enabled
301
328
  */
302
- readonly logprobs: LanguageModelV1LogProbs | undefined;
329
+ readonly logprobs: LogProbs | undefined;
303
330
  constructor(options: {
304
331
  object: T;
305
- finishReason: LanguageModelV1FinishReason;
332
+ finishReason: FinishReason;
306
333
  usage: TokenUsage;
307
- warnings: LanguageModelV1CallWarning[] | undefined;
334
+ warnings: CallWarning[] | undefined;
308
335
  rawResponse?: {
309
336
  headers?: Record<string, string>;
310
337
  };
311
- logprobs: LanguageModelV1LogProbs | undefined;
338
+ logprobs: LogProbs | undefined;
312
339
  });
313
340
  }
314
341
 
@@ -370,7 +397,7 @@ declare function experimental_streamObject<T>({ model, schema, mode, system, pro
370
397
  /**
371
398
  The language model to use.
372
399
  */
373
- model: LanguageModelV1;
400
+ model: LanguageModel;
374
401
  /**
375
402
  The schema of the object that the model should generate.
376
403
  */
@@ -387,8 +414,8 @@ type ObjectStreamPartInput = {
387
414
  error: unknown;
388
415
  } | {
389
416
  type: 'finish';
390
- finishReason: LanguageModelV1FinishReason;
391
- logprobs?: LanguageModelV1LogProbs;
417
+ finishReason: FinishReason;
418
+ logprobs?: LogProbs;
392
419
  usage: {
393
420
  promptTokens: number;
394
421
  completionTokens: number;
@@ -407,7 +434,7 @@ declare class StreamObjectResult<T> {
407
434
  /**
408
435
  Warnings from the model provider (e.g. unsupported settings)
409
436
  */
410
- readonly warnings: LanguageModelV1CallWarning[] | undefined;
437
+ readonly warnings: CallWarning[] | undefined;
411
438
  /**
412
439
  Optional raw response data.
413
440
  */
@@ -419,7 +446,7 @@ declare class StreamObjectResult<T> {
419
446
  };
420
447
  constructor({ stream, warnings, rawResponse, }: {
421
448
  stream: ReadableStream<string | ObjectStreamPartInput>;
422
- warnings: LanguageModelV1CallWarning[] | undefined;
449
+ warnings: CallWarning[] | undefined;
423
450
  rawResponse?: {
424
451
  headers?: Record<string, string>;
425
452
  };
@@ -574,7 +601,7 @@ declare function experimental_generateText<TOOLS extends Record<string, Experime
574
601
  /**
575
602
  The language model to use.
576
603
  */
577
- model: LanguageModelV1;
604
+ model: LanguageModel;
578
605
  /**
579
606
  The tools that the model can call. The model needs to support calling tools.
580
607
  */
@@ -600,7 +627,7 @@ declare class GenerateTextResult<TOOLS extends Record<string, ExperimentalTool>>
600
627
  /**
601
628
  The reason why the generation finished.
602
629
  */
603
- readonly finishReason: LanguageModelV1FinishReason;
630
+ readonly finishReason: FinishReason;
604
631
  /**
605
632
  The token usage of the generated text.
606
633
  */
@@ -608,7 +635,7 @@ declare class GenerateTextResult<TOOLS extends Record<string, ExperimentalTool>>
608
635
  /**
609
636
  Warnings from the model provider (e.g. unsupported settings)
610
637
  */
611
- readonly warnings: LanguageModelV1CallWarning[] | undefined;
638
+ readonly warnings: CallWarning[] | undefined;
612
639
  /**
613
640
  Optional raw response data.
614
641
  */
@@ -622,18 +649,18 @@ declare class GenerateTextResult<TOOLS extends Record<string, ExperimentalTool>>
622
649
  Logprobs for the completion.
623
650
  `undefined` if the mode does not support logprobs or if was not enabled
624
651
  */
625
- readonly logprobs: LanguageModelV1LogProbs | undefined;
652
+ readonly logprobs: LogProbs | undefined;
626
653
  constructor(options: {
627
654
  text: string;
628
655
  toolCalls: ToToolCallArray<TOOLS>;
629
656
  toolResults: ToToolResultArray<TOOLS>;
630
- finishReason: LanguageModelV1FinishReason;
657
+ finishReason: FinishReason;
631
658
  usage: TokenUsage;
632
- warnings: LanguageModelV1CallWarning[] | undefined;
659
+ warnings: CallWarning[] | undefined;
633
660
  rawResponse?: {
634
661
  headers?: Record<string, string>;
635
662
  };
636
- logprobs: LanguageModelV1LogProbs | undefined;
663
+ logprobs: LogProbs | undefined;
637
664
  });
638
665
  }
639
666
 
@@ -675,7 +702,7 @@ declare function experimental_streamText<TOOLS extends Record<string, Experiment
675
702
  /**
676
703
  The language model to use.
677
704
  */
678
- model: LanguageModelV1;
705
+ model: LanguageModel;
679
706
  /**
680
707
  The tools that the model can call. The model needs to support calling tools.
681
708
  */
@@ -693,8 +720,8 @@ type TextStreamPart<TOOLS extends Record<string, ExperimentalTool>> = {
693
720
  type: 'tool-result';
694
721
  } & ToToolResult<TOOLS>) | {
695
722
  type: 'finish';
696
- finishReason: LanguageModelV1FinishReason;
697
- logprobs?: LanguageModelV1LogProbs;
723
+ finishReason: FinishReason;
724
+ logprobs?: LogProbs;
698
725
  usage: {
699
726
  promptTokens: number;
700
727
  completionTokens: number;
@@ -709,7 +736,7 @@ declare class StreamTextResult<TOOLS extends Record<string, ExperimentalTool>> {
709
736
  /**
710
737
  Warnings from the model provider (e.g. unsupported settings)
711
738
  */
712
- readonly warnings: LanguageModelV1CallWarning[] | undefined;
739
+ readonly warnings: CallWarning[] | undefined;
713
740
  /**
714
741
  Optional raw response data.
715
742
  */
@@ -721,7 +748,7 @@ declare class StreamTextResult<TOOLS extends Record<string, ExperimentalTool>> {
721
748
  };
722
749
  constructor({ stream, warnings, rawResponse, }: {
723
750
  stream: ReadableStream<TextStreamPart<TOOLS>>;
724
- warnings: LanguageModelV1CallWarning[] | undefined;
751
+ warnings: CallWarning[] | undefined;
725
752
  rawResponse?: {
726
753
  headers?: Record<string, string>;
727
754
  };
@@ -1893,4 +1920,4 @@ declare function streamToResponse(res: ReadableStream, response: ServerResponse,
1893
1920
  status?: number;
1894
1921
  }): void;
1895
1922
 
1896
- export { AIStream, AIStreamCallbacksAndOptions, AIStreamParser, AIStreamParserOptions, AWSBedrockAnthropicMessagesStream, AWSBedrockAnthropicStream, AWSBedrockCohereStream, AWSBedrockLlama2Stream, AWSBedrockStream, AnthropicStream, AssistantContent, AssistantMessage, AssistantResponse, ChatRequest, ChatRequestOptions, CohereStream, CompletionUsage, CreateMessage, DataContent, DataMessage, DeepPartial, ExperimentalAssistantMessage, ExperimentalMessage, ExperimentalTool, ExperimentalToolMessage, ExperimentalUserMessage, Function, FunctionCall$1 as FunctionCall, FunctionCallHandler, FunctionCallPayload, GenerateObjectResult, GenerateTextResult, GoogleGenerativeAIStream, HuggingFaceStream, IdGenerator, ImagePart, InkeepAIStreamCallbacksAndOptions, InkeepChatResultCallbacks, InkeepOnFinalMetadata, InkeepStream, JSONValue, LangChainStream, Message$1 as Message, MistralStream, ObjectStreamPart, ObjectStreamPartInput, OpenAIStream, OpenAIStreamCallbacks, ReactResponseRow, ReplicateStream, RequestOptions, StreamData, StreamObjectResult, StreamPart, StreamString, StreamTextResult, StreamingTextResponse, TextPart$1 as TextPart, TextStreamPart, Tool, ToolCall, ToolCallHandler, ToolCallPart, ToolCallPayload, ToolChoice, ToolContent, ToolResultPart, UseChatOptions, UseCompletionOptions, UserContent, convertDataContentToBase64String, convertDataContentToUint8Array, createCallbacksTransformer, createChunkDecoder, createEventStreamTransformer, createStreamDataTransformer, experimental_AssistantResponse, experimental_StreamData, experimental_StreamingReactResponse, experimental_generateObject, experimental_generateText, experimental_streamObject, experimental_streamText, formatStreamPart, generateId, isStreamStringEqualToType, generateId as nanoid, parseStreamPart, readDataStream, readableFromAsyncIterable, streamToResponse, tool, trimStartOfStreamHelper };
1923
+ export { AIStream, AIStreamCallbacksAndOptions, AIStreamParser, AIStreamParserOptions, AWSBedrockAnthropicMessagesStream, AWSBedrockAnthropicStream, AWSBedrockCohereStream, AWSBedrockLlama2Stream, AWSBedrockStream, AnthropicStream, AssistantContent, AssistantMessage, AssistantResponse, CallWarning, ChatRequest, ChatRequestOptions, CohereStream, CompletionUsage, CreateMessage, DataContent, DataMessage, DeepPartial, ExperimentalAssistantMessage, ExperimentalMessage, ExperimentalTool, ExperimentalToolMessage, ExperimentalUserMessage, FinishReason, Function, FunctionCall$1 as FunctionCall, FunctionCallHandler, FunctionCallPayload, GenerateObjectResult, GenerateTextResult, GoogleGenerativeAIStream, HuggingFaceStream, IdGenerator, ImagePart, InkeepAIStreamCallbacksAndOptions, InkeepChatResultCallbacks, InkeepOnFinalMetadata, InkeepStream, JSONValue, LangChainStream, LanguageModel, LogProbs, Message$1 as Message, MistralStream, ObjectStreamPart, ObjectStreamPartInput, OpenAIStream, OpenAIStreamCallbacks, ReactResponseRow, ReplicateStream, RequestOptions, StreamData, StreamObjectResult, StreamPart, StreamString, StreamTextResult, StreamingTextResponse, TextPart$1 as TextPart, TextStreamPart, Tool, ToolCall, ToolCallHandler, ToolCallPart, ToolCallPayload, ToolChoice, ToolContent, ToolResultPart, UseChatOptions, UseCompletionOptions, UserContent, convertDataContentToBase64String, convertDataContentToUint8Array, createCallbacksTransformer, createChunkDecoder, createEventStreamTransformer, createStreamDataTransformer, experimental_AssistantResponse, experimental_StreamData, experimental_StreamingReactResponse, experimental_generateObject, experimental_generateText, experimental_streamObject, experimental_streamText, formatStreamPart, generateId, isStreamStringEqualToType, generateId as nanoid, parseStreamPart, readDataStream, readableFromAsyncIterable, streamToResponse, tool, trimStartOfStreamHelper };
package/dist/index.js CHANGED
@@ -31,6 +31,7 @@ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: tru
31
31
  var streams_exports = {};
32
32
  __export(streams_exports, {
33
33
  AIStream: () => AIStream,
34
+ APICallError: () => import_provider8.APICallError,
34
35
  AWSBedrockAnthropicMessagesStream: () => AWSBedrockAnthropicMessagesStream,
35
36
  AWSBedrockAnthropicStream: () => AWSBedrockAnthropicStream,
36
37
  AWSBedrockCohereStream: () => AWSBedrockCohereStream,
@@ -39,19 +40,34 @@ __export(streams_exports, {
39
40
  AnthropicStream: () => AnthropicStream,
40
41
  AssistantResponse: () => AssistantResponse,
41
42
  CohereStream: () => CohereStream,
43
+ EmptyResponseBodyError: () => import_provider8.EmptyResponseBodyError,
42
44
  GenerateObjectResult: () => GenerateObjectResult,
43
45
  GenerateTextResult: () => GenerateTextResult,
44
46
  GoogleGenerativeAIStream: () => GoogleGenerativeAIStream,
45
47
  HuggingFaceStream: () => HuggingFaceStream,
46
48
  InkeepStream: () => InkeepStream,
49
+ InvalidArgumentError: () => import_provider8.InvalidArgumentError,
50
+ InvalidDataContentError: () => import_provider8.InvalidDataContentError,
51
+ InvalidPromptError: () => import_provider8.InvalidPromptError,
52
+ InvalidResponseDataError: () => import_provider8.InvalidResponseDataError,
53
+ InvalidToolArgumentsError: () => import_provider8.InvalidToolArgumentsError,
54
+ JSONParseError: () => import_provider8.JSONParseError,
47
55
  LangChainStream: () => LangChainStream,
56
+ LoadAPIKeyError: () => import_provider8.LoadAPIKeyError,
48
57
  MistralStream: () => MistralStream,
58
+ NoObjectGeneratedError: () => import_provider8.NoObjectGeneratedError,
59
+ NoSuchToolError: () => import_provider8.NoSuchToolError,
49
60
  OpenAIStream: () => OpenAIStream,
50
61
  ReplicateStream: () => ReplicateStream,
62
+ RetryError: () => import_provider8.RetryError,
51
63
  StreamData: () => StreamData,
52
64
  StreamObjectResult: () => StreamObjectResult,
53
65
  StreamTextResult: () => StreamTextResult,
54
66
  StreamingTextResponse: () => StreamingTextResponse,
67
+ ToolCallParseError: () => import_provider8.ToolCallParseError,
68
+ TypeValidationError: () => import_provider8.TypeValidationError,
69
+ UnsupportedFunctionalityError: () => import_provider8.UnsupportedFunctionalityError,
70
+ UnsupportedJSONSchemaError: () => import_provider8.UnsupportedJSONSchemaError,
55
71
  convertDataContentToBase64String: () => convertDataContentToBase64String,
56
72
  convertDataContentToUint8Array: () => convertDataContentToUint8Array,
57
73
  createCallbacksTransformer: () => createCallbacksTransformer,
@@ -469,7 +485,7 @@ async function experimental_generateObject({
469
485
  });
470
486
  });
471
487
  if (generateResult.text === void 0) {
472
- throw new import_provider5.NoTextGeneratedError();
488
+ throw new import_provider5.NoObjectGeneratedError();
473
489
  }
474
490
  result = generateResult.text;
475
491
  finishReason = generateResult.finishReason;
@@ -495,7 +511,7 @@ async function experimental_generateObject({
495
511
  })
496
512
  );
497
513
  if (generateResult.text === void 0) {
498
- throw new import_provider5.NoTextGeneratedError();
514
+ throw new import_provider5.NoObjectGeneratedError();
499
515
  }
500
516
  result = generateResult.text;
501
517
  finishReason = generateResult.finishReason;
@@ -530,7 +546,7 @@ async function experimental_generateObject({
530
546
  );
531
547
  const functionArgs = (_b = (_a = generateResult.toolCalls) == null ? void 0 : _a[0]) == null ? void 0 : _b.args;
532
548
  if (functionArgs === void 0) {
533
- throw new import_provider5.NoTextGeneratedError();
549
+ throw new import_provider5.NoObjectGeneratedError();
534
550
  }
535
551
  result = functionArgs;
536
552
  finishReason = generateResult.finishReason;
@@ -1624,6 +1640,9 @@ function tool(tool2) {
1624
1640
  return tool2;
1625
1641
  }
1626
1642
 
1643
+ // core/types/errors.ts
1644
+ var import_provider8 = require("@ai-sdk/provider");
1645
+
1627
1646
  // shared/stream-parts.ts
1628
1647
  var textStreamPart = {
1629
1648
  code: "0",
@@ -3011,6 +3030,7 @@ function streamToResponse(res, response, init) {
3011
3030
  // Annotate the CommonJS export names for ESM import in node:
3012
3031
  0 && (module.exports = {
3013
3032
  AIStream,
3033
+ APICallError,
3014
3034
  AWSBedrockAnthropicMessagesStream,
3015
3035
  AWSBedrockAnthropicStream,
3016
3036
  AWSBedrockCohereStream,
@@ -3019,19 +3039,34 @@ function streamToResponse(res, response, init) {
3019
3039
  AnthropicStream,
3020
3040
  AssistantResponse,
3021
3041
  CohereStream,
3042
+ EmptyResponseBodyError,
3022
3043
  GenerateObjectResult,
3023
3044
  GenerateTextResult,
3024
3045
  GoogleGenerativeAIStream,
3025
3046
  HuggingFaceStream,
3026
3047
  InkeepStream,
3048
+ InvalidArgumentError,
3049
+ InvalidDataContentError,
3050
+ InvalidPromptError,
3051
+ InvalidResponseDataError,
3052
+ InvalidToolArgumentsError,
3053
+ JSONParseError,
3027
3054
  LangChainStream,
3055
+ LoadAPIKeyError,
3028
3056
  MistralStream,
3057
+ NoObjectGeneratedError,
3058
+ NoSuchToolError,
3029
3059
  OpenAIStream,
3030
3060
  ReplicateStream,
3061
+ RetryError,
3031
3062
  StreamData,
3032
3063
  StreamObjectResult,
3033
3064
  StreamTextResult,
3034
3065
  StreamingTextResponse,
3066
+ ToolCallParseError,
3067
+ TypeValidationError,
3068
+ UnsupportedFunctionalityError,
3069
+ UnsupportedJSONSchemaError,
3035
3070
  convertDataContentToBase64String,
3036
3071
  convertDataContentToUint8Array,
3037
3072
  createCallbacksTransformer,