ai 3.3.26 → 3.3.27

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.d.ts CHANGED
@@ -44,7 +44,7 @@ type TelemetrySettings = {
44
44
  /**
45
45
  Represents the number of tokens used in a prompt and completion.
46
46
  */
47
- type CompletionTokenUsage$1 = {
47
+ type LanguageModelUsage$1 = {
48
48
  /**
49
49
  The number of tokens used in the prompt.
50
50
  */
@@ -61,7 +61,7 @@ type CompletionTokenUsage$1 = {
61
61
  /**
62
62
  Represents the number of tokens used in an embedding.
63
63
  */
64
- type EmbeddingTokenUsage = {
64
+ type EmbeddingModelUsage$1 = {
65
65
  /**
66
66
  The number of tokens used in the embedding.
67
67
  */
@@ -95,6 +95,8 @@ Can be one of the following:
95
95
  type FinishReason = LanguageModelV1FinishReason;
96
96
  /**
97
97
  Log probabilities for each token and its top log probabilities.
98
+
99
+ @deprecated Will become a provider extension in the future.
98
100
  */
99
101
  type LogProbs = LanguageModelV1LogProbs;
100
102
  /**
@@ -114,6 +116,23 @@ type CoreToolChoice<TOOLS extends Record<string, unknown>> = 'auto' | 'none' | '
114
116
  type: 'tool';
115
117
  toolName: keyof TOOLS;
116
118
  };
119
+ type LanguageModelResponseMetadata = {
120
+ /**
121
+ ID for the generated response.
122
+ */
123
+ id: string;
124
+ /**
125
+ Timestamp for the start of the generated response.
126
+ */
127
+ timestamp: Date;
128
+ /**
129
+ The ID of the response model that was used to generate the response.
130
+ */
131
+ modelId: string;
132
+ };
133
+ type LanguageModelResponseMetadataWithHeaders = LanguageModelResponseMetadata & {
134
+ headers?: Record<string, string>;
135
+ };
117
136
 
118
137
  /**
119
138
  * Provider for language and text embedding models.
@@ -151,10 +170,19 @@ functionality that can be fully encapsulated in the provider.
151
170
  type ProviderMetadata = LanguageModelV1ProviderMetadata;
152
171
 
153
172
  /**
154
- * @deprecated Use CompletionTokenUsage instead.
173
+ * @deprecated Use LanguageModelUsage instead.
174
+ */
175
+ type TokenUsage = LanguageModelUsage$1;
176
+ /**
177
+ * @deprecated Use LanguageModelUsage instead.
155
178
  */
156
- type TokenUsage = CompletionTokenUsage$1;
157
- type CompletionTokenUsage = CompletionTokenUsage$1;
179
+ type CompletionTokenUsage = LanguageModelUsage$1;
180
+ type LanguageModelUsage = LanguageModelUsage$1;
181
+ /**
182
+ * @deprecated Use EmbeddingModelUsage instead.
183
+ */
184
+ type EmbeddingTokenUsage = EmbeddingModelUsage$1;
185
+ type EmbeddingModelUsage = EmbeddingModelUsage$1;
158
186
 
159
187
  /**
160
188
  The result of a `embed` call.
@@ -172,7 +200,7 @@ interface EmbedResult<VALUE> {
172
200
  /**
173
201
  The embedding token usage.
174
202
  */
175
- readonly usage: EmbeddingTokenUsage;
203
+ readonly usage: EmbeddingModelUsage$1;
176
204
  /**
177
205
  Optional raw response data.
178
206
  */
@@ -242,7 +270,7 @@ interface EmbedManyResult<VALUE> {
242
270
  /**
243
271
  The embedding token usage.
244
272
  */
245
- readonly usage: EmbeddingTokenUsage;
273
+ readonly usage: EmbeddingModelUsage$1;
246
274
  }
247
275
 
248
276
  /**
@@ -430,6 +458,12 @@ interface ToolCallPart {
430
458
  Arguments of the tool call. This is a JSON-serializable object that matches the tool's input schema.
431
459
  */
432
460
  args: unknown;
461
+ /**
462
+ Additional provider-specific metadata. They are passed through
463
+ to the provider from the AI SDK and enable provider-specific
464
+ functionality that can be fully encapsulated in the provider.
465
+ */
466
+ experimental_providerMetadata?: ProviderMetadata;
433
467
  }
434
468
  /**
435
469
  Tool result content part of a prompt. It contains the result of the tool call with the matching ID.
@@ -583,13 +617,15 @@ interface GenerateObjectResult<T> {
583
617
  /**
584
618
  The token usage of the generated text.
585
619
  */
586
- readonly usage: CompletionTokenUsage$1;
620
+ readonly usage: LanguageModelUsage$1;
587
621
  /**
588
622
  Warnings from the model provider (e.g. unsupported settings)
589
623
  */
590
624
  readonly warnings: CallWarning[] | undefined;
591
625
  /**
592
- Optional raw response data.
626
+ Optional raw response data.
627
+
628
+ @deprecated Use `response.headers` instead.
593
629
  */
594
630
  readonly rawResponse?: {
595
631
  /**
@@ -598,8 +634,14 @@ interface GenerateObjectResult<T> {
598
634
  headers?: Record<string, string>;
599
635
  };
600
636
  /**
601
- Logprobs for the completion.
602
- `undefined` if the mode does not support logprobs or if was not enabled
637
+ Additional response information.
638
+ */
639
+ readonly response: LanguageModelResponseMetadataWithHeaders;
640
+ /**
641
+ Logprobs for the completion.
642
+ `undefined` if the mode does not support logprobs or if was not enabled.
643
+
644
+ @deprecated Will become a provider extension in the future.
603
645
  */
604
646
  readonly logprobs: LogProbs | undefined;
605
647
  /**
@@ -663,6 +705,13 @@ Default and recommended: 'auto' (best mode for the model).
663
705
  Optional telemetry configuration (experimental).
664
706
  */
665
707
  experimental_telemetry?: TelemetrySettings;
708
+ /**
709
+ * Internal. For test use only. May change without notice.
710
+ */
711
+ _internal?: {
712
+ generateId?: () => string;
713
+ currentDate?: () => Date;
714
+ };
666
715
  }): Promise<GenerateObjectResult<OBJECT>>;
667
716
  /**
668
717
  Generate an array with structured, typed elements for a given prompt and element schema using a language model.
@@ -712,6 +761,13 @@ Default and recommended: 'auto' (best mode for the model).
712
761
  Optional telemetry configuration (experimental).
713
762
  */
714
763
  experimental_telemetry?: TelemetrySettings;
764
+ /**
765
+ * Internal. For test use only. May change without notice.
766
+ */
767
+ _internal?: {
768
+ generateId?: () => string;
769
+ currentDate?: () => Date;
770
+ };
715
771
  }): Promise<GenerateObjectResult<Array<ELEMENT>>>;
716
772
  /**
717
773
  Generate JSON with any schema for a given prompt using a language model.
@@ -735,6 +791,13 @@ The mode to use for object generation. Must be "json" for no-schema output.
735
791
  Optional telemetry configuration (experimental).
736
792
  */
737
793
  experimental_telemetry?: TelemetrySettings;
794
+ /**
795
+ * Internal. For test use only. May change without notice.
796
+ */
797
+ _internal?: {
798
+ generateId?: () => string;
799
+ currentDate?: () => Date;
800
+ };
738
801
  }): Promise<GenerateObjectResult<JSONValue>>;
739
802
  /**
740
803
  * @deprecated Use `generateObject` instead.
@@ -754,7 +817,7 @@ interface StreamObjectResult<PARTIAL, RESULT, ELEMENT_STREAM> {
754
817
  /**
755
818
  The token usage of the generated response. Resolved when the response is finished.
756
819
  */
757
- readonly usage: Promise<CompletionTokenUsage$1>;
820
+ readonly usage: Promise<LanguageModelUsage$1>;
758
821
  /**
759
822
  Additional provider-specific metadata. They are passed through
760
823
  from the provider to the AI SDK and enable provider-specific
@@ -762,7 +825,9 @@ interface StreamObjectResult<PARTIAL, RESULT, ELEMENT_STREAM> {
762
825
  */
763
826
  readonly experimental_providerMetadata: Promise<ProviderMetadata | undefined>;
764
827
  /**
765
- Optional raw response data.
828
+ Optional raw response data.
829
+
830
+ @deprecated Use `response` instead.
766
831
  */
767
832
  readonly rawResponse?: {
768
833
  /**
@@ -771,6 +836,10 @@ interface StreamObjectResult<PARTIAL, RESULT, ELEMENT_STREAM> {
771
836
  headers?: Record<string, string>;
772
837
  };
773
838
  /**
839
+ Additional response information.
840
+ */
841
+ readonly response: Promise<LanguageModelResponseMetadataWithHeaders>;
842
+ /**
774
843
  The generated object (typed according to the schema). Resolved when the response is finished.
775
844
  */
776
845
  readonly object: Promise<RESULT>;
@@ -817,33 +886,29 @@ interface StreamObjectResult<PARTIAL, RESULT, ELEMENT_STREAM> {
817
886
  */
818
887
  toTextStreamResponse(init?: ResponseInit): Response;
819
888
  }
820
- type ObjectStreamInputPart = {
889
+ type ObjectStreamPart<PARTIAL> = {
890
+ type: 'object';
891
+ object: PARTIAL;
892
+ } | {
893
+ type: 'text-delta';
894
+ textDelta: string;
895
+ } | {
821
896
  type: 'error';
822
897
  error: unknown;
823
898
  } | {
824
899
  type: 'finish';
825
900
  finishReason: FinishReason;
826
901
  logprobs?: LogProbs;
827
- usage: {
828
- promptTokens: number;
829
- completionTokens: number;
830
- totalTokens: number;
831
- };
902
+ usage: LanguageModelUsage$1;
903
+ response: LanguageModelResponseMetadata;
832
904
  providerMetadata?: ProviderMetadata;
833
905
  };
834
- type ObjectStreamPart<PARTIAL> = ObjectStreamInputPart | {
835
- type: 'object';
836
- object: PARTIAL;
837
- } | {
838
- type: 'text-delta';
839
- textDelta: string;
840
- };
841
906
 
842
907
  type OnFinishCallback<RESULT> = (event: {
843
908
  /**
844
909
  The token usage of the generated response.
845
910
  */
846
- usage: CompletionTokenUsage$1;
911
+ usage: LanguageModelUsage$1;
847
912
  /**
848
913
  The generated object. Can be undefined if the final object does not match the schema.
849
914
  */
@@ -854,14 +919,20 @@ type OnFinishCallback<RESULT> = (event: {
854
919
  error: unknown | undefined;
855
920
  /**
856
921
  Optional raw response data.
857
- */
922
+
923
+ @deprecated Use `response` instead.
924
+ */
858
925
  rawResponse?: {
859
926
  /**
860
927
  Response headers.
861
- */
928
+ */
862
929
  headers?: Record<string, string>;
863
930
  };
864
931
  /**
932
+ Response metadata.
933
+ */
934
+ response: LanguageModelResponseMetadataWithHeaders;
935
+ /**
865
936
  Warnings from the model provider (e.g. unsupported settings).
866
937
  */
867
938
  warnings?: CallWarning[];
@@ -928,6 +999,8 @@ Callback that is called when the LLM response and the final object validation ar
928
999
  * Internal. For test use only. May change without notice.
929
1000
  */
930
1001
  _internal?: {
1002
+ generateId?: () => string;
1003
+ currentDate?: () => Date;
931
1004
  now?: () => number;
932
1005
  };
933
1006
  }): Promise<StreamObjectResult<DeepPartial<OBJECT>, OBJECT, never>>;
@@ -987,6 +1060,8 @@ Callback that is called when the LLM response and the final object validation ar
987
1060
  * Internal. For test use only. May change without notice.
988
1061
  */
989
1062
  _internal?: {
1063
+ generateId?: () => string;
1064
+ currentDate?: () => Date;
990
1065
  now?: () => number;
991
1066
  };
992
1067
  }): Promise<StreamObjectResult<Array<ELEMENT>, Array<ELEMENT>, AsyncIterableStream<ELEMENT>>>;
@@ -1020,6 +1095,8 @@ Callback that is called when the LLM response and the final object validation ar
1020
1095
  * Internal. For test use only. May change without notice.
1021
1096
  */
1022
1097
  _internal?: {
1098
+ generateId?: () => string;
1099
+ currentDate?: () => Date;
1023
1100
  now?: () => number;
1024
1101
  };
1025
1102
  }): Promise<StreamObjectResult<JSONValue, JSONValue, never>>;
@@ -1179,7 +1256,7 @@ interface GenerateTextResult<TOOLS extends Record<string, CoreTool>> {
1179
1256
  /**
1180
1257
  The token usage of the generated text.
1181
1258
  */
1182
- readonly usage: CompletionTokenUsage$1;
1259
+ readonly usage: LanguageModelUsage$1;
1183
1260
  /**
1184
1261
  Warnings from the model provider (e.g. unsupported settings)
1185
1262
  */
@@ -1216,7 +1293,7 @@ interface GenerateTextResult<TOOLS extends Record<string, CoreTool>> {
1216
1293
  /**
1217
1294
  The token usage of the generated text.
1218
1295
  */
1219
- readonly usage: CompletionTokenUsage$1;
1296
+ readonly usage: LanguageModelUsage$1;
1220
1297
  /**
1221
1298
  Warnings from the model provider (e.g. unsupported settings)
1222
1299
  */
@@ -1227,17 +1304,25 @@ interface GenerateTextResult<TOOLS extends Record<string, CoreTool>> {
1227
1304
  */
1228
1305
  readonly logprobs: LogProbs | undefined;
1229
1306
  /**
1230
- Optional raw response data.
1231
- */
1307
+ Optional raw response data.
1308
+
1309
+ @deprecated Use `response.headers` instead.
1310
+ */
1232
1311
  readonly rawResponse?: {
1233
1312
  /**
1234
- Response headers.
1235
- */
1313
+ Response headers.
1314
+ */
1236
1315
  readonly headers?: Record<string, string>;
1237
1316
  };
1317
+ /**
1318
+ Additional response information.
1319
+ */
1320
+ readonly response: LanguageModelResponseMetadataWithHeaders;
1238
1321
  }>;
1239
1322
  /**
1240
- Optional raw response data.
1323
+ Optional raw response data.
1324
+
1325
+ @deprecated Use `response.headers` instead.
1241
1326
  */
1242
1327
  readonly rawResponse?: {
1243
1328
  /**
@@ -1246,8 +1331,14 @@ interface GenerateTextResult<TOOLS extends Record<string, CoreTool>> {
1246
1331
  readonly headers?: Record<string, string>;
1247
1332
  };
1248
1333
  /**
1249
- Logprobs for the completion.
1250
- `undefined` if the mode does not support logprobs or if was not enabled.
1334
+ Additional response information.
1335
+ */
1336
+ readonly response: LanguageModelResponseMetadataWithHeaders;
1337
+ /**
1338
+ Logprobs for the completion.
1339
+ `undefined` if the mode does not support logprobs or if was not enabled.
1340
+
1341
+ @deprecated Will become a provider extension in the future.
1251
1342
  */
1252
1343
  readonly logprobs: LogProbs | undefined;
1253
1344
  /**
@@ -1302,7 +1393,7 @@ If set and supported by the model, calls will generate deterministic results.
1302
1393
  @returns
1303
1394
  A result object that contains the generated text, the results of the tool calls, and additional information.
1304
1395
  */
1305
- declare function generateText<TOOLS extends Record<string, CoreTool>>({ model, tools, toolChoice, system, prompt, messages, maxRetries, abortSignal, headers, maxAutomaticRoundtrips, maxToolRoundtrips, experimental_telemetry: telemetry, ...settings }: CallSettings & Prompt & {
1396
+ declare function generateText<TOOLS extends Record<string, CoreTool>>({ model, tools, toolChoice, system, prompt, messages, maxRetries, abortSignal, headers, maxAutomaticRoundtrips, maxToolRoundtrips, experimental_telemetry: telemetry, _internal: { generateId, currentDate, }, ...settings }: CallSettings & Prompt & {
1306
1397
  /**
1307
1398
  The language model to use.
1308
1399
  */
@@ -1336,7 +1427,18 @@ By default, it's set to 0, which will disable the feature.
1336
1427
  * Optional telemetry configuration (experimental).
1337
1428
  */
1338
1429
  experimental_telemetry?: TelemetrySettings;
1430
+ /**
1431
+ * Internal. For test use only. May change without notice.
1432
+ */
1433
+ _internal?: {
1434
+ generateId?: () => string;
1435
+ currentDate?: () => Date;
1436
+ };
1339
1437
  }): Promise<GenerateTextResult<TOOLS>>;
1438
+ /**
1439
+ * @deprecated Use `generateText` instead.
1440
+ */
1441
+ declare const experimental_generateText: typeof generateText;
1340
1442
 
1341
1443
  /**
1342
1444
  A result object for accessing different stream types and additional information.
@@ -1352,7 +1454,7 @@ interface StreamTextResult<TOOLS extends Record<string, CoreTool>> {
1352
1454
 
1353
1455
  Resolved when the response is finished.
1354
1456
  */
1355
- readonly usage: Promise<CompletionTokenUsage$1>;
1457
+ readonly usage: Promise<LanguageModelUsage$1>;
1356
1458
  /**
1357
1459
  The reason why the generation finished. Taken from the last roundtrip.
1358
1460
 
@@ -1385,6 +1487,8 @@ interface StreamTextResult<TOOLS extends Record<string, CoreTool>> {
1385
1487
  readonly toolResults: Promise<ToToolResult<TOOLS>[]>;
1386
1488
  /**
1387
1489
  Optional raw response data.
1490
+
1491
+ @deprecated Use `response` instead.
1388
1492
  */
1389
1493
  readonly rawResponse?: {
1390
1494
  /**
@@ -1393,6 +1497,10 @@ interface StreamTextResult<TOOLS extends Record<string, CoreTool>> {
1393
1497
  headers?: Record<string, string>;
1394
1498
  };
1395
1499
  /**
1500
+ Additional response information.
1501
+ */
1502
+ readonly response: Promise<LanguageModelResponseMetadataWithHeaders>;
1503
+ /**
1396
1504
  A text stream that returns only the generated text deltas. You can use it
1397
1505
  as either an AsyncIterable or a ReadableStream. When an error occurs, the
1398
1506
  stream will throw the error.
@@ -1513,21 +1621,15 @@ type TextStreamPart<TOOLS extends Record<string, CoreTool>> = {
1513
1621
  type: 'roundtrip-finish';
1514
1622
  finishReason: FinishReason;
1515
1623
  logprobs?: LogProbs;
1516
- usage: {
1517
- promptTokens: number;
1518
- completionTokens: number;
1519
- totalTokens: number;
1520
- };
1624
+ usage: LanguageModelUsage$1;
1625
+ response: LanguageModelResponseMetadata;
1521
1626
  experimental_providerMetadata?: ProviderMetadata;
1522
1627
  } | {
1523
1628
  type: 'finish';
1524
1629
  finishReason: FinishReason;
1525
1630
  logprobs?: LogProbs;
1526
- usage: {
1527
- promptTokens: number;
1528
- completionTokens: number;
1529
- totalTokens: number;
1530
- };
1631
+ usage: LanguageModelUsage$1;
1632
+ response: LanguageModelResponseMetadata;
1531
1633
  experimental_providerMetadata?: ProviderMetadata;
1532
1634
  } | {
1533
1635
  type: 'error';
@@ -1580,7 +1682,7 @@ If set and supported by the model, calls will generate deterministic results.
1580
1682
  @return
1581
1683
  A result object for accessing different stream types and additional information.
1582
1684
  */
1583
- declare function streamText<TOOLS extends Record<string, CoreTool>>({ model, tools, toolChoice, system, prompt, messages, maxRetries, abortSignal, headers, maxToolRoundtrips, experimental_telemetry: telemetry, experimental_toolCallStreaming: toolCallStreaming, onChunk, onFinish, _internal: { now }, ...settings }: CallSettings & Prompt & {
1685
+ declare function streamText<TOOLS extends Record<string, CoreTool>>({ model, tools, toolChoice, system, prompt, messages, maxRetries, abortSignal, headers, maxToolRoundtrips, experimental_telemetry: telemetry, experimental_toolCallStreaming: toolCallStreaming, onChunk, onFinish, _internal: { now, generateId, currentDate, }, ...settings }: CallSettings & Prompt & {
1584
1686
  /**
1585
1687
  The language model to use.
1586
1688
  */
@@ -1634,7 +1736,7 @@ Callback that is called when the LLM response and all request tool executions
1634
1736
  /**
1635
1737
  The token usage of the generated response.
1636
1738
  */
1637
- usage: CompletionTokenUsage$1;
1739
+ usage: LanguageModelUsage$1;
1638
1740
  /**
1639
1741
  The full text that has been generated.
1640
1742
  */
@@ -1649,6 +1751,8 @@ Callback that is called when the LLM response and all request tool executions
1649
1751
  toolResults?: ToToolResult<TOOLS>[];
1650
1752
  /**
1651
1753
  Optional raw response data.
1754
+
1755
+ @deprecated Use `response` instead.
1652
1756
  */
1653
1757
  rawResponse?: {
1654
1758
  /**
@@ -1657,6 +1761,10 @@ Callback that is called when the LLM response and all request tool executions
1657
1761
  headers?: Record<string, string>;
1658
1762
  };
1659
1763
  /**
1764
+ Response metadata.
1765
+ */
1766
+ response: LanguageModelResponseMetadataWithHeaders;
1767
+ /**
1660
1768
  Warnings from the model provider (e.g. unsupported settings).
1661
1769
  */
1662
1770
  warnings?: CallWarning[];
@@ -1672,8 +1780,14 @@ Callback that is called when the LLM response and all request tool executions
1672
1780
  */
1673
1781
  _internal?: {
1674
1782
  now?: () => number;
1783
+ generateId?: () => string;
1784
+ currentDate?: () => Date;
1675
1785
  };
1676
1786
  }): Promise<StreamTextResult<TOOLS>>;
1787
+ /**
1788
+ * @deprecated Use `streamText` instead.
1789
+ */
1790
+ declare const experimental_streamText: typeof streamText;
1677
1791
 
1678
1792
  /**
1679
1793
  * Creates a custom provider with specified language models, text embedding models, and an optional fallback provider.
@@ -2781,10 +2895,10 @@ declare class StreamingTextResponse extends Response {
2781
2895
  constructor(res: ReadableStream, init?: ResponseInit, data?: StreamData);
2782
2896
  }
2783
2897
 
2784
- declare const generateId: (size?: number) => string;
2898
+ declare const generateId: () => string;
2785
2899
  /**
2786
2900
  @deprecated Use `generateId` instead.
2787
2901
  */
2788
- declare const nanoid: (size?: number) => string;
2902
+ declare const nanoid: () => string;
2789
2903
 
2790
- export { AIStream, AIStreamCallbacksAndOptions, AIStreamParser, AIStreamParserOptions, AWSBedrockAnthropicMessagesStream, AWSBedrockAnthropicStream, AWSBedrockCohereStream, AWSBedrockLlama2Stream, AWSBedrockStream, AnthropicStream, AssistantContent, AssistantResponse, CallWarning, CohereStream, CompletionTokenUsage, CompletionUsage, CoreAssistantMessage, CoreMessage, CoreSystemMessage, CoreTool, CoreToolChoice, CoreToolMessage, CoreUserMessage, DataContent, DownloadError, EmbedManyResult, EmbedResult, Embedding, EmbeddingModel, EmbeddingTokenUsage, ExperimentalAssistantMessage, ExperimentalMessage, ExperimentalTool, ExperimentalToolMessage, ExperimentalUserMessage, FinishReason, FunctionCallPayload, GenerateObjectResult, GenerateTextResult, GoogleGenerativeAIStream, HuggingFaceStream, ImagePart, InkeepAIStreamCallbacksAndOptions, InkeepChatResultCallbacks, InkeepOnFinalMetadata, InkeepStream, InvalidArgumentError, InvalidDataContentError, InvalidMessageRoleError, InvalidToolArgumentsError, langchainAdapter as LangChainAdapter, LangChainStream, LanguageModel, LogProbs, MessageConversionError, MistralStream, NoObjectGeneratedError, NoSuchProviderError, NoSuchToolError, ObjectStreamInputPart, ObjectStreamPart, OpenAIStream, OpenAIStreamCallbacks, Provider, ProviderMetadata, ReplicateStream, RetryError, StreamData, StreamObjectResult, StreamTextResult, StreamingTextResponse, TextPart$1 as TextPart, TextStreamPart, TokenUsage, ToolCallPart, ToolCallPayload, ToolContent, ToolResultPart, UserContent, convertToCoreMessages, cosineSimilarity, createCallbacksTransformer, createEventStreamTransformer, createStreamDataTransformer, embed, embedMany, experimental_AssistantResponse, experimental_ModelRegistry, experimental_Provider, experimental_ProviderRegistry, experimental_StreamData, experimental_createModelRegistry, experimental_createProviderRegistry, experimental_customProvider, experimental_generateObject, experimental_streamObject, generateId, generateObject, generateText, nanoid, readableFromAsyncIterable, streamObject, streamText, streamToResponse, tool, trimStartOfStreamHelper };
2904
+ export { AIStream, AIStreamCallbacksAndOptions, AIStreamParser, AIStreamParserOptions, AWSBedrockAnthropicMessagesStream, AWSBedrockAnthropicStream, AWSBedrockCohereStream, AWSBedrockLlama2Stream, AWSBedrockStream, AnthropicStream, AssistantContent, AssistantResponse, CallWarning, CohereStream, CompletionTokenUsage, CompletionUsage, CoreAssistantMessage, CoreMessage, CoreSystemMessage, CoreTool, CoreToolChoice, CoreToolMessage, CoreUserMessage, DataContent, DownloadError, EmbedManyResult, EmbedResult, Embedding, EmbeddingModel, EmbeddingModelUsage, EmbeddingTokenUsage, ExperimentalAssistantMessage, ExperimentalMessage, ExperimentalTool, ExperimentalToolMessage, ExperimentalUserMessage, FinishReason, FunctionCallPayload, GenerateObjectResult, GenerateTextResult, GoogleGenerativeAIStream, HuggingFaceStream, ImagePart, InkeepAIStreamCallbacksAndOptions, InkeepChatResultCallbacks, InkeepOnFinalMetadata, InkeepStream, InvalidArgumentError, InvalidDataContentError, InvalidMessageRoleError, InvalidToolArgumentsError, langchainAdapter as LangChainAdapter, LangChainStream, LanguageModel, LanguageModelResponseMetadata, LanguageModelResponseMetadataWithHeaders, LanguageModelUsage, LogProbs, MessageConversionError, MistralStream, NoObjectGeneratedError, NoSuchProviderError, NoSuchToolError, ObjectStreamPart, OpenAIStream, OpenAIStreamCallbacks, Provider, ProviderMetadata, ReplicateStream, RetryError, StreamData, StreamObjectResult, StreamTextResult, StreamingTextResponse, TextPart$1 as TextPart, TextStreamPart, TokenUsage, ToolCallPart, ToolCallPayload, ToolContent, ToolResultPart, UserContent, convertToCoreMessages, cosineSimilarity, createCallbacksTransformer, createEventStreamTransformer, createStreamDataTransformer, embed, embedMany, experimental_AssistantResponse, experimental_ModelRegistry, experimental_Provider, experimental_ProviderRegistry, experimental_StreamData, experimental_createModelRegistry, experimental_createProviderRegistry, experimental_customProvider, experimental_generateObject, experimental_generateText, experimental_streamObject, experimental_streamText, generateId, generateObject, generateText, nanoid, readableFromAsyncIterable, streamObject, streamText, streamToResponse, tool, trimStartOfStreamHelper };