ai 3.3.26 → 3.3.28
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +22 -0
- package/dist/index.d.mts +237 -57
- package/dist/index.d.ts +237 -57
- package/dist/index.js +246 -41
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +239 -37
- package/dist/index.mjs.map +1 -1
- package/package.json +8 -8
- package/rsc/dist/index.d.ts +8 -2
- package/rsc/dist/rsc-server.d.mts +8 -2
- package/rsc/dist/rsc-server.mjs +10 -4
- package/rsc/dist/rsc-server.mjs.map +1 -1
package/CHANGELOG.md
CHANGED
@@ -1,5 +1,27 @@
|
|
1
1
|
# ai
|
2
2
|
|
3
|
+
## 3.3.28
|
4
|
+
|
5
|
+
### Patch Changes
|
6
|
+
|
7
|
+
- db61c53: feat (ai/core): middleware support
|
8
|
+
|
9
|
+
## 3.3.27
|
10
|
+
|
11
|
+
### Patch Changes
|
12
|
+
|
13
|
+
- 03313cd: feat (ai): expose response id, response model, response timestamp in telemetry and api
|
14
|
+
- 3be7c1c: fix (provider/anthropic): support prompt caching on assistant messages
|
15
|
+
- Updated dependencies [03313cd]
|
16
|
+
- Updated dependencies [3be7c1c]
|
17
|
+
- @ai-sdk/provider-utils@1.0.18
|
18
|
+
- @ai-sdk/provider@0.0.23
|
19
|
+
- @ai-sdk/react@0.0.55
|
20
|
+
- @ai-sdk/solid@0.0.44
|
21
|
+
- @ai-sdk/svelte@0.0.46
|
22
|
+
- @ai-sdk/ui-utils@0.0.41
|
23
|
+
- @ai-sdk/vue@0.0.46
|
24
|
+
|
3
25
|
## 3.3.26
|
4
26
|
|
5
27
|
### Patch Changes
|
package/dist/index.d.mts
CHANGED
@@ -1,8 +1,8 @@
|
|
1
1
|
import { Schema, DeepPartial, ToolInvocation, Attachment, JSONValue as JSONValue$1, CreateMessage, FunctionCall as FunctionCall$1, AssistantMessage, DataMessage } from '@ai-sdk/ui-utils';
|
2
2
|
export { AssistantMessage, AssistantStatus, ChatRequest, ChatRequestOptions, CreateMessage, DataMessage, DeepPartial, Function, FunctionCall, FunctionCallHandler, IdGenerator, JSONValue, Message, RequestOptions, Schema, StreamPart, Tool, ToolCall, ToolCallHandler, ToolChoice, ToolInvocation, UseAssistantOptions, formatStreamPart, jsonSchema, parseStreamPart, processDataProtocolResponse, readDataStream } from '@ai-sdk/ui-utils';
|
3
3
|
import { AttributeValue } from '@opentelemetry/api';
|
4
|
-
import { EmbeddingModelV1, EmbeddingModelV1Embedding, LanguageModelV1, LanguageModelV1FinishReason, LanguageModelV1LogProbs, LanguageModelV1CallWarning, LanguageModelV1ProviderMetadata, JSONValue, NoSuchModelError, AISDKError } from '@ai-sdk/provider';
|
5
|
-
export { AISDKError, APICallError, EmptyResponseBodyError, InvalidPromptError, InvalidResponseDataError, JSONParseError, LoadAPIKeyError, NoContentGeneratedError, NoSuchModelError, TypeValidationError, UnsupportedFunctionalityError } from '@ai-sdk/provider';
|
4
|
+
import { EmbeddingModelV1, EmbeddingModelV1Embedding, LanguageModelV1, LanguageModelV1FinishReason, LanguageModelV1LogProbs, LanguageModelV1CallWarning, LanguageModelV1ProviderMetadata, JSONValue, LanguageModelV1CallOptions, NoSuchModelError, AISDKError } from '@ai-sdk/provider';
|
5
|
+
export { AISDKError, APICallError, EmptyResponseBodyError, InvalidPromptError, InvalidResponseDataError, JSONParseError, LanguageModelV1, LanguageModelV1CallOptions, LanguageModelV1Prompt, LanguageModelV1StreamPart, LoadAPIKeyError, NoContentGeneratedError, NoSuchModelError, TypeValidationError, UnsupportedFunctionalityError } from '@ai-sdk/provider';
|
6
6
|
import { z } from 'zod';
|
7
7
|
import { ServerResponse } from 'http';
|
8
8
|
import { ServerResponse as ServerResponse$1 } from 'node:http';
|
@@ -44,7 +44,7 @@ type TelemetrySettings = {
|
|
44
44
|
/**
|
45
45
|
Represents the number of tokens used in a prompt and completion.
|
46
46
|
*/
|
47
|
-
type
|
47
|
+
type LanguageModelUsage$1 = {
|
48
48
|
/**
|
49
49
|
The number of tokens used in the prompt.
|
50
50
|
*/
|
@@ -61,7 +61,7 @@ type CompletionTokenUsage$1 = {
|
|
61
61
|
/**
|
62
62
|
Represents the number of tokens used in an embedding.
|
63
63
|
*/
|
64
|
-
type
|
64
|
+
type EmbeddingModelUsage$1 = {
|
65
65
|
/**
|
66
66
|
The number of tokens used in the embedding.
|
67
67
|
*/
|
@@ -95,6 +95,8 @@ Can be one of the following:
|
|
95
95
|
type FinishReason = LanguageModelV1FinishReason;
|
96
96
|
/**
|
97
97
|
Log probabilities for each token and its top log probabilities.
|
98
|
+
|
99
|
+
@deprecated Will become a provider extension in the future.
|
98
100
|
*/
|
99
101
|
type LogProbs = LanguageModelV1LogProbs;
|
100
102
|
/**
|
@@ -114,6 +116,23 @@ type CoreToolChoice<TOOLS extends Record<string, unknown>> = 'auto' | 'none' | '
|
|
114
116
|
type: 'tool';
|
115
117
|
toolName: keyof TOOLS;
|
116
118
|
};
|
119
|
+
type LanguageModelResponseMetadata = {
|
120
|
+
/**
|
121
|
+
ID for the generated response.
|
122
|
+
*/
|
123
|
+
id: string;
|
124
|
+
/**
|
125
|
+
Timestamp for the start of the generated response.
|
126
|
+
*/
|
127
|
+
timestamp: Date;
|
128
|
+
/**
|
129
|
+
The ID of the response model that was used to generate the response.
|
130
|
+
*/
|
131
|
+
modelId: string;
|
132
|
+
};
|
133
|
+
type LanguageModelResponseMetadataWithHeaders = LanguageModelResponseMetadata & {
|
134
|
+
headers?: Record<string, string>;
|
135
|
+
};
|
117
136
|
|
118
137
|
/**
|
119
138
|
* Provider for language and text embedding models.
|
@@ -151,10 +170,19 @@ functionality that can be fully encapsulated in the provider.
|
|
151
170
|
type ProviderMetadata = LanguageModelV1ProviderMetadata;
|
152
171
|
|
153
172
|
/**
|
154
|
-
* @deprecated Use
|
173
|
+
* @deprecated Use LanguageModelUsage instead.
|
174
|
+
*/
|
175
|
+
type TokenUsage = LanguageModelUsage$1;
|
176
|
+
/**
|
177
|
+
* @deprecated Use LanguageModelUsage instead.
|
155
178
|
*/
|
156
|
-
type
|
157
|
-
type
|
179
|
+
type CompletionTokenUsage = LanguageModelUsage$1;
|
180
|
+
type LanguageModelUsage = LanguageModelUsage$1;
|
181
|
+
/**
|
182
|
+
* @deprecated Use EmbeddingModelUsage instead.
|
183
|
+
*/
|
184
|
+
type EmbeddingTokenUsage = EmbeddingModelUsage$1;
|
185
|
+
type EmbeddingModelUsage = EmbeddingModelUsage$1;
|
158
186
|
|
159
187
|
/**
|
160
188
|
The result of a `embed` call.
|
@@ -172,7 +200,7 @@ interface EmbedResult<VALUE> {
|
|
172
200
|
/**
|
173
201
|
The embedding token usage.
|
174
202
|
*/
|
175
|
-
readonly usage:
|
203
|
+
readonly usage: EmbeddingModelUsage$1;
|
176
204
|
/**
|
177
205
|
Optional raw response data.
|
178
206
|
*/
|
@@ -242,7 +270,7 @@ interface EmbedManyResult<VALUE> {
|
|
242
270
|
/**
|
243
271
|
The embedding token usage.
|
244
272
|
*/
|
245
|
-
readonly usage:
|
273
|
+
readonly usage: EmbeddingModelUsage$1;
|
246
274
|
}
|
247
275
|
|
248
276
|
/**
|
@@ -430,6 +458,12 @@ interface ToolCallPart {
|
|
430
458
|
Arguments of the tool call. This is a JSON-serializable object that matches the tool's input schema.
|
431
459
|
*/
|
432
460
|
args: unknown;
|
461
|
+
/**
|
462
|
+
Additional provider-specific metadata. They are passed through
|
463
|
+
to the provider from the AI SDK and enable provider-specific
|
464
|
+
functionality that can be fully encapsulated in the provider.
|
465
|
+
*/
|
466
|
+
experimental_providerMetadata?: ProviderMetadata;
|
433
467
|
}
|
434
468
|
/**
|
435
469
|
Tool result content part of a prompt. It contains the result of the tool call with the matching ID.
|
@@ -583,13 +617,15 @@ interface GenerateObjectResult<T> {
|
|
583
617
|
/**
|
584
618
|
The token usage of the generated text.
|
585
619
|
*/
|
586
|
-
readonly usage:
|
620
|
+
readonly usage: LanguageModelUsage$1;
|
587
621
|
/**
|
588
622
|
Warnings from the model provider (e.g. unsupported settings)
|
589
623
|
*/
|
590
624
|
readonly warnings: CallWarning[] | undefined;
|
591
625
|
/**
|
592
|
-
|
626
|
+
Optional raw response data.
|
627
|
+
|
628
|
+
@deprecated Use `response.headers` instead.
|
593
629
|
*/
|
594
630
|
readonly rawResponse?: {
|
595
631
|
/**
|
@@ -598,8 +634,14 @@ interface GenerateObjectResult<T> {
|
|
598
634
|
headers?: Record<string, string>;
|
599
635
|
};
|
600
636
|
/**
|
601
|
-
|
602
|
-
|
637
|
+
Additional response information.
|
638
|
+
*/
|
639
|
+
readonly response: LanguageModelResponseMetadataWithHeaders;
|
640
|
+
/**
|
641
|
+
Logprobs for the completion.
|
642
|
+
`undefined` if the mode does not support logprobs or if was not enabled.
|
643
|
+
|
644
|
+
@deprecated Will become a provider extension in the future.
|
603
645
|
*/
|
604
646
|
readonly logprobs: LogProbs | undefined;
|
605
647
|
/**
|
@@ -663,6 +705,13 @@ Default and recommended: 'auto' (best mode for the model).
|
|
663
705
|
Optional telemetry configuration (experimental).
|
664
706
|
*/
|
665
707
|
experimental_telemetry?: TelemetrySettings;
|
708
|
+
/**
|
709
|
+
* Internal. For test use only. May change without notice.
|
710
|
+
*/
|
711
|
+
_internal?: {
|
712
|
+
generateId?: () => string;
|
713
|
+
currentDate?: () => Date;
|
714
|
+
};
|
666
715
|
}): Promise<GenerateObjectResult<OBJECT>>;
|
667
716
|
/**
|
668
717
|
Generate an array with structured, typed elements for a given prompt and element schema using a language model.
|
@@ -712,6 +761,13 @@ Default and recommended: 'auto' (best mode for the model).
|
|
712
761
|
Optional telemetry configuration (experimental).
|
713
762
|
*/
|
714
763
|
experimental_telemetry?: TelemetrySettings;
|
764
|
+
/**
|
765
|
+
* Internal. For test use only. May change without notice.
|
766
|
+
*/
|
767
|
+
_internal?: {
|
768
|
+
generateId?: () => string;
|
769
|
+
currentDate?: () => Date;
|
770
|
+
};
|
715
771
|
}): Promise<GenerateObjectResult<Array<ELEMENT>>>;
|
716
772
|
/**
|
717
773
|
Generate JSON with any schema for a given prompt using a language model.
|
@@ -735,6 +791,13 @@ The mode to use for object generation. Must be "json" for no-schema output.
|
|
735
791
|
Optional telemetry configuration (experimental).
|
736
792
|
*/
|
737
793
|
experimental_telemetry?: TelemetrySettings;
|
794
|
+
/**
|
795
|
+
* Internal. For test use only. May change without notice.
|
796
|
+
*/
|
797
|
+
_internal?: {
|
798
|
+
generateId?: () => string;
|
799
|
+
currentDate?: () => Date;
|
800
|
+
};
|
738
801
|
}): Promise<GenerateObjectResult<JSONValue>>;
|
739
802
|
/**
|
740
803
|
* @deprecated Use `generateObject` instead.
|
@@ -754,7 +817,7 @@ interface StreamObjectResult<PARTIAL, RESULT, ELEMENT_STREAM> {
|
|
754
817
|
/**
|
755
818
|
The token usage of the generated response. Resolved when the response is finished.
|
756
819
|
*/
|
757
|
-
readonly usage: Promise<
|
820
|
+
readonly usage: Promise<LanguageModelUsage$1>;
|
758
821
|
/**
|
759
822
|
Additional provider-specific metadata. They are passed through
|
760
823
|
from the provider to the AI SDK and enable provider-specific
|
@@ -762,7 +825,9 @@ interface StreamObjectResult<PARTIAL, RESULT, ELEMENT_STREAM> {
|
|
762
825
|
*/
|
763
826
|
readonly experimental_providerMetadata: Promise<ProviderMetadata | undefined>;
|
764
827
|
/**
|
765
|
-
|
828
|
+
Optional raw response data.
|
829
|
+
|
830
|
+
@deprecated Use `response` instead.
|
766
831
|
*/
|
767
832
|
readonly rawResponse?: {
|
768
833
|
/**
|
@@ -771,6 +836,10 @@ interface StreamObjectResult<PARTIAL, RESULT, ELEMENT_STREAM> {
|
|
771
836
|
headers?: Record<string, string>;
|
772
837
|
};
|
773
838
|
/**
|
839
|
+
Additional response information.
|
840
|
+
*/
|
841
|
+
readonly response: Promise<LanguageModelResponseMetadataWithHeaders>;
|
842
|
+
/**
|
774
843
|
The generated object (typed according to the schema). Resolved when the response is finished.
|
775
844
|
*/
|
776
845
|
readonly object: Promise<RESULT>;
|
@@ -817,33 +886,29 @@ interface StreamObjectResult<PARTIAL, RESULT, ELEMENT_STREAM> {
|
|
817
886
|
*/
|
818
887
|
toTextStreamResponse(init?: ResponseInit): Response;
|
819
888
|
}
|
820
|
-
type
|
889
|
+
type ObjectStreamPart<PARTIAL> = {
|
890
|
+
type: 'object';
|
891
|
+
object: PARTIAL;
|
892
|
+
} | {
|
893
|
+
type: 'text-delta';
|
894
|
+
textDelta: string;
|
895
|
+
} | {
|
821
896
|
type: 'error';
|
822
897
|
error: unknown;
|
823
898
|
} | {
|
824
899
|
type: 'finish';
|
825
900
|
finishReason: FinishReason;
|
826
901
|
logprobs?: LogProbs;
|
827
|
-
usage:
|
828
|
-
|
829
|
-
completionTokens: number;
|
830
|
-
totalTokens: number;
|
831
|
-
};
|
902
|
+
usage: LanguageModelUsage$1;
|
903
|
+
response: LanguageModelResponseMetadata;
|
832
904
|
providerMetadata?: ProviderMetadata;
|
833
905
|
};
|
834
|
-
type ObjectStreamPart<PARTIAL> = ObjectStreamInputPart | {
|
835
|
-
type: 'object';
|
836
|
-
object: PARTIAL;
|
837
|
-
} | {
|
838
|
-
type: 'text-delta';
|
839
|
-
textDelta: string;
|
840
|
-
};
|
841
906
|
|
842
907
|
type OnFinishCallback<RESULT> = (event: {
|
843
908
|
/**
|
844
909
|
The token usage of the generated response.
|
845
910
|
*/
|
846
|
-
usage:
|
911
|
+
usage: LanguageModelUsage$1;
|
847
912
|
/**
|
848
913
|
The generated object. Can be undefined if the final object does not match the schema.
|
849
914
|
*/
|
@@ -854,14 +919,20 @@ type OnFinishCallback<RESULT> = (event: {
|
|
854
919
|
error: unknown | undefined;
|
855
920
|
/**
|
856
921
|
Optional raw response data.
|
857
|
-
|
922
|
+
|
923
|
+
@deprecated Use `response` instead.
|
924
|
+
*/
|
858
925
|
rawResponse?: {
|
859
926
|
/**
|
860
927
|
Response headers.
|
861
|
-
|
928
|
+
*/
|
862
929
|
headers?: Record<string, string>;
|
863
930
|
};
|
864
931
|
/**
|
932
|
+
Response metadata.
|
933
|
+
*/
|
934
|
+
response: LanguageModelResponseMetadataWithHeaders;
|
935
|
+
/**
|
865
936
|
Warnings from the model provider (e.g. unsupported settings).
|
866
937
|
*/
|
867
938
|
warnings?: CallWarning[];
|
@@ -928,6 +999,8 @@ Callback that is called when the LLM response and the final object validation ar
|
|
928
999
|
* Internal. For test use only. May change without notice.
|
929
1000
|
*/
|
930
1001
|
_internal?: {
|
1002
|
+
generateId?: () => string;
|
1003
|
+
currentDate?: () => Date;
|
931
1004
|
now?: () => number;
|
932
1005
|
};
|
933
1006
|
}): Promise<StreamObjectResult<DeepPartial<OBJECT>, OBJECT, never>>;
|
@@ -987,6 +1060,8 @@ Callback that is called when the LLM response and the final object validation ar
|
|
987
1060
|
* Internal. For test use only. May change without notice.
|
988
1061
|
*/
|
989
1062
|
_internal?: {
|
1063
|
+
generateId?: () => string;
|
1064
|
+
currentDate?: () => Date;
|
990
1065
|
now?: () => number;
|
991
1066
|
};
|
992
1067
|
}): Promise<StreamObjectResult<Array<ELEMENT>, Array<ELEMENT>, AsyncIterableStream<ELEMENT>>>;
|
@@ -1020,6 +1095,8 @@ Callback that is called when the LLM response and the final object validation ar
|
|
1020
1095
|
* Internal. For test use only. May change without notice.
|
1021
1096
|
*/
|
1022
1097
|
_internal?: {
|
1098
|
+
generateId?: () => string;
|
1099
|
+
currentDate?: () => Date;
|
1023
1100
|
now?: () => number;
|
1024
1101
|
};
|
1025
1102
|
}): Promise<StreamObjectResult<JSONValue, JSONValue, never>>;
|
@@ -1179,7 +1256,7 @@ interface GenerateTextResult<TOOLS extends Record<string, CoreTool>> {
|
|
1179
1256
|
/**
|
1180
1257
|
The token usage of the generated text.
|
1181
1258
|
*/
|
1182
|
-
readonly usage:
|
1259
|
+
readonly usage: LanguageModelUsage$1;
|
1183
1260
|
/**
|
1184
1261
|
Warnings from the model provider (e.g. unsupported settings)
|
1185
1262
|
*/
|
@@ -1216,7 +1293,7 @@ interface GenerateTextResult<TOOLS extends Record<string, CoreTool>> {
|
|
1216
1293
|
/**
|
1217
1294
|
The token usage of the generated text.
|
1218
1295
|
*/
|
1219
|
-
readonly usage:
|
1296
|
+
readonly usage: LanguageModelUsage$1;
|
1220
1297
|
/**
|
1221
1298
|
Warnings from the model provider (e.g. unsupported settings)
|
1222
1299
|
*/
|
@@ -1227,17 +1304,25 @@ interface GenerateTextResult<TOOLS extends Record<string, CoreTool>> {
|
|
1227
1304
|
*/
|
1228
1305
|
readonly logprobs: LogProbs | undefined;
|
1229
1306
|
/**
|
1230
|
-
|
1231
|
-
|
1307
|
+
Optional raw response data.
|
1308
|
+
|
1309
|
+
@deprecated Use `response.headers` instead.
|
1310
|
+
*/
|
1232
1311
|
readonly rawResponse?: {
|
1233
1312
|
/**
|
1234
|
-
|
1235
|
-
|
1313
|
+
Response headers.
|
1314
|
+
*/
|
1236
1315
|
readonly headers?: Record<string, string>;
|
1237
1316
|
};
|
1317
|
+
/**
|
1318
|
+
Additional response information.
|
1319
|
+
*/
|
1320
|
+
readonly response: LanguageModelResponseMetadataWithHeaders;
|
1238
1321
|
}>;
|
1239
1322
|
/**
|
1240
|
-
|
1323
|
+
Optional raw response data.
|
1324
|
+
|
1325
|
+
@deprecated Use `response.headers` instead.
|
1241
1326
|
*/
|
1242
1327
|
readonly rawResponse?: {
|
1243
1328
|
/**
|
@@ -1246,8 +1331,14 @@ interface GenerateTextResult<TOOLS extends Record<string, CoreTool>> {
|
|
1246
1331
|
readonly headers?: Record<string, string>;
|
1247
1332
|
};
|
1248
1333
|
/**
|
1249
|
-
|
1250
|
-
|
1334
|
+
Additional response information.
|
1335
|
+
*/
|
1336
|
+
readonly response: LanguageModelResponseMetadataWithHeaders;
|
1337
|
+
/**
|
1338
|
+
Logprobs for the completion.
|
1339
|
+
`undefined` if the mode does not support logprobs or if was not enabled.
|
1340
|
+
|
1341
|
+
@deprecated Will become a provider extension in the future.
|
1251
1342
|
*/
|
1252
1343
|
readonly logprobs: LogProbs | undefined;
|
1253
1344
|
/**
|
@@ -1302,7 +1393,7 @@ If set and supported by the model, calls will generate deterministic results.
|
|
1302
1393
|
@returns
|
1303
1394
|
A result object that contains the generated text, the results of the tool calls, and additional information.
|
1304
1395
|
*/
|
1305
|
-
declare function generateText<TOOLS extends Record<string, CoreTool>>({ model, tools, toolChoice, system, prompt, messages, maxRetries, abortSignal, headers, maxAutomaticRoundtrips, maxToolRoundtrips, experimental_telemetry: telemetry, ...settings }: CallSettings & Prompt & {
|
1396
|
+
declare function generateText<TOOLS extends Record<string, CoreTool>>({ model, tools, toolChoice, system, prompt, messages, maxRetries, abortSignal, headers, maxAutomaticRoundtrips, maxToolRoundtrips, experimental_telemetry: telemetry, _internal: { generateId, currentDate, }, ...settings }: CallSettings & Prompt & {
|
1306
1397
|
/**
|
1307
1398
|
The language model to use.
|
1308
1399
|
*/
|
@@ -1336,7 +1427,18 @@ By default, it's set to 0, which will disable the feature.
|
|
1336
1427
|
* Optional telemetry configuration (experimental).
|
1337
1428
|
*/
|
1338
1429
|
experimental_telemetry?: TelemetrySettings;
|
1430
|
+
/**
|
1431
|
+
* Internal. For test use only. May change without notice.
|
1432
|
+
*/
|
1433
|
+
_internal?: {
|
1434
|
+
generateId?: () => string;
|
1435
|
+
currentDate?: () => Date;
|
1436
|
+
};
|
1339
1437
|
}): Promise<GenerateTextResult<TOOLS>>;
|
1438
|
+
/**
|
1439
|
+
* @deprecated Use `generateText` instead.
|
1440
|
+
*/
|
1441
|
+
declare const experimental_generateText: typeof generateText;
|
1340
1442
|
|
1341
1443
|
/**
|
1342
1444
|
A result object for accessing different stream types and additional information.
|
@@ -1352,7 +1454,7 @@ interface StreamTextResult<TOOLS extends Record<string, CoreTool>> {
|
|
1352
1454
|
|
1353
1455
|
Resolved when the response is finished.
|
1354
1456
|
*/
|
1355
|
-
readonly usage: Promise<
|
1457
|
+
readonly usage: Promise<LanguageModelUsage$1>;
|
1356
1458
|
/**
|
1357
1459
|
The reason why the generation finished. Taken from the last roundtrip.
|
1358
1460
|
|
@@ -1385,6 +1487,8 @@ interface StreamTextResult<TOOLS extends Record<string, CoreTool>> {
|
|
1385
1487
|
readonly toolResults: Promise<ToToolResult<TOOLS>[]>;
|
1386
1488
|
/**
|
1387
1489
|
Optional raw response data.
|
1490
|
+
|
1491
|
+
@deprecated Use `response` instead.
|
1388
1492
|
*/
|
1389
1493
|
readonly rawResponse?: {
|
1390
1494
|
/**
|
@@ -1393,6 +1497,10 @@ interface StreamTextResult<TOOLS extends Record<string, CoreTool>> {
|
|
1393
1497
|
headers?: Record<string, string>;
|
1394
1498
|
};
|
1395
1499
|
/**
|
1500
|
+
Additional response information.
|
1501
|
+
*/
|
1502
|
+
readonly response: Promise<LanguageModelResponseMetadataWithHeaders>;
|
1503
|
+
/**
|
1396
1504
|
A text stream that returns only the generated text deltas. You can use it
|
1397
1505
|
as either an AsyncIterable or a ReadableStream. When an error occurs, the
|
1398
1506
|
stream will throw the error.
|
@@ -1513,21 +1621,15 @@ type TextStreamPart<TOOLS extends Record<string, CoreTool>> = {
|
|
1513
1621
|
type: 'roundtrip-finish';
|
1514
1622
|
finishReason: FinishReason;
|
1515
1623
|
logprobs?: LogProbs;
|
1516
|
-
usage:
|
1517
|
-
|
1518
|
-
completionTokens: number;
|
1519
|
-
totalTokens: number;
|
1520
|
-
};
|
1624
|
+
usage: LanguageModelUsage$1;
|
1625
|
+
response: LanguageModelResponseMetadata;
|
1521
1626
|
experimental_providerMetadata?: ProviderMetadata;
|
1522
1627
|
} | {
|
1523
1628
|
type: 'finish';
|
1524
1629
|
finishReason: FinishReason;
|
1525
1630
|
logprobs?: LogProbs;
|
1526
|
-
usage:
|
1527
|
-
|
1528
|
-
completionTokens: number;
|
1529
|
-
totalTokens: number;
|
1530
|
-
};
|
1631
|
+
usage: LanguageModelUsage$1;
|
1632
|
+
response: LanguageModelResponseMetadata;
|
1531
1633
|
experimental_providerMetadata?: ProviderMetadata;
|
1532
1634
|
} | {
|
1533
1635
|
type: 'error';
|
@@ -1580,7 +1682,7 @@ If set and supported by the model, calls will generate deterministic results.
|
|
1580
1682
|
@return
|
1581
1683
|
A result object for accessing different stream types and additional information.
|
1582
1684
|
*/
|
1583
|
-
declare function streamText<TOOLS extends Record<string, CoreTool>>({ model, tools, toolChoice, system, prompt, messages, maxRetries, abortSignal, headers, maxToolRoundtrips, experimental_telemetry: telemetry, experimental_toolCallStreaming: toolCallStreaming, onChunk, onFinish, _internal: { now }, ...settings }: CallSettings & Prompt & {
|
1685
|
+
declare function streamText<TOOLS extends Record<string, CoreTool>>({ model, tools, toolChoice, system, prompt, messages, maxRetries, abortSignal, headers, maxToolRoundtrips, experimental_telemetry: telemetry, experimental_toolCallStreaming: toolCallStreaming, onChunk, onFinish, _internal: { now, generateId, currentDate, }, ...settings }: CallSettings & Prompt & {
|
1584
1686
|
/**
|
1585
1687
|
The language model to use.
|
1586
1688
|
*/
|
@@ -1634,7 +1736,7 @@ Callback that is called when the LLM response and all request tool executions
|
|
1634
1736
|
/**
|
1635
1737
|
The token usage of the generated response.
|
1636
1738
|
*/
|
1637
|
-
usage:
|
1739
|
+
usage: LanguageModelUsage$1;
|
1638
1740
|
/**
|
1639
1741
|
The full text that has been generated.
|
1640
1742
|
*/
|
@@ -1649,6 +1751,8 @@ Callback that is called when the LLM response and all request tool executions
|
|
1649
1751
|
toolResults?: ToToolResult<TOOLS>[];
|
1650
1752
|
/**
|
1651
1753
|
Optional raw response data.
|
1754
|
+
|
1755
|
+
@deprecated Use `response` instead.
|
1652
1756
|
*/
|
1653
1757
|
rawResponse?: {
|
1654
1758
|
/**
|
@@ -1657,6 +1761,10 @@ Callback that is called when the LLM response and all request tool executions
|
|
1657
1761
|
headers?: Record<string, string>;
|
1658
1762
|
};
|
1659
1763
|
/**
|
1764
|
+
Response metadata.
|
1765
|
+
*/
|
1766
|
+
response: LanguageModelResponseMetadataWithHeaders;
|
1767
|
+
/**
|
1660
1768
|
Warnings from the model provider (e.g. unsupported settings).
|
1661
1769
|
*/
|
1662
1770
|
warnings?: CallWarning[];
|
@@ -1672,8 +1780,80 @@ Callback that is called when the LLM response and all request tool executions
|
|
1672
1780
|
*/
|
1673
1781
|
_internal?: {
|
1674
1782
|
now?: () => number;
|
1783
|
+
generateId?: () => string;
|
1784
|
+
currentDate?: () => Date;
|
1675
1785
|
};
|
1676
1786
|
}): Promise<StreamTextResult<TOOLS>>;
|
1787
|
+
/**
|
1788
|
+
* @deprecated Use `streamText` instead.
|
1789
|
+
*/
|
1790
|
+
declare const experimental_streamText: typeof streamText;
|
1791
|
+
|
1792
|
+
/**
|
1793
|
+
* Experimental middleware for LanguageModelV1.
|
1794
|
+
* This type defines the structure for middleware that can be used to modify
|
1795
|
+
* the behavior of LanguageModelV1 operations.
|
1796
|
+
*/
|
1797
|
+
type Experimental_LanguageModelV1Middleware = {
|
1798
|
+
/**
|
1799
|
+
* Transforms the parameters before they are passed to the language model.
|
1800
|
+
* @param options - Object containing the type of operation and the parameters.
|
1801
|
+
* @param options.type - The type of operation ('generate' or 'stream').
|
1802
|
+
* @param options.params - The original parameters for the language model call.
|
1803
|
+
* @returns A promise that resolves to the transformed parameters.
|
1804
|
+
*/
|
1805
|
+
transformParams?: (options: {
|
1806
|
+
type: 'generate' | 'stream';
|
1807
|
+
params: LanguageModelV1CallOptions;
|
1808
|
+
}) => PromiseLike<LanguageModelV1CallOptions>;
|
1809
|
+
/**
|
1810
|
+
* Wraps the generate operation of the language model.
|
1811
|
+
* @param options - Object containing the generate function, parameters, and model.
|
1812
|
+
* @param options.doGenerate - The original generate function.
|
1813
|
+
* @param options.params - The parameters for the generate call. If the
|
1814
|
+
* `transformParams` middleware is used, this will be the transformed parameters.
|
1815
|
+
* @param options.model - The language model instance.
|
1816
|
+
* @returns A promise that resolves to the result of the generate operation.
|
1817
|
+
*/
|
1818
|
+
wrapGenerate?: (options: {
|
1819
|
+
doGenerate: () => ReturnType<LanguageModelV1['doGenerate']>;
|
1820
|
+
params: LanguageModelV1CallOptions;
|
1821
|
+
model: LanguageModelV1;
|
1822
|
+
}) => Promise<Awaited<ReturnType<LanguageModelV1['doGenerate']>>>;
|
1823
|
+
/**
|
1824
|
+
* Wraps the stream operation of the language model.
|
1825
|
+
* @param options - Object containing the stream function, parameters, and model.
|
1826
|
+
* @param options.doStream - The original stream function.
|
1827
|
+
* @param options.params - The parameters for the stream call. If the
|
1828
|
+
* `transformParams` middleware is used, this will be the transformed parameters.
|
1829
|
+
* @param options.model - The language model instance.
|
1830
|
+
* @returns A promise that resolves to the result of the stream operation.
|
1831
|
+
*/
|
1832
|
+
wrapStream?: (options: {
|
1833
|
+
doStream: () => ReturnType<LanguageModelV1['doStream']>;
|
1834
|
+
params: LanguageModelV1CallOptions;
|
1835
|
+
model: LanguageModelV1;
|
1836
|
+
}) => PromiseLike<Awaited<ReturnType<LanguageModelV1['doStream']>>>;
|
1837
|
+
};
|
1838
|
+
|
1839
|
+
/**
|
1840
|
+
* Wraps a LanguageModelV1 instance with middleware functionality.
|
1841
|
+
* This function allows you to apply middleware to transform parameters,
|
1842
|
+
* wrap generate operations, and wrap stream operations of a language model.
|
1843
|
+
*
|
1844
|
+
* @param options - Configuration options for wrapping the language model.
|
1845
|
+
* @param options.model - The original LanguageModelV1 instance to be wrapped.
|
1846
|
+
* @param options.middleware - The middleware to be applied to the language model.
|
1847
|
+
* @param options.modelId - Optional custom model ID to override the original model's ID.
|
1848
|
+
* @param options.providerId - Optional custom provider ID to override the original model's provider.
|
1849
|
+
* @returns A new LanguageModelV1 instance with middleware applied.
|
1850
|
+
*/
|
1851
|
+
declare const experimental_wrapLanguageModel: ({ model, middleware: { transformParams, wrapGenerate, wrapStream }, modelId, providerId, }: {
|
1852
|
+
model: LanguageModelV1;
|
1853
|
+
middleware: Experimental_LanguageModelV1Middleware;
|
1854
|
+
modelId?: string;
|
1855
|
+
providerId?: string;
|
1856
|
+
}) => LanguageModelV1;
|
1677
1857
|
|
1678
1858
|
/**
|
1679
1859
|
* Creates a custom provider with specified language models, text embedding models, and an optional fallback provider.
|
@@ -2781,10 +2961,10 @@ declare class StreamingTextResponse extends Response {
|
|
2781
2961
|
constructor(res: ReadableStream, init?: ResponseInit, data?: StreamData);
|
2782
2962
|
}
|
2783
2963
|
|
2784
|
-
declare const generateId: (
|
2964
|
+
declare const generateId: () => string;
|
2785
2965
|
/**
|
2786
2966
|
@deprecated Use `generateId` instead.
|
2787
2967
|
*/
|
2788
|
-
declare const nanoid: (
|
2968
|
+
declare const nanoid: () => string;
|
2789
2969
|
|
2790
|
-
export { AIStream, AIStreamCallbacksAndOptions, AIStreamParser, AIStreamParserOptions, AWSBedrockAnthropicMessagesStream, AWSBedrockAnthropicStream, AWSBedrockCohereStream, AWSBedrockLlama2Stream, AWSBedrockStream, AnthropicStream, AssistantContent, AssistantResponse, CallWarning, CohereStream, CompletionTokenUsage, CompletionUsage, CoreAssistantMessage, CoreMessage, CoreSystemMessage, CoreTool, CoreToolChoice, CoreToolMessage, CoreUserMessage, DataContent, DownloadError, EmbedManyResult, EmbedResult, Embedding, EmbeddingModel, EmbeddingTokenUsage, ExperimentalAssistantMessage, ExperimentalMessage, ExperimentalTool, ExperimentalToolMessage, ExperimentalUserMessage, FinishReason, FunctionCallPayload, GenerateObjectResult, GenerateTextResult, GoogleGenerativeAIStream, HuggingFaceStream, ImagePart, InkeepAIStreamCallbacksAndOptions, InkeepChatResultCallbacks, InkeepOnFinalMetadata, InkeepStream, InvalidArgumentError, InvalidDataContentError, InvalidMessageRoleError, InvalidToolArgumentsError, langchainAdapter as LangChainAdapter, LangChainStream, LanguageModel, LogProbs, MessageConversionError, MistralStream, NoObjectGeneratedError, NoSuchProviderError, NoSuchToolError,
|
2970
|
+
export { AIStream, AIStreamCallbacksAndOptions, AIStreamParser, AIStreamParserOptions, AWSBedrockAnthropicMessagesStream, AWSBedrockAnthropicStream, AWSBedrockCohereStream, AWSBedrockLlama2Stream, AWSBedrockStream, AnthropicStream, AssistantContent, AssistantResponse, CallWarning, CohereStream, CompletionTokenUsage, CompletionUsage, CoreAssistantMessage, CoreMessage, CoreSystemMessage, CoreTool, CoreToolChoice, CoreToolMessage, CoreUserMessage, DataContent, DownloadError, EmbedManyResult, EmbedResult, Embedding, EmbeddingModel, EmbeddingModelUsage, EmbeddingTokenUsage, ExperimentalAssistantMessage, ExperimentalMessage, ExperimentalTool, ExperimentalToolMessage, ExperimentalUserMessage, Experimental_LanguageModelV1Middleware, FinishReason, FunctionCallPayload, GenerateObjectResult, GenerateTextResult, GoogleGenerativeAIStream, HuggingFaceStream, ImagePart, InkeepAIStreamCallbacksAndOptions, InkeepChatResultCallbacks, InkeepOnFinalMetadata, InkeepStream, InvalidArgumentError, InvalidDataContentError, InvalidMessageRoleError, InvalidToolArgumentsError, langchainAdapter as LangChainAdapter, LangChainStream, LanguageModel, LanguageModelResponseMetadata, LanguageModelResponseMetadataWithHeaders, LanguageModelUsage, LogProbs, MessageConversionError, MistralStream, NoObjectGeneratedError, NoSuchProviderError, NoSuchToolError, ObjectStreamPart, OpenAIStream, OpenAIStreamCallbacks, Provider, ProviderMetadata, ReplicateStream, RetryError, StreamData, StreamObjectResult, StreamTextResult, StreamingTextResponse, TextPart$1 as TextPart, TextStreamPart, TokenUsage, ToolCallPart, ToolCallPayload, ToolContent, ToolResultPart, UserContent, convertToCoreMessages, cosineSimilarity, createCallbacksTransformer, createEventStreamTransformer, createStreamDataTransformer, embed, embedMany, experimental_AssistantResponse, experimental_ModelRegistry, experimental_Provider, experimental_ProviderRegistry, experimental_StreamData, experimental_createModelRegistry, experimental_createProviderRegistry, experimental_customProvider, experimental_generateObject, experimental_generateText, experimental_streamObject, experimental_streamText, experimental_wrapLanguageModel, generateId, generateObject, generateText, nanoid, readableFromAsyncIterable, streamObject, streamText, streamToResponse, tool, trimStartOfStreamHelper };
|